Approximation of High-Dimensional Rank One Tensors

Markus Bachmayr, Wolfgang Dahmen, Ronald DeVore, Lars Grasedyck

Research output: Contribution to journalArticlepeer-review

15 Scopus citations


Many real world problems are high-dimensional in that their solution is a function which depends on many variables or parameters. This presents a computational challenge since traditional numerical techniques are built on model classes for functions based solely on smoothness. It is known that the approximation of smoothness classes of functions suffers from the so-called 'curse of dimensionality'. Avoiding this curse requires new model classes for real world functions that match applications. This has led to the introduction of notions such as sparsity, variable reduction, and reduced modeling. One theme that is particularly common is to assume a tensor structure for the target function. This paper investigates how well a rank one function f(x 1,...,x d)=f 1(x 1)⋯f d(x d), defined on Ω=[0,1]d can be captured through point queries. It is shown that such a rank one function with component functions f j in W∞ r([0,1]) can be captured (in L ∞) to accuracy O(C(d,r)N -r) from N well-chosen point evaluations. The constant C(d,r) scales like d dr. The queries in our algorithms have two ingredients, a set of points built on the results from discrepancy theory and a second adaptive set of queries dependent on the information drawn from the first set. Under the assumption that a point z∈Ω with nonvanishing f(z) is known, the accuracy improves to O(dN -r). © 2013 Springer Science+Business Media New York.
Original languageEnglish (US)
Pages (from-to)385-395
Number of pages11
JournalConstructive Approximation
Issue number2
StatePublished - Nov 12 2013
Externally publishedYes


Dive into the research topics of 'Approximation of High-Dimensional Rank One Tensors'. Together they form a unique fingerprint.

Cite this