779 research outputs found
Spectral tensor-train decomposition
The accurate approximation of high-dimensional functions is an essential task
in uncertainty quantification and many other fields. We propose a new function
approximation scheme based on a spectral extension of the tensor-train (TT)
decomposition. We first define a functional version of the TT decomposition and
analyze its properties. We obtain results on the convergence of the
decomposition, revealing links between the regularity of the function, the
dimension of the input space, and the TT ranks. We also show that the
regularity of the target function is preserved by the univariate functions
(i.e., the "cores") comprising the functional TT decomposition. This result
motivates an approximation scheme employing polynomial approximations of the
cores. For functions with appropriate regularity, the resulting
\textit{spectral tensor-train decomposition} combines the favorable
dimension-scaling of the TT decomposition with the spectral convergence rate of
polynomial approximations, yielding efficient and accurate surrogates for
high-dimensional functions. To construct these decompositions, we use the
sampling algorithm \texttt{TT-DMRG-cross} to obtain the TT decomposition of
tensors resulting from suitable discretizations of the target function. We
assess the performance of the method on a range of numerical examples: a
modifed set of Genz functions with dimension up to , and functions with
mixed Fourier modes or with local features. We observe significant improvements
in performance over an anisotropic adaptive Smolyak approach. The method is
also used to approximate the solution of an elliptic PDE with random input
data. The open source software and examples presented in this work are
available online.Comment: 33 pages, 19 figure
An Incremental Tensor Train Decomposition Algorithm
We present a new algorithm for incrementally updating the tensor-train
decomposition of a stream of tensor data. This new algorithm, called the
tensor-train incremental core expansion (TT-ICE) improves upon the current
state-of-the-art algorithms for compressing in tensor-train format by
developing a new adaptive approach that incurs significantly slower rank growth
and guarantees compression accuracy. This capability is achieved by limiting
the number of new vectors appended to the TT-cores of an existing accumulation
tensor after each data increment. These vectors represent directions orthogonal
to the span of existing cores and are limited to those needed to represent a
newly arrived tensor to a target accuracy. We provide two versions of the
algorithm: TT-ICE and TT-ICE accelerated with heuristics (TT-ICE). We
provide a proof of correctness for TT-ICE and empirically demonstrate the
performance of the algorithms in compressing large-scale video and scientific
simulation datasets. Compared to existing approaches that also use rank
adaptation, TT-ICE achieves 57 higher compression and up to 95%
reduction in computational time.Comment: 22 pages, 7 figures, for the python code of TT-ICE and TT-ICE
algorithms see https://github.com/dorukaks/TT-IC
- …