821 research outputs found
A continuous analogue of the tensor-train decomposition
We develop new approximation algorithms and data structures for representing
and computing with multivariate functions using the functional tensor-train
(FT), a continuous extension of the tensor-train (TT) decomposition. The FT
represents functions using a tensor-train ansatz by replacing the
three-dimensional TT cores with univariate matrix-valued functions. The main
contribution of this paper is a framework to compute the FT that employs
adaptive approximations of univariate fibers, and that is not tied to any
tensorized discretization. The algorithm can be coupled with any univariate
linear or nonlinear approximation procedure. We demonstrate that this approach
can generate multivariate function approximations that are several orders of
magnitude more accurate, for the same cost, than those based on the
conventional approach of compressing the coefficient tensor of a tensor-product
basis. Our approach is in the spirit of other continuous computation packages
such as Chebfun, and yields an algorithm which requires the computation of
"continuous" matrix factorizations such as the LU and QR decompositions of
vector-valued functions. To support these developments, we describe continuous
versions of an approximate maximum-volume cross approximation algorithm and of
a rounding algorithm that re-approximates an FT by one of lower ranks. We
demonstrate that our technique improves accuracy and robustness, compared to TT
and quantics-TT approaches with fixed parameterizations, of high-dimensional
integration, differentiation, and approximation of functions with local
features such as discontinuities and other nonlinearities
On Training Efficiency and Computational Costs of a Feed Forward Neural Network: A Review
A comprehensive review on the problem of choosing a suitable activation function for the hidden layer of a feed forward neural network has been widely investigated. Since the nonlinear component of a neural network is the main contributor to the network mapping capabilities, the different choices that may lead to enhanced performances, in terms of training, generalization, or computational costs, are analyzed, both in general-purpose and in embedded computing environments. Finally, a strategy to convert a network configuration between different activation functions without altering the network mapping capabilities will be presented
Mixtures of Spatial Spline Regressions
We present an extension of the functional data analysis framework for
univariate functions to the analysis of surfaces: functions of two variables.
The spatial spline regression (SSR) approach developed can be used to model
surfaces that are sampled over a rectangular domain. Furthermore, combining SSR
with linear mixed effects models (LMM) allows for the analysis of populations
of surfaces, and combining the joint SSR-LMM method with finite mixture models
allows for the analysis of populations of surfaces with sub-family structures.
Through the mixtures of spatial splines regressions (MSSR) approach developed,
we present methodologies for clustering surfaces into sub-families, and for
performing surface-based discriminant analysis. The effectiveness of our
methodologies, as well as the modeling capabilities of the SSR model are
assessed through an application to handwritten character recognition
- …