2,964 research outputs found
A literature survey of low-rank tensor approximation techniques
During the last years, low-rank tensor approximation has been established as
a new tool in scientific computing to address large-scale linear and
multilinear algebra problems, which would be intractable by classical
techniques. This survey attempts to give a literature overview of current
developments in this area, with an emphasis on function-related tensors
Kernels for Vector-Valued Functions: a Review
Kernel methods are among the most popular techniques in machine learning.
From a frequentist/discriminative perspective they play a central role in
regularization theory as they provide a natural choice for the hypotheses space
and the regularization functional through the notion of reproducing kernel
Hilbert spaces. From a Bayesian/generative perspective they are the key in the
context of Gaussian processes, where the kernel function is also known as the
covariance function. Traditionally, kernel methods have been used in supervised
learning problem with scalar outputs and indeed there has been a considerable
amount of work devoted to designing and learning kernels. More recently there
has been an increasing interest in methods that deal with multiple outputs,
motivated partly by frameworks like multitask learning. In this paper, we
review different methods to design or learn valid kernel functions for multiple
outputs, paying particular attention to the connection between probabilistic
and functional methods
Kernel Mean Shrinkage Estimators
A mean function in a reproducing kernel Hilbert space (RKHS), or a kernel
mean, is central to kernel methods in that it is used by many classical
algorithms such as kernel principal component analysis, and it also forms the
core inference step of modern kernel methods that rely on embedding probability
distributions in RKHSs. Given a finite sample, an empirical average has been
used commonly as a standard estimator of the true kernel mean. Despite a
widespread use of this estimator, we show that it can be improved thanks to the
well-known Stein phenomenon. We propose a new family of estimators called
kernel mean shrinkage estimators (KMSEs), which benefit from both theoretical
justifications and good empirical performance. The results demonstrate that the
proposed estimators outperform the standard one, especially in a "large d,
small n" paradigm.Comment: 41 page
Using separable non-negative matrix factorization techniques for the analysis of time-resolved Raman spectra
The key challenge of time-resolved Raman spectroscopy is the identification
of the constituent species and the analysis of the kinetics of the underlying
reaction network. In this work we present an integral approach that allows for
determining both the component spectra and the rate constants simultaneously
from a series of vibrational spectra. It is based on an algorithm for
non-negative matrix factorization which is applied to the experimental data set
following a few pre-processing steps. As a prerequisite for physically
unambiguous solutions, each component spectrum must include one vibrational
band that does not significantly interfere with vibrational bands of other
species. The approach is applied to synthetic "experimental" spectra derived
from model systems comprising a set of species with component spectra differing
with respect to their degree of spectral interferences and signal-to-noise
ratios. In each case, the species involved are connected via monomolecular
reaction pathways. The potential and limitations of the approach for recovering
the respective rate constants and component spectra are discussed
- …