834 research outputs found
Parametric Regression on the Grassmannian
We address the problem of fitting parametric curves on the Grassmann manifold
for the purpose of intrinsic parametric regression. As customary in the
literature, we start from the energy minimization formulation of linear
least-squares in Euclidean spaces and generalize this concept to general
nonflat Riemannian manifolds, following an optimal-control point of view. We
then specialize this idea to the Grassmann manifold and demonstrate that it
yields a simple, extensible and easy-to-implement solution to the parametric
regression problem. In fact, it allows us to extend the basic geodesic model to
(1) a time-warped variant and (2) cubic splines. We demonstrate the utility of
the proposed solution on different vision problems, such as shape regression as
a function of age, traffic-speed estimation and crowd-counting from
surveillance video clips. Most notably, these problems can be conveniently
solved within the same framework without any specifically-tailored steps along
the processing pipeline.Comment: 14 pages, 11 figure
Stochastic Development Regression on Non-Linear Manifolds
We introduce a regression model for data on non-linear manifolds. The model
describes the relation between a set of manifold valued observations, such as
shapes of anatomical objects, and Euclidean explanatory variables. The approach
is based on stochastic development of Euclidean diffusion processes to the
manifold. Defining the data distribution as the transition distribution of the
mapped stochastic process, parameters of the model, the non-linear analogue of
design matrix and intercept, are found via maximum likelihood. The model is
intrinsically related to the geometry encoded in the connection of the
manifold. We propose an estimation procedure which applies the Laplace
approximation of the likelihood function. A simulation study of the performance
of the model is performed and the model is applied to a real dataset of Corpus
Callosum shapes
Recommended from our members
Statistical Learning in Wasserstein Space
We seek a generalization of regression and principle component analysis (PCA) in a metric space where data points are distributions metrized by the Wasserstein metric. We recast these analyses as multimarginal optimal transport problems. The particular formulation allows efficient computation, ensures existence of optimal solutions, and admits a probabilistic interpretation over the space of paths (line segments). Application of the theory to the interpolation of empirical distributions, images, power spectra, as well as assessing uncertainty in experimental designs, is envisioned
Principal arc analysis on direct product manifolds
We propose a new approach to analyze data that naturally lie on manifolds. We
focus on a special class of manifolds, called direct product manifolds, whose
intrinsic dimension could be very high. Our method finds a low-dimensional
representation of the manifold that can be used to find and visualize the
principal modes of variation of the data, as Principal Component Analysis (PCA)
does in linear spaces. The proposed method improves upon earlier manifold
extensions of PCA by more concisely capturing important nonlinear modes. For
the special case of data on a sphere, variation following nongeodesic arcs is
captured in a single mode, compared to the two modes needed by previous
methods. Several computational and statistical challenges are resolved. The
development on spheres forms the basis of principal arc analysis on more
complicated manifolds. The benefits of the method are illustrated by a data
example using medial representations in image analysis.Comment: Published in at http://dx.doi.org/10.1214/10-AOAS370 the Annals of
Applied Statistics (http://www.imstat.org/aoas/) by the Institute of
Mathematical Statistics (http://www.imstat.org
A variational model for data fitting on manifolds by minimizing the acceleration of a B\'ezier curve
We derive a variational model to fit a composite B\'ezier curve to a set of
data points on a Riemannian manifold. The resulting curve is obtained in such a
way that its mean squared acceleration is minimal in addition to remaining
close the data points. We approximate the acceleration by discretizing the
squared second order derivative along the curve. We derive a closed-form,
numerically stable and efficient algorithm to compute the gradient of a
B\'ezier curve on manifolds with respect to its control points, expressed as a
concatenation of so-called adjoint Jacobi fields. Several examples illustrate
the capabilites and validity of this approach both for interpolation and
approximation. The examples also illustrate that the approach outperforms
previous works tackling this problem
Rank-preserving geometric means of positive semi-definite matrices
The generalization of the geometric mean of positive scalars to positive
definite matrices has attracted considerable attention since the seminal work
of Ando. The paper generalizes this framework of matrix means by proposing the
definition of a rank-preserving mean for two or an arbitrary number of positive
semi-definite matrices of fixed rank. The proposed mean is shown to be
geometric in that it satisfies all the expected properties of a rank-preserving
geometric mean. The work is motivated by operations on low-rank approximations
of positive definite matrices in high-dimensional spaces.Comment: To appear in Linear Algebra and its Application
- …