15,232 research outputs found
Regression on fixed-rank positive semidefinite matrices: a Riemannian approach
The paper addresses the problem of learning a regression model parameterized
by a fixed-rank positive semidefinite matrix. The focus is on the nonlinear
nature of the search space and on scalability to high-dimensional problems. The
mathematical developments rely on the theory of gradient descent algorithms
adapted to the Riemannian geometry that underlies the set of fixed-rank
positive semidefinite matrices. In contrast with previous contributions in the
literature, no restrictions are imposed on the range space of the learned
matrix. The resulting algorithms maintain a linear complexity in the problem
size and enjoy important invariance properties. We apply the proposed
algorithms to the problem of learning a distance function parameterized by a
positive semidefinite matrix. Good performance is observed on classical
benchmarks
On the optimality of shape and data representation in the spectral domain
A proof of the optimality of the eigenfunctions of the Laplace-Beltrami
operator (LBO) in representing smooth functions on surfaces is provided and
adapted to the field of applied shape and data analysis. It is based on the
Courant-Fischer min-max principle adapted to our case. % The theorem we present
supports the new trend in geometry processing of treating geometric structures
by using their projection onto the leading eigenfunctions of the decomposition
of the LBO. Utilisation of this result can be used for constructing numerically
efficient algorithms to process shapes in their spectrum. We review a couple of
applications as possible practical usage cases of the proposed optimality
criteria. % We refer to a scale invariant metric, which is also invariant to
bending of the manifold. This novel pseudo-metric allows constructing an LBO by
which a scale invariant eigenspace on the surface is defined. We demonstrate
the efficiency of an intermediate metric, defined as an interpolation between
the scale invariant and the regular one, in representing geometric structures
while capturing both coarse and fine details. Next, we review a numerical
acceleration technique for classical scaling, a member of a family of
flattening methods known as multidimensional scaling (MDS). There, the
optimality is exploited to efficiently approximate all geodesic distances
between pairs of points on a given surface, and thereby match and compare
between almost isometric surfaces. Finally, we revisit the classical principal
component analysis (PCA) definition by coupling its variational form with a
Dirichlet energy on the data manifold. By pairing the PCA with the LBO we can
handle cases that go beyond the scope defined by the observation set that is
handled by regular PCA
Spectral Generalized Multi-Dimensional Scaling
Multidimensional scaling (MDS) is a family of methods that embed a given set
of points into a simple, usually flat, domain. The points are assumed to be
sampled from some metric space, and the mapping attempts to preserve the
distances between each pair of points in the set. Distances in the target space
can be computed analytically in this setting. Generalized MDS is an extension
that allows mapping one metric space into another, that is, multidimensional
scaling into target spaces in which distances are evaluated numerically rather
than analytically. Here, we propose an efficient approach for computing such
mappings between surfaces based on their natural spectral decomposition, where
the surfaces are treated as sampled metric-spaces. The resulting spectral-GMDS
procedure enables efficient embedding by implicitly incorporating smoothness of
the mapping into the problem, thereby substantially reducing the complexity
involved in its solution while practically overcoming its non-convex nature.
The method is compared to existing techniques that compute dense correspondence
between shapes. Numerical experiments of the proposed method demonstrate its
efficiency and accuracy compared to state-of-the-art approaches
DIMAL: Deep Isometric Manifold Learning Using Sparse Geodesic Sampling
This paper explores a fully unsupervised deep learning approach for computing
distance-preserving maps that generate low-dimensional embeddings for a certain
class of manifolds. We use the Siamese configuration to train a neural network
to solve the problem of least squares multidimensional scaling for generating
maps that approximately preserve geodesic distances. By training with only a
few landmarks, we show a significantly improved local and nonlocal
generalization of the isometric mapping as compared to analogous non-parametric
counterparts. Importantly, the combination of a deep-learning framework with a
multidimensional scaling objective enables a numerical analysis of network
architectures to aid in understanding their representation power. This provides
a geometric perspective to the generalizability of deep learning.Comment: 10 pages, 11 Figure
Relative Comparison Kernel Learning with Auxiliary Kernels
In this work we consider the problem of learning a positive semidefinite
kernel matrix from relative comparisons of the form: "object A is more similar
to object B than it is to C", where comparisons are given by humans. Existing
solutions to this problem assume many comparisons are provided to learn a high
quality kernel. However, this can be considered unrealistic for many real-world
tasks since relative assessments require human input, which is often costly or
difficult to obtain. Because of this, only a limited number of these
comparisons may be provided. In this work, we explore methods for aiding the
process of learning a kernel with the help of auxiliary kernels built from more
easily extractable information regarding the relationships among objects. We
propose a new kernel learning approach in which the target kernel is defined as
a conic combination of auxiliary kernels and a kernel whose elements are
learned directly. We formulate a convex optimization to solve for this target
kernel that adds only minor overhead to methods that use no auxiliary
information. Empirical results show that in the presence of few training
relative comparisons, our method can learn kernels that generalize to more
out-of-sample comparisons than methods that do not utilize auxiliary
information, as well as similar methods that learn metrics over objects
Construction of embedded fMRI resting state functional connectivity networks using manifold learning
We construct embedded functional connectivity networks (FCN) from benchmark
resting-state functional magnetic resonance imaging (rsfMRI) data acquired from
patients with schizophrenia and healthy controls based on linear and nonlinear
manifold learning algorithms, namely, Multidimensional Scaling (MDS), Isometric
Feature Mapping (ISOMAP) and Diffusion Maps. Furthermore, based on key global
graph-theoretical properties of the embedded FCN, we compare their
classification potential using machine learning techniques. We also assess the
performance of two metrics that are widely used for the construction of FCN
from fMRI, namely the Euclidean distance and the lagged cross-correlation
metric. We show that the FCN constructed with Diffusion Maps and the lagged
cross-correlation metric outperform the other combinations
- …