251 research outputs found
DIMAL: Deep Isometric Manifold Learning Using Sparse Geodesic Sampling
This paper explores a fully unsupervised deep learning approach for computing
distance-preserving maps that generate low-dimensional embeddings for a certain
class of manifolds. We use the Siamese configuration to train a neural network
to solve the problem of least squares multidimensional scaling for generating
maps that approximately preserve geodesic distances. By training with only a
few landmarks, we show a significantly improved local and nonlocal
generalization of the isometric mapping as compared to analogous non-parametric
counterparts. Importantly, the combination of a deep-learning framework with a
multidimensional scaling objective enables a numerical analysis of network
architectures to aid in understanding their representation power. This provides
a geometric perspective to the generalizability of deep learning.Comment: 10 pages, 11 Figure
An efficient high-order algorithm for acoustic scattering from penetrable thin structures in three dimensions
This paper presents a high-order accelerated algorithm for the solution of the integral-equation formulation of volumetric scattering problems. The scheme is particularly well suited to the analysis of “thin” structures as they arise in certain applications (e.g., material coatings); in addition, it is also designed to be used in conjunction with existing low-order FFT-based codes to upgrade their order of accuracy through a suitable treatment of material interfaces. The high-order convergence of the new procedure is attained through a combination of changes of parametric variables (to resolve the singularities of the Green function) and “partitions of unity” (to allow for a simple implementation of spectrally accurate quadratures away from singular points). Accelerated evaluations of the interaction between degrees of freedom, on the other hand, are accomplished by incorporating (two-face) equivalent source approximations on Cartesian grids. A detailed account of the main algorithmic components of the scheme are presented, together with a brief review of the corresponding error and performance analyses which are exemplified with a variety of numerical results
Online Tensor Methods for Learning Latent Variable Models
We introduce an online tensor decomposition based approach for two latent
variable modeling problems namely, (1) community detection, in which we learn
the latent communities that the social actors in social networks belong to, and
(2) topic modeling, in which we infer hidden topics of text articles. We
consider decomposition of moment tensors using stochastic gradient descent. We
conduct optimization of multilinear operations in SGD and avoid directly
forming the tensors, to save computational and storage costs. We present
optimized algorithm in two platforms. Our GPU-based implementation exploits the
parallelism of SIMD architectures to allow for maximum speed-up by a careful
optimization of storage and data transfer, whereas our CPU-based implementation
uses efficient sparse matrix computations and is suitable for large sparse
datasets. For the community detection problem, we demonstrate accuracy and
computational efficiency on Facebook, Yelp and DBLP datasets, and for the topic
modeling problem, we also demonstrate good performance on the New York Times
dataset. We compare our results to the state-of-the-art algorithms such as the
variational method, and report a gain of accuracy and a gain of several orders
of magnitude in the execution time.Comment: JMLR 201
The Bregman Variational Dual-Tree Framework
Graph-based methods provide a powerful tool set for many non-parametric
frameworks in Machine Learning. In general, the memory and computational
complexity of these methods is quadratic in the number of examples in the data
which makes them quickly infeasible for moderate to large scale datasets. A
significant effort to find more efficient solutions to the problem has been
made in the literature. One of the state-of-the-art methods that has been
recently introduced is the Variational Dual-Tree (VDT) framework. Despite some
of its unique features, VDT is currently restricted only to Euclidean spaces
where the Euclidean distance quantifies the similarity. In this paper, we
extend the VDT framework beyond the Euclidean distance to more general Bregman
divergences that include the Euclidean distance as a special case. By
exploiting the properties of the general Bregman divergence, we show how the
new framework can maintain all the pivotal features of the VDT framework and
yet significantly improve its performance in non-Euclidean domains. We apply
the proposed framework to different text categorization problems and
demonstrate its benefits over the original VDT.Comment: Appears in Proceedings of the Twenty-Ninth Conference on Uncertainty
in Artificial Intelligence (UAI2013
Preconditioning Kernel Matrices
The computational and storage complexity of kernel machines presents the
primary barrier to their scaling to large, modern, datasets. A common way to
tackle the scalability issue is to use the conjugate gradient algorithm, which
relieves the constraints on both storage (the kernel matrix need not be stored)
and computation (both stochastic gradients and parallelization can be used).
Even so, conjugate gradient is not without its own issues: the conditioning of
kernel matrices is often such that conjugate gradients will have poor
convergence in practice. Preconditioning is a common approach to alleviating
this issue. Here we propose preconditioned conjugate gradients for kernel
machines, and develop a broad range of preconditioners particularly useful for
kernel matrices. We describe a scalable approach to both solving kernel
machines and learning their hyperparameters. We show this approach is exact in
the limit of iterations and outperforms state-of-the-art approximations for a
given computational budget
Perspectives on Beam-Shaping Optimization for Thermal-Noise Reduction in Advanced Gravitational-Wave Interferometric Detectors: Bounds, Profiles, and Critical Parameters
Suitable shaping (in particular, flattening and broadening) of the laser beam
has recently been proposed as an effective device to reduce internal (mirror)
thermal noise in advanced gravitational wave interferometric detectors. Based
on some recently published analytic approximations (valid in the
infinite-test-mass limit) for the Brownian and thermoelastic mirror noises in
the presence of arbitrary-shaped beams, this paper addresses certain
preliminary issues related to the optimal beam-shaping problem. In particular,
with specific reference to the Laser Interferometer Gravitational-wave
Observatory (LIGO) experiment, absolute and realistic lower-bounds for the
various thermal noise constituents are obtained and compared with the current
status (Gaussian beams) and trends ("mesa" beams), indicating fairly ample
margins for further reduction. In this framework, the effective dimension of
the related optimization problem, and its relationship to the critical design
parameters are identified, physical-feasibility and model-consistency issues
are considered, and possible additional requirements and/or prior information
exploitable to drive the subsequent optimization process are highlighted.Comment: 12 pages, 9 figures, 2 table
- …