8,596 research outputs found
A sparse decomposition of low rank symmetric positive semi-definite matrices
Suppose that is symmetric positive
semidefinite with rank . Our goal is to decompose into
rank-one matrices where the modes
are required to be as sparse as possible. In contrast to eigen decomposition,
these sparse modes are not required to be orthogonal. Such a problem arises in
random field parametrization where is the covariance function and is
intractable to solve in general. In this paper, we partition the indices from 1
to into several patches and propose to quantify the sparseness of a vector
by the number of patches on which it is nonzero, which is called patch-wise
sparseness. Our aim is to find the decomposition which minimizes the total
patch-wise sparseness of the decomposed modes. We propose a
domain-decomposition type method, called intrinsic sparse mode decomposition
(ISMD), which follows the "local-modes-construction + patching-up" procedure.
The key step in the ISMD is to construct local pieces of the intrinsic sparse
modes by a joint diagonalization problem. Thereafter a pivoted Cholesky
decomposition is utilized to glue these local pieces together. Optimal sparse
decomposition, consistency with different domain decomposition and robustness
to small perturbation are proved under the so called regular-sparse assumption
(see Definition 1.2). We provide simulation results to show the efficiency and
robustness of the ISMD. We also compare the ISMD to other existing methods,
e.g., eigen decomposition, pivoted Cholesky decomposition and convex relaxation
of sparse principal component analysis [25] and [40]
Improving Sampling from Generative Autoencoders with Markov Chains
We focus on generative autoencoders, such as variational or adversarial autoencoders, which jointly learn a generative model alongside an inference model. We define generative autoencoders as autoencoders which are trained to softly enforce a prior on the latent distribution learned by the model. However, the model does not necessarily learn to match the prior. We formulate a Markov chain Monte Carlo (MCMC) sampling process, equivalent to iteratively encoding and decoding, which allows us to sample from the learned latent distribution. Using this we can improve the quality of samples drawn from the model, especially when the learned distribution is far from the prior. Using MCMC sampling, we also reveal previously unseen differences between generative autoencoders trained either with or without the denoising criterion
Beating the Perils of Non-Convexity: Guaranteed Training of Neural Networks using Tensor Methods
Training neural networks is a challenging non-convex optimization problem,
and backpropagation or gradient descent can get stuck in spurious local optima.
We propose a novel algorithm based on tensor decomposition for guaranteed
training of two-layer neural networks. We provide risk bounds for our proposed
method, with a polynomial sample complexity in the relevant parameters, such as
input dimension and number of neurons. While learning arbitrary target
functions is NP-hard, we provide transparent conditions on the function and the
input for learnability. Our training method is based on tensor decomposition,
which provably converges to the global optimum, under a set of mild
non-degeneracy conditions. It consists of simple embarrassingly parallel linear
and multi-linear operations, and is competitive with standard stochastic
gradient descent (SGD), in terms of computational complexity. Thus, we propose
a computationally efficient method with guaranteed risk bounds for training
neural networks with one hidden layer.Comment: The tensor decomposition analysis is expanded, and the analysis of
ridge regression is added for recovering the parameters of last layer of
neural networ
- β¦