67,200 research outputs found
Image Restoration using Autoencoding Priors
We propose to leverage denoising autoencoder networks as priors to address
image restoration problems. We build on the key observation that the output of
an optimal denoising autoencoder is a local mean of the true data density, and
the autoencoder error (the difference between the output and input of the
trained autoencoder) is a mean shift vector. We use the magnitude of this mean
shift vector, that is, the distance to the local mean, as the negative log
likelihood of our natural image prior. For image restoration, we maximize the
likelihood using gradient descent by backpropagating the autoencoder error. A
key advantage of our approach is that we do not need to train separate networks
for different image restoration tasks, such as non-blind deconvolution with
different kernels, or super-resolution at different magnification factors. We
demonstrate state of the art results for non-blind deconvolution and
super-resolution using the same autoencoding prior
Learning compressed representations of blood samples time series with missing data
Clinical measurements collected over time are naturally represented as
multivariate time series (MTS), which often contain missing data. An
autoencoder can learn low dimensional vectorial representations of MTS that
preserve important data characteristics, but cannot deal explicitly with
missing data. In this work, we propose a new framework that combines an
autoencoder with the Time series Cluster Kernel (TCK), a kernel that accounts
for missingness patterns in MTS. Via kernel alignment, we incorporate TCK in
the autoencoder to improve the learned representations in presence of missing
data. We consider a classification problem of MTS with missing values,
representing blood samples of patients with surgical site infection. With our
approach, rather than with a standard autoencoder, we learn representations in
low dimensions that can be classified better
Denoising without access to clean data using a partitioned autoencoder
Training a denoising autoencoder neural network requires access to truly
clean data, a requirement which is often impractical. To remedy this, we
introduce a method to train an autoencoder using only noisy data, having
examples with and without the signal class of interest. The autoencoder learns
a partitioned representation of signal and noise, learning to reconstruct each
separately. We illustrate the method by denoising birdsong audio (available
abundantly in uncontrolled noisy datasets) using a convolutional autoencoder
Adversarial Autoencoders
In this paper, we propose the "adversarial autoencoder" (AAE), which is a
probabilistic autoencoder that uses the recently proposed generative
adversarial networks (GAN) to perform variational inference by matching the
aggregated posterior of the hidden code vector of the autoencoder with an
arbitrary prior distribution. Matching the aggregated posterior to the prior
ensures that generating from any part of prior space results in meaningful
samples. As a result, the decoder of the adversarial autoencoder learns a deep
generative model that maps the imposed prior to the data distribution. We show
how the adversarial autoencoder can be used in applications such as
semi-supervised classification, disentangling style and content of images,
unsupervised clustering, dimensionality reduction and data visualization. We
performed experiments on MNIST, Street View House Numbers and Toronto Face
datasets and show that adversarial autoencoders achieve competitive results in
generative modeling and semi-supervised classification tasks
Gaussian AutoEncoder
Generative AutoEncoders require a chosen probability distribution in latent
space, usually multivariate Gaussian. The original Variational AutoEncoder
(VAE) uses randomness in encoder - causing problematic distortion, and overlaps
in latent space for distinct inputs. It turned out unnecessary: we can instead
use deterministic encoder with additional regularizer to ensure that sample
distribution in latent space is close to the required. The original approach
(WAE) uses Wasserstein metric, what required comparing with random sample and
using an arbitrarily chosen kernel. Later CWAE finally derived a non-random
analytic formula by averaging distance of Gaussian-smoothened sample over
all 1D projections. However, these arbitrarily chosen regularizers do not lead
to Gaussian distribution.
This article proposes approach for regularizers directly optimizing agreement
between empirical distribution function and its desired CDF for chosen
properties, for example radii and distances for Gaussian distribution, or
coordinate-wise, to directly attract this distribution in latent space of
AutoEncoder. We can also attract different distributions with this general
approach, for example latent space uniform distribution on hypercube
or torus would allow for data compression without entropy coding, increased
density near codewords would optimize for the required quantization.Comment: 6 pages, 2 figure
Autoencoder Trees
We discuss an autoencoder model in which the encoding and decoding functions
are implemented by decision trees. We use the soft decision tree where internal
nodes realize soft multivariate splits given by a gating function and the
overall output is the average of all leaves weighted by the gating values on
their path. The encoder tree takes the input and generates a lower dimensional
representation in the leaves and the decoder tree takes this and reconstructs
the original input. Exploiting the continuity of the trees, autoencoder trees
are trained with stochastic gradient descent. On handwritten digit and news
data, we see that the autoencoder trees yield good reconstruction error
compared to traditional autoencoder perceptrons. We also see that the
autoencoder tree captures hierarchical representations at different
granularities of the data on its different levels and the leaves capture the
localities in the input space.Comment: 9 page
Toroidal AutoEncoder
Enforcing distributions of latent variables in neural networks is an active
subject. It is vital in all kinds of generative models, where we want to be
able to interpolate between points in the latent space, or sample from it.
Modern generative AutoEncoders (AE) like WAE, SWAE, CWAE add a regularizer to
the standard (deterministic) AE, which allows to enforce Gaussian distribution
in the latent space. Enforcing different distributions, especially
topologically nontrivial, might bring some new interesting possibilities, but
this subject seems unexplored so far.
This article proposes a new approach to enforce uniform distribution on
d-dimensional torus. We introduce a circular spring loss, which enforces
minibatch points to be equally spaced and satisfy cyclic boundary conditions.
As example of application we propose multiple-path morphing. Minimal distance
geodesic between two points in uniform distribution on latent space of angles
becomes a line, however, torus topology allows us to choose such lines in
alternative ways, going through different edges of .
Further applications to explore can be for example trying to learn real-life
topologically nontrivial spaces of features, like rotations to automatically
recognize 2D rotation of an object in picture by training on relative angles,
or even 3D rotations by additionally using spherical features - this way
morphing should be close to object rotation.Comment: 5 pages, 5 figure
Residual Codean Autoencoder for Facial Attribute Analysis
Facial attributes can provide rich ancillary information which can be
utilized for different applications such as targeted marketing, human computer
interaction, and law enforcement. This research focuses on facial attribute
prediction using a novel deep learning formulation, termed as R-Codean
autoencoder. The paper first presents Cosine similarity based loss function in
an autoencoder which is then incorporated into the Euclidean distance based
autoencoder to formulate R-Codean. The proposed loss function thus aims to
incorporate both magnitude and direction of image vectors during feature
learning. Further, inspired by the utility of shortcut connections in deep
models to facilitate learning of optimal parameters, without incurring the
problem of vanishing gradient, the proposed formulation is extended to
incorporate shortcut connections in the architecture. The proposed R-Codean
autoencoder is utilized in facial attribute prediction framework which
incorporates patch-based weighting mechanism for assigning higher weights to
relevant patches for each attribute. The experimental results on publicly
available CelebA and LFWA datasets demonstrate the efficacy of the proposed
approach in addressing this challenging problem.Comment: Accepted in Pattern Recognition Letter
A Generalized Data Representation and Training-Performance Analysis for Deep Learning-Based Communications Systems
Deep learning (DL)-based autoencoder is a potential architecture to implement
end-to-end communication systems. In this letter, we first give a brief
introduction to the autoencoder-represented communication system. Then, we
propose a novel generalized data representation (GDR) aiming to improve the
data rate of DL-based communication systems. Finally, simulation results show
that the proposed GDR scheme has lower training complexity, comparable block
error rate performance and higher channel capacity than the conventional
one-hot vector scheme. Furthermore, we investigate the effect of
signal-to-noise ratio (SNR) in DL-based communication systems and prove that
training at a high SNR could produce a good training performance for
autoencoder
Deep Learning of Part-based Representation of Data Using Sparse Autoencoders with Nonnegativity Constraints
We demonstrate a new deep learning autoencoder network, trained by a
nonnegativity constraint algorithm (NCAE), that learns features which show
part-based representation of data. The learning algorithm is based on
constraining negative weights. The performance of the algorithm is assessed
based on decomposing data into parts and its prediction performance is tested
on three standard image data sets and one text dataset. The results indicate
that the nonnegativity constraint forces the autoencoder to learn features that
amount to a part-based representation of data, while improving sparsity and
reconstruction quality in comparison with the traditional sparse autoencoder
and Nonnegative Matrix Factorization. It is also shown that this newly acquired
representation improves the prediction performance of a deep neural network.Comment: Accepted for publication in IEEE Transactions of Neural Networks and
Learning System
- …
