153,122 research outputs found
Deep Generative Models for Library Augmentation in Multiple Endmember Spectral Mixture Analysis
Multiple Endmember Spectral Mixture Analysis (MESMA) is one of the leading
approaches to perform spectral unmixing (SU) considering variability of the
endmembers (EMs). It represents each EM in the image using libraries of
spectral signatures acquired a priori. However, existing spectral libraries are
often small and unable to properly capture the variability of each EM in
practical scenes, which compromises the performance of MESMA. In this paper, we
propose a library augmentation strategy to increase the diversity of existing
spectral libraries, thus improving their ability to represent the materials in
real images. First, we leverage the power of deep generative models to learn
the statistical distribution of the EMs based on the spectral signatures
available in the existing libraries. Afterwards, new samples can be drawn from
the learned EM distributions and used to augment the spectral libraries,
improving the overall quality of the SU process. Experimental results using
synthetic and real data attest the superior performance of the proposed method
even under library mismatch conditions
Multi-Source Neural Variational Inference
Learning from multiple sources of information is an important problem in
machine-learning research. The key challenges are learning representations and
formulating inference methods that take into account the complementarity and
redundancy of various information sources. In this paper we formulate a
variational autoencoder based multi-source learning framework in which each
encoder is conditioned on a different information source. This allows us to
relate the sources via the shared latent variables by computing divergence
measures between individual source's posterior approximations. We explore a
variety of options to learn these encoders and to integrate the beliefs they
compute into a consistent posterior approximation. We visualise learned beliefs
on a toy dataset and evaluate our methods for learning shared representations
and structured output prediction, showing trade-offs of learning separate
encoders for each information source. Furthermore, we demonstrate how conflict
detection and redundancy can increase robustness of inference in a multi-source
setting.Comment: AAAI 2019, Association for the Advancement of Artificial Intelligence
(AAAI) 201
Generative Adversarial Networks (GANs): Challenges, Solutions, and Future Directions
Generative Adversarial Networks (GANs) is a novel class of deep generative
models which has recently gained significant attention. GANs learns complex and
high-dimensional distributions implicitly over images, audio, and data.
However, there exists major challenges in training of GANs, i.e., mode
collapse, non-convergence and instability, due to inappropriate design of
network architecture, use of objective function and selection of optimization
algorithm. Recently, to address these challenges, several solutions for better
design and optimization of GANs have been investigated based on techniques of
re-engineered network architectures, new objective functions and alternative
optimization algorithms. To the best of our knowledge, there is no existing
survey that has particularly focused on broad and systematic developments of
these solutions. In this study, we perform a comprehensive survey of the
advancements in GANs design and optimization solutions proposed to handle GANs
challenges. We first identify key research issues within each design and
optimization technique and then propose a new taxonomy to structure solutions
by key research issues. In accordance with the taxonomy, we provide a detailed
discussion on different GANs variants proposed within each solution and their
relationships. Finally, based on the insights gained, we present the promising
research directions in this rapidly growing field.Comment: 42 pages, Figure 13, Table
- …