347,707 research outputs found
Generative-Discriminative Complementary Learning
Majority of state-of-the-art deep learning methods are discriminative
approaches, which model the conditional distribution of labels given inputs
features. The success of such approaches heavily depends on high-quality
labeled instances, which are not easy to obtain, especially as the number of
candidate classes increases. In this paper, we study the complementary learning
problem. Unlike ordinary labels, complementary labels are easy to obtain
because an annotator only needs to provide a yes/no answer to a randomly chosen
candidate class for each instance. We propose a generative-discriminative
complementary learning method that estimates the ordinary labels by modeling
both the conditional (discriminative) and instance (generative) distributions.
Our method, we call Complementary Conditional GAN (CCGAN), improves the
accuracy of predicting ordinary labels and can generate high-quality instances
in spite of weak supervision. In addition to the extensive empirical studies,
we also theoretically show that our model can retrieve the true conditional
distribution from the complementarily-labeled data
Auxiliary Deep Generative Models
Deep generative models parameterized by neural networks have recently
achieved state-of-the-art performance in unsupervised and semi-supervised
learning. We extend deep generative models with auxiliary variables which
improves the variational approximation. The auxiliary variables leave the
generative model unchanged but make the variational distribution more
expressive. Inspired by the structure of the auxiliary variable we also propose
a model with two stochastic layers and skip connections. Our findings suggest
that more expressive and properly specified deep generative models converge
faster with better results. We show state-of-the-art performance within
semi-supervised learning on MNIST, SVHN and NORB datasets.Comment: Proceedings of the 33rd International Conference on Machine Learning,
New York, NY, USA, 2016, JMLR: Workshop and Conference Proceedings volume 48,
Proceedings of the 33rd International Conference on Machine Learning, New
York, NY, USA, 201
Quantum generative adversarial learning
Generative adversarial networks (GANs) represent a powerful tool for
classical machine learning: a generator tries to create statistics for data
that mimics those of a true data set, while a discriminator tries to
discriminate between the true and fake data. The learning process for generator
and discriminator can be thought of as an adversarial game, and under
reasonable assumptions, the game converges to the point where the generator
generates the same statistics as the true data and the discriminator is unable
to discriminate between the true and the generated data. This paper introduces
the notion of quantum generative adversarial networks (QuGANs), where the data
consists either of quantum states, or of classical data, and the generator and
discriminator are equipped with quantum information processors. We show that
the unique fixed point of the quantum adversarial game also occurs when the
generator produces the same statistics as the data. Since quantum systems are
intrinsically probabilistic the proof of the quantum case is different from -
and simpler than - the classical case. We show that when the data consists of
samples of measurements made on high-dimensional spaces, quantum adversarial
networks may exhibit an exponential advantage over classical adversarial
networks.Comment: 5 pages, 1 figur
Learning Generative Models across Incomparable Spaces
Generative Adversarial Networks have shown remarkable success in learning a
distribution that faithfully recovers a reference distribution in its entirety.
However, in some cases, we may want to only learn some aspects (e.g., cluster
or manifold structure), while modifying others (e.g., style, orientation or
dimension). In this work, we propose an approach to learn generative models
across such incomparable spaces, and demonstrate how to steer the learned
distribution towards target properties. A key component of our model is the
Gromov-Wasserstein distance, a notion of discrepancy that compares
distributions relationally rather than absolutely. While this framework
subsumes current generative models in identically reproducing distributions,
its inherent flexibility allows application to tasks in manifold learning,
relational learning and cross-domain learning.Comment: International Conference on Machine Learning (ICML
- …
