17 research outputs found
Generative Temporal Models with Spatial Memory for Partially Observed Environments
In model-based reinforcement learning, generative and temporal models of
environments can be leveraged to boost agent performance, either by tuning the
agent's representations during training or via use as part of an explicit
planning mechanism. However, their application in practice has been limited to
simplistic environments, due to the difficulty of training such models in
larger, potentially partially-observed and 3D environments. In this work we
introduce a novel action-conditioned generative model of such challenging
environments. The model features a non-parametric spatial memory system in
which we store learned, disentangled representations of the environment.
Low-dimensional spatial updates are computed using a state-space model that
makes use of knowledge on the prior dynamics of the moving agent, and
high-dimensional visual observations are modelled with a Variational
Auto-Encoder. The result is a scalable architecture capable of performing
coherent predictions over hundreds of time steps across a range of partially
observed 2D and 3D environments.Comment: ICML 201
Leveraging the Exact Likelihood of Deep Latent Variable Models
Deep latent variable models (DLVMs) combine the approximation abilities of
deep neural networks and the statistical foundations of generative models.
Variational methods are commonly used for inference; however, the exact
likelihood of these models has been largely overlooked. The purpose of this
work is to study the general properties of this quantity and to show how they
can be leveraged in practice. We focus on important inferential problems that
rely on the likelihood: estimation and missing data imputation. First, we
investigate maximum likelihood estimation for DLVMs: in particular, we show
that most unconstrained models used for continuous data have an unbounded
likelihood function. This problematic behaviour is demonstrated to be a source
of mode collapse. We also show how to ensure the existence of maximum
likelihood estimates, and draw useful connections with nonparametric mixture
models. Finally, we describe an algorithm for missing data imputation using the
exact conditional likelihood of a deep latent variable model. On several data
sets, our algorithm consistently and significantly outperforms the usual
imputation scheme used for DLVMs
Memory-Based Learning of Latent Structures for Generative Adversarial Networks
ํ์๋
ผ๋ฌธ (์์ฌ)-- ์์ธ๋ํ๊ต ๋ํ์ : ๊ณต๊ณผ๋ํ ์ปดํจํฐ๊ณตํ๋ถ, 2019. 2. ๊น๊ฑดํฌ.๋ณธ ์ฐ๊ตฌ๋ Generative Adversarial Network (GAN) ๋ชจ๋ธ์ ํ์ต ๊ณผ์ ์์ ๋ฐ์ํ๋ ๋๊ฐ์ง ๋ฌธ์ ์ ์ ํด๊ฒฐํ๋ ๋ฐฉ์์ ์ ์ํ์๋ค. ๋จผ์ , ์ผ๋ฐ์ ์ธ GAN ๋ชจ๋ธ์ ์ฌ์ง๊ณผ ๊ฐ์ ๋ณต์กํ ํ๋ฅ ๋ณ์์ ๋ถํฌ๋ฅผ ๋ชจ๋ธ๋งํ ๋ ์ ์ฌ๋ณ์์ ์ฌ์ ํ๋ฅ ๋ถํฌ๋ก ํ์ค์ ๊ท๋ถํฌ๋ฅผ ์ฌ์ฉํ๋ค. ๊ทธ๋ฌ๋ ์ด๋ฐ ์ฐ์์ ์ธ ์ ์ฌ๋ณ์๋ฅผ ์ฌ์ฉํ ๊ฒฝ์ฐ ์๋ก ๋ค๋ฅธ ๋ฐ์ดํฐ ์ํ๊ฐ์ ๊ตฌ์กฐ์ ๋ถ์ฐ์์ฑ์ ๋ฐ์ํ๊ธฐ ์ด๋ ต๋ค. ๋ ๋ค๋ฅธ ๋ฌธ์ ์ ์ผ๋ก, GAN ๋ชจ๋ธ์์ ํ๋ณ์๋ ํ์ต ๊ณผ์ ์์ ๊ณผ๊ฑฐ์ ์์ฑ์ ๋ชจ๋ธ์ด ์์ฑํ๋ ๋ฐ์ดํฐ ์ํ์ ๋ํ ์ ๋ณด๋ฅผ ๋ง๊ฐํ๋ฉฐ, ์ด๋ก์ธํด ํ์ต ๊ณผ์ ์ด ๋ถ์์ ํด์ง๋ค. ์ด ๋๊ฐ์ง ๋ฌธ์ ์ ์ ์์ฑ์๊ฐ ํ๋ณ์๊ฐ ๊ณต์ ํ๋ memory network๋ฅผ ๋์์ ํ์ตํจ์ผ๋ก์จ ํฌ๊ฒ ์ํํ ์ ์๋ค. ์์ฑ์๊ฐ ํ์ต ๋ฐ์ดํฐ์ ๋ด์ฌ๋ ๊ตฐ์ง์ ๋ถํฌ๋ฅผ ํ์ตํ๋ค๋ฉด ์ด๋ฅผ ํตํด ๊ตฌ์กฐ์ ๋ถ์ฐ์์ฑ์ผ๋ก ์ธํ ์ฑ๋ฅ ํ๋ฝ์ ํผํ ์ ์์ผ๋ฉฐ, ํ๋ณ์๊ฐ ์ฃผ์ด์ง ์
๋ ฅ ๋ฐ์ดํฐ์ ๋ํ ํ๋ณ์ ํ ๋ ํ์ต ์ ๊ณผ์ ์ ๊ฑธ์ณ ์์ฑ์๊ฐ ์์ฑํ๋ ๋ฐ์ดํฐ ์ํ๋ค๋ก๋ถํฐ ํ์ต๋ ๊ตฐ์ง ๋ถํฌ๋ฅผ ์ฐธ์กฐํ๋ค๋ฉด ๋ง๊ฐ ๋ฌธ์ ๋ก ์ธํ ์ํฅ์ ๋ ๋ฐ๊ฒ ๋๋ค. ๋ณธ ์ฐ๊ตฌ์์ ์ ์ํ memoryGAN ๋ชจ๋ธ์ ๋น์ง๋ํ์ต์ ํตํด ๋ฐ์ดํฐ์ ๋ด์ฌ๋ ๊ตฐ์ง์ ๋ถํฌ๋ฅผ ํ์ตํ์ฌ ๊ตฌ์กฐ์ ๋ถ์ฐ์์ฑ ๋ฌธ์ ์ ๋ง๊ฐ ๋ฌธ์ ๋ฅผ ์ํํ๋ฉฐ, ๋๋ถ๋ถ์ GAN ๋ชจ๋ธ์ ์ ์ฉํ ์ ์๋ค. Fashion-MNIST, CelebA, CIFAR10, ๊ทธ๋ฆฌ๊ณ Chairs ๋ฐ์ดํฐ์
์ ๋ํ ์ฑ๋ฅ ํ๊ฐ ๋ฐ ์๊ฐํ ์คํ์ ํตํด memoryGAN์ด ํ๋ฅ ๋ก ์ ์ผ๋ก ํด์ ๊ฐ๋ฅํ ๋ชจ๋ธ์ด๋ฉฐ, ๋์ ์์ค์ ์ฌ์ง ์ํ์ ์์ฑํ๋ค๋ ๊ฒ์ ๋ณด์๋ค. ํนํ memoryGAN์ ๊ฐ์ ๋ ์ต์ ํ ๋ฐฉ๋ฒ์ด๋ Weaker divergence๋ฅผ ๋์
ํ์ง ์๊ณ ๋ CIFAR10 ๋ฐ์ดํฐ์
์์ Inception Score๋ฅผ ๊ธฐ์ค์ผ๋ก ๋น์ง๋ํ์ต ๋ฐฉ์์ GAN ๋ชจ๋ธ ์ค ๋์ ์ฑ๋ฅ์ ๋ฌ์ฑํ๋ค.We propose an approach to address two issues that commonly occur during training of unsupervised GANs. First, since GANs use only a continuous latent distribution to embed multiple classes or clusters of data, they often do not correctly handle the structural discontinuity between disparate classes in a latent space. Second, discriminators of GANs easily forget about past generated samples by generators, incurring instability during adversarial training. We argue that these two infamous problems of unsupervised GAN training can be largely alleviated by a learnable memory network to which both generators and discriminators can access.
Generators can effectively learn representation of training samples to understand underlying cluster distributions of data, which ease the structure discontinuity problem. At the same time, discriminators can better memorize clusters of previously generated samples, which mitigate the forgetting problem. We propose a novel end-to-end GAN model named memoryGAN, which involves a memory network that is unsupervisedly trainable and integrable to many existing GAN models. With evaluations on multiple datasets such as Fashion-MNIST, CelebA, CIFAR10, and Chairs, we show that our model is probabilistically interpretable, and generates realistic image samples of high visual fidelity. The memoryGAN also achieves the state-of-the-art inception scores over unsupervised GAN models on the CIFAR10 dataset, without any optimization tricks and weaker divergences.Introduction
Related Works
The MemoryGAN
Experiments
ConclusionMaste
Learning to Learn Variational Semantic Memory
In this paper, we introduce variational semantic memory into meta-learning to
acquire long-term knowledge for few-shot learning. The variational semantic
memory accrues and stores semantic information for the probabilistic inference
of class prototypes in a hierarchical Bayesian framework. The semantic memory
is grown from scratch and gradually consolidated by absorbing information from
tasks it experiences. By doing so, it is able to accumulate long-term, general
knowledge that enables it to learn new concepts of objects. We formulate memory
recall as the variational inference of a latent memory variable from addressed
contents, which offers a principled way to adapt the knowledge to individual
tasks. Our variational semantic memory, as a new long-term memory module,
confers principled recall and update mechanisms that enable semantic
information to be efficiently accrued and adapted for few-shot learning.
Experiments demonstrate that the probabilistic modelling of prototypes achieves
a more informative representation of object classes compared to deterministic
vectors. The consistent new state-of-the-art performance on four benchmarks
shows the benefit of variational semantic memory in boosting few-shot
recognition.Comment: accepted to NeurIPS 2020; code is available in
https://github.com/YDU-uva/VS