310,475 research outputs found
Continual Learning in Practice
This paper describes a reference architecture for self-maintaining systems
that can learn continually, as data arrives. In environments where data
evolves, we need architectures that manage Machine Learning (ML) models in
production, adapt to shifting data distributions, cope with outliers, retrain
when necessary, and adapt to new tasks. This represents continual AutoML or
Automatically Adaptive Machine Learning. We describe the challenges and
proposes a reference architecture.Comment: Presented at the NeurIPS 2018 workshop on Continual Learning
https://sites.google.com/view/continual2018/hom
Differentially Private Continual Learning
Catastrophic forgetting can be a significant problem for institutions that
must delete historic data for privacy reasons. For example, hospitals might not
be able to retain patient data permanently. But neural networks trained on
recent data alone will tend to forget lessons learned on old data. We present a
differentially private continual learning framework based on variational
inference. We estimate the likelihood of past data given the current model
using differentially private generative models of old datasets.Comment: Presented at the Privacy in Machine Learning and AI workshop at ICML
201
Continual Learning Through Synaptic Intelligence
While deep learning has led to remarkable advances across diverse
applications, it struggles in domains where the data distribution changes over
the course of learning. In stark contrast, biological neural networks
continually adapt to changing domains, possibly by leveraging complex molecular
machinery to solve many tasks simultaneously. In this study, we introduce
intelligent synapses that bring some of this biological complexity into
artificial neural networks. Each synapse accumulates task relevant information
over time, and exploits this information to rapidly store new memories without
forgetting old ones. We evaluate our approach on continual learning of
classification tasks, and show that it dramatically reduces forgetting while
maintaining computational efficiency.Comment: ICML 201
Meta-Learning Representations for Continual Learning
A continual learning agent should be able to build on top of existing
knowledge to learn on new data quickly while minimizing forgetting. Current
intelligent systems based on neural network function approximators arguably do
the opposite---they are highly prone to forgetting and rarely trained to
facilitate future learning. One reason for this poor behavior is that they
learn from a representation that is not explicitly trained for these two goals.
In this paper, we propose OML, an objective that directly minimizes
catastrophic interference by learning representations that accelerate future
learning and are robust to forgetting under online updates in continual
learning. We show that it is possible to learn naturally sparse representations
that are more effective for online updating. Moreover, our algorithm is
complementary to existing continual learning strategies, such as MER and GEM.
Finally, we demonstrate that a basic online updating strategy on
representations learned by OML is competitive with rehearsal based methods for
continual learning. We release an implementation of our method at
https://github.com/khurramjaved96/mrcl .Comment: Accepted at NeurIPS19, 15 pages, 10 figures, open-source,
representation learning, continual learning, online learnin
Experience Replay for Continual Learning
Continual learning is the problem of learning new tasks or knowledge while
protecting old knowledge and ideally generalizing from old experience to learn
new tasks faster. Neural networks trained by stochastic gradient descent often
degrade on old tasks when trained successively on new tasks with different data
distributions. This phenomenon, referred to as catastrophic forgetting, is
considered a major hurdle to learning with non-stationary data or sequences of
new tasks, and prevents networks from continually accumulating knowledge and
skills. We examine this issue in the context of reinforcement learning, in a
setting where an agent is exposed to tasks in a sequence. Unlike most other
work, we do not provide an explicit indication to the model of task boundaries,
which is the most general circumstance for a learning agent exposed to
continuous experience. While various methods to counteract catastrophic
forgetting have recently been proposed, we explore a straightforward, general,
and seemingly overlooked solution - that of using experience replay buffers for
all past events - with a mixture of on- and off-policy learning, leveraging
behavioral cloning. We show that this strategy can still learn new tasks
quickly yet can substantially reduce catastrophic forgetting in both Atari and
DMLab domains, even matching the performance of methods that require task
identities. When buffer storage is constrained, we confirm that a simple
mechanism for randomly discarding data allows a limited size buffer to perform
almost as well as an unbounded one.Comment: NeurIPS 201
Bayesian Optimized Continual Learning with Attention Mechanism
Though neural networks have achieved much progress in various applications,
it is still highly challenging for them to learn from a continuous stream of
tasks without forgetting. Continual learning, a new learning paradigm, aims to
solve this issue. In this work, we propose a new model for continual learning,
called Bayesian Optimized Continual Learning with Attention Mechanism (BOCL)
that dynamically expands the network capacity upon the arrival of new tasks by
Bayesian optimization and selectively utilizes previous knowledge (e.g. feature
maps of previous tasks) via attention mechanism. Our experiments on variants of
MNIST and CIFAR-100 demonstrate that our methods outperform the
state-of-the-art in preventing catastrophic forgetting and fitting new tasks
better.Comment: 8 page
Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference
Lack of performance when it comes to continual learning over non-stationary
distributions of data remains a major challenge in scaling neural network
learning to more human realistic settings. In this work we propose a new
conceptualization of the continual learning problem in terms of a temporally
symmetric trade-off between transfer and interference that can be optimized by
enforcing gradient alignment across examples. We then propose a new algorithm,
Meta-Experience Replay (MER), that directly exploits this view by combining
experience replay with optimization based meta-learning. This method learns
parameters that make interference based on future gradients less likely and
transfer based on future gradients more likely. We conduct experiments across
continual lifelong supervised learning benchmarks and non-stationary
reinforcement learning environments demonstrating that our approach
consistently outperforms recently proposed baselines for continual learning.
Our experiments show that the gap between the performance of MER and baseline
algorithms grows both as the environment gets more non-stationary and as the
fraction of the total experiences stored gets smaller.Comment: ICLR 201
Facilitating Bayesian Continual Learning by Natural Gradients and Stein Gradients
Continual learning aims to enable machine learning models to learn a general
solution space for past and future tasks in a sequential manner. Conventional
models tend to forget the knowledge of previous tasks while learning a new
task, a phenomenon known as catastrophic forgetting. When using Bayesian models
in continual learning, knowledge from previous tasks can be retained in two
ways: 1). posterior distributions over the parameters, containing the knowledge
gained from inference in previous tasks, which then serve as the priors for the
following task; 2). coresets, containing knowledge of data distributions of
previous tasks. Here, we show that Bayesian continual learning can be
facilitated in terms of these two means through the use of natural gradients
and Stein gradients respectively
Learn to Grow: A Continual Structure Learning Framework for Overcoming Catastrophic Forgetting
Addressing catastrophic forgetting is one of the key challenges in continual
learning where machine learning systems are trained with sequential or
streaming tasks. Despite recent remarkable progress in state-of-the-art deep
learning, deep neural networks (DNNs) are still plagued with the catastrophic
forgetting problem. This paper presents a conceptually simple yet general and
effective framework for handling catastrophic forgetting in continual learning
with DNNs. The proposed method consists of two components: a neural structure
optimization component and a parameter learning and/or fine-tuning component.
By separating the explicit neural structure learning and the parameter
estimation, not only is the proposed method capable of evolving neural
structures in an intuitively meaningful way, but also shows strong capabilities
of alleviating catastrophic forgetting in experiments. Furthermore, the
proposed method outperforms all other baselines on the permuted MNIST dataset,
the split CIFAR100 dataset and the Visual Domain Decathlon dataset in continual
learning setting
Few-Shot Self Reminder to Overcome Catastrophic Forgetting
Deep neural networks are known to suffer the catastrophic forgetting problem,
where they tend to forget the knowledge from the previous tasks when
sequentially learning new tasks. Such failure hinders the application of deep
learning based vision system in continual learning settings. In this work, we
present a simple yet surprisingly effective way of preventing catastrophic
forgetting. Our method, called Few-shot Self Reminder (FSR), regularizes the
neural net from changing its learned behaviour by performing logit matching on
selected samples kept in episodic memory from the old tasks. Surprisingly, this
simplistic approach only requires to retrain a small amount of data in order to
outperform previous methods in knowledge retention. We demonstrate the
superiority of our method to the previous ones in two different continual
learning settings on popular benchmarks, as well as a new continual learning
problem where tasks are designed to be more dissimilar
- …
