4,789 research outputs found
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
We propose an algorithm for meta-learning that is model-agnostic, in the
sense that it is compatible with any model trained with gradient descent and
applicable to a variety of different learning problems, including
classification, regression, and reinforcement learning. The goal of
meta-learning is to train a model on a variety of learning tasks, such that it
can solve new learning tasks using only a small number of training samples. In
our approach, the parameters of the model are explicitly trained such that a
small number of gradient steps with a small amount of training data from a new
task will produce good generalization performance on that task. In effect, our
method trains the model to be easy to fine-tune. We demonstrate that this
approach leads to state-of-the-art performance on two few-shot image
classification benchmarks, produces good results on few-shot regression, and
accelerates fine-tuning for policy gradient reinforcement learning with neural
network policies.Comment: ICML 2017. Code at https://github.com/cbfinn/maml, Videos of RL
results at https://sites.google.com/view/maml, Blog post at
http://bair.berkeley.edu/blog/2017/07/18/learning-to-learn
MetaPred: Meta-Learning for Clinical Risk Prediction with Limited Patient Electronic Health Records
In recent years, increasingly augmentation of health data, such as patient
Electronic Health Records (EHR), are becoming readily available. This provides
an unprecedented opportunity for knowledge discovery and data mining algorithms
to dig insights from them, which can, later on, be helpful to the improvement
of the quality of care delivery. Predictive modeling of clinical risk,
including in-hospital mortality, hospital readmission, chronic disease onset,
condition exacerbation, etc., from patient EHR, is one of the health data
analytic problems that attract most of the interests. The reason is not only
because the problem is important in clinical settings, but also there are
challenges working with EHR such as sparsity, irregularity, temporality, etc.
Different from applications in other domains such as computer vision and
natural language processing, the labeled data samples in medicine (patients)
are relatively limited, which creates lots of troubles for effective predictive
model learning, especially for complicated models such as deep learning. In
this paper, we propose MetaPred, a meta-learning for clinical risk prediction
from longitudinal patient EHRs. In particular, in order to predict the target
risk where there are limited data samples, we train a meta-learner from a set
of related risk prediction tasks which learns how a good predictor is learned.
The meta-learned can then be directly used in target risk prediction, and the
limited available samples can be used for further fine-tuning the model
performance. The effectiveness of MetaPred is tested on a real patient EHR
repository from Oregon Health & Science University. We are able to demonstrate
that with CNN and RNN as base predictors, MetaPred can achieve much better
performance for predicting target risk with low resources comparing with the
predictor trained on the limited samples available for this risk
RelationNet2: Deep Comparison Columns for Few-Shot Learning
Few-shot deep learning is a topical challenge area for scaling visual
recognition to open ended growth of unseen new classes with limited labeled
examples. A promising approach is based on metric learning, which trains a deep
embedding to support image similarity matching. Our insight is that effective
general purpose matching requires non-linear comparison of features at multiple
abstraction levels. We thus propose a new deep comparison network comprised of
embedding and relation modules that learn multiple non-linear distance metrics
based on different levels of features simultaneously. Furthermore, to reduce
over-fitting and enable the use of deeper embeddings, we represent images as
distributions rather than vectors via learning parameterized Gaussian noise
regularization. The resulting network achieves excellent performance on both
miniImageNet and tieredImageNet.Comment: 10 pages, 5 figures, Published in IJCNN 202
Learning to Adapt for Stereo
Real world applications of stereo depth estimation require models that are
robust to dynamic variations in the environment. Even though deep learning
based stereo methods are successful, they often fail to generalize to unseen
variations in the environment, making them less suitable for practical
applications such as autonomous driving. In this work, we introduce a
"learning-to-adapt" framework that enables deep stereo methods to continuously
adapt to new target domains in an unsupervised manner. Specifically, our
approach incorporates the adaptation procedure into the learning objective to
obtain a base set of parameters that are better suited for unsupervised online
adaptation. To further improve the quality of the adaptation, we learn a
confidence measure that effectively masks the errors introduced during the
unsupervised adaptation. We evaluate our method on synthetic and real-world
stereo datasets and our experiments evidence that learning-to-adapt is, indeed
beneficial for online adaptation on vastly different domains.Comment: Accepted at CVPR2019. Code available at
https://github.com/CVLAB-Unibo/Learning2AdaptForStere
Meta-Learning of Neural Architectures for Few-Shot Learning
The recent progress in neural architecture search (NAS) has allowed scaling
the automated design of neural architectures to real-world domains, such as
object detection and semantic segmentation. However, one prerequisite for the
application of NAS are large amounts of labeled data and compute resources.
This renders its application challenging in few-shot learning scenarios, where
many related tasks need to be learned, each with limited amounts of data and
compute time. Thus, few-shot learning is typically done with a fixed neural
architecture. To improve upon this, we propose MetaNAS, the first method which
fully integrates NAS with gradient-based meta-learning. MetaNAS optimizes a
meta-architecture along with the meta-weights during meta-training. During
meta-testing, architectures can be adapted to a novel task with a few steps of
the task optimizer, that is: task adaptation becomes computationally cheap and
requires only little data per task. Moreover, MetaNAS is agnostic in that it
can be used with arbitrary model-agnostic meta-learning algorithms and
arbitrary gradient-based NAS methods. %We present encouraging results for
MetaNAS with a combination of DARTS and REPTILE on few-shot classification
benchmarks. Empirical results on standard few-shot classification benchmarks
show that MetaNAS with a combination of DARTS and REPTILE yields
state-of-the-art results
Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning
Although reinforcement learning methods can achieve impressive results in
simulation, the real world presents two major challenges: generating samples is
exceedingly expensive, and unexpected perturbations or unseen situations cause
proficient but specialized policies to fail at test time. Given that it is
impractical to train separate policies to accommodate all situations the agent
may see in the real world, this work proposes to learn how to quickly and
effectively adapt online to new tasks. To enable sample-efficient learning, we
consider learning online adaptation in the context of model-based reinforcement
learning. Our approach uses meta-learning to train a dynamics model prior such
that, when combined with recent data, this prior can be rapidly adapted to the
local context. Our experiments demonstrate online adaptation for continuous
control tasks on both simulated and real-world agents. We first show simulated
agents adapting their behavior online to novel terrains, crippled body parts,
and highly-dynamic environments. We also illustrate the importance of
incorporating online adaptation into autonomous agents that operate in the real
world by applying our method to a real dynamic legged millirobot. We
demonstrate the agent's learned ability to quickly adapt online to a missing
leg, adjust to novel terrains and slopes, account for miscalibration or errors
in pose estimation, and compensate for pulling payloads.Comment: First 2 authors contributed equally. Website:
https://sites.google.com/berkeley.edu/metaadaptivecontro
A Meta-Learning Approach for Custom Model Training
Transfer-learning and meta-learning are two effective methods to apply
knowledge learned from large data sources to new tasks. In few-class, few-shot
target task settings (i.e. when there are only a few classes and training
examples available in the target task), meta-learning approaches that optimize
for future task learning have outperformed the typical transfer approach of
initializing model weights from a pre-trained starting point. But as we
experimentally show, meta-learning algorithms that work well in the few-class
setting do not generalize well in many-shot and many-class cases. In this
paper, we propose a joint training approach that combines both
transfer-learning and meta-learning. Benefiting from the advantages of each,
our method obtains improved generalization performance on unseen target tasks
in both few- and many-class and few- and many-shot scenarios.Comment: AAAI 201
Domain-Invariant Speaker Vector Projection by Model-Agnostic Meta-Learning
Domain generalization remains a critical problem for speaker recognition,
even with the state-of-the-art architectures based on deep neural nets. For
example, a model trained on reading speech may largely fail when applied to
scenarios of singing or movie. In this paper, we propose a domain-invariant
projection to improve the generalizability of speaker vectors. This projection
is a simple neural net and is trained following the Model-Agnostic
Meta-Learning (MAML) principle, for which the objective is to classify speakers
in one domain if it had been updated with speech data in another domain. We
tested the proposed method on CNCeleb, a new dataset consisting of
single-speaker multi-condition (SSMC) data. The results demonstrated that the
MAML-based domain-invariant projection can produce more generalizable speaker
vectors, and effectively improve the performance in unseen domains.Comment: submitted to INTERSPEECH 202
NoRML: No-Reward Meta Learning
Efficiently adapting to new environments and changes in dynamics is critical
for agents to successfully operate in the real world. Reinforcement learning
(RL) based approaches typically rely on external reward feedback for
adaptation. However, in many scenarios this reward signal might not be readily
available for the target task, or the difference between the environments can
be implicit and only observable from the dynamics. To this end, we introduce a
method that allows for self-adaptation of learned policies: No-Reward Meta
Learning (NoRML). NoRML extends Model Agnostic Meta Learning (MAML) for RL and
uses observable dynamics of the environment instead of an explicit reward
function in MAML's finetune step. Our method has a more expressive update step
than MAML, while maintaining MAML's gradient based foundation. Additionally, in
order to allow more targeted exploration, we implement an extension to MAML
that effectively disconnects the meta-policy parameters from the fine-tuned
policies' parameters. We first study our method on a number of synthetic
control problems and then validate our method on common benchmark environments,
showing that NoRML outperforms MAML when the dynamics change between tasks
Unsupervised Meta-Learning For Few-Shot Image Classification
Few-shot or one-shot learning of classifiers requires a significant inductive
bias towards the type of task to be learned. One way to acquire this is by
meta-learning on tasks similar to the target task. In this paper, we propose
UMTRA, an algorithm that performs unsupervised, model-agnostic meta-learning
for classification tasks. The meta-learning step of UMTRA is performed on a
flat collection of unlabeled images. While we assume that these images can be
grouped into a diverse set of classes and are relevant to the target task, no
explicit information about the classes or any labels are needed. UMTRA uses
random sampling and augmentation to create synthetic training tasks for
meta-learning phase. Labels are only needed at the final target task learning
step, and they can be as little as one sample per class. On the Omniglot and
Mini-Imagenet few-shot learning benchmarks, UMTRA outperforms every tested
approach based on unsupervised learning of representations, while alternating
for the best performance with the recent CACTUs algorithm. Compared to
supervised model-agnostic meta-learning approaches, UMTRA trades off some
classification accuracy for a reduction in the required labels of several
orders of magnitude
- …