9,281 research outputs found
Progressive Neural Networks
Learning to solve complex sequences of tasks--while both leveraging transfer
and avoiding catastrophic forgetting--remains a key obstacle to achieving
human-level intelligence. The progressive networks approach represents a step
forward in this direction: they are immune to forgetting and can leverage prior
knowledge via lateral connections to previously learned features. We evaluate
this architecture extensively on a wide variety of reinforcement learning tasks
(Atari and 3D maze games), and show that it outperforms common baselines based
on pretraining and finetuning. Using a novel sensitivity measure, we
demonstrate that transfer occurs at both low-level sensory and high-level
control layers of the learned policy
Modular knowledge systems accelerate human migration in asymmetric random environments
Migration is a key mechanism for expansion of communities. In spatially
heterogeneous environments, rapidly gaining knowledge about the local
environment is key to the evolutionary success of a migrating population. For
historical human migration, environmental heterogeneity was naturally
asymmetric in the north-south (NS) and east-west (EW) directions. We here
consider the human migration process in the Americas, modeled as random,
asymmetric, modularly correlated environments. Knowledge about the environments
determines the fitness of each individual. We present a phase diagram for
asymmetry of migration as a function of carrying capacity and fitness
threshold. We find that the speed of migration is proportional to the inverse
complement of the spatial environmental gradient, and in particular we find
that north-south migration rates are lower than east-west migration rates when
the environmental gradient is higher in the north-south direction.
Communication of knowledge between individuals can help to spread beneficial
knowledge within the population. The speed of migration increases when
communication transmits pieces of knowledge that contribute in a modular way to
the fitness of individuals. The results for the dependence of migration rate on
asymmetry and modularity are consistent with existing archaeological
observations. The results for asymmetry of genetic divergence are consistent
with patterns of human gene flow.Comment: 13 pages, 6 figures, 1 table in Proc. Roy. Soc. Interface 201
Learning Generative Models across Incomparable Spaces
Generative Adversarial Networks have shown remarkable success in learning a
distribution that faithfully recovers a reference distribution in its entirety.
However, in some cases, we may want to only learn some aspects (e.g., cluster
or manifold structure), while modifying others (e.g., style, orientation or
dimension). In this work, we propose an approach to learn generative models
across such incomparable spaces, and demonstrate how to steer the learned
distribution towards target properties. A key component of our model is the
Gromov-Wasserstein distance, a notion of discrepancy that compares
distributions relationally rather than absolutely. While this framework
subsumes current generative models in identically reproducing distributions,
its inherent flexibility allows application to tasks in manifold learning,
relational learning and cross-domain learning.Comment: International Conference on Machine Learning (ICML
- …