12 research outputs found
Grounding Neuroscience in Behavioral Changes using Artificial Neural Networks
Connecting neural activity to function is a common aim in neuroscience. How
to define and conceptualize function, however, can vary. Here I focus on
grounding this goal in the specific question of how a given change in behavior
is produced by a change in neural circuits or activity. Artificial neural
network models offer a particularly fruitful format for tackling such questions
because they use neural mechanisms to perform complex transformations and
produce appropriate behavior. Therefore, they can be a means of causally
testing the extent to which a neural change can be responsible for an
experimentally observed behavioral change. Furthermore, because the field of
interpretability in artificial intelligence has similar aims, neuroscientists
can look to interpretability methods for new ways of identifying neural
features that drive performance and behaviors.Comment: Final Accepted Manuscrip
Catalyzing next-generation Artificial Intelligence through NeuroAI
Neuroscience has long been an essential driver of progress in artificial intelligence (AI). We propose that to accelerate progress in AI, we must invest in fundamental research in NeuroAI. A core component of this is the embodied Turing test, which challenges AI animal models to interact with the sensorimotor world at skill levels akin to their living counterparts. The embodied Turing test shifts the focus from those capabilities like game playing and language that are especially well-developed or uniquely human to those capabilities - inherited from over 500 million years of evolution - that are shared with all animals. Building models that can pass the embodied Turing test will provide a roadmap for the next generation of AI
Learning not to learn: Nature versus nurture in silico
Animals are equipped with a rich innate repertoire of sensory, behavioral and
motor skills, which allows them to interact with the world immediately after
birth. At the same time, many behaviors are highly adaptive and can be tailored
to specific environments by means of learning. In this work, we use
mathematical analysis and the framework of meta-learning (or 'learning to
learn') to answer when it is beneficial to learn such an adaptive strategy and
when to hard-code a heuristic behavior. We find that the interplay of
ecological uncertainty, task complexity and the agents' lifetime has crucial
effects on the meta-learned amortized Bayesian inference performed by an agent.
There exist two regimes: One in which meta-learning yields a learning algorithm
that implements task-dependent information-integration and a second regime in
which meta-learning imprints a heuristic or 'hard-coded' behavior. Further
analysis reveals that non-adaptive behaviors are not only optimal for aspects
of the environment that are stable across individuals, but also in situations
where an adaptation to the environment would in fact be highly beneficial, but
could not be done quickly enough to be exploited within the remaining lifetime.
Hard-coded behaviors should hence not only be those that always work, but also
those that are too complex to be learned within a reasonable time frame
Learning Task-Agnostic Action Spaces for Movement Optimization
We propose a novel method for exploring the dynamics of physically based
animated characters, and learning a task-agnostic action space that makes
movement optimization easier. Like several previous papers, we parameterize
actions as target states, and learn a short-horizon goal-conditioned low-level
control policy that drives the agent's state towards the targets. Our novel
contribution is that with our exploration data, we are able to learn the
low-level policy in a generic manner and without any reference movement data.
Trained once for each agent or simulation environment, the policy improves the
efficiency of optimizing both trajectories and high-level policies across
multiple tasks and optimization algorithms. We also contribute novel
visualizations that show how using target states as actions makes optimized
trajectories more robust to disturbances; this manifests as wider optima that
are easy to find. Due to its simplicity and generality, our proposed approach
should provide a building block that can improve a large variety of movement
optimization methods and applications.Comment: Accepted as a regular paper by IEEE Transactions on Visualization and
Computer Graphics (TVCG) in July 202