25 research outputs found
Combined Reinforcement Learning via Abstract Representations
In the quest for efficient and robust reinforcement learning methods, both
model-free and model-based approaches offer advantages. In this paper we
propose a new way of explicitly bridging both approaches via a shared
low-dimensional learned encoding of the environment, meant to capture
summarizing abstractions. We show that the modularity brought by this approach
leads to good generalization while being computationally efficient, with
planning happening in a smaller latent state space. In addition, this approach
recovers a sufficient low-dimensional representation of the environment, which
opens up new strategies for interpretable AI, exploration and transfer
learning.Comment: Accepted to the Thirty-Third AAAI Conference On Artificial
Intelligence, 201
Multi-Source Transfer Learning for Deep Model-Based Reinforcement Learning
A crucial challenge in reinforcement learning is to reduce the number of
interactions with the environment that an agent requires to master a given
task. Transfer learning proposes to address this issue by re-using knowledge
from previously learned tasks. However, determining which source task qualifies
as optimal for knowledge extraction, as well as the choice regarding which
algorithm components to transfer, represent severe obstacles to its application
in reinforcement learning. The goal of this paper is to alleviate these issues
with modular multi-source transfer learning techniques. Our proposed
methodologies automatically learn how to extract useful information from source
tasks, regardless of the difference in state-action space and reward function.
We support our claims with extensive and challenging cross-domain experiments
for visual control.Comment: 15 pages, 6 figures, 8 tables. arXiv admin note: text overlap with
arXiv:2108.0652