1,109,654 research outputs found
Time-Contrastive Networks: Self-Supervised Learning from Video
We propose a self-supervised approach for learning representations and
robotic behaviors entirely from unlabeled videos recorded from multiple
viewpoints, and study how this representation can be used in two robotic
imitation settings: imitating object interactions from videos of humans, and
imitating human poses. Imitation of human behavior requires a
viewpoint-invariant representation that captures the relationships between
end-effectors (hands or robot grippers) and the environment, object attributes,
and body pose. We train our representations using a metric learning loss, where
multiple simultaneous viewpoints of the same observation are attracted in the
embedding space, while being repelled from temporal neighbors which are often
visually similar but functionally different. In other words, the model
simultaneously learns to recognize what is common between different-looking
images, and what is different between similar-looking images. This signal
causes our model to discover attributes that do not change across viewpoint,
but do change across time, while ignoring nuisance variables such as
occlusions, motion blur, lighting and background. We demonstrate that this
representation can be used by a robot to directly mimic human poses without an
explicit correspondence, and that it can be used as a reward function within a
reinforcement learning algorithm. While representations are learned from an
unlabeled collection of task-related videos, robot behaviors such as pouring
are learned by watching a single 3rd-person demonstration by a human. Reward
functions obtained by following the human demonstrations under the learned
representation enable efficient reinforcement learning that is practical for
real-world robotic systems. Video results, open-source code and dataset are
available at https://sermanet.github.io/imitat
Similarity of Pre-trained and Fine-tuned Representations
In transfer learning, only the last part of the networks - the so-called head
- is often fine-tuned. Representation similarity analysis shows that the most
significant change still occurs in the head even if all weights are updatable.
However, recent results from few-shot learning have shown that representation
change in the early layers, which are mostly convolutional, is beneficial,
especially in the case of cross-domain adaption. In our paper, we find out
whether that also holds true for transfer learning. In addition, we analyze the
change of representation in transfer learning, both during pre-training and
fine-tuning, and find out that pre-trained structure is unlearned if not
usable.Comment: Workshop of Updatable Machine Learning at ICML 202
Recommended from our members
Sharing practice, problems and solutions for institutional change
This chapter critiques the roles of different forms of representation of practice as part of an institutional change process. It discusses how these representations can be used both to design and to share learning activities at the various levels of decision-making in a university. We illustrate our arguments with empirical data gathered on change processes associated with an institution-wide change programme: the introduction of a new virtual learning environment (VLE). In particular, we describe a case study of the introduction of the VLE tools in a business course. We focus on two particular forms of representations to describe the essence of the innovation: a pedagogical pattern and a visual learning design. We argue that pedagogical patterns and learning design have emerged as parallel approaches to describing practice in recent years. Despite their very different origins, both provide complementary representations, which emphasize different aspects of the practice being described. We are attempting to combine these approaches. We briefly outline the Open University Learning Design initiative, of which this work is part, and describe its key underpinning philosophies. We believe our approach provides a vehicle for enabling a better articulation of design principles and the discussion of issues concerning the re-use of educational resources and activities
Building and Refining Abstract Planning Cases by Change of Representation Language
ion is one of the most promising approaches to improve the performance of
problem solvers. In several domains abstraction by dropping sentences of a
domain description -- as used in most hierarchical planners -- has proven
useful. In this paper we present examples which illustrate significant
drawbacks of abstraction by dropping sentences. To overcome these drawbacks, we
propose a more general view of abstraction involving the change of
representation language. We have developed a new abstraction methodology and a
related sound and complete learning algorithm that allows the complete change
of representation language of planning cases from concrete to abstract.
However, to achieve a powerful change of the representation language, the
abstract language itself as well as rules which describe admissible ways of
abstracting states must be provided in the domain model. This new abstraction
approach is the core of Paris (Plan Abstraction and Refinement in an Integrated
System), a system in which abstract planning cases are automatically learned
from given concrete cases. An empirical study in the domain of process planning
in mechanical engineering shows significant advantages of the proposed
reasoning from abstract cases over classical hierarchical planning.Comment: See http://www.jair.org/ for an online appendix and other files
accompanying this articl
- тАж