33 research outputs found
CASE: Learning Conditional Adversarial Skill Embeddings for Physics-based Characters
We present CASE, an efficient and effective framework that learns
conditional Adversarial Skill Embeddings for physics-based characters. Our
physically simulated character can learn a diverse repertoire of skills while
providing controllability in the form of direct manipulation of the skills to
be performed. CASE divides the heterogeneous skill motions into distinct
subsets containing homogeneous samples for training a low-level conditional
model to learn conditional behavior distribution. The skill-conditioned
imitation learning naturally offers explicit control over the character's
skills after training. The training course incorporates the focal skill
sampling, skeletal residual forces, and element-wise feature masking to balance
diverse skills of varying complexities, mitigate dynamics mismatch to master
agile motions and capture more general behavior characteristics, respectively.
Once trained, the conditional model can produce highly diverse and realistic
skills, outperforming state-of-the-art models, and can be repurposed in various
downstream tasks. In particular, the explicit skill control handle allows a
high-level policy or user to direct the character with desired skill
specifications, which we demonstrate is advantageous for interactive character
animation.Comment: SIGGRAPH Asia 202
Active Visual Localization in Partially Calibrated Environments
Humans can robustly localize themselves without a map after they get lost
following prominent visual cues or landmarks. In this work, we aim at endowing
autonomous agents the same ability. Such ability is important in robotics
applications yet very challenging when an agent is exposed to partially
calibrated environments, where camera images with accurate 6 Degree-of-Freedom
pose labels only cover part of the scene. To address the above challenge, we
explore using Reinforcement Learning to search for a policy to generate
intelligent motions so as to actively localize the agent given visual
information in partially calibrated environments. Our core contribution is to
formulate the active visual localization problem as a Partially Observable
Markov Decision Process and propose an algorithmic framework based on Deep
Reinforcement Learning to solve it. We further propose an indoor scene dataset
ACR-6, which consists of both synthetic and real data and simulates challenging
scenarios for active visual localization. We benchmark our algorithm against
handcrafted baselines for localization and demonstrate that our approach
significantly outperforms them on localization success rate.Comment: https://www.youtube.com/watch?v=DIH-GbytCPM&feature=youtu.b
3D-Aware Object Goal Navigation via Simultaneous Exploration and Identification
Object goal navigation (ObjectNav) in unseen environments is a fundamental
task for Embodied AI. Agents in existing works learn ObjectNav policies based
on 2D maps, scene graphs, or image sequences. Considering this task happens in
3D space, a 3D-aware agent can advance its ObjectNav capability via learning
from fine-grained spatial information. However, leveraging 3D scene
representation can be prohibitively unpractical for policy learning in this
floor-level task, due to low sample efficiency and expensive computational
cost. In this work, we propose a framework for the challenging 3D-aware
ObjectNav based on two straightforward sub-policies. The two sub-polices,
namely corner-guided exploration policy and category-aware identification
policy, simultaneously perform by utilizing online fused 3D points as
observation. Through extensive experiments, we show that this framework can
dramatically improve the performance in ObjectNav through learning from 3D
scene representation. Our framework achieves the best performance among all
modular-based methods on the Matterport3D and Gibson datasets, while requiring
(up to 30x) less computational cost for training.Comment: To appear in CVPR 202