4 research outputs found
TRiPOD: Human Trajectory and Pose Dynamics Forecasting in the Wild
Joint forecasting of human trajectory and pose dynamics is a fundamental
building block of various applications ranging from robotics and autonomous
driving to surveillance systems. Predicting body dynamics requires capturing
subtle information embedded in the humans' interactions with each other and
with the objects present in the scene. In this paper, we propose a novel
TRajectory and POse Dynamics (nicknamed TRiPOD) method based on graph
attentional networks to model the human-human and human-object interactions
both in the input space and the output space (decoded future output). The model
is supplemented by a message passing interface over the graphs to fuse these
different levels of interactions efficiently. Furthermore, to incorporate a
real-world challenge, we propound to learn an indicator representing whether an
estimated body joint is visible/invisible at each frame, e.g. due to occlusion
or being outside the sensor field of view. Finally, we introduce a new
benchmark for this joint task based on two challenging datasets (PoseTrack and
3DPW) and propose evaluation metrics to measure the effectiveness of
predictions in the global space, even when there are invisible cases of joints.
Our evaluation shows that TRiPOD outperforms all prior work and
state-of-the-art specifically designed for each of the trajectory and pose
forecasting tasks
From Perception to Navigation in Environments with Persons: An Indoor Evaluation of the State of the Art
Research in the field of social robotics is allowing service robots to operate in environments with people. In the aim of realizing the vision of humans and robots coexisting in the same environment, several solutions have been proposed to (1) perceive persons and objects in the immediate environment; (2) predict the movements of humans; as well as (3) plan the navigation in agreement with socially accepted rules. In this work, we discuss the different aspects related to social navigation in the context of our experience in an indoor environment. We describe state-of-the-art approaches and experiment with existing methods to analyze their performance in practice. From this study, we gather first-hand insights into the limitations of current solutions and identify possible research directions to address the open challenges. In particular, this paper focuses on topics related to perception at the hardware and application levels, including 2D and 3D sensors, geometric and mainly semantic mapping, the prediction of people trajectories (physics-, pattern- and planning-based), and social navigation (reactive and predictive) in indoor environments