41,370 research outputs found
Human Motion Trajectory Prediction: A Survey
With growing numbers of intelligent autonomous systems in human environments,
the ability of such systems to perceive, understand and anticipate human
behavior becomes increasingly important. Specifically, predicting future
positions of dynamic agents and planning considering such predictions are key
tasks for self-driving vehicles, service robots and advanced surveillance
systems. This paper provides a survey of human motion trajectory prediction. We
review, analyze and structure a large selection of work from different
communities and propose a taxonomy that categorizes existing methods based on
the motion modeling approach and level of contextual information used. We
provide an overview of the existing datasets and performance metrics. We
discuss limitations of the state of the art and outline directions for further
research.Comment: Submitted to the International Journal of Robotics Research (IJRR),
37 page
Social Attention: Modeling Attention in Human Crowds
Robots that navigate through human crowds need to be able to plan safe,
efficient, and human predictable trajectories. This is a particularly
challenging problem as it requires the robot to predict future human
trajectories within a crowd where everyone implicitly cooperates with each
other to avoid collisions. Previous approaches to human trajectory prediction
have modeled the interactions between humans as a function of proximity.
However, that is not necessarily true as some people in our immediate vicinity
moving in the same direction might not be as important as other people that are
further away, but that might collide with us in the future. In this work, we
propose Social Attention, a novel trajectory prediction model that captures the
relative importance of each person when navigating in the crowd, irrespective
of their proximity. We demonstrate the performance of our method against a
state-of-the-art approach on two publicly available crowd datasets and analyze
the trained attention model to gain a better understanding of which surrounding
agents humans attend to, when navigating in a crowd
Modeling Cooperative Navigation in Dense Human Crowds
For robots to be a part of our daily life, they need to be able to navigate
among crowds not only safely but also in a socially compliant fashion. This is
a challenging problem because humans tend to navigate by implicitly cooperating
with one another to avoid collisions, while heading toward their respective
destinations. Previous approaches have used hand-crafted functions based on
proximity to model human-human and human-robot interactions. However, these
approaches can only model simple interactions and fail to generalize for
complex crowded settings. In this paper, we develop an approach that models the
joint distribution over future trajectories of all interacting agents in the
crowd, through a local interaction model that we train using real human
trajectory data. The interaction model infers the velocity of each agent based
on the spatial orientation of other agents in his vicinity. During prediction,
our approach infers the goal of the agent from its past trajectory and uses the
learned model to predict its future trajectory. We demonstrate the performance
of our method against a state-of-the-art approach on a public dataset and show
that our model outperforms when predicting future trajectories for longer
horizons.Comment: Accepted at ICRA 201
Robust Modeling of Epistemic Mental States
This work identifies and advances some research challenges in the analysis of
facial features and their temporal dynamics with epistemic mental states in
dyadic conversations. Epistemic states are: Agreement, Concentration,
Thoughtful, Certain, and Interest. In this paper, we perform a number of
statistical analyses and simulations to identify the relationship between
facial features and epistemic states. Non-linear relations are found to be more
prevalent, while temporal features derived from original facial features have
demonstrated a strong correlation with intensity changes. Then, we propose a
novel prediction framework that takes facial features and their nonlinear
relation scores as input and predict different epistemic states in videos. The
prediction of epistemic states is boosted when the classification of emotion
changing regions such as rising, falling, or steady-state are incorporated with
the temporal features. The proposed predictive models can predict the epistemic
states with significantly improved accuracy: correlation coefficient (CoERR)
for Agreement is 0.827, for Concentration 0.901, for Thoughtful 0.794, for
Certain 0.854, and for Interest 0.913.Comment: Accepted for Publication in Multimedia Tools and Application, Special
Issue: Socio-Affective Technologie
Forecasting People Trajectories and Head Poses by Jointly Reasoning on Tracklets and Vislets
In this work, we explore the correlation between people trajectories and
their head orientations. We argue that people trajectory and head pose
forecasting can be modelled as a joint problem. Recent approaches on trajectory
forecasting leverage short-term trajectories (aka tracklets) of pedestrians to
predict their future paths. In addition, sociological cues, such as expected
destination or pedestrian interaction, are often combined with tracklets. In
this paper, we propose MiXing-LSTM (MX-LSTM) to capture the interplay between
positions and head orientations (vislets) thanks to a joint unconstrained
optimization of full covariance matrices during the LSTM backpropagation. We
additionally exploit the head orientations as a proxy for the visual attention,
when modeling social interactions. MX-LSTM predicts future pedestrians location
and head pose, increasing the standard capabilities of the current approaches
on long-term trajectory forecasting. Compared to the state-of-the-art, our
approach shows better performances on an extensive set of public benchmarks.
MX-LSTM is particularly effective when people move slowly, i.e. the most
challenging scenario for all other models. The proposed approach also allows
for accurate predictions on a longer time horizon.Comment: Accepted at IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE
INTELLIGENCE 2019. arXiv admin note: text overlap with arXiv:1805.0065
Fundamental structures of dynamic social networks
Social systems are in a constant state of flux with dynamics spanning from
minute-by-minute changes to patterns present on the timescale of years.
Accurate models of social dynamics are important for understanding spreading of
influence or diseases, formation of friendships, and the productivity of teams.
While there has been much progress on understanding complex networks over the
past decade, little is known about the regularities governing the
micro-dynamics of social networks. Here we explore the dynamic social network
of a densely-connected population of approximately 1000 individuals and their
interactions in the network of real-world person-to-person proximity measured
via Bluetooth, as well as their telecommunication networks, online social media
contacts, geo-location, and demographic data. These high-resolution data allow
us to observe social groups directly, rendering community detection
unnecessary. Starting from 5-minute time slices we uncover dynamic social
structures expressed on multiple timescales. On the hourly timescale, we find
that gatherings are fluid, with members coming and going, but organized via a
stable core of individuals. Each core represents a social context. Cores
exhibit a pattern of recurring meetings across weeks and months, each with
varying degrees of regularity. Taken together, these findings provide a
powerful simplification of the social network, where cores represent
fundamental structures expressed with strong temporal and spatial regularity.
Using this framework, we explore the complex interplay between social and
geospatial behavior, documenting how the formation of cores are preceded by
coordination behavior in the communication networks, and demonstrating that
social behavior can be predicted with high precision.Comment: Main Manuscript: 16 pages, 4 figures. Supplementary Information: 39
pages, 34 figure
How simple rules determine pedestrian behavior and crowd disasters
With the increasing size and frequency of mass events, the study of crowd
disasters and the simulation of pedestrian flows have become important research
areas. Yet, even successful modeling approaches such as those inspired by
Newtonian force models are still not fully consistent with empirical
observations and are sometimes hard to calibrate. Here, a novel cognitive
science approach is proposed, which is based on behavioral heuristics. We
suggest that, guided by visual information, namely the distance of obstructions
in candidate lines of sight, pedestrians apply two simple cognitive procedures
to adapt their walking speeds and directions. While simpler than previous
approaches, this model predicts individual trajectories and collective patterns
of motion in good quantitative agreement with a large variety of empirical and
experimental data. This includes the emergence of self-organization phenomena,
such as the spontaneous formation of unidirectional lanes or stop-and-go waves.
Moreover, the combination of pedestrian heuristics with body collisions
generates crowd turbulence at extreme densities-a phenomenon that has been
observed during recent crowd disasters. By proposing an integrated treatment of
simultaneous interactions between multiple individuals, our approach overcomes
limitations of current physics-inspired pair interaction models. Understanding
crowd dynamics through cognitive heuristics is therefore not only crucial for a
better preparation of safe mass events. It also clears the way for a more
realistic modeling of collective social behaviors, in particular of human
crowds and biological swarms. Furthermore, our behavioral heuristics may serve
to improve the navigation of autonomous robots.Comment: Article accepted for publication in PNA
- …