27 research outputs found
A Primer on Motion Capture with Deep Learning: Principles, Pitfalls and Perspectives
Extracting behavioral measurements non-invasively from video is stymied by
the fact that it is a hard computational problem. Recent advances in deep
learning have tremendously advanced predicting posture from videos directly,
which quickly impacted neuroscience and biology more broadly. In this primer we
review the budding field of motion capture with deep learning. In particular,
we will discuss the principles of those novel algorithms, highlight their
potential as well as pitfalls for experimentalists, and provide a glimpse into
the future.Comment: Review, 21 pages, 8 figures and 5 boxe
AmadeusGPT: a natural language interface for interactive animal behavioral analysis
The process of quantifying and analyzing animal behavior involves translating
the naturally occurring descriptive language of their actions into
machine-readable code. Yet, codifying behavior analysis is often challenging
without deep understanding of animal behavior and technical machine learning
knowledge. To limit this gap, we introduce AmadeusGPT: a natural language
interface that turns natural language descriptions of behaviors into
machine-executable code. Large-language models (LLMs) such as GPT3.5 and GPT4
allow for interactive language-based queries that are potentially well suited
for making interactive behavior analysis. However, the comprehension capability
of these LLMs is limited by the context window size, which prevents it from
remembering distant conversations. To overcome the context window limitation,
we implement a novel dual-memory mechanism to allow communication between
short-term and long-term memory using symbols as context pointers for retrieval
and saving. Concretely, users directly use language-based definitions of
behavior and our augmented GPT develops code based on the core AmadeusGPT API,
which contains machine learning, computer vision, spatio-temporal reasoning,
and visualization modules. Users then can interactively refine results, and
seamlessly add new behavioral modules as needed. We benchmark AmadeusGPT and
show we can produce state-of-the-art performance on the MABE 2022 behavior
challenge tasks. Note, an end-user would not need to write any code to achieve
this. Thus, collectively AmadeusGPT presents a novel way to merge deep
biological knowledge, large-language models, and core computer vision modules
into a more naturally intelligent system. Code and demos can be found at:
https://github.com/AdaptiveMotorControlLab/AmadeusGPT.Comment: demo available https://github.com/AdaptiveMotorControlLab/AmadeusGP
AcinoSet: A 3D Pose Estimation Dataset and Baseline Models for Cheetahs in the Wild
Animals are capable of extreme agility, yet understanding their complex
dynamics, which have ecological, biomechanical and evolutionary implications,
remains challenging. Being able to study this incredible agility will be
critical for the development of next-generation autonomous legged robots. In
particular, the cheetah (acinonyx jubatus) is supremely fast and maneuverable,
yet quantifying its whole-body 3D kinematic data during locomotion in the wild
remains a challenge, even with new deep learning-based methods. In this work we
present an extensive dataset of free-running cheetahs in the wild, called
AcinoSet, that contains 119,490 frames of multi-view synchronized high-speed
video footage, camera calibration files and 7,588 human-annotated frames. We
utilize markerless animal pose estimation to provide 2D keypoints. Then, we use
three methods that serve as strong baselines for 3D pose estimation tool
development: traditional sparse bundle adjustment, an Extended Kalman Filter,
and a trajectory optimization-based method we call Full Trajectory Estimation.
The resulting 3D trajectories, human-checked 3D ground truth, and an
interactive tool to inspect the data is also provided. We believe this dataset
will be useful for a diverse range of fields such as ecology, neuroscience,
robotics, biomechanics as well as computer vision.Comment: Code and data can be found at:
https://github.com/African-Robotics-Unit/AcinoSe
A land-to-ocean perspective on the magnitude, source and implication of DIC flux from major Arctic rivers to the Arctic Ocean
Author Posting. © American Geophysical Union, 2012. This article is posted here by permission of American Geophysical Union for personal use, not for redistribution. The definitive version was published in Global Biogeochemical Cycles 26 (2012): GB4018, doi:10.1029/2011GB004192.A series of seasonally distributed measurements from the six largest Arctic rivers (the Ob', Yenisey, Lena, Kolyma, Yukon and Mackenzie) was used to examine the magnitude and significance of Arctic riverine DIC flux to larger scale C dynamics within the Arctic system. DIC concentration showed considerable, and synchronous, seasonal variation across these six large Arctic rivers, which have an estimated combined annual DIC flux of 30 Tg C yr−1. By examining the relationship between DIC flux and landscape variables known to regulate riverine DIC, we extrapolate to a DIC flux of 57 ± 9.9 Tg C yr−1for the full pan-arctic basin, and show that DIC export increases with runoff, the extent of carbonate rocks and glacial coverage, but decreases with permafrost extent. This pan-arctic riverine DIC estimate represents 13–15% of the total global DIC flux. The annual flux of selected ions (HCO3−, Na+, Ca2+, Mg2+, Sr2+, and Cl−) from the six largest Arctic rivers confirms that chemical weathering is dominated by inputs from carbonate rocks in the North American watersheds, but points to a more important role for silicate rocks in Siberian watersheds. In the coastal ocean, river water-induced decreases in aragonite saturation (i.e., an ocean acidification effect) appears to be much more pronounced in Siberia than in the North American Arctic, and stronger in the winter and spring than in the late summer. Accounting for seasonal variation in the flux of DIC and other major ions gives a much clearer understanding of the importance of riverine DIC within the broader pan-arctic C cycle.Funding for this work was provided through
NSF-OPP-0229302 and NSF-OPP-0732985. Additional support to SET
was provided by an NSERC Postdoctoral Fellowship.2013-06-1
Motor control: Neural correlates of optimal feedback control theory
Recent work is revealing neural correlates of a leading theory of motor control. By linking an elegant series of behavioral experiments with neural inactivation in macaques with computational models, a new study shows that premotor and parietal areas can be mapped onto a model for optimal feedback control
Rethinking pose estimation in crowds: overcoming the detection information bottleneck and ambiguity
<p>#############</p><p>Rethinking pose estimation in crowds: overcoming the detection information-bottleneck and ambiguity, ICCV 2023</p><p>#############</p><p> </p><p>Authors: Zhou, Mu and Stoffl, Lucas and Mathis, Mackenzie Weygandt and Mathis, Alexander</p><p>Affiliation: EPFL</p><p>Date: October, 2023</p><p> </p><p>Here we provide neural networks weights for the best models in our article "Rethinking pose estimation in crowds: overcoming the detection information-bottleneck and ambiguity", ICCV 2023. Each model has the naming convention "dataset"-"modeltype".pth</p><p>These pth files can be loaded with PyTorch. The code to load and use the models is available at The code to load and use the models is available at: <a href="https://github.com/amathislab/BUCTD">https://github.com/amathislab/BUCTD</a></p><p><strong>Note: The weights for OCHuman</strong>, are called COCO-* as one only trains on COCO. So OCHuman-X := COCO-X</p><p>We also share the predictions from various bottom-up models to reproduce the training stored in *.json format (compressed as zip files). See our repository for more details.</p><p> </p><p>Link to the ICCV article: </p><p><a href="https://openaccess.thecvf.com/content/ICCV2023/papers/Zhou_Rethinking_Pose_Estimation_in_Crowds_Overcoming_the_Detection_Information_Bottleneck_ICCV_2023_paper.pdf">https://openaccess.thecvf.com/content/ICCV2023/papers/Zhou_Rethinking_Pose_Estimation_in_Crowds_Overcoming_the_Detection_Information_Bottleneck_ICCV_2023_paper.pdf</a></p><p> </p><p>The weights and predictions are released with Creative Commons Attribution 4.0 license. The code is released under the Apache 2.0 license, see https://github.com/amathislab/BUCTD </p><p><i>If you find our weights, code or ideas useful, please cite:</i></p><p> </p><p>@InProceedings{Zhou_2023_ICCV,</p><p> author = {Zhou, Mu and Stoffl, Lucas and Mathis, Mackenzie Weygandt and Mathis, Alexander},</p><p> title = {Rethinking Pose Estimation in Crowds: Overcoming the Detection Information Bottleneck and Ambiguity},</p><p> booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},</p><p> month = {October},</p><p> year = {2023},</p><p> pages = {14689-14699}</p><p>}</p><p> </p>