93 research outputs found
SmartMocap: Joint Estimation of Human and Camera Motion using Uncalibrated RGB Cameras
Markerless human motion capture (mocap) from multiple RGB cameras is a widely
studied problem. Existing methods either need calibrated cameras or calibrate
them relative to a static camera, which acts as the reference frame for the
mocap system. The calibration step has to be done a priori for every capture
session, which is a tedious process, and re-calibration is required whenever
cameras are intentionally or accidentally moved. In this paper, we propose a
mocap method which uses multiple static and moving extrinsically uncalibrated
RGB cameras. The key components of our method are as follows. First, since the
cameras and the subject can move freely, we select the ground plane as a common
reference to represent both the body and the camera motions unlike existing
methods which represent bodies in the camera coordinate. Second, we learn a
probability distribution of short human motion sequences (1sec) relative
to the ground plane and leverage it to disambiguate between the camera and
human motion. Third, we use this distribution as a motion prior in a novel
multi-stage optimization approach to fit the SMPL human body model and the
camera poses to the human body keypoints on the images. Finally, we show that
our method can work on a variety of datasets ranging from aerial cameras to
smartphones. It also gives more accurate results compared to the
state-of-the-art on the task of monocular human mocap with a static camera. Our
code is available for research purposes on
https://github.com/robot-perception-group/SmartMocap
ActiveMoCap: Optimized Viewpoint Selection for Active Human Motion Capture
The accuracy of monocular 3D human pose estimation depends on the viewpoint
from which the image is captured. While freely moving cameras, such as on
drones, provide control over this viewpoint, automatically positioning them at
the location which will yield the highest accuracy remains an open problem.
This is the problem that we address in this paper. Specifically, given a short
video sequence, we introduce an algorithm that predicts which viewpoints should
be chosen to capture future frames so as to maximize 3D human pose estimation
accuracy. The key idea underlying our approach is a method to estimate the
uncertainty of the 3D body pose estimates. We integrate several sources of
uncertainty, originating from deep learning based regressors and temporal
smoothness. Our motion planner yields improved 3D body pose estimates and
outperforms or matches existing ones that are based on person following and
orbiting.Comment: For associated video, see https://youtu.be/i58Bu-hbZHs Published in
CVPR 202
Enabling Multi-LiDAR Sensing in GNSS-Denied Environments: SLAM Dataset, Benchmark, and UAV Tracking with LiDAR-as-a-camera
The rise of Light Detection and Ranging (LiDAR) sensors has profoundly impacted industries ranging from automotive to urban planning. As these sensors become increasingly affordable and compact, their applications are diversifying, driving precision, and innovation. This thesis delves into LiDAR's advancements in autonomous robotic systems, with a focus on its role in simultaneous localization and mapping (SLAM) methodologies and LiDAR as a camera-based tracking for Unmanned Aerial Vehicles (UAV).
Our contributions span two primary domains: the Multi-Modal LiDAR SLAM Benchmark, and the LiDAR-as-a-camera UAV Tracking. In the former, we have expanded our previous multi-modal LiDAR dataset by adding more data sequences from various scenarios. In contrast to the previous dataset, we employ different ground truth-generating approaches. We propose a new multi-modal multi-lidar SLAM-assisted and ICP-based sensor fusion method for generating ground truth maps. Additionally, we also supplement our data with new open road sequences with GNSS-RTK. This enriched dataset, supported by high-resolution LiDAR, provides detailed insights through an evaluation of ten configurations, pairing diverse LiDAR sensors with state-of-the-art SLAM algorithms. In the latter contribution, we leverage a custom YOLOv5 model trained on panoramic low-resolution images from LiDAR reflectivity (LiDAR-as-a-camera) to detect UAVs, demonstrating the superiority of this approach over point cloud or image-only methods. Additionally, we evaluated the real-time performance of our approach on the Nvidia Jetson Nano, a popular mobile computing platform.
Overall, our research underscores the transformative potential of integrating advanced LiDAR sensors with autonomous robotics. By bridging the gaps between different technological approaches, we pave the way for more versatile and efficient applications in the future
Modern facilities for experimental measurement of dynamic loads induced by humans: a literature review.
This paper provides a critical overview of available technology and facilities for determining human-induced dynamic forces of civil engineering structures, such as due to walking, running, jumping and bouncing. In addition to traditional equipment for direct force measurements comprising force plate(s), foot pressure insoles and instrumented treadmills, the review also investigates possibility of using optical motion tracking systems (marker-based and marker-free optoelectronic technology) and non-optical motion tracking systems (inertial sensors) to reproduce contact forces between humans and structures based on body kinematics data and known body mass distribution. Although significant technological advancements have been made in the last decade, the literature survey showed that the state-of-the-art force measurements are often limited to individuals in artificial laboratory environments. Experimental identification of seriously needed group- and crowd-induced force data recorded on as-built structures, such as footbridges, grandstands and floors, still remains a challenge due to the complexity of human actions and the lack of adequate equipment
Kinematic State Estimation using Multiple DGPS/MEMS-IMU Sensors
Animals have evolved over billions of years and understanding these complex and intertwined systems have potential to advance the technology in the field of sports science, robotics and more. As such, a gait analysis using Motion Capture (MOCAP) technology is the subject of a number of research and development projects aimed at obtaining quantitative measurements. Existing MOCAP technology has limited the majority of studies to the analysis of the steady-state locomotion in a controlled (indoor) laboratory environment. MOCAP systems such as the optical, non-optical acoustic and non-optical magnetic MOCAP systems require predefined capture volume and controlled environmental conditions whilst the non-optical mechanical MOCAP system impedes the motion of the subject. Although the non-optical inertial MOCAP system allows MOCAP in an outdoor environment, it suffers from measurement noise and drift and lacks global trajectory information. The accuracy of these MOCAP systems are known to decrease during the tracking of the transient locomotion. Quantifying the manoeuvrability of animals in their natural habitat to answer the question âWhy are animals so manoeuvrable?â remains a challenge. This research aims to develop an outdoor MOCAP system that will allow tracking of the steady-state as well as the transient locomotion of an animal in its natural habitat outside a controlled laboratory condition. A number of researchers have developed novel MOCAP systems with the same aim of creating an outdoor MOCAP system that is aimed at tracking the motion outside a controlled laboratory (indoor) environment with unlimited capture volume. These novel MOCAP systems are either not validated against the commercial MOCAP systems or do not have comparable sub-millimetre accuracy as the commercial MOCAP systems. The developed DGPS/MEMS-IMU multi-receiver fusion MOCAP system was assessed to have global trajectory accuracy of _0:0394m, relative limb position accuracy of _0:006497m. To conclude the research, several recommendations are made to improve the developed MOCAP system and to prepare for a field-testing with a wild animal from a family of a terrestrial megafauna
Autonomous 3D mapping and surveillance of mines with MAVs
A dissertation Submitted to the Faculty of Science, University of the
Witwatersrand, Johannesburg, for the degree of Master of Science.
12 July 2017.The mapping of mines, both operational and abandoned, is a long, di cult and occasionally
dangerous task especially in the latter case. Recent developments in active and passive consumer
grade sensors, as well as quadcopter drones present the opportunity to automate these
challenging tasks providing cost and safety bene ts. The goal of this research is to develop an
autonomous vision-based mapping system that employs quadrotor drones to explore and map
sections of mine tunnels. The system is equipped with inexpensive, structured light, depth cameras
in place of traditional laser scanners, making the quadrotor setup more viable to produce in
bulk. A modi ed version of Microsoft's Kinect Fusion algorithm is used to construct 3D point
clouds in real-time as the agents traverse the scene. Finally, the generated and merged point
clouds from the system are compared with those produced by current Lidar scanners.LG201
- âŠ