22 research outputs found
Recommended from our members
Hâ LIDAR odometry for spacecraft relative navigation
Current light detection and ranging (LIDAR) based odometry solutions that are used for spacecraft relative navigation suffer from quite a few deficiencies. These include an off-line training requirement and relying on the iterative closest point (ICP) that does not guarantee a globally optimum solution. To encounter this, the authors suggest a robust architecture that overcomes the problems of current proposals by combining the concepts of 3D local feature matching with an adaptive variant of the Hâ recursive filtering process. Trials on real laser scans of an EnviSat model demonstrate that the proposed architecture affords at least one order of magnitude better accuracy compared to ICP
Cosmological distance indicators
We review three distance measurement techniques beyond the local universe:
(1) gravitational lens time delays, (2) baryon acoustic oscillation (BAO), and
(3) HI intensity mapping. We describe the principles and theory behind each
method, the ingredients needed for measuring such distances, the current
observational results, and future prospects. Time delays from strongly lensed
quasars currently provide constraints on with < 4% uncertainty, and with
1% within reach from ongoing surveys and efforts. Recent exciting discoveries
of strongly lensed supernovae hold great promise for time-delay cosmography.
BAO features have been detected in redshift surveys up to z <~ 0.8 with
galaxies and z ~ 2 with Ly- forest, providing precise distance
measurements and with < 2% uncertainty in flat CDM. Future BAO
surveys will probe the distance scale with percent-level precision. HI
intensity mapping has great potential to map BAO distances at z ~ 0.8 and
beyond with precisions of a few percent. The next years ahead will be exciting
as various cosmological probes reach 1% uncertainty in determining , to
assess the current tension in measurements that could indicate new
physics.Comment: Review article accepted for publication in Space Science Reviews
(Springer), 45 pages, 10 figures. Chapter of a special collection resulting
from the May 2016 ISSI-BJ workshop on Astronomical Distance Determination in
the Space Ag
OUR-CVFH - Oriented, Unique and Repeatable Clustered Viewpoint Feature Histogram for Object Recognition and 6DOF Pose Estimation
We propose a novel method to estimate a unique and repeatable reference frame in the context of 3D object recognition from a single viewpoint based on global descriptors. We show that the ability of defining a robust reference frame on both model and scene views allows creating descriptive global representations of the object view, with the beneficial effect of enhancing the spatial descriptiveness of the feature and its ability to recognize objects by means of a simple nearest neighbor classifier computed on the descriptor space. Moreover, the definition of repeatable directions can be deployed to efficiently retrieve the 6DOF pose of the objects in a scene. We experimentally demonstrate the effectiveness of the proposed method on a dataset including 23 scenes acquired with the Microsoft Kinect sensor and 25 full-3D models by comparing the proposed approach with state-of-the-art global descriptors. A substantial improvement is presented regarding accuracy in recognition and 6DOF pose estimation, as well as in terms of computational performanc
Human Action Recognition from RGB-D Frames Based on Real-Time 3D Optical Flow Estimation
Modern advances in the area of intelligent agents have led to the concept of cognitive robots. A cognitive robot is not only able to perceive complex stimuli from the environment, but also to reason about them and to act coherently. Computer vision-based recognition systems serve the perception task, but they also go beyond it by finding challenging applications in other fields such as video surveillance, HCI, content-based video analysis and motion capture. In this context, we propose an automatic system for real-time human action recognition. We use the Kinect sensor and the tracking system in [1] to robustly detect and track people in the scene. Next, we estimate the 3D optical flow related to the tracked people from point cloud data only and we summarize it by means of a 3D grid-based descriptor. Finally, temporal sequences of descriptors are classified with the Nearest Neighbor technique and the overall application is tested on a newly created dataset. Experimental results show the effectiveness of the proposed approach
Point Cloud Library: Three-Dimensional Object Recognition and 6DOF Pose Estimation
With the advent of new-generation depth sensors, the use of three-dimensional (3-D) data is becoming increasingly popular. As these sensors are commodity hardware and sold at low cost, a rapidly growing group of people can acquire 3- D data cheaply and in real time