841 research outputs found

    Integration of Absolute Orientation Measurements in the KinectFusion Reconstruction pipeline

    Full text link
    In this paper, we show how absolute orientation measurements provided by low-cost but high-fidelity IMU sensors can be integrated into the KinectFusion pipeline. We show that integration improves both runtime, robustness and quality of the 3D reconstruction. In particular, we use this orientation data to seed and regularize the ICP registration technique. We also present a technique to filter the pairs of 3D matched points based on the distribution of their distances. This filter is implemented efficiently on the GPU. Estimating the distribution of the distances helps control the number of iterations necessary for the convergence of the ICP algorithm. Finally, we show experimental results that highlight improvements in robustness, a speed-up of almost 12%, and a gain in tracking quality of 53% for the ATE metric on the Freiburg benchmark.Comment: CVPR Workshop on Visual Odometry and Computer Vision Applications Based on Location Clues 201

    Exploitation of time-of-flight (ToF) cameras

    Get PDF
    This technical report reviews the state-of-the art in the field of ToF cameras, their advantages, their limitations, and their present-day applications sometimes in combination with other sensors. Even though ToF cameras provide neither higher resolution nor larger ambiguity-free range compared to other range map estimation systems, advantages such as registered depth and intensity data at a high frame rate, compact design, low weight and reduced power consumption have motivated their use in numerous areas of research. In robotics, these areas range from mobile robot navigation and map building to vision-based human motion capture and gesture recognition, showing particularly a great potential in object modeling and recognition.Preprin

    Performance Evaluation of Two Indoor Mapping Systems: Low-Cost UWB-Aided Photogrammetry and Backpack Laser Scanning

    Get PDF
    During the past dozen years, several mobile mapping systems based on the use of imaging and positioning sensors mounted on terrestrial (and aerial) vehicles have been developed. Recently, systems characterized by an increased portability have been proposed in order to enable mobile mapping in environments that are difficult to access for vehicles, in particular for indoor environments. In this work the performance of a low-cost mobile mapping system is compared with that of: (i) a state-of-the-art terrestrial laser scanning (TLS), considered as the control; (ii) a mobile mapping backpack system (Leica Pegasus), which can be considered as the state-of-the-art of commercial mobile mapping backpack systems. The aim of this paper is two-fold: first, assessing the reconstruction accuracy of the proposed low-cost mobile mapping system, based on photogrammetry and ultra-wide band (UWB) for relative positioning (and a GNSS receiver if georeferencing is needed), with respect to a TLS survey in an indoor environment, where the global navigation satellite system (GNSS) signal is not available; second, comparing such performance with that obtained with the Leica backpack. Both mobile mapping systems are designed to work without any control point, to enable an easy and quick survey (e.g., few minutes) and to be easily portable (relatively low weight and small size). The case study deals with the 3D reconstruction of a medieval bastion in Padua, Italy. Reconstruction using the Leica Pegasus backpack allowed obtaining a smaller absolute error with respect to the UWB-based photogrammetric system. In georeferenced coordinates, the root mean square (RMS) error was respectively 16.1 cm and 50.3 cm; relative error in local coordinates was more similar, respectively 8.2 cm and 6.1 cm. Given the much lower cost (approximately $6k), the proposed photogrammetric-based system can be an interesting alternative when decimetric reconstruction accuracy in georeferenced coordinates is sufficient. \ua9 2018 by the authors

    Depth sensors in augmented reality solutions. Literature review

    Get PDF
    The emergence of depth sensors has made it possible to track – not only monocular cues – but also the actual depth values of the environment. This is especially useful in augmented reality solutions, where the position and orientation (pose) of the observer need to be accurately determined. This allows virtual objects to be installed to the view of the user through, for example, a screen of a tablet or augmented reality glasses (e.g. Google glass, etc.). Although the early 3D sensors have been physically quite large, the size of these sensors is decreasing, and possibly – eventually – a 3D sensor could be embedded – for example – to augmented reality glasses. The wider subject area considered in this review is 3D SLAM methods, which take advantage of the 3D information available by modern RGB-D sensors, such as Microsoft Kinect. Thus the review for SLAM (Simultaneous Localization and Mapping) and 3D tracking in augmented reality is a timely subject. We also try to find out the limitations and possibilities of different tracking methods, and how they should be improved, in order to allow efficient integration of the methods to the augmented reality solutions of the future.Siirretty Doriast

    Depth sensors in augmented reality solutions. Literature review

    Get PDF
    The emergence of depth sensors has made it possible to track – not only monocular cues – but also the actual depth values of the environment. This is especially useful in augmented reality solutions, where the position and orientation (pose) of the observer need to be accurately determined. This allows virtual objects to be installed to the view of the user through, for example, a screen of a tablet or augmented reality glasses (e.g. Google glass, etc.). Although the early 3D sensors have been physically quite large, the size of these sensors is decreasing, and possibly – eventually – a 3D sensor could be embedded – for example – to augmented reality glasses. The wider subject area considered in this review is 3D SLAM methods, which take advantage of the 3D information available by modern RGB-D sensors, such as Microsoft Kinect. Thus the review for SLAM (Simultaneous Localization and Mapping) and 3D tracking in augmented reality is a timely subject. We also try to find out the limitations and possibilities of different tracking methods, and how they should be improved, in order to allow efficient integration of the methods to the augmented reality solutions of the future.Siirretty Doriast

    MICP-L: Mesh-based ICP for Robot Localization using Hardware-Accelerated Ray Casting

    Full text link
    Triangle mesh maps have proven to be a versatile 3D environment representation for robots to navigate in challenging indoor and outdoor environments exhibiting tunnels, hills and varying slopes. To make use of these mesh maps, methods are needed that allow robots to accurately localize themselves to perform typical tasks like path planning and navigation. We present Mesh ICP Localization (MICP-L), a novel and computationally efficient method for registering one or more range sensors to a triangle mesh map to continuously localize a robot in 6D, even in GPS-denied environments. We accelerate the computation of ray casting correspondences (RCC) between range sensors and mesh maps by supporting different parallel computing devices like multicore CPUs, GPUs and the latest NVIDIA RTX hardware. By additionally transforming the covariance computation into a reduction operation, we can optimize the initial guessed poses in parallel on CPUs or GPUs, making our implementation applicable in real-time on a variety of target architectures. We demonstrate the robustness of our localization approach with datasets from agriculture, drones, and automotive domains

    Toward Automated Aerial Refueling: Relative Navigation with Structure from Motion

    Get PDF
    The USAF\u27s use of UAS has expanded from reconnaissance to hunter/killer missions. As the UAS mission further expands into aerial combat, better performance and larger payloads will have a negative correlation with range and loiter times. Additionally, the Air Force Future Operating Concept calls for \formations of uninhabited refueling aircraft...[that] enable refueling operations partway inside threat areas. However, a lack of accurate relative positioning information prevents the ability to safely maintain close formation flight and contact between a tanker and a UAS. The inclusion of cutting edge vision systems on present refueling platforms may provide the information necessary to support a AAR mission by estimating the position of a trailing aircraft to provide inputs to a UAS controller capable of maintaining a given position. This research examines the ability of SfM to generate relative navigation information. Previous AAR research efforts involved the use of differential GPS, LiDAR, and vision systems. This research aims to leverage current and future imaging technology to compliment these solutions. The algorithm used in this thesis generates a point cloud by determining 3D structure from a sequence of 2D images. The algorithm then utilizes PCA to register the point cloud to a reference model. The algorithm was tested in a real world environment using a 1:7 scale F-15 model. Additionally, this thesis studies common 3D rigid registration algorithms in an effort characterize their performance in the AAR domain. Three algorithms are tested for runtime and registration accuracy with four data sets
    • …
    corecore