8 research outputs found

    A hybrid visual-based SLAM architecture: local filter-based SLAM with keyframe-based global mapping

    Get PDF
    This work presents a hybrid visual-based SLAM architecture that aims to take advantage of the strengths of each of the two main methodologies currently available for implementing visual-based SLAM systems, while at the same time minimizing some of their drawbacks. The main idea is to implement a local SLAM process using a filter-based technique, and enable the tasks of building and maintaining a consistent global map of the environment, including the loop closure problem, to use the processes implemented using optimization-based techniques. Different variants of visual-based SLAM systems can be implemented using the proposed architecture. This work also presents the implementation case of a full monocular-based SLAM system for unmanned aerial vehicles that integrates additional sensory inputs. Experiments using real data obtained from the sensors of a quadrotor are presented to validate the feasibility of the proposed approachPostprint (published version

    RGB-D Odometry and SLAM

    Full text link
    The emergence of modern RGB-D sensors had a significant impact in many application fields, including robotics, augmented reality (AR) and 3D scanning. They are low-cost, low-power and low-size alternatives to traditional range sensors such as LiDAR. Moreover, unlike RGB cameras, RGB-D sensors provide the additional depth information that removes the need of frame-by-frame triangulation for 3D scene reconstruction. These merits have made them very popular in mobile robotics and AR, where it is of great interest to estimate ego-motion and 3D scene structure. Such spatial understanding can enable robots to navigate autonomously without collisions and allow users to insert virtual entities consistent with the image stream. In this chapter, we review common formulations of odometry and Simultaneous Localization and Mapping (known by its acronym SLAM) using RGB-D stream input. The two topics are closely related, as the former aims to track the incremental camera motion with respect to a local map of the scene, and the latter to jointly estimate the camera trajectory and the global map with consistency. In both cases, the standard approaches minimize a cost function using nonlinear optimization techniques. This chapter consists of three main parts: In the first part, we introduce the basic concept of odometry and SLAM and motivate the use of RGB-D sensors. We also give mathematical preliminaries relevant to most odometry and SLAM algorithms. In the second part, we detail the three main components of SLAM systems: camera pose tracking, scene mapping and loop closing. For each component, we describe different approaches proposed in the literature. In the final part, we provide a brief discussion on advanced research topics with the references to the state-of-the-art.Comment: This is the pre-submission version of the manuscript that was later edited and published as a chapter in RGB-D Image Analysis and Processin

    Online Synthesis Of Speculative Building Information Models For Robot Motion Planning

    Get PDF
    Autonomous mobile robots today still lack the necessary understanding of indoor environments for making informed decisions about the state of the world beyond their immediate field of view. As a result, they are forced to make conservative and often inaccurate assumptions about unexplored space, inhibiting the degree of performance being increasingly expected of them in the areas of high-speed navigation and mission planning. In order to address this limitation, this thesis explores the use of Building Information Models (BIMs) for providing the existing ecosystem of local and global planning algorithms with informative compact higher-level representations of indoor environments. Although BIMs have long been used in architecture, engineering, and construction for a number of different purposes, to our knowledge, this is the first instance of them being used in robotics. Given the technical constraints accompanying this domain, including a limited and incomplete set of observations which grows over time, the systems we present are designed such that together they produce BIMs capable of providing explanations of both the explored and unexplored space in an online fashion. The first is a SLAM system that uses the structural regularity of buildings in order to mitigate drift and provide the simplest explanation of architectural features such as floors, walls, and ceilings. The planar model generated is then passed to a secondary system that then reasons about their mutual relationships in order to provide a water-tight model of the observed and inferred freespace. Our experimental results demonstrate this to be an accurate and efficient approach towards this end

    ์‹ค๋‚ด ๋กœ๋ด‡์„ ์œ„ํ•œ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ฑฐ๋ฆฌ ๊ธฐ๋ก๊ณ„

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ๊ธฐ๊ณ„ํ•ญ๊ณต๊ณตํ•™๋ถ€, 2019. 2. ๊น€ํ˜„์ง„.๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์นด๋ฉ”๋ผ๋กœ๋ถ€ํ„ฐ ์ดฌ์˜๋˜๋Š” ์ผ๋ จ์˜ ์—ฐ์†์ ์ธ ์ด๋ฏธ์ง€๋“ค๋กœ๋ถ€ํ„ฐ, 3์ฐจ์› ๊ณต๊ฐ„์ƒ์—์„œ ์ž๊ธฐ ์ž์‹ ์˜ 6 ์ž์œ ๋„ ์›€์ง์ž„์„ ์ถ”์ •ํ•˜๋Š” ๊ธฐ๋ฒ•์ธ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ฑฐ๋ฆฌ ๊ธฐ๋ก๊ณ„ (VO) ๊ทธ๋ฆฌ๊ณ  ๋™์‹œ์  ์œ„์น˜ ์ธ์‹ ๋ฐ ์ง€๋„ ์ž‘์„ฑ (vSLAM) ๊ธฐ๋ฒ•๋“ค์— ๋Œ€ํ•ด ํƒ๊ตฌํ•˜์˜€๊ณ , ํŠนํžˆ ์ฃผ๋ณ€ ํ™˜๊ฒฝ ์กฐ๊ฑด์— ๋Œ€ํ•ด ๊ฐ•๊ฑดํ•˜๊ณ , ์ •ํ™•ํ•œ ์œ„์น˜ ์ถ”์ • ๊ธฐ๋ฒ•๋“ค์„ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•˜์˜€๋‹ค. ์˜์ƒ ๋‚ด ๊ฐ‘์ž‘์Šค๋Ÿฝ๊ฒŒ ๋ฐœ์ƒํ•˜๋Š” ๋น› ๋ณ€ํ™”๋ฅผ ๊ฐ„๋‹จํ•œ ์•„ํ•€ ๋ณ€ํ™” ๋ชจ๋ธ๋กœ ๋ชจ์‚ฌํ•จ์œผ๋กœ์จ ์กฐ๋„ ๋ณ€ํ™”์—๋„ ๊ฐ•๊ฑดํ•œ ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์œ„์น˜ ์ถ”์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•˜์˜€๋‹ค. ๋˜ํ•œ ์„ , ๋ฉด๊ณผ ๊ฐ™์€ ์‹ค๋‚ด ๊ตฌ์กฐ์  ํŠน์ง•๋“ค์„ ํšจ๊ณผ์ ์œผ๋กœ ๋™์‹œ์— ์ด์šฉํ•จ์œผ๋กœ์จ ๋งค์šฐ ์ •ํ™•ํ•œ ์œ„์น˜ ์ธ์‹ ๊ธฐ๋ฒ•์„ ์ƒˆ๋กญ๊ฒŒ ์—ฐ๊ตฌํ•˜์˜€๋‹ค. ์ด๋ ‡๊ฒŒ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•œ ๊ธฐ๋ฒ•๋“ค์€ ๊ฐ‘์ž‘์Šค๋Ÿฌ์šด ์กฐ๋„ ๋ณ€ํ™”๊ฐ€ ์ผ์–ด๋‚˜๋Š” ์‹ค๋‚ด ๋ฐ ์‹ค์™ธ, ๊ทธ๋ฆฌ๊ณ  ์นด๋ฉ”๋ผ์˜ ์ˆœ์ˆ˜ํ•œ ์ œ์ž๋ฆฌ ํšŒ์ „ ์šด๋™๊ณผ ๊ฐ™์€ ๋„์ „์ ์ธ ํ™˜๊ฒฝ๊ณผ ์›€์ง์ž„์—์„œ๋„ ์ •ํ™•๋„๋ฅผ ์žƒ์ง€ ์•Š๋Š”๋‹ค๋Š” ๊ฒƒ์ด ๊ฐ€์žฅ ์ค‘์š”ํ•œ ํŠน์ง•์ด๋‹ค. ์ฒซ ๋ฒˆ์งธ๋กœ, ์˜์ƒ ๋‚ด ์กฐ๋„ ๋ณ€ํ™”์— ๋Œ€ํ•ด ๊ฐ•๊ฑดํ•˜๊ฒŒ ๋Œ€์‘ํ•˜๋Š” ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ฑฐ๋ฆฌ ๊ธฐ๋ก๊ณ„ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•˜์˜€๋‹ค. ์ผ๋ฐ˜์ ์ธ ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์˜์ƒ ๋‚ด ๋™์ผ ๋ฌผ์ฒด๋Š” ๋™์ผ ๋ฐ๊ธฐ๋ฅผ ๊ฐ€์ง„๋‹ค๋Š” ๋ฐ๊ธฐ ๋ถˆ๋ณ€๋Ÿ‰ ๊ฐ€์ •์„ ๊ธฐ๋ฐ˜์œผ๋กœ ์œ„์น˜ ์ถ”์ •์„ ์ˆ˜ํ–‰ํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์กฐ๊ทธ๋งˆํ•œ ๋น› ๋ณ€ํ™”์—๋„ ๋งค์šฐ ์ทจ์•ฝํ•˜๊ณ  ๋ถˆ์•ˆ์ •ํ•œ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ค€๋‹ค. ์ œ์•ˆํ•œ ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์˜์ƒ ๋‚ด ์—ฌ๋Ÿฌ ๊ฐœ์˜ ํŒจ์น˜๋ฅผ ์ƒ์„ฑํ•œ ๋’ค, ๊ฐ ํŒจ์น˜๋ณ„๋กœ ๋…๋ฆฝ์ ์ธ ์•„ํ•€ ๋น› ๋ณ€ํ™” ๋ชจ๋ธ์„ ์ ์šฉํ•˜๊ณ  ์ด๋ฅผ ์‹ค์‹œ๊ฐ„์œผ๋กœ ์ถ”์ •, ๋ณด์ƒํ•จ์œผ๋กœ์จ ์ „์—ญ์  ๋ฐ ๋ถ€๋ถ„์  ์กฐ๋„ ๋ณ€ํ™”์— ๋งค์šฐ ๊ฐ•๊ฑดํ•œ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ค€๋‹ค. ์ถ”๊ฐ€์ ์œผ๋กœ ํŠน์ง•์  ๊ธฐ๋ฐ˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์‚ฌ์ „ ์›€์ง์ž„์œผ๋กœ ๊ฒฐํ•ฉํ•˜๊ณ  ์ด๋ฅผ ์œ„์น˜ ์ถ”์ • ์ตœ์ ํ™” ์‹œ์— ์‚ฌ์šฉํ•˜์—ฌ, ๊ธฐ์กด ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋Œ€๋น„ ๋” ๋†’์€ ์ •ํ™•๋„์™€ ๋” ์•ˆ์ •์ ์ธ ์ž๊ฐ€ ์œ„์น˜ ์ถ”์ • ์„ฑ๋Šฅ์„ ์–ป์„ ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋‹ค์–‘ํ•˜๊ณ  ๋„์ „์ ์ธ ํ™˜๊ฒฝ์—์„œ ์ดฌ์˜๋œ ์˜์ƒ ๋ฐ์ดํ„ฐ์…‹์„ ์ด์šฉํ•˜์—ฌ, ํƒ€ ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋Œ€๋น„ ์ œ์•ˆํ•œ ๋ฐฉ๋ฒ•์˜ ํšจ์œจ์„ฑ ๋ฐ ์œ„์น˜ ์ถ”์ • ์„ฑ๋Šฅ์„ ์ •๋Ÿ‰์ ์œผ๋กœ ํ‰๊ฐ€ํ•ด๋ณด์•˜์œผ๋ฉฐ, ์ถ”๊ฐ€์ ์œผ๋กœ ์‹ค๋‚ด ๋น„ํ–‰ ๋“œ๋ก ์— ํƒ‘์žฌํ•˜์—ฌ ๋„์ „์ ์ธ ์กฐ๋„ ๋ณ€ํ™” ํ™˜๊ฒฝ ๋‚ด์—์„œ๋„ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ž์œจ ๋น„ํ–‰์ด ๊ฐ€๋Šฅํ•จ์„ ์ œ์‹œํ•˜์˜€๋‹ค. ๋‘ ๋ฒˆ์งธ๋กœ๋Š”, ์ปฌ๋Ÿฌ ๋ฐ ๊นŠ์ด ์˜์ƒ์—์„œ ๋ฐœ๊ฒฌ๋˜๋Š” ์„  ๋ฐ ๋ฉด๊ณผ ๊ฐ™์€ ๊ณต๊ฐ„ ๊ตฌ์กฐ์  ํŠน์ง•์„ ์ ๊ทน์ ์œผ๋กœ ํ™œ์šฉํ•œ ๊ณ ์ •๋ฐ€ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ธฐ๋ก๊ณ„ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ƒˆ๋กญ๊ฒŒ ์—ฐ๊ตฌํ•˜์˜€๋‹ค. ๋Œ€๋ถ€๋ถ„์˜ ์˜์ƒ ๊ธฐ๋ฐ˜ ์œ„์น˜ ์ถ”์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์€ ์‹œ๊ฐ„์ด ์ง€๋‚จ์ด ๋”ฐ๋ผ ๋ˆ„์ ๋˜๋Š” ํฐ ์œ„์น˜ ์ถ”์ • ์˜ค์ฐจ๋ฅผ ํ”ผํ•  ์ˆ˜ ์—†๋Š”๋ฐ, ์ด๋Š” ๋Œ€๋ถ€๋ถ„ ๋ถ€์ •ํ™•ํ•œ ํšŒ์ „ ์šด๋™ ์ถ”์ •์œผ๋กœ ์ธํ•ด ์œ ๋ฐœ๋œ๋‹ค๊ณ  ์•Œ๋ ค์ ธ ์žˆ๋‹ค. ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์‹ค๋‚ด ๊ตฌ์กฐ์  ํŠน์ง•์ธ ์„ ๊ณผ ๋ฉด์„ ์ถ”์ ํ•˜์—ฌ, ์ด๋Ÿฌํ•œ ์œ„์น˜ ์ถ”์ • ์˜ค์ฐจ์˜ ์ฃผ์š” ์›์ธ์ธ ํšŒ์ „์šด๋™์„ ๋งค์šฐ ์ •ํ™•ํ•˜๊ณ  ๋“œ๋ฆฌํ”„ํŠธ ์—†์ด ์ถ”์ •ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์šฐ๋ฆฌ๋Š” ํšจ์œจ์ ์ธ SO(3) ๊ณต๊ฐ„์ƒ์—์„œ ์ œํ•œ๋œ ํ‰๊ท  ์ด๋™ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์‚ฌ์šฉํ•˜์—ฌ ํ™˜๊ฒฝ์˜ ๊ตฌ์กฐ์  ๊ทœ์น™์„ฑ์„ ์ธ์ง€ํ•˜๊ณ  ์ถ”์ ํ•˜์˜€๋‹ค. ์ด๋ ‡๊ฒŒ ์ •ํ™•ํ•œ ์นด๋ฉ”๋ผ์˜ ํšŒ์ „ ์›€์ง์ž„์ด ์–ป์–ด์ง„ ํ›„, ์ด๋ฅผ ์ด์šฉํ•˜์—ฌ ํšŒ์ „ ์„ฑ๋ถ„์ด ์ œ๊ฑฐ๋œ ์žฌํˆฌ์˜ ์˜ค์ฐจ๋ฅผ ์ตœ์†Œํ™”ํ•˜์—ฌ ์นด๋ฉ”๋ผ์˜ ๋ณ‘์ง„ ์›€์ง์ž„์„ ์ถ”์ •ํ•œ๋‹ค. ์ œ์•ˆ๋œ ์œ„์น˜ ์ถ”์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์ˆœ์ˆ˜ ํšŒ์ „ ์›€์ง์ž„๊ณผ ๊ฐ™์ด ์ถ”์ •ํ•˜๊ธฐ ์–ด๋ ค์šด ์นด๋ฉ”๋ผ ์›€์ง์ž„์ด ํฌํ•จ๋œ ๋‹ค์–‘ํ•œ ์˜์ƒ ๋ฐ์ดํ„ฐ์…‹์—์„œ ํ‰๊ฐ€๋˜์—ˆ์œผ๋ฉฐ, ํŠนํžˆ ํƒ€ ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋Œ€๋น„ ๋งค์šฐ ๋†’์€ ์ •ํ™•์„ฑ๊ณผ ๊ฐ•๊ฑด์„ฑ, ๊ทธ๋ฆฌ๊ณ  ๋‚ฎ์€ ๋“œ๋ฆฌํ”„ํŠธ ์˜ค์ฐจ๋ฅผ ์‹คํ—˜์ ์œผ๋กœ ๋ณด์—ฌ์ฃผ์—ˆ๋‹ค.This thesis explores the robust and accurate 6-DoF camera motion estimation from a sequence of images, called visual odometry (VO) or visual simultaneous localization and mapping (vSLAM). We focus on the robustness and high accuracy of the VO and visual localization by explicitly modeling the light changes as an affine illumination model, and utilizing the indoor environmental structures such as lines and planes. This brings the significant advantage to VO that it does not lose estimation accuracy in challenging environments such as light-changing conditions or pure, on the spot rotations. The first part of the thesis proposes a novel patch-based illumination invariant visual odometry algorithm (PIVO). PIVO employs an affine illumination change model per each patch in the image to compensate unexpected, abrupt, and irregular illumination changes during the direct motion estimation. PIVO infers camera geometry directly from the images, i.e., the raw sensor measurements, without intermediate abstraction, for instance in the form of keypoint matches. We furthermore incorporate a motion prior from feature-based stereo visual odometry in the optimization, resulting in higher accuracy and more stable motion estimates. We evaluate the proposed VO algorithm on a variety of datasets, and demonstrate autonomous flight experiments with an aerial robot, showing that the proposed method successfully estimates 6-DoF pose under significant illumination changes. In the second part of the thesis, we propose a low-drift VO that separately estimates rotational and translational motion from lines, planes, and points found in RGB-D images. To estimate the rotational motion that is a main source of drift in VO in an accurate and drift-free manner, we exploit both lines and planes jointly from environmental regularities. We recognize and track the structural regularities with an efficient SO(3)-manifold constrained mean shift algorithm. Once the absolute camera orientation is found, we recover the translational motion from all tracked points with and without depth by minimizing the de-rotated reprojection error. We compare the proposed algorithm to other state-of-the-art VO methods on a variety of RGB-D datasets that include especially challenging pure rotations, and demonstrate improved accuracy and lower drift error.1 Introduction 2 Background 3 Robust Visual Odometry to Irregular Illumination Changes with RGB-D Camera 4 Robust Visual Localization in Changing Lighting Conditions 5 Autonomous Flight with Robust Visual Odometry under Dynamic Lighting Conditions 6 Visual Odometry with Drift-Free Rotation Estimation Using Indoor Scene Regularities 7 Low-Drift Visual Odometry in Structured Environments by Decoupling Rotational and Translational Motion 8 Indoor RGB-D Compass from a Single Line and Plane 9 Linear RGB-D SLAM for Planar Environments 10 ConclusionDocto
    corecore