7 research outputs found

    RGB-D Odometry and SLAM

    Full text link
    The emergence of modern RGB-D sensors had a significant impact in many application fields, including robotics, augmented reality (AR) and 3D scanning. They are low-cost, low-power and low-size alternatives to traditional range sensors such as LiDAR. Moreover, unlike RGB cameras, RGB-D sensors provide the additional depth information that removes the need of frame-by-frame triangulation for 3D scene reconstruction. These merits have made them very popular in mobile robotics and AR, where it is of great interest to estimate ego-motion and 3D scene structure. Such spatial understanding can enable robots to navigate autonomously without collisions and allow users to insert virtual entities consistent with the image stream. In this chapter, we review common formulations of odometry and Simultaneous Localization and Mapping (known by its acronym SLAM) using RGB-D stream input. The two topics are closely related, as the former aims to track the incremental camera motion with respect to a local map of the scene, and the latter to jointly estimate the camera trajectory and the global map with consistency. In both cases, the standard approaches minimize a cost function using nonlinear optimization techniques. This chapter consists of three main parts: In the first part, we introduce the basic concept of odometry and SLAM and motivate the use of RGB-D sensors. We also give mathematical preliminaries relevant to most odometry and SLAM algorithms. In the second part, we detail the three main components of SLAM systems: camera pose tracking, scene mapping and loop closing. For each component, we describe different approaches proposed in the literature. In the final part, we provide a brief discussion on advanced research topics with the references to the state-of-the-art.Comment: This is the pre-submission version of the manuscript that was later edited and published as a chapter in RGB-D Image Analysis and Processin

    On the Enhancement of the Localization of Autonomous Mobile Platforms

    Get PDF
    The focus of many industrial and research entities on achieving full robotic autonomy increased in the past few years. In order to achieve full robotic autonomy, a fundamental problem is the localization, which is the ability of a mobile platform to determine its position and orientation in the environment. In this thesis, several problems related to the localization of autonomous platforms are addressed, namely, visual odometry accuracy and robustness; uncertainty estimation in odometries; and accurate multi-sensor fusion-based localization. Beside localization, the control of mobile manipulators is also tackled in this thesis. First, a generic image processing pipeline is proposed which, when integrated with a feature-based Visual Odometry (VO), can enhance robustness, accuracy and reduce the accumulation of errors (drift) in the pose estimation. Afterwards, since odometries (e.g. wheel odometry, LiDAR odometry, or VO) suffer from drift errors due to integration, and because such errors need to be quantified in order to achieve accurate localization through multi-sensor fusion schemes (e.g. extended or unscented kalman filters). A covariance estimation algorithm is proposed, which estimates the uncertainty of odometry measurements using another sensor which does not rely on integration. Furthermore, optimization-based multi-sensor fusion techniques are known to achieve better localization results compared to filtering techniques, but with higher computational cost. Consequently, an efficient and generic multi-sensor fusion scheme, based on Moving Horizon Estimation (MHE), is developed. The proposed multi-sensor fusion scheme: is capable of operating with any number of sensors; and considers different sensors measurements rates, missing measurements, and outliers. Moreover, the proposed multi-sensor scheme is based on a multi-threading architecture, in order to reduce its computational cost, making it more feasible for practical applications. Finally, the main purpose of achieving accurate localization is navigation. Hence, the last part of this thesis focuses on developing a stabilization controller of a 10-DOF mobile manipulator based on Model Predictive Control (MPC). All of the aforementioned works are validated using numerical simulations; real data from: EU Long-term Dataset, KITTI Dataset, TUM Dataset; and/or experimental sequences using an omni-directional mobile robot. The results show the efficacy and importance of each part of the proposed work

    ์‹ค๋‚ด ๋กœ๋ด‡์„ ์œ„ํ•œ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ฑฐ๋ฆฌ ๊ธฐ๋ก๊ณ„

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ๊ธฐ๊ณ„ํ•ญ๊ณต๊ณตํ•™๋ถ€, 2019. 2. ๊น€ํ˜„์ง„.๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์นด๋ฉ”๋ผ๋กœ๋ถ€ํ„ฐ ์ดฌ์˜๋˜๋Š” ์ผ๋ จ์˜ ์—ฐ์†์ ์ธ ์ด๋ฏธ์ง€๋“ค๋กœ๋ถ€ํ„ฐ, 3์ฐจ์› ๊ณต๊ฐ„์ƒ์—์„œ ์ž๊ธฐ ์ž์‹ ์˜ 6 ์ž์œ ๋„ ์›€์ง์ž„์„ ์ถ”์ •ํ•˜๋Š” ๊ธฐ๋ฒ•์ธ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ฑฐ๋ฆฌ ๊ธฐ๋ก๊ณ„ (VO) ๊ทธ๋ฆฌ๊ณ  ๋™์‹œ์  ์œ„์น˜ ์ธ์‹ ๋ฐ ์ง€๋„ ์ž‘์„ฑ (vSLAM) ๊ธฐ๋ฒ•๋“ค์— ๋Œ€ํ•ด ํƒ๊ตฌํ•˜์˜€๊ณ , ํŠนํžˆ ์ฃผ๋ณ€ ํ™˜๊ฒฝ ์กฐ๊ฑด์— ๋Œ€ํ•ด ๊ฐ•๊ฑดํ•˜๊ณ , ์ •ํ™•ํ•œ ์œ„์น˜ ์ถ”์ • ๊ธฐ๋ฒ•๋“ค์„ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•˜์˜€๋‹ค. ์˜์ƒ ๋‚ด ๊ฐ‘์ž‘์Šค๋Ÿฝ๊ฒŒ ๋ฐœ์ƒํ•˜๋Š” ๋น› ๋ณ€ํ™”๋ฅผ ๊ฐ„๋‹จํ•œ ์•„ํ•€ ๋ณ€ํ™” ๋ชจ๋ธ๋กœ ๋ชจ์‚ฌํ•จ์œผ๋กœ์จ ์กฐ๋„ ๋ณ€ํ™”์—๋„ ๊ฐ•๊ฑดํ•œ ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์œ„์น˜ ์ถ”์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•˜์˜€๋‹ค. ๋˜ํ•œ ์„ , ๋ฉด๊ณผ ๊ฐ™์€ ์‹ค๋‚ด ๊ตฌ์กฐ์  ํŠน์ง•๋“ค์„ ํšจ๊ณผ์ ์œผ๋กœ ๋™์‹œ์— ์ด์šฉํ•จ์œผ๋กœ์จ ๋งค์šฐ ์ •ํ™•ํ•œ ์œ„์น˜ ์ธ์‹ ๊ธฐ๋ฒ•์„ ์ƒˆ๋กญ๊ฒŒ ์—ฐ๊ตฌํ•˜์˜€๋‹ค. ์ด๋ ‡๊ฒŒ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•œ ๊ธฐ๋ฒ•๋“ค์€ ๊ฐ‘์ž‘์Šค๋Ÿฌ์šด ์กฐ๋„ ๋ณ€ํ™”๊ฐ€ ์ผ์–ด๋‚˜๋Š” ์‹ค๋‚ด ๋ฐ ์‹ค์™ธ, ๊ทธ๋ฆฌ๊ณ  ์นด๋ฉ”๋ผ์˜ ์ˆœ์ˆ˜ํ•œ ์ œ์ž๋ฆฌ ํšŒ์ „ ์šด๋™๊ณผ ๊ฐ™์€ ๋„์ „์ ์ธ ํ™˜๊ฒฝ๊ณผ ์›€์ง์ž„์—์„œ๋„ ์ •ํ™•๋„๋ฅผ ์žƒ์ง€ ์•Š๋Š”๋‹ค๋Š” ๊ฒƒ์ด ๊ฐ€์žฅ ์ค‘์š”ํ•œ ํŠน์ง•์ด๋‹ค. ์ฒซ ๋ฒˆ์งธ๋กœ, ์˜์ƒ ๋‚ด ์กฐ๋„ ๋ณ€ํ™”์— ๋Œ€ํ•ด ๊ฐ•๊ฑดํ•˜๊ฒŒ ๋Œ€์‘ํ•˜๋Š” ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ฑฐ๋ฆฌ ๊ธฐ๋ก๊ณ„ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•˜์˜€๋‹ค. ์ผ๋ฐ˜์ ์ธ ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์˜์ƒ ๋‚ด ๋™์ผ ๋ฌผ์ฒด๋Š” ๋™์ผ ๋ฐ๊ธฐ๋ฅผ ๊ฐ€์ง„๋‹ค๋Š” ๋ฐ๊ธฐ ๋ถˆ๋ณ€๋Ÿ‰ ๊ฐ€์ •์„ ๊ธฐ๋ฐ˜์œผ๋กœ ์œ„์น˜ ์ถ”์ •์„ ์ˆ˜ํ–‰ํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์กฐ๊ทธ๋งˆํ•œ ๋น› ๋ณ€ํ™”์—๋„ ๋งค์šฐ ์ทจ์•ฝํ•˜๊ณ  ๋ถˆ์•ˆ์ •ํ•œ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ค€๋‹ค. ์ œ์•ˆํ•œ ์ง์ ‘์  ๋ฐฉ์‹ ๊ธฐ๋ฐ˜์˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์˜์ƒ ๋‚ด ์—ฌ๋Ÿฌ ๊ฐœ์˜ ํŒจ์น˜๋ฅผ ์ƒ์„ฑํ•œ ๋’ค, ๊ฐ ํŒจ์น˜๋ณ„๋กœ ๋…๋ฆฝ์ ์ธ ์•„ํ•€ ๋น› ๋ณ€ํ™” ๋ชจ๋ธ์„ ์ ์šฉํ•˜๊ณ  ์ด๋ฅผ ์‹ค์‹œ๊ฐ„์œผ๋กœ ์ถ”์ •, ๋ณด์ƒํ•จ์œผ๋กœ์จ ์ „์—ญ์  ๋ฐ ๋ถ€๋ถ„์  ์กฐ๋„ ๋ณ€ํ™”์— ๋งค์šฐ ๊ฐ•๊ฑดํ•œ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ค€๋‹ค. ์ถ”๊ฐ€์ ์œผ๋กœ ํŠน์ง•์  ๊ธฐ๋ฐ˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์‚ฌ์ „ ์›€์ง์ž„์œผ๋กœ ๊ฒฐํ•ฉํ•˜๊ณ  ์ด๋ฅผ ์œ„์น˜ ์ถ”์ • ์ตœ์ ํ™” ์‹œ์— ์‚ฌ์šฉํ•˜์—ฌ, ๊ธฐ์กด ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋Œ€๋น„ ๋” ๋†’์€ ์ •ํ™•๋„์™€ ๋” ์•ˆ์ •์ ์ธ ์ž๊ฐ€ ์œ„์น˜ ์ถ”์ • ์„ฑ๋Šฅ์„ ์–ป์„ ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋‹ค์–‘ํ•˜๊ณ  ๋„์ „์ ์ธ ํ™˜๊ฒฝ์—์„œ ์ดฌ์˜๋œ ์˜์ƒ ๋ฐ์ดํ„ฐ์…‹์„ ์ด์šฉํ•˜์—ฌ, ํƒ€ ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋Œ€๋น„ ์ œ์•ˆํ•œ ๋ฐฉ๋ฒ•์˜ ํšจ์œจ์„ฑ ๋ฐ ์œ„์น˜ ์ถ”์ • ์„ฑ๋Šฅ์„ ์ •๋Ÿ‰์ ์œผ๋กœ ํ‰๊ฐ€ํ•ด๋ณด์•˜์œผ๋ฉฐ, ์ถ”๊ฐ€์ ์œผ๋กœ ์‹ค๋‚ด ๋น„ํ–‰ ๋“œ๋ก ์— ํƒ‘์žฌํ•˜์—ฌ ๋„์ „์ ์ธ ์กฐ๋„ ๋ณ€ํ™” ํ™˜๊ฒฝ ๋‚ด์—์„œ๋„ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ž์œจ ๋น„ํ–‰์ด ๊ฐ€๋Šฅํ•จ์„ ์ œ์‹œํ•˜์˜€๋‹ค. ๋‘ ๋ฒˆ์งธ๋กœ๋Š”, ์ปฌ๋Ÿฌ ๋ฐ ๊นŠ์ด ์˜์ƒ์—์„œ ๋ฐœ๊ฒฌ๋˜๋Š” ์„  ๋ฐ ๋ฉด๊ณผ ๊ฐ™์€ ๊ณต๊ฐ„ ๊ตฌ์กฐ์  ํŠน์ง•์„ ์ ๊ทน์ ์œผ๋กœ ํ™œ์šฉํ•œ ๊ณ ์ •๋ฐ€ ์˜์ƒ ๊ธฐ๋ฐ˜ ์ฃผํ–‰ ๊ธฐ๋ก๊ณ„ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ƒˆ๋กญ๊ฒŒ ์—ฐ๊ตฌํ•˜์˜€๋‹ค. ๋Œ€๋ถ€๋ถ„์˜ ์˜์ƒ ๊ธฐ๋ฐ˜ ์œ„์น˜ ์ถ”์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์€ ์‹œ๊ฐ„์ด ์ง€๋‚จ์ด ๋”ฐ๋ผ ๋ˆ„์ ๋˜๋Š” ํฐ ์œ„์น˜ ์ถ”์ • ์˜ค์ฐจ๋ฅผ ํ”ผํ•  ์ˆ˜ ์—†๋Š”๋ฐ, ์ด๋Š” ๋Œ€๋ถ€๋ถ„ ๋ถ€์ •ํ™•ํ•œ ํšŒ์ „ ์šด๋™ ์ถ”์ •์œผ๋กœ ์ธํ•ด ์œ ๋ฐœ๋œ๋‹ค๊ณ  ์•Œ๋ ค์ ธ ์žˆ๋‹ค. ์ƒˆ๋กญ๊ฒŒ ์ œ์•ˆํ•œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์‹ค๋‚ด ๊ตฌ์กฐ์  ํŠน์ง•์ธ ์„ ๊ณผ ๋ฉด์„ ์ถ”์ ํ•˜์—ฌ, ์ด๋Ÿฌํ•œ ์œ„์น˜ ์ถ”์ • ์˜ค์ฐจ์˜ ์ฃผ์š” ์›์ธ์ธ ํšŒ์ „์šด๋™์„ ๋งค์šฐ ์ •ํ™•ํ•˜๊ณ  ๋“œ๋ฆฌํ”„ํŠธ ์—†์ด ์ถ”์ •ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์šฐ๋ฆฌ๋Š” ํšจ์œจ์ ์ธ SO(3) ๊ณต๊ฐ„์ƒ์—์„œ ์ œํ•œ๋œ ํ‰๊ท  ์ด๋™ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์‚ฌ์šฉํ•˜์—ฌ ํ™˜๊ฒฝ์˜ ๊ตฌ์กฐ์  ๊ทœ์น™์„ฑ์„ ์ธ์ง€ํ•˜๊ณ  ์ถ”์ ํ•˜์˜€๋‹ค. ์ด๋ ‡๊ฒŒ ์ •ํ™•ํ•œ ์นด๋ฉ”๋ผ์˜ ํšŒ์ „ ์›€์ง์ž„์ด ์–ป์–ด์ง„ ํ›„, ์ด๋ฅผ ์ด์šฉํ•˜์—ฌ ํšŒ์ „ ์„ฑ๋ถ„์ด ์ œ๊ฑฐ๋œ ์žฌํˆฌ์˜ ์˜ค์ฐจ๋ฅผ ์ตœ์†Œํ™”ํ•˜์—ฌ ์นด๋ฉ”๋ผ์˜ ๋ณ‘์ง„ ์›€์ง์ž„์„ ์ถ”์ •ํ•œ๋‹ค. ์ œ์•ˆ๋œ ์œ„์น˜ ์ถ”์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์ˆœ์ˆ˜ ํšŒ์ „ ์›€์ง์ž„๊ณผ ๊ฐ™์ด ์ถ”์ •ํ•˜๊ธฐ ์–ด๋ ค์šด ์นด๋ฉ”๋ผ ์›€์ง์ž„์ด ํฌํ•จ๋œ ๋‹ค์–‘ํ•œ ์˜์ƒ ๋ฐ์ดํ„ฐ์…‹์—์„œ ํ‰๊ฐ€๋˜์—ˆ์œผ๋ฉฐ, ํŠนํžˆ ํƒ€ ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋Œ€๋น„ ๋งค์šฐ ๋†’์€ ์ •ํ™•์„ฑ๊ณผ ๊ฐ•๊ฑด์„ฑ, ๊ทธ๋ฆฌ๊ณ  ๋‚ฎ์€ ๋“œ๋ฆฌํ”„ํŠธ ์˜ค์ฐจ๋ฅผ ์‹คํ—˜์ ์œผ๋กœ ๋ณด์—ฌ์ฃผ์—ˆ๋‹ค.This thesis explores the robust and accurate 6-DoF camera motion estimation from a sequence of images, called visual odometry (VO) or visual simultaneous localization and mapping (vSLAM). We focus on the robustness and high accuracy of the VO and visual localization by explicitly modeling the light changes as an affine illumination model, and utilizing the indoor environmental structures such as lines and planes. This brings the significant advantage to VO that it does not lose estimation accuracy in challenging environments such as light-changing conditions or pure, on the spot rotations. The first part of the thesis proposes a novel patch-based illumination invariant visual odometry algorithm (PIVO). PIVO employs an affine illumination change model per each patch in the image to compensate unexpected, abrupt, and irregular illumination changes during the direct motion estimation. PIVO infers camera geometry directly from the images, i.e., the raw sensor measurements, without intermediate abstraction, for instance in the form of keypoint matches. We furthermore incorporate a motion prior from feature-based stereo visual odometry in the optimization, resulting in higher accuracy and more stable motion estimates. We evaluate the proposed VO algorithm on a variety of datasets, and demonstrate autonomous flight experiments with an aerial robot, showing that the proposed method successfully estimates 6-DoF pose under significant illumination changes. In the second part of the thesis, we propose a low-drift VO that separately estimates rotational and translational motion from lines, planes, and points found in RGB-D images. To estimate the rotational motion that is a main source of drift in VO in an accurate and drift-free manner, we exploit both lines and planes jointly from environmental regularities. We recognize and track the structural regularities with an efficient SO(3)-manifold constrained mean shift algorithm. Once the absolute camera orientation is found, we recover the translational motion from all tracked points with and without depth by minimizing the de-rotated reprojection error. We compare the proposed algorithm to other state-of-the-art VO methods on a variety of RGB-D datasets that include especially challenging pure rotations, and demonstrate improved accuracy and lower drift error.1 Introduction 2 Background 3 Robust Visual Odometry to Irregular Illumination Changes with RGB-D Camera 4 Robust Visual Localization in Changing Lighting Conditions 5 Autonomous Flight with Robust Visual Odometry under Dynamic Lighting Conditions 6 Visual Odometry with Drift-Free Rotation Estimation Using Indoor Scene Regularities 7 Low-Drift Visual Odometry in Structured Environments by Decoupling Rotational and Translational Motion 8 Indoor RGB-D Compass from a Single Line and Plane 9 Linear RGB-D SLAM for Planar Environments 10 ConclusionDocto
    corecore