595 research outputs found

    Efficient 3D Segmentation, Registration and Mapping for Mobile Robots

    Get PDF
    Sometimes simple is better! For certain situations and tasks, simple but robust methods can achieve the same or better results in the same or less time than related sophisticated approaches. In the context of robots operating in real-world environments, key challenges are perceiving objects of interest and obstacles as well as building maps of the environment and localizing therein. The goal of this thesis is to carefully analyze such problem formulations, to deduce valid assumptions and simplifications, and to develop simple solutions that are both robust and fast. All approaches make use of sensors capturing 3D information, such as consumer RGBD cameras. Comparative evaluations show the performance of the developed approaches. For identifying objects and regions of interest in manipulation tasks, a real-time object segmentation pipeline is proposed. It exploits several common assumptions of manipulation tasks such as objects being on horizontal support surfaces (and well separated). It achieves real-time performance by using particularly efficient approximations in the individual processing steps, subsampling the input data where possible, and processing only relevant subsets of the data. The resulting pipeline segments 3D input data with up to 30Hz. In order to obtain complete segmentations of the 3D input data, a second pipeline is proposed that approximates the sampled surface, smooths the underlying data, and segments the smoothed surface into coherent regions belonging to the same geometric primitive. It uses different primitive models and can reliably segment input data into planes, cylinders and spheres. A thorough comparative evaluation shows state-of-the-art performance while computing such segmentations in near real-time. The second part of the thesis addresses the registration of 3D input data, i.e., consistently aligning input captured from different view poses. Several methods are presented for different types of input data. For the particular application of mapping with micro aerial vehicles where the 3D input data is particularly sparse, a pipeline is proposed that uses the same approximate surface reconstruction to exploit the measurement topology and a surface-to-surface registration algorithm that robustly aligns the data. Optimization of the resulting graph of determined view poses then yields globally consistent 3D maps. For sequences of RGBD data this pipeline is extended to include additional subsampling steps and an initial alignment of the data in local windows in the pose graph. In both cases, comparative evaluations show a robust and fast alignment of the input data

    Nonholonomic Motion Planning for Automated Vehicles in Dense Scenarios

    Get PDF

    Fault-Tolerant Vision for Vehicle Guidance in Agriculture

    Get PDF

    Feasible, Robust and Reliable Automation and Control for Autonomous Systems

    Get PDF
    The Special Issue book focuses on highlighting current research and developments in the automation and control field for autonomous systems as well as showcasing state-of-the-art control strategy approaches for autonomous platforms. The book is co-edited by distinguished international control system experts currently based in Sweden, the United States of America, and the United Kingdom, with contributions from reputable researchers from China, Austria, France, the United States of America, Poland, and Hungary, among many others. The editors believe the ten articles published within this Special Issue will be highly appealing to control-systems-related researchers in applications typified in the fields of ground, aerial, maritime vehicles, and robotics as well as industrial audiences

    Model-Based Environmental Visual Perception for Humanoid Robots

    Get PDF
    The visual perception of a robot should answer two fundamental questions: What? and Where? In order to properly and efficiently reply to these questions, it is essential to establish a bidirectional coupling between the external stimuli and the internal representations. This coupling links the physical world with the inner abstraction models by sensor transformation, recognition, matching and optimization algorithms. The objective of this PhD is to establish this sensor-model coupling

    Challenges in 3D scanning: Focusing on Ears and Multiple View Stereopsis

    Get PDF

    Mapping of complex marine environments using an unmanned surface craft

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2011.Cataloged from PDF version of thesis.Includes bibliographical references (p. 185-199).Recent technology has combined accurate GPS localization with mapping to build 3D maps in a diverse range of terrestrial environments, but the mapping of marine environments lags behind. This is particularly true in shallow water and coastal areas with man-made structures such as bridges, piers, and marinas, which can pose formidable challenges to autonomous underwater vehicle (AUV) operations. In this thesis, we propose a new approach for mapping shallow water marine environments, combining data from both above and below the water in a robust probabilistic state estimation framework. The ability to rapidly acquire detailed maps of these environments would have many applications, including surveillance, environmental monitoring, forensic search, and disaster recovery. Whereas most recent AUV mapping research has been limited to open waters, far from man-made surface structures, in our work we focus on complex shallow water environments, such as rivers and harbors, where man-made structures block GPS signals and pose hazards to navigation. Our goal is to enable an autonomous surface craft to combine data from the heterogeneous environments above and below the water surface - as if the water were drained, and we had a complete integrated model of the marine environment, with full visibility. To tackle this problem, we propose a new framework for 3D SLAM in marine environments that combines data obtained concurrently from above and below the water in a robust probabilistic state estimation framework. Our work makes systems, algorithmic, and experimental contributions in perceptual robotics for the marine environment. We have created a novel Autonomous Surface Vehicle (ASV), equipped with substantial onboard computation and an extensive sensor suite that includes three SICK lidars, a Blueview MB2250 imaging sonar, a Doppler Velocity Log, and an integrated global positioning system/inertial measurement unit (GPS/IMU) device. The data from these sensors is processed in a hybrid metric/topological SLAM state estimation framework. A key challenge to mapping is extracting effective constraints from 3D lidar data despite GPS loss and reacquisition. This was achieved by developing a GPS trust engine that uses a semi-supervised learning classifier to ascertain the validity of GPS information for different segments of the vehicle trajectory. This eliminates the troublesome effects of multipath on the vehicle trajectory estimate, and provides cues for submap decomposition. Localization from lidar point clouds is performed using octrees combined with Iterative Closest Point (ICP) matching, which provides constraints between submaps both within and across different mapping sessions. Submap positions are optimized via least squares optimization of the graph of constraints, to achieve global alignment. The global vehicle trajectory is used for subsea sonar bathymetric map generation and for mesh reconstruction from lidar data for 3D visualization of above-water structures. We present experimental results in the vicinity of several structures spanning or along the Charles River between Boston and Cambridge, MA. The Harvard and Longfellow Bridges, three sailing pavilions and a yacht club provide structures of interest, having both extensive superstructure and subsurface foundations. To quantitatively assess the mapping error, we compare against a georeferenced model of the Harvard Bridge using blueprints from the Library of Congress. Our results demonstrate the potential of this new approach to achieve robust and efficient model capture for complex shallow-water marine environments. Future work aims to incorporate autonomy for path planning of a region of interest while performing collision avoidance to enable fully autonomous surveys that achieve full sensor coverage of a complete marine environment.by Jacques Chadwick Leedekerken.Ph.D

    Trajectory bundle estimation For perception-driven planning

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2013.This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.Cataloged from student-submitted PDF version of thesis.Includes bibliographical references (p. 113-122).When operating in unknown environments, autonomous vehicles must perceive and understand the environment ahead in order to make effective navigation decisions. Long range perception can enable a vehicle to choose actions that take it directly toward its goal, avoiding dead ends. In addition, the perception range is critically important for ensuring the safety of vehicles with constrained dynamics. In general, the faster a vehicle moves, the more constrained its dynamics become due to acceleration limits imposed by its actuators. This means that the speed at which an autonomous agent can safely travel is often governed by its ability to perceive and understand the environment ahead. Overall, perception range is one of the most important factors that determines the performance of an autonomous vehicle. Today, autonomous vehicles tend to rely exclusively on metric representations built using range sensors to plan paths. However, such sensors are limited by their maximum range, field of view, and occluding obstacles in the foreground. Together, these limitations make up what we call the metric sensing horizon of the vehicle. The first two limitations are generally determined by the weight, size, power, and cost budget allocated to sensing. However, range sensors will always be limited by occlusions. If we wish to develop autonomous vehicles that are able to navigate directly toward a goal at high speeds through unknown environments, then we must move beyond the simple range-sensor based techniques. We must develop algorithms that enable autonomous agents to harness knowledge about the structure of the world to interpret additional sensor information (such as appearance information provided by cameras), and make inferences about parts of the world that cannot be directly observed. We develop a new representation based around trajectory bundles, that makes this challenging task more tractable. Rather than attempt to explicitly model the geometry of the world in front of the vehicle (which can be incredibly complex), we reason about the world in terms of what the vehicle can and cannot do. Trajectory bundles are designed to capture an abstract concept such as the command "go straight and then turn towards the right" in a concrete and actionable manner. We employ a library of trajectory bundles to reason about the layout of obstacles in the environment based on which bundles in the library are predicted to be feasible. Trajectory bundles provide a lens through which we can look at perception tasks, allowing us to leverage machine learning tools in much more effective ways for navigation. In this thesis we introduce trajectory bundles, and develop algorithms that use them to enable perception-driven planning. We develop a trajectory clustering algorithm that enables us to construct a set of trajectory bundles. We then develop a Bayesian filtering framework that enables us to estimate a belief over which trajectory bundles are feasible based on the history of actions and observations of the vehicle. We test our algorithms by using them to navigate a simulated fixed wing air vehicle at high speeds through an unknown environment using a monocular camera sensor.by Abraham Galton Bachrach.Ph.D
    • …
    corecore