1,470 research outputs found

    Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age

    Get PDF
    Simultaneous Localization and Mapping (SLAM)consists in the concurrent construction of a model of the environment (the map), and the estimation of the state of the robot moving within it. The SLAM community has made astonishing progress over the last 30 years, enabling large-scale real-world applications, and witnessing a steady transition of this technology to industry. We survey the current state of SLAM. We start by presenting what is now the de-facto standard formulation for SLAM. We then review related work, covering a broad set of topics including robustness and scalability in long-term mapping, metric and semantic representations for mapping, theoretical performance guarantees, active SLAM and exploration, and other new frontiers. This paper simultaneously serves as a position paper and tutorial to those who are users of SLAM. By looking at the published research with a critical eye, we delineate open challenges and new research issues, that still deserve careful scientific investigation. The paper also contains the authors' take on two questions that often animate discussions during robotics conferences: Do robots need SLAM? and Is SLAM solved

    Real-time performance-focused on localisation techniques for autonomous vehicle: a review

    Get PDF

    Map building, localization and exploration for multi-robot systems

    Get PDF
    The idea of having robots performing the task for which they have been designed completely autonomously and interacting with the environment has been the main objective since the beginning of mobile robotics. In order to achieve such a degree of autonomy, it is indispensable for the robot to have a map of the environment and to know its location in it, in addition to being able to solve other problems such as motion control and path planning towards its goal. During the fulfillment of certain missions without a prior knowledge of its environment, the robot must use the inaccurate information provided by its on-board sensors to build a map at the same time it is located in it, arising the problem of Simultaneous Localization and Mapping (SLAM) extensively studied in mobile robotics. In recent years, there has been a growing interest in the use of robot teams due to their multiple benefits with respect to single-robot systems such as higher robustness, accuracy, efficiency and the possibility to cooperate to perform a task or to cover larger environments in less time. Robot formations also belongs to this field of cooperative robots, where they have to maintain a predefined structure while navigating in the environment. Despite their advantages, the complexity of autonomous multi-robot systems increases with the number of robots as a consequence of the larger amount of information available that must be handled, stored and transmitted through the communications network. Therefore, the development of these systems presents new difficulties when solving the aforementioned problems which, instead of being addressed individually for each robot, must be solved cooperatively to efficiently exploit all the information collected by the team. The design of algorithms in this multi-robot context should be directed to obtain greater scalability and performance to allow their online execution. This thesis is developed in the field of multi-robot systems and proposes solutions to the navigation, localization, mapping and path planning processes which form an autonomous system. The first part of contributions presented in this thesis is developed in the context of robot formations, which require greater team cooperation and synchronization, although they can be extended to systems without this navigation constraint. We propose localization, map refinement and exploration techniques under the assumption that the formation is provided with a map of the environment, possibly partial and inaccurate, wherein it has to carry out its commanded mission. In a second part, we propose a multi-robot SLAM approach without any assumption about the prior knowledge of a map nor the relationships between robots in which we make use of state of the art methodologies to efficiently manage the resources available in the system. The performance and efficiency of the proposed robot formation and multi-robot SLAM systems have been demonstrated through their implementation and testing both in simulations and with real robots

    Localization Algorithms for GNSS-denied and Challenging Environments

    Get PDF
    In this dissertation, the problem about localization in GNSS-denied and challenging environments is addressed. Specifically, the challenging environments discussed in this dissertation include two different types, environments including only low-resolution features and environments containing moving objects. To achieve accurate pose estimates, the errors are always bounded through matching observations from sensors with surrounding environments. These challenging environments, unfortunately, would bring troubles into matching related methods, such as fingerprint matching, and ICP. For instance, in environments with low-resolution features, the on-board sensor measurements could match to multiple positions on a map, which creates ambiguity; in environments with moving objects included, the accuracy of the estimated localization is affected by the moving objects when performing matching. In this dissertation, two sensor fusion based strategies are proposed to solve localization problems with respect to these two types of challenging environments, respectively. For environments with only low-resolution features, such as flying over sea or desert, a multi-agent localization algorithm using pairwise communication with ranging and magnetic anomaly measurements is proposed in this dissertation. A scalable framework is then presented to extend the multi-agent localization algorithm to be suitable for a large group of agents (e.g., 128 agents) through applying CI algorithm. The simulation results show that the proposed algorithm is able to deal with large group sizes, achieve 10 meters level localization performance with 180 km traveling distance, while under restrictive communication constraints. For environments including moving objects, lidar-inertial-based solutions are proposed and tested in this dissertation. Inspired by the CI algorithm presented above, a potential solution using multiple features motions estimate and tracking is analyzed. In order to improve the performance and effectiveness of the potential solution, a lidar-inertial based SLAM algorithm is then proposed. In this method, an efficient tightly-coupled iterated Kalman filter with a build-in dynamic object filter is designed as the front-end of the SLAM algorithm, and the factor graph strategy using a scan context technology as the loop closure detection is utilized as the back-end. The performance of the proposed lidar-inertial based SLAM algorithm is evaluated with several data sets collected in environments including moving objects, and compared with the state-of-the-art lidar-inertial based SLAM algorithms

    Viewfinder: final activity report

    Get PDF
    The VIEW-FINDER project (2006-2009) is an 'Advanced Robotics' project that seeks to apply a semi-autonomous robotic system to inspect ground safety in the event of a fire. Its primary aim is to gather data (visual and chemical) in order to assist rescue personnel. A base station combines the gathered information with information retrieved from off-site sources. The project addresses key issues related to map building and reconstruction, interfacing local command information with external sources, human-robot interfaces and semi-autonomous robot navigation. The VIEW-FINDER system is a semi-autonomous; the individual robot-sensors operate autonomously within the limits of the task assigned to them, that is, they will autonomously navigate through and inspect an area. Human operators monitor their operations and send high level task requests as well as low level commands through the interface to any nodes in the entire system. The human interface has to ensure the human supervisor and human interveners are provided a reduced but good and relevant overview of the ground and the robots and human rescue workers therein

    Efficient Factor Graph Fusion for Multi-robot Mapping

    Get PDF
    This work presents a novel method to efficiently factorize the combination of multiple factor graphs having common variables of estimation. The fast-paced innovation in the algebraic graph theory has enabled new tools of state estimation like factor graphs. Recent factor graph formulation for Simultaneous Localization and Mapping (SLAM) like Incremental Smoothing and Mapping using the Bayes tree (ISAM2) has been very successful and garnered much attention. Variable ordering, a well-known technique in linear algebra is employed for solving the factor graph. Our primary contribution in this work is to reuse the variable ordering of the graphs being combined to find the ordering of the fused graph. In the case of mapping, multiple robots provide a great advantage over single robot by providing a faster map coverage and better estimation quality. This coupled with an inevitable increase in the number of robots around us produce a demand for faster algorithms. For example, a city full of self-driving cars could pool their observation measurements rapidly to plan a traffic free navigation. By reusing the variable ordering of the parent graphs we were able to produce an order-of-magnitude difference in the time required for solving the fused graph. We also provide a formal verification to show that the proposed strategy does not violate any of the relevant standards. A common problem in multi-robot SLAM is relative pose graph initialization to produce a globally consistent map. The other contribution addresses this by minimizing a specially formulated error function as a part of solving the factor graph. The performance is illustrated on a publicly available SuiteSparse dataset and the multi-robot AP Hill dataset
    corecore