2,204 research outputs found

    SkiMap: An Efficient Mapping Framework for Robot Navigation

    Full text link
    We present a novel mapping framework for robot navigation which features a multi-level querying system capable to obtain rapidly representations as diverse as a 3D voxel grid, a 2.5D height map and a 2D occupancy grid. These are inherently embedded into a memory and time efficient core data structure organized as a Tree of SkipLists. Compared to the well-known Octree representation, our approach exhibits a better time efficiency, thanks to its simple and highly parallelizable computational structure, and a similar memory footprint when mapping large workspaces. Peculiarly within the realm of mapping for robot navigation, our framework supports realtime erosion and re-integration of measurements upon reception of optimized poses from the sensor tracker, so as to improve continuously the accuracy of the map.Comment: Accepted by International Conference on Robotics and Automation (ICRA) 2017. This is the submitted version. The final published version may be slightly differen

    User interface and function library for ground robot navigation

    Get PDF
    Master's Project (M.S.) University of Alaska Fairbanks, 2017A web application user interface and function library were developed to enable a user to program a ground robot to navigate autonomously. The user interface includes modules for generating a grid of obstacles from a map image, setting waypoints for a path through the map, and programming a robot in a code editor to navigate autonomously. The algorithm used for navigation is an A* algorithm modified with obstacle padding to accommodate the width of the robot and path smoothing to simplify the paths. The user interface and functions were designed to be simple so that users without technical backgrounds can use them, and by doing so they can engage in the development process of human-centered robots. The navigation functions were successful in finding paths in test configurations, and the performance of the algorithms was fast enough for user interactivity up to a certain limit of grid cell sizes

    User interface and function library for ground robot navigation

    Get PDF
    Master's Project (M.S.) University of Alaska Fairbanks, 2017A web application user interface and function library were developed to enable a user to program a ground robot to navigate autonomously. The user interface includes modules for generating a grid of obstacles from a map image, setting waypoints for a path through the map, and programming a robot in a code editor to navigate autonomously. The algorithm used for navigation is an A* algorithm modified with obstacle padding to accommodate the width of the robot and path smoothing to simplify the paths. The user interface and functions were designed to be simple so that users without technical backgrounds can use them, and by doing so they can engage in the development process of human-centered robots. The navigation functions were successful in finding paths in test configurations, and the performance of the algorithms was fast enough for user interactivity up to a certain limit of grid cell sizes

    Probabilistic Human-Robot Information Fusion

    Get PDF
    This thesis is concerned with combining the perceptual abilities of mobile robots and human operators to execute tasks cooperatively. It is generally agreed that a synergy of human and robotic skills offers an opportunity to enhance the capabilities of today’s robotic systems, while also increasing their robustness and reliability. Systems which incorporate both human and robotic information sources have the potential to build complex world models, essential for both automated and human decision making. In this work, humans and robots are regarded as equal team members who interact and communicate on a peer-to-peer basis. Human-robot communication is addressed using probabilistic representations common in robotics. While communication can in general be bidirectional, this work focuses primarily on human-to-robot information flow. More specifically, the approach advocated in this thesis is to let robots fuse their sensor observations with observations obtained from human operators. While robotic perception is well-suited for lower level world descriptions such as geometric properties, humans are able to contribute perceptual information on higher abstraction levels. Human input is translated into the machine representation via Human Sensor Models. A common mathematical framework for humans and robots reinforces the notion of true peer-to-peer interaction. Human-robot information fusion is demonstrated in two application domains: (1) scalable information gathering, and (2) cooperative decision making. Scalable information gathering is experimentally demonstrated on a system comprised of a ground vehicle, an unmanned air vehicle, and two human operators in a natural environment. Information from humans and robots was fused in a fully decentralised manner to build a shared environment representation on multiple abstraction levels. Results are presented in the form of information exchange patterns, qualitatively demonstrating the benefits of human-robot information fusion. The second application domain adds decision making to the human-robot task. Rational decisions are made based on the robots’ current beliefs which are generated by fusing human and robotic observations. Since humans are considered a valuable resource in this context, operators are only queried for input when the expected benefit of an observation exceeds the cost of obtaining it. The system can be seen as adjusting its autonomy at run-time based on the uncertainty in the robots’ beliefs. A navigation task is used to demonstrate the adjustable autonomy system experimentally. Results from two experiments are reported: a quantitative evaluation of human-robot team effectiveness, and a user study to compare the system to classical teleoperation. Results show the superiority of the system with respect to performance, operator workload, and usability

    Teleoperated visual inspection and surveillance with unmanned ground and aerial vehicles,” Int

    Get PDF
    Abstract—This paper introduces our robotic system named UGAV (Unmanned Ground-Air Vehicle) consisting of two semi-autonomous robot platforms, an Unmanned Ground Vehicle (UGV) and an Unmanned Aerial Vehicles (UAV). The paper focuses on three topics of the inspection with the combined UGV and UAV: (A) teleoperated control by means of cell or smart phones with a new concept of automatic configuration of the smart phone based on a RKI-XML description of the vehicles control capabilities, (B) the camera and vision system with the focus to real time feature extraction e.g. for the tracking of the UAV and (C) the architecture and hardware of the UAV

    CrowdFusion: Multi-Signal Fusion SLAM Positioning Leveraging Visible Light

    Get PDF
    With the fast development of location-based services, an ubiquitous indoor positioning approach with high accuracy and low calibration has become increasingly important. In this work, we target on a crowdsourcing approach with zero calibration effort based on visible light, magnetic field and WiFi to achieve sub-meter accuracy. We propose a CrowdFusion Simultaneous Localization and Mapping (SLAM) comprised of coarse-grained and fine-grained trace merging respectively based on the Iterative Closest Point (ICP) SLAM and GraphSLAM. ICP SLAM is proposed to correct the relative locations and directions of crowdsourcing traces and GraphSLAM is further adopted for fine-grained pose optimization. In CrowdFusion SLAM, visible light is used to accurately detect loop closures and magnetic field to extend the coverage. According to the merged traces, we construct a radio map with visible light and WiFi fingerprints. An enhanced particle filter fusing inertial sensors, visible light, WiFi and floor plan is designed, in which visible light fingerprinting is used to improve the accuracy and increase the resampling/rebooting efficiency. We evaluate CrowdFusion based on comprehensive experiments. The evaluation results show a mean accuracy of 0.67m for the merged traces and 0.77m for positioning, merely replying on crowdsourcing traces without professional calibration
    • 

    corecore