921 research outputs found

    NASA Center for Intelligent Robotic Systems for Space Exploration

    Get PDF
    NASA's program for the civilian exploration of space is a challenge to scientists and engineers to help maintain and further develop the United States' position of leadership in a focused sphere of space activity. Such an ambitious plan requires the contribution and further development of many scientific and technological fields. One research area essential for the success of these space exploration programs is Intelligent Robotic Systems. These systems represent a class of autonomous and semi-autonomous machines that can perform human-like functions with or without human interaction. They are fundamental for activities too hazardous for humans or too distant or complex for remote telemanipulation. To meet this challenge, Rensselaer Polytechnic Institute (RPI) has established an Engineering Research Center for Intelligent Robotic Systems for Space Exploration (CIRSSE). The Center was created with a five year $5.5 million grant from NASA submitted by a team of the Robotics and Automation Laboratories. The Robotics and Automation Laboratories of RPI are the result of the merger of the Robotics and Automation Laboratory of the Department of Electrical, Computer, and Systems Engineering (ECSE) and the Research Laboratory for Kinematics and Robotic Mechanisms of the Department of Mechanical Engineering, Aeronautical Engineering, and Mechanics (ME,AE,&M), in 1987. This report is an examination of the activities that are centered at CIRSSE

    Role Playing Learning for Socially Concomitant Mobile Robot Navigation

    Full text link
    In this paper, we present the Role Playing Learning (RPL) scheme for a mobile robot to navigate socially with its human companion in populated environments. Neural networks (NN) are constructed to parameterize a stochastic policy that directly maps sensory data collected by the robot to its velocity outputs, while respecting a set of social norms. An efficient simulative learning environment is built with maps and pedestrians trajectories collected from a number of real-world crowd data sets. In each learning iteration, a robot equipped with the NN policy is created virtually in the learning environment to play itself as a companied pedestrian and navigate towards a goal in a socially concomitant manner. Thus, we call this process Role Playing Learning, which is formulated under a reinforcement learning (RL) framework. The NN policy is optimized end-to-end using Trust Region Policy Optimization (TRPO), with consideration of the imperfectness of robot's sensor measurements. Simulative and experimental results are provided to demonstrate the efficacy and superiority of our method

    Multi-Robot Path Planning Combining Heuristics and Multi-Agent Reinforcement Learning

    Full text link
    Multi-robot path finding in dynamic environments is a highly challenging classic problem. In the movement process, robots need to avoid collisions with other moving robots while minimizing their travel distance. Previous methods for this problem either continuously replan paths using heuristic search methods to avoid conflicts or choose appropriate collision avoidance strategies based on learning approaches. The former may result in long travel distances due to frequent replanning, while the latter may have low learning efficiency due to low sample exploration and utilization, and causing high training costs for the model. To address these issues, we propose a path planning method, MAPPOHR, which combines heuristic search, empirical rules, and multi-agent reinforcement learning. The method consists of two layers: a real-time planner based on the multi-agent reinforcement learning algorithm, MAPPO, which embeds empirical rules in the action output layer and reward functions, and a heuristic search planner used to create a global guiding path. During movement, the heuristic search planner replans new paths based on the instructions of the real-time planner. We tested our method in 10 different conflict scenarios. The experiments show that the planning performance of MAPPOHR is better than that of existing learning and heuristic methods. Due to the utilization of empirical knowledge and heuristic search, the learning efficiency of MAPPOHR is higher than that of existing learning methods

    Motion Primitives and Planning for Robots with Closed Chain Systems and Changing Topologies

    Get PDF
    When operating in human environments, a robot should use predictable motions that allow humans to trust and anticipate its behavior. Heuristic search-based planning offers predictable motions and guarantees on completeness and sub-optimality of solutions. While search-based planning on motion primitive-based (lattice-based) graphs has been used extensively in navigation, application to high-dimensional state-spaces has, until recently, been thought impractical. This dissertation presents methods we have developed for applying these graphs to mobile manipulation, specifically for systems which contain closed chains. The formation of closed chains in tasks that involve contacts with the environment may reduce the number of available degrees-of-freedom but adds complexity in terms of constraints in the high-dimensional state-space. We exploit the dimensionality reduction inherent in closed kinematic chains to get efficient search-based planning. Our planner handles changing topologies (switching between open and closed-chains) in a single plan, including what transitions to include and when to include them. Thus, we can leverage existing results for search-based planning for open chains, combining open and closed chain manipulation planning into one framework. Proofs regarding the framework are introduced for the application to graph-search and its theoretical guarantees of optimality. The dimensionality-reduction is done in a manner that enables finding optimal solutions to low-dimensional problems which map to correspondingly optimal full-dimensional solutions. We apply this framework to planning for opening and navigating through non-spring and spring-loaded doors using a Willow Garage PR2. The framework motivates our approaches to the Atlas humanoid robot from Boston Dynamics for both stationary manipulation and quasi-static walking, as a closed chain is formed when both feet are on the ground

    Safe, Scalable, and Complete Motion Planning of Large Teams of Interchangeable Robots

    Get PDF
    Large teams of mobile robots have an unprecedented potential to assist humans in a number of roles ranging from humanitarian efforts to e-commerce order fulfillment. Utilizing a team of robots provides an inherent parallelism in computation and task completion while providing redundancy to isolated robot failures. Whether a mission requires all robots to stay close to each other in a formation, navigate to a preselected set of goal locations, or to actively try to spread out to gain as much information as possible, the team must be able to successfully navigate the robots to desired locations. While there is a rich literature on motion planning for teams of robots, the problem is sufficiently challenging that in general all methods trade off one of the following properties: completeness, computational scalability, safety, or optimality. This dissertation proposes robot interchangeability as an additional trade-off consideration. Specifically, the work presented here leverages the total interchangeability of robots and develops a series of novel, complete, computationally tractable algorithms to control a team of robots and avoid collisions while retaining a notion of optimality. This dissertation begins by presenting a robust decentralized formation control algorithm for control of robots operating in tight proximity to one another. Next, a series of complete, computationally tractable multiple robot planning algorithms are presented. These planners preserve optimality, completeness, and computationally tractability by leveraging robot interchangeability. Finally, a polynomial time approximation algorithm is proposed that routes teams of robots to visit a large number of specified locations while bounding the suboptimality of total mission completion time. Each algorithm is verified in simulation and when applicable, on a team of dynamic aerial robots

    Learning Team-Based Navigation: A Review of Deep Reinforcement Learning Techniques for Multi-Agent Pathfinding

    Full text link
    Multi-agent pathfinding (MAPF) is a critical field in many large-scale robotic applications, often being the fundamental step in multi-agent systems. The increasing complexity of MAPF in complex and crowded environments, however, critically diminishes the effectiveness of existing solutions. In contrast to other studies that have either presented a general overview of the recent advancements in MAPF or extensively reviewed Deep Reinforcement Learning (DRL) within multi-agent system settings independently, our work presented in this review paper focuses on highlighting the integration of DRL-based approaches in MAPF. Moreover, we aim to bridge the current gap in evaluating MAPF solutions by addressing the lack of unified evaluation metrics and providing comprehensive clarification on these metrics. Finally, our paper discusses the potential of model-based DRL as a promising future direction and provides its required foundational understanding to address current challenges in MAPF. Our objective is to assist readers in gaining insight into the current research direction, providing unified metrics for comparing different MAPF algorithms and expanding their knowledge of model-based DRL to address the existing challenges in MAPF.Comment: 36 pages, 10 figures, published in Artif Intell Rev 57, 41 (2024

    Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning

    Get PDF
    Finding feasible, collision-free paths for multiagent systems can be challenging, particularly in non-communicating scenarios where each agent's intent (e.g. goal) is unobservable to the others. In particular, finding time efficient paths often requires anticipating interaction with neighboring agents, the process of which can be computationally prohibitive. This work presents a decentralized multiagent collision avoidance algorithm based on a novel application of deep reinforcement learning, which effectively offloads the online computation (for predicting interaction patterns) to an offline learning procedure. Specifically, the proposed approach develops a value network that encodes the estimated time to the goal given an agent's joint configuration (positions and velocities) with its neighbors. Use of the value network not only admits efficient (i.e., real-time implementable) queries for finding a collision-free velocity vector, but also considers the uncertainty in the other agents' motion. Simulation results show more than 26% improvement in paths quality (i.e., time to reach the goal) when compared with optimal reciprocal collision avoidance (ORCA), a state-of-the-art collision avoidance strategy.Ford Motor Compan

    Cooperative Path-Planning for Multi-Vehicle Systems

    Get PDF
    In this paper, we propose a collision avoidance algorithm for multi-vehicle systems, which is a common problem in many areas, including navigation and robotics. In dynamic environments, vehicles may become involved in potential collisions with each other, particularly when the vehicle density is high and the direction of travel is unrestricted. Cooperatively planning vehicle movement can effectively reduce and fairly distribute the detour inconvenience before subsequently returning vehicles to their intended paths. We present a novel method of cooperative path planning for multi-vehicle systems based on reinforcement learning to address this problem as a decision process. A dynamic system is described as a multi-dimensional space formed by vectors as states to represent all participating vehicles’ position and orientation, whilst considering the kinematic constraints of the vehicles. Actions are defined for the system to transit from one state to another. In order to select appropriate actions whilst satisfying the constraints of path smoothness, constant speed and complying with a minimum distance between vehicles, an approximate value function is iteratively developed to indicate the desirability of every state-action pair from the continuous state space and action space. The proposed scheme comprises two phases. The convergence of the value function takes place in the former learning phase, and it is then used as a path planning guideline in the subsequent action phase. This paper summarizes the concept and methodologies used to implement this online cooperative collision avoidance algorithm and presents results and analysis regarding how this cooperative scheme improves upon two baseline schemes where vehicles make movement decisions independently
    corecore