10,443 research outputs found
Balancing Global Exploration and Local-connectivity Exploitation with Rapidly-exploring Random disjointed-Trees
Sampling efficiency in a highly constrained environment has long been a major
challenge for sampling-based planners. In this work, we propose
Rapidly-exploring Random disjointed-Trees* (RRdT*), an incremental optimal
multi-query planner. RRdT* uses multiple disjointed-trees to exploit
local-connectivity of spaces via Markov Chain random sampling, which utilises
neighbourhood information derived from previous successful and failed samples.
To balance local exploitation, RRdT* actively explore unseen global spaces when
local-connectivity exploitation is unsuccessful. The active trade-off between
local exploitation and global exploration is formulated as a multi-armed bandit
problem. We argue that the active balancing of global exploration and local
exploitation is the key to improving sample efficient in sampling-based motion
planners. We provide rigorous proofs of completeness and optimal convergence
for this novel approach. Furthermore, we demonstrate experimentally the
effectiveness of RRdT*'s locally exploring trees in granting improved
visibility for planning. Consequently, RRdT* outperforms existing
state-of-the-art incremental planners, especially in highly constrained
environments.Comment: Submitted to IEEE International Conference on Robotics and Automation
(ICRA) 201
Motion Memory: Leveraging Past Experiences to Accelerate Future Motion Planning
When facing a new motion-planning problem, most motion planners solve it from
scratch, e.g., via sampling and exploration or starting optimization from a
straight-line path. However, most motion planners have to experience a variety
of planning problems throughout their lifetimes, which are yet to be leveraged
for future planning. In this paper, we present a simple but efficient method
called Motion Memory, which allows different motion planners to accelerate
future planning using past experiences. Treating existing motion planners as
either a closed or open box, we present a variety of ways that Motion Memory
can contribute to reduce the planning time when facing a new planning problem.
We provide extensive experiment results with three different motion planners on
three classes of planning problems with over 30,000 problem instances and show
that planning speed can be significantly reduced by up to 89% with the proposed
Motion Memory technique and with increasing past planning experiences
Reset-free Trial-and-Error Learning for Robot Damage Recovery
The high probability of hardware failures prevents many advanced robots
(e.g., legged robots) from being confidently deployed in real-world situations
(e.g., post-disaster rescue). Instead of attempting to diagnose the failures,
robots could adapt by trial-and-error in order to be able to complete their
tasks. In this situation, damage recovery can be seen as a Reinforcement
Learning (RL) problem. However, the best RL algorithms for robotics require the
robot and the environment to be reset to an initial state after each episode,
that is, the robot is not learning autonomously. In addition, most of the RL
methods for robotics do not scale well with complex robots (e.g., walking
robots) and either cannot be used at all or take too long to converge to a
solution (e.g., hours of learning). In this paper, we introduce a novel
learning algorithm called "Reset-free Trial-and-Error" (RTE) that (1) breaks
the complexity by pre-generating hundreds of possible behaviors with a dynamics
simulator of the intact robot, and (2) allows complex robots to quickly recover
from damage while completing their tasks and taking the environment into
account. We evaluate our algorithm on a simulated wheeled robot, a simulated
six-legged robot, and a real six-legged walking robot that are damaged in
several ways (e.g., a missing leg, a shortened leg, faulty motor, etc.) and
whose objective is to reach a sequence of targets in an arena. Our experiments
show that the robots can recover most of their locomotion abilities in an
environment with obstacles, and without any human intervention.Comment: 18 pages, 16 figures, 3 tables, 6 pseudocodes/algorithms, video at
https://youtu.be/IqtyHFrb3BU, code at
https://github.com/resibots/chatzilygeroudis_2018_rt
Planning hand-arm grasping motions with human-like appearance
© 2018 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting /republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other worksFinalista de l’IROS Best Application Paper Award a la 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, ICROS.This paper addresses the problem of obtaining human-like motions on hand-arm robotic systems performing pick-and-place actions. The focus is set on the coordinated movements of the robotic arm and the anthropomorphic mechanical hand, with which the arm is equipped. For this, human movements performing different grasps are captured and mapped to the robot in order to compute the human hand synergies. These synergies are used to reduce the complexity of the planning phase by reducing the dimension of the search space. In addition, the paper proposes a sampling-based planner, which guides the motion planning ollowing the synergies. The introduced approach is tested in an application example and thoroughly compared with other state-of-the-art planning algorithms, obtaining better results.Peer ReviewedAward-winningPostprint (author's final draft
- …