92,044 research outputs found
Symbiotic Navigation in Multi-Robot Systems with Remote Obstacle Knowledge Sharing
Large scale operational areas often require multiple service robots for coverage and task parallelism. In such scenarios, each robot keeps its individual map of the environment and serves specific areas of the map at different times. We propose a knowledge sharing mechanism for multiple robots in which one robot can inform other robots about the changes in map, like path blockage, or new static obstacles, encountered at specific areas of the map. This symbiotic information sharing allows the robots to update remote areas of the map without having to explicitly navigate those areas, and plan efficient paths. A node representation of paths is presented for seamless sharing of blocked path information. The transience of obstacles is modeled to track obstacles which might have been removed. A lazy information update scheme is presented in which only relevant information affecting the current task is updated for efficiency. The advantages of the proposed method for path planning are discussed against traditional method with experimental results in both simulation and real environments
Positional Encoding by Robots with Non-Rigid Movements
Consider a set of autonomous computational entities, called \emph{robots},
operating inside a polygonal enclosure (possibly with holes), that have to
perform some collaborative tasks. The boundary of the polygon obstructs both
visibility and mobility of a robot. Since the polygon is initially unknown to
the robots, the natural approach is to first explore and construct a map of the
polygon. For this, the robots need an unlimited amount of persistent memory to
store the snapshots taken from different points inside the polygon. However, it
has been shown by Di Luna et al. [DISC 2017] that map construction can be done
even by oblivious robots by employing a positional encoding strategy where a
robot carefully positions itself inside the polygon to encode information in
the binary representation of its distance from the closest polygon vertex. Of
course, to execute this strategy, it is crucial for the robots to make accurate
movements. In this paper, we address the question whether this technique can be
implemented even when the movements of the robots are unpredictable in the
sense that the robot can be stopped by the adversary during its movement before
reaching its destination. However, there exists a constant ,
unknown to the robot, such that the robot can always reach its destination if
it has to move by no more than amount. This model is known in
literature as \emph{non-rigid} movement. We give a partial answer to the
question in the affirmative by presenting a map construction algorithm for
robots with non-rigid movement, but having bits of persistent memory and
ability to make circular moves
Multi-Robot Transfer Learning: A Dynamical System Perspective
Multi-robot transfer learning allows a robot to use data generated by a
second, similar robot to improve its own behavior. The potential advantages are
reducing the time of training and the unavoidable risks that exist during the
training phase. Transfer learning algorithms aim to find an optimal transfer
map between different robots. In this paper, we investigate, through a
theoretical study of single-input single-output (SISO) systems, the properties
of such optimal transfer maps. We first show that the optimal transfer learning
map is, in general, a dynamic system. The main contribution of the paper is to
provide an algorithm for determining the properties of this optimal dynamic map
including its order and regressors (i.e., the variables it depends on). The
proposed algorithm does not require detailed knowledge of the robots' dynamics,
but relies on basic system properties easily obtainable through simple
experimental tests. We validate the proposed algorithm experimentally through
an example of transfer learning between two different quadrotor platforms.
Experimental results show that an optimal dynamic map, with correct properties
obtained from our proposed algorithm, achieves 60-70% reduction of transfer
learning error compared to the cases when the data is directly transferred or
transferred using an optimal static map.Comment: 7 pages, 6 figures, accepted at the 2017 IEEE/RSJ International
Conference on Intelligent Robots and System
Robots that can adapt like animals
As robots leave the controlled environments of factories to autonomously
function in more complex, natural environments, they will have to respond to
the inevitable fact that they will become damaged. However, while animals can
quickly adapt to a wide variety of injuries, current robots cannot "think
outside the box" to find a compensatory behavior when damaged: they are limited
to their pre-specified self-sensing abilities, can diagnose only anticipated
failure modes, and require a pre-programmed contingency plan for every type of
potential damage, an impracticality for complex robots. Here we introduce an
intelligent trial and error algorithm that allows robots to adapt to damage in
less than two minutes, without requiring self-diagnosis or pre-specified
contingency plans. Before deployment, a robot exploits a novel algorithm to
create a detailed map of the space of high-performing behaviors: This map
represents the robot's intuitions about what behaviors it can perform and their
value. If the robot is damaged, it uses these intuitions to guide a
trial-and-error learning algorithm that conducts intelligent experiments to
rapidly discover a compensatory behavior that works in spite of the damage.
Experiments reveal successful adaptations for a legged robot injured in five
different ways, including damaged, broken, and missing legs, and for a robotic
arm with joints broken in 14 different ways. This new technique will enable
more robust, effective, autonomous robots, and suggests principles that animals
may use to adapt to injury
Learning Models for Following Natural Language Directions in Unknown Environments
Natural language offers an intuitive and flexible means for humans to
communicate with the robots that we will increasingly work alongside in our
homes and workplaces. Recent advancements have given rise to robots that are
able to interpret natural language manipulation and navigation commands, but
these methods require a prior map of the robot's environment. In this paper, we
propose a novel learning framework that enables robots to successfully follow
natural language route directions without any previous knowledge of the
environment. The algorithm utilizes spatial and semantic information that the
human conveys through the command to learn a distribution over the metric and
semantic properties of spatially extended environments. Our method uses this
distribution in place of the latent world model and interprets the natural
language instruction as a distribution over the intended behavior. A novel
belief space planner reasons directly over the map and behavior distributions
to solve for a policy using imitation learning. We evaluate our framework on a
voice-commandable wheelchair. The results demonstrate that by learning and
performing inference over a latent environment model, the algorithm is able to
successfully follow natural language route directions within novel, extended
environments.Comment: ICRA 201
- …