3 research outputs found

    A Scalable Information Theoretic Approach to Distributed Robot Coordination

    Get PDF
    This paper presents a scalable information theoretic approach to infer the state of an environment by distributively controlling robots equipped with sensors. The robots iteratively estimate the environment state using a recursive Bayesian filter, while continuously moving to improve the quality of the estimate by following the gradient of mutual information. Both the filter and the controller use a novel algorithm for approximating the robots' joint measurement probabilities, which combines consensus (for decentralization) and sampling (for scalability). The approximations are shown to approach the true joint measurement probabilities as the size of the consensus rounds grows or as the network becomes complete. The resulting gradient controller runs in constant time with respect to the number of robots, and linear time with respect to the number of sensor measurements and environment discretization cells, while traditional mutual information methods are exponential in all of these quantities. Furthermore, the controller is proven to be convergent between consensus rounds and, under certain conditions, is locally optimal. The complete distributed inference and coordination algorithm is demonstrated in experiments with five quad-rotor flying robots and simulations with 100 robots.This work is sponsored by the Department of the Air Force under Air Force contract number FA8721-05-C-0002. The opinions, interpretations, recommendations, and conclusions are those of the authors and are not necessarily endorsed by the United States Government. This work is also supported in part by ARO grant number W911NF-05-1-0219, ONR grant number N00014-09-1-1051, NSF grant number EFRI-0735953, ARL grant number W911NF-08-2-0004, MIT Lincoln Laboratory, the European Commission, and the Boeing Company

    Mutual information-based gradient-ascent control for distributed robotics

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2013.Cataloged from PDF version of thesis.Includes bibliographical references (pages 167-179).This thesis presents the derivation, analysis, and implementation of a novel class of decentralized mutual information-based gradient-ascent controllers that continuously move robots equipped with sensors to better observe their environment. We begin with the fundamental problem of deploying a single ground robot equipped with a range sensor and tasked to build an occupancy grid map. The desired explorative behaviors of the robot for occupancy grid mapping highlight the correlation between the information content and the spatial realization of the robot's range measurements. We prove that any occupancy grid controller tasked to maximize a mutual information reward function is eventually attracted to unexplored space, i.e., areas of highest uncertainty. We show that mutual information encodes geometric relationships that are fundamental to robot control and yields geometrically relevant reward surfaces on which robots can navigate. Taking inspiration from geometric-based approaches to distributed robot coordination, we show that many multi-robot inference tasks can be cast in terms of an optimization problem. This optimization problem defines the task of minimizing the conditional entropy associated with the robots' inferred beliefs of the environment, which is equivalent to maximizing the mutual information between the environment state and the robots' next joint observation. Given simple robot dynamics and few probabilistic assumptions, none of which involve Gaussianity, we derive a gradientascent solution approach to these optimization problems that is convergent between sensor observations and locally optimal. More formally, we invoke LaSalle's Invariance Principle to prove that, given enough time between consecutive joint observations, robots following the gradient of mutual information will converge to goal positions that locally maximize the expected information gain resulting from the next observation. We show that the algorithmic implementation of the generalized gradient-ascent controller is not readily distributed among multiple robots, and thus sample-based methods are introduced to distributively approximate the likelihoods of the robots' joint observations. Not only are the involved non-parametric representations compatible with any type of Bayesian filter, but the computational complexities of the resulting decentralized controllers are independent with respect to the number of robots. Concerning the distributed approximations, we give two example consensus-based algorithms that run on an undirected network graph. The first consensus-based algorithm approximates discrete measurement probabilities, while the second approximates continuous likelihood distributions. We show that these anytime approximations provably converge to the correct values on a static and connected network graph without knowledge of the number of robots in the network or the corresponding graph's topology. Lastly, we incorporate the resulting consensus-based algorithms into both a hardware system and a simulation environment to allow for decentralized controller evaluation under non-ideal network settings. For the hardware experiments, the task is to infer the state of a bounded, planar environment by deploying five quadrotor flying robots with simulated sensors in both indoor and outdoor settings. For the numerical simulations, Monte Carlo-based analyses are performed for 100 robots, where each robot is simulated on an independent computer node within a computer cluster system. Simulations are also performed for 1000 robots using a single workstation computer equipped with a multicore GPU-enabled graphics card. The results from both the hardware experiments and numerical simulations validate our theoretical and computational claims throughout the thesis.by Brian John Julian.Ph.D

    A scalable information theoretic approach to distributed robot coordination

    No full text
    This paper presents a scalable information theoretic approach to infer the state of an environment by distributively controlling robots equipped with sensors. The robots iteratively estimate the environment state using a recursive Bayesian filter, while continuously moving to improve the quality of the estimate by following the gradient of mutual information. Both the filter and the controller use a novel algorithm for approximating the robots' joint measurement probabilities, which combines consensus (for decentralization) and sampling (for scalability). The approximations are shown to approach the true joint measurement probabilities as the size of the consensus rounds grows or as the network becomes complete. The resulting gradient controller runs in constant time with respect to the number of robots, and linear time with respect to the number of sensor measurements and environment discretization cells, while traditional mutual information methods are exponential in all of these quantities. Furthermore, the controller is proven to be convergent between consensus rounds and, under certain conditions, is locally optimal. The complete distributed inference and coordination algorithm is demonstrated in experiments with five quad-rotor flying robots and simulations with 100 robots. © 2011 IEEE
    corecore