27,567 research outputs found
Particle Swarm Optimization Based Source Seeking
Signal source seeking using autonomous vehicles is a complex problem. The
complexity increases manifold when signal intensities captured by physical
sensors onboard are noisy and unreliable. Added to the fact that signal
strength decays with distance, noisy environments make it extremely difficult
to describe and model a decay function. This paper addresses our work with
seeking maximum signal strength in a continuous electromagnetic signal source
with mobile robots, using Particle Swarm Optimization (PSO). A one to one
correspondence with swarm members in a PSO and physical Mobile robots is
established and the positions of the robots are iteratively updated as the PSO
algorithm proceeds forward. Since physical robots are responsive to swarm
position updates, modifications were required to implement the interaction
between real robots and the PSO algorithm. The development of modifications
necessary to implement PSO on mobile robots, and strategies to adapt to real
life environments such as obstacles and collision objects are presented in this
paper. Our findings are also validated using experimental testbeds.Comment: 13 pages, 12 figure
Airborne chemical sensing with mobile robots
Airborne chemical sensing with mobile robots has been an active research areasince the beginning of the 1990s. This article presents a review of research work in this field,including gas distribution mapping, trail guidance, and the different subtasks of gas sourcelocalisation. Due to the difficulty of modelling gas distribution in a real world environmentwith currently available simulation techniques, we focus largely on experimental work and donot consider publications that are purely based on simulations
Wavefront Propagation and Fuzzy Based Autonomous Navigation
Path planning and obstacle avoidance are the two major issues in any
navigation system. Wavefront propagation algorithm, as a good path planner, can
be used to determine an optimal path. Obstacle avoidance can be achieved using
possibility theory. Combining these two functions enable a robot to
autonomously navigate to its destination. This paper presents the approach and
results in implementing an autonomous navigation system for an indoor mobile
robot. The system developed is based on a laser sensor used to retrieve data to
update a two dimensional world model of therobot environment. Waypoints in the
path are incorporated into the obstacle avoidance. Features such as ageing of
objects and smooth motion planning are implemented to enhance efficiency and
also to cater for dynamic environments
GUARDIANS final report
Emergencies in industrial warehouses are a major concern for firefghters. The large dimensions together with the development of dense smoke that drastically reduces visibility, represent major challenges. The Guardians robot swarm is designed to assist fire fighters in searching a
large warehouse. In this report we discuss the technology developed for a swarm of robots searching and assisting fire fighters. We explain the swarming algorithms which provide the functionality by which the robots react to and follow humans while no communication is required. Next we
discuss the wireless communication system, which is a so-called mobile ad-hoc network. The communication network provides also one of the means to locate the robots and humans. Thus the robot swarm is able to locate itself and provide guidance information to the humans. Together with
the re ghters we explored how the robot swarm should feed information back to the human fire fighter. We have designed and experimented with interfaces for presenting swarm based information to human beings
Reinforcement Learning: A Survey
This paper surveys the field of reinforcement learning from a
computer-science perspective. It is written to be accessible to researchers
familiar with machine learning. Both the historical basis of the field and a
broad selection of current work are summarized. Reinforcement learning is the
problem faced by an agent that learns behavior through trial-and-error
interactions with a dynamic environment. The work described here has a
resemblance to work in psychology, but differs considerably in the details and
in the use of the word ``reinforcement.'' The paper discusses central issues of
reinforcement learning, including trading off exploration and exploitation,
establishing the foundations of the field via Markov decision theory, learning
from delayed reinforcement, constructing empirical models to accelerate
learning, making use of generalization and hierarchy, and coping with hidden
state. It concludes with a survey of some implemented systems and an assessment
of the practical utility of current methods for reinforcement learning.Comment: See http://www.jair.org/ for any accompanying file
- …