1,943 research outputs found

    A general framework of multi-population methods with clustering in undetectable dynamic environments

    Get PDF
    Copyright @ 2011 IEEETo solve dynamic optimization problems, multiple population methods are used to enhance the population diversity for an algorithm with the aim of maintaining multiple populations in different sub-areas in the fitness landscape. Many experimental studies have shown that locating and tracking multiple relatively good optima rather than a single global optimum is an effective idea in dynamic environments. However, several challenges need to be addressed when multi-population methods are applied, e.g., how to create multiple populations, how to maintain them in different sub-areas, and how to deal with the situation where changes can not be detected or predicted. To address these issues, this paper investigates a hierarchical clustering method to locate and track multiple optima for dynamic optimization problems. To deal with undetectable dynamic environments, this paper applies the random immigrants method without change detection based on a mechanism that can automatically reduce redundant individuals in the search space throughout the run. These methods are implemented into several research areas, including particle swarm optimization, genetic algorithm, and differential evolution. An experimental study is conducted based on the moving peaks benchmark to test the performance with several other algorithms from the literature. The experimental results show the efficiency of the clustering method for locating and tracking multiple optima in comparison with other algorithms based on multi-population methods on the moving peaks benchmark

    Combining Subgoal Graphs with Reinforcement Learning to Build a Rational Pathfinder

    Full text link
    In this paper, we present a hierarchical path planning framework called SG-RL (subgoal graphs-reinforcement learning), to plan rational paths for agents maneuvering in continuous and uncertain environments. By "rational", we mean (1) efficient path planning to eliminate first-move lags; (2) collision-free and smooth for agents with kinematic constraints satisfied. SG-RL works in a two-level manner. At the first level, SG-RL uses a geometric path-planning method, i.e., Simple Subgoal Graphs (SSG), to efficiently find optimal abstract paths, also called subgoal sequences. At the second level, SG-RL uses an RL method, i.e., Least-Squares Policy Iteration (LSPI), to learn near-optimal motion-planning policies which can generate kinematically feasible and collision-free trajectories between adjacent subgoals. The first advantage of the proposed method is that SSG can solve the limitations of sparse reward and local minima trap for RL agents; thus, LSPI can be used to generate paths in complex environments. The second advantage is that, when the environment changes slightly (i.e., unexpected obstacles appearing), SG-RL does not need to reconstruct subgoal graphs and replan subgoal sequences using SSG, since LSPI can deal with uncertainties by exploiting its generalization ability to handle changes in environments. Simulation experiments in representative scenarios demonstrate that, compared with existing methods, SG-RL can work well on large-scale maps with relatively low action-switching frequencies and shorter path lengths, and SG-RL can deal with small changes in environments. We further demonstrate that the design of reward functions and the types of training environments are important factors for learning feasible policies.Comment: 20 page

    Particle Swarm Optimization Based Source Seeking

    Get PDF
    Signal source seeking using autonomous vehicles is a complex problem. The complexity increases manifold when signal intensities captured by physical sensors onboard are noisy and unreliable. Added to the fact that signal strength decays with distance, noisy environments make it extremely difficult to describe and model a decay function. This paper addresses our work with seeking maximum signal strength in a continuous electromagnetic signal source with mobile robots, using Particle Swarm Optimization (PSO). A one to one correspondence with swarm members in a PSO and physical Mobile robots is established and the positions of the robots are iteratively updated as the PSO algorithm proceeds forward. Since physical robots are responsive to swarm position updates, modifications were required to implement the interaction between real robots and the PSO algorithm. The development of modifications necessary to implement PSO on mobile robots, and strategies to adapt to real life environments such as obstacles and collision objects are presented in this paper. Our findings are also validated using experimental testbeds.Comment: 13 pages, 12 figure

    Intelligent Robotics Navigation System: Problems, Methods, and Algorithm

    Get PDF
    This paper set out to supplement new studies with a brief and comprehensible review of the advanced development in the area of the navigation system, starting from a single robot, multi-robot, and swarm robots from a particular perspective by taking insights from these biological systems. The inspiration is taken from nature by observing the human and the social animal that is believed to be very beneficial for this purpose. The intelligent navigation system is developed based on an individual characteristic or a social animal biological structure. The discussion of this paper will focus on how simple agent’s structure utilizes flexible and potential outcomes in order to navigate in a productive and unorganized surrounding. The combination of the navigation system and biologically inspired approach has attracted considerable attention, which makes it an important research area in the intelligent robotic system. Overall, this paper explores the implementation, which is resulted from the simulation performed by the embodiment of robots operating in real environments

    A novel improved elephant herding optimization for path planning of a mobile robot

    Get PDF
    Swarm intelligence algorithms have been in recent years one of the most used tools for planning the trajectory of a mobile robot. Researchers are applying those algorithms to find the optimal path, which reduces the time required to perform a task by the mobile robot. In this paper, we propose a new method based on the grey wolf optimizer algorithm (GWO) and the improved elephant herding optimization algorithm (IEHO) for planning the optimal trajectory of a mobile robot. The proposed solution consists of developing an IEHO algorithm by improving the basic EHO algorithm and then hybridizing it with the GWO algorithm to take advantage of the exploration and exploitation capabilities of both algorithms. The comparison of the IEHO-GWO hybrid proposed in this work with the GWO, EHO, and cuckoo-search (CS) algorithms via simulation shows its effectiveness in finding an optimal trajectory by avoiding obstacles around the mobile robot

    Adaptive and learning-based formation control of swarm robots

    Get PDF
    Autonomous aerial and wheeled mobile robots play a major role in tasks such as search and rescue, transportation, monitoring, and inspection. However, these operations are faced with a few open challenges including robust autonomy, and adaptive coordination based on the environment and operating conditions, particularly in swarm robots with limited communication and perception capabilities. Furthermore, the computational complexity increases exponentially with the number of robots in the swarm. This thesis examines two different aspects of the formation control problem. On the one hand, we investigate how formation could be performed by swarm robots with limited communication and perception (e.g., Crazyflie nano quadrotor). On the other hand, we explore human-swarm interaction (HSI) and different shared-control mechanisms between human and swarm robots (e.g., BristleBot) for artistic creation. In particular, we combine bio-inspired (i.e., flocking, foraging) techniques with learning-based control strategies (using artificial neural networks) for adaptive control of multi- robots. We first review how learning-based control and networked dynamical systems can be used to assign distributed and decentralized policies to individual robots such that the desired formation emerges from their collective behavior. We proceed by presenting a novel flocking control for UAV swarm using deep reinforcement learning. We formulate the flocking formation problem as a partially observable Markov decision process (POMDP), and consider a leader-follower configuration, where consensus among all UAVs is used to train a shared control policy, and each UAV performs actions based on the local information it collects. In addition, to avoid collision among UAVs and guarantee flocking and navigation, a reward function is added with the global flocking maintenance, mutual reward, and a collision penalty. We adapt deep deterministic policy gradient (DDPG) with centralized training and decentralized execution to obtain the flocking control policy using actor-critic networks and a global state space matrix. In the context of swarm robotics in arts, we investigate how the formation paradigm can serve as an interaction modality for artists to aesthetically utilize swarms. In particular, we explore particle swarm optimization (PSO) and random walk to control the communication between a team of robots with swarming behavior for musical creation

    A Hierarchal Planning Framework for AUV Mission Management in a Spatio-Temporal Varying Ocean

    Full text link
    The purpose of this paper is to provide a hierarchical dynamic mission planning framework for a single autonomous underwater vehicle (AUV) to accomplish task-assign process in a limited time interval while operating in an uncertain undersea environment, where spatio-temporal variability of the operating field is taken into account. To this end, a high level reactive mission planner and a low level motion planning system are constructed. The high level system is responsible for task priority assignment and guiding the vehicle toward a target of interest considering on-time termination of the mission. The lower layer is in charge of generating optimal trajectories based on sequence of tasks and dynamicity of operating terrain. The mission planner is able to reactively re-arrange the tasks based on mission/terrain updates while the low level planner is capable of coping unexpected changes of the terrain by correcting the old path and re-generating a new trajectory. As a result, the vehicle is able to undertake the maximum number of tasks with certain degree of maneuverability having situational awareness of the operating field. The computational engine of the mentioned framework is based on the biogeography based optimization (BBO) algorithm that is capable of providing efficient solutions. To evaluate the performance of the proposed framework, firstly, a realistic model of undersea environment is provided based on realistic map data, and then several scenarios, treated as real experiments, are designed through the simulation study. Additionally, to show the robustness and reliability of the framework, Monte-Carlo simulation is carried out and statistical analysis is performed. The results of simulations indicate the significant potential of the two-level hierarchical mission planning system in mission success and its applicability for real-time implementation
    corecore