9 research outputs found

    Past, present and future of path-planning algorithms for mobile robot navigation in dynamic environments

    Get PDF
    Mobile robots have been making a significant contribution to the advancement of many sectors including automation of mining, space, surveillance, military, health, agriculture and many more. Safe and efficient navigation is a fundamental requirement of mobile robots, thus, the demand for advanced algorithms rapidly increased. Mobile robot navigation encompasses the following four requirements: perception, localization, path-planning and motion control. Among those, path-planning is a vital part of a fast, secure operation. During the last couple of decades, many path-planning algorithms were developed. Despite most of the mobile robot applications being in dynamic environments, the number of algorithms capable of navigating robots in dynamic environments is limited. This paper presents a qualitative comparative study of the up-to-date mobile robot path-planning methods capable of navigating robots in dynamic environments. The paper discusses both classical and heuristic methods including artificial potential field, genetic algorithm, fuzzy logic, neural networks, artificial bee colony, particle swarm optimization, bacterial foraging optimization, ant-colony and Agoraphilic algorithm. The general advantages and disadvantages of each method are discussed. Furthermore, the commonly used state-of-the-art methods are critically analyzed based on six performance criteria: algorithm's ability to navigate in dynamically cluttered areas, moving goal hunting ability, object tracking ability, object path prediction ability, incorporating the obstacle velocity in the decision, validation by simulation and experimentation. This investigation benefits researchers in choosing suitable path-planning methods for different applications as well as identifying gaps in this field. © 2020 IEEE

    Using reinforcement learning for optimizing the reproduction of tasks in robot programming by demonstration

    Get PDF
    As robots start pervading human environments, the need for new interfaces that would simplify human-robot interaction has become more pressing. Robot Programming by Demonstration (RbD) develops intuitive ways of programming robots, taking inspiration in strategies used by humans to transmit knowledge to apprentices. The user-friendliness of RbD is meant to allow lay users with no prior knowledge in computer science, electronics or mechanics to train robots to accomplish tasks the same way as they would with a co-worker. When a trainer teaches a task to a robot, he/she shows a particular way of fulfilling the task. For a robot to be able to learn from observing the trainer, it must be able to learn what the task entails (i.e. answer the so-called "What-to-imitate?" question), by inferring the user's intentions. But most importantly, the robot must be able to adapt its own controller to fit at best the demonstration (the so-called "How-to-imitate?" question) despite different setups and embodiments. The latter is the question that interested us in this thesis. It relates to the problem of optimizing the reproduction of the task under environmental constraints. The "How-to-imitate?" question is subdivided into two problems. The first problem, also known as the "correspondence problem", relates to resolving the discrepancy between the human demonstrator and robot's body that prevent the robot from doing an identical reproduction of the task. Even though we helped ourselves by considering solely humanoid platforms, that is platforms that have a joint configuration similar to that of the human, discrepancies in the number of degrees of freedom and range of motion remained. We resolved these by exploiting the redundant information conveyed through the demonstrations by collecting data through different frames of reference. By exploiting these redundancies in an algorithm comparable to the damped least square algorithm, we are able to reproduce a trajectory that minimizes the error between the desired trajectory and the reproduced trajectory across each frame of reference. The second problem consists in reproducing a trajectory in an unknown setup while respecting the task constraints learned during training. When the information learned from the demonstration no longer suffice to generalize the task constraints to a new set-up, the robot must re-learn the task; this time through trial-and-error. Here we considered the combination of trial-and-error learning to complement RbD. By adding a trial-and-error module to the original Imitation Learning algorithm, the robot can find a solution that is more adapted to the context and to its embodiment than the solution found using RbD. Specifically, we compared Reinforcement Learning (RL) – to other classical optimization techniques. We show that the system is advantageous in that: a) learning is more robust to unexpected events that have not been encountered during the demonstrations and b) the robot is able to optimize its own model of the task according to its own embodiment
    corecore