2,508 research outputs found
Quantum Robot: Structure, Algorithms and Applications
A kind of brand-new robot, quantum robot, is proposed through fusing quantum
theory with robot technology. Quantum robot is essentially a complex quantum
system and it is generally composed of three fundamental parts: MQCU (multi
quantum computing units), quantum controller/actuator, and information
acquisition units. Corresponding to the system structure, several learning
control algorithms including quantum searching algorithm and quantum
reinforcement learning are presented for quantum robot. The theoretic results
show that quantum robot can reduce the complexity of O(N^2) in traditional
robot to O(N^(3/2)) using quantum searching algorithm, and the simulation
results demonstrate that quantum robot is also superior to traditional robot in
efficient learning by novel quantum reinforcement learning algorithm.
Considering the advantages of quantum robot, its some potential important
applications are also analyzed and prospected.Comment: 19 pages, 4 figures, 2 table
Fuzzy Predictive Controller for Mobile Robot Path Tracking
IFAC Intelligent Components and Instruments for Control Applications, Annecy, France 1997This paper presents a way of implementing a Model Based Predictive Controller (MBPC) for mobile robot path-tracking. The method uses a non-linear model of mobile robot dynamics and thus allows an accurate prediction of the future trajectories. Constraints on the maximum attainable angular velocity is also considered by the algorithm. A fuzzy approach is used to implement the MBPC. The fuzzy controller has been trained using a lookup-table scheme, where the database of fuzzy-rules has been obtained automatically from a set of input-output training patterns, computed with the predictive controller. Experimental results obtained when applying the fuzzy controller to a TRC labmate mobile platform are given in the paper.Ministerio de Ciencia y Tecnología TAP95-0307Ministerio de Ciencia y Tecnología TAP96-884C
Neural network controller against environment: A coevolutive approach to generalize robot navigation behavior
In this paper, a new coevolutive method, called Uniform Coevolution, is introduced to learn weights of a neural network controller in autonomous robots. An evolutionary strategy is used to learn high-performance reactive behavior for navigation and collisions avoidance. The introduction of coevolutive over evolutionary strategies allows evolving the environment, to learn a general behavior able to solve the problem in different environments. Using a traditional evolutionary strategy method, without coevolution, the learning process obtains a specialized behavior. All the behaviors obtained, with/without coevolution have been tested in a set of environments and the capability of generalization is shown for each learned behavior. A simulator based on a mini-robot Khepera has been used to learn each behavior. The results show that Uniform Coevolution obtains better generalized solutions to examples-based problems.Publicad
Robotic Olfactory-Based Navigation with Mobile Robots
Robotic odor source localization (OSL) is a technology that enables mobile robots or autonomous vehicles to find an odor source in unknown environments. It has been viewed as challenging due to the turbulent nature of airflows and the resulting odor plume characteristics. The key to correctly finding an odor source is designing an effective olfactory-based navigation algorithm, which guides the robot to detect emitted odor plumes as cues in finding the source. This dissertation proposes three kinds of olfactory-based navigation methods to improve search efficiency while maintaining a low computational cost, incorporating different machine learning and artificial intelligence methods.
A. Adaptive Bio-inspired Navigation via Fuzzy Inference Systems.
In nature, animals use olfaction to perform many life-essential activities, such as homing, foraging, mate-seeking, and evading predators. Inspired by the mate-seeking behaviors of male moths, this method presents a behavior-based navigation algorithm for using on a mobile robot to locate an odor source. Unlike traditional bio-inspired methods, which use fixed parameters to formulate robot search trajectories, a fuzzy inference system is designed to perceive the environment and adjust trajectory parameters based on the current search situation. The robot can automatically adapt the scale of search trajectories to fit environmental changes and balance the exploration and exploitation of the search.
B. Olfactory-based Navigation via Model-based Reinforcement Learning Methods.
This method analogizes the odor source localization as a reinforcement learning problem. During the odor plume tracing process, the belief state in a partially observable Markov decision process model is adapted to generate a source probability map that estimates possible odor source locations. A hidden Markov model is employed to produce a plume distribution map that premises plume propagation areas. Both source and plume estimates are fed to the robot. A decision-making model based on a fuzzy inference system is designed to dynamically fuse information from two maps and balance the exploitation and exploration of the search. After assigning the fused information to reward functions, a value iteration-based path planning algorithm solves the optimal action policy.
C. Robotic Odor Source Localization via Deep Learning-based Methods.
This method investigates the viability of implementing deep learning algorithms to solve the odor source localization problem. The primary objective is to obtain a deep learning model that guides a mobile robot to find an odor source without explicating search strategies. To achieve this goal, two kinds of deep learning models, including adaptive neuro-fuzzy inference system (ANFIS) and deep neural networks (DNNs), are employed to generate the olfactory-based navigation strategies. Multiple training data sets are acquired by applying two traditional methods in both simulation and on-vehicle tests to train deep learning models. After the supervised training, the deep learning models are verified with unseen search situations in simulation and real-world environments.
All proposed algorithms are implemented in simulation and on-vehicle tests to verify their effectiveness. Compared to traditional methods, experiment results show that the proposed algorithms outperform them in terms of the success rate and average search time. Finally, the future research directions are presented at the end of the dissertation
Discussion on Different Controllers Used for the Navigation of Mobile Robot
Robots that can comprehend and navigate their surroundings independently on their own are considered intelligent mobile robots (MR). Using a sophisticated set of controllers, artificial intelligence (AI), deep learning (DL), machine learning (ML), sensors, and computation for navigation, MR\u27s can understand and navigate around their environments without even being connected to a cabled source of power. Mobility and intelligence are fundamental drivers of autonomous robots that are intended for their planned operations. They are becoming popular in a variety of fields, including business, industry, healthcare, education, government, agriculture, military operations, and even domestic settings, to optimize everyday activities. We describe different controllers, including proportional integral derivative (PID) controllers, model predictive controllers (MPCs), fuzzy logic controllers (FLCs), and reinforcement learning controllers used in robotics science. The main objective of this article is to demonstrate a comprehensive idea and basic working principle of controllers utilized by mobile robots (MR) for navigation. This work thoroughly investigates several available books and literature to provide a better understanding of the navigation strategies taken by MR. Future research trends and possible challenges to optimizing the MR navigation system are also discussed
Decision tree learning for intelligent mobile robot navigation
The replication of human intelligence, learning and reasoning by means of computer
algorithms is termed Artificial Intelligence (Al) and the interaction of such
algorithms with the physical world can be achieved using robotics. The work described in
this thesis investigates the applications of concept learning (an approach which takes its
inspiration from biological motivations and from survival instincts in particular) to robot
control and path planning. The methodology of concept learning has been applied using
learning decision trees (DTs) which induce domain knowledge from a finite set of training
vectors which in turn describe systematically a physical entity and are used to train a robot
to learn new concepts and to adapt its behaviour.
To achieve behaviour learning, this work introduces the novel approach of hierarchical
learning and knowledge decomposition to the frame of the reactive robot architecture.
Following the analogy with survival instincts, the robot is first taught how to survive in
very simple and homogeneous environments, namely a world without any disturbances or
any kind of "hostility". Once this simple behaviour, named a primitive, has been established, the robot is trained to adapt new knowledge to cope with increasingly complex
environments by adding further worlds to its existing knowledge. The repertoire of the
robot behaviours in the form of symbolic knowledge is retained in a hierarchy of clustered
decision trees (DTs) accommodating a number of primitives. To classify robot perceptions,
control rules are synthesised using symbolic knowledge derived from searching the
hierarchy of DTs.
A second novel concept is introduced, namely that of multi-dimensional fuzzy associative
memories (MDFAMs). These are clustered fuzzy decision trees (FDTs) which are trained
locally and accommodate specific perceptual knowledge. Fuzzy logic is incorporated to
deal with inherent noise in sensory data and to merge conflicting behaviours of the DTs.
In this thesis, the feasibility of the developed techniques is illustrated in the robot
applications, their benefits and drawbacks are discussed
- …