35,215 research outputs found
Quantum Robot: Structure, Algorithms and Applications
A kind of brand-new robot, quantum robot, is proposed through fusing quantum
theory with robot technology. Quantum robot is essentially a complex quantum
system and it is generally composed of three fundamental parts: MQCU (multi
quantum computing units), quantum controller/actuator, and information
acquisition units. Corresponding to the system structure, several learning
control algorithms including quantum searching algorithm and quantum
reinforcement learning are presented for quantum robot. The theoretic results
show that quantum robot can reduce the complexity of O(N^2) in traditional
robot to O(N^(3/2)) using quantum searching algorithm, and the simulation
results demonstrate that quantum robot is also superior to traditional robot in
efficient learning by novel quantum reinforcement learning algorithm.
Considering the advantages of quantum robot, its some potential important
applications are also analyzed and prospected.Comment: 19 pages, 4 figures, 2 table
MOTION CONTROL SIMULATION OF A HEXAPOD ROBOT
This thesis addresses hexapod robot motion control. Insect morphology and locomotion patterns inform the design of a robotic model, and motion control is achieved via trajectory planning and bio-inspired principles. Additionally, deep learning and multi-agent reinforcement learning are employed to train the robot motion control strategy with leg coordination achieves using a multi-agent deep reinforcement learning framework. The thesis makes the following contributions:
First, research on legged robots is synthesized, with a focus on hexapod robot motion control. Insect anatomy analysis informs the hexagonal robot body and three-joint single robotic leg design, which is assembled using SolidWorks. Different gaits are studied and compared, and robot leg kinematics are derived and experimentally verified, culminating in a three-legged gait for motion control.
Second, an animal-inspired approach employs a central pattern generator (CPG) control unit based on the Hopf oscillator, facilitating robot motion control in complex environments such as stable walking and climbing. The robot\u27s motion process is quantitatively evaluated in terms of displacement change and body pitch angle.
Third, a value function decomposition algorithm, QPLEX, is applied to hexapod robot motion control. The QPLEX architecture treats each leg as a separate agent with local control modules, that are trained using reinforcement learning. QPLEX outperforms decentralized approaches, achieving coordinated rhythmic gaits and increased robustness on uneven terrain. The significant of terrain curriculum learning is assessed, with QPLEX demonstrating superior stability and faster consequence.
The foot-end trajectory planning method enables robot motion control through inverse kinematic solutions but has limited generalization capabilities for diverse terrains. The animal-inspired CPG-based method offers a versatile control strategy but is constrained to core aspects. In contrast, the multi-agent deep reinforcement learning-based approach affords adaptable motion strategy adjustments, rendering it a superior control policy. These methods can be combined to develop a customized robot motion control policy for specific scenarios
Bolts detection and a combination of conventional and reinforcement learning based control of UR5 industrial robot
Master's thesis in Mechatronics (MAS500)The main objective of this paper is to investigate the possibilities for using reinforcement learning to control a UR-5 robot. The paper also looks at how well reinforcement learning works to control a UR-5 robot. These questions are answered by constructing of matlab and simulink programes. Based on different mathworks example programs and scripts. In this study, reinforcement learning only works in the situation it is trained to perform. The author believe that it could work better if it were given other configurations/parameters. This will still be an interesting subject for further studies. According to the research done in this paper, the conventional control have the best control accuracy
Recommended from our members
Reinforcement learning for human-robot shared control
This paper aims at proposing a general framework of shared control for human-robot interaction. Human dynamics are considered in analysis of the coupled human-robot system. Motion intentions of both human and robot are taken into account in the control objective of the robot. Reinforcement learning is developed to achieve the control objective subject to unknown dynamics of human and robot. The closed-loop system performance is discussed through a rigorous proof. Simulations are conducted to demonstrate the learning capability of the proposed method and its feasibility in handling various situations. Compared to existing works, the proposed framework combines motion intentions of both human and robot in a human-robot shared control system, without the requirement of the knowledge of humans and robots dynamics
Vision-based reinforcement learning using approximate policy iteration
A major issue for reinforcement learning (RL) applied to robotics is the time required to learn a new skill. While RL has been used to learn mobile robot control in many simulated domains, applications involving learning on real
robots are still relatively rare. In this paper, the Least-Squares Policy Iteration (LSPI) reinforcement learning algorithm and a new model-based algorithm Least-Squares Policy Iteration with Prioritized Sweeping (LSPI+), are implemented on a mobile robot to acquire new skills quickly and efficiently. LSPI+ combines the benefits of LSPI and prioritized sweeping, which uses all previous experience to focus the computational effort on the most āinterestingā or dynamic parts of the state space.
The proposed algorithms are tested on a household vacuum
cleaner robot for learning a docking task using vision as the only sensor modality. In experiments these algorithms are compared to other model-based and model-free RL algorithms. The results show that the number of trials required to learn the docking task is significantly reduced using LSPI compared to the other RL algorithms investigated, and that LSPI+ further improves on the performance of LSPI
- ā¦