35,215 research outputs found

    Quantum Robot: Structure, Algorithms and Applications

    Full text link
    A kind of brand-new robot, quantum robot, is proposed through fusing quantum theory with robot technology. Quantum robot is essentially a complex quantum system and it is generally composed of three fundamental parts: MQCU (multi quantum computing units), quantum controller/actuator, and information acquisition units. Corresponding to the system structure, several learning control algorithms including quantum searching algorithm and quantum reinforcement learning are presented for quantum robot. The theoretic results show that quantum robot can reduce the complexity of O(N^2) in traditional robot to O(N^(3/2)) using quantum searching algorithm, and the simulation results demonstrate that quantum robot is also superior to traditional robot in efficient learning by novel quantum reinforcement learning algorithm. Considering the advantages of quantum robot, its some potential important applications are also analyzed and prospected.Comment: 19 pages, 4 figures, 2 table

    MOTION CONTROL SIMULATION OF A HEXAPOD ROBOT

    Get PDF
    This thesis addresses hexapod robot motion control. Insect morphology and locomotion patterns inform the design of a robotic model, and motion control is achieved via trajectory planning and bio-inspired principles. Additionally, deep learning and multi-agent reinforcement learning are employed to train the robot motion control strategy with leg coordination achieves using a multi-agent deep reinforcement learning framework. The thesis makes the following contributions: First, research on legged robots is synthesized, with a focus on hexapod robot motion control. Insect anatomy analysis informs the hexagonal robot body and three-joint single robotic leg design, which is assembled using SolidWorks. Different gaits are studied and compared, and robot leg kinematics are derived and experimentally verified, culminating in a three-legged gait for motion control. Second, an animal-inspired approach employs a central pattern generator (CPG) control unit based on the Hopf oscillator, facilitating robot motion control in complex environments such as stable walking and climbing. The robot\u27s motion process is quantitatively evaluated in terms of displacement change and body pitch angle. Third, a value function decomposition algorithm, QPLEX, is applied to hexapod robot motion control. The QPLEX architecture treats each leg as a separate agent with local control modules, that are trained using reinforcement learning. QPLEX outperforms decentralized approaches, achieving coordinated rhythmic gaits and increased robustness on uneven terrain. The significant of terrain curriculum learning is assessed, with QPLEX demonstrating superior stability and faster consequence. The foot-end trajectory planning method enables robot motion control through inverse kinematic solutions but has limited generalization capabilities for diverse terrains. The animal-inspired CPG-based method offers a versatile control strategy but is constrained to core aspects. In contrast, the multi-agent deep reinforcement learning-based approach affords adaptable motion strategy adjustments, rendering it a superior control policy. These methods can be combined to develop a customized robot motion control policy for specific scenarios

    Bolts detection and a combination of conventional and reinforcement learning based control of UR5 industrial robot

    Get PDF
    Master's thesis in Mechatronics (MAS500)The main objective of this paper is to investigate the possibilities for using reinforcement learning to control a UR-5 robot. The paper also looks at how well reinforcement learning works to control a UR-5 robot. These questions are answered by constructing of matlab and simulink programes. Based on different mathworks example programs and scripts. In this study, reinforcement learning only works in the situation it is trained to perform. The author believe that it could work better if it were given other configurations/parameters. This will still be an interesting subject for further studies. According to the research done in this paper, the conventional control have the best control accuracy

    Vision-based reinforcement learning using approximate policy iteration

    Get PDF
    A major issue for reinforcement learning (RL) applied to robotics is the time required to learn a new skill. While RL has been used to learn mobile robot control in many simulated domains, applications involving learning on real robots are still relatively rare. In this paper, the Least-Squares Policy Iteration (LSPI) reinforcement learning algorithm and a new model-based algorithm Least-Squares Policy Iteration with Prioritized Sweeping (LSPI+), are implemented on a mobile robot to acquire new skills quickly and efficiently. LSPI+ combines the benefits of LSPI and prioritized sweeping, which uses all previous experience to focus the computational effort on the most ā€œinterestingā€ or dynamic parts of the state space. The proposed algorithms are tested on a household vacuum cleaner robot for learning a docking task using vision as the only sensor modality. In experiments these algorithms are compared to other model-based and model-free RL algorithms. The results show that the number of trials required to learn the docking task is significantly reduced using LSPI compared to the other RL algorithms investigated, and that LSPI+ further improves on the performance of LSPI
    • ā€¦
    corecore