1,284 research outputs found

    From Rolling Over to Walking: Enabling Humanoid Robots to Develop Complex Motor Skills

    Full text link
    This paper presents an innovative method for humanoid robots to acquire a comprehensive set of motor skills through reinforcement learning. The approach utilizes an achievement-triggered multi-path reward function rooted in developmental robotics principles, facilitating the robot to learn gross motor skills typically mastered by human infants within a single training phase. The proposed method outperforms standard reinforcement learning techniques in success rates and learning speed within a simulation environment. By leveraging the principles of self-discovery and exploration integral to infant learning, this method holds the potential to significantly advance humanoid robot motor skill acquisition.Comment: 8 pages, 9 figures. Submitted to IEEE Robotics and Automation Letters. Video available at https://youtu.be/d0RqrW1Ezj

    Robot Learning from Human Demonstration: Interpretation, Adaptation, and Interaction

    Get PDF
    Robot Learning from Demonstration (LfD) is a research area that focuses on how robots can learn new skills by observing how people perform various activities. As humans, we have a remarkable ability to imitate other humanโ€™s behaviors and adapt to new situations. Endowing robots with these critical capabilities is a significant but very challenging problem considering the complexity and variation of human activities in highly dynamic environments. This research focuses on how robots can learn new skills by interpreting human activities, adapting the learned skills to new situations, and naturally interacting with humans. This dissertation begins with a discussion of challenges in each of these three problems. A new unified representation approach is introduced to enable robots to simultaneously interpret the high-level semantic meanings and generalize the low-level trajectories of a broad range of human activities. An adaptive framework based on feature space decomposition is then presented for robots to not only reproduce skills, but also autonomously and efficiently adjust the learned skills to new environments that are significantly different from demonstrations. To achieve natural Human Robot Interaction (HRI), this dissertation presents a Recurrent Neural Network based deep perceptual control approach, which is capable of integrating multi-modal perception sequences with actions for robots to interact with humans in long-term tasks. Overall, by combining the above approaches, an autonomous system is created for robots to acquire important skills that can be applied to human-centered applications. Finally, this dissertation concludes with a discussion of future directions that could accelerate the upcoming technological revolution of robot learning from human demonstration

    ์‹ฌ์ธต ๊ฐ•ํ™”ํ•™์Šต์„ ์ด์šฉํ•œ ์‚ฌ๋žŒ์˜ ๋ชจ์…˜์„ ํ†ตํ•œ ์ดํ˜•์  ์บ๋ฆญํ„ฐ ์ œ์–ด๊ธฐ ๊ฐœ๋ฐœ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2022. 8. ์„œ์ง„์šฑ.์‚ฌ๋žŒ์˜ ๋ชจ์…˜์„ ์ด์šฉํ•œ ๋กœ๋ด‡ ์ปจํŠธ๋กค ์ธํ„ฐํŽ˜์ด์Šค๋Š” ์‚ฌ์šฉ์ž์˜ ์ง๊ด€๊ณผ ๋กœ๋ด‡์˜ ๋ชจํ„ฐ ๋Šฅ๋ ฅ์„ ํ•ฉํ•˜์—ฌ ์œ„ํ—˜ํ•œ ํ™˜๊ฒฝ์—์„œ ๋กœ๋ด‡์˜ ์œ ์—ฐํ•œ ์ž‘๋™์„ ๋งŒ๋“ค์–ด๋‚ธ๋‹ค. ํ•˜์ง€๋งŒ, ํœด๋จธ๋…ธ์ด๋“œ ์™ธ์˜ ์‚ฌ์กฑ๋ณดํ–‰ ๋กœ๋ด‡์ด๋‚˜ ์œก์กฑ๋ณดํ–‰ ๋กœ๋ด‡์„ ์œ„ํ•œ ๋ชจ์…˜ ์ธํ„ฐํŽ˜์ด์Šค๋ฅผ ๋””์ž์ธ ํ•˜๋Š” ๊ฒƒ์€ ์‰ฌ์šด์ผ์ด ์•„๋‹ˆ๋‹ค. ์ด๊ฒƒ์€ ์‚ฌ๋žŒ๊ณผ ๋กœ๋ด‡ ์‚ฌ์ด์˜ ํ˜•ํƒœ ์ฐจ์ด๋กœ ์˜ค๋Š” ๋‹ค์ด๋‚˜๋ฏน์Šค ์ฐจ์ด์™€ ์ œ์–ด ์ „๋žต์ด ํฌ๊ฒŒ ์ฐจ์ด๋‚˜๊ธฐ ๋•Œ๋ฌธ์ด๋‹ค. ์šฐ๋ฆฌ๋Š” ์‚ฌ๋žŒ ์‚ฌ์šฉ์ž๊ฐ€ ์›€์ง์ž„์„ ํ†ตํ•˜์—ฌ ์‚ฌ์กฑ๋ณดํ–‰ ๋กœ๋ด‡์—์„œ ๋ถ€๋“œ๋Ÿฝ๊ฒŒ ์—ฌ๋Ÿฌ ๊ณผ์ œ๋ฅผ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ๊ฒŒ๋” ํ•˜๋Š” ์ƒˆ๋กœ์šด ๋ชจ์…˜ ์ œ์–ด ์‹œ์Šคํ…œ์„ ์ œ์•ˆํ•œ๋‹ค. ์šฐ๋ฆฌ๋Š” ์šฐ์„  ์บก์ณํ•œ ์‚ฌ๋žŒ์˜ ๋ชจ์…˜์„ ์ƒ์‘ํ•˜๋Š” ๋กœ๋ด‡์˜ ๋ชจ์…˜์œผ๋กœ ๋ฆฌํƒ€๊ฒŸ ์‹œํ‚จ๋‹ค. ์ด๋•Œ ์ƒ์‘ํ•˜๋Š” ๋กœ๋ด‡์˜ ๋ชจ์…˜์€ ์œ ์ €๊ฐ€ ์˜๋„ํ•œ ์˜๋ฏธ๋ฅผ ๋‚ดํฌํ•˜๊ฒŒ ๋˜๋ฉฐ, ์šฐ๋ฆฌ๋Š” ์ด๋ฅผ ์ง€๋„ํ•™์Šต ๋ฐฉ๋ฒ•๊ณผ ํ›„์ฒ˜๋ฆฌ ๊ธฐ์ˆ ์„ ์ด์šฉํ•˜์—ฌ ๊ฐ€๋Šฅ์ผ€ ํ•˜์˜€๋‹ค. ๊ทธ ๋’ค ์šฐ๋ฆฌ๋Š” ๋ชจ์…˜์„ ๋ชจ์‚ฌํ•˜๋Š” ํ•™์Šต์„ ์ปค๋ฆฌํ˜๋Ÿผ ํ•™์Šต๊ณผ ๋ณ‘ํ–‰ํ•˜์—ฌ ์ฃผ์–ด์ง„ ๋ฆฌํƒ€๊ฒŸ๋œ ์ฐธ์กฐ ๋ชจ์…˜์„ ๋”ฐ๋ผ๊ฐ€๋Š” ์ œ์–ด ์ •์ฑ…์„ ์ƒ์„ฑํ•˜์˜€๋‹ค. ์šฐ๋ฆฌ๋Š” "์ „๋ฌธ๊ฐ€ ์ง‘๋‹จ"์„ ํ•™์Šตํ•จ์œผ๋กœ ๋ชจ์…˜ ๋ฆฌํƒ€๊ฒŒํŒ… ๋ชจ๋“ˆ๊ณผ ๋ชจ์…˜ ๋ชจ์‚ฌ ๋ชจ๋“ˆ์˜ ์„ฑ๋Šฅ์„ ํฌ๊ฒŒ ์ฆ๊ฐ€์‹œ์ผฐ๋‹ค. ๊ฒฐ๊ณผ์—์„œ ๋ณผ ์ˆ˜ ์žˆ๋“ฏ, ์šฐ๋ฆฌ์˜ ์‹œ์Šคํ…œ์„ ์ด์šฉํ•˜์—ฌ ์‚ฌ์šฉ์ž๊ฐ€ ์‚ฌ์กฑ๋ณดํ–‰ ๋กœ๋ด‡์˜ ์„œ์žˆ๊ธฐ, ์•‰๊ธฐ, ๊ธฐ์šธ์ด๊ธฐ, ํŒ” ๋ป—๊ธฐ, ๊ฑท๊ธฐ, ๋Œ๊ธฐ์™€ ๊ฐ™์€ ๋‹ค์–‘ํ•œ ๋ชจํ„ฐ ๊ณผ์ œ๋“ค์„ ์‹œ๋ฎฌ๋ ˆ์ด์…˜ ํ™˜๊ฒฝ๊ณผ ํ˜„์‹ค์—์„œ ๋‘˜ ๋‹ค ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์šฐ๋ฆฌ๋Š” ์—ฐ๊ตฌ์˜ ์„ฑ๋Šฅ์„ ํ‰๊ฐ€ํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ๋‹ค์–‘ํ•œ ๋ถ„์„์„ ํ•˜์˜€์œผ๋ฉฐ, ํŠนํžˆ ์šฐ๋ฆฌ ์‹œ์Šคํ…œ์˜ ๊ฐ๊ฐ์˜ ์š”์†Œ๋“ค์˜ ์ค‘์š”์„ฑ์„ ๋ณด์—ฌ์ค„ ์ˆ˜ ์žˆ๋Š” ์‹คํ—˜๋“ค์„ ์ง„ํ–‰ํ•˜์˜€๋‹ค.A human motion-based interface fuses operator intuitions with the motor capabilities of robots, enabling adaptable robot operations in dangerous environments. However, the challenge of designing a motion interface for non-humanoid robots, such as quadrupeds or hexapods, is emerged from the different morphology and dynamics of a human controller, leading to an ambiguity of control strategy. We propose a novel control framework that allows human operators to execute various motor skills on a quadrupedal robot by their motion. Our system first retargets the captured human motion into the corresponding robot motion with the operator's intended semantics. The supervised learning and post-processing techniques allow this retargeting skill which is ambiguity-free and suitable for control policy training. To enable a robot to track a given retargeted motion, we then obtain the control policy from reinforcement learning that imitates the given reference motion with designed curriculums. We additionally enhance the system's performance by introducing a set of experts. Finally, we randomize the domain parameters to adapt the physically simulated motor skills to real-world tasks. We demonstrate that a human operator can perform various motor tasks using our system including standing, tilting, manipulating, sitting, walking, and steering on both physically simulated and real quadruped robots. We also analyze the performance of each system component ablation study.1 Introduction 1 2 Related Work 5 2.1 Legged Robot Control 5 2.2 Motion Imitation 6 2.3 Motion-based Control 7 3 Overview 9 4 Motion Retargeting Module 11 4.1 Motion Retargeting Network 12 4.2 Post-processing for Consistency 14 4.3 A Set of Experts for Multi-task Support 15 5 Motion Imitation Module 17 5.1 Background: Reinforcement Learning 18 5.2 Formulation of Motion Imitation 18 5.3 Curriculum Learning over Tasks and Difficulties 21 5.4 Hierarchical Control with States 21 5.5 Domain Randomization 22 6 Results and Analysis 23 6.1 Experimental Setup 23 6.2 Motion Performance 24 6.3 Analysis 28 6.4 Comparison to Other Methods 31 7 Conclusion And Future Work 32 Bibliography 34 Abstract (In Korean) 44 ๊ฐ์‚ฌ์˜ ๊ธ€ 45์„

    CasIL: Cognizing and Imitating Skills via a Dual Cognition-Action Architecture

    Full text link
    Enabling robots to effectively imitate expert skills in longhorizon tasks such as locomotion, manipulation, and more, poses a long-standing challenge. Existing imitation learning (IL) approaches for robots still grapple with sub-optimal performance in complex tasks. In this paper, we consider how this challenge can be addressed within the human cognitive priors. Heuristically, we extend the usual notion of action to a dual Cognition (high-level)-Action (low-level) architecture by introducing intuitive human cognitive priors, and propose a novel skill IL framework through human-robot interaction, called Cognition-Action-based Skill Imitation Learning (CasIL), for the robotic agent to effectively cognize and imitate the critical skills from raw visual demonstrations. CasIL enables both cognition and action imitation, while high-level skill cognition explicitly guides low-level primitive actions, providing robustness and reliability to the entire skill IL process. We evaluated our method on MuJoCo and RLBench benchmarks, as well as on the obstacle avoidance and point-goal navigation tasks for quadrupedal robot locomotion. Experimental results show that our CasIL consistently achieves competitive and robust skill imitation capability compared to other counterparts in a variety of long-horizon robotic tasks

    Leveraging Sequentiality in Reinforcement Learning from a Single Demonstration

    Full text link
    Deep Reinforcement Learning has been successfully applied to learn robotic control. However, the corresponding algorithms struggle when applied to problems where the agent is only rewarded after achieving a complex task. In this context, using demonstrations can significantly speed up the learning process, but demonstrations can be costly to acquire. In this paper, we propose to leverage a sequential bias to learn control policies for complex robotic tasks using a single demonstration. To do so, our method learns a goal-conditioned policy to control a system between successive low-dimensional goals. This sequential goal-reaching approach raises a problem of compatibility between successive goals: we need to ensure that the state resulting from reaching a goal is compatible with the achievement of the following goals. To tackle this problem, we present a new algorithm called DCIL-II. We show that DCIL-II can solve with unprecedented sample efficiency some challenging simulated tasks such as humanoid locomotion and stand-up as well as fast running with a simulated Cassie robot. Our method leveraging sequentiality is a step towards the resolution of complex robotic tasks under minimal specification effort, a key feature for the next generation of autonomous robots

    A Continuous Grasp Representation for the Imitation Learning of Grasps on Humanoid Robots

    Get PDF
    Models and methods are presented which enable a humanoid robot to learn reusable, adaptive grasping skills. Mechanisms and principles in human grasp behavior are studied. The findings are used to develop a grasp representation capable of retaining specific motion characteristics and of adapting to different objects and tasks. Based on the representation a framework is proposed which enables the robot to observe human grasping, learn grasp representations, and infer executable grasping actions

    Fractional Control of a Humanoid Robot Reduced Model with Model Disturbances

    Get PDF
    There is an open discussion between those who defend mass-distributed models for humanoid robots and those in favor of simple concentrated models. Even though each of them has its advantages and disadvantages, little research has been conducted analyzing the control performance due to the mismatch between the model and the real robot, and how the simplifications affect the controller's output. In this article we address this problem by combining a reduced model of the humanoid robot, which has an easier mathematical formulation and implementation, with a fractional order controller, which is robust to changes in the model parameters. This controller is a generalization of the well-known proportional-integral-derivative (PID) structure obtained from the application of Fractional Calculus for control, as will be discussed in this article. This control strategy guarantees the robustness of the system, minimizing the effects from the assumption that the robot has a simple mass distribution. The humanoid robot is modeled and identified as a triple inverted pendulum and, using a gain scheduling strategy, the performances of a classical PID controller and a fractional order PID controller are compared, tuning the controller parameters with a genetic algorithm.The research leading to these results has received funding from the ARCADIA project DPI2010-21047- C02-01, funded by CICYT project grant on behalf of Spanish Ministry of Economy and Competitiveness, and from the RoboCity2030-II-CM project (S2009/DPI-1559), funded by the Research and Development Work Programme of the Community of Madrid and cofunded by Structural Funds of the EU.Publicad
    • โ€ฆ
    corecore