59 research outputs found

    Proprioceptive Robot Collision Detection through Gaussian Process Regression

    Full text link
    This paper proposes a proprioceptive collision detection algorithm based on Gaussian Regression. Compared to sensor-based collision detection and other proprioceptive algorithms, the proposed approach has minimal sensing requirements, since only the currents and the joint configurations are needed. The algorithm extends the standard Gaussian Process models adopted in learning the robot inverse dynamics, using a more rich set of input locations and an ad-hoc kernel structure to model the complex and non-linear behaviors due to frictions in quasi-static configurations. Tests performed on a Universal Robots UR10 show the effectiveness of the proposed algorithm to detect when a collision has occurred.Comment: Published at ACC 201

    Contact aware robust semi-autonomous teleoperation of mobile manipulators

    Get PDF
    In the context of human-robot collaboration, cooperation and teaming, the use of mobile manipulators is widespread on applications involving unpredictable or hazardous environments for humans operators, like space operations, waste management and search and rescue on disaster scenarios. Applications where the manipulator's motion is controlled remotely by specialized operators. Teleoperation of manipulators is not a straightforward task, and in many practical cases represent a common source of failures. Common issues during the remote control of manipulators are: increasing control complexity with respect the mechanical degrees of freedom; inadequate or incomplete feedback to the user (i.e. limited visualization or knowledge of the environment); predefined motion directives may be incompatible with constraints or obstacles imposed by the environment. In the latter case, part of the manipulator may get trapped or blocked by some obstacle in the environment, failure that cannot be easily detected, isolated nor counteracted remotely. While control complexity can be reduced by the introduction of motion directives or by abstraction of the robot motion, the real-time constraint of the teleoperation task requires the transfer of the least possible amount of data over the system's network, thus limiting the number of physical sensors that can be used to model the environment. Therefore, it is of fundamental to define alternative perceptive strategies to accurately characterize different interaction with the environment without relying on specific sensory technologies. In this work, we present a novel approach for safe teleoperation, that takes advantage of model based proprioceptive measurement of the robot dynamics to robustly identify unexpected collisions or contact events with the environment. Each identified collision is translated on-the-fly into a set of local motion constraints, allowing the exploitation of the system redundancies for the computation of intelligent control laws for automatic reaction, without requiring human intervention and minimizing the disturbance of the task execution (or, equivalently, the operator efforts). More precisely, the described system consist in two different building blocks. The first, for detecting unexpected interactions with the environment (perceptive block). The second, for intelligent and autonomous reaction after the stimulus (control block). The perceptive block is responsible of the contact event identification. In short, the approach is based on the claim that a sensorless collision detection method for robot manipulators can be extended to the field of mobile manipulators, by embedding it within a statistical learning framework. The control deals with the intelligent and autonomous reaction after the contact or impact with the environment occurs, and consist on an motion abstraction controller with a prioritized set of constrains, where the highest priority correspond to the robot reconfiguration after a collision is detected; when all related dynamical effects have been compensated, the controller switch again to the basic control mode

    Comparison of interaction modalities for mobile indoor robot guidance : direct physical interaction, person following, and pointing control

    Get PDF
    © 2015 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other worksThree advanced natural interaction modalities for mobile robot guidance in an indoor environment were developed and compared using two tasks and quantitative metrics to measure performance and workload. The first interaction modality is based on direct physical interaction requiring the human user to push the robot in order to displace it. The second and third interaction modalities exploit a 3-D vision-based human-skeleton tracking allowing the user to guide the robot by either walking in front of it or by pointing toward a desired location. In the first task, the participants were asked to guide the robot between different rooms in a simulated physical apartment requiring rough movement of the robot through designated areas. The second task evaluated robot guidance in the same environment through a set of waypoints, which required accurate movements. The three interaction modalities were implemented on a generic differential drive mobile platform equipped with a pan-tilt system and a Kinect camera. Task completion time and accuracy were used as metrics to assess the users’ performance, while the NASA-TLX questionnaire was used to evaluate the users’ workload. A study with 24 participants indicated that choice of interaction modality had significant effect on completion time (F(2,61)=84.874, p<0.001), accuracy (F(2,29)=4.937, p=0.016), and workload (F(2,68)=11.948, p<0.001). The direct physical interaction required less time, provided more accuracy and less workload than the two contactless interaction modalities. Between the two contactless interaction modalities, the person-following interaction mod- lity was systematically better than the pointing-control one: The participants completed the tasks faster with less workloadPeer ReviewedPostprint (author's final draft

    Force Sensorless Admittance Control for Teleoperation of Uncertain Robot Manipulator Using Neural Networks

    Get PDF
    In this paper, a force sensorless control scheme based on neural networks (NNs) is developed for interaction between robot manipulators and human arms in physical collision. In this scheme, the trajectory is generated by using geometry vector method with Kinect sensor. To comply with the external torque from the environment, this paper presents a sensorless admittance control approach in joint space based on an observer approach, which is used to estimate external torques applied by the operator. To deal with the tracking problem of the uncertain manipulator, an adaptive controller combined with the radial basis function NN (RBFNN) is designed. The RBFNN is used to compensate for uncertainties in the system. In order to achieve the prescribed tracking precision, an error transformation algorithm is integrated into the controller. The Lyapunov functions are used to analyze the stability of the control system. The experiments on the Baxter robot are carried out to demonstrate the effectiveness and correctness of the proposed control scheme

    Design and Evolution of a Modular Tensegrity Robot Platform

    Get PDF
    NASA Ames Research Center is developing a compliant modular tensegrity robotic platform for planetary exploration. In this paper we present the design and evolution of the platform's main hardware component, an untethered, robust tensegrity strut, with rich sensor feedback and cable actuation. Each strut is a complete robot, and multiple struts can be combined together to form a wide range of complex tensegrity robots. Our current goal for the tensegrity robotic platform is the development of SUPERball, a 6-strut icosahedron underactuated tensegrity robot aimed at dynamic locomotion for planetary exploration rovers and landers, but the aim is for the modular strut to enable a wide range of tensegrity morphologies. SUPERball is a second generation prototype, evolving from the tensegrity robot ReCTeR, which is also a modular, lightweight, highly compliant 6-strut tensegrity robot that was used to validate our physics based NASA Tensegrity Robot Toolkit (NTRT) simulator. Many hardware design parameters of the SUPERball were driven by locomotion results obtained in our validated simulator. These evolutionary explorations helped constrain motor torque and speed parameters, along with strut and string stress. As construction of the hardware has finalized, we have also used the same evolutionary framework to evolve controllers that respect the built hardware parameters

    Autonomous Robotic Systems in a Variable World:A Task-Centric approach based on Explainable Models

    Get PDF

    Autonomous Robotic Systems in a Variable World:A Task-Centric approach based on Explainable Models

    Get PDF

    Development of an autonomous mobile robot with planning and location in a structured environment

    Get PDF
    Mestrado de dupla diplomação com a UTFPR - Universidade Tecnológica Federal do ParanáWith the advance of technology mobile robots have been increasingly applied in the industry, performing repetitive work with high performance, and in environments that pose risks to human health. The present work plans and develops a mobile robot platform for the micromouse competition. The micromouse consists of a small autonomous mobile robot that, when placed in an unknown labyrinth, is able to map it, search for the best path between the starting point and the goal and travel it in the shortest possible time. To accomplish these tasks, the robot must be able to self-locate, map the maze as it traverses it and plan paths based on the map obtained. The developed self-localization method is based on the odometry, the laser sensors present in the robot and on a previous knowledge of the start point and the configuration of the environment. Several methodologies of locomotion in unknown environment and route planning are analyzed in order to obtain the combination with the best performance. In order to verify the results, the present work is developed in real environment, in 3D simulation and also with a hardware in the loop capability. Labyrinths from previous competitions are used as basis for comparing methodologies and validating results. At the end it presents the algorithm capable of fulfilling all the requirements of the micromouse competition together with the results of its evaluation run.Com o avanço da tecnologia, os robôs móveis têm sido cada vez mais aplicados na indústria, realizando trabalhos repetitivos com alto desempenho e em ambientes que expõem riscos à saúde humana. O presente trabalho planeja e desenvolve um robô móvel para a competição micromouse. O micromouse consiste em um pequeno robô autônomo que, ao ser colocado em um labirinto desconhecido, é capaz de mapeá-lo, procurar o melhor caminho entre o ponto de partida e o objetivo, e percorrê-lo no menor tempo possível. Para realizar estas tarefas, o robô deve ser capaz de se auto-localizar, mapear o labirinto enquanto o percorre e planejar caminhos com base no mapa obtido. O método de auto-localização desenvolvido baseia-se na odometria, nos sensores a laser presentes no robô e em um prévio conhecimento do ponto de início e da configuração do ambiente. Diversas metodologias de locomoção em ambiente desconhecido e planejamento de rotas são analisadas buscando-se obter a combinação com o melhor desempenho. Para averiguação de resultados o presente trabalho desenvolve-se em ambiente real e em simulação 3D com hardware in the loop. Labirintos de competições anteriores são utilizados de base para o comparativo de metodologias e validação de resultados. Ao final apresenta-se o algoritmo capaz de cumprir todas as exigências da competição micromouse juntamente com os resultados em sua corrida de avaliação
    • …
    corecore