391 research outputs found

    Motion planning for mobile manipulator with keeping manipulability

    Get PDF
    Our research goal is to realize a motion planning for an intelligent mobile manipulator. To plan a mobile manipulator's motion, it is popular that the base robot motion is regarded as manipulator's extra joints, and the whole system is considered as a redundant manipulator. In this case, the locomotion controller is a part of the manipulator controller. However, it is difficult to implement both controllers as one controller, in our implementation experience, because of difference of actuators' character. In this research, we focus on a path planning algorithm for a mobile base with keeping manipulability at the tip of the mounted manipulator. In this case, the locomotion controller is independent from the manipulator controller, and a cooperative motion is realized by a communication between both controllers. In this paper, we propose a motion planning algorithm for a mobile manipulator, and report several experimental results. </p

    End-effector vibrations reduction in trajectory tracking for mobile manipulator

    Get PDF
    A method of motion planning for a mobile manipulator taking into account damping the end-effector vibrations is presented. The primary task of the robot is to trace a given end-effector trajectory. The redundant degrees of freedom are used to fulfil secondary objectives such as minimisation of platform kinetic energy and maximisation of holonomic manipulability measure, which leads to reduction of the end-effector vibrations. The method is based on Jacobian pseudo inverse at the acceleration level. Nonholonomic constraints in a Pfaffian form are explicitly incorporated to the control algorithm. A computer example involving a mobile manipulator consisting of a nonholonomic platform (2, 0) class and SCARA-type holonomic manipulator operating in two-dimensional task space is also presented

    Contact aware robust semi-autonomous teleoperation of mobile manipulators

    Get PDF
    In the context of human-robot collaboration, cooperation and teaming, the use of mobile manipulators is widespread on applications involving unpredictable or hazardous environments for humans operators, like space operations, waste management and search and rescue on disaster scenarios. Applications where the manipulator's motion is controlled remotely by specialized operators. Teleoperation of manipulators is not a straightforward task, and in many practical cases represent a common source of failures. Common issues during the remote control of manipulators are: increasing control complexity with respect the mechanical degrees of freedom; inadequate or incomplete feedback to the user (i.e. limited visualization or knowledge of the environment); predefined motion directives may be incompatible with constraints or obstacles imposed by the environment. In the latter case, part of the manipulator may get trapped or blocked by some obstacle in the environment, failure that cannot be easily detected, isolated nor counteracted remotely. While control complexity can be reduced by the introduction of motion directives or by abstraction of the robot motion, the real-time constraint of the teleoperation task requires the transfer of the least possible amount of data over the system's network, thus limiting the number of physical sensors that can be used to model the environment. Therefore, it is of fundamental to define alternative perceptive strategies to accurately characterize different interaction with the environment without relying on specific sensory technologies. In this work, we present a novel approach for safe teleoperation, that takes advantage of model based proprioceptive measurement of the robot dynamics to robustly identify unexpected collisions or contact events with the environment. Each identified collision is translated on-the-fly into a set of local motion constraints, allowing the exploitation of the system redundancies for the computation of intelligent control laws for automatic reaction, without requiring human intervention and minimizing the disturbance of the task execution (or, equivalently, the operator efforts). More precisely, the described system consist in two different building blocks. The first, for detecting unexpected interactions with the environment (perceptive block). The second, for intelligent and autonomous reaction after the stimulus (control block). The perceptive block is responsible of the contact event identification. In short, the approach is based on the claim that a sensorless collision detection method for robot manipulators can be extended to the field of mobile manipulators, by embedding it within a statistical learning framework. The control deals with the intelligent and autonomous reaction after the contact or impact with the environment occurs, and consist on an motion abstraction controller with a prioritized set of constrains, where the highest priority correspond to the robot reconfiguration after a collision is detected; when all related dynamical effects have been compensated, the controller switch again to the basic control mode

    Safety-related Tasks within the Set-Based Task-Priority Inverse Kinematics Framework

    Full text link
    In this paper we present a framework that allows the motion control of a robotic arm automatically handling different kinds of safety-related tasks. The developed controller is based on a Task-Priority Inverse Kinematics algorithm that allows the manipulator's motion while respecting constraints defined either in the joint or in the operational space in the form of equality-based or set-based tasks. This gives the possibility to define, among the others, tasks as joint-limits, obstacle avoidance or limiting the workspace in the operational space. Additionally, an algorithm for the real-time computation of the minimum distance between the manipulator and other objects in the environment using depth measurements has been implemented, effectively allowing obstacle avoidance tasks. Experiments with a Jaco2^2 manipulator, operating in an environment where an RGB-D sensor is used for the obstacles detection, show the effectiveness of the developed system

    Path and Motion Planning for Autonomous Mobile 3D Printing

    Get PDF
    Autonomous robotic construction was envisioned as early as the ‘90s, and yet, con- struction sites today look much alike ones half a century ago. Meanwhile, highly automated and efficient fabrication methods like Additive Manufacturing, or 3D Printing, have seen great success in conventional production. However, existing efforts to transfer printing technology to construction applications mainly rely on manufacturing-like machines and fail to utilise the capabilities of modern robotics. This thesis considers using Mobile Manipulator robots to perform large-scale Additive Manufacturing tasks. Comprised of an articulated arm and a mobile base, Mobile Manipulators, are unique in their simultaneous mobility and agility, which enables printing-in-motion, or Mobile 3D Printing. This is a 3D printing modality, where a robot deposits material along larger-than-self trajectories while in motion. Despite profound potential advantages over existing static manufacturing-like large- scale printers, Mobile 3D printing is underexplored. Therefore, this thesis tack- les Mobile 3D printing-specific challenges and proposes path and motion planning methodologies that allow this printing modality to be realised. The work details the development of Task-Consistent Path Planning that solves the problem of find- ing a valid robot-base path needed to print larger-than-self trajectories. A motion planning and control strategy is then proposed, utilising the robot-base paths found to inform an optimisation-based whole-body motion controller. Several Mobile 3D Printing robot prototypes are built throughout this work, and the overall path and motion planning strategy proposed is holistically evaluated in a series of large-scale 3D printing experiments
    • …
    corecore