10,900 research outputs found
Object Manipulation and Control with Robotic Hand
From industrial robots to nursing robots, object manipulation has become a growing area of robotics research. This Major Qualifying Project explores methods of teleoperation through the use of a wireless data glove able to detect multiple degrees of freedom. Our project also explored methods for autonomous control. We developed a computer vision model by integrating two state-of-the-art Mask Region Convolutional Neural Networks (Mask-RCNN) models to create a final model for determining both object location and grasp angle. This modeling allows the Baxter Robot to autonomously detect and reach towards the object. Using learning by demonstration, the robot can learn how to grasp and manipulate said objects
Improving Task-Parameterised Movement Learning Generalisation with Frame-Weighted Trajectory Generation
Learning from Demonstration depends on a robot learner generalising its
learned model to unseen conditions, as it is not feasible for a person to
provide a demonstration set that accounts for all possible variations in
non-trivial tasks. While there are many learning methods that can handle
interpolation of observed data effectively, extrapolation from observed data
offers a much greater challenge. To address this problem of generalisation,
this paper proposes a modified Task-Parameterised Gaussian Mixture Regression
method that considers the relevance of task parameters during trajectory
generation, as determined by variance in the data. The benefits of the proposed
method are first explored using a simulated reaching task data set. Here it is
shown that the proposed method offers far-reaching, low-error extrapolation
abilities that are different in nature to existing learning methods. Data
collected from novice users for a real-world manipulation task is then
considered, where it is shown that the proposed method is able to effectively
reduce grasping performance errors by and extrapolate to unseen
grasp targets under real-world conditions. These results indicate the proposed
method serves to benefit novice users by placing less reliance on the user to
provide high quality demonstration data sets.Comment: 8 pages, 6 figures, submitted to 2019 IEEE/RSJ International
Conference on Intelligent Robots and Systems (IROS
Data-Driven Grasp Synthesis - A Survey
We review the work on data-driven grasp synthesis and the methodologies for
sampling and ranking candidate grasps. We divide the approaches into three
groups based on whether they synthesize grasps for known, familiar or unknown
objects. This structure allows us to identify common object representations and
perceptual processes that facilitate the employed data-driven grasp synthesis
technique. In the case of known objects, we concentrate on the approaches that
are based on object recognition and pose estimation. In the case of familiar
objects, the techniques use some form of a similarity matching to a set of
previously encountered objects. Finally for the approaches dealing with unknown
objects, the core part is the extraction of specific features that are
indicative of good grasps. Our survey provides an overview of the different
methodologies and discusses open problems in the area of robot grasping. We
also draw a parallel to the classical approaches that rely on analytic
formulations.Comment: 20 pages, 30 Figures, submitted to IEEE Transactions on Robotic
- …