648 research outputs found
Shape and Pose Recovery from Planar Pushing
Tactile exploration refers to the use of physical interaction to infer object properties. In this work, we study the feasibility of recovering the shape and pose of a movable object from observing a series of contacts. In particular, we approach the problem of estimating the shape and trajectory of a planar object lying on a frictional surface, and being pushed by a frictional probe. The probe, when in contact with the object, makes observations of the location of contact and the contact normal.
Our approach draws inspiration from the SLAM problem, where noisy observations of the location of landmarks are used to reconstruct and locate a static environment. In tactile exploration, analogously, we can think of the object as a rigid but moving environment, and of the pusher as a sensor that reports contact points on the boundary of the object.
A key challenge to tactile exploration is that, unlike visual feedback, sensing by touch is intrusive in nature. The object moves by the action of sensing. In the 2D version of the problem that we study in this paper, the well understood mechanics of planar frictional pushing provides a motion model that plays the role of odometry. The conjecture we investigate in this paper is whether the models of frictional pushing are sufficiently descriptive to simultaneously estimate the shape and pose of an object from the cumulative effect of a sequence of pushes.National Science Foundation (U.S.) (Award IIS-1427050
More than a Million Ways to Be Pushed: A High-Fidelity Experimental Dataset of Planar Pushing
Pushing is a motion primitive useful to handle objects that are too large,
too heavy, or too cluttered to be grasped. It is at the core of much of robotic
manipulation, in particular when physical interaction is involved. It seems
reasonable then to wish for robots to understand how pushed objects move.
In reality, however, robots often rely on approximations which yield models
that are computable, but also restricted and inaccurate. Just how close are
those models? How reasonable are the assumptions they are based on? To help
answer these questions, and to get a better experimental understanding of
pushing, we present a comprehensive and high-fidelity dataset of planar pushing
experiments. The dataset contains timestamped poses of a circular pusher and a
pushed object, as well as forces at the interaction.We vary the push
interaction in 6 dimensions: surface material, shape of the pushed object,
contact position, pushing direction, pushing speed, and pushing acceleration.
An industrial robot automates the data capturing along precisely controlled
position-velocity-acceleration trajectories of the pusher, which give dense
samples of positions and forces of uniform quality.
We finish the paper by characterizing the variability of friction, and
evaluating the most common assumptions and simplifications made by models of
frictional pushing in robotics.Comment: 8 pages, 10 figure
Real-Time Physics-Based Object Pose Tracking during Non-Prehensile Manipulation
We propose a method to track the 6D pose of an object over time, while the
object is under non-prehensile manipulation by a robot. At any given time
during the manipulation of the object, we assume access to the robot joint
controls and an image from a camera looking at the scene. We use the robot
joint controls to perform a physics-based prediction of how the object might be
moving. We then combine this prediction with the observation coming from the
camera, to estimate the object pose as accurately as possible. We use a
particle filtering approach to combine the control information with the visual
information. We compare the proposed method with two baselines: (i) using only
an image-based pose estimation system at each time-step, and (ii) a particle
filter which does not perform the computationally expensive physics
predictions, but assumes the object moves with constant velocity. Our results
show that making physics-based predictions is worth the computational cost,
resulting in more accurate tracking, and estimating object pose even when the
object is not clearly visible to the camera
Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age
Simultaneous Localization and Mapping (SLAM)consists in the concurrent
construction of a model of the environment (the map), and the estimation of the
state of the robot moving within it. The SLAM community has made astonishing
progress over the last 30 years, enabling large-scale real-world applications,
and witnessing a steady transition of this technology to industry. We survey
the current state of SLAM. We start by presenting what is now the de-facto
standard formulation for SLAM. We then review related work, covering a broad
set of topics including robustness and scalability in long-term mapping, metric
and semantic representations for mapping, theoretical performance guarantees,
active SLAM and exploration, and other new frontiers. This paper simultaneously
serves as a position paper and tutorial to those who are users of SLAM. By
looking at the published research with a critical eye, we delineate open
challenges and new research issues, that still deserve careful scientific
investigation. The paper also contains the authors' take on two questions that
often animate discussions during robotics conferences: Do robots need SLAM? and
Is SLAM solved
- …