500 research outputs found
More than a Million Ways to Be Pushed: A High-Fidelity Experimental Dataset of Planar Pushing
Pushing is a motion primitive useful to handle objects that are too large,
too heavy, or too cluttered to be grasped. It is at the core of much of robotic
manipulation, in particular when physical interaction is involved. It seems
reasonable then to wish for robots to understand how pushed objects move.
In reality, however, robots often rely on approximations which yield models
that are computable, but also restricted and inaccurate. Just how close are
those models? How reasonable are the assumptions they are based on? To help
answer these questions, and to get a better experimental understanding of
pushing, we present a comprehensive and high-fidelity dataset of planar pushing
experiments. The dataset contains timestamped poses of a circular pusher and a
pushed object, as well as forces at the interaction.We vary the push
interaction in 6 dimensions: surface material, shape of the pushed object,
contact position, pushing direction, pushing speed, and pushing acceleration.
An industrial robot automates the data capturing along precisely controlled
position-velocity-acceleration trajectories of the pusher, which give dense
samples of positions and forces of uniform quality.
We finish the paper by characterizing the variability of friction, and
evaluating the most common assumptions and simplifications made by models of
frictional pushing in robotics.Comment: 8 pages, 10 figure
A probabilistic data-driven model for planar pushing
This paper presents a data-driven approach to model planar pushing
interaction to predict both the most likely outcome of a push and its expected
variability. The learned models rely on a variation of Gaussian processes with
input-dependent noise called Variational Heteroscedastic Gaussian processes
(VHGP) that capture the mean and variance of a stochastic function. We show
that we can learn accurate models that outperform analytical models after less
than 100 samples and saturate in performance with less than 1000 samples. We
validate the results against a collected dataset of repeated trajectories, and
use the learned models to study questions such as the nature of the variability
in pushing, and the validity of the quasi-static assumption.Comment: 8 pages, 11 figures, ICRA 201
Friction Variability in Planar Pushing Data: Anisotropic Friction and Data-collection Bias
Friction plays a key role in manipulating objects. Most of what we do with
our hands, and most of what robots do with their grippers, is based on the
ability to control frictional forces. This paper aims to better understand the
variability and predictability of planar friction. In particular, we focus on
the analysis of a recent dataset on planar pushing by Yu et al. [1] devised to
create a data-driven footprint of planar friction.
We show in this paper how we can explain a significant fraction of the
observed unconventional phenomena, e.g., stochasticity and multi-modality, by
combining the effects of material non-homogeneity, anisotropy of friction and
biases due to data collection dynamics, hinting that the variability is
explainable but inevitable in practice.
We introduce an anisotropic friction model and conduct simulation experiments
comparing with more standard isotropic friction models. The anisotropic
friction between object and supporting surface results in convergence of
initial condition during the automated data collection. Numerical results
confirm that the anisotropic friction model explains the bias in the dataset
and the apparent stochasticity in the outcome of a push. The fact that the data
collection process itself can originate biases in the collected datasets,
resulting in deterioration of trained models, calls attention to the data
collection dynamics.Comment: 8 pages, 13 figure
Realtime State Estimation with Tactile and Visual sensing. Application to Planar Manipulation
Accurate and robust object state estimation enables successful object
manipulation. Visual sensing is widely used to estimate object poses. However,
in a cluttered scene or in a tight workspace, the robot's end-effector often
occludes the object from the visual sensor. The robot then loses visual
feedback and must fall back on open-loop execution.
In this paper, we integrate both tactile and visual input using a framework
for solving the SLAM problem, incremental smoothing and mapping (iSAM), to
provide a fast and flexible solution. Visual sensing provides global pose
information but is noisy in general, whereas contact sensing is local, but its
measurements are more accurate relative to the end-effector. By combining them,
we aim to exploit their advantages and overcome their limitations. We explore
the technique in the context of a pusher-slider system. We adapt iSAM's
measurement cost and motion cost to the pushing scenario, and use an
instrumented setup to evaluate the estimation quality with different object
shapes, on different surface materials, and under different contact modes
Experimental Validation of Contact Dynamics for In-Hand Manipulation
This paper evaluates state-of-the-art contact models at predicting the
motions and forces involved in simple in-hand robotic manipulations. In
particular it focuses on three primitive actions --linear sliding, pivoting,
and rolling-- that involve contacts between a gripper, a rigid object, and
their environment. The evaluation is done through thousands of controlled
experiments designed to capture the motion of object and gripper, and all
contact forces and torques at 250Hz. We demonstrate that a contact modeling
approach based on Coulomb's friction law and maximum energy principle is
effective at reasoning about interaction to first order, but limited for making
accurate predictions. We attribute the major limitations to 1) the
non-uniqueness of force resolution inherent to grasps with multiple hard
contacts of complex geometries, 2) unmodeled dynamics due to contact
compliance, and 3) unmodeled geometries dueto manufacturing defects.Comment: International Symposium on Experimental Robotics, ISER 2016, Tokyo,
Japa
Combining Physical Simulators and Object-Based Networks for Control
Physics engines play an important role in robot planning and control;
however, many real-world control problems involve complex contact dynamics that
cannot be characterized analytically. Most physics engines therefore employ .
approximations that lead to a loss in precision. In this paper, we propose a
hybrid dynamics model, simulator-augmented interaction networks (SAIN),
combining a physics engine with an object-based neural network for dynamics
modeling. Compared with existing models that are purely analytical or purely
data-driven, our hybrid model captures the dynamics of interacting objects in a
more accurate and data-efficient manner.Experiments both in simulation and on a
real robot suggest that it also leads to better performance when used in
complex control tasks. Finally, we show that our model generalizes to novel
environments with varying object shapes and materials.Comment: ICRA 2019; Project page: http://sain.csail.mit.ed
- …