91 research outputs found
Robust Legged Robot State Estimation Using Factor Graph Optimization
Legged robots, specifically quadrupeds, are becoming increasingly attractive
for industrial applications such as inspection. However, to leave the
laboratory and to become useful to an end user requires reliability in harsh
conditions. From the perspective of state estimation, it is essential to be
able to accurately estimate the robot's state despite challenges such as uneven
or slippery terrain, textureless and reflective scenes, as well as dynamic
camera occlusions. We are motivated to reduce the dependency on foot contact
classifications, which fail when slipping, and to reduce position drift during
dynamic motions such as trotting. To this end, we present a factor graph
optimization method for state estimation which tightly fuses and smooths
inertial navigation, leg odometry and visual odometry. The effectiveness of the
approach is demonstrated using the ANYmal quadruped robot navigating in a
realistic outdoor industrial environment. This experiment included trotting,
walking, crossing obstacles and ascending a staircase. The proposed approach
decreased the relative position error by up to 55% and absolute position error
by 76% compared to kinematic-inertial odometry.Comment: 8 pages, 12 figures. Accepted to RA-L + IROS 2019, July 201
Learning to See the Wood for the Trees: Deep Laser Localization in Urban and Natural Environments on a CPU
Localization in challenging, natural environments such as forests or
woodlands is an important capability for many applications from guiding a robot
navigating along a forest trail to monitoring vegetation growth with handheld
sensors. In this work we explore laser-based localization in both urban and
natural environments, which is suitable for online applications. We propose a
deep learning approach capable of learning meaningful descriptors directly from
3D point clouds by comparing triplets (anchor, positive and negative examples).
The approach learns a feature space representation for a set of segmented point
clouds that are matched between a current and previous observations. Our
learning method is tailored towards loop closure detection resulting in a small
model which can be deployed using only a CPU. The proposed learning method
would allow the full pipeline to run on robots with limited computational
payload such as drones, quadrupeds or UGVs.Comment: Accepted for publication at RA-L/ICRA 2019. More info:
https://ori.ox.ac.uk/esm-localizatio
Preintegrated Velocity Bias Estimation to Overcome Contact Nonlinearities in Legged Robot Odometry
In this paper, we present a novel factor graph formulation to estimate the
pose and velocity of a quadruped robot on slippery and deformable terrain. The
factor graph introduces a preintegrated velocity factor that incorporates
velocity inputs from leg odometry and also estimates related biases. From our
experimentation we have seen that it is difficult to model uncertainties at the
contact point such as slip or deforming terrain, as well as leg flexibility. To
accommodate for these effects and to minimize leg odometry drift, we extend the
robot's state vector with a bias term for this preintegrated velocity factor.
The bias term can be accurately estimated thanks to the tight fusion of the
preintegrated velocity factor with stereo vision and IMU factors, without which
it would be unobservable. The system has been validated on several scenarios
that involve dynamic motions of the ANYmal robot on loose rocks, slopes and
muddy ground. We demonstrate a 26% improvement of relative pose error compared
to our previous work and 52% compared to a state-of-the-art proprioceptive
state estimator.Comment: Accepted to ICRA 2020. Video: youtu.be/w1Sx6dIqgQ
Haptic Sequential Monte Carlo Localization for Quadrupedal Locomotion in Vision-Denied Scenarios
Continuous robot operation in extreme scenarios such as underground mines or
sewers is difficult because exteroceptive sensors may fail due to fog,
darkness, dirt or malfunction. So as to enable autonomous navigation in these
kinds of situations, we have developed a type of proprioceptive localization
which exploits the foot contacts made by a quadruped robot to localize against
a prior map of an environment, without the help of any camera or LIDAR sensor.
The proposed method enables the robot to accurately re-localize itself after
making a sequence of contact events over a terrain feature. The method is based
on Sequential Monte Carlo and can support both 2.5D and 3D prior map
representations. We have tested the approach online and onboard the ANYmal
quadruped robot in two different scenarios: the traversal of a custom built
wooden terrain course and a wall probing and following task. In both scenarios,
the robot is able to effectively achieve a localization match and to execute a
desired pre-planned path. The method keeps the localization error down to 10cm
on feature rich terrain by only using its feet, kinematic and inertial sensing.Comment: 7 pages, 8 figures, 1 table. Accepted at IEEE/RSJ IROS 202
Actively Mapping Industrial Structures with Information Gain-Based Planning on a Quadruped Robot
In this paper, we develop an online active mapping system to enable a
quadruped robot to autonomously survey large physical structures. We describe
the perception, planning and control modules needed to scan and reconstruct an
object of interest, without requiring a prior model. The system builds a voxel
representation of the object, and iteratively determines the Next-Best-View
(NBV) to extend the representation, according to both the reconstruction itself
and to avoid collisions with the environment. By computing the expected
information gain of a set of candidate scan locations sampled on the as-sensed
terrain map, as well as the cost of reaching these candidates, the robot
decides the NBV for further exploration. The robot plans an optimal path
towards the NBV, avoiding obstacles and un-traversable terrain. Experimental
results on both simulated and real-world environments show the capability and
efficiency of our system. Finally we present a full system demonstration on the
real robot, the ANYbotics ANYmal, autonomously reconstructing a building facade
and an industrial structure
Efficient scene simulation for robust monte carlo localization using an RGB-D camera
This paper presents Kinect Monte Carlo Localization (KMCL), a new method for localization in three dimensional indoor environments using RGB-D cameras, such as the Microsoft Kinect. The approach makes use of a low fidelity a priori 3-D model of the area of operation composed of large planar segments, such as walls and ceilings, which are assumed to remain static. Using this map as input, the KMCL algorithm employs feature-based visual odometry as the particle propagation mechanism and utilizes the 3-D map and the underlying sensor image formation model to efficiently simulate RGB-D camera views at the location of particle poses, using a graphical processing unit (GPU). The generated 3D views of the scene are then used to evaluate the likelihood of the particle poses. This GPU implementation provides a factor of ten speedup over a pure distance-based method, yet provides comparable accuracy. Experimental results are presented for five different configurations, including: (1) a robotic wheelchair, (2) a sensor mounted on a person, (3) an Ascending Technologies quadrotor, (4) a Willow Garage PR2, and (5) an RWI B21 wheeled mobile robot platform. The results demonstrate that the system can perform robust localization with 3D information for motions as fast as 1.5 meters per second. The approach is designed to be applicable not just for robotics but other applications such as wearable computing
IMU-based Online Multi-lidar Calibration
Modern autonomous systems typically use several sensors for perception. For
best performance, accurate and reliable extrinsic calibration is necessary. In
this research, we propose a reliable technique for the extrinsic calibration of
several lidars on a vehicle without the need for odometry estimation or
fiducial markers. First, our method generates an initial guess of the
extrinsics by matching the raw signals of IMUs co-located with each lidar. This
initial guess is then used in ICP and point cloud feature matching which
refines and verifies this estimate. Furthermore, we can use observability
criteria to choose a subset of the IMU measurements that have the highest
mutual information -- rather than comparing all the readings. We have
successfully validated our methodology using data gathered from Scania test
vehicles.Comment: For associated video, see https://youtu.be/HJ0CBWTFOh
Language-EXtended Indoor SLAM (LEXIS): A Versatile System for Real-time Visual Scene Understanding
Versatile and adaptive semantic understanding would enable autonomous systems
to comprehend and interact with their surroundings. Existing fixed-class models
limit the adaptability of indoor mobile and assistive autonomous systems. In
this work, we introduce LEXIS, a real-time indoor Simultaneous Localization and
Mapping (SLAM) system that harnesses the open-vocabulary nature of Large
Language Models (LLMs) to create a unified approach to scene understanding and
place recognition. The approach first builds a topological SLAM graph of the
environment (using visual-inertial odometry) and embeds Contrastive
Language-Image Pretraining (CLIP) features in the graph nodes. We use this
representation for flexible room classification and segmentation, serving as a
basis for room-centric place recognition. This allows loop closure searches to
be directed towards semantically relevant places. Our proposed system is
evaluated using both public, simulated data and real-world data, covering
office and home environments. It successfully categorizes rooms with varying
layouts and dimensions and outperforms the state-of-the-art (SOTA). For place
recognition and trajectory estimation tasks we achieve equivalent performance
to the SOTA, all also utilizing the same pre-trained model. Lastly, we
demonstrate the system's potential for planning
- …