301 research outputs found
Review and classification of vision-based localisation techniques in unknown environments
International audienceThis study presents a review of the state-of-the-art and a novel classification of current vision-based localisation techniques in unknown environments. Indeed, because of progresses made in computer vision, it is now possible to consider vision-based systems as promising navigation means that can complement traditional navigation sensors like global navigation satellite systems (GNSSs) and inertial navigation systems. This study aims to review techniques employing a camera as a localisation sensor, provide a classification of techniques and introduce schemes that exploit the use of video information within a multi-sensor system. In fact, a general model is needed to better compare existing techniques in order to decide which approach is appropriate and which are the innovation axes. In addition, existing classifications only consider techniques based on vision as a standalone tool and do not consider video as a sensor among others. The focus is addressed to scenarios where no a priori knowledge of the environment is provided. In fact, these scenarios are the most challenging since the system has to cope with objects as they appear in the scene without any prior information about their expected position
A two-dimensional pedestrian navigation solution aided with a visual gyroscope and a visual odometer
Advanced Integration of GNSS and External Sensors for Autonomous Mobility Applications
L'abstract è presente nell'allegato / the abstract is in the attachmen
Vision-Aided Pedestrian Navigation for Challenging GNSS Environments
There is a strong need for an accurate pedestrian navigation system, functional also in GNSS challenging environments, namely urban areas and indoors, for improved safety and to enhance everyday life. Pedestrian navigation is mainly needed in these environments that are challenging for GNSS but also for other RF positioning systems and some non-RF systems such as the magnetometry used for heading due to the presence of ferrous material. Indoor and urban navigation has been an active research area for years. There is no individual system at this time that can address all needs set for pedestrian navigation in these environments, but a fused solution of different sensors can provide better accuracy, availability and continuity. Self-contained sensors, namely digital compasses for measuring heading, gyroscopes for heading changes and accelerometers for the user speed, constitute a good option for pedestrian navigation. However, their performance suffers from noise and biases that result in large position errors increasing with time. Such errors can however be mitigated using information about the user motion obtained from consecutive images taken by a camera carried by the user, provided that its position and orientation with respect to the user’s body are known. The motion of the features in the images may then be transformed into information about the user’s motion. Due to its distinctive characteristics, this vision-aiding complements other positioning technologies in order to provide better pedestrian navigation accuracy and reliability. This thesis discusses the concepts of a visual gyroscope that provides the relative user heading and a visual odometer that provides the translation of the user between the consecutive images. Both methods use a monocular camera carried by the user. The visual gyroscope monitors the motion of virtual features, called vanishing points, arising from parallel straight lines in the scene, and from the change of their location that resolves heading, roll and pitch. The method is applicable to the human environments as the straight lines in the structures enable the vanishing point perception. For the visual odometer, the ambiguous scale arising when using the homography between consecutive images to observe the translation is solved using two different methods. First, the scale is computed using a special configuration intended for indoors. Secondly, the scale is resolved using differenced GNSS carrier phase measurements of the camera in a method aimed at urban environments, where GNSS can’t perform alone due to tall buildings blocking the required line-of-sight to four satellites. However, the use of visual perception provides position information by exploiting a minimum of two satellites and therefore the availability of navigation solution is substantially increased. Both methods are sufficiently tolerant for the challenges of visual perception in indoor and urban environments, namely low lighting and dynamic objects hindering the view. The heading and translation are further integrated with other positioning systems and a navigation solution is obtained. The performance of the proposed vision-aided navigation was tested in various environments, indoors and urban canyon environments to demonstrate its effectiveness. These experiments, although of limited durations, show that visual processing efficiently complements other positioning technologies in order to provide better pedestrian navigation accuracy and reliability
GPS-aided Visual Wheel Odometry
This paper introduces a novel GPS-aided visual-wheel odometry (GPS-VWO) for
ground robots. The state estimation algorithm tightly fuses visual, wheeled
encoder and GPS measurements in the way of Multi-State Constraint Kalman Filter
(MSCKF). To avoid accumulating calibration errors over time, the proposed
algorithm calculates the extrinsic rotation parameter between the GPS global
coordinate frame and the VWO reference frame online as part of the estimation
process. The convergence of this extrinsic parameter is guaranteed by the
observability analysis and verified by using real-world visual and wheel
encoder measurements as well as simulated GPS measurements. Moreover, a novel
theoretical finding is presented that the variance of unobservable state could
converge to zero for specific Kalman filter system. We evaluate the proposed
system extensively in large-scale urban driving scenarios. The results
demonstrate that better accuracy than GPS is achieved through the fusion of GPS
and VWO. The comparison between extrinsic parameter calibration and
non-calibration shows significant improvement in localization accuracy thanks
to the online calibration.Comment: Accepted by IEEE ITSC 202
Low cost inertial-based localization system for a service robot
Dissertation presented at Faculty of Sciences and Technology of the New University of Lisbon
to attain the Master degree in Electrical and Computer Science EngineeringThe knowledge of a robot’s location it’s fundamental for most part of service robots. The success of tasks such as mapping and planning depend on a good robot’s position knowledge.
The main goal of this dissertation is to present a solution that provides a estimation of the robot’s location. This is, a tracking system that can run either inside buildings or outside them, not taking into account just structured environments. Therefore, the localization system takes into
account only measurements relative.
In the presented solution is used an AHRS device and digital encoders placed on wheels to make a estimation of robot’s position. It also relies on the use of Kalman Filter to integrate sensorial information and deal with estimate errors.
The developed system was testes in real environments through its integration on real robot. The results revealed that is not possible to attain a good position estimation using only low-cost inertial
sensors. Thus, is required the integration of more sensorial information, through absolute or relative measurements technologies, to provide a more accurate position estimation
A LiDAR-Inertial SLAM Tightly-Coupled with Dropout-Tolerant GNSS Fusion for Autonomous Mine Service Vehicles
Multi-modal sensor integration has become a crucial prerequisite for the
real-world navigation systems. Recent studies have reported successful
deployment of such system in many fields. However, it is still challenging for
navigation tasks in mine scenes due to satellite signal dropouts, degraded
perception, and observation degeneracy. To solve this problem, we propose a
LiDAR-inertial odometry method in this paper, utilizing both Kalman filter and
graph optimization. The front-end consists of multiple parallel running
LiDAR-inertial odometries, where the laser points, IMU, and wheel odometer
information are tightly fused in an error-state Kalman filter. Instead of the
commonly used feature points, we employ surface elements for registration. The
back-end construct a pose graph and jointly optimize the pose estimation
results from inertial, LiDAR odometry, and global navigation satellite system
(GNSS). Since the vehicle has a long operation time inside the tunnel, the
largely accumulated drift may be not fully by the GNSS measurements. We hereby
leverage a loop closure based re-initialization process to achieve full
alignment. In addition, the system robustness is improved through handling data
loss, stream consistency, and estimation error. The experimental results show
that our system has a good tolerance to the long-period degeneracy with the
cooperation different LiDARs and surfel registration, achieving meter-level
accuracy even for tens of minutes running during GNSS dropouts
- …