490 research outputs found
Egomotion from event-based SNN optical flow
We present a method for computing egomotion using event cameras with a pre-trained optical flow spiking neural network (SNN). To address the aperture problem encountered in the sparse and noisy normal flow of the initial SNN layers, our method includes a sliding-window bin-based pooling layer that computes a fused full flow estimate. To add robustness to noisy flow estimates, instead of computing the egomotion from vector averages, our method optimizes the intersection of constraints. The method also includes a RANSAC step to robustly deal with outlier flow estimates in the pooling layer. We validate our approach on both simulated and real scenes and compare our results favorably to the state-of-the-art methods. However, our method may be sensitive to datasets and motion speeds different from those used for training, limiting its generalizability.This work received support from projects EBCON (PID2020-119244GBI00) and AUDEL (TED2021-131759A-I00) funded by MCIN/ AEI/
10.13039/ 501100011033 and by the "European Union NextGenerationEU/PRTR"; the Consolidated Research Group RAIG (2021 SGR
00510) of the Departament de Recerca i Universitats de la Generalitat de Catalunya; and by an FI AGAUR PhD grant to Yi Tian.Peer ReviewedPostprint (author's final draft
Multimotion Visual Odometry (MVO)
Visual motion estimation is a well-studied challenge in autonomous
navigation. Recent work has focused on addressing multimotion estimation in
highly dynamic environments. These environments not only comprise multiple,
complex motions but also tend to exhibit significant occlusion.
Estimating third-party motions simultaneously with the sensor egomotion is
difficult because an object's observed motion consists of both its true motion
and the sensor motion. Most previous works in multimotion estimation simplify
this problem by relying on appearance-based object detection or
application-specific motion constraints. These approaches are effective in
specific applications and environments but do not generalize well to the full
multimotion estimation problem (MEP).
This paper presents Multimotion Visual Odometry (MVO), a multimotion
estimation pipeline that estimates the full SE(3) trajectory of every motion in
the scene, including the sensor egomotion, without relying on appearance-based
information. MVO extends the traditional visual odometry (VO) pipeline with
multimotion segmentation and tracking techniques. It uses physically founded
motion priors to extrapolate motions through temporary occlusions and identify
the reappearance of motions through motion closure. Evaluations on real-world
data from the Oxford Multimotion Dataset (OMD) and the KITTI Vision Benchmark
Suite demonstrate that MVO achieves good estimation accuracy compared to
similar approaches and is applicable to a variety of multimotion estimation
challenges.Comment: Under review for the International Journal of Robotics Research
(IJRR), Manuscript #IJR-21-4311. 25 pages, 14 figures, 11 tables. Videos
available at https://www.youtube.com/watch?v=mNj3s1nf-6A and
https://www.youtube.com/playlist?list=PLbaQBz4TuPcxMIXKh5Q80s0N9ISezFcp
Joint on-manifold self-calibration of odometry model and sensor extrinsics using pre-integration
© 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.This paper describes a self-calibration procedure that jointly estimates the extrinsic parameters of an exteroceptive sensor able to observe ego-motion, and the intrinsic parameters of an odometry motion model, consisting of wheel radii and wheel separation. We use iterative nonlinear onmanifold optimization with a graphical representation of the state, and resort to an adaptation of the pre-integration theory, initially developed for the IMU motion sensor, to be applied to the differential drive motion model. For this, we describe the construction of a pre-integrated factor for the differential drive motion model, which includes the motion increment, its covariance, and a first-order approximation of its dependence with the calibration parameters. As the calibration parameters change at each solver iteration, this allows a posteriori factor correction without the need of re-integrating the motion data. We validate our proposal in simulations and on a real robot and show the convergence of the calibration towards the true values of the parameters. It is then tested online in simulation and is shown to accommodate to variations in the calibration parameters when the vehicle is subject to physical changes such as loading and unloading a freight.Peer ReviewedPostprint (author's final draft
DeepSpatial: Intelligent Spatial Sensor to Perception of Things
This paper discusses a spatial sensor to identify and track objects in the environment. The sensor is composed of an RGB-D camera that provides point cloud and RGB images and an egomotion sensor able to identify its displacement in the environment. The proposed sensor also incorporates a data processing strategy developed by the authors to conferring to the sensor different skills. The adopted approach is based on four analysis steps: egomotive, lexical, syntax, and prediction analysis. As a result, the proposed sensor can identify objects in the environment, track these objects, calculate their direction, speed, and acceleration, and also predict their future positions. The on-line detector YOLO is used as a tool to identify objects, and its output is combined with the point cloud information to obtain the spatial location of each identified object. The sensor can operate with higher precision and a lower update rate, using YOLOv2, or with a higher update rate, and a smaller accuracy using YOLOv3-tiny. The object tracking, egomotion, and collision prediction skills are tested and validated using a mobile robot having a precise speed control. The presented results show that the proposed sensor (hardware + software) achieves a satisfactory accuracy and usage rate, powering its use to mobile robotic. This paper's contribution is developing an algorithm for identifying, tracking, and predicting the future position of objects embedded in a compact hardware. Thus, the contribution of this paper is to convert raw data from traditional sensors into useful information.info:eu-repo/semantics/publishedVersio
- …