297 research outputs found
Direct Sparse Odometry with Rolling Shutter
Neglecting the effects of rolling-shutter cameras for visual odometry (VO)
severely degrades accuracy and robustness. In this paper, we propose a novel
direct monocular VO method that incorporates a rolling-shutter model. Our
approach extends direct sparse odometry which performs direct bundle adjustment
of a set of recent keyframe poses and the depths of a sparse set of image
points. We estimate the velocity at each keyframe and impose a
constant-velocity prior for the optimization. In this way, we obtain a near
real-time, accurate direct VO method. Our approach achieves improved results on
challenging rolling-shutter sequences over state-of-the-art global-shutter VO
Rolling-Shutter Modelling for Direct Visual-Inertial Odometry
We present a direct visual-inertial odometry (VIO) method which estimates the
motion of the sensor setup and sparse 3D geometry of the environment based on
measurements from a rolling-shutter camera and an inertial measurement unit
(IMU).
The visual part of the system performs a photometric bundle adjustment on a
sparse set of points. This direct approach does not extract feature points and
is able to track not only corners, but any pixels with sufficient gradient
magnitude. Neglecting rolling-shutter effects in the visual part severely
degrades accuracy and robustness of the system. In this paper, we incorporate a
rolling-shutter model into the photometric bundle adjustment that estimates a
set of recent keyframe poses and the inverse depth of a sparse set of points.
IMU information is accumulated between several frames using measurement
preintegration, and is inserted into the optimization as an additional
constraint between selected keyframes. For every keyframe we estimate not only
the pose but also velocity and biases to correct the IMU measurements. Unlike
systems with global-shutter cameras, we use both IMU measurements and
rolling-shutter effects of the camera to estimate velocity and biases for every
state.
Last, we evaluate our system on a novel dataset that contains global-shutter
and rolling-shutter images, IMU data and ground-truth poses for ten different
sequences, which we make publicly available. Evaluation shows that the proposed
method outperforms a system where rolling shutter is not modelled and achieves
similar accuracy to the global-shutter method on global-shutter data
Visual SLAM algorithms: a survey from 2010 to 2016
SLAM is an abbreviation for simultaneous localization and mapping, which is a technique for estimating sensor motion and reconstructing structure in an unknown environment. Especially, Simultaneous Localization and Mapping (SLAM) using cameras is referred to as visual SLAM (vSLAM) because it is based on visual information only. vSLAM can be used as a fundamental technology for various types of applications and has been discussed in the field of computer vision, augmented reality, and robotics in the literature. This paper aims to categorize and summarize recent vSLAM algorithms proposed in different research communities from both technical and historical points of views. Especially, we focus on vSLAM algorithms proposed mainly from 2010 to 2016 because major advance occurred in that period. The technical categories are summarized as follows: feature-based, direct, and RGB-D camera-based approaches
Revisiting Rolling Shutter Bundle Adjustment: Toward Accurate and Fast Solution
We propose a robust and fast bundle adjustment solution that estimates the
6-DoF pose of the camera and the geometry of the environment based on
measurements from a rolling shutter (RS) camera. This tackles the challenges in
the existing works, namely relying on additional sensors, high frame rate video
as input, restrictive assumptions on camera motion, readout direction, and poor
efficiency. To this end, we first investigate the influence of normalization to
the image point on RSBA performance and show its better approximation in
modelling the real 6-DoF camera motion. Then we present a novel analytical
model for the visual residual covariance, which can be used to standardize the
reprojection error during the optimization, consequently improving the overall
accuracy. More importantly, the combination of normalization and covariance
standardization weighting in RSBA (NW-RSBA) can avoid common planar degeneracy
without needing to constrain the filming manner. Besides, we propose an
acceleration strategy for NW-RSBA based on the sparsity of its Jacobian matrix
and Schur complement. The extensive synthetic and real data experiments verify
the effectiveness and efficiency of the proposed solution over the
state-of-the-art works. We also demonstrate the proposed method can be easily
implemented and plug-in famous GSSfM and GSSLAM systems as completed RSSfM and
RSSLAM solutions
Unsupervised Odometry and Depth Learning for Endoscopic Capsule Robots
In the last decade, many medical companies and research groups have tried to
convert passive capsule endoscopes as an emerging and minimally invasive
diagnostic technology into actively steerable endoscopic capsule robots which
will provide more intuitive disease detection, targeted drug delivery and
biopsy-like operations in the gastrointestinal(GI) tract. In this study, we
introduce a fully unsupervised, real-time odometry and depth learner for
monocular endoscopic capsule robots. We establish the supervision by warping
view sequences and assigning the re-projection minimization to the loss
function, which we adopt in multi-view pose estimation and single-view depth
estimation network. Detailed quantitative and qualitative analyses of the
proposed framework performed on non-rigidly deformable ex-vivo porcine stomach
datasets proves the effectiveness of the method in terms of motion estimation
and depth recovery.Comment: submitted to IROS 201
- …