11,511 research outputs found
Multiframe Scene Flow with Piecewise Rigid Motion
We introduce a novel multiframe scene flow approach that jointly optimizes
the consistency of the patch appearances and their local rigid motions from
RGB-D image sequences. In contrast to the competing methods, we take advantage
of an oversegmentation of the reference frame and robust optimization
techniques. We formulate scene flow recovery as a global non-linear least
squares problem which is iteratively solved by a damped Gauss-Newton approach.
As a result, we obtain a qualitatively new level of accuracy in RGB-D based
scene flow estimation which can potentially run in real-time. Our method can
handle challenging cases with rigid, piecewise rigid, articulated and moderate
non-rigid motion, and does not rely on prior knowledge about the types of
motions and deformations. Extensive experiments on synthetic and real data show
that our method outperforms state-of-the-art.Comment: International Conference on 3D Vision (3DV), Qingdao, China, October
201
Multiframe Scene Flow with Piecewise Rigid Motion
We introduce a novel multiframe scene flow approach that jointly optimizes
the consistency of the patch appearances and their local rigid motions from
RGB-D image sequences. In contrast to the competing methods, we take advantage
of an oversegmentation of the reference frame and robust optimization
techniques. We formulate scene flow recovery as a global non-linear least
squares problem which is iteratively solved by a damped Gauss-Newton approach.
As a result, we obtain a qualitatively new level of accuracy in RGB-D based
scene flow estimation which can potentially run in real-time. Our method can
handle challenging cases with rigid, piecewise rigid, articulated and moderate
non-rigid motion, and does not rely on prior knowledge about the types of
motions and deformations. Extensive experiments on synthetic and real data show
that our method outperforms state-of-the-art.Comment: International Conference on 3D Vision (3DV), Qingdao, China, October
201
SO(3)-invariant asymptotic observers for dense depth field estimation based on visual data and known camera motion
In this paper, we use known camera motion associated to a video sequence of a
static scene in order to estimate and incrementally refine the surrounding
depth field. We exploit the SO(3)-invariance of brightness and depth fields
dynamics to customize standard image processing techniques. Inspired by the
Horn-Schunck method, we propose a SO(3)-invariant cost to estimate the depth
field. At each time step, this provides a diffusion equation on the unit
Riemannian sphere that is numerically solved to obtain a real time depth field
estimation of the entire field of view. Two asymptotic observers are derived
from the governing equations of dynamics, respectively based on optical flow
and depth estimations: implemented on noisy sequences of synthetic images as
well as on real data, they perform a more robust and accurate depth estimation.
This approach is complementary to most methods employing state observers for
range estimation, which uniquely concern single or isolated feature points.Comment: Submitte
Enhancing retinal images by nonlinear registration
Being able to image the human retina in high resolution opens a new era in
many important fields, such as pharmacological research for retinal diseases,
researches in human cognition, nervous system, metabolism and blood stream, to
name a few. In this paper, we propose to share the knowledge acquired in the
fields of optics and imaging in solar astrophysics in order to improve the
retinal imaging at very high spatial resolution in the perspective to perform a
medical diagnosis. The main purpose would be to assist health care
practitioners by enhancing retinal images and detect abnormal features. We
apply a nonlinear registration method using local correlation tracking to
increase the field of view and follow structure evolutions using correlation
techniques borrowed from solar astronomy technique expertise. Another purpose
is to define the tracer of movements after analyzing local correlations to
follow the proper motions of an image from one moment to another, such as
changes in optical flows that would be of high interest in a medical diagnosis.Comment: 21 pages, 7 figures, submitted to Optics Communication
Data Fusion of Objects Using Techniques Such as Laser Scanning, Structured Light and Photogrammetry for Cultural Heritage Applications
In this paper we present a semi-automatic 2D-3D local registration pipeline
capable of coloring 3D models obtained from 3D scanners by using uncalibrated
images. The proposed pipeline exploits the Structure from Motion (SfM)
technique in order to reconstruct a sparse representation of the 3D object and
obtain the camera parameters from image feature matches. We then coarsely
register the reconstructed 3D model to the scanned one through the Scale
Iterative Closest Point (SICP) algorithm. SICP provides the global scale,
rotation and translation parameters, using minimal manual user intervention. In
the final processing stage, a local registration refinement algorithm optimizes
the color projection of the aligned photos on the 3D object removing the
blurring/ghosting artefacts introduced due to small inaccuracies during the
registration. The proposed pipeline is capable of handling real world cases
with a range of characteristics from objects with low level geometric features
to complex ones
Inertial-sensor bias estimation from brightness/depth images and based on SO(3)-invariant integro/partial-differential equations on the unit sphere
Constant biases associated to measured linear and angular velocities of a
moving object can be estimated from measurements of a static scene by embedded
brightness and depth sensors. We propose here a Lyapunov-based observer taking
advantage of the SO(3)-invariance of the partial differential equations
satisfied by the measured brightness and depth fields. The resulting asymptotic
observer is governed by a non-linear integro/partial differential system where
the two independent scalar variables indexing the pixels live on the unit
sphere of the 3D Euclidian space. The observer design and analysis are strongly
simplified by coordinate-free differential calculus on the unit sphere equipped
with its natural Riemannian structure. The observer convergence is investigated
under C^1 regularity assumptions on the object motion and its scene. It relies
on Ascoli-Arzela theorem and pre-compactness of the observer trajectories. It
is proved that the estimated biases converge towards the true ones, if and only
if, the scene admits no cylindrical symmetry. The observer design can be
adapted to realistic sensors where brightness and depth data are only available
on a subset of the unit sphere. Preliminary simulations with synthetic
brightness and depth images (corrupted by noise around 10%) indicate that such
Lyapunov-based observers should be robust and convergent for much weaker
regularity assumptions.Comment: 30 pages, 6 figures, submitte
- …