181 research outputs found
Large Scale SfM with the Distributed Camera Model
We introduce the distributed camera model, a novel model for
Structure-from-Motion (SfM). This model describes image observations in terms
of light rays with ray origins and directions rather than pixels. As such, the
proposed model is capable of describing a single camera or multiple cameras
simultaneously as the collection of all light rays observed. We show how the
distributed camera model is a generalization of the standard camera model and
describe a general formulation and solution to the absolute camera pose problem
that works for standard or distributed cameras. The proposed method computes a
solution that is up to 8 times more efficient and robust to rotation
singularities in comparison with gDLS. Finally, this method is used in an novel
large-scale incremental SfM pipeline where distributed cameras are accurately
and robustly merged together. This pipeline is a direct generalization of
traditional incremental SfM; however, instead of incrementally adding one
camera at a time to grow the reconstruction the reconstruction is grown by
adding a distributed camera. Our pipeline produces highly accurate
reconstructions efficiently by avoiding the need for many bundle adjustment
iterations and is capable of computing a 3D model of Rome from over 15,000
images in just 22 minutes.Comment: Published at 2016 3DV Conferenc
Cluster-Wise Ratio Tests for Fast Camera Localization
Feature point matching for camera localization suffers from scalability
problems. Even when feature descriptors associated with 3D scene points are
locally unique, as coverage grows, similar or repeated features become
increasingly common. As a result, the standard distance ratio-test used to
identify reliable image feature points is overly restrictive and rejects many
good candidate matches. We propose a simple coarse-to-fine strategy that uses
conservative approximations to robust local ratio-tests that can be computed
efficiently using global approximate k-nearest neighbor search. We treat these
forward matches as votes in camera pose space and use them to prioritize
back-matching within candidate camera pose clusters, exploiting feature
co-visibility captured by clustering the 3D model camera pose graph. This
approach achieves state-of-the-art camera localization results on a variety of
popular benchmarks, outperforming several methods that use more complicated
data structures and that make more restrictive assumptions on camera pose. We
also carry out diagnostic analyses on a difficult test dataset containing
globally repetitive structure that suggest our approach successfully adapts to
the challenges of large-scale image localization
P1AC: Revisiting Absolute Pose From a Single Affine Correspondence
We introduce a novel solution to the problem of estimating the pose of a
calibrated camera given a single observation of an oriented point and an affine
correspondence to a reference image. Affine correspondences have traditionally
been used to improve feature matching over wide baselines; however, little
previous work has considered the use of such correspondences for absolute
camera pose computation. The advantage of our approach (P1AC) is that it
requires only a single correspondence in the minimal case in comparison to the
traditional point-based approach (P3P) which requires at least three points.
Our method removes the limiting assumptions made in previous work and provides
a general solution that is applicable to large-scale image-based localization.
Our evaluation on synthetic data shows that our approach is numerically stable
and more robust to point observation noise than P3P. We also evaluate the
application of our approach for large-scale image-based localization and
demonstrate a practical reduction in the number of iterations and computation
time required to robustly localize an image
Efficient solutions to the relative pose of three calibrated cameras from four points using virtual correspondences
We study the challenging problem of estimating the relative pose of three
calibrated cameras. We propose two novel solutions to the notoriously difficult
configuration of four points in three views, known as the 4p3v problem. Our
solutions are based on the simple idea of generating one additional virtual
point correspondence in two views by using the information from the locations
of the four input correspondences in the three views. For the first solver, we
train a network to predict this point correspondence. The second solver uses a
much simpler and more efficient strategy based on the mean points of three
corresponding input points. The new solvers are efficient and easy to implement
since they are based on the existing efficient minimal solvers, i.e., the
well-known 5-point relative pose and the P3P solvers. The solvers achieve
state-of-the-art results on real data. The idea of solving minimal problems
using virtual correspondences is general and can be applied to other problems,
e.g., the 5-point relative pose problem. In this way, minimal problems can be
solved using simpler non-minimal solvers or even using sub-minimal samples
inside RANSAC.
In addition, we compare different variants of 4p3v solvers with the baseline
solver for the minimal configuration consisting of three triplets of points and
two points visible in two views. We discuss which configuration of points is
potentially the most practical in real applications
Calibrated and Partially Calibrated Semi-Generalized Homographies
In this paper, we propose the first minimal solutions for estimating the
semi-generalized homography given a perspective and a generalized camera. The
proposed solvers use five 2D-2D image point correspondences induced by a scene
plane. One of them assumes the perspective camera to be fully calibrated, while
the other solver estimates the unknown focal length together with the absolute
pose parameters. This setup is particularly important in structure-from-motion
and image-based localization pipelines, where a new camera is localized in each
step with respect to a set of known cameras and 2D-3D correspondences might not
be available. As a consequence of a clever parametrization and the elimination
ideal method, our approach only needs to solve a univariate polynomial of
degree five or three. The proposed solvers are stable and efficient as
demonstrated by a number of synthetic and real-world experiments
Accurate and linear time pose estimation from points and lines
The final publication is available at link.springer.comThe Perspective-n-Point (PnP) problem seeks to estimate the pose of a calibrated camera from n 3Dto-2D point correspondences. There are situations, though, where PnP solutions are prone to fail because feature point correspondences cannot be reliably estimated (e.g. scenes with repetitive patterns or with low texture). In such
scenarios, one can still exploit alternative geometric entities, such as lines, yielding the so-called Perspective-n-Line (PnL) algorithms. Unfortunately, existing PnL solutions are not as accurate and efficient as their point-based
counterparts. In this paper we propose a novel approach to introduce 3D-to-2D line correspondences into a PnP formulation, allowing to simultaneously process points and lines. For this purpose we introduce an algebraic line error
that can be formulated as linear constraints on the line endpoints, even when these are not directly observable. These constraints can then be naturally integrated within the linear formulations of two state-of-the-art point-based algorithms,
the OPnP and the EPnP, allowing them to indistinctly handle points, lines, or a combination of them. Exhaustive experiments show that the proposed formulation brings remarkable boost in performance compared to only point or
only line based solutions, with a negligible computational overhead compared to the original OPnP and EPnP.Peer ReviewedPostprint (author's final draft
- âŠ