401 research outputs found
A -adic RanSaC algorithm for stereo vision using Hensel lifting
A -adic variation of the Ran(dom) Sa(mple) C(onsensus) method for solving
the relative pose problem in stereo vision is developped. From two 2-adically
encoded images a random sample of five pairs of corresponding points is taken,
and the equations for the essential matrix are solved by lifting solutions
modulo 2 to the 2-adic integers. A recently devised -adic hierarchical
classification algorithm imitating the known LBG quantisation method classifies
the solutions for all the samples after having determined the number of
clusters using the known intra-inter validity of clusterings. In the successful
case, a cluster ranking will determine the cluster containing a 2-adic
approximation to the "true" solution of the problem.Comment: 15 pages; typos removed, abstract changed, computation error remove
RPNet: an End-to-End Network for Relative Camera Pose Estimation
This paper addresses the task of relative camera pose estimation from raw
image pixels, by means of deep neural networks. The proposed RPNet network
takes pairs of images as input and directly infers the relative poses, without
the need of camera intrinsic/extrinsic. While state-of-the-art systems based on
SIFT + RANSAC, are able to recover the translation vector only up to scale,
RPNet is trained to produce the full translation vector, in an end-to-end way.
Experimental results on the Cambridge Landmark dataset show very promising
results regarding the recovery of the full translation vector. They also show
that RPNet produces more accurate and more stable results than traditional
approaches, especially for hard images (repetitive textures, textureless
images, etc). To the best of our knowledge, RPNet is the first attempt to
recover full translation vectors in relative pose estimation
Investigation on the automatic geo-referencing of archaeological UAV photographs by correlation with pre-existing ortho-photos
We present a method for the automatic geo-referencing of archaeological photographs captured aboard unmanned aerial vehicles (UAVs), termed UPs. We do so by help of pre-existing ortho-photo maps (OPMs) and digital surface models (DSMs). Typically, these pre-existing data sets are based on data that were captured at a widely different point in time. This renders the detection (and hence the matching) of homologous feature points in the UPs and OPMs infeasible mainly due to temporal variations of vegetation and illumination. Facing this difficulty, we opt for the normalized cross correlation coefficient of perspectively transformed image patches as the measure of image similarity. Applying a threshold to this measure, we detect candidates for homologous image points, resulting in a distinctive, but computationally intensive method. In order to lower computation times, we reduce the dimensionality and extents of the search space by making use of a priori knowledge of the data sets. By assigning terrain heights interpolated in the DSM to the image points found in the OPM, we generate control points. We introduce respective observations into a bundle block, from which gross errors i.e. false matches are eliminated during its robust adjustment. A test of our approach on a UAV image data set demonstrates its potential and raises hope to successfully process large image archives
DeMoN: Depth and Motion Network for Learning Monocular Stereo
In this paper we formulate structure from motion as a learning problem. We
train a convolutional network end-to-end to compute depth and camera motion
from successive, unconstrained image pairs. The architecture is composed of
multiple stacked encoder-decoder networks, the core part being an iterative
network that is able to improve its own predictions. The network estimates not
only depth and motion, but additionally surface normals, optical flow between
the images and confidence of the matching. A crucial component of the approach
is a training loss based on spatial relative differences. Compared to
traditional two-frame structure from motion methods, results are more accurate
and more robust. In contrast to the popular depth-from-single-image networks,
DeMoN learns the concept of matching and, thus, better generalizes to
structures not seen during training.Comment: Camera ready version for CVPR 2017. Supplementary material included.
Project page:
http://lmb.informatik.uni-freiburg.de/people/ummenhof/depthmotionnet
Self-Calibration of Cameras with Euclidean Image Plane in Case of Two Views and Known Relative Rotation Angle
The internal calibration of a pinhole camera is given by five parameters that
are combined into an upper-triangular calibration matrix. If the
skew parameter is zero and the aspect ratio is equal to one, then the camera is
said to have Euclidean image plane. In this paper, we propose a non-iterative
self-calibration algorithm for a camera with Euclidean image plane in case the
remaining three internal parameters --- the focal length and the principal
point coordinates --- are fixed but unknown. The algorithm requires a set of point correspondences in two views and also the measured relative
rotation angle between the views. We show that the problem generically has six
solutions (including complex ones).
The algorithm has been implemented and tested both on synthetic data and on
publicly available real dataset. The experiments demonstrate that the method is
correct, numerically stable and robust.Comment: 13 pages, 7 eps-figure
- …