52 research outputs found

    Camera motion estimation through planar deformation determination

    Get PDF
    In this paper, we propose a global method for estimating the motion of a camera which films a static scene. Our approach is direct, fast and robust, and deals with adjacent frames of a sequence. It is based on a quadratic approximation of the deformation between two images, in the case of a scene with constant depth in the camera coordinate system. This condition is very restrictive but we show that provided translation and depth inverse variations are small enough, the error on optical flow involved by the approximation of depths by a constant is small. In this context, we propose a new model of camera motion, that allows to separate the image deformation in a similarity and a ``purely'' projective application, due to change of optical axis direction. This model leads to a quadratic approximation of image deformation that we estimate with an M-estimator; we can immediatly deduce camera motion parameters.Comment: 21 pages, version modifi\'ee accept\'e le 20 mars 200

    Exploring Vision-Based Interfaces: How to Use Your Head in Dual Pointing Tasks

    Get PDF
    The utility of vision-based face tracking for dual pointing tasks is evaluated. We first describe a 3-D face tracking technique based on real-time parametric motion-stereo, which is non-invasive, robust, and self-initialized. The tracker provides a real-time estimate of a ?frontal face ray? whose intersection with the display surface plane is used as a second stream of input for scrolling or pointing, in paral-lel with hand input. We evaluated the performance of com-bined head/hand input on a box selection and coloring task: users selected boxes with one pointer and colors with a second pointer, or performed both tasks with a single pointer. We found that performance with head and one hand was intermediate between single hand performance and dual hand performance. Our results are consistent with previously reported dual hand conflict in symmetric pointing tasks, and suggest that a head-based input stream should be used for asymmetric control

    Direct Estimation of Motion and Extended Scene Structure from a Moving Stereo Rig

    Get PDF
    We describe a new method for motion estimation and 3D reconstruction from stereo image sequences obtained by a stereo rig moving through a rigid world. We show that given two stereo pairs one can compute the motion of the stereo rig directly from the image derivatives (spatial and temporal). Correspondences are not required. One can then use the images from both pairs combined to compute a dense depth map. The motion estimates between stereo pairs enable us to combine depth maps from all the pairs in the sequence to form an extended scene reconstruction and we show results from a real image sequence. The motion computation is a linear least squares computation using all the pixels in the image. Areas with little or no contrast are implicitly weighted less so one does not have to explicitly apply a confidence measure

    A Variational Framework for Structure from Motion inOmnidirectional Image Sequences

    Get PDF
    We address the problem of depth and ego-motion estimation from omnidirectional images. We propose a correspondence-free structure-from-motion problem for sequences of images mapped on the 2-sphere. A novel graph-based variational framework is first proposed for depth estimation between pairs of images. The estimation is cast as a TV-L1 optimization problem that is solved by a fast graph-based algorithm. The ego-motion is then estimated directly from the depth information without explicit computation of the optical flow. Both problems are finally addressed together in an iterative algorithm that alternates between depth and ego-motion estimation for fast computation of 3D information from motion in image sequences. Experimental results demonstrate the effective performance of the proposed algorithm for 3D reconstruction from synthetic and natural omnidirectional image

    A Kalman filter approach to direct depth estimation incorporating surface structure

    Get PDF
    The problem of depth-from-motion using a monocular image sequence is considered. A pixel-based model is developed for direct depth estimation within a Kaiman filtering framework. A method is proposed for incorporating local surface structure into the Kaiman filter. Experimental results are provided to illustrate the effect of structural information on depth estimation. ©1999 IEEE.published_or_final_versio

    An ECC Based Iterative Algorithm For Photometric Invariant Projective Registration

    Get PDF
    International audienceThe ability of an algorithm to accurately estimate the parameters of the geometric trans- formation which aligns two image profiles even in the presence of photometric distortions can be considered as a basic requirement in many computer vision applications. Projec- tive transformations constitute a general class which includes as special cases the affine, as well as the metric subclasses of transformations. In this paper the applicability of a recently proposed iterative algorithm, which uses the Enhanced Correlation Coefficient as a performance criterion, in the projective image registration problem is investigated. The main theoretical results concerning the proposed iterative algorithm are presented. Furthermore, the performance of the iterative algorithm in the presence of nonlinear photometric distortions is compared against the leading Lucas-Kanade algorithm and its simultaneous inverse compositional variant with the help of a series of experiments involving strong or weak geometric deformations, ideal and noisy conditions and even over-modelling of the warping process. Although under ideal conditions the proposed al- gorithm and simultaneous inverse compositional algorithm exhibit a similar performance and both outperform the Lucas-Kanade algorithm, under noisy conditions the proposed algorithm outperforms the other algorithms in convergence speed and accuracy, and exhibits robustness against photometric distortions

    Correspondenceless Structure from Motion

    Get PDF
    We present a novel approach for the estimation of 3D-motion directly from two images using the Radon transform. The feasibility of any camera motion is computed by integrating over all feature pairs that satisfy the epipolar constraint. This integration is equivalent to taking the inner product of a similarity function on feature pairs with a Dirac function embedding the epipolar constraint. The maxima in this five dimensional motion space will correspond to compatible rigid motions. The main novelty is in the realization that the Radon transform is a filtering operator: If we assume that the similarity and Dirac functions are defined on spheres and the epipolar constraint is a group action of rotations on spheres, then the Radon transform is a correlation integral. We propose a new algorithm to compute this integral from the spherical Fourier transform of the similarity and Dirac functions. Generating the similarity function now becomes a preprocessing step which reduces the complexity of the Radon computation by a factor equal to the number of feature pairs processed. The strength of the algorithm is in avoiding a commitment to correspondences, thus being robust to erroneous feature detection, outliers, and multiple motions

    Correspondence and Affine Shape from Two Orthographic Views: Motion and Recognition

    Get PDF
    The paper presents a simple model for recovering affine shape and correspondence from two orthographic views of a 3D object. It is shown that four corresponding points along two orthographic views, taken under similar illumination conditions, determine affine shape and correspondence for all other points. The scheme is useful for purposes of visual recognition by generating novel views of an object given two model views. It is also shown that the scheme can handle objects with smooth boundaries, to a good approximation, without introducing any modifications or additional model views

    "Determining optical flow": a retrospective

    Full text link
    Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/30981/1/0000654.pd

    Direct recovering of surface structure characterized by an Nth degree polynomial equation using the UOFF approach

    Get PDF
    There are two different approaches for estimation of structure and/or motion of objects in the computer vision community today. One is the feature correspondence method, and the other is the optical flow method [1]. There are many difficulties and limitations encountered with the feature correspondence method, while the optical flow method is more feasible, but requires a substantial amount of extra calculations if the optical flow is to be computed as an intermediate step. Direct methods have been developed [2-4], that use the optical flow approach, but avoid computing the full optical flow field as an intermediate step for recovering structure and motion. The unified optical flow field theory was recently established in [5]. It is an extension of the optical flow (UOFF) [1] to stereo imagery. Based on the UOFF, a direct method is developed to reconstruct an Alpha shape surface structure characterized by an third degree polynomial equation, and a Sphere surface characterized by a second degree polynomial [6]. This thesis work uses the methods developed in [5,6], to reconstruct the third degree polynomial describing a surface. The main difference from the simulation results obtained in [6], is that in this case, one of the two surfaces tested is a third order, unbounded surface, and that tbe image gradients are computed directly from the image data, with no prior knowledge of the surface gray function distribution. Another important difference is that the gray levels of the surface are quantized in this work; i.e., the computations are done using integer image data, not the continuous gray levels as in [6]. These differences contribute to proving that the UOFF technique can be used in a practical manner, and with good results. Further discussions of the contributions of this work are included in the last chapter
    • …
    corecore