11 research outputs found

    Recovering Heading for Visually-Guided Navigation

    Get PDF
    We present a model for recovering the direction of heading of an observer who is moving relative to a scene that may contain self-moving objects. The model builds upon an algorithm proposed by Rieger and Lawton (1985), which is based on earlier work by Longuet-Higgens and Prazdny (1981). The algorithm uses velocity differences computed in regions of high depth variation to estimate the location of the focus of expansion, which indicates the observer's heading direction. We relate the behavior of the proposed model to psychophysical observations regarding the ability of human observers to judge their heading direction, and show how the model can cope with self-moving objects in the environment. We also discuss this model in the broader context of a navigational system that performs tasks requiring rapid sensing and response through the interaction of simple task-specific routines

    3D object reconstruction using stereo and motion

    Get PDF
    The extraction of reliable range data from images is investigated, considering, as a possible solution, the integration of different sensor modalities. Two different algorithms are used to obtain independent estimates of depth from a sequence of stereo images. The results are integrated on the basis of the uncertainty of each measure. The stereo algorithm uses a coarse-to-fine control strategy to compute disparity. An algorithm for depth-from-motion is used, exploiting the constraint imposed by active motion of the cameras. To obtain a 3D description of the objects, the motion of the cameras is purposefully controlled, in such a manner as to move around the objects in view while the gaze is directed toward a fixed point in space. This egomotion strategy, which is similar to that adopted by the human visuomotor system, allows a better exploration of partially occluded objects and simplifies the motion equations. When tested on real scenes, the algorithm demonstrated a low sensitivity to image noise, mainly due to the integration of independent measures. An experiment performed on a real scene containing several objects is presented

    Models for Motion Perception

    Get PDF
    As observers move through the environment or shift their direction of gaze, the world moves past them. In addition, there may be objects that are moving differently from the static background, either rigid-body motions or nonrigid (e.g., turbulent) ones. This dissertation discusses several models for motion perception. The models rely on first measuring motion energy, a multi-resolution representation of motion information extracted from image sequences. The image flow model combines the outputs of a set of spatiotemporal motion-energy filters to estimate image velocity, consonant with current views regarding the neurophysiology and psychophysics of motion perception. A parallel implementation computes a distributed representation of image velocity that encodes both a velocity estimate and the uncertainty in that estimate. In addition, a numerical measure of image-flow uncertainty is derived. The egomotion model poses the detection of moving objects and the recovery of depth from motion as sensor fusion problems that necessitate combining information from different sensors in the presence of noise and uncertainty. Image sequences are segmented by finding image regions corresponding to entire objects that are moving differently from the stationary background. The turbulent flow model utilizes a fractal-based model of turbulence, and estimates the fractal scaling parameter of fractal image sequences from the outputs of motion-energy filters. Some preliminary results demonstrate the model\u27s potential for discriminating image regions based on fractal scaling

    The Analysis Of Visual Motion: From Computational Theory To Neuronal Mechanisms

    Get PDF

    Objective measurement of motion in the orbit

    Get PDF
    The research described in the thesis had two major aims: to find methods for objective measurement of motion in the orbit, and to determine the clinical use of these methods in patients with orbital disorders. This implied that a number of research questions had to be answered in the fields of both image science and of ophthalmology and orbitology. The results have established that measurement of the two- and three-dimensional motion of tissues in the orbit is feasible in humans. It can be imaged in 2-D and 3-D with Magnetic Resonance Imaging (MRI) sequences. It can be calculated objectively from these sequences using optical flow methods. First order techniques were found to be superior for this purpose as they are less sensitive to noise, partial volume effects and aliasing. A new first-order 3-D optical flow algorithm was developed for robust and computationally efficient three-dimensional optical flow estimation. A number of techniques were developed to visualize 2- and 3-D motion combined with the anatomy in 3-D space, using color-coding. The results of two clinical studies have established the clinical use of objective measurement of motion in the orbit in a number of orbital and motion disorders. The studies have resulted in an explanation for the persistent pain that may occur after enucleation of the globe, based on motion measurements in patients after enucleation. They have also shown that orbital tumors can be differentiated using motion studies. The results have also shown that the position of the rectus muscle pulleys is the same in patients with Graves disease and in normals. After surgical decompression of the orbit, the muscles and their pulleys are usually not displaced, except in cases where specific muscle paths and pulleys are displaced in specific patterns, resulting in specific motility disturbances. Thus, a new explanation has been found for the motility disturbances that may occur in some patients after decompression surgery for Graves orbitopathy. These last findings have led to the notion central to this thesis, namely that the orbital tissues are an organ in their own right, the organ of gaze, and do not need their bony orbit to function normally

    Model-based Optical Flow: Layers, Learning, and Geometry

    Get PDF
    The estimation of motion in video sequences establishes temporal correspondences between pixels and surfaces and allows reasoning about a scene using multiple frames. Despite being a focus of research for over three decades, computing motion, or optical flow, remains challenging due to a number of difficulties, including the treatment of motion discontinuities and occluded regions, and the integration of information from more than two frames. One reason for these issues is that most optical flow algorithms only reason about the motion of pixels on the image plane, while not taking the image formation pipeline or the 3D structure of the world into account. One approach to address this uses layered models, which represent the occlusion structure of a scene and provide an approximation to the geometry. The goal of this dissertation is to show ways to inject additional knowledge about the scene into layered methods, making them more robust, faster, and more accurate. First, this thesis demonstrates the modeling power of layers using the example of motion blur in videos, which is caused by fast motion relative to the exposure time of the camera. Layers segment the scene into regions that move coherently while preserving their occlusion relationships. The motion of each layer therefore directly determines its motion blur. At the same time, the layered model captures complex blur overlap effects at motion discontinuities. Using layers, we can thus formulate a generative model for blurred video sequences, and use this model to simultaneously deblur a video and compute accurate optical flow for highly dynamic scenes containing motion blur. Next, we consider the representation of the motion within layers. Since, in a layered model, important motion discontinuities are captured by the segmentation into layers, the flow within each layer varies smoothly and can be approximated using a low dimensional subspace. We show how this subspace can be learned from training data using principal component analysis (PCA), and that flow estimation using this subspace is computationally efficient. The combination of the layered model and the low-dimensional subspace gives the best of both worlds, sharp motion discontinuities from the layers and computational efficiency from the subspace. Lastly, we show how layered methods can be dramatically improved using simple semantics. Instead of treating all layers equally, a semantic segmentation divides the scene into its static parts and moving objects. Static parts of the scene constitute a large majority of what is shown in typical video sequences; yet, in such regions optical flow is fully constrained by the depth structure of the scene and the camera motion. After segmenting out moving objects, we consider only static regions, and explicitly reason about the structure of the scene and the camera motion, yielding much better optical flow estimates. Furthermore, computing the structure of the scene allows to better combine information from multiple frames, resulting in high accuracies even in occluded regions. For moving regions, we compute the flow using a generic optical flow method, and combine it with the flow computed for the static regions to obtain a full optical flow field. By combining layered models of the scene with reasoning about the dynamic behavior of the real, three-dimensional world, the methods presented herein push the envelope of optical flow computation in terms of robustness, speed, and accuracy, giving state-of-the-art results on benchmarks and pointing to important future research directions for the estimation of motion in natural scenes
    corecore