11 research outputs found

    Object Segmentation from Motion Discontinuities and Temporal Occlusions–A Biologically Inspired Model

    Get PDF
    BACKGROUND: Optic flow is an important cue for object detection. Humans are able to perceive objects in a scene using only kinetic boundaries, and can perform the task even when other shape cues are not provided. These kinetic boundaries are characterized by the presence of motion discontinuities in a local neighbourhood. In addition, temporal occlusions appear along the boundaries as the object in front covers the background and the objects that are spatially behind it. METHODOLOGY/PRINCIPAL FINDINGS: From a technical point of view, the detection of motion boundaries for segmentation based on optic flow is a difficult task. This is due to the problem that flow detected along such boundaries is generally not reliable. We propose a model derived from mechanisms found in visual areas V1, MT, and MSTl of human and primate cortex that achieves robust detection along motion boundaries. It includes two separate mechanisms for both the detection of motion discontinuities and of occlusion regions based on how neurons respond to spatial and temporal contrast, respectively. The mechanisms are embedded in a biologically inspired architecture that integrates information of different model components of the visual processing due to feedback connections. In particular, mutual interactions between the detection of motion discontinuities and temporal occlusions allow a considerable improvement of the kinetic boundary detection. CONCLUSIONS/SIGNIFICANCE: A new model is proposed that uses optic flow cues to detect motion discontinuities and object occlusion. We suggest that by combining these results for motion discontinuities and object occlusion, object segmentation within the model can be improved. This idea could also be applied in other models for object segmentation. In addition, we discuss how this model is related to neurophysiological findings. The model was successfully tested both with artificial and real sequences including self and object motion

    Sketch of the biologically inspired model.

    No full text
    <p>V1<sub>Model</sub> Motion and MT<sub>Model</sub> Motion represent the basic modules for optic flow estimation. In TO<sub>Model</sub> regions that have been occluded or disoccluded are estimated. In MSTl<sub>Model</sub> motion discontinuities are computed based on MT<sub>Model</sub> input due to spatial on-center-off-surround receptive fields. The information of areas MSTl<sub>Model</sub>, TO<sub>Model</sub>, and V2<sub>Model</sub> is combined in a higher level processing area (HLP<sub>Model</sub>). Feedforward connections are depicted with dark blue arrows, feedback connections with light blue arrows. The interactions between MSTl<sub>Model</sub> and TO<sub>Model</sub> are depicted with green arrows.</p

    Detection of motion discontinuities.

    No full text
    <p>Some examples for motion discontinuities are given on the left bottom. We use a motion discontinuity detector built of an on-center-off-surround RF that will respond very strongly if center and surround motion differ. If a homogeneous flow field is presented, only a weak response is produced.</p

    Overview of mechanisms for scene interpretation.

    No full text
    <p>Top row: The optic flow of the input image is computed in V1<sub>Model</sub> and MT<sub>Model</sub>, spatial contrast neurons in MSTl<sub>Model</sub> compute the motion discontinuities. Based on the detected motion boundaries a simple filling-in mechanism provides a scene segmentation. Bottom row: In TO<sub>Model</sub> input from V1<sub>Model</sub> neurons is used for a temporal on-center-off-surround processing step to detect occlusion and disocclusion regions. In HLP<sub>Model</sub> these regions are restricted to the motion discontinuities or luminance contours provided from V2<sub>Model</sub> to find the corresponding object that is adjacent to the occlusion region, namely the occluder. The results of the object segmentation are used to find the label of the corresponding object (indicated by the arrow from the top row, third column). Based on these data, the corresponding depth order can be computed. Interactions between MSTl<sub>Model</sub> and TO<sub>Model</sub> are not depicted in this figure.</p

    3D scenario with two objects.

    No full text
    <p>This figure depicts a typical scenario for a person moving in a room. A static object (green) and a moving object (blue) are located in the room in front of the background. On the left, static occlusion regions with respect to the observer perspective are marked with gray overlay. Due to the spatial configuration the green object is partly covering the blue one, both objects are occluding the background texture. When the observer is moving forward, an expansional flow field is generated that is partly superimposed by the translational movement of the blue object. The optic flow, i.e. the projection of the 3D flow is shown on the projection plane. The alignment of the objects in the 2D projection is shown on the right. Here, also the kinetic occlusions generated by the movement of the blue object are depicted. On its left side, background texture is uncovered (disocclusion), on the right side it is temporarily covered (occlusion). Note, that the expansional flow leads to further kinetic occlusion regions along the outline of both objects, for simplicity this is not included in the sketch.</p

    Experiment 3: Independently moving object in a scene with a moving observer.

    No full text
    <p>A) Input image of the sequence (generated in the XVR environment, download at <a href="http://www.vrmedia.it" target="_blank">www.vrmedia.it</a>), the gray arrow indicates the movement of the independently moving object. B) The optic flow in area MT<sub>Model</sub> is depicted, the object movement is correctly indicating a translation to the right. C) Occlusions and disocclusions are correctly detected on the right and left side of the object, respectively. The result shown here include feedback from MSTl<sub>Model</sub>. D) Motion discontinuities as computed by MSTl<sub>Model</sub> on-center-off-surround neurons show the object boundary, E) after the grouping and filling-in step the object can be segmented.</p

    Experiment 4: City view through a window.

    No full text
    <p>Artificially generated scene with a background moving to the left while the aperture is fixed. A) One image of the input sequence. B) The mean optic flow as detected in MT<sub>Model</sub>. C) The movement generates occlusions on the left (black positions) and disocclusions on the right side (white positions). D) The motion discontinuities show the complete object boundary. E) After segmentation two objects are detected depicted in different colors, the aperture (gray) and the region within the window (white). F) The corresponding occluder to the occlusion positions with respect to the objects segmented like shown in E), the colors indicate the assignment. Most positions correctly indicate the aperture as the object causing the occlusion.</p

    Experiment 2: Moving boxes.

    No full text
    <p>Results for an input sequence with 5 boxes and the background all moving in different directions. A) Input image with arrows indicating the movement of the objects. The background is slowly moving to the left. B) Mean optic flow estimations in area MT<sub>Model</sub> marked with a color code that is superimposed on the input image. In C) the detected occlusion (black) and disocclusion (white) regions are shown. Note that depending on the direction of the object movement these regions appear all along the object boundaries or just on two sides (for a movement in vertical or horizontal direction). D) Contours of the objects as provided by V2<sub>Model</sub> Form. This activity is used to achieve a clear localization of the occlusion boundary to the corresponding occluder. E) A clear segmentation of the object boundaries is achieved using the motion discontinuities detected with MSTl<sub>Model</sub> on-center-off-surround neurons. F) After the detected boundaries have been grouped and filled, the image is segmented in different regions representing the objects of the scene. G) Classification of object movement. The difference of object and background motion is computed as explained in the <a href="http://www.plosone.org/article/info:doi/10.1371/journal.pone.0003807#s2" target="_blank">Methods</a> section. Light object boundaries indicate a strong difference, darker outlines represent a movement similar to the background. Note, that object 5 and 2 have a strong motion contrast to the background despite the similar movement direction due to a much higher speed than the background. H) The results of the relative depth order derived automatically from the scene. A confidence value is applied to get a probability for the correctness of the depth order (indicated in percent). This is derived from the number of positions belonging to the object that indicate that the object is in front (#pos<sub>front</sub>) and the number of positions that indicate that the object is in the background (#pos<sub>bg</sub>) (conf = max(#pos<sub>front</sub>, #pos<sub>bg</sub>)/(#pos<sub>front</sub>+#pos<sub>bg</sub>).</p

    Experiment 5: Rotating rectangle.

    No full text
    <p>A bar is rotating around its center in front of a stationary background. A) Input image of the sequence. B) The motion estimates of area MT<sub>Model</sub>, C) Discclusion regions appear on the upper left and the lower right, in contrast occlusions are found at the lower left and the upper right, this diagonal appearance is due to the rotational movement of the object. The result indicated here is without feedback from motion discontinuities. D) The motion boundary is correctly detected using the motion discontinuities, however, also in the object center MSTl<sub>Model</sub> neurons respond strongly when the movement switches from zero movement to the smallest movement that can be detected with the model. E) When including the interaction between occlusion and motion discontinuity detection, the erroneously detected central part is erased. F) Occlusion regions are correctly restricted due to feedback from motion discontinuity neurons as shown in D. The feedback is slightly blurred as occlusion regions may be significantly bigger than motion discontinuities.</p
    corecore