9 research outputs found
Detection of motion discontinuities.
<p>Some examples for motion discontinuities are given on the left bottom. We use a motion discontinuity detector built of an on-center-off-surround RF that will respond very strongly if center and surround motion differ. If a homogeneous flow field is presented, only a weak response is produced.</p
Sketch of the biologically inspired model.
<p>V1<sub>Model</sub> Motion and MT<sub>Model</sub> Motion represent the basic modules for optic flow estimation. In TO<sub>Model</sub> regions that have been occluded or disoccluded are estimated. In MSTl<sub>Model</sub> motion discontinuities are computed based on MT<sub>Model</sub> input due to spatial on-center-off-surround receptive fields. The information of areas MSTl<sub>Model</sub>, TO<sub>Model</sub>, and V2<sub>Model</sub> is combined in a higher level processing area (HLP<sub>Model</sub>). Feedforward connections are depicted with dark blue arrows, feedback connections with light blue arrows. The interactions between MSTl<sub>Model</sub> and TO<sub>Model</sub> are depicted with green arrows.</p
Overview of mechanisms for scene interpretation.
<p>Top row: The optic flow of the input image is computed in V1<sub>Model</sub> and MT<sub>Model</sub>, spatial contrast neurons in MSTl<sub>Model</sub> compute the motion discontinuities. Based on the detected motion boundaries a simple filling-in mechanism provides a scene segmentation. Bottom row: In TO<sub>Model</sub> input from V1<sub>Model</sub> neurons is used for a temporal on-center-off-surround processing step to detect occlusion and disocclusion regions. In HLP<sub>Model</sub> these regions are restricted to the motion discontinuities or luminance contours provided from V2<sub>Model</sub> to find the corresponding object that is adjacent to the occlusion region, namely the occluder. The results of the object segmentation are used to find the label of the corresponding object (indicated by the arrow from the top row, third column). Based on these data, the corresponding depth order can be computed. Interactions between MSTl<sub>Model</sub> and TO<sub>Model</sub> are not depicted in this figure.</p
3D scenario with two objects.
<p>This figure depicts a typical scenario for a person moving in a room. A static object (green) and a moving object (blue) are located in the room in front of the background. On the left, static occlusion regions with respect to the observer perspective are marked with gray overlay. Due to the spatial configuration the green object is partly covering the blue one, both objects are occluding the background texture. When the observer is moving forward, an expansional flow field is generated that is partly superimposed by the translational movement of the blue object. The optic flow, i.e. the projection of the 3D flow is shown on the projection plane. The alignment of the objects in the 2D projection is shown on the right. Here, also the kinetic occlusions generated by the movement of the blue object are depicted. On its left side, background texture is uncovered (disocclusion), on the right side it is temporarily covered (occlusion). Note, that the expansional flow leads to further kinetic occlusion regions along the outline of both objects, for simplicity this is not included in the sketch.</p
Experiment 3: Independently moving object in a scene with a moving observer.
<p>A) Input image of the sequence (generated in the XVR environment, download at <a href="http://www.vrmedia.it" target="_blank">www.vrmedia.it</a>), the gray arrow indicates the movement of the independently moving object. B) The optic flow in area MT<sub>Model</sub> is depicted, the object movement is correctly indicating a translation to the right. C) Occlusions and disocclusions are correctly detected on the right and left side of the object, respectively. The result shown here include feedback from MSTl<sub>Model</sub>. D) Motion discontinuities as computed by MSTl<sub>Model</sub> on-center-off-surround neurons show the object boundary, E) after the grouping and filling-in step the object can be segmented.</p
Experiment 4: City view through a window.
<p>Artificially generated scene with a background moving to the left while the aperture is fixed. A) One image of the input sequence. B) The mean optic flow as detected in MT<sub>Model</sub>. C) The movement generates occlusions on the left (black positions) and disocclusions on the right side (white positions). D) The motion discontinuities show the complete object boundary. E) After segmentation two objects are detected depicted in different colors, the aperture (gray) and the region within the window (white). F) The corresponding occluder to the occlusion positions with respect to the objects segmented like shown in E), the colors indicate the assignment. Most positions correctly indicate the aperture as the object causing the occlusion.</p
Experiment 2: Moving boxes.
<p>Results for an input sequence with 5 boxes and the background all moving in different directions. A) Input image with arrows indicating the movement of the objects. The background is slowly moving to the left. B) Mean optic flow estimations in area MT<sub>Model</sub> marked with a color code that is superimposed on the input image. In C) the detected occlusion (black) and disocclusion (white) regions are shown. Note that depending on the direction of the object movement these regions appear all along the object boundaries or just on two sides (for a movement in vertical or horizontal direction). D) Contours of the objects as provided by V2<sub>Model</sub> Form. This activity is used to achieve a clear localization of the occlusion boundary to the corresponding occluder. E) A clear segmentation of the object boundaries is achieved using the motion discontinuities detected with MSTl<sub>Model</sub> on-center-off-surround neurons. F) After the detected boundaries have been grouped and filled, the image is segmented in different regions representing the objects of the scene. G) Classification of object movement. The difference of object and background motion is computed as explained in the <a href="http://www.plosone.org/article/info:doi/10.1371/journal.pone.0003807#s2" target="_blank">Methods</a> section. Light object boundaries indicate a strong difference, darker outlines represent a movement similar to the background. Note, that object 5 and 2 have a strong motion contrast to the background despite the similar movement direction due to a much higher speed than the background. H) The results of the relative depth order derived automatically from the scene. A confidence value is applied to get a probability for the correctness of the depth order (indicated in percent). This is derived from the number of positions belonging to the object that indicate that the object is in front (#pos<sub>front</sub>) and the number of positions that indicate that the object is in the background (#pos<sub>bg</sub>) (conf = max(#pos<sub>front</sub>, #pos<sub>bg</sub>)/(#pos<sub>front</sub>+#pos<sub>bg</sub>).</p
Experiment 5: Rotating rectangle.
<p>A bar is rotating around its center in front of a stationary background. A) Input image of the sequence. B) The motion estimates of area MT<sub>Model</sub>, C) Discclusion regions appear on the upper left and the lower right, in contrast occlusions are found at the lower left and the upper right, this diagonal appearance is due to the rotational movement of the object. The result indicated here is without feedback from motion discontinuities. D) The motion boundary is correctly detected using the motion discontinuities, however, also in the object center MSTl<sub>Model</sub> neurons respond strongly when the movement switches from zero movement to the smallest movement that can be detected with the model. E) When including the interaction between occlusion and motion discontinuity detection, the erroneously detected central part is erased. F) Occlusion regions are correctly restricted due to feedback from motion discontinuity neurons as shown in D. The feedback is slightly blurred as occlusion regions may be significantly bigger than motion discontinuities.</p
Detection of occlusion regions.
<p>To detect occlusions and disocclusions in the motion sequence, we compare the motion energy at each spatial position that was estimated using the past frame pair t<sub>−1</sub>/t<sub>0</sub> and using the future frame pair t<sub>0</sub>/t<sub>1</sub>. A high difference typically occurs at occlusion and disocclusion positions due to regions that are only visible in t<sub>−1</sub> or t<sub>1</sub> and thus entail very ambiguous motion estimates.</p