558 research outputs found
A Neural Model of Surface Perception: Lightness, Anchoring, and Filling-in
This article develops a neural model of how the visual system processes natural images under variable illumination conditions to generate surface lightness percepts. Previous models have clarified how the brain can compute the relative contrast of images from variably illuminate scenes. How the brain determines an absolute lightness scale that "anchors" percepts of surface lightness to us the full dynamic range of neurons remains an unsolved problem. Lightness anchoring properties include articulation, insulation, configuration, and are effects. The model quantatively simulates these and other lightness data such as discounting the illuminant, the double brilliant illusion, lightness constancy and contrast, Mondrian contrast constancy, and the Craik-O'Brien-Cornsweet illusion. The model also clarifies the functional significance for lightness perception of anatomical and neurophysiological data, including gain control at retinal photoreceptors, and spatioal contrast adaptation at the negative feedback circuit between the inner segment of photoreceptors and interacting horizontal cells. The model retina can hereby adjust its sensitivity to input intensities ranging from dim moonlight to dazzling sunlight. A later model cortical processing stages, boundary representations gate the filling-in of surface lightness via long-range horizontal connections. Variants of this filling-in mechanism run 100-1000 times faster than diffusion mechanisms of previous biological filling-in models, and shows how filling-in can occur at realistic speeds. A new anchoring mechanism called the Blurred-Highest-Luminance-As-White (BHLAW) rule helps simulate how surface lightness becomes sensitive to the spatial scale of objects in a scene. The model is also able to process natural images under variable lighting conditions.Air Force Office of Scientific Research (F49620-01-1-0397); Defense Advanced Research Projects Agency and the Office of Naval Research (N00014-95-1-0409); Office of Naval Research (N00014-01-1-0624
From Stereogram to Surface: How the Brain Sees the World in Depth
When we look at a scene, how do we consciously see surfaces infused with lightness and color at the correct depths? Random Dot Stereograms (RDS) probe how binocular disparity between the two eyes can generate such conscious surface percepts. Dense RDS do so despite the fact that they include multiple false binocular matches. Sparse stereograms do so even across large contrast-free regions with no binocular matches. Stereograms that define occluding and occluded surfaces lead to surface percepts wherein partially occluded textured surfaces are completed behind occluding textured surfaces at a spatial scale much larger than that of the texture elements themselves. Earlier models suggest how the brain detects binocular disparity, but not how RDS generate conscious percepts of 3D surfaces. A neural model predicts how the layered circuits of visual cortex generate these 3D surface percepts using interactions between visual boundary and surface representations that obey complementary computational rules.Air Force Office of Scientific Research (F49620-01-1-0397); National Science Foundation (EIA-01-30851, SBE-0354378); Office of Naval Research (N00014-01-1-0624
A Contrast/Filling-In Model of 3-D Lightness Perception
Wallach's ratio hypothesis states that local luminance ratios clr!termine lightness perception under variable illumination. While local luminance ratios successfully discount gradual variations in illumination (illumination constancy or Type I constancy), they fail to explain lightness constancy in general. Some examples of failures of the ratio hypothesis include effects suggesting the coplanar ratio hypothesis (Gilchrist 1977), "assimilation" effects, and configural effects such as the Benary cross, and White's illusion. The present article extends the Boundary Contour System/Feature Contour System (BCS/FCS) approach to provide an explanation of these effects in terms of a neural model of 3-D lightness perception. Lightness constancy of objects in front of different backgrounds (background constancy or Type II constancy) is used to provide functional constraints to the theory and suggest a contrast negation hypothesis which states that ratio measures between coplanar regions are given more weight in the determination of lightness. Simulations of the model applied to several stimuli including Benary cross and White's illusion show that contrast negation mechanisms modulate illumination constancy mechanisms to extend the explanatory power of the model. The model is also used to devise new stimuli that test theoretical predictions
A Contrast- and Luminance-Driven Multiscale Netowrk Model of Brightness Perception
A neural network model of brightness perception is developed to account for a wide variety of data, including the classical phenomenon of Mach bands, low- and high-contrast missing fundamental, luminance staircases, and non-linear contrast effects associated with sinusoidal waveforms. The model builds upon previous work on filling-in models that produce brightness profiles through the interaction of boundary and feature signals. Boundary computations that are sensitive to luminance steps and to continuous lumi- nance gradients are presented. A new interpretation of feature signals through the explicit representation of contrast-driven and luminance-driven information is provided and directly addresses the issue of brightness "anchoring." Computer simulations illustrate the model's competencies.Air Force Office of Scientific Research (F49620-92-J-0334); Northeast Consortium for Engineering Education (NCEE-A303-21-93); Office of Naval Research (N00014-91-J-4100); German BMFT grant (413-5839-01 1N 101 C/1); CNPq and NUTES/UFRJ, Brazi
Mach Bands: How Many Models are Possible? Recent Experiemental Findings and Modeling Attempts
Mach bands are illusory bright and dark bands seen where a luminance plateau meets a ramp, as in half-shadows or penumbras. A tremendous amount of work has been devoted to studying the psychophysics and the potential underlying neural circuitry concerning this phenomenon. A number of theoretical models have also been proposed, originating in the seminal studies of Mach himself. The present article reviews the main experimental findings after 1965 and the main recent theories of early vision that have attempted to discount for the effect. It is shown that the different theories share working principles and can be grouped in three clsses: a) feature-based; b) rule-based; and c) filling-in. In order to evaluate individual proposals it is necessary to consider them in the larger picture of visual science and to determine how they contribute to the understanding of vision in general.Air Force Office of Scientific Research (F49620-92-J-0334); Office of Naval Research (N00014-J-4100); COPPE/UFRJ, Brazi
Neural dynamics of invariant object recognition: relative disparity, binocular fusion, and predictive eye movements
How does the visual cortex learn invariant object categories as an observer scans
a depthful scene? Two neural processes that contribute to this ability are modeled in this
thesis.
The first model clarifies how an object is represented in depth. Cortical area V1
computes absolute disparity, which is the horizontal difference in retinal location of an
image in the left and right foveas. Many cells in cortical area V2 compute relative
disparity, which is the difference in absolute disparity of two visible features. Relative,
but not absolute, disparity is unaffected by the distance of visual stimuli from an
observer, and by vergence eye movements. A laminar cortical model of V2 that includes
shunting lateral inhibition of disparity-sensitive layer 4 cells causes a peak shift in cell
responses that transforms absolute disparity from V1 into relative disparity in V2.
The second model simulates how the brain maintains stable percepts of a 3D
scene during binocular movements. The visual cortex initiates the formation of a 3D boundary and surface representation by binocularly fusing corresponding features from
the left and right retinotopic images. However, after each saccadic eye movement, every
scenic feature projects to a different combination of retinal positions than before the
saccade. Yet the 3D representation, resulting from the prior fusion, is stable through the
post-saccadic re-fusion. One key to stability is predictive remapping: the system
anticipates the new retinal positions of features entailed by eye movements by using gain
fields that are updated by eye movement commands. The 3D ARTSCAN model
developed here simulates how perceptual, attentional, and cognitive interactions across
different brain regions within the What and Where visual processing streams interact to
coordinate predictive remapping, stable 3D boundary and surface perception, spatial
attention, and the learning of object categories that are invariant to changes in an object's
retinal projections. Such invariant learning helps the system to avoid treating each new
view of the same object as a distinct object to be learned. The thesis hereby shows how a
process that enables invariant object category learning can be extended to also enable
stable 3D scene perception
Bioplausible multiscale filtering in retino-cortical processing as a mechanism in perceptual grouping
Why does our visual system fail to reconstruct reality, when we look at
certain patterns? Where do Geometrical illusions start to emerge in the visual
pathway? How far should we take computational models of vision with the same
visual ability to detect illusions as we do? This study addresses these
questions, by focusing on a specific underlying neural mechanism involved in
our visual experiences that affects our final perception. Among many types of
visual illusion, Geometrical and, in particular, Tilt Illusions are rather
important, being characterized by misperception of geometric patterns involving
lines and tiles in combination with contrasting orientation, size or position.
Over the last decade, many new neurophysiological experiments have led to new
insights as to how, when and where retinal processing takes place, and the
encoding nature of the retinal representation that is sent to the cortex for
further processing. Based on these neurobiological discoveries, we provide
computer simulation evidence from modelling retinal ganglion cells responses to
some complex Tilt Illusions, suggesting that the emergence of tilt in these
illusions is partially related to the interaction of multiscale visual
processing performed in the retina. The output of our low-level filtering model
is presented for several types of Tilt Illusion, predicting that the final tilt
percept arises from multiple-scale processing of the Differences of Gaussians
and the perceptual interaction of foreground and background elements. Our
results suggest that this model has a high potential in revealing the
underlying mechanism connecting low-level filtering approaches to mid- and
high-level explanations such as Anchoring theory and Perceptual grouping.Comment: 23 pages, 8 figures, Brain Informatics journal: Full text access:
https://link.springer.com/article/10.1007/s40708-017-0072-
- …