5,933 research outputs found
Information recovery from rank-order encoded images
The time to detection of a visual stimulus by the primate eye is recorded at
100 – 150ms. This near instantaneous recognition is in spite of the considerable
processing required by the several stages of the visual pathway to recognise and
react to a visual scene. How this is achieved is still a matter of speculation.
Rank-order codes have been proposed as a means of encoding by the primate
eye in the rapid transmission of the initial burst of information from the sensory
neurons to the brain. We study the efficiency of rank-order codes in encoding
perceptually-important information in an image. VanRullen and Thorpe built a
model of the ganglion cell layers of the retina to simulate and study the viability
of rank-order as a means of encoding by retinal neurons. We validate their model
and quantify the information retrieved from rank-order encoded images in terms
of the visually-important information recovered. Towards this goal, we apply
the ‘perceptual information preservation algorithm’, proposed by Petrovic and
Xydeas after slight modification. We observe a low information recovery due
to losses suffered during the rank-order encoding and decoding processes. We
propose to minimise these losses to recover maximum information in minimum
time from rank-order encoded images. We first maximise information recovery by
using the pseudo-inverse of the filter-bank matrix to minimise losses during rankorder
decoding. We then apply the biological principle of lateral inhibition to
minimise losses during rank-order encoding. In doing so, we propose the Filteroverlap
Correction algorithm. To test the perfomance of rank-order codes in
a biologically realistic model, we design and simulate a model of the foveal-pit
ganglion cells of the retina keeping close to biological parameters. We use this
as a rank-order encoder and analyse its performance relative to VanRullen and
Thorpe’s retinal model
Large-scale multielectrode recording and stimulation of neural activity
Large circuits of neurons are employed by the brain to encode and process information. How this encoding and processing is carried out is one of the central questions in neuroscience. Since individual neurons communicate with each other through electrical signals (action potentials), the recording of neural activity with arrays of extracellular electrodes is uniquely suited for the investigation of this question. Such recordings provide the combination of the best spatial (individual neurons) and temporal (individual action-potentials) resolutions compared to other large-scale imaging methods. Electrical stimulation of neural activity in turn has two very important applications: it enhances our understanding of neural circuits by allowing active interactions with them, and it is a basis for a large variety of neural prosthetic devices. Until recently, the state-of-the-art in neural activity recording systems consisted of several dozen electrodes with inter-electrode spacing ranging from tens to hundreds of microns. Using silicon microstrip detector expertise acquired in the field of high-energy physics, we created a unique neural activity readout and stimulation framework that consists of high-density electrode arrays, multi-channel custom-designed integrated circuits, a data acquisition system, and data-processing software. Using this framework we developed a number of neural readout and stimulation systems: (1) a 512-electrode system for recording the simultaneous activity of as many as hundreds of neurons, (2) a 61-electrode system for electrical stimulation and readout of neural activity in retinas and brain-tissue slices, and (3) a system with telemetry capabilities for recording neural activity in the intact brain of awake, naturally behaving animals. We will report on these systems, their various applications to the field of neurobiology, and novel scientific results obtained with some of them. We will also outline future directions
Recommended from our members
Functional evidence for cone-specific connectivity in the human retina
NoPhysiological studies of colour vision have not yet resolved the controversial issue of how chromatic opponency is constructed at a neuronal level. Two competing theories, the cone-selective hypothesis and the random-wiring hypothesis, are currently equivocal to the architecture of the primate retina. In central vision, both schemes are capable of producing colour opponency due to the fact that receptive field centres receive input from a single bipolar cell ¿ the so called `private line arrangement¿. However, in peripheral vision this single-cone input to the receptive field centre is lost, so that any random cone connectivity would result in a predictable reduction in the quality of colour vision. Behavioural studies thus far have indeed suggested a selective loss of chromatic sensitivity in peripheral vision. We investigated chromatic sensitivity as a function of eccentricity for the cardinal chromatic (L/M and S/(L + M)) and achromatic (L + M) pathways, adopting stimulus size as the critical variable. Results show that performance can be equated across the visual field simply by a change of scale (size). In other words, there exists no qualitative loss of chromatic sensitivity across the visual field. Critically, however, the quantitative nature of size dependency for each of the cardinal chromatic and achromatic mechanisms is very specific, reinforcing their independence in terms of anatomy and genetics. Our data provide clear evidence for a physiological model of primate colour vision that retains chromatic quality in peripheral vision, thus supporting the cone-selective hypothesis
Evidence for an additive inhibitory component of contrast adaptation
The latency of visual responses generally decreases as contrast increases.
Recording in the lateral geniculate nucleus (LGN), we find that response
latency increases with increasing contrast in ON cells for some visual stimuli.
We propose that this surprising latency trend can be explained if ON cells rest
further from threshold at higher contrasts. Indeed, while contrast changes
caused a combination of multiplicative gain change and additive shift in LGN
cells, the additive shift predominated in ON cells. Modeling results supported
this theory: the ON cell latency trend was found when the distance-to-threshold
shifted with contrast, but not when distance-to-threshold was fixed across
contrasts. In the model, latency also increases as surround-to-center ratios
increase, which has been shown to occur at higher contrasts. We propose that
higher-contrast full-field stimuli can evoke more surround inhibition, shifting
the potential further from spiking threshold and thereby increasing response
latency
A Neural Model of How the Brain Computes Heading from Optic Flow in Realistic Scenes
Animals avoid obstacles and approach goals in novel cluttered environments using visual information, notably optic flow, to compute heading, or direction of travel, with respect to objects in the environment. We present a neural model of how heading is computed that describes interactions among neurons in several visual areas of the primate magnocellular pathway, from retina through V1, MT+, and MSTd. The model produces outputs which are qualitatively and quantitatively similar to human heading estimation data in response to complex natural scenes. The model estimates heading to within 1.5° in random dot or photo-realistically rendered scenes and within 3° in video streams from driving in real-world environments. Simulated rotations of less than 1 degree per second do not affect model performance, but faster simulated rotation rates deteriorate performance, as in humans. The model is part of a larger navigational system that identifies and tracks objects while navigating in cluttered environments.National Science Foundation (SBE-0354378, BCS-0235398); Office of Naval Research (N00014-01-1-0624); National-Geospatial Intelligence Agency (NMA201-01-1-2016
- …