12,959 research outputs found
The Gray-code filter kernels
Abstract In this paper we introduce a family of filter kernels -the Gray-Code Kernels (GCK) and demonstrate their use in image analysis. Filtering an image with a sequence of Gray-Code Kernels is highly efficient and requires only 2 operations per pixel for each filter kernel, independent of the size or dimension of the kernel. We show that the family of kernels is large and includes the Walsh-Hadamard kernels amongst others. The GCK can be used to approximate any desired kernel and as such forms a complete representation. The efficiency of computation using a sequence of GCK filters can be exploited for various real-time applications, such as, pattern detection, feature extraction, texture analysis, texture synthesis, and more
Transformation of stimulus correlations by the retina
Redundancies and correlations in the responses of sensory neurons seem to
waste neural resources but can carry cues about structured stimuli and may help
the brain to correct for response errors. To assess how the retina negotiates
this tradeoff, we measured simultaneous responses from populations of ganglion
cells presented with natural and artificial stimuli that varied greatly in
correlation structure. We found that pairwise correlations in the retinal
output remained similar across stimuli with widely different spatio-temporal
correlations including white noise and natural movies. Meanwhile, purely
spatial correlations tended to increase correlations in the retinal response.
Responding to more correlated stimuli, ganglion cells had faster temporal
kernels and tended to have stronger surrounds. These properties of individual
cells, along with gain changes that opposed changes in effective contrast at
the ganglion cell input, largely explained the similarity of pairwise
correlations across stimuli where receptive field measurements were possible.Comment: author list corrected in metadat
Color Constancy Using CNNs
In this work we describe a Convolutional Neural Network (CNN) to accurately
predict the scene illumination. Taking image patches as input, the CNN works in
the spatial domain without using hand-crafted features that are employed by
most previous methods. The network consists of one convolutional layer with max
pooling, one fully connected layer and three output nodes. Within the network
structure, feature learning and regression are integrated into one optimization
process, which leads to a more effective model for estimating scene
illumination. This approach achieves state-of-the-art performance on a standard
dataset of RAW images. Preliminary experiments on images with spatially varying
illumination demonstrate the stability of the local illuminant estimation
ability of our CNN.Comment: Accepted at DeepVision: Deep Learning in Computer Vision 2015 (CVPR
2015 workshop
- …