3,502 research outputs found
A real-time neural system for color constancy
A neural network approach to the problem of color constancy is presented. Various algorithms based on Land's retinex theory are discussed with respect to neurobiological parallels, computational efficiency, and suitability for VLSI implementation. The efficiency of one algorithm is improved by the application of resistive grids and is tested in computer simulations; the simulations make clear the strengths and weaknesses of the algorithm. A novel extension to the algorithm is developed to address its weaknesses. An electronic system that is based on the original algorithm and that operates at video rates was built using subthreshold analog CMOS VLSI resistive grids. The system displays color constancy abilities and qualitatively mimics aspects of human color perception
A real-time neural system for color constancy
A neural network approach to the problem of color constancy is presented. Various algorithms based on Land's retinex theory are discussed with respect to neurobiological parallels, computational efficiency, and suitability for VLSI implementation. The efficiency of one algorithm is improved by the application of resistive grids and is tested in computer simulations; the simulations make clear the strengths and weaknesses of the algorithm. A novel extension to the algorithm is developed to address its weaknesses. An electronic system that is based on the original algorithm and that operates at video rates was built using subthreshold analog CMOS VLSI resistive grids. The system displays color constancy abilities and qualitatively mimics aspects of human color perception
Scene illumination classification based on histogram quartering of CIE-Y component
Despite the rapidly expanding research into various aspects of illumination estimation
methods, there are limited number of studies addressing illumination classification for
different purposes. The increasing demand for color constancy process, wide application
of it and high dependency of color constancy to illumination estimation makes this
research topic challenging. Definitely, an accurate estimation of illumination in the
image will provide a better platform for doing correction and finally will lead in better
color constancy performance. The main purpose of any illumination estimation
algorithm from any type and class is to estimate an accurate number as illumination. In
scene illumination estimation dealing with large range of illumination and small
variation of it is critical. Those algorithms which performed estimation carrying out lots
of calculation that leads in expensive methods in terms of computing resources. There
are several technical limitations in estimating an accurate number as illumination. In
addition using light temperature in all previous studies leads to have complicated and
computationally expensive methods. On the other hand classification is appropriate for
applications like photography when most of the images have been captured in a small set
of illuminants like scene illuminant. This study aims to develop a framework of image
illumination classifier that is capable of classifying images under different illumination
levels with an acceptable accuracy. The method will be tested on real scene images
captured with illumination level is measured. This method is a combination of physic
based methods and data driven (statistical) methods that categorize the images based on
statistical features extracted from illumination histogram of image. The result of
categorization will be validated using inherent illumination data of scene. Applying the
improving algorithm for characterizing histograms (histogram quartering) handed out
the advantages of high accuracy. A trained neural network which is the parameters are
tuned for this specific application has taken into account in order to sort out the image
into predefined groups. Finally, for performance and accuracy evaluation
misclassification error percentages, Mean Square Error (MSE), regression analysis and response time are used. This developed method finally will result in a high accuracy and
straightforward classification system especially for illumination concept. The results of
this study strongly demonstrate that light intensity with the help of a perfectly tuned
neural network can be used as the light property to establish a scene illumination
classification system
Fully Point-wise Convolutional Neural Network for Modeling Statistical Regularities in Natural Images
Modeling statistical regularity plays an essential role in ill-posed image
processing problems. Recently, deep learning based methods have been presented
to implicitly learn statistical representation of pixel distributions in
natural images and leverage it as a constraint to facilitate subsequent tasks,
such as color constancy and image dehazing. However, the existing CNN
architecture is prone to variability and diversity of pixel intensity within
and between local regions, which may result in inaccurate statistical
representation. To address this problem, this paper presents a novel fully
point-wise CNN architecture for modeling statistical regularities in natural
images. Specifically, we propose to randomly shuffle the pixels in the origin
images and leverage the shuffled image as input to make CNN more concerned with
the statistical properties. Moreover, since the pixels in the shuffled image
are independent identically distributed, we can replace all the large
convolution kernels in CNN with point-wise () convolution kernels while
maintaining the representation ability. Experimental results on two
applications: color constancy and image dehazing, demonstrate the superiority
of our proposed network over the existing architectures, i.e., using
1/101/100 network parameters and computational cost while achieving
comparable performance.Comment: 9 pages, 7 figures. To appear in ACM MM 201
Learning Invariant Color Features for Person Re-Identification
Matching people across multiple camera views known as person
re-identification, is a challenging problem due to the change in visual
appearance caused by varying lighting conditions. The perceived color of the
subject appears to be different with respect to illumination. Previous works
use color as it is or address these challenges by designing color spaces
focusing on a specific cue. In this paper, we propose a data driven approach
for learning color patterns from pixels sampled from images across two camera
views. The intuition behind this work is that, even though pixel values of same
color would be different across views, they should be encoded with the same
values. We model color feature generation as a learning problem by jointly
learning a linear transformation and a dictionary to encode pixel values. We
also analyze different photometric invariant color spaces. Using color as the
only cue, we compare our approach with all the photometric invariant color
spaces and show superior performance over all of them. Combining with other
learned low-level and high-level features, we obtain promising results in
ViPER, Person Re-ID 2011 and CAVIAR4REID datasets
Nonlinearities and Adaptation of Color Vision from Sequential Principal Curves Analysis
Mechanisms of human color vision are characterized by two phenomenological
aspects: the system is nonlinear and adaptive to changing environments.
Conventional attempts to derive these features from statistics use separate
arguments for each aspect. The few statistical approaches that do consider both
phenomena simultaneously follow parametric formulations based on empirical
models. Therefore, it may be argued that the behavior does not come directly
from the color statistics but from the convenient functional form adopted. In
addition, many times the whole statistical analysis is based on simplified
databases that disregard relevant physical effects in the input signal, as for
instance by assuming flat Lambertian surfaces. Here we address the simultaneous
statistical explanation of (i) the nonlinear behavior of achromatic and
chromatic mechanisms in a fixed adaptation state, and (ii) the change of such
behavior. Both phenomena emerge directly from the samples through a single
data-driven method: the Sequential Principal Curves Analysis (SPCA) with local
metric. SPCA is a new manifold learning technique to derive a set of sensors
adapted to the manifold using different optimality criteria. A new database of
colorimetrically calibrated images of natural objects under these illuminants
was collected. The results obtained by applying SPCA show that the
psychophysical behavior on color discrimination thresholds, discount of the
illuminant and corresponding pairs in asymmetric color matching, emerge
directly from realistic data regularities assuming no a priori functional form.
These results provide stronger evidence for the hypothesis of a statistically
driven organization of color sensors. Moreover, the obtained results suggest
that color perception at this low abstraction level may be guided by an error
minimization strategy rather than by the information maximization principle
Unsupervised Learning for Color Constancy
Most digital camera pipelines use color constancy methods to reduce the
influence of illumination and camera sensor on the colors of scene objects. The
highest accuracy of color correction is obtained with learning-based color
constancy methods, but they require a significant amount of calibrated training
images with known ground-truth illumination. Such calibration is time
consuming, preferably done for each sensor individually, and therefore a major
bottleneck in acquiring high color constancy accuracy. Statistics-based methods
do not require calibrated training images, but they are less accurate. In this
paper an unsupervised learning-based method is proposed that learns its
parameter values after approximating the unknown ground-truth illumination of
the training images, thus avoiding calibration. In terms of accuracy the
proposed method outperforms all statistics-based and many learning-based
methods. An extension of the method is also proposed, which learns the needed
parameters from non-calibrated images taken with one sensor and which can then
be successfully applied to images taken with another sensor. This effectively
enables inter-camera unsupervised learning for color constancy. Additionally, a
new high quality color constancy benchmark dataset with 1707 calibrated images
is created, used for testing, and made publicly available. The results are
presented and discussed. The source code and the dataset are available at
http://www.fer.unizg.hr/ipg/resources/color_constancy/.Comment: 15 pages, 16 figure
A Lightweight Approach for On-the-Fly Reflectance Estimation
Estimating surface reflectance (BRDF) is one key component for complete 3D
scene capture, with wide applications in virtual reality, augmented reality,
and human computer interaction. Prior work is either limited to controlled
environments (\eg gonioreflectometers, light stages, or multi-camera domes), or
requires the joint optimization of shape, illumination, and reflectance, which
is often computationally too expensive (\eg hours of running time) for
real-time applications. Moreover, most prior work requires HDR images as input
which further complicates the capture process. In this paper, we propose a
lightweight approach for surface reflectance estimation directly from -bit
RGB images in real-time, which can be easily plugged into any 3D
scanning-and-fusion system with a commodity RGBD sensor. Our method is
learning-based, with an inference time of less than 90ms per scene and a model
size of less than 340K bytes. We propose two novel network architectures,
HemiCNN and Grouplet, to deal with the unstructured input data from multiple
viewpoints under unknown illumination. We further design a loss function to
resolve the color-constancy and scale ambiguity. In addition, we have created a
large synthetic dataset, SynBRDF, which comprises a total of K RGBD images
rendered with a physically-based ray tracer under a variety of natural
illumination, covering materials and shapes. SynBRDF is the first
large-scale benchmark dataset for reflectance estimation. Experiments on both
synthetic data and real data show that the proposed method effectively recovers
surface reflectance, and outperforms prior work for reflectance estimation in
uncontrolled environments.Comment: ICCV 201
DeepIlluminance: Contextual Illuminance Estimation via Deep Neural Networks
Computational color constancy refers to the estimation of the scene
illumination and makes the perceived color relatively stable under varying
illumination. In the past few years, deep Convolutional Neural Networks (CNNs)
have delivered superior performance in illuminant estimation. Several
representative methods formulate it as a multi-label prediction problem by
learning the local appearance of image patches using CNNs. However, these
approaches inevitably make incorrect estimations for the ambiguous patches
affected by their neighborhood contexts. Inaccurate local estimates are likely
to bring in degraded performance when combining into a global prediction. To
address the above issues, we propose a contextual deep network for patch-based
illuminant estimation equipped with refinement. First, the contextual net with
a center-surround architecture extracts local contextual features from image
patches, and generates initial illuminant estimates and the corresponding color
corrected patches. The patches are sampled based on the observation that pixels
with large color differences describe the illumination well. Then, the
refinement net integrates the input patches with the corrected patches in
conjunction with the use of intermediate features to improve the performance.
To train such a network with numerous parameters, we propose a stage-wise
training strategy, in which the features and the predicted illuminant from
previous stages are provided to the next learning stage with more finer
estimates recovered. Experiments show that our approach obtains competitive
performance on two illuminant estimation benchmarks.Comment: 12 pages, 7 figure
Learn to Model Motion from Blurry Footages
It is difficult to recover the motion field from a real-world footage given a
mixture of camera shake and other photometric effects. In this paper we propose
a hybrid framework by interleaving a Convolutional Neural Network (CNN) and a
traditional optical flow energy. We first conduct a CNN architecture using a
novel learnable directional filtering layer. Such layer encodes the angle and
distance similarity matrix between blur and camera motion, which is able to
enhance the blur features of the camera-shake footages. The proposed CNNs are
then integrated into an iterative optical flow framework, which enable the
capability of modelling and solving both the blind deconvolution and the
optical flow estimation problems simultaneously. Our framework is trained
end-to-end on a synthetic dataset and yields competitive precision and
performance against the state-of-the-art approaches.Comment: Preprint of our paper accepted by Pattern Recognitio
- …