26 research outputs found

    A real-time neural system for color constancy

    Get PDF
    A neural network approach to the problem of color constancy is presented. Various algorithms based on Land's retinex theory are discussed with respect to neurobiological parallels, computational efficiency, and suitability for VLSI implementation. The efficiency of one algorithm is improved by the application of resistive grids and is tested in computer simulations; the simulations make clear the strengths and weaknesses of the algorithm. A novel extension to the algorithm is developed to address its weaknesses. An electronic system that is based on the original algorithm and that operates at video rates was built using subthreshold analog CMOS VLSI resistive grids. The system displays color constancy abilities and qualitatively mimics aspects of human color perception

    A VLSI Neural Network for Color Constancy

    Get PDF
    A system for color correction has been designed, built, and tested successfully; the essential components are three custom chips built using sub-threshold analog CMOS VLSI. The system, based on Land's Retinex theory of color constancy, produces colors similar in many respects to those produced by the visual system. Resistive grids implemented in analog VLSI perform the smoothing operation central to the algorithm at video rates. With the electronic system, the strengths and weaknesses of the algorithm are explored

    Computational mechanisms for colour and lightness constancy

    Get PDF
    Attributes of colour images have been found which allow colour and lightness constancy to be computed without prior knowledge of the illumination, even in complex scenes with three -dimensional objects and multiple light sources of different colours. The ratio of surface reflectance colour can be immediately determined between any two image points, however distant. It is possible to determine the number of spectrally independent light sources, and to isolate the effect of each. Reflectance edges across which the illumination remains constant can be correctly identified.In a scene illuminated by multiple distant point sources of distinguishalbe colours, the spatial angle between the sources and their brightness ratios can be computed from the image alone. If there are three or more sources then reflectance constancy is immediately possible without use of additional knowledge.The results are an extension of Edwin Land's Retinex algorithm. They account for previously unexplained data such as Gilchrist's veiling luminances and his single- colour rooms.The validity of the algorithms has been demonstrated by implementing them in a series of computer programs. The computational methods do not follow the edge or region finding paradigms of previous vision mechanisms. Although the new reflectance constancy cues occur in all normal scenes, it is likely that human vision makes use of only some of them.In a colour image all the pixels of a single surface colour lie in a single structure in flux space. The dimension of the structure equals the number of illumination colours. The reflectance ratio between two regions is determined by the transformation between their structures. Parallel tracing of edge pairs in their respective structures identifies an edge of constant illumination, and gives the lightness ratio of each such edge. Enhanced noise reduction techniques for colour pictures follow from the natural constraints on the flux structures

    Wavelet-Based Enhancement Technique for Visibility Improvement of Digital Images

    Get PDF
    Image enhancement techniques for visibility improvement of color digital images based on wavelet transform domain are investigated in this dissertation research. In this research, a novel, fast and robust wavelet-based dynamic range compression and local contrast enhancement (WDRC) algorithm to improve the visibility of digital images captured under non-uniform lighting conditions has been developed. A wavelet transform is mainly used for dimensionality reduction such that a dynamic range compression with local contrast enhancement algorithm is applied only to the approximation coefficients which are obtained by low-pass filtering and down-sampling the original intensity image. The normalized approximation coefficients are transformed using a hyperbolic sine curve and the contrast enhancement is realized by tuning the magnitude of the each coefficient with respect to surrounding coefficients. The transformed coefficients are then de-normalized to their original range. The detail coefficients are also modified to prevent edge deformation. The inverse wavelet transform is carried out resulting in a lower dynamic range and contrast enhanced intensity image. A color restoration process based on the relationship between spectral bands and the luminance of the original image is applied to convert the enhanced intensity image back to a color image. Although the colors of the enhanced images produced by the proposed algorithm are consistent with the colors of the original image, the proposed algorithm fails to produce color constant results for some pathological scenes that have very strong spectral characteristics in a single band. The linear color restoration process is the main reason for this drawback. Hence, a different approach is required for tackling the color constancy problem. The illuminant is modeled having an effect on the image histogram as a linear shift and adjust the image histogram to discount the illuminant. The WDRC algorithm is then applied with a slight modification, i.e. instead of using a linear color restoration, a non-linear color restoration process employing the spectral context relationships of the original image is applied. The proposed technique solves the color constancy issue and the overall enhancement algorithm provides attractive results improving visibility even for scenes with near-zero visibility conditions. In this research, a new wavelet-based image interpolation technique that can be used for improving the visibility of tiny features in an image is presented. In wavelet domain interpolation techniques, the input image is usually treated as the low-pass filtered subbands of an unknown wavelet-transformed high-resolution (HR) image, and then the unknown high-resolution image is produced by estimating the wavelet coefficients of the high-pass filtered subbands. The same approach is used to obtain an initial estimate of the high-resolution image by zero filling the high-pass filtered subbands. Detail coefficients are estimated via feeding this initial estimate to an undecimated wavelet transform (UWT). Taking an inverse transform after replacing the approximation coefficients of the UWT with initially estimated HR image, results in the final interpolated image. Experimental results of the proposed algorithms proved their superiority over the state-of-the-art enhancement and interpolation techniques

    A real-time neural system for color constancy

    Get PDF
    A neural network approach to the problem of color constancy is presented. Various algorithms based on Land's retinex theory are discussed with respect to neurobiological parallels, computational efficiency, and suitability for VLSI implementation. The efficiency of one algorithm is improved by the application of resistive grids and is tested in computer simulations; the simulations make clear the strengths and weaknesses of the algorithm. A novel extension to the algorithm is developed to address its weaknesses. An electronic system that is based on the original algorithm and that operates at video rates was built using subthreshold analog CMOS VLSI resistive grids. The system displays color constancy abilities and qualitatively mimics aspects of human color perception

    Fusion of Visual and Thermal Images Using Genetic Algorithms

    Get PDF
    Demands for reliable person identification systems have increased significantly due to highly security risks in our daily life. Recently, person identification systems are built upon the biometrics techniques such as face recognition. Although face recognition systems have reached a certain level of maturity, their accomplishments in practical applications are restricted by some challenges, such as illumination variations. Current visual face recognition systems perform relatively well under controlled illumination conditions while thermal face recognition systems are more advantageous for detecting disguised faces or when there is no illumination control. A hybrid system utilizing both visual and thermal images for face recognition will be beneficial. The overall goal of this research is to develop computational methods that improve image quality by fusing visual and thermal face images. First, three novel algorithms were proposed to enhance visual face images. In those techniques, specifical nonlinear image transfer functions were developed and parameters associated with the functions were determined by image statistics, making the algorithms adaptive. Second, methods were developed for registering the enhanced visual images to their corresponding thermal images. Landmarks in the images were first detected and a subset of those landmarks were selected to compute a transformation matrix for the registration. Finally, A Genetic algorithm was proposed to fuse the registered visual and thermal images. Experimental results showed that image quality can be significantly improved using the proposed framework

    Learning as a Nonlinear Line of Attraction for Pattern Association, Classification and Recognition

    Get PDF
    Development of a mathematical model for learning a nonlinear line of attraction is presented in this dissertation, in contrast to the conventional recurrent neural network model in which the memory is stored in an attractive fixed point at discrete location in state space. A nonlinear line of attraction is the encapsulation of attractive fixed points scattered in state space as an attractive nonlinear line, describing patterns with similar characteristics as a family of patterns. It is usually of prime imperative to guarantee the convergence of the dynamics of the recurrent network for associative learning and recall. We propose to alter this picture. That is, if the brain remembers by converging to the state representing familiar patterns, it should also diverge from such states when presented by an unknown encoded representation of a visual image. The conception of the dynamics of the nonlinear line attractor network to operate between stable and unstable states is the second contribution in this dissertation research. These criteria can be used to circumvent the plasticity-stability dilemma by using the unstable state as an indicator to create a new line for an unfamiliar pattern. This novel learning strategy utilizes stability (convergence) and instability (divergence) criteria of the designed dynamics to induce self-organizing behavior. The self-organizing behavior of the nonlinear line attractor model can manifest complex dynamics in an unsupervised manner. The third contribution of this dissertation is the introduction of the concept of manifold of color perception. The fourth contribution of this dissertation is the development of a nonlinear dimensionality reduction technique by embedding a set of related observations into a low-dimensional space utilizing the result attained by the learned memory matrices of the nonlinear line attractor network. Development of a system for affective states computation is also presented in this dissertation. This system is capable of extracting the user\u27s mental state in real time using a low cost computer. It is successfully interfaced with an advanced learning environment for human-computer interaction

    Cognitive Image Fusion and Assessment

    Get PDF

    Real-Time Full Color Multiband Night Vision

    Get PDF

    TESTING COLOR APPEARANCE MODELS IN COMPLEX SCENE

    Get PDF
    The sensation of sight is our primary mechanism to perceive the world around us. However it is not yet perfectly clear how the human visual system works. The images of the world are formed on the retina, captured by sensors and converted in signals sent to the brain. Here the signals are processed and somehow interpreted, thus we are able to see. A lot of information, hypothesis, hints come from a field of the optical (or visual) illusions. These illusions have led many scientists and researchers to ask themselves why we are not able to interpret in a correct way some particular scenes. The word \u201cinterpret\u201d underlines the fact that the brain, and not only the eye, is involved in the process of vision. If our sight worked as a measurement tool, similar to a spectrophotometer, we would not perceive, for example, the simultaneous contrast phenomenon, in which a grey patch placed on a black background appears lighter than an identical coloured patch on a white background. So, why do we perceive the patches as different, while the light that reaches the eyes is the same? In the same way we would not be able to distinguish a white paper seen in a room lit with a red light from a red paper seen under a white light, however humans can do this. These phenomena are called colour appearance phenomena. Simulating the appearance is the objective of a range of computational models called colour appearance models. In this dissertation themes about colour appearance models are addressed. Specific experiments, performed by human observers, aim to evaluate and measure the appearance. Different algorithms are tested in order to compare the results of the computational model with the human sensations about colours. From these data, a new printing pipeline is developed, able to simulate the appearance of advertising billboard in different context
    corecore