8 research outputs found
Evaluation of the color image and video processing chain and visual quality management for consumer systems
With the advent of novel digital display technologies, color processing is increasingly becoming a key aspect in consumer video applications. Today’s state-of-the-art displays require sophisticated color and image reproduction techniques in order to achieve larger screen size, higher luminance and higher resolution than ever before. However, from color science perspective, there are clearly opportunities for improvement in the color reproduction capabilities of various emerging and conventional display technologies. This research seeks to identify potential areas for improvement in color processing in a video processing chain. As part of this research, various processes involved in a typical video processing chain in consumer video applications were reviewed. Several published color and contrast enhancement algorithms were evaluated, and a novel algorithm was developed to enhance color and contrast in images and videos in an effective and coordinated manner. Further, a psychophysical technique was developed and implemented for performing visual evaluation of color image and consumer video quality. Based on the performance analysis and visual experiments involving various algorithms, guidelines were proposed for the development of an effective color and contrast enhancement method for images and video applications. It is hoped that the knowledge gained from this research will help build a better understanding of color processing and color quality management methods in consumer video
Recommended from our members
Case Studies in Invertebrate Visual Processing: I. Spectral and Spatial Processing in the Early Visual System of Drosophila melanogaster II. Binocular Stereopsis in Sepia officinalis
This thesis addresses two aspects of visual processing in two different invertebrate organisms.
The fruit fly, Drosophila melanogaster, has emerged as a key model for invertebrate vision research. Despite extensive characterisation of motion vision, very little is known about how flies process colour information, or how the spectral content of light affects other visual modalities. With the aim to accurately dissect the different components of the Drosophila visual system responsible for processing colour, I have developed a versatile visual stimulation setup to probe for the combinations of spatial, temporal and spectral visual response properties. Using flies that express neural activity indicators, I can track visual responses to a colour stimulus (i.e. narrow bands of light across the spectrum) via a two-photon imaging system. The visual stimulus is projected on a specialised screen material that scatters wavelengths of light across the spectrum equally at all locations of the screen, thus enabling presentation of spatially structured stimuli. Using this setup, I have characterised spectral responses, intensity-response relationships, and receptive fields of neurons in the early visual system of a variety of genetically modified strains of Drosophila. Specifically, I compared visual responses in the medulla of flies expressing either a subset or all photoreceptor opsins, with differing levels of screening pigment present in the eye. I found layer-specific shifts of spectral response properties correlating with projection regions of photoreceptor terminals. I also
3
found that a reduction in screening pigment shifts the general spectral response in the neuropil towards the longer wavelengths of light. I have also mapped receptive fields across the different layers of the medulla for the peak spectral response wavelength. My results suggest that receptive field dimensions match the expected size predicted by the conservation of a columnar organisation in the medulla, with little variation from layer to layer. In a subset of these cells, we see an elongated receptive field suggestive of static orientation selectivity with an apparent split in the preferred axis of orientation of these receptive fields, with a near-orthogonal angle between the summed vectors of the split populations.
The camera type eyes of vertebrates and cephalopods exhibit remarkable convergence, but it is currently unknown if the mechanisms for visual information processing in these brains, which exhibit wildly disparate architecture, is also shared. I chose to investigate the visual processing mechanism known as stereopsis in the cuttlefish Sepia officinalis. Stereoscopic vision is used to assess depth information by comparing the disparity between left and right visual fields. This strategy is commonplace in vertebrates having evolved multiple times independently but has only been demonstrated in one invertebrate: the praying mantis. Cuttlefish require precise distance estimation during their predatory hunt when they extend two tentacles in a ballistic strike to catch their target. Using a 3D perception paradigm whereby the cuttlefish were fitted with anaglyph glasses, I show that these animals use stereopsis to resolve distance to their prey. Although this is not an exclusive depth perception mechanism for hunting, it does shorten the time and distance covered prior to striking at a target. Furthermore, stereopsis in cuttlefish works differently to vertebrates, as cuttlefish can extract stereopsis cues from anti-correlated stimuli.BBSRC Doctoral Training Partnershi
Rendering and display for multi-viewer tele-immersion
Video teleconferencing systems are widely deployed for business, education and personal use to enable face-to-face communication between people at distant sites. Unfortunately, the two-dimensional video of conventional systems does not correctly convey several important non-verbal communication cues such as eye contact and gaze awareness. Tele-immersion refers to technologies aimed at providing distant users with a more compelling sense of remote presence than conventional video teleconferencing. This dissertation is concerned with the particular challenges of interaction between groups of users at remote sites. The problems of video teleconferencing are exacerbated when groups of people communicate. Ideally, a group tele-immersion system would display views of the remote site at the right size and location, from the correct viewpoint for each local user. However, is is not practical to put a camera in every possible eye location, and it is not clear how to provide each viewer with correct and unique imagery. I introduce rendering techniques and multi-view display designs to support eye contact and gaze awareness between groups of viewers at two distant sites. With a shared 2D display, virtual camera views can improve local spatial cues while preserving scene continuity, by rendering the scene from novel viewpoints that may not correspond to a physical camera. I describe several techniques, including a compact light field, a plane sweeping algorithm, a depth dependent camera model, and video-quality proxies, suitable for producing useful views of a remote scene for a group local viewers. The first novel display provides simultaneous, unique monoscopic views to several users, with fewer user position restrictions than existing autostereoscopic displays. The second is a random hole barrier autostereoscopic display that eliminates the viewing zones and user position requirements of conventional autostereoscopic displays, and provides unique 3D views for multiple users in arbitrary locations
Stereoscopic high dynamic range imaging
Two modern technologies show promise to dramatically increase immersion in
virtual environments. Stereoscopic imaging captures two images representing
the views of both eyes and allows for better depth perception. High dynamic
range (HDR) imaging accurately represents real world lighting as opposed to
traditional low dynamic range (LDR) imaging. HDR provides a better contrast
and more natural looking scenes. The combination of the two technologies in
order to gain advantages of both has been, until now, mostly unexplored due to
the current limitations in the imaging pipeline. This thesis reviews both fields,
proposes stereoscopic high dynamic range (SHDR) imaging pipeline outlining the
challenges that need to be resolved to enable SHDR and focuses on capture and
compression aspects of that pipeline.
The problems of capturing SHDR images that would potentially require two
HDR cameras and introduce ghosting, are mitigated by capturing an HDR and
LDR pair and using it to generate SHDR images. A detailed user study compared
four different methods of generating SHDR images. Results demonstrated that
one of the methods may produce images perceptually indistinguishable from the
ground truth.
Insights obtained while developing static image operators guided the design
of SHDR video techniques. Three methods for generating SHDR video from an
HDR-LDR video pair are proposed and compared to the ground truth SHDR
videos. Results showed little overall error and identified a method with the least
error.
Once captured, SHDR content needs to be efficiently compressed. Five SHDR
compression methods that are backward compatible are presented. The proposed
methods can encode SHDR content to little more than that of a traditional single
LDR image (18% larger for one method) and the backward compatibility property
encourages early adoption of the format.
The work presented in this thesis has introduced and advanced capture and
compression methods for the adoption of SHDR imaging. In general, this research
paves the way for a novel field of SHDR imaging which should lead to improved
and more realistic representation of captured scenes
Low-level visual processing and its relation to neurological disease
Retinal neurons extract changes in image intensity across space, time, and wavelength. Retinal signal is transmitted to the early visual cortex, where the processing of low-level visual information occurs. The fundamental nature of these early visual pathways means that they are often compromised by neurological disease. This thesis had two aims. First, it aimed to investigate changes in visual processing in response to Parkinson’s disease (PD) by using electrophysiological recordings from animal models. Second, it aimed to use functional magnetic resonance imaging (fMRI) to investigate how low-level visual processes are represented in healthy human visual cortex, focusing on two pathways often compromised in disease; the magnocellular pathway and chromatic S-cone pathway. First, we identified a pathological mechanism of excitotoxicity in the visual system of Drosophila PD models. Next, we found that we could apply machine learning classifiers to multivariate visual response profiles recorded from the eye and brain of Drosophila and rodent PD models to accurately classify these animals into their correct class. Using fMRI and psychophysics, found that measurements of temporal contrast sensitivity differ as a function of visual space, with peripherally tuned voxels in early visual areas showing increased contrast sensitivity at a high temporal frequency. Finally, we used 7T fMRI to investigate systematic differences in achromatic and S-cone population receptive field (pRF) size estimates in the visual cortex of healthy humans. Unfortunately, we could not replicate the fundamental effect of pRF size increasing with eccentricity, indicating complications with our data and stimulus