6 research outputs found
Estimation of multiple illuminants from a single image of arbitrary known geometry
Abstract. We present a new method for the detection and estimation of multiple illuminants, using one image of any object withknown geometry and Lambertian reflectance. Our method obviates the need to modify the imaged scene by inserting calibration objects of any particular geometry, relying instead on partial knowledge of the geometry of the scene. Thus, the recovered multiple illuminants can be used both for image-based rendering and for shape reconstruction. We first develop our method for the case of a sphere with known size, illuminated by a set of directional light sources. In general, eachpoint of sucha sphere will be illuminated by a subset of these sources. We propose a novel, robust way to segment the surface into regions, witheachregion illuminated by a different set of sources. The regions are separated by boundaries consisting of critical points (points where one illuminant is perpendicular to the normal). Our region-based recursive least-squares method is impervious to noise and missing data and significantly outperforms a previous boundary-based method using spheres[21]. This robustness to missing data is crucial to extending the method to surfaces of arbitrary smooth geometry, other than spheres. We map the normals of the arbitrary shape onto a sphere, which we can then segment, even when only a subset of the normals is available on the scene. We demonstrate experimentally the accuracy of our method, both in detecting the number of light sources and in estimating their directions, by testing on images of a variety of synthetic and real objects. 1
Algorithms for the enhancement of dynamic range and colour constancy of digital images & video
One of the main objectives in digital imaging is to mimic the capabilities of the human eye, and perhaps, go beyond in certain aspects. However, the human visual system is so versatile, complex, and only partially understood that no up-to-date imaging technology has been able to accurately reproduce the capabilities of the it. The extraordinary capabilities of the human eye have become a crucial shortcoming in digital imaging, since digital photography, video recording, and computer vision applications have continued to demand more realistic and accurate imaging reproduction and analytic capabilities.
Over decades, researchers have tried to solve the colour constancy problem, as well as extending the dynamic range of digital imaging devices by proposing a number of algorithms and instrumentation approaches. Nevertheless, no unique solution has been identified; this is partially due to the wide range of computer vision applications that require colour constancy and high dynamic range imaging, and the complexity of the human visual system to achieve effective colour constancy and dynamic range capabilities.
The aim of the research presented in this thesis is to enhance the overall image quality within an image signal processor of digital cameras by achieving colour constancy and extending dynamic range capabilities. This is achieved by developing a set of advanced image-processing algorithms that are robust to a number of practical challenges and feasible to be implemented within an image signal processor used in consumer electronics imaging devises.
The experiments conducted in this research show that the proposed algorithms supersede state-of-the-art methods in the fields of dynamic range and colour constancy. Moreover, this unique set of image processing algorithms show that if they are used within an image signal processor, they enable digital camera devices to mimic the human visual system s dynamic range and colour constancy capabilities; the ultimate goal of any state-of-the-art technique, or commercial imaging device
A PDE approach to Shape from Shading via Photometric Stereo
We present a new analytic and numerical approach to the shape from shading using photometric stereo technique. That is, we solve the problem to find the 3D surface of an object starting from its several 2D pictures taken from the same point of view, but changing, for every image, the direction of the light source
Factor Graphs for Computer Vision and Image Processing
Factor graphs have been used extensively in the decoding of error
correcting codes such as turbo codes, and in signal processing.
However, while computer vision and pattern recognition are awash
with graphical model usage, it is some-what surprising that
factor graphs are still somewhat under-researched in these
communities. This is surprising because factor graphs naturally
generalise both Markov random fields and Bayesian networks.
Moreover, they are useful in modelling relationships between
variables that are not necessarily probabilistic and allow for
efficient marginalisation via a sum-product of probabilities.
In this thesis, we present and illustrate the utility of factor
graphs in the vision community through some of the field’s
popular problems. The thesis does so with a particular focus on
maximum a posteriori (MAP) inference in graphical
structures with layers. To this end, we are able to break-down
complex problems into factored representations and more
computationally realisable constructions. Firstly, we present a
sum-product framework that uses the explicit factorisation
in local subgraphs from the partitioned factor graph of a layered
structure to perform inference. This provides an efficient method
to perform inference since exact inference is attainable in the
resulting local subtrees. Secondly, we extend this framework to
the entire graphical structure without partitioning, and discuss
preliminary ways to combine outputs from a multilevel
construction. Lastly, we further our endeavour to combine
evidence from different methods through
a simplicial spanning tree reparameterisation of the factor graph
in a way that ensures consistency, to produce an ensembled and
improved result. Throughout the thesis, the underlying feature we
make use of is to enforce adjacency constraints using Delaunay
triangulations computed by adding points dynamically, or using a
convex hull algorithm. The adjacency relationships from Delaunay
triangulations aid the factor graph approaches in this thesis to
be both efficient and
competitive for computer vision tasks. This is because of the low
treewidth they provide in local subgraphs, as well as the
reparameterised interpretation of the graph they form through the
spanning tree of simplexes. While exact inference is known to be
intractable for junction trees obtained from the loopy graphs in
computer vision, in this thesis we are able to effect exact
inference on our spanning tree of simplexes. More importantly,
the approaches presented here are not restricted to the computer
vision and image processing fields, but are extendable to more
general applications that involve distributed computations
Photometric Reconstruction from Images: New Scenarios and Approaches for Uncontrolled Input Data
The changes in surface shading caused by varying illumination constitute an important cue to discern fine details and recognize the shape of textureless objects.
Humans perform this task subconsciously, but it is challenging for a computer because several variables are unknown and intermix in the light distribution that actually reaches the eye or camera.
In this work, we study algorithms and techniques to automatically recover the surface orientation and reflectance properties from multiple images of a scene.
Photometric reconstruction techniques have been investigated for decades but are still restricted to industrial applications and research laboratories.
Making these techniques work on more general, uncontrolled input without specialized capture setups has to be the next step but is not yet solved.
We explore the current limits of photometric shape recovery in terms of input data and propose ways to overcome some of its restrictions.
Many approaches, especially for non-Lambertian surfaces, rely on the illumination and the radiometric response function of the camera to be known.
The accuracy such algorithms are able to achieve depends a lot on the quality of an a priori calibration of these parameters.
We propose two techniques to estimate the position of a point light source, experimentally compare their performance with the commonly employed method, and draw conclusions which one to use in practice.
We also discuss how well an absolute radiometric calibration can be performed on uncontrolled consumer images and show the application of a simple radiometric model to re-create night-time impressions from color images.
A focus of this thesis is on Internet images which are an increasingly important source of data for computer vision and graphics applications.
Concerning reconstructions in this setting we present novel approaches that are able to recover surface orientation from Internet webcam images.
We explore two different strategies to overcome the challenges posed by this kind of input data.
One technique exploits orientation consistency and matches appearance profiles on the target with a partial reconstruction of the scene.
This avoids an explicit light calibration and works for any reflectance that is observed on the partial reference geometry.
The other technique employs an outdoor lighting model and reflectance properties represented as parametric basis materials.
It yields a richer scene representation consisting of shape and reflectance.
This is very useful for the simulation of new impressions or editing operations, e.g. relighting.
The proposed approach is the first that achieves such a reconstruction on webcam data.
Both presentations are accompanied by evaluations on synthetic and real-world data showing qualitative and quantitative results.
We also present a reconstruction approach for more controlled data in terms of the target scene.
It relies on a reference object to relax a constraint common to many photometric stereo approaches: the fixed camera assumption.
The proposed technique allows the camera and light source to vary freely in each image.
It again avoids a light calibration step and can be applied to non-Lambertian surfaces.
In summary, this thesis contributes to the calibration and to the reconstruction aspects of photometric techniques.
We overcome challenges in both controlled and uncontrolled settings, with a focus on the latter.
All proposed approaches are shown to operate also on non-Lambertian objects