23 research outputs found
Recommended from our members
Virtual viewpoint three-dimensional panorama
Conventional panoramic images are known to provide for an enhanced field of view in which the scene
always has a fixed appearance. The idea presented in this paper focuses on the use of the concept of virtual
viewpoint creation to generate different panoramic images of the same scene with three-dimensional
component. Three-dimensional effect in a resultant panorama is realized by superimposing a stereo-pair of
panoramic images
Livrable D4.2 of the PERSEE project : Représentation et codage 3D - Rapport intermédiaire - Définitions des softs et architecture
51Livrable D4.2 du projet ANR PERSEECe rapport a été réalisé dans le cadre du projet ANR PERSEE (n° ANR-09-BLAN-0170). Exactement il correspond au livrable D4.2 du projet. Son titre : Représentation et codage 3D - Rapport intermédiaire - Définitions des softs et architectur
Recommended from our members
Camera positioning for 3D panoramic image rendering
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University London.Virtual camera realisation and the proposition of trapezoidal camera architecture are the two broad contributions of this thesis. Firstly, multiple camera and their arrangement constitute a critical component which affect the integrity of visual content acquisition for multi-view video. Currently, linear, convergence, and divergence arrays are the prominent camera topologies adopted. However, the large number of cameras required and their synchronisation are two of prominent challenges usually encountered. The use of virtual cameras can significantly reduce the number of physical cameras used with respect to any of the known
camera structures, hence adequately reducing some of the other implementation issues. This thesis explores to use image-based rendering with and without geometry in the implementations leading to the realisation of virtual cameras. The virtual camera implementation was carried out from the perspective of depth map (geometry) and use of multiple image samples (no geometry). Prior to the virtual camera realisation, the generation of depth map was investigated using region match measures widely known for solving image point correspondence problem. The constructed depth maps have been compare with the ones generated
using the dynamic programming approach. In both the geometry and no geometry approaches, the virtual cameras lead to the rendering of views from a textured depth map, construction of 3D panoramic image of a scene by stitching multiple image samples and performing superposition on them, and computation
of virtual scene from a stereo pair of panoramic images. The quality of these rendered images were assessed through the use of either objective or subjective analysis in Imatest software. Further more, metric reconstruction of a scene was performed by re-projection of the pixel points from multiple image samples with
a single centre of projection. This was done using sparse bundle adjustment algorithm. The statistical summary obtained after the application of this algorithm provides a gauge for the efficiency of the optimisation step. The optimised data was then visualised in Meshlab software environment, hence providing the reconstructed scene. Secondly, with any of the well-established camera arrangements, all cameras are usually constrained to the same horizontal plane. Therefore, occlusion becomes an extremely challenging problem, and a robust camera set-up is required in order to resolve strongly the hidden part of any scene objects.
To adequately meet the visibility condition for scene objects and given that occlusion of the same scene objects can occur, a multi-plane camera structure is highly desirable. Therefore, this thesis also explore trapezoidal camera structure for image acquisition. The approach here is to assess the feasibility and potential
of several physical cameras of the same model being sparsely arranged on the edge of an efficient trapezoid graph. This is implemented both Matlab and Maya. The quality of the depth maps rendered in Matlab are better in Quality
Recommended from our members
A Novel Inpainting Framework for Virtual View Synthesis
Multi-view imaging has stimulated significant research to enhance the user experience of free viewpoint video, allowing interactive navigation between views and the freedom to select a desired view to watch. This usually involves transmitting both textural and depth information captured from different viewpoints to the receiver, to enable the synthesis of an arbitrary view. In rendering these virtual views, perceptual holes can appear due to certain regions, hidden in the original view by a closer object, becoming visible in the virtual view. To provide a high quality experience these holes must be filled in a visually plausible way, in a process known as inpainting. This is challenging because the missing information is generally unknown and the hole-regions can be large. Recently depth-based inpainting techniques have been proposed to address this challenge and while these generally perform better than non-depth assisted methods, they are not very robust and can produce perceptual artefacts.
This thesis presents a new inpainting framework that innovatively exploits depth and textural self-similarity characteristics to construct subjectively enhanced virtual viewpoints. The framework makes three significant contributions to the field: i) the exploitation of view information to jointly inpaint textural and depth hole regions; ii) the introduction of the novel concept of self-similarity characterisation which is combined with relevant depth information; and iii) an advanced self-similarity characterising scheme that automatically determines key spatial transform parameters for effective and flexible inpainting.
The presented inpainting framework has been critically analysed and shown to provide superior performance both perceptually and numerically compared to existing techniques, especially in terms of lower visual artefacts. It provides a flexible robust framework to develop new inpainting strategies for the next generation of interactive multi-view technologies
Robust density modelling using the student's t-distribution for human action recognition
The extraction of human features from videos is often inaccurate and prone to outliers. Such outliers can severely affect density modelling when the Gaussian distribution is used as the model since it is highly sensitive to outliers. The Gaussian distribution is also often used as base component of graphical models for recognising human actions in the videos (hidden Markov model and others) and the presence of outliers can significantly affect the recognition accuracy. In contrast, the Student's t-distribution is more robust to outliers and can be exploited to improve the recognition rate in the presence of abnormal data. In this paper, we present an HMM which uses mixtures of t-distributions as observation probabilities and show how experiments over two well-known datasets (Weizmann, MuHAVi) reported a remarkable improvement in classification accuracy. © 2011 IEEE
Analysis of Disparity Maps for Detecting Saliency in Stereoscopic Video
We present a system for automatically detecting salient image regions in stereoscopic videos. This report extends our previous system and provides additional details about its implementation. Our proposed algorithm considers information based on three dimensions: salient colors in individual frames, salient information derived from camera and object motion, and depth saliency. These three components are dynamically combined into one final saliency map based on the reliability of the individual saliency detectors. Such a combination allows using more efficient algorithms even if the quality of one detector degrades. For example, we use a computationally efficient stereo correspondence algorithm that might cause noisy disparity maps for certain scenarios. In this case, however, a more reliable saliency detection algorithm such as the image saliency is preferred. To evaluate the quality of the saliency detection, we created modified versions of stereoscopic videos with the non-salient regions blurred. Having users rate the quality of these videos, the results show that most users do not detect the blurred regions and that the automatic saliency detection is very reliable
Electronic Imaging & the Visual Arts. EVA 2018 Florence
The Publication is following the yearly Editions of EVA FLORENCE. The State of Art is presented regarding the Application of Technologies (in particular of digital type) to Cultural Heritage. The more recent results of the Researches in the considered Area are presented. Information Technologies of interest for Culture Heritage are presented: multimedia systems, data-bases, data protection, access to digital content, Virtual Galleries. Particular reference is reserved to digital images (Electronic Imaging & the Visual Arts), regarding Cultural Institutions (Museums, Libraries, Palace - Monuments, Archaeological Sites). The International Conference includes the following Sessions: Strategic Issues; New Sciences and Culture Developments and Applications; New Technical Developments & Applications; Museums - Virtual Galleries and Related Initiatives; Art and Humanities Ecosystem & Applications; Access to the Culture Information. Two Workshops regard: Innovation and Enterprise; the Cloud Systems connected to the Culture (eCulture Cloud) in the Smart Cities context. The more recent results of the Researches at national and international are reported in the Area of Technologies and Culture Heritage, also with experimental demonstrations of developed Activities
Multi-scale metric for objective synthesized image quality assessment for FTV
Основни допринос ове докторске дисертације је развој алгоритама за објективну процену визуелног квалитета слике синтетизоване применом ДИБР (Depth Image Based Rendering) техника које узрокују неуниформна изобличења у области ивица. Применом нелинеарних морфолошких филтара у мултирезолуционој декомпозицији слика код израчунавања предложене метрике, важне геометријске информације као што су ивице су добро очуване без помака и замућења у сликама на различитим скалама мултирезолуционе репрезентације. Израчунавањем МСЕ по подопсезима који садрже ивице, пиксел по пиксел, прецизно се мери разлика две мултирезолуционе репрезентације. Тако се највећи значај у процени квалитета додељује области ивица. Процене предложене метрике се добро поклапају са субјективним оценама.Osnovni doprinos ove doktorske disertacije je razvoj algoritama za objektivnu procenu vizuelnog kvaliteta slike sintetizovane primenom DIBR (Depth Image Based Rendering) tehnika koje uzrokuju neuniformna izobličenja u oblasti ivica. Primenom nelinearnih morfoloških filtara u multirezolucionoj dekompoziciji slika kod izračunavanja predložene metrike, važne geometrijske informacije kao što su ivice su dobro očuvane bez pomaka i zamućenja u slikama na različitim skalama multirezolucione reprezentacije. Izračunavanjem MSE po podopsezima koji sadrže ivice, piksel po piksel, precizno se meri razlika dve multirezolucione reprezentacije. Tako se najveći značaj u proceni kvaliteta dodeljuje oblasti ivica. Procene predložene metrike se dobro poklapaju sa subjektivnim ocenama.The main contribution of this doctoral thesis is the development of algorithms for objective DIBR-synthesized view quality assessment. DIBR algorithms introduce nonuniform geometric distortions affecting the edge coherency in the synthesized images.The non-linear morphological filters used in multi-scale image decompositions of the proposed metric maintain important geometric information such as edges across different resolution levels.Calculating MSE pixel-by-pixel through subbands in which the edges are extracted, the difference of the two multiresolution representations, the reference and the synthesized image, is precisely measured. In that way the importance of edge areas which are prone to synthesis artifacts is emphasized in the image quality assessment. The proposed metric has very good agreement with human judgment
Visual attention models and applications to 3D computer graphics
Ankara : The Department of Computer Engineering and the Graduate School of Engineering and Science of Bilkent University, 2012.Thesis (Ph. D.) -- Bilkent University, 2012.Includes bibliographical refences.3D computer graphics, with the increasing technological and computational
opportunities, have advanced to very high levels that it is possible to generate very
realistic computer-generated scenes in real-time for games and other interactive
environments. However, we cannot claim that computer graphics research has
reached to its limits. Rendering photo-realistic scenes still cannot be achieved in
real-time; and improving visual quality and decreasing computational costs are
still research areas of great interest.
Recent e orts in computer graphics have been directed towards exploiting
principles of human visual perception to increase visual quality of rendering.
This is natural since in computer graphics, the main source of evaluation is the
judgment of people, which is based on their perception. In this thesis, our aim is
to extend the use of perceptual principles in computer graphics. Our contribution
is two-fold: First, we present several models to determine the visually important,
salient, regions in a 3D scene. Secondly, we contribute to use of de nition of
saliency metrics in computer graphics.
Human visual attention is composed of two components, the rst component
is the stimuli-oriented, bottom-up, visual attention; and the second component
is task-oriented, top-down visual attention. The main di erence between these
components is the role of the user. In the top-down component, viewer's intention
and task a ect perception of the visual scene as opposed to the bottom-up component.
We mostly investigate the bottom-up component where saliency resides.
We de ne saliency computation metrics for two types of graphical contents.
Our rst metric is applicable to 3D mesh models that are possibly animating, and
it extracts saliency values for each vertex of the mesh models. The second metric we propose is applicable to animating objects and nds visually important objects
due to their motion behaviours. In a third model, we present how to adapt the
second metric for the animated 3D meshes.
Along with the metrics of saliency, we also present possible application areas
and a perceptual method to accelerate stereoscopic rendering, which is based on
binocular vision principles and makes use of saliency information in a stereoscopic
rendering scene.
Each of the proposed models are evaluated with formal experiments. The
proposed saliency metrics are evaluated via eye-tracker based experiments and
the computationally salient regions are found to attract more attention in practice
too. For the stereoscopic optimization part, we have performed a detailed
experiment and veri ed our model of optimization.
In conclusion, this thesis extends the use of human visual system principles
in 3D computer graphics, especially in terms of saliency.Bülbül, Muhammed AbdullahPh.D