3,180 research outputs found
The Modelling of Stereoscopic 3D Scene Acquisition
The main goal of this work is to find a suitable method for calculating the best setting of a stereo pair of cameras that are viewing the scene to enable spatial imaging. The method is based on a geometric model of a stereo pair cameras currently used for the acquisition of 3D scenes. Based on selectable camera parameters and object positions in the scene, the resultant model allows calculating the parameters of the stereo pair of images that influence the quality of spatial imaging. For the purpose of presenting the properties of the model of a simple 3D scene, an interactive application was created that allows, in addition to setting the cameras and scene parameters and displaying the calculated parameters, also displaying the modelled scene using perspective views and the stereo pair modelled with the aid of anaglyphic images. The resulting modelling method can be used in practice to determine appropriate parameters of the camera configuration based on the known arrangement of the objects in the scene. Analogously, it can, for a given camera configuration, determine appropriate geometrical limits of arranging the objects in the scene being displayed. This method ensures that the resulting stereoscopic recording will be of good quality and observer-friendly
Recommended from our members
Virtual viewpoint three-dimensional panorama
Conventional panoramic images are known to provide for an enhanced field of view in which the scene
always has a fixed appearance. The idea presented in this paper focuses on the use of the concept of virtual
viewpoint creation to generate different panoramic images of the same scene with three-dimensional
component. Three-dimensional effect in a resultant panorama is realized by superimposing a stereo-pair of
panoramic images
Visual Distortions in 360-degree Videos.
Omnidirectional (or 360°) images and videos are emergent signals being used in many areas, such as robotics and virtual/augmented reality. In particular, for virtual reality applications, they allow an immersive experience in which the user can interactively navigate through a scene with three degrees of freedom, wearing a head-mounted display. Current approaches for capturing, processing, delivering, and displaying 360° content, however, present many open technical challenges and introduce several types of distortions in the visual signal. Some of the distortions are specific to the nature of 360° images and often differ from those encountered in classical visual communication frameworks. This paper provides a first comprehensive review of the most common visual distortions that alter 360° signals going through the different processing elements of the visual communication pipeline. While their impact on viewers' visual perception and the immersive experience at large is still unknown-thus, it is an open research topic-this review serves the purpose of proposing a taxonomy of the visual distortions that can be encountered in 360° signals. Their underlying causes in the end-to-end 360° content distribution pipeline are identified. This taxonomy is essential as a basis for comparing different processing techniques, such as visual enhancement, encoding, and streaming strategies, and allowing the effective design of new algorithms and applications. It is also a useful resource for the design of psycho-visual studies aiming to characterize human perception of 360° content in interactive and immersive applications
Quality Assessment of Stereoscopic 360-degree Images from Multi-viewports
Objective quality assessment of stereoscopic panoramic images becomes a
challenging problem owing to the rapid growth of 360-degree contents. Different
from traditional 2D image quality assessment (IQA), more complex aspects are
involved in 3D omnidirectional IQA, especially unlimited field of view (FoV)
and extra depth perception, which brings difficulty to evaluate the quality of
experience (QoE) of 3D omnidirectional images. In this paper, we propose a
multi-viewport based fullreference stereo 360 IQA model. Due to the freely
changeable viewports when browsing in the head-mounted display (HMD), our
proposed approach processes the image inside FoV rather than the projected one
such as equirectangular projection (ERP). In addition, since overall QoE
depends on both image quality and depth perception, we utilize the features
estimated by the difference map between left and right views which can reflect
disparity. The depth perception features along with binocular image qualities
are employed to further predict the overall QoE of 3D 360 images. The
experimental results on our public Stereoscopic OmnidirectionaL Image quality
assessment Database (SOLID) show that the proposed method achieves a
significant improvement over some well-known IQA metrics and can accurately
reflect the overall QoE of perceived images
New visual coding exploration in MPEG: Super-MultiView and free navigation in free viewpoint TV
ISO/IEC MPEG and ITU-T VCEG have recently jointly issued
a new multiview video compression standard, called 3D-HEVC,
which reaches unpreceded compression performances for linear,dense camera arrangements. In view of supporting future highquality,auto-stereoscopic 3D displays and Free Navigation virtual/augmented reality applications with sparse, arbitrarily arranged camera setups, innovative depth estimation and virtual view synthesis techniques with global optimizations over all camera views should be developed. Preliminary studies in response to the MPEG-FTV (Free viewpoint TV) Call for Evidence suggest these
targets are within reach, with at least 6% bitrate gains over 3DHEVC
technology
D-SAV360: A Dataset of Gaze Scanpaths on 360° Ambisonic Videos
Understanding human visual behavior within virtual reality environments is crucial to fully leverage their potential. While previous research has provided rich visual data from human observers, existing gaze datasets often suffer from the absence of multimodal stimuli. Moreover, no dataset has yet gathered eye gaze trajectories (i.e., scanpaths) for dynamic content with directional ambisonic sound, which is a critical aspect of sound perception by humans. To address this gap, we introduce D-SAV360, a dataset of 4,609 head and eye scanpaths for 360° videos with first-order ambisonics. This dataset enables a more comprehensive study of multimodal interaction on visual behavior in virtual reality environments. We analyze our collected scanpaths from a total of 87 participants viewing 85 different videos and show that various factors such as viewing mode, content type, and gender significantly impact eye movement statistics. We demonstrate the potential of D-SAV360 as a benchmarking resource for state-of-the-art attention prediction models and discuss its possible applications in further research. By providing a comprehensive dataset of eye movement data for dynamic, multimodal virtual environments, our work can facilitate future investigations of visual behavior and attention in virtual reality
Use of functional near-infrared spectroscopy to evaluate cognitive change when using healthcare simulation tools
This is an accepted manuscript of an article published by BMJ on 01/11/2020, available online: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8936993/ The accepted version of the publication may differ from the final published version.Background The use of brain imaging techniques in healthcare simulation is relatively rare. However, the use of mobile, wireless technique, such as functional nearinfrared spectroscopy (fNIRS), is becoming a useful tool for assessing the unique demands of simulation learning. For this study, this imaging technique was used to evaluate cognitive load during simulation learning events.
Methods This study took place in relation to six simulation activities, paired for similarity, and evaluated
comparative cognitive change between the three task pairs. The three paired tasks were: receiving a (1) face-toface and (2) video patient handover; observing a simulated scene in (1) two dimensions and (2) 360° field of vision; and on a simulated patient (1) taking a pulse and (2) taking a pulse and respiratory rate simultaneously. The total number of participants was n=12.
Results In this study, fNIRS was sensitive to variations in task difficulty in common simulation tools and scenarios, showing an increase in oxygenated haemoglobin concentration and a decrease in deoxygenated haemoglobin concentration, as tasks increased in cognitive load.
Conclusion Overall, findings confirmed the usefulness of neurohaemoglobin concentration markers as an evaluation tool of cognitive change in healthcare simulation. Study findings suggested that cognitive load increases in more complex cognitive tasks in simulation learning events. Task performance that increased in complexity therefore affected cognitive markers, with increase in mental effort required
Stereoscopic Omnidirectional Image Quality Assessment Based on Predictive Coding Theory
Objective quality assessment of stereoscopic omnidirectional images is a
challenging problem since it is influenced by multiple aspects such as
projection deformation, field of view (FoV) range, binocular vision, visual
comfort, etc. Existing studies show that classic 2D or 3D image quality
assessment (IQA) metrics are not able to perform well for stereoscopic
omnidirectional images. However, very few research works have focused on
evaluating the perceptual visual quality of omnidirectional images, especially
for stereoscopic omnidirectional images. In this paper, based on the predictive
coding theory of the human vision system (HVS), we propose a stereoscopic
omnidirectional image quality evaluator (SOIQE) to cope with the
characteristics of 3D 360-degree images. Two modules are involved in SOIQE:
predictive coding theory based binocular rivalry module and multi-view fusion
module. In the binocular rivalry module, we introduce predictive coding theory
to simulate the competition between high-level patterns and calculate the
similarity and rivalry dominance to obtain the quality scores of viewport
images. Moreover, we develop the multi-view fusion module to aggregate the
quality scores of viewport images with the help of both content weight and
location weight. The proposed SOIQE is a parametric model without necessary of
regression learning, which ensures its interpretability and generalization
performance. Experimental results on our published stereoscopic omnidirectional
image quality assessment database (SOLID) demonstrate that our proposed SOIQE
method outperforms state-of-the-art metrics. Furthermore, we also verify the
effectiveness of each proposed module on both public stereoscopic image
datasets and panoramic image datasets
- …