1,490 research outputs found
A Bayesian fusion model for space-time reconstruction of finely resolved velocities in turbulent flows from low resolution measurements
The study of turbulent flows calls for measurements with high resolution both
in space and in time. We propose a new approach to reconstruct
High-Temporal-High-Spatial resolution velocity fields by combining two sources
of information that are well-resolved either in space or in time, the
Low-Temporal-High-Spatial (LTHS) and the High-Temporal-Low-Spatial (HTLS)
resolution measurements. In the framework of co-conception between sensing and
data post-processing, this work extensively investigates a Bayesian
reconstruction approach using a simulated database. A Bayesian fusion model is
developed to solve the inverse problem of data reconstruction. The model uses a
Maximum A Posteriori estimate, which yields the most probable field knowing the
measurements. The DNS of a wall-bounded turbulent flow at moderate Reynolds
number is used to validate and assess the performances of the present approach.
Low resolution measurements are subsampled in time and space from the fully
resolved data. Reconstructed velocities are compared to the reference DNS to
estimate the reconstruction errors. The model is compared to other conventional
methods such as Linear Stochastic Estimation and cubic spline interpolation.
Results show the superior accuracy of the proposed method in all
configurations. Further investigations of model performances on various range
of scales demonstrate its robustness. Numerical experiments also permit to
estimate the expected maximum information level corresponding to limitations of
experimental instruments.Comment: 15 pages, 6 figure
Wavelet based stereo images reconstruction using depth images
It is believed by many that three-dimensional (3D) television will be the next logical development toward a more natural and vivid home entertaiment experience. While classical 3D approach requires the transmission of two video streams, one for each view, 3D TV systems based on depth image rendering (DIBR) require a single stream of monoscopic images and a second stream of associated images usually termed depth images or depth maps, that contain per-pixel depth information. Depth map is a two-dimensional function that contains information about distance from camera to a certain point of the object as a function of the image coordinates. By using this depth information and the original image it is possible to reconstruct a virtual image of a nearby viewpoint by projecting the pixels of available image to their locations in 3D space and finding their position in the desired view plane. One of the most significant advantages of the DIBR is that depth maps can be coded more efficiently than two streams corresponding to left and right view of the scene, thereby reducing the bandwidth required for transmission, which makes it possible to reuse existing transmission channels for the transmission of 3D TV. This technique can also be applied for other 3D technologies such as multimedia systems.
In this paper we propose an advanced wavelet domain scheme for the reconstruction of stereoscopic images, which solves some of the shortcommings of the existing methods discussed above. We perform the wavelet transform of both the luminance and depth images in order to obtain significant geometric features, which enable more sensible reconstruction of the virtual view. Motion estimation employed in our approach uses Markov random field smoothness prior for regularization of the estimated motion field.
The evaluation of the proposed reconstruction method is done on two video sequences which are typically used for comparison of stereo reconstruction algorithms. The results demonstrate advantages of the proposed approach with respect to the state-of-the-art methods, in terms of both objective and subjective performance measures
Stereoscopic visual saliency prediction based on stereo contrast and stereo focus
© 2017, The Author(s). In this paper, we exploit two characteristics of stereoscopic vision: the pop-out effect and the comfort zone. We propose a visual saliency prediction model for stereoscopic images based on stereo contrast and stereo focus models. The stereo contrast model measures stereo saliency based on the color/depth contrast and the pop-out effect. The stereo focus model describes the degree of focus based on monocular focus and the comfort zone. After obtaining the values of the stereo contrast and stereo focus models in parallel, an enhancement based on clustering is performed on both values. We then apply a multi-scale fusion to form the respective maps of the two models. Last, we use a Bayesian integration scheme to integrate the two maps (the stereo contrast and stereo focus maps) into the stereo saliency map. Experimental results on two eye-tracking databases show that our proposed method outperforms the state-of-the-art saliency models
Visualization and Correction of Automated Segmentation, Tracking and Lineaging from 5-D Stem Cell Image Sequences
Results: We present an application that enables the quantitative analysis of
multichannel 5-D (x, y, z, t, channel) and large montage confocal fluorescence
microscopy images. The image sequences show stem cells together with blood
vessels, enabling quantification of the dynamic behaviors of stem cells in
relation to their vascular niche, with applications in developmental and cancer
biology. Our application automatically segments, tracks, and lineages the image
sequence data and then allows the user to view and edit the results of
automated algorithms in a stereoscopic 3-D window while simultaneously viewing
the stem cell lineage tree in a 2-D window. Using the GPU to store and render
the image sequence data enables a hybrid computational approach. An
inference-based approach utilizing user-provided edits to automatically correct
related mistakes executes interactively on the system CPU while the GPU handles
3-D visualization tasks. Conclusions: By exploiting commodity computer gaming
hardware, we have developed an application that can be run in the laboratory to
facilitate rapid iteration through biological experiments. There is a pressing
need for visualization and analysis tools for 5-D live cell image data. We
combine accurate unsupervised processes with an intuitive visualization of the
results. Our validation interface allows for each data set to be corrected to
100% accuracy, ensuring that downstream data analysis is accurate and
verifiable. Our tool is the first to combine all of these aspects, leveraging
the synergies obtained by utilizing validation information from stereo
visualization to improve the low level image processing tasks.Comment: BioVis 2014 conferenc
Digital ocular fundus imaging: a review
Ocular fundus imaging plays a key role in monitoring the health status of the human eye. Currently, a large number of imaging modalities allow the assessment and/or quantification of ocular changes from a healthy status. This review focuses on the main digital fundus imaging modality, color fundus photography, with a brief overview of complementary techniques, such as fluorescein angiography. While focusing on two-dimensional color fundus photography, the authors address the evolution from nondigital to digital imaging and its impact on diagnosis. They also compare several studies performed along the transitional path of this technology. Retinal image processing and analysis, automated disease detection and identification of the stage of diabetic retinopathy (DR) are addressed as well. The authors emphasize the problems of image segmentation, focusing on the major landmark structures of the ocular fundus: the vascular network, optic disk and the fovea. Several proposed approaches for the automatic detection of signs of disease onset and progression, such as microaneurysms, are surveyed. A thorough comparison is conducted among different studies with regard to the number of eyes/subjects, imaging modality, fundus camera used, field of view and image resolution to identify the large variation in characteristics from one study to another. Similarly, the main features of the proposed classifications and algorithms for the automatic detection of DR are compared, thereby addressing computer-aided diagnosis and computer-aided detection for use in screening programs.Fundação para a Ciência e TecnologiaFEDErPrograma COMPET
Optical techniques for 3D surface reconstruction in computer-assisted laparoscopic surgery
One of the main challenges for computer-assisted surgery (CAS) is to determine the intra-opera- tive morphology and motion of soft-tissues. This information is prerequisite to the registration of multi-modal patient-specific data for enhancing the surgeon’s navigation capabilites by observ- ing beyond exposed tissue surfaces and for providing intelligent control of robotic-assisted in- struments. In minimally invasive surgery (MIS), optical techniques are an increasingly attractive approach for in vivo 3D reconstruction of the soft-tissue surface geometry. This paper reviews the state-of-the-art methods for optical intra-operative 3D reconstruction in laparoscopic surgery and discusses the technical challenges and future perspectives towards clinical translation. With the recent paradigm shift of surgical practice towards MIS and new developments in 3D opti- cal imaging, this is a timely discussion about technologies that could facilitate complex CAS procedures in dynamic and deformable anatomical regions
Study on Cosmic Ray Background Rejection with a 30 m Stand-Alone IACT using Non-parametric Multivariate Methods in a sub-100 GeV Energy Range
During the last decade ground-based very high-energy gamma-ray astronomy
achieved a remarkable advancement in the development of the observational
technique for the registration and study of gamma-ray emission above 100 GeV.
It is widely believed that the next step in its future development will be the
construction of telescopes of substantially larger size than the currently used
10 m class telescopes. This can drastically improve the sensitivity of the
ground-based detectors for gamma rays of energy from 10 to 100 GeV. Based on
Monte Carlo simulations of the response of a single stand-alone 30 m imaging
atmospheric Cherenkov telescope (IACT) the maximal rejection power against
background cosmic ray showers for low energy gamma-rays was investigated in
great detail. An advanced Bayesian multivariate analysis has been applied to
the simulated Cherenkov light images of the gamma-ray- and proton-induced air
showers. The results obtained here quantitatively testify that the separation
between the signal and background images degrades substantially at low
energies, and consequently the maximum overall quality factor can only be about
3.1 for gamma rays in the 10-30 GeV energy range. Various selection criteria as
well as optimal combinations of the standard image parameters utilized for
effective image separation have been also evaluated.Comment: Accepted for publication in the Journal of Physics
The assessment of visual behaviour and depth perception in surgery
Imperial Users onl
- …