8,008 research outputs found
Transmission of 3D Scenes over Lossy Channels
This paper introduces a novel error correction scheme for the transmission of three-dimensional scenes over unreliable networks. We propose a novel Unequal Error Protection scheme for the transmission of depth and texture information that distributes a prefixed amount of redundancy among the various elements of the scene description in order to maximize the quality of the rendered views. This target is achieved exploiting also a new model for the estimation of the impact on the rendered views of the various geometry and texture packets which takes into account their relevance in the coded bitstream and the viewpoint required by the user. Experimental results show how the proposed scheme effectively enhances the quality of the rendered images in a typical depth-image-based rendering scenario as packets are progressively decoded/recovered by the receiver
Deformable Shape Completion with Graph Convolutional Autoencoders
The availability of affordable and portable depth sensors has made scanning
objects and people simpler than ever. However, dealing with occlusions and
missing parts is still a significant challenge. The problem of reconstructing a
(possibly non-rigidly moving) 3D object from a single or multiple partial scans
has received increasing attention in recent years. In this work, we propose a
novel learning-based method for the completion of partial shapes. Unlike the
majority of existing approaches, our method focuses on objects that can undergo
non-rigid deformations. The core of our method is a variational autoencoder
with graph convolutional operations that learns a latent space for complete
realistic shapes. At inference, we optimize to find the representation in this
latent space that best fits the generated shape to the known partial input. The
completed shape exhibits a realistic appearance on the unknown part. We show
promising results towards the completion of synthetic and real scans of human
body and face meshes exhibiting different styles of articulation and
partiality.Comment: CVPR 201
Random Linear Network Coding for 5G Mobile Video Delivery
An exponential increase in mobile video delivery will continue with the
demand for higher resolution, multi-view and large-scale multicast video
services. Novel fifth generation (5G) 3GPP New Radio (NR) standard will bring a
number of new opportunities for optimizing video delivery across both 5G core
and radio access networks. One of the promising approaches for video quality
adaptation, throughput enhancement and erasure protection is the use of
packet-level random linear network coding (RLNC). In this review paper, we
discuss the integration of RLNC into the 5G NR standard, building upon the
ideas and opportunities identified in 4G LTE. We explicitly identify and
discuss in detail novel 5G NR features that provide support for RLNC-based
video delivery in 5G, thus pointing out to the promising avenues for future
research.Comment: Invited paper for Special Issue "Network and Rateless Coding for
Video Streaming" - MDPI Informatio
Graph Spectral Image Processing
Recent advent of graph signal processing (GSP) has spurred intensive studies
of signals that live naturally on irregular data kernels described by graphs
(e.g., social networks, wireless sensor networks). Though a digital image
contains pixels that reside on a regularly sampled 2D grid, if one can design
an appropriate underlying graph connecting pixels with weights that reflect the
image structure, then one can interpret the image (or image patch) as a signal
on a graph, and apply GSP tools for processing and analysis of the signal in
graph spectral domain. In this article, we overview recent graph spectral
techniques in GSP specifically for image / video processing. The topics covered
include image compression, image restoration, image filtering and image
segmentation
Recommended from our members
3D multiple description coding for error resilience over wireless networks
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Mobile communications has gained a growing interest from both customers and service providers alike in the last 1-2 decades. Visual information is used in many application domains such as remote health care, video –on demand, broadcasting, video surveillance etc. In order to enhance the visual effects of digital video content, the depth perception needs to be provided with the actual visual content. 3D video has earned a significant interest from the research community in recent years, due to the tremendous impact it leaves on viewers and its enhancement of the user’s quality of experience (QoE). In the near future, 3D video is likely to be used in most video applications, as it offers a greater sense of immersion and perceptual experience. When 3D video is compressed and transmitted over error prone channels, the associated packet loss leads to visual quality degradation. When a picture is lost or corrupted so severely that the concealment result is not acceptable, the receiver typically pauses video playback and waits for the next INTRA picture to resume decoding. Error propagation caused by employing predictive coding may degrade the video quality severely. There are several ways used to mitigate the effects of such transmission errors. One widely used technique in International Video Coding Standards is error resilience.
The motivation behind this research work is that, existing schemes for 2D colour video compression such as MPEG, JPEG and H.263 cannot be applied to 3D video content. 3D video signals contain depth as well as colour information and are bandwidth demanding, as they require the transmission of multiple high-bandwidth 3D video streams. On the other hand, the capacity of wireless channels is limited and wireless links are prone to various types of errors caused by noise, interference, fading, handoff, error burst and network congestion. Given the maximum bit rate budget to represent the 3D scene, optimal bit-rate allocation between texture and depth information rendering distortion/losses should be minimised. To mitigate the effect of these errors on the perceptual 3D video quality, error resilience video coding needs to be investigated further to offer better quality of experience (QoE) to end users.
This research work aims at enhancing the error resilience capability of compressed 3D video, when transmitted over mobile channels, using Multiple Description Coding (MDC) in order to improve better user’s quality of experience (QoE).
Furthermore, this thesis examines the sensitivity of the human visual system (HVS) when employed to view 3D video scenes. The approach used in this study is to use subjective testing in order to rate people’s perception of 3D video under error free and error prone conditions through the use of a carefully designed bespoke questionnaire.Petroleum Technology Development Fund (PTDF
DeshadowGAN: a deep learning approach to remove shadows from optical coherence tomography images
Purpose: To remove blood vessel shadows from optical coherence tomography (OCT) images of the optic nerve head (ONH).
Methods: Volume scans consisting of 97 horizontal B-scans were acquired through the center of the ONH using a commercial OCT device for both eyes of 13 subjects. A custom generative adversarial network (named DeshadowGAN) was designed and trained with 2328 B-scans in order to remove blood vessel shadows in unseen B-scans. Image quality was assessed qualitatively (for artifacts) and quantitatively using the intralayer contrast—a measure of shadow visibility ranging from 0 (shadow-free) to 1 (strong shadow). This was computed in the retinal nerve fiber layer (RNFL), the inner plexiform layer (IPL), the photoreceptor (PR) layer, and the retinal pigment epithelium (RPE) layer. The performance of DeshadowGAN was also compared with that of compensation, the standard for shadow removal.
Results: DeshadowGAN decreased the intralayer contrast in all tissue layers. On average, the intralayer contrast decreased by 33.7 ± 6.81%, 28.8 ± 10.4%, 35.9 ± 13.0%, and 43.0 ± 19.5% for the RNFL, IPL, PR layer, and RPE layer, respectively, indicating successful shadow removal across all depths. Output images were also free from artifacts commonly observed with compensation.
Conclusions: DeshadowGAN significantly corrected blood vessel shadows in OCT images of the ONH. Our algorithm may be considered as a preprocessing step to improve the performance of a wide range of algorithms including those currently being used for OCT segmentation, denoising, and classification.
Translational Relevance: DeshadowGAN could be integrated to existing OCT devices to improve the diagnosis and prognosis of ocular pathologies
Recommended from our members
LUNG LESION SEGMENTATION USING DEEP LEARNING APPROACHES
The amount of data generated in the medical imaging field, especially in a modern context, is growing significantly. As the amount of data grows, it\u27s prudent to make use of automated techniques that can leverage datasets to solve problems that are error-prone or have inconsistent solutions.
Deep learning approaches have gained traction in medical imaging tasks due to their superior performance with larger datasets and ability to discern the intricate features of 3D volumes, a task inefficient if done manually. Specifically for the task of lung nodule segmentation, several different methods have been tried before such as region growing etc. but this project focuses on using an Attention U-Net model to automatically segment the nodule boundaries. Specifically, this is done on the LUNA16 dataset as a benchmark which is a popular reference point for comparison. To achieve this, specifically, the Attention U-Net was trained with 5-fold cross-validation on the training dataset.
In addition to the segmentation outputs, averaged training and validation curves over all folds were also shown as the model is trained for 70 epochs. To conclude, these results present a useful automated method to segment the lung nodules. In practical situations, this would be of significant help to radiologists as it is less error-prone and not as susceptible to inter-observer variability. These automated tools along with other radiologist interactions could potentially significantly improve patient outcomes
- …