182 research outputs found
A fully scalable wavelet video coding scheme with homologous inter-scale prediction
In this paper, we present a fully scalable wavelet-based video coding architecture called STP-Tool, in which motion-compensated temporal-filtered subbands of spatially scaled versions of a video sequence can be used as a base layer for inter-scale predictions. These predictions take place in a pyramidal closed-loop structure between homologous resolution data, i.e., without the need of spatial interpolation. The presented implementation of the STP-Tool architecture is based on the reference software of the Wavelet Video Coding MPEG Ad-Hoc Group. The STP-Tool architecture makes it possible to compensate for some of the typical drawbacks of current wavelet-based scalable video coding architectures and shows interesting objective and visual results even when compared with other wavelet-based or MPEG-4 AVC/H.264-based scalable video coding systems
Generic techniques to reduce SVC enhancement layer encoding complexity
Scalable video coding is an important mechanism to provide several types of end-user devices with different versions of the same encoded bitstream. However, scalable video encoding remains a computationally expensive operation. To decrease the complexity we propose generic techniques. These techniques are generic in a sense that they can be combined with existing fast mode decision methods and optimizations. We show that extending such an existing fast mode decision technique yields an average complexity reduction of 87.27%, while only an additional 0.74% of bit rate increase and a decrease of 0.11dB in PSNR is required, compared to the original fast mode decision method(1)
No tradeoff between confidentiality and performance: An analysis on H.264/SVC partial encryption
Agency for Science, Technology and Research (A*STAR
Transparent encryption with scalable video communication: Lower-latency, CABAC-based schemes
Selective encryption masks all of the content without completely hiding it, as full encryption would do at a cost in encryption delay and increased bandwidth. Many commercial applications of video encryption do not even require selective encryption, because greater utility can be gained from transparent encryption, i.e. allowing prospective viewers to glimpse a reduced quality version of the content as a taster. Our lightweight selective encryption scheme when applied to scalable video coding is well suited to transparent encryption. The paper illustrates the gains in reducing delay and increased distortion arising from a transparent encryption that leaves reduced quality base layer in the clear. Reduced encryption of B-frames is a further step beyond transparent encryption in which the computational overhead reduction is traded against content security and limited distortion. This spectrum of video encryption possibilities is analyzed in this paper, though all of the schemes maintain decoder compatibility and add no bitrate overhead as a result of jointly encoding and encrypting the input video by virtue of carefully selecting the entropy coding parameters that are encrypted. The schemes are suitable both for H.264 and HEVC codecs, though demonstrated in the paper for H.264. Selected Content Adaptive Binary Arithmetic Coding (CABAC) parameters are encrypted by a lightweight Exclusive OR technique, which is chosen for practicality
Error resilient H.264 coded video transmission over wireless channels
The H.264/AVC recommendation was first published in 2003 and builds on the concepts of earlier standards such as MPEG-2 and MPEG-4. The H.264 recommendation represents an evolution of the existing video coding standards and
was developed in response to the growing need for higher compression. Even though H.264 provides for greater compression, H.264 compressed video streams are very
prone to channel errors in mobile wireless fading channels such as 3G due to high error rates experienced.
Common video compression techniques include motion compensation, prediction methods, transformation, quantization and entropy coding, which are the common
elements of a hybrid video codecs. The ITU-T recommendation H.264 introduces several new error resilience tools, as well as several new features such as Intra Prediction and Deblocking Filter.
The channel model used for the testing was the Rayleigh Fading channel with the noise component simulated as Additive White Gaussian Noise (AWGN) using QPSK as the modulation technique. The channel was used over several Eb/N0 values to provide similar bit error rates as those found in the literature.
Though further research needs to be conducted, results have shown that when using the H.264 error resilience tools in protecting encoded bitstreams to minor channel errors improvement in the decoded video quality can be observed. The tools did not perform as well with mild and severe channel errors significant as the resultant bitstream was too corrupted. From this, further research in channel coding techniques is needed to determine if the bitstream can be protected from these sorts of error rate
GRACE: Loss-Resilient Real-Time Video through Neural Codecs
In real-time video communication, retransmitting lost packets over
high-latency networks is not viable due to strict latency requirements. To
counter packet losses without retransmission, two primary strategies are
employed -- encoder-based forward error correction (FEC) and decoder-based
error concealment. The former encodes data with redundancy before transmission,
yet determining the optimal redundancy level in advance proves challenging. The
latter reconstructs video from partially received frames, but dividing a frame
into independently coded partitions inherently compromises compression
efficiency, and the lost information cannot be effectively recovered by the
decoder without adapting the encoder.
We present a loss-resilient real-time video system called GRACE, which
preserves the user's quality of experience (QoE) across a wide range of packet
losses through a new neural video codec. Central to GRACE's enhanced loss
resilience is its joint training of the neural encoder and decoder under a
spectrum of simulated packet losses. In lossless scenarios, GRACE achieves
video quality on par with conventional codecs (e.g., H.265). As the loss rate
escalates, GRACE exhibits a more graceful, less pronounced decline in quality,
consistently outperforming other loss-resilient schemes. Through extensive
evaluation on various videos and real network traces, we demonstrate that GRACE
reduces undecodable frames by 95% and stall duration by 90% compared with FEC,
while markedly boosting video quality over error concealment methods. In a user
study with 240 crowdsourced participants and 960 subjective ratings, GRACE
registers a 38% higher mean opinion score (MOS) than other baselines
Recommended from our members
3D multiple description coding for error resilience over wireless networks
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Mobile communications has gained a growing interest from both customers and service providers alike in the last 1-2 decades. Visual information is used in many application domains such as remote health care, video âon demand, broadcasting, video surveillance etc. In order to enhance the visual effects of digital video content, the depth perception needs to be provided with the actual visual content. 3D video has earned a significant interest from the research community in recent years, due to the tremendous impact it leaves on viewers and its enhancement of the userâs quality of experience (QoE). In the near future, 3D video is likely to be used in most video applications, as it offers a greater sense of immersion and perceptual experience. When 3D video is compressed and transmitted over error prone channels, the associated packet loss leads to visual quality degradation. When a picture is lost or corrupted so severely that the concealment result is not acceptable, the receiver typically pauses video playback and waits for the next INTRA picture to resume decoding. Error propagation caused by employing predictive coding may degrade the video quality severely. There are several ways used to mitigate the effects of such transmission errors. One widely used technique in International Video Coding Standards is error resilience.
The motivation behind this research work is that, existing schemes for 2D colour video compression such as MPEG, JPEG and H.263 cannot be applied to 3D video content. 3D video signals contain depth as well as colour information and are bandwidth demanding, as they require the transmission of multiple high-bandwidth 3D video streams. On the other hand, the capacity of wireless channels is limited and wireless links are prone to various types of errors caused by noise, interference, fading, handoff, error burst and network congestion. Given the maximum bit rate budget to represent the 3D scene, optimal bit-rate allocation between texture and depth information rendering distortion/losses should be minimised. To mitigate the effect of these errors on the perceptual 3D video quality, error resilience video coding needs to be investigated further to offer better quality of experience (QoE) to end users.
This research work aims at enhancing the error resilience capability of compressed 3D video, when transmitted over mobile channels, using Multiple Description Coding (MDC) in order to improve better userâs quality of experience (QoE).
Furthermore, this thesis examines the sensitivity of the human visual system (HVS) when employed to view 3D video scenes. The approach used in this study is to use subjective testing in order to rate peopleâs perception of 3D video under error free and error prone conditions through the use of a carefully designed bespoke questionnaire.Petroleum Technology Development Fund (PTDF
- âŠ