27 research outputs found

    Adaptive deblocking filter for transform domain Wyner-Ziv video coding

    Get PDF
    Wyner-Ziv (WZ) video coding is a particular case of distributed video coding, the recent video coding paradigm based on the Slepian-Wolf and Wyner-Ziv theorems that exploits the source correlation at the decoder and not at the encoder as in predictive video coding. Although many improvements have been done over the last years, the performance of the state-of-the-art WZ video codecs still did not reach the performance of state-of-the-art predictive video codecs, especially for high and complex motion video content. This is also true in terms of subjective image quality mainly because of a considerable amount of blocking artefacts present in the decoded WZ video frames. This paper proposes an adaptive deblocking filter to improve both the subjective and objective qualities of the WZ frames in a transform domain WZ video codec. The proposed filter is an adaptation of the advanced deblocking filter defined in the H.264/AVC (advanced video coding) standard to a WZ video codec. The results obtained confirm the subjective quality improvement and objective quality gains that can go up to 0.63 dB in the overall for sequences with high motion content when large group of pictures are used

    Temporal scalability comparison of the H.264/SVC and distributed video codec

    Get PDF

    THE VISNET II DVC CODEC: ARCHITECTURE, TOOLS AND PERFORMANCE

    Get PDF
    ABSTRACT This paper introduces the VISNET II DVC codec. This codec achieves very high RD performance thanks to the efficient combination of many state-of-the-art coding tools into INTRODUCTION With the wide deployment of mobile and wireless networks, a growing number of emerging applications, such as lowpower sensor networks, video surveillance cameras and mobile communications, rely on an up-link model rather than the typical down-link communication model. Typically, these applications are characterized by many senders transmitting data to a central receiver. In this context, light encoding or a flexible distribution of the codec complexity, robustness to packet losses, high compression efficiency and low latency/delay are important requirements. To address the needs of these up-link applications, the usual predictive video coding paradigm has been revisited based on Information Theory theorems from the 70s. The Slepian-Wolf (SW) theorem [1] establishes lower bounds on the achievable rates for the lossless coding of two or more correlated sources. More specifically, considering two statistically dependent random signals X and Y, it is well-known that the lower bound for the rate is given by the joint entropy H(X,Y) when these two signals are jointly encoded (as in conventional predictive coding). Conversely, when these two signals are independently encoded but jointly decoded (distributed coding), the SW theorem states that the minimum rate is still H(X,Y) with a residual error probability which tends towards 0 for long sequences. Later, Wyner and Ziv (WZ) have extended the SW theorem and showed that the result holds for the lossy coding case under the assumptions that the sources are jointly Gaussian and a mean square error distortion measure is used [2]. Subsequently, it was shown that this result remains valid as long as the difference between X and Y is Gaussian. Video coding schemes based on these theorems are referred to as Distributed Video Coding (DVC) solutions. Since the new coding paradigm is based on a statistical framework and does not rely on joint encoding, DVC architectures may provide several functional benefits which are rather important for many emerging applications: i) flexible allocation of the global video codec complexity; ii) improved error resilience; iii) codec independent scalability; and iv) exploitation of multiview correlation. Based on these theoretical results, practical implementations of DVC have been proposed since 2002. The PRISM (Power-efficient, Robust, hIgh compression Syndrome-based Multimedia coding) [3] solution works at the block level and performs motion estimation at the decoder. Based on the amount of temporal correlation, estimated using a zeromotion block difference, each block can either be conventionally (intra) coded, skipped or coded using distributed coding principles. Another DVC architecture working at frame level has been proposed in In this paper, the DVC codec developed within the European Network of Excellence VISNET II project [5] is described. This codec is based on the early architecture in VISNET II CODEC ARCHITECTURE AND TOOLS This section provides a description of the VISNET II DVC codec architecture and tools illustrated i

    Side information exploitation, quality control and low complexity implementation for distributed video coding

    Get PDF
    Distributed video coding (DVC) is a new video coding methodology that shifts the highly complex motion search components from the encoder to the decoder, such a video coder would have a great advantage in encoding speed and it is still able to achieve similar rate-distortion performance as the conventional coding solutions. Applications include wireless video sensor networks, mobile video cameras and wireless video surveillance, etc. Although many progresses have been made in DVC over the past ten years, there is still a gap in RD performance between conventional video coding solutions and DVC. The latest development of DVC is still far from standardization and practical use. The key problems remain in the areas such as accurate and efficient side information generation and refinement, quality control between Wyner-Ziv frames and key frames, correlation noise modelling and decoder complexity, etc. Under this context, this thesis proposes solutions to improve the state-of-the-art side information refinement schemes, enable consistent quality control over decoded frames during coding process and implement highly efficient DVC codec. This thesis investigates the impact of reference frames on side information generation and reveals that reference frames have the potential to be better side information than the extensively used interpolated frames. Based on this investigation, we also propose a motion range prediction (MRP) method to exploit reference frames and precisely guide the statistical motion learning process. Extensive simulation results show that choosing reference frames as SI performs competitively, and sometimes even better than interpolated frames. Furthermore, the proposed MRP method is shown to significantly reduce the decoding complexity without degrading any RD performance. To minimize the block artifacts and achieve consistent improvement in both subjective and objective quality of side information, we propose a novel side information synthesis framework working on pixel granularity. We synthesize the SI at pixel level to minimize the block artifacts and adaptively change the correlation noise model according to the new SI. Furthermore, we have fully implemented a state-of-the-art DVC decoder with the proposed framework using serial and parallel processing technologies to identify bottlenecks and areas to further reduce the decoding complexity, which is another major challenge for future practical DVC system deployments. The performance is evaluated based on the latest transform domain DVC codec and compared with different standard codecs. Extensive experimental results show substantial and consistent rate-distortion gains over standard video codecs and significant speedup over serial implementation. In order to bring the state-of-the-art DVC one step closer to practical use, we address the problem of distortion variation introduced by typical rate control algorithms, especially in a variable bit rate environment. Simulation results show that the proposed quality control algorithm is capable to meet user defined target distortion and maintain a rather small variation for sequence with slow motion and performs similar to fixed quantization for fast motion sequence at the cost of some RD performance. Finally, we propose the first implementation of a distributed video encoder on a Texas Instruments TMS320DM6437 digital signal processor. The WZ encoder is efficiently implemented, using rate adaptive low-density-parity-check accumulative (LDPCA) codes, exploiting the hardware features and optimization techniques to improve the overall performance. Implementation results show that the WZ encoder is able to encode at 134M instruction cycles per QCIF frame on a TMS320DM6437 DSP running at 700MHz. This results in encoder speed 29 times faster than non-optimized encoder implementation. We also implemented a highly efficient DVC decoder using both serial and parallel technology based on a PC-HPC (high performance cluster) architecture, where the encoder is running in a general purpose PC and the decoder is running in a multicore HPC. The experimental results show that the parallelized decoder can achieve about 10 times speedup under various bit-rates and GOP sizes compared to the serial implementation and significant RD gains with regards to the state-of-the-art DISCOVER codec

    Resource-Constrained Low-Complexity Video Coding for Wireless Transmission

    Get PDF

    Distributed video coding in wireless multimedia sensor network for multimedia broadcasting

    Get PDF
    Recently the development of Distributed Video Coding (DVC) has provided the promising theory support to realize the infrastructure of Wireless Multimedia Sensor Network (WMSN), which composed of autonomous hardware for capturing and transmission of quality audio-visual content. The implementation of DVC in WMSN can better solve the problem of energy constraint of the sensor nodes due to the benefit of lower computational encoder in DVC. In this paper, a practical DVC scheme, pixel-domain Wyner-Ziv(PDWZ) video coding, with slice structure and adaptive rate selection(ARS) is proposed to solve the certain problems when applying DVC into WMSN. Firstly, the proposed slice structure in PDWZ has extended the feasibility of PDWZ to work with any interleaver size used in Slepian-wolf turbo codec for heterogeneous applications. Meanwhile, based on the slice structure, an adaptive code rate selection has been proposed aiming at reduce the system delay occurred in feedback request. The simulation results clearly showed the enhancement in R-D performance and perceptual quality. It also can be observed that system delay caused by frequent feedback is greatly reduced, which gives a promising support for WMSN with low latency and facilitates the QoS management

    Achieving H.264-like compression efficiency with distributed video coding

    Full text link

    On the Effectiveness of Video Recolouring as an Uplink-model Video Coding Technique

    Get PDF
    For decades, conventional video compression formats have advanced via incremental improvements with each subsequent standard achieving better rate-distortion (RD) efficiency at the cost of increased encoder complexity compared to its predecessors. Design efforts have been driven by common multi-media use cases such as video-on-demand, teleconferencing, and video streaming, where the most important requirements are low bandwidth and low video playback latency. Meeting these requirements involves the use of computa- tionally expensive block-matching algorithms which produce excellent compression rates and quick decoding times. However, emerging use cases such as Wireless Video Sensor Networks, remote surveillance, and mobile video present new technical challenges in video compression. In these scenarios, the video capture and encoding devices are often power-constrained and have limited computational resources available, while the decoder devices have abundant resources and access to a dedicated power source. To address these use cases, codecs must be power-aware and offer a reasonable trade-off between video quality, bitrate, and encoder complexity. Balancing these constraints requires a complete rethinking of video compression technology. The uplink video-coding model represents a new paradigm to address these low-power use cases, providing the ability to redistribute computational complexity by offloading the motion estimation and compensation steps from encoder to decoder. Distributed Video Coding (DVC) follows this uplink model of video codec design, and maintains high quality video reconstruction through innovative channel coding techniques. The field of DVC is still early in its development, with many open problems waiting to be solved, and no defined video compression or distribution standards. Due to the experimental nature of the field, most DVC codec to date have focused on encoding and decoding the Luma plane only, which produce grayscale reconstructed videos. In this thesis, a technique called “video recolouring” is examined as an alternative to DVC. Video recolour- ing exploits the temporal redundancies between colour planes, reducing video bitrate by removing Chroma information from specific frames and then recolouring them at the decoder. A novel video recolouring algorithm called Motion-Compensated Recolouring (MCR) is proposed, which uses block motion estimation and bi-directional weighted motion-compensation to reconstruct Chroma planes at the decoder. MCR is used to enhance a conventional base-layer codec, and shown to reduce bitrate by up to 16% with only a slight decrease in objective quality. MCR also outperforms other video recolouring algorithms in terms of objective video quality, demonstrating up to 2 dB PSNR improvement in some cases

    Depth-based Multi-View 3D Video Coding

    Get PDF
    corecore