64 research outputs found

    Enhancing the error detection capabilities of the standard video decoder using pixel domain dissimilarity metrics

    Get PDF
    The video compression standards commonly adopted in wireless multimedia services utilize variable length codes (VLC) in order to attain high compression ratios. While providing the high data rates required, this technique makes the system more susceptible to transmission errors. Thus the end-to-end quality of the video stream transmitted over an error-prone channel depends on the detection, and concealment of the corrupted macroblocks. The error detection capability of standard decoders is quite limited, for example, in the case of the H.263+ codec around 40.54% of the corrupted macroblocks are undetected, placing a bound on the perceived quality of the reconstructed video sequence. This paper presents a novel solution using eight pixel domain dissimilarity metrics computed in the CIE LUV color space which can be used at decode time to improve the error detection rate of the standard decoder. The spatial dissimilarity metric has been found to perform the best with an average increase in error detection rate of 60.38% when compared to the standard decoder (about 20% more than other published results) with 0% of false detection and a gain in peak signal-to-noise ratio (PSNR) of 3.94 dB.peer-reviewe

    Enhancing the error detection capabilities of DCT based codecs using compressed domain dissimilarity metrics

    Get PDF
    Video compression standards are implemented in wireless data transmission technologies to provide multimedia services efficiently. These compression standards generally utilize the Discrete Cosine Transform (DCT) in conjunction with variable length codes (VLC) in order to achieve the required high compression ratios. While providing the necessary high data rates, this technique has the disadvantage of making the system more susceptible to transmission errors. The standard decoders do not manage to detect a large number of corrupted macroblocks, 40.54% not detected for H.263+, contributing to a significant reduction in the end-to-end video quality as perceived by the end-user. This paper presents three dissimilarity metrics which contain both color and texture information and that can be extracted directly from the compressed DCT coefficients. These metrics can be used to enhance the error-detection capabilities of standard DCT based codecs. Simulation results show that the proposed algorithm increases the error detection rate by 54.06% with a gain in peak signal-to-noise ratio (PSNR) of 3.21 dB. This improvement in performance is superior to other solutions found in literature.peer-reviewe

    Enhancing error resilience in wireless transmitted compressed video sequences through a probabilistic neural network core

    Get PDF
    Video compression standards commonly employed in the delivery of real-time wireless multimedia services regularly adopt variable length codes (VLCs) for efficient transmission. This coding technique achieves the necessary high compression ratios at the expense of an increased system’s vulnerability to transmission errors. The more frequent presence of transmission errors in wireless channels requires video compression standards to accurately detect, localize and conceal any corrupted macroblocks (MBs) present in the video sequence. Unfortunately, standard decoders offer limited error detection and localization capabilities posing a bound on the perceived video quality of the reconstructed video sequence. This paper presents a novel solution which enhances the error detection and localization capabilities of standard decoders through the application of a Probabilistic Neural Network (PNN). The proposed solution generally outperforms other error detection mechanisms present in literature, as it manages to improve the standard decoder’s error detection rate by up to 95.74%. Index Terms — Error detection coding, learning systems, multimedia communications, video coding, wireless networks.peer-reviewe

    Resilient transmission of H.264/AVC video sequences using probabilistic neural networks

    Get PDF
    H.264/AVC is expected to become an essential component in the delivery of wireless multimedia content. While achieving high compression ratios, this codec is extremely vulnerable to transmission errors. These errors generally result in spatio-temporal propagation of distorted macroblocks (MBs) which significantly degrade the perceptual quality of the reconstructed video sequences. This paper presents a scheme for resilient transmission of H.264/AVC streams in noisy environments. The proposed algorithm exploits the redundant information which is inherent in the neighboring MBs and applies a Probabilistic Neural Network (PNN) classifier to detect visually impaired MBs. This algorithm achieves Peak Signal-to-Noise Ratio (PSNR) gains of up to 14.29 dB when compared to the standard decoder. Moreover, this significant gain in quality is achieved with minimal overheads and no additional bandwidth requirement, thus making it suitable for conversational and multicast/ broadcast services where feedback-based transport protocols cannot be applied.peer-reviewe

    A support vector machine approach for detection and localization of transmission errors within standard H.263++ decoders

    Get PDF
    Wireless multimedia services are increasingly becoming popular boosting the need for better quality-of-experience (QoE) with minimal costs. The standard codecs employed by these systems remove spatio-temporal redundancies to minimize the bandwidth required. However, this increases the exposure of the system to transmission errors, thus presenting a significant degradation in perceptual quality of the reconstructed video sequences. A number of mechanisms were investigated in the past to make these codecs more robust against transmission errors. Nevertheless, these techniques achieved little success, forcing the transmission to be held at lower bit-error rates (BERs) to guarantee acceptable quality. This paper presents a novel solution to this problem based on the error detection capabilities of the transport protocols to identify potentially corrupted group-of-blocks (GOBs). The algorithm uses a support vector machine (SVM) at its core to localize visually impaired macroblocks (MBs) that require concealment within these GOBs. Hence, this method drastically reduces the region to be concealed compared to state-of-the-art error resilient strategies which assume a packet loss scenario. Testing on a standard H.263++ codec confirms that a significant gain in quality is achieved with error detection rates of 97.8% and peak signal-to-noise ratio (PSNR) gains of up to 5.33 dB. Moreover, most of the undetected errors provide minimal visual artifacts and are thus of little influence to the perceived quality of the reconstructed sequences.peer-reviewe

    Robust error detection methods for H.264/AVC videos

    Get PDF
    The 3rd generation of mobile systems is mainly focused on enabling multimedia services such as video streaming, video call and conferencing. In order to achieve this, the Universal Mobile Telecommunications System (UMTS), is the standard that has been developed by the 3rd Generation Partnership ect (3GPP) in Europe, including the baseline profile of H.264/AVC in the specification. With the union of both technologies a great improvement on video transmission over mobile networks, and even modification of the user habits towards the use of the mobile phone is expected. Nevertheless, video transmission has always been related to wired networks and unfortunately the migration to wireless networks is not as easy as it seems. In real time applications the delay is a critical constraint. Usually, transmission protocols without delivery warranties, like the User Network Protocol (UDP) for IP based networks, are used. This works under the assumption that in real time applications dropped packets are preferable to delayed packets. Moreover, in UMTS the network needs to be treated in a different way, thus the wireless channel is a prone error channel due to its high time variance. Typically, when transmitting video, the receiver checks whether the information packet is corrupted (by means of a checksum) or if its temporal mark exceeds the specified delay. This approach is suboptimal, due to the fact that perhaps the video information is not damaged and could still be used. Instead, residual redundancy on the video stream can be used to locate the errors in the corrupted packet, increasing the granularity of the typical upper-layer checksum error detection. Based on this, the amount of information previous to the error detection can be decoded as usually. The aim of this thesis is to combine some of the more effective methods concretely, Syntax check, Watermarking and Checksum schemes have been reformulated, combined and simulated

    Indexing, browsing and searching of digital video

    Get PDF
    Video is a communications medium that normally brings together moving pictures with a synchronised audio track into a discrete piece or pieces of information. The size of a “piece ” of video can variously be referred to as a frame, a shot, a scene, a clip, a programme or an episode, and these are distinguished by their lengths and by their composition. We shall return to the definition of each of these in section 4 this chapter. In modern society, video is ver

    Robust decoder-based error control strategy for recovery of H.264/AVC video content

    Get PDF
    Real-time wireless conversational and broadcasting multimedia applications offer particular transmission challenges as reliable content delivery cannot be guaranteed. The undelivered and erroneous content causes significant degradation in quality of experience. The H.264/AVC standard includes several error resilient tools to mitigate this effect on video quality. However, the methods implemented by the standard are based on a packet-loss scenario, where corrupted slices are dropped and the lost information concealed. Partially damaged slices still contain valuable information that can be used to enhance the quality of the recovered video. This study presents a novel error recovery solution that relies on a joint source-channel decoder to recover only feasible slices. A major advantage of this decoder-based strategy is that it grants additional robustness while keeping the same transmission data rate. Simulation results show that the proposed approach manages to completely recover 30.79% of the corrupted slices. This provides frame-by-frame peak signal-to-noise ratio (PSNR) gains of up to 18.1%dB, a result which, to the knowledge of the authors, is superior to all other joint source-channel decoding methods found in literature. Furthermore, this error resilient strategy can be combined with other error resilient tools adopted by the standard to enhance their performance.peer-reviewe

    Improved quality block-based low bit rate video coding.

    Get PDF
    The aim of this research is to develop algorithms for enhancing the subjective quality and coding efficiency of standard block-based video coders. In the past few years, numerous video coding standards based on motion-compensated block-transform structure have been established where block-based motion estimation is used for reducing the correlation between consecutive images and block transform is used for coding the resulting motion-compensated residual images. Due to the use of predictive differential coding and variable length coding techniques, the output data rate exhibits extreme fluctuations. A rate control algorithm is devised for achieving a stable output data rate. This rate control algorithm, which is essentially a bit-rate estimation algorithm, is then employed in a bit-allocation algorithm for improving the visual quality of the coded images, based on some prior knowledge of the images. Block-based hybrid coders achieve high compression ratio mainly due to the employment of a motion estimation and compensation stage in the coding process. The conventional bit-allocation strategy for these coders simply assigns the bits required by the motion vectors and the rest to the residual image. However, at very low bit-rates, this bit-allocation strategy is inadequate as the motion vector bits takes up a considerable portion of the total bit-rate. A rate-constrained selection algorithm is presented where an analysis-by-synthesis approach is used for choosing the best motion vectors in term of resulting bit rate and image quality. This selection algorithm is then implemented for mode selection. A simple algorithm based on the above-mentioned bit-rate estimation algorithm is developed for the latter to reduce the computational complexity. For very low bit-rate applications, it is well-known that block-based coders suffer from blocking artifacts. A coding mode is presented for reducing these annoying artifacts by coding a down-sampled version of the residual image with a smaller quantisation step size. Its applications for adaptive source/channel coding and for coding fast changing sequences are examined

    Digital rights management techniques for H.264 video

    Get PDF
    This work aims to present a number of low-complexity digital rights management (DRM) methodologies for the H.264 standard. Initially, requirements to enforce DRM are analyzed and understood. Based on these requirements, a framework is constructed which puts forth different possibilities that can be explored to satisfy the objective. To implement computationally efficient DRM methods, watermarking and content based copy detection are then chosen as the preferred methodologies. The first approach is based on robust watermarking which modifies the DC residuals of 4×4 macroblocks within I-frames. Robust watermarks are appropriate for content protection and proving ownership. Experimental results show that the technique exhibits encouraging rate-distortion (R-D) characteristics while at the same time being computationally efficient. The problem of content authentication is addressed with the help of two methodologies: irreversible and reversible watermarks. The first approach utilizes the highest frequency coefficient within 4×4 blocks of the I-frames after CAVLC en- tropy encoding to embed a watermark. The technique was found to be very effect- ive in detecting tampering. The second approach applies the difference expansion (DE) method on IPCM macroblocks within P-frames to embed a high-capacity reversible watermark. Experiments prove the technique to be not only fragile and reversible but also exhibiting minimal variation in its R-D characteristics. The final methodology adopted to enforce DRM for H.264 video is based on the concept of signature generation and matching. Specific types of macroblocks within each predefined region of an I-, B- and P-frame are counted at regular intervals in a video clip and an ordinal matrix is constructed based on their count. The matrix is considered to be the signature of that video clip and is matched with longer video sequences to detect copies within them. Simulation results show that the matching methodology is capable of not only detecting copies but also its location within a longer video sequence. Performance analysis depict acceptable false positive and false negative rates and encouraging receiver operating charac- teristics. Finally, the time taken to match and locate copies is significantly low which makes it ideal for use in broadcast and streaming applications
    • 

    corecore