415,846 research outputs found
Multiple description video coding for stereoscopic 3D
In this paper, we propose an MDC schemes for stereoscopic 3D video. In the literature, MDC has previously been applied in 2D video but not so much in 3D video. The proposed algorithm enhances the error resilience of the 3D video using the combination of even and odd frame based MDC while retaining good temporal prediction efficiency for video over error-prone networks. Improvements are made to the original even and odd frame MDC scheme by adding a controllable amount of side information to improve frame interpolation at the decoder. The side information is also sent according to the video sequence motion for further improvement. The performance of the proposed algorithms is evaluated in error free and error prone environments especially for wireless channels. Simulation results show improved performance using the proposed MDC at high error rates compared to the single description coding (SDC) and the original even and odd frame MDC
Enabling error-resilient internet broadcasting using motion compensated spatial partitioning and packet FEC for the dirac video codec
Video transmission over the wireless or wired
network require protection from channel errors since compressed video bitstreams are very sensitive to transmission errors because of the use of predictive coding and variable length coding. In this paper, a simple, low complexity and patent free error-resilient coding is proposed. It is based upon the idea of using spatial partitioning on the motion compensated residual frame without employing the transform coefficient coding. The proposed scheme is intended for open source Dirac video codec in order to enable the codec to be used for Internet
broadcasting. By partitioning the wavelet transform coefficients of the motion compensated residual frame into groups and independently processing each group using arithmetic coding and Forward Error Correction (FEC), robustness to transmission errors over the packet erasure
wired network could be achieved. Using the Rate
Compatibles Punctured Code (RCPC) and Turbo Code
(TC) as the FEC, the proposed technique provides
gracefully decreasing perceptual quality over packet loss rates up to 30%. The PSNR performance is much better when compared with the conventional data partitioning only methods. Simulation results show that the use of multiple
partitioning of wavelet coefficient in Dirac can achieve up to 8 dB PSNR gain over its existing un-partitioned method
IBVC: Interpolation-driven B-frame Video Compression
Learned B-frame video compression aims to adopt bi-directional motion
estimation and motion compensation (MEMC) coding for middle frame
reconstruction. However, previous learned approaches often directly extend
neural P-frame codecs to B-frame relying on bi-directional optical-flow
estimation or video frame interpolation. They suffer from inaccurate quantized
motions and inefficient motion compensation. To address these issues, we
propose a simple yet effective structure called Interpolation-driven B-frame
Video Compression (IBVC). Our approach only involves two major operations:
video frame interpolation and artifact reduction compression. IBVC introduces a
bit-rate free MEMC based on interpolation, which avoids optical-flow
quantization and additional compression distortions. Later, to reduce duplicate
bit-rate consumption and focus on unaligned artifacts, a residual guided
masking encoder is deployed to adaptively select the meaningful contexts with
interpolated multi-scale dependencies. In addition, a conditional
spatio-temporal decoder is proposed to eliminate location errors and artifacts
instead of using MEMC coding in other methods. The experimental results on
B-frame coding demonstrate that IBVC has significant improvements compared to
the relevant state-of-the-art methods. Meanwhile, our approach can save bit
rates compared with the random access (RA) configuration of H.266 (VTM). The
code will be available at https://github.com/ruhig6/IBVC.Comment: Submitted to IEEE TCSV
360-degree Video Stitching for Dual-fisheye Lens Cameras Based On Rigid Moving Least Squares
Dual-fisheye lens cameras are becoming popular for 360-degree video capture,
especially for User-generated content (UGC), since they are affordable and
portable. Images generated by the dual-fisheye cameras have limited overlap and
hence require non-conventional stitching techniques to produce high-quality
360x180-degree panoramas. This paper introduces a novel method to align these
images using interpolation grids based on rigid moving least squares.
Furthermore, jitter is the critical issue arising when one applies the
image-based stitching algorithms to video. It stems from the unconstrained
movement of stitching boundary from one frame to another. Therefore, we also
propose a new algorithm to maintain the temporal coherence of stitching
boundary to provide jitter-free 360-degree videos. Results show that the method
proposed in this paper can produce higher quality stitched images and videos
than prior work.Comment: Preprint versio
- ā¦