69 research outputs found
Rate-Distortion Analysis of Multiview Coding in a DIBR Framework
Depth image based rendering techniques for multiview applications have been
recently introduced for efficient view generation at arbitrary camera
positions. Encoding rate control has thus to consider both texture and depth
data. Due to different structures of depth and texture images and their
different roles on the rendered views, distributing the available bit budget
between them however requires a careful analysis. Information loss due to
texture coding affects the value of pixels in synthesized views while errors in
depth information lead to shift in objects or unexpected patterns at their
boundaries. In this paper, we address the problem of efficient bit allocation
between textures and depth data of multiview video sequences. We adopt a
rate-distortion framework based on a simplified model of depth and texture
images. Our model preserves the main features of depth and texture images.
Unlike most recent solutions, our method permits to avoid rendering at encoding
time for distortion estimation so that the encoding complexity is not
augmented. In addition to this, our model is independent of the underlying
inpainting method that is used at decoder. Experiments confirm our theoretical
results and the efficiency of our rate allocation strategy
Optimized learned entropy coding parameters for practical neural-based image and video compression
Neural-based image and video codecs are significantly more power-efficient
when weights and activations are quantized to low-precision integers. While
there are general-purpose techniques for reducing quantization effects, large
losses can occur when specific entropy coding properties are not considered.
This work analyzes how entropy coding is affected by parameter quantizations,
and provides a method to minimize losses. It is shown that, by using a certain
type of coding parameters to be learned, uniform quantization becomes
practically optimal, also simplifying the minimization of code memory
requirements. The mathematical properties of the new representation are
presented, and its effectiveness is demonstrated by coding experiments, showing
that good results can be obtained with precision as low as 4~bits per network
output, and practically no loss with 8~bits.Comment: 2022 IEEE International Conference on Image Processing (ICIP
Boosting neural video codecs by exploiting hierarchical redundancy
In video compression, coding efficiency is improved by reusing pixels from
previously decoded frames via motion and residual compensation. We define two
levels of hierarchical redundancy in video frames: 1) first-order: redundancy
in pixel space, i.e., similarities in pixel values across neighboring frames,
which is effectively captured using motion and residual compensation, 2)
second-order: redundancy in motion and residual maps due to smooth motion in
natural videos. While most of the existing neural video coding literature
addresses first-order redundancy, we tackle the problem of capturing
second-order redundancy in neural video codecs via predictors. We introduce
generic motion and residual predictors that learn to extrapolate from
previously decoded data. These predictors are lightweight, and can be employed
with most neural video codecs in order to improve their rate-distortion
performance. Moreover, while RGB is the dominant colorspace in neural video
coding literature, we introduce general modifications for neural video codecs
to embrace the YUV420 colorspace and report YUV420 results. Our experiments
show that using our predictors with a well-known neural video codec leads to
38% and 34% bitrate savings in RGB and YUV420 colorspaces measured on the UVG
dataset.Comment: WACV 202
- …