10 research outputs found
Content-Adaptive Variable Framerate Encoding Scheme for Green Live Streaming
Adaptive live video streaming applications use a fixed predefined
configuration for the bitrate ladder with constant framerate and encoding
presets in a session. However, selecting optimized framerates and presets for
every bitrate ladder representation can enhance perceptual quality, improve
computational resource allocation, and thus, the streaming energy efficiency.
In particular, low framerates for low-bitrate representations reduce
compression artifacts and decrease encoding energy consumption. In addition, an
optimized preset may lead to improved compression efficiency. To this light,
this paper proposes a Content-adaptive Variable Framerate (CVFR) encoding
scheme, which offers two modes of operation: ecological (ECO) and high-quality
(HQ). CVFR-ECO optimizes for the highest encoding energy savings by predicting
the optimized framerate for each representation in the bitrate ladder. CVFR-HQ
takes it further by predicting each representation's optimized
framerate-encoding preset pair using low-complexity discrete cosine transform
energy-based spatial and temporal features for compression efficiency and
sustainable storage. We demonstrate the advantage of CVFR using the x264
open-source video encoder. The results show that CVFR-ECO yields an average
PSNR and VMAF increase of 0.02 dB and 2.50 points, respectively, for the same
bitrate, compared to the fastest preset highest framerate encoding. CVFR-ECO
also yields an average encoding and storage energy consumption reduction of
34.54% and 76.24%, considering a just noticeable difference (JND) of six VMAF
points. In comparison, CVFR-HQ yields an average increase in PSNR and VMAF of
2.43 dB and 10.14 points, respectively, for the same bitrate. Finally, CVFR-HQ
resulted in an average reduction in storage energy consumption of 83.18%,
considering a JND of six VMAF points
BVI-VFI: A Video Quality Database for Video Frame Interpolation
Video frame interpolation (VFI) is a fundamental research topic in video
processing, which is currently attracting increased attention across the
research community. While the development of more advanced VFI algorithms has
been extensively researched, there remains little understanding of how humans
perceive the quality of interpolated content and how well existing objective
quality assessment methods perform when measuring the perceived quality. In
order to narrow this research gap, we have developed a new video quality
database named BVI-VFI, which contains 540 distorted sequences generated by
applying five commonly used VFI algorithms to 36 diverse source videos with
various spatial resolutions and frame rates. We collected more than 10,800
quality ratings for these videos through a large scale subjective study
involving 189 human subjects. Based on the collected subjective scores, we
further analysed the influence of VFI algorithms and frame rates on the
perceptual quality of interpolated videos. Moreover, we benchmarked the
performance of 33 classic and state-of-the-art objective image/video quality
metrics on the new database, and demonstrated the urgent requirement for more
accurate bespoke quality assessment methods for VFI. To facilitate further
research in this area, we have made BVI-VFI publicly available at
https://github.com/danier97/BVI-VFI-database