34 research outputs found
JND-Based Perceptual Video Coding for 4:4:4 Screen Content Data in HEVC
The JCT-VC standardized Screen Content Coding (SCC) extension in the HEVC HM
RExt + SCM reference codec offers an impressive coding efficiency performance
when compared with HM RExt alone; however, it is not significantly perceptually
optimized. For instance, it does not include advanced HVS-based perceptual
coding methods, such as JND-based spatiotemporal masking schemes. In this
paper, we propose a novel JND-based perceptual video coding technique for HM
RExt + SCM. The proposed method is designed to further improve the compression
performance of HM RExt + SCM when applied to YCbCr 4:4:4 SC video data. In the
proposed technique, luminance masking and chrominance masking are exploited to
perceptually adjust the Quantization Step Size (QStep) at the Coding Block (CB)
level. Compared with HM RExt 16.10 + SCM 8.0, the proposed method considerably
reduces bitrates (Kbps), with a maximum reduction of 48.3%. In addition to
this, the subjective evaluations reveal that SC-PAQ achieves visually lossless
coding at very low bitrates.Comment: Preprint: 2018 IEEE International Conference on Acoustics, Speech and
Signal Processing (ICASSP 2018
Geometry-based spherical JND modeling for 360 display
360 videos have received widespread attention due to its realistic
and immersive experiences for users. To date, how to accurately model the user
perceptions on 360 display is still a challenging issue. In this paper,
we exploit the visual characteristics of 360 projection and display and
extend the popular just noticeable difference (JND) model to spherical JND
(SJND). First, we propose a quantitative 2D-JND model by jointly considering
spatial contrast sensitivity, luminance adaptation and texture masking effect.
In particular, our model introduces an entropy-based region classification and
utilizes different parameters for different types of regions for better
modeling performance. Second, we extend our 2D-JND model to SJND by jointly
exploiting latitude projection and field of view during 360 display.
With this operation, SJND reflects both the characteristics of human vision
system and the 360 display. Third, our SJND model is more consistent
with user perceptions during subjective test and also shows more tolerance in
distortions with fewer bit rates during 360 video compression. To
further examine the effectiveness of our SJND model, we embed it in Versatile
Video Coding (VVC) compression. Compared with the state-of-the-arts, our
SJND-VVC framework significantly reduced the bit rate with negligible loss in
visual quality
High-fidelity imaging : the computational models of the human visual system in high dynamic range video compression, visible difference prediction and image processing
As new displays and cameras offer enhanced color capabilities, there is a need to extend the precision of digital content. High Dynamic Range (HDR) imaging encodes images and video with higher than normal bit-depth precision, enabling representation of the complete color gamut and the full visible range of luminance. This thesis addresses three problems of HDR imaging: the measurement of visible distortions in HDR images, lossy compression for HDR video, and artifact-free image processing. To measure distortions in HDR images, we develop a visual difference predictor for HDR images that is based on a computational model of the human visual system. To address the problem of HDR image encoding and compression, we derive a perceptually motivated color space for HDR pixels that can efficiently encode all perceivable colors and distinguishable shades of brightness. We use the derived color space to extend the MPEG-4 video compression standard for encoding HDR movie sequences. We also propose a backward-compatible HDR MPEG compression algorithm that encodes both a low-dynamic range and an HDR video sequence into a single MPEG stream. Finally, we propose a framework for image processing in the contrast domain. The framework transforms an image into multi-resolution physical contrast images (maps), which are then rescaled in just-noticeable-difference (JND) units. The application of the framework is demonstrated with a contrast-enhancing tone mapping and a color to gray conversion that preserves color saliency.Aktuelle Innovationen in der Farbverarbeitung bei Bildschirmen und Kameras erzwingen eine Präzisionserweiterung bei digitalen Medien. High Dynamic Range (HDR) kodieren Bilder und Video mit einer grösseren Bittiefe pro Pixel, und ermöglichen damit die Darstellung des kompletten Farbraums und aller sichtbaren Helligkeitswerte. Diese Arbeit konzentriert sich auf drei Probleme in der HDR-Verarbeitung: Messung von für den Menschen störenden Fehlern in HDR-Bildern, verlustbehaftete Kompression von HDR-Video, und visuell verlustfreie HDR-Bildverarbeitung. Die Messung von HDR-Bildfehlern geschieht mittels einer Vorhersage von sichtbaren Unterschieden zweier HDR-Bilder. Die Vorhersage basiert dabei auf einer Modellierung der menschlichen Sehens. Wir addressieren die Kompression und Kodierung von HDR-Bildern mit der Ableitung eines perzeptuellen Farbraums für HDR-Pixel, der alle wahrnehmbaren Farben und deren unterscheidbaren Helligkeitsnuancen effizient abbildet. Danach verwenden wir diesen Farbraum für die Erweiterung des MPEG-4 Videokompressionsstandards, welcher sich hinfort auch für die Kodierung von HDR-Videosequenzen eignet. Wir unterbreiten weiters eine rückwärts-kompatible MPEG-Kompression von HDR-Material, welche die übliche YUV-Bildsequenz zusammen mit dessen HDRVersion in einen gemeinsamen MPEG-Strom bettet. Abschliessend erklären wir unser Framework zur Bildverarbeitung in der Kontrastdomäne. Das Framework transformiert Bilder in mehrere physikalische Kontrastauflösungen, um sie danach in Einheiten von just-noticeable-difference (JND, noch erkennbarem Unterschied) zu reskalieren. Wir demonstrieren den Nutzen dieses Frameworks anhand von einem kontrastverstärkenden Tone Mapping-Verfahren und einer Graukonvertierung, die die urspr ünglichen Farbkontraste bestmöglich beibehält
Watermarking of HDR images in the spatial domain with HVS-imperceptibility
This paper presents a watermarking method in the spatial domain with HVS-imperceptibility for High Dynamic Range (HDR) images. The proposed method combines the content readability afforded by invisible watermarking with the visual ownership identification afforded by visible watermarking. The HVS-imperceptibility is guaranteed thanks to a Luma Variation Tolerance (LVT) curve, which is associated with the transfer function (TF) used for HDR encoding and provides the information needed to embed an imperceptible watermark in the spatial domain. The LVT curve is based on the inaccuracies between the non-linear digital representation of the linear luminance acquired by an HDR sensor and the brightness perceived by the Human Visual System (HVS) from the linear luminance displayed on an HDR screen. The embedded watermarks remain imperceptible to the HVS as long as the TF is not altered or the normal calibration and colorimetry conditions of the HDR screen remain unchanged. Extensive qualitative and quantitative evaluations on several HDR images encoded by two widely-used TFs confirm the strong HVSimperceptibility capabilities of the method, as well as the robustness of the embedded watermarks to tone mapping, lossy compression, and common signal processing operations