32 research outputs found
Efficient HEVC-based video adaptation using transcoding
In a video transmission system, it is important to take into account the great diversity of the network/end-user constraints. On the one hand, video content is typically streamed over a network that is characterized by different bandwidth capacities. In many cases, the bandwidth is insufficient to transfer the video at its original quality. On the other hand, a single video is often played by multiple devices like PCs, laptops, and cell phones. Obviously, a single video would not satisfy their different constraints.
These diversities of the network and devices capacity lead to the need for video adaptation techniques, e.g., a reduction of the bit rate or spatial resolution. Video transcoding, which modifies a property of the video without the change of the coding format, has been well-known as an efficient adaptation solution. However, this approach comes along with a high computational complexity, resulting in huge energy consumption in the network and possibly network latency.
This presentation provides several optimization strategies for the transcoding process of HEVC (the latest High Efficiency Video Coding standard) video streams. First, the computational complexity of a bit rate transcoder (transrater) is reduced. We proposed several techniques to speed-up the encoder of a transrater, notably a machine-learning-based approach and a novel coding-mode evaluation strategy have been proposed. Moreover, the motion estimation process of the encoder has been optimized with the use of decision theory and the proposed fast search patterns. Second, the issues and challenges of a spatial transcoder have been solved by using machine-learning algorithms. Thanks to their great performance, the proposed techniques are expected to significantly help HEVC gain popularity in a wide range of modern multimedia applications
Advanced heterogeneous video transcoding
PhDVideo transcoding is an essential tool to promote inter-operability
between different video communication systems. This thesis presents
two novel video transcoders, both operating on bitstreams of the cur-
rent H.264/AVC standard. The first transcoder converts H.264/AVC
bitstreams to a Wavelet Scalable Video Codec (W-SVC), while the second targets the emerging High Efficiency Video Coding (HEVC).
Scalable Video Coding (SVC) enables low complexity adaptation
of compressed video, providing an efficient solution for content delivery
through heterogeneous networks. The transcoder proposed here aims at
exploiting the advantages offered by SVC technology when dealing with
conventional coders and legacy video, efficiently reusing information
found in the H.264/AVC bitstream to achieve a high rate-distortion
performance at a low complexity cost. Its main features include new
mode mapping algorithms that exploit the W-SVC larger macroblock
sizes, and a new state-of-the-art motion vector composition algorithm
that is able to tackle different coding configurations in the H.264/AVC
bitstream, including IPP or IBBP with multiple reference frames.
The emerging video coding standard, HEVC, is currently approaching the final stage of development prior to standardization. This thesis
proposes and evaluates several transcoding algorithms for the HEVC
codec. In particular, a transcoder based on a new method that is capable of complexity scalability, trading off rate-distortion performance
for complexity reduction, is proposed. Furthermore, other transcoding solutions are explored, based on a novel content-based modeling
approach, in which the transcoder adapts its parameters based on the
contents of the sequence being encoded.
Finally, the application of this research is not constrained to these
transcoders, as many of the techniques developed aim to contribute
to advance the research on this field, and have the potential to be
incorporated in different video transcoding architectures
PokroÄilĂ© metody detekce steganografickĂ©ho obsahu
Steganography can be used for illegal activities. It is essential to be prepared. To detect steganography images, we have a counter-technique known as steganalysis. There are different steganalysis types, depending on if the original artifact (cover work) is known or not, or we know which algorithm was used for embedding. In terms of practical use, the most important are âblind steganalysisâ methods that can be applied to image files because we do not have the original cover work for comparison. This philosophiĂŠ doctor thesis describes the methodology to the issues of image steganalysis.In this work, it is crucial to understand the behavior of the targeted steganography algorithm. Then we can use it is weaknesses to increase the detection capability and success of categorization. We are primarily focusing on breaking the steganography algorithm OutGuess2.0. and secondary on breaking the F5 algorithm. We are analyzing the detector's ability, which utilizes a calibration process, blockiness calculation, and shallow neural network, to detect the presence of steganography message in the suspected image. The new approach and results are discussed in this Ph.D. thesis.Steganografie mĆŻĆŸe bĂœt vyuĆŸita k nelegĂĄlnĂm aktivitĂĄm. Proto je velmi dĆŻleĆŸitĂ© bĂœt pĆipraven. K detekci steganografickĂ©ho obrĂĄzku mĂĄme k dispozici techniku znĂĄmou jako stegoanalĂœza. ExistujĂ rĆŻznĂ© typy stegoanalĂœzy v zĂĄvislosti na tom, zda je znĂĄm originĂĄlnĂ nosiÄ nebo zdali vĂme, jakĂœ byl pouĆŸit algoritmus pro vloĆŸenĂ tajnĂ© zprĂĄvy. Z hlediska praktickĂ©ho pouĆŸitĂ jsou nejdĆŻleĆŸitÄjĆĄĂ metody "slepĂ© stagoanalĂœzy", kterĂ© zle aplikovat na obrazovĂ© soubory a jelikoĆŸ nemĂĄme originĂĄlnĂ nosiÄ pro srovnĂĄnĂ. Tato doktorskĂĄ prĂĄce popisuje metodologii obrazovĂ© stegoanalĂœzy. V tĂ©to prĂĄci je dĆŻleĆŸitĂ© porozumÄt chovĂĄnĂ cĂlenĂ©ho steganografickĂ©ho algoritmu. Pak mĆŻĆŸeme vyuĆŸĂt jeho slabiny ke zvĂœĆĄenĂ detekÄnĂ schopnosti a ĂșspÄĆĄnosti kategorizace. PrimĂĄrnÄ se zamÄĆujeme na prolomenĂ steganografickĂ©ho algoritmu OutGuess2.0 a sekundĂĄrnÄ na algoritmus F5. Analyzujeme schopnost detektoru, kterĂœ vyuĆŸĂvĂĄ proces kalibrace, vĂœpoÄtu shlukovĂĄnĂ a mÄlkou neuronovou sĂĆ„ k detekci pĆĂtomnosti steganografickĂ© zprĂĄvy na podezĆelĂ©m snĂmku. NovĂœ pĆĂstup a vĂœsledky jsou sepsĂĄny v tĂ©to doktorskĂ© prĂĄci.460 - Katedra informatikyvyhovÄ
A NOVEL JOINT PERCEPTUAL ENCRYPTION AND WATERMARKING SCHEME (JPEW) WITHIN JPEG FRAMEWORK
Due to the rapid growth in internet and multimedia technologies, many new
commercial applications like video on demand (VOD), pay-per-view and real-time
multimedia broadcast etc, have emerged. To ensure the integrity and confidentiality of
the multimedia content, the content is usually watermarked and then encrypted or vice
versa. If the multimedia content needs to be watermarked and encrypted at the same
time, the watermarking function needs to be performed first followed by encryption
function. Hence, if the watermark needs to be extracted then the multimedia data
needs to be decrypted first followed by extraction of the watermark. This results in
large computational overhead. The solution provided in the literature for this problem
is by using what is called partial encryption, in which media data are partitioned into
two parts - one to be watermarked and the other is encrypted. In addition, some
multimedia applications i.e. video on demand (VOD), Pay-TV, pay-per-view etc,
allow multimedia content preview which involves âperceptualâ encryption wherein all
or some selected part of the content is, perceptually speaking, distorted with an
encryption key. Up till now no joint perceptual encryption and watermarking scheme
has been proposed in the literature.
In this thesis, a novel Joint Perceptual Encryption and Watermarking (JPEW)
scheme is proposed that is integrated within JPEG standard. The design of JPEW
involves the design and development of both perceptual encryption and watermarking
schemes that are integrated in JPEG and feasible within the âpartialâ encryption
framework. The perceptual encryption scheme exploits the energy distribution of AC
components and DC components bitplanes of continuous-tone images and is carried
out by selectively encrypting these AC coefficients and DC components bitplanes.
The encryption itself is based on a chaos-based permutation reported in an earlier
work. Similarly, in contrast to the traditional watermarking schemes, the proposed
watermarking scheme makes use of DC component of the image and it is carried out
by selectively substituting certain bitplanes of DC components with watermark bits.
vi ii
Apart from the aforesaid JPEW, additional perceptual encryption scheme, integrated
in JPEG, has also been proposed. The scheme is outside of joint framework and
implements perceptual encryption on region of interest (ROI) by scrambling the DCT
blocks of the chosen ROI.
The performances of both, perceptual encryption and watermarking schemes are
evaluated and compared with Quantization Index modulation (QIM) based
watermarking scheme and reversible Histogram Spreading (RHS) based perceptual
encryption scheme. The results show that the proposed watermarking scheme is
imperceptible and robust, and suitable for authentication. Similarly, the proposed
perceptual encryption scheme outperforms the RHS based scheme in terms of number
of operations required to achieve a given level of perceptual encryption and provides
control over the amount of perceptual encryption. The overall security of the JPEW
has also been evaluated. Additionally, the performance of proposed separate
perceptual encryption scheme has been thoroughly evaluated in terms of security and
compression efficiency. The scheme is found to be simpler in implementation, have
insignificant effect on compression ratios and provide more options for the selection
of control factor
Virtual Reality Games for Motor Rehabilitation
This paper presents a fuzzy logic based method to track user satisfaction without the need for devices to monitor users physiological conditions. User satisfaction is the key to any productâs acceptance; computer applications and video games provide a unique opportunity to provide a tailored environment for each user to better suit their needs. We have implemented a non-adaptive fuzzy logic model of emotion, based on the emotional component of the Fuzzy Logic Adaptive Model of Emotion (FLAME) proposed by El-Nasr, to estimate player emotion in UnrealTournament 2004. In this paper we describe the implementation of this system and present the results of one of several play tests. Our research contradicts the current literature that suggests physiological measurements are needed. We show that it is possible to use a software only method to estimate user emotion
MediaSync: Handbook on Multimedia Synchronization
This book provides an approachable overview of the most recent advances in the fascinating field of media synchronization (mediasync), gathering contributions from the most representative and influential experts. Understanding the challenges of this field in the current multi-sensory, multi-device, and multi-protocol world is not an easy task. The book revisits the foundations of mediasync, including theoretical frameworks and models, highlights ongoing research efforts, like hybrid broadband broadcast (HBB) delivery and users' perception modeling (i.e., Quality of Experience or QoE), and paves the way for the future (e.g., towards the deployment of multi-sensory and ultra-realistic experiences). Although many advances around mediasync have been devised and deployed, this area of research is getting renewed attention to overcome remaining challenges in the next-generation (heterogeneous and ubiquitous) media ecosystem. Given the significant advances in this research area, its current relevance and the multiple disciplines it involves, the availability of a reference book on mediasync becomes necessary. This book fills the gap in this context. In particular, it addresses key aspects and reviews the most relevant contributions within the mediasync research space, from different perspectives. Mediasync: Handbook on Multimedia Synchronization is the perfect companion for scholars and practitioners that want to acquire strong knowledge about this research area, and also approach the challenges behind ensuring the best mediated experiences, by providing the adequate synchronization between the media elements that constitute these experiences