1,447 research outputs found

    STUDY ON AUDIO AND VIDEO WATERMARKING

    Get PDF
    This paper gives the overview of audio and video watermarking. This paper introduces the basic requirements that affect the algorithms for audio and video watermarking which are perceptibility, robustness and security. The attacks which cause manipulations of the audio and video signals are also discussed. The common group of attacks on audio and video data is dynamics, filtering, conversion, compression, noise, modulation, time stretch and pitch shift, multiple watermark, cropping, rotation etc. The applications of audio and video watermarking are Fingerprinting, copyright protection, authentication, copy control etc. The audio watermarking techniques can be classified into Time-domain and Frequencydomain methods and video watermarking techniques are classified into spatial domain, frequency domain and formatspecific domain

    Variable bit rate video time-series and scene modeling using discrete-time statistically self-similar systems

    Get PDF
    This thesis investigates the application of discrete-time statistically self-similar (DTSS) systems to modeling of variable bit rate (VBR) video traffic data. The work is motivated by the fact that while VBR video has been characterized as self-similar by various researchers, models based on self-similarity considerations have not been previously studied. Given the relationship between self-similarity and long-range dependence the potential for using DTSS model in applications involving modeling of VBR MPEG video traffic data is presented. This thesis initially explores the characteristic properties of the model and then establishes relationships between the discrete-time self-similar model and fractional order transfer function systems. Using white noise as the input, the modeling approach is presented using least-square fitting technique of the output autocorrelations to the correlations of various VBR video trace sequences. This measure is used to compare the model performance with the performance of other existing models such as Markovian, long-range dependent and M/G/(infinity) . The study shows that using heavy-tailed inputs the output of these models can be used to match both the scene time-series correlations as well as scene density functions. Furthermore, the discrete-time self-similar model is applied to scene classification in VBR MPEG video to provide a demonstration of potential application of discrete-time self-similar models in modeling self-similar and long-range dependent data. Simulation results have shown that the proposed modeling technique is indeed a better approach than several earlier approaches and finds application is areas such as automatic scene classification, estimation of motion intensity and metadata generation for MPEG-7 applications

    Spatial prediction based on self-similarity compensation for 3D holoscopic image and video coding

    Get PDF
    WOS:000298962501022 (Nº de Acesso Web of Science)Holoscopic imaging, also known as integral imaging, provides a solution for glassless 3D, and is promising to change the market for 3D television. To start, this paper briefly describes the general concepts of holoscopic imaging, focusing mainly on the spatial correlations inherent to this new type of content, which appear due to the micro-lens array that is used for both acquisition and display. The micro-images that are formed behind each micro-lens, from which only one pixel is viewed from a given observation point, have a high cross-correlation between them, which can be exploited for coding. A novel scheme for spatial prediction, exploring the particular arrangement of holoscopic images, is proposed. The proposed scheme can be used for both still image coding and intra-coding of video. Experimental results based on an H.264/AVC video codec modified to handle 3D holoscopic images and video are presented, showing the superior performance of this approach

    Surveillance centric coding

    Get PDF
    PhDThe research work presented in this thesis focuses on the development of techniques specific to surveillance videos for efficient video compression with higher processing speed. The Scalable Video Coding (SVC) techniques are explored to achieve higher compression efficiency. The framework of SVC is modified to support Surveillance Centric Coding (SCC). Motion estimation techniques specific to surveillance videos are proposed in order to speed up the compression process of the SCC. The main contributions of the research work presented in this thesis are divided into two groups (i) Efficient Compression and (ii) Efficient Motion Estimation. The paradigm of Surveillance Centric Coding (SCC) is introduced, in which coding aims to achieve bit-rate optimisation and adaptation of surveillance videos for storing and transmission purposes. In the proposed approach the SCC encoder communicates with the Video Content Analysis (VCA) module that detects events of interest in video captured by the CCTV. Bit-rate optimisation and adaptation are achieved by exploiting the scalability properties of the employed codec. Time segments containing events relevant to surveillance application are encoded using high spatiotemporal resolution and quality while the irrelevant portions from the surveillance standpoint are encoded at low spatio-temporal resolution and / or quality. Thanks to the scalability of the resulting compressed bit-stream, additional bit-rate adaptation is possible; for instance for the transmission purposes. Experimental evaluation showed that significant reduction in bit-rate can be achieved by the proposed approach without loss of information relevant to surveillance applications. In addition to more optimal compression strategy, novel approaches to performing efficient motion estimation specific to surveillance videos are proposed and implemented with experimental results. A real-time background subtractor is used to detect the presence of any motion activity in the sequence. Different approaches for selective motion estimation, GOP based, Frame based and Block based, are implemented. In the former, motion estimation is performed for the whole group of pictures (GOP) only when a moving object is detected for any frame of the GOP. iii While for the Frame based approach; each frame is tested for the motion activity and consequently for selective motion estimation. The selective motion estimation approach is further explored at a lower level as Block based selective motion estimation. Experimental evaluation showed that significant reduction in computational complexity can be achieved by applying the proposed strategy. In addition to selective motion estimation, a tracker based motion estimation and fast full search using multiple reference frames has been proposed for the surveillance videos. Extensive testing on different surveillance videos shows benefits of application of proposed approaches to achieve the goals of the SCC

    Multi-Stream Management for Supporting Multi-Party 3D Tele-Immersive Environments

    Get PDF
    Three-dimensional tele-immersive (3DTI) environments have great potential to promote collaborative work among geographically distributed participants. However, extensive application of 3DTI environments is still hindered by problems pertaining to scalability, manageability and reliance of special-purpose components. Thus, one critical question is how to organize the acquisition, transmission and display of large volume real-time 3D visual data over commercially available computing and networking infrastructures so that .everybody. would be able to install and enjoy 3DTI environments for high quality tele-collaboration. In the thesis, we explore the design space from the angle of multi-stream Quality-of-Service (QoS) management to support multi-party 3DTI communication. In 3DTI environments, multiple correlated 3D video streams are deployed to provide a comprehensive representation of the physical scene. Traditional QoS approach in 2D and single-stream scenario has become inadequate. On the other hand, the existence of multiple streams provides unique opportunity for QoS provisioning. We propose an innovative cross-layer hierarchical and distributed multi-stream management middleware framework for QoS provisioning to fully enable multi-party 3DTI communication over general delivery infrastructure. The major contributions are as follows. First, we introduce the view model for representing the user interest in the application layer. The design revolves around the concept of view-aware multi-stream coordination, which leverages the central role of view semantics in 3D video systems. Second, in the stream differentiation layer we present the design of view to stream mapping, where a subset of relevant streams are selected based on the relative importance of each stream to the current view. Conventional streaming controllers focus on a fixed set of streams specified by the application. Different from all the others, in our management framework the application layer only specifies the view information while the underlying controller dynamically determines the set of streams to be managed. Third, in the stream coordination layer we present two designs applicable in different situations. In the case of end-to-end 3DTI communication, a learning-based controller is embedded which provides bandwidth allocation for relevant streams. In the case of multi-party 3DTI communication, we propose a novel ViewCast protocol to coordinate the multi-stream content dissemination upon an end-system overlay network
    corecore