175 research outputs found

    Audio Signal Processing Using Time-Frequency Approaches: Coding, Classification, Fingerprinting, and Watermarking

    Get PDF
    Audio signals are information rich nonstationary signals that play an important role in our day-to-day communication, perception of environment, and entertainment. Due to its non-stationary nature, time- or frequency-only approaches are inadequate in analyzing these signals. A joint time-frequency (TF) approach would be a better choice to efficiently process these signals. In this digital era, compression, intelligent indexing for content-based retrieval, classification, and protection of digital audio content are few of the areas that encapsulate a majority of the audio signal processing applications. In this paper, we present a comprehensive array of TF methodologies that successfully address applications in all of the above mentioned areas. A TF-based audio coding scheme with novel psychoacoustics model, music classification, audio classification of environmental sounds, audio fingerprinting, and audio watermarking will be presented to demonstrate the advantages of using time-frequency approaches in analyzing and extracting information from audio signals.</p

    AUDIO WATERMARKING WITH ANGLE QUANTIZATION BASED ON DISCRETE WAVELET TRANSFORM AND SINGULAR VALUE DECOMPOSITION

    Get PDF
    An&nbsp;audio watermark&nbsp;is a unique electronic identifier embedded in an&nbsp;audio&nbsp;signal, typically used to identify ownership of copyright. Proposed work is a new method of audio watermark hiding inside another bigger cover standard audio cover. The method includes ‘harr’ wavelet based Discrete Wavelet Transform decomposition of frequencies hence the audio samples of watermark gets hidden only those parts of cover audio where human ears are less sensible according to Human Auditory System. Proposed method also includes the Singular Value Decomposition, which is required for making our method robust against the various communication of processing attacks like compression, filtering, fading or noise addition. Proposed work is also using the concept of angular modulation which initially modifies the audio watermark in to provide extra security and also extra robustness in communication. The design is been develop on MATLAB 2013b version and verification of design o the same.&nbsp

    Robust Video Watermarking Algorithm based on DCT-SVD approach and Encryption

    Get PDF
    Sharing of digital media content over the internet is increasing everyday .Digital watermarking is a technique used to protect the intellectual property rights of multimedia content owners. In this paper, we propose a robust video watermarking scheme that utilizes Discrete Cosine Transform (DCT) and Singular Value Decomposition (SVD) for embedding a watermark into video frames. The proposed method uses encryption to make the watermark more robust against malicious attacks. The encryption key is used to modify the watermark before it is embedded in the video frames. The modified watermark is then embedded in the DCT and SVD coefficients of the video frames. The experimental results show that the proposed method provides better robustness against various attacks such as compression, noise addition, and filtering, while maintaining good perceptual quality of the watermarked video. The proposed method also shows better resistance against geometric attacks such as cropping, rotation, and scaling. Overall, the proposed method provides an effective solution for protecting the intellectual property rights of multimedia content owners in video distribution and transmission scenarios

    A Framework for Multimedia Data Hiding (Security)

    Get PDF
    With the proliferation of multimedia data such as images, audio, and video, robust digital watermarking and data hiding techniques are needed for copyright protection, copy control, annotation, and authentication. While many techniques have been proposed for digital color and grayscale images, not all of them can be directly applied to binary document images. The difficulty lies in the fact that changing pixel values in a binary document could introduce Irregularities that is very visually noticeable. We have seen but limited number of papers proposing new techniques and ideas for document image watermarking and data hiding. In this paper, we present an overview and summary of recent developments on this important topic, and discuss important issues such as robustness and data hiding capacity of the different techniques

    Informed Multiple-F0 Estimation Applied to Monaural Audio Source Separation

    No full text
    International audienceThis paper proposes a new informed source separation technique which combines music transcription with source separation. The presented system is based on a coder / decoder configuration where a classic (not informed) multiple-F0 estimation is applied on each separated source signal assumed known at the coder before the mixing process. Thus, the extra information required to recover the reference transcription of each isolated instrument is computed and inaudibly embedded into the mixture using a watermarking technique. At the decoder, where the original source signals are unknown, instruments are separated from the mixture using the informed transcription of each source signal. In this paper, we show that a classic (non-informed) F0 estimator can be used to reduce the amount of bits necessary to transmit the exact transcription of each isolated instrument

    Source Separation in the Presence of Side-information

    Get PDF
    The source separation problem involves the separation of unknown signals from their mixture. This problem is relevant in a wide range of applications from audio signal processing, communication, biomedical signal processing and art investigation to name a few. There is a vast literature on this problem which is based on either making strong assumption on the source signals or availability of additional data. This thesis proposes new algorithms for source separation with side information where one observes the linear superposition of two source signals plus two additional signals that are correlated with the mixed ones. The first algorithm is based on two ingredients: first, we learn a Gaussian mixture model (GMM) for the joint distribution of a source signal and the corresponding correlated side information signal; second, we separate the signals using standard computationally efficient conditional mean estimators. This also puts forth new recovery guarantees for this source separation algorithm. In particular, under the assumption that the signals can be perfectly described by a GMM model, we characterize necessary and sufficient conditions for reliable source separation in the asymptotic regime of low-noise as a function of the geometry of the underlying signals and their interaction. It is shown that if the subspaces spanned by the innovation components of the source signals with respect to the side information signals have zero intersection, provided that we observe a certain number of linear measurements from the mixture, then we can reliably separate the sources; otherwise we cannot. The second algorithms is based on deep learning where we introduce a novel self-supervised algorithm for the source separation problem. Source separation is intrinsically unsupervised and the lack of training data makes it a difficult task for artificial intelligence to solve. The proposed framework takes advantage of the available data and delivers near perfect separation results in real data scenarios. Our proposed frameworks – which provide new ways to incorporate side information to aid the solution of the source separation problem – are also employed in a real-world art investigation application involving the separation of mixtures of X-Ray images. The simulation results showcase the superiority of our algorithm against other state-of-the-art algorithms

    Lifting dual tree complex wavelets transform

    Get PDF
    We describe the lifting dual tree complex wavelet transform (LDTCWT), a type of lifting wavelets remodeling that produce complex coefficients by employing a dual tree of lifting wavelets filters to get its real part and imaginary part. Permits the remodel to produce approximate shift invariance, directionally selective filters and reduces the computation time (properties lacking within the classical wavelets transform). We describe a way to estimate the accuracy of this approximation and style appropriate filters to attain this. These benefits are often exploited among applications like denoising, segmentation, image fusion and compression. The results of applications shrinkage denoising demonstrate objective and subjective enhancements over the dual tree complex wavelet transform (DTCWT). The results of the shrinkage denoising example application indicate empirical and subjective enhancements over the DTCWT. The new transform with the DTCWT provide a trade-off between denoising computational competence of performance, and memory necessities. We tend to use the PSNR (peak signal to noise ratio) alongside the structural similarity index measure (SSIM) and the SSIM map to estimate denoised image quality

    An overview of informed audio source separation

    Get PDF
    International audienceAudio source separation consists in recovering different unknown signals called sources by filtering their observed mixtures. In music processing, most mixtures are stereophonic songs and the sources are the individual signals played by the instruments, e.g. bass, vocals, guitar, etc. Source separation is often achieved through a classical generalized Wiener filtering, which is controlled by parameters such as the power spectrograms and the spatial locations of the sources. For an efficient filtering, those parameters need to be available and their estimation is the main challenge faced by separation algorithms. In the blind scenario, only the mixtures are available and performance strongly depends on the mixtures considered. In recent years, much research has focused on informed separation, which consists in using additional available information about the sources to improve the separation quality. In this paper, we review some recent trends in this direction

    Improved Convolutive and Under-Determined Blind Audio Source Separation with MRF Smoothing

    Get PDF
    Convolutive and under-determined blind audio source separation from noisy recordings is a challenging problem. Several computational strategies have been proposed to address this problem. This study is concerned with several modifications to the expectation-minimization-based algorithm, which iteratively estimates the mixing and source parameters. This strategy assumes that any entry in each source spectrogram is modeled using superimposed Gaussian components, which are mutually and individually independent across frequency and time bins. In our approach, we resolve this issue by considering a locally smooth temporal and frequency structure in the power source spectrograms. Local smoothness is enforced by incorporating a Gibbs prior in the complete data likelihood function, which models the interactions between neighboring spectrogram bins using a Markov random field. Simulations using audio files derived from stereo audio source separation evaluation campaign 2008 demonstrate high efficiency with the proposed improvement
    • …
    corecore