250 research outputs found

    Modeling and frequency tracking of marine mammal whistle calls

    Get PDF
    Submitted in partial fulfillment of the requirements for the degree of Master of Science at the Massachusetts Institute of Technology and the Woods Hole Oceanographic Institution February 2009Marine mammal whistle calls present an attractive medium for covert underwater communications. High quality models of the whistle calls are needed in order to synthesize natural-sounding whistles with embedded information. Since the whistle calls are composed of frequency modulated harmonic tones, they are best modeled as a weighted superposition of harmonically related sinusoids. Previous research with bottlenose dolphin whistle calls has produced synthetic whistles that sound too “clean” for use in a covert communications system. Due to the sensitivity of the human auditory system, watermarking schemes that slightly modify the fundamental frequency contour have good potential for producing natural-sounding whistles embedded with retrievable watermarks. Structured total least squares is used with linear prediction analysis to track the time-varying fundamental frequency and harmonic amplitude contours throughout a whistle call. Simulation and experimental results demonstrate the capability to accurately model bottlenose dolphin whistle calls and retrieve embedded information from watermarked synthetic whistle calls. Different fundamental frequency watermarking schemes are proposed based on their ability to produce natural sounding synthetic whistles and yield suitable watermark detection and retrieval

    Text-based Editing of Talking-head Video

    No full text
    Editing talking-head video to change the speech content or to remove filler words is challenging. We propose a novel method to edit talking-head video based on its transcript to produce a realistic output video in which the dialogue of the speaker has been modified, while maintaining a seamless audio-visual flow (i.e. no jump cuts). Our method automatically annotates an input talking-head video with phonemes, visemes, 3D face pose and geometry, reflectance, expression and scene illumination per frame. To edit a video, the user has to only edit the transcript, and an optimization strategy then chooses segments of the input corpus as base material. The annotated parameters corresponding to the selected segments are seamlessly stitched together and used to produce an intermediate video representation in which the lower half of the face is rendered with a parametric face model. Finally, a recurrent video generation network transforms this representation to a photorealistic video that matches the edited transcript. We demonstrate a large variety of edits, such as the addition, removal, and alteration of words, as well as convincing language translation and full sentence synthesis

    Source Separation in the Presence of Side-information

    Get PDF
    The source separation problem involves the separation of unknown signals from their mixture. This problem is relevant in a wide range of applications from audio signal processing, communication, biomedical signal processing and art investigation to name a few. There is a vast literature on this problem which is based on either making strong assumption on the source signals or availability of additional data. This thesis proposes new algorithms for source separation with side information where one observes the linear superposition of two source signals plus two additional signals that are correlated with the mixed ones. The first algorithm is based on two ingredients: first, we learn a Gaussian mixture model (GMM) for the joint distribution of a source signal and the corresponding correlated side information signal; second, we separate the signals using standard computationally efficient conditional mean estimators. This also puts forth new recovery guarantees for this source separation algorithm. In particular, under the assumption that the signals can be perfectly described by a GMM model, we characterize necessary and sufficient conditions for reliable source separation in the asymptotic regime of low-noise as a function of the geometry of the underlying signals and their interaction. It is shown that if the subspaces spanned by the innovation components of the source signals with respect to the side information signals have zero intersection, provided that we observe a certain number of linear measurements from the mixture, then we can reliably separate the sources; otherwise we cannot. The second algorithms is based on deep learning where we introduce a novel self-supervised algorithm for the source separation problem. Source separation is intrinsically unsupervised and the lack of training data makes it a difficult task for artificial intelligence to solve. The proposed framework takes advantage of the available data and delivers near perfect separation results in real data scenarios. Our proposed frameworks – which provide new ways to incorporate side information to aid the solution of the source separation problem – are also employed in a real-world art investigation application involving the separation of mixtures of X-Ray images. The simulation results showcase the superiority of our algorithm against other state-of-the-art algorithms

    Information Analysis for Steganography and Steganalysis in 3D Polygonal Meshes

    Get PDF
    Information hiding, which embeds a watermark/message over a cover signal, has recently found extensive applications in, for example, copyright protection, content authentication and covert communication. It has been widely considered as an appealing technology to complement conventional cryptographic processes in the field of multimedia security by embedding information into the signal being protected. Generally, information hiding can be classified into two categories: steganography and watermarking. While steganography attempts to embed as much information as possible into a cover signal, watermarking tries to emphasize the robustness of the embedded information at the expense of embedding capacity. In contrast to information hiding, steganalysis aims at detecting whether a given medium has hidden message in it, and, if possible, recover that hidden message. It can be used to measure the security performance of information hiding techniques, meaning a steganalysis resistant steganographic/watermarking method should be imperceptible not only to Human Vision Systems (HVS), but also to intelligent analysis. As yet, 3D information hiding and steganalysis has received relatively less attention compared to image information hiding, despite the proliferation of 3D computer graphics models which are fairly promising information carriers. This thesis focuses on this relatively neglected research area and has the following primary objectives: 1) to investigate the trade-off between embedding capacity and distortion by considering the correlation between spatial and normal/curvature noise in triangle meshes; 2) to design satisfactory 3D steganographic algorithms, taking into account this trade-off; 3) to design robust 3D watermarking algorithms; 4) to propose a steganalysis framework for detecting the existence of the hidden information in 3D models and introduce a universal 3D steganalytic method under this framework. %and demonstrate the performance of the proposed steganalysis by testing it against six well-known 3D steganographic/watermarking methods. The thesis is organized as follows. Chapter 1 describes in detail the background relating to information hiding and steganalysis, as well as the research problems this thesis will be studying. Chapter 2 conducts a survey on the previous information hiding techniques for digital images, 3D models and other medium and also on image steganalysis algorithms. Motivated by the observation that the knowledge of the spatial accuracy of the mesh vertices does not easily translate into information related to the accuracy of other visually important mesh attributes such as normals, Chapters 3 and 4 investigate the impact of modifying vertex coordinates of 3D triangle models on the mesh normals. Chapter 3 presents the results of an empirical investigation, whereas Chapter 4 presents the results of a theoretical study. Based on these results, a high-capacity 3D steganographic algorithm capable of controlling embedding distortion is also presented in Chapter 4. In addition to normal information, several mesh interrogation, processing and rendering algorithms make direct or indirect use of curvature information. Motivated by this, Chapter 5 studies the relation between Discrete Gaussian Curvature (DGC) degradation and vertex coordinate modifications. Chapter 6 proposes a robust watermarking algorithm for 3D polygonal models, based on modifying the histogram of the distances from the model vertices to a point in 3D space. That point is determined by applying Principal Component Analysis (PCA) to the cover model. The use of PCA makes the watermarking method robust against common 3D operations, such as rotation, translation and vertex reordering. In addition, Chapter 6 develops a 3D specific steganalytic algorithm to detect the existence of the hidden messages embedded by one well-known watermarking method. By contrast, the focus of Chapter 7 will be on developing a 3D watermarking algorithm that is resistant to mesh editing or deformation attacks that change the global shape of the mesh. By adopting a framework which has been successfully developed for image steganalysis, Chapter 8 designs a 3D steganalysis method to detect the existence of messages hidden in 3D models with existing steganographic and watermarking algorithms. The efficiency of this steganalytic algorithm has been evaluated on five state-of-the-art 3D watermarking/steganographic methods. Moreover, being a universal steganalytic algorithm can be used as a benchmark for measuring the anti-steganalysis performance of other existing and most importantly future watermarking/steganographic algorithms. Chapter 9 concludes this thesis and also suggests some potential directions for future work

    Đ˜ĐĐąĐ•Đ›Đ›Đ•ĐšĐąĐŁĐĐ›ĐŹĐĐ«Đ™ Ń‡ĐžŃĐ»ĐŸĐČŃ‹ĐŒ ĐżŃ€ĐŸĐłŃ€Đ°ĐŒĐŒĐœŃ‹ĐŒ ДЛЯ MIMD-ĐșĐŸĐŒĐżŃŒŃŽŃ‚Đ”Ń€

    Get PDF
    For most scientific and engineering problems simulated on computers the solving of problems of the computational mathematics with approximately given initial data constitutes an intermediate or a final stage. Basic problems of the computational mathematics include the investigating and solving of linear algebraic systems, evaluating of eigenvalues and eigenvectors of matrices, the solving of systems of non-linear equations, numerical integration of initial- value problems for systems of ordinary differential equations.Đ”Đ»Ń Đ±Ń–Đ»ŃŒŃˆĐŸŃŃ‚Ń– ĐœĐ°ŃƒĐșĐŸĐČох та Ń–ĐœĐ¶Đ”ĐœĐ”Ń€ĐœĐžŃ… заЎач ĐŒĐŸĐŽĐ”Đ»ŃŽĐČĐ°ĐœĐœŃ ĐœĐ° ЕОМ Ń€Ń–ŃˆĐ”ĐœĐœŃ заЎач ĐŸĐ±Ń‡ĐžŃĐ»ŃŽĐČĐ°Đ»ŃŒĐœĐŸŃ— ĐŒĐ°Ń‚Đ”ĐŒĐ°Ń‚ĐžĐșĐž Đ· ĐœĐ°Đ±Đ»ĐžĐ¶Đ”ĐœĐŸ Đ·Đ°ĐŽĐ°ĐœĐžĐŒĐž ĐČĐžŃ…Ń–ĐŽĐœĐžĐŒĐž ĐŽĐ°ĐœĐžĐŒĐž сĐșлаЎає ĐżŃ€ĐŸĐŒŃ–Đ¶ĐœĐžĐč Đ°Đ±ĐŸ ĐŸŃŃ‚Đ°Ń‚ĐŸŃ‡ĐœĐžĐč Дтап. ĐžŃĐœĐŸĐČĐœŃ– ĐżŃ€ĐŸĐ±Đ»Đ”ĐŒĐž ĐŸĐ±Ń‡ĐžŃĐ»ŃŽĐČĐ°Đ»ŃŒĐœĐŸŃ— ĐŒĐ°Ń‚Đ”ĐŒĐ°Ń‚ĐžĐșĐž ĐČŃ–ĐŽĐœĐŸŃŃŃ‚ŃŒŃŃ ĐŽĐŸŃĐ»Ń–ĐŽĐ¶Đ”ĐœĐœŃ і Ń€Ń–ŃˆĐ”ĐœĐœŃ Đ»Ń–ĐœŃ–ĐčĐœĐžŃ… Đ°Đ»ĐłĐ”Đ±Ń€Đ°Ń—Ń‡ĐœĐžŃ… ŃĐžŃŃ‚Đ”ĐŒ ĐŸŃ†Ń–ĐœĐșĐž ĐČĐ»Đ°ŃĐœĐžŃ… Đ·ĐœĐ°Ń‡Đ”ĐœŃŒ і ĐČĐ»Đ°ŃĐœĐžŃ… ĐČĐ”ĐșŃ‚ĐŸŃ€Ń–ĐČ ĐŒĐ°Ń‚Ń€ĐžŃ†ŃŒ, Ń€Ń–ŃˆĐ”ĐœĐœŃ ŃĐžŃŃ‚Đ”ĐŒ ĐœĐ”Đ»Ń–ĐœŃ–ĐčĐœĐžŃ… ріĐČĐœŃĐœŃŒ, Ń‡ĐžŃĐ”Đ»ŃŒĐœĐŸĐłĐŸ Ń–ĐœŃ‚Đ”ĐłŃ€ŃƒĐČĐ°ĐœĐœŃ ĐżĐŸŃ‡Đ°Ń‚ĐșĐŸĐČĐŸ заЎач ĐŽĐ»Ń ŃĐžŃŃ‚Đ”ĐŒ Đ·ĐČочаĐčĐœĐžŃ… ĐŽĐžŃ„Đ”Ń€Đ”ĐœŃ†Ń–Đ°Đ»ŃŒĐœĐžŃ… ріĐČĐœŃĐœŃŒ.Đ”Đ»Ń Đ±ĐŸĐ»ŃŒŃˆĐžĐœŃŃ‚ĐČĐ° ĐœĐ°ŃƒŃ‡ĐœŃ‹Ń… Đž ĐžĐœĐ¶Đ”ĐœĐ”Ń€ĐœŃ‹Ń… заЎач ĐŒĐŸĐŽĐ”Đ»ĐžŃ€ĐŸĐČĐ°ĐœĐžŃ ĐœĐ° ЭВМ Ń€Đ”ŃˆĐ”ĐœĐžĐ” заЎач ĐČŃ‹Ń‡ĐžŃĐ»ĐžŃ‚Đ”Đ»ŃŒĐœĐŸĐč ĐŒĐ°Ń‚Đ”ĐŒĐ°Ń‚ĐžĐșĐž с ĐżŃ€ĐžĐ±Đ»ĐžĐ¶Đ”ĐœĐœĐŸ Đ·Đ°ĐŽĐ°ĐœĐœŃ‹ĐŒ ĐžŃŃ…ĐŸĐŽĐœŃ‹ĐŒ ĐŽĐ°ĐœĐœŃ‹ĐŒ ŃĐŸŃŃ‚Đ°ĐČĐ»ŃĐ”Ń‚ ĐżŃ€ĐŸĐŒĐ”Đ¶ŃƒŃ‚ĐŸŃ‡ĐœŃ‹Đč ОлО ĐŸĐșĐŸĐœŃ‡Đ°Ń‚Đ”Đ»ŃŒĐœŃ‹Đč этап. ĐžŃĐœĐŸĐČĐœŃ‹Đ” ĐżŃ€ĐŸĐ±Đ»Đ”ĐŒŃ‹ ĐČŃ‹Ń‡ĐžŃĐ»ĐžŃ‚Đ”Đ»ŃŒĐœĐŸĐč ĐŒĐ°Ń‚Đ”ĐŒĐ°Ń‚ĐžĐșĐž ĐŸŃ‚ĐœĐŸŃŃŃ‚ŃŃ ĐžŃŃĐ»Đ”ĐŽĐŸĐČĐ°ĐœĐžŃ Đž Ń€Đ”ŃˆĐ”ĐœĐžŃ Đ»ĐžĐœĐ”ĐčĐœŃ‹Ń… алгДбраОчДсĐșох ŃĐžŃŃ‚Đ”ĐŒ ĐŸŃ†Đ”ĐœĐșĐž ŃĐŸĐ±ŃŃ‚ĐČĐ”ĐœĐœŃ‹Ń… Đ·ĐœĐ°Ń‡Đ”ĐœĐžĐč Đž ŃĐŸĐ±ŃŃ‚ĐČĐ”ĐœĐœŃ‹Ń… ĐČĐ”ĐșŃ‚ĐŸŃ€ĐŸĐČ ĐŒĐ°Ń‚Ń€ĐžŃ†, Ń€Đ”ŃˆĐ”ĐœĐžĐ” ŃĐžŃŃ‚Đ”ĐŒ ĐœĐ”Đ»ĐžĐœĐ”ĐčĐœŃ‹Ń… ураĐČĐœĐ”ĐœĐžĐč, Ń‡ĐžŃĐ»Đ”ĐœĐœĐŸĐłĐŸ ĐžĐœŃ‚Đ”ĐłŃ€ĐžŃ€ĐŸĐČĐ°ĐœĐžŃ ĐœĐ°Ń‡Đ°Đ»ŃŒĐœĐŸ заЎач ĐŽĐ»Ń ŃĐžŃŃ‚Đ”ĐŒ ĐŸĐ±Ń‹ĐșĐœĐŸĐČĐ”ĐœĐœŃ‹Ń… ĐŽĐžŃ„Ń„Đ”Ń€Đ”ĐœŃ†ĐžĐ°Đ»ŃŒĐœŃ‹Ń… ураĐČĐœĐ”ĐœĐžĐč

    Application and Theory of Multimedia Signal Processing Using Machine Learning or Advanced Methods

    Get PDF
    This Special Issue is a book composed by collecting documents published through peer review on the research of various advanced technologies related to applications and theories of signal processing for multimedia systems using ML or advanced methods. Multimedia signals include image, video, audio, character recognition and optimization of communication channels for networks. The specific contents included in this book are data hiding, encryption, object detection, image classification, and character recognition. Academics and colleagues who are interested in these topics will find it interesting to read

    Information embedding and retrieval in 3D printed objects

    Get PDF
    Deep learning and convolutional neural networks have become the main tools of computer vision. These techniques are good at using supervised learning to learn complex representations from data. In particular, under limited settings, the image recognition model now performs better than the human baseline. However, computer vision science aims to build machines that can see. It requires the model to be able to extract more valuable information from images and videos than recognition. Generally, it is much more challenging to apply these deep learning models from recognition to other problems in computer vision. This thesis presents end-to-end deep learning architectures for a new computer vision field: watermark retrieval from 3D printed objects. As it is a new area, there is no state-of-the-art on many challenging benchmarks. Hence, we first define the problems and introduce the traditional approach, Local Binary Pattern method, to set our baseline for further study. Our neural networks seem useful but straightfor- ward, which outperform traditional approaches. What is more, these networks have good generalization. However, because our research field is new, the problems we face are not only various unpredictable parameters but also limited and low-quality training data. To address this, we make two observations: (i) we do not need to learn everything from scratch, we know a lot about the image segmentation area, and (ii) we cannot know everything from data, our models should be aware what key features they should learn. This thesis explores these ideas and even explore more. We show how to use end-to-end deep learning models to learn to retrieve watermark bumps and tackle covariates from a few training images data. Secondly, we introduce ideas from synthetic image data and domain randomization to augment training data and understand various covariates that may affect retrieve real-world 3D watermark bumps. We also show how the illumination in synthetic images data to effect and even improve retrieval accuracy for real-world recognization applications

    Multibiometric security in wireless communication systems

    Get PDF
    This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University, 05/08/2010.This thesis has aimed to explore an application of Multibiometrics to secured wireless communications. The medium of study for this purpose included Wi-Fi, 3G, and WiMAX, over which simulations and experimental studies were carried out to assess the performance. In specific, restriction of access to authorized users only is provided by a technique referred to hereafter as multibiometric cryptosystem. In brief, the system is built upon a complete challenge/response methodology in order to obtain a high level of security on the basis of user identification by fingerprint and further confirmation by verification of the user through text-dependent speaker recognition. First is the enrolment phase by which the database of watermarked fingerprints with memorable texts along with the voice features, based on the same texts, is created by sending them to the server through wireless channel. Later is the verification stage at which claimed users, ones who claim are genuine, are verified against the database, and it consists of five steps. Initially faced by the identification level, one is asked to first present one’s fingerprint and a memorable word, former is watermarked into latter, in order for system to authenticate the fingerprint and verify the validity of it by retrieving the challenge for accepted user. The following three steps then involve speaker recognition including the user responding to the challenge by text-dependent voice, server authenticating the response, and finally server accepting/rejecting the user. In order to implement fingerprint watermarking, i.e. incorporating the memorable word as a watermark message into the fingerprint image, an algorithm of five steps has been developed. The first three novel steps having to do with the fingerprint image enhancement (CLAHE with 'Clip Limit', standard deviation analysis and sliding neighborhood) have been followed with further two steps for embedding, and extracting the watermark into the enhanced fingerprint image utilising Discrete Wavelet Transform (DWT). In the speaker recognition stage, the limitations of this technique in wireless communication have been addressed by sending voice feature (cepstral coefficients) instead of raw sample. This scheme is to reap the advantages of reducing the transmission time and dependency of the data on communication channel, together with no loss of packet. Finally, the obtained results have verified the claims
    • 

    corecore