110 research outputs found

    Modeling and frequency tracking of marine mammal whistle calls

    Get PDF
    Submitted in partial fulfillment of the requirements for the degree of Master of Science at the Massachusetts Institute of Technology and the Woods Hole Oceanographic Institution February 2009Marine mammal whistle calls present an attractive medium for covert underwater communications. High quality models of the whistle calls are needed in order to synthesize natural-sounding whistles with embedded information. Since the whistle calls are composed of frequency modulated harmonic tones, they are best modeled as a weighted superposition of harmonically related sinusoids. Previous research with bottlenose dolphin whistle calls has produced synthetic whistles that sound too “clean” for use in a covert communications system. Due to the sensitivity of the human auditory system, watermarking schemes that slightly modify the fundamental frequency contour have good potential for producing natural-sounding whistles embedded with retrievable watermarks. Structured total least squares is used with linear prediction analysis to track the time-varying fundamental frequency and harmonic amplitude contours throughout a whistle call. Simulation and experimental results demonstrate the capability to accurately model bottlenose dolphin whistle calls and retrieve embedded information from watermarked synthetic whistle calls. Different fundamental frequency watermarking schemes are proposed based on their ability to produce natural sounding synthetic whistles and yield suitable watermark detection and retrieval

    Extraction of vocal-tract system characteristics from speechsignals

    Get PDF
    We propose methods to track natural variations in the characteristics of the vocal-tract system from speech signals. We are especially interested in the cases where these characteristics vary over time, as happens in dynamic sounds such as consonant-vowel transitions. We show that the selection of appropriate analysis segments is crucial in these methods, and we propose a selection based on estimated instants of significant excitation. These instants are obtained by a method based on the average group-delay property of minimum-phase signals. In voiced speech, they correspond to the instants of glottal closure. The vocal-tract system is characterized by its formant parameters, which are extracted from the analysis segments. Because the segments are always at the same relative position in each pitch period, in voiced speech the extracted formants are consistent across successive pitch periods. We demonstrate the results of the analysis for several difficult cases of speech signals

    A unified approach to sparse signal processing

    Get PDF
    A unified view of the area of sparse signal processing is presented in tutorial form by bringing together various fields in which the property of sparsity has been successfully exploited. For each of these fields, various algorithms and techniques, which have been developed to leverage sparsity, are described succinctly. The common potential benefits of significant reduction in sampling rate and processing manipulations through sparse signal processing are revealed. The key application domains of sparse signal processing are sampling, coding, spectral estimation, array processing, compo-nent analysis, and multipath channel estimation. In terms of the sampling process and reconstruction algorithms, linkages are made with random sampling, compressed sensing and rate of innovation. The redundancy introduced by channel coding i

    Zolotarev polynomials utilization in spectral analysis

    Get PDF
    Tato práce je zaměřena na vybrané problémy Zolotarevových polynomů a jejich vyuľití ke spektrální analýze. Pokud jde o Zolotarevovy polynomy, jsou popsány základní vlastnosti symetrických Zolotarevových polynomů včetně ortogonality. Rovněľ se provádí prozkoumání numerických vlastností algoritmů generujících dokonce Zolotarevovy polynomy. Pokud jde o aplikaci Zolotarevových polynomů na spektrální analýzu, je implementována aproximovaná diskrétní Zolotarevova transformace, která umoľňuje výpočet spektrogramu (zologramu) v reálném čase. Aproximovaná diskrétní zolotarevská transformace je navíc upravena tak, aby lépe fungovala při analýze tlumených exponenciálních signálů. A nakonec je navrľena nová diskrétní Zolotarevova transformace implementovaná plně v časové oblasti. Tato transformace také ukazuje, ľe některé rysy pozorované u aproximované diskrétní Zolotarevovy transformace jsou důsledkem pouľití Zolotarevových polynomů.This thesis is focused on selected problems of symmetrical Zolotarev polynomials and their use in spectral analysis. Basic properties of symmetrical Zolotarev polynomials including orthogonality are described. Also, the exploration of numerical properties of algorithms generating even Zolotarev polynomials is performed. As regards to the application of Zolotarev polynomials to spectral analysis the Approximated Discrete Zolotarev Transform is implemented so that it enables computing of zologram in real–time. Moreover, the Approximated Discrete Zolotarev Transform is modified to perform better in the analysis of damped exponential signals. And finally, a novel Discrete Zolotarev Transform implemented fully in the time domain is suggested. This transform also shows that some features observed using the Approximated Discrete Zolotarev Transform are a consequence of using Zolotarev polynomials

    ИНТЕЛЛЕКТУАЛЬНЫЙ числовым программным ДЛЯ MIMD-компьютер

    Get PDF
    For most scientific and engineering problems simulated on computers the solving of problems of the computational mathematics with approximately given initial data constitutes an intermediate or a final stage. Basic problems of the computational mathematics include the investigating and solving of linear algebraic systems, evaluating of eigenvalues and eigenvectors of matrices, the solving of systems of non-linear equations, numerical integration of initial- value problems for systems of ordinary differential equations.Для більшості наукових та інженерних задач моделювання на ЕОМ рішення задач обчислювальної математики з наближено заданими вихідними даними складає проміжний або остаточний етап. Основні проблеми обчислювальної математики відносяться дослідження і рішення лінійних алгебраїчних систем оцінки власних значень і власних векторів матриць, рішення систем нелінійних рівнянь, чисельного інтегрування початково задач для систем звичайних диференціальних рівнянь.Для большинства научных и инженерных задач моделирования на ЭВМ решение задач вычислительной математики с приближенно заданным исходным данным составляет промежуточный или окончательный этап. Основные проблемы вычислительной математики относятся исследования и решения линейных алгебраических систем оценки собственных значений и собственных векторов матриц, решение систем нелинейных уравнений, численного интегрирования начально задач для систем обыкновенных дифференциальных уравнений

    Comparison of modelled pursuits with ESPRIT and the matrix pencil method in the modelling of medical percussion signals

    Get PDF
    The objective of this paper is to compare Modelled Pursuits (MoP), a recently developed iterative signal decomposition method, with more established matrix based subspace methods used to aid or automate medical percussion diagnoses. Medical percussion is a technique used by clinicians to aid the diagnosis of pulmonary disease. It requires considerable expertise, so it is desirable to automate this process where possible. Previous work has examined the application of modal decomposition techniques, since medical percussion signals (MPS) can be intuitively characterised as combinations of exponentially decaying sinusoidal (EDS) vibrations. Best results have typically been reported with matrix based subspace methods such as Estimation of Signal Parameters via Rotational Invariance Techniques (ESPRIT) and the Matrix Pencil Method (MPM). Since ESPRIT and MPM are computationally expensive, this paper investigates whether an iterative method such as MoP can produce similar results with less computation and/or memory overheads. Using randomly generated synthetic signals designed to replicate typical ‘tympanic’ and ‘resonant’ percussion signals, we compared each method: MoP, ESPRIT, and MPM, for accuracy, speed and memory usage. We find that for low Signal to Noise Ratios (SNRs) MoP gives less accuracy than both ESPRIT and MPM, however for high SNRs (as would be typically encountered in a clinical setting) it is more accurate than MPM but less accurate than ESPRIT. We conclude that in embedded clinical applications where both operations-per-second and memory-usage are a factor, MoP is less computationally intensive than ESPRIT and thus is worth considering for use in those contexts

    Automatic music genre classification

    Get PDF
    A dissertation submitted to the Faculty of Science, University of the Witwatersrand, in fulfillment of the requirements for the degree of Master of Science. 2014.No abstract provided

    Multi modal multi-semantic image retrieval

    Get PDF
    PhDThe rapid growth in the volume of visual information, e.g. image, and video can overwhelm users’ ability to find and access the specific visual information of interest to them. In recent years, ontology knowledge-based (KB) image information retrieval techniques have been adopted into in order to attempt to extract knowledge from these images, enhancing the retrieval performance. A KB framework is presented to promote semi-automatic annotation and semantic image retrieval using multimodal cues (visual features and text captions). In addition, a hierarchical structure for the KB allows metadata to be shared that supports multi-semantics (polysemy) for concepts. The framework builds up an effective knowledge base pertaining to a domain specific image collection, e.g. sports, and is able to disambiguate and assign high level semantics to ‘unannotated’ images. Local feature analysis of visual content, namely using Scale Invariant Feature Transform (SIFT) descriptors, have been deployed in the ‘Bag of Visual Words’ model (BVW) as an effective method to represent visual content information and to enhance its classification and retrieval. Local features are more useful than global features, e.g. colour, shape or texture, as they are invariant to image scale, orientation and camera angle. An innovative approach is proposed for the representation, annotation and retrieval of visual content using a hybrid technique based upon the use of an unstructured visual word and upon a (structured) hierarchical ontology KB model. The structural model facilitates the disambiguation of unstructured visual words and a more effective classification of visual content, compared to a vector space model, through exploiting local conceptual structures and their relationships. The key contributions of this framework in using local features for image representation include: first, a method to generate visual words using the semantic local adaptive clustering (SLAC) algorithm which takes term weight and spatial locations of keypoints into account. Consequently, the semantic information is preserved. Second a technique is used to detect the domain specific ‘non-informative visual words’ which are ineffective at representing the content of visual data and degrade its categorisation ability. Third, a method to combine an ontology model with xi a visual word model to resolve synonym (visual heterogeneity) and polysemy problems, is proposed. The experimental results show that this approach can discover semantically meaningful visual content descriptions and recognise specific events, e.g., sports events, depicted in images efficiently. Since discovering the semantics of an image is an extremely challenging problem, one promising approach to enhance visual content interpretation is to use any associated textual information that accompanies an image, as a cue to predict the meaning of an image, by transforming this textual information into a structured annotation for an image e.g. using XML, RDF, OWL or MPEG-7. Although, text and image are distinct types of information representation and modality, there are some strong, invariant, implicit, connections between images and any accompanying text information. Semantic analysis of image captions can be used by image retrieval systems to retrieve selected images more precisely. To do this, a Natural Language Processing (NLP) is exploited firstly in order to extract concepts from image captions. Next, an ontology-based knowledge model is deployed in order to resolve natural language ambiguities. To deal with the accompanying text information, two methods to extract knowledge from textual information have been proposed. First, metadata can be extracted automatically from text captions and restructured with respect to a semantic model. Second, the use of LSI in relation to a domain-specific ontology-based knowledge model enables the combined framework to tolerate ambiguities and variations (incompleteness) of metadata. The use of the ontology-based knowledge model allows the system to find indirectly relevant concepts in image captions and thus leverage these to represent the semantics of images at a higher level. Experimental results show that the proposed framework significantly enhances image retrieval and leads to narrowing of the semantic gap between lower level machinederived and higher level human-understandable conceptualisation
    corecore