98 research outputs found

    Indexing Audio-Visual Sequences by Joint Audio and Video Processing

    Get PDF
    The focus of this work is oriented to the creation of a content-based hierarchical organisation of audio-visual data (a description scheme) and to the creation of meta-data (descriptors) to associate with audio and/or visual signals. The generation of efficient indices to access audio-visual databases is strictly connected to the generation of content descriptors and to the hierarchical representation of audio-visual material. Once a hierarchy can be extracted from the data analysis, a nested indexing structure can be created to access relevant information at a specific level of detail. Accordingly, a query can be made very specific in relationship to the level of detail that is required by the user. In order to construct the hierarchy, we describe how to extract information content from audio-visual sequences so as to have different hierarchical indicators (or descriptors), which can be associated to each media (audio, video). At this stage, video and audio signals can be separated into temporally consistent elements. At the lowest level, information is organised in frames (groups of pixels for visual information, groups of consecutive samples for audio information). At a higher level, low-level consistent temporal entities are identified: in case of digital image sequences, these consist of shots (or continuous camera records) which can be obtained by detecting cuts or special effects such as dissolves, fade in and fade out; in case of audio information, these represent consistent audio segments belonging to one specific audio type (such as speech, music, silence, ...). One more level up, patterns of video shots or audio segments can be recognised so as to reflect more meaningful structures such as dialogues, actions, ... At the highest level, information is organised so as to establish correlation beyond the temporal organisation of information, allowing to reflect classes of visual or audio types: we call these classes idioms. The paper ends with a description of possible solutions to allow a cross-modal analysis of audio and video information, which may validate or invalidate the proposed hierarchy, and in some cases enable more sophisticated levels of representation of information content

    Query by Example of Speaker Audio Signals using Power Spectrum and MFCCs

    Get PDF
    Search engine is the popular term for an information retrieval (IR) system. Typically, search engine can be based on full-text indexing. Changing the presentation from the text data to multimedia data types make an information retrieval process more complex such as a retrieval of image or sounds in large databases. This paper introduces the use of language and text independent speech as input queries in a large sound database by using Speaker identification algorithm. The method consists of 2 main processing first steps, we separate vocal and non-vocal identification after that vocal be used to speaker identification for audio query by speaker voice. For the speaker identification and audio query by process, we estimate the similarity of the example signal and the samples in the queried database by calculating the Euclidian distance between the Mel frequency cepstral coefficients (MFCC) and Energy spectrum of acoustic features. The simulations show that the good performance with a sustainable computational cost and obtained the average accuracy rate more than 90%

    Robust Sound Event Classification using Deep Neural Networks

    Get PDF
    The automatic recognition of sound events by computers is an important aspect of emerging applications such as automated surveillance, machine hearing and auditory scene understanding. Recent advances in machine learning, as well as in computational models of the human auditory system, have contributed to advances in this increasingly popular research field. Robust sound event classification, the ability to recognise sounds under real-world noisy conditions, is an especially challenging task. Classification methods translated from the speech recognition domain, using features such as mel-frequency cepstral coefficients, have been shown to perform reasonably well for the sound event classification task, although spectrogram-based or auditory image analysis techniques reportedly achieve superior performance in noise. This paper outlines a sound event classification framework that compares auditory image front end features with spectrogram image-based front end features, using support vector machine and deep neural network classifiers. Performance is evaluated on a standard robust classification task in different levels of corrupting noise, and with several system enhancements, and shown to compare very well with current state-of-the-art classification techniques

    Image segmentation in the wavelet domain using N-cut framework

    Full text link
    We introduce a wavelet domain image segmentation algorithm based on Normalized Cut (NCut) framework in this thesis. By employing the NCut algorithm we solve the perceptual grouping problem of image segmentation which aims at the extraction of the global impression of an image. We capitalize on the reduced set of data to be processed and statistical features derived from the wavelet-transformed images to solve graph partitioning more efficiently than before. Five orientation histograms are computed to evaluate similarity/dissimilarity measure of local structure. We use properties of the wavelet transform filtering to capture edge information in vertical, horizontal and diagonal orientations. This approach allows for direct processing of compressed data and results in faster implementation of NCut framework than that in the spatial domain and also decent quality of segmentation of natural scene images

    The Space and Earth Science Data Compression Workshop

    Get PDF
    This document is the proceedings from a Space and Earth Science Data Compression Workshop, which was held on March 27, 1992, at the Snowbird Conference Center in Snowbird, Utah. This workshop was held in conjunction with the 1992 Data Compression Conference (DCC '92), which was held at the same location, March 24-26, 1992. The workshop explored opportunities for data compression to enhance the collection and analysis of space and Earth science data. The workshop consisted of eleven papers presented in four sessions. These papers describe research that is integrated into, or has the potential of being integrated into, a particular space and/or Earth science data information system. Presenters were encouraged to take into account the scientists's data requirements, and the constraints imposed by the data collection, transmission, distribution, and archival system

    Texture representation using wavelet filterbanks

    Get PDF
    Texture analysis is a fundamental issue in image analysis and computer vision. While considerable research has been carried out in the texture analysis domain, problems relating to texture representation have been addressed only partially and active research is continuing. The vast majority of algorithms for texture analysis make either an explicit or implicit assumption that all images are captured under the same measurement conditions, such as orientation and illumination. These assumptions are often unrealistic in many practical applications;This dissertation addresses the viewpoint-invariance problem in texture classification by introducing a rotated wavelet filterbank. The proposed filterbank, in conjunction with a standard wavelet filterbank, provides better freedom of orientation tuning for texture analysis. This allows one to obtain texture features that are invariant with respect to texture rotation and linear grayscale transformation. In this study, energy estimates of channel outputs that are commonly used as texture features in texture classification are transformed into a set of viewpoint-invariant features. Texture properties that have a physical connection with human perception are taken into account in the transformation of the energy estimates;Experiments using natural texture image sets that have been used for evaluating other successful approaches were conducted in order to facilitate comparison. We observe that the proposed feature set outperformed methods proposed by others in the past. A channel selection method is also proposed to minimize the computational complexity and improve performance in a texture segmentation algorithm. Results demonstrating the validity of the approach are presented using experimental ultrasound tendon images

    High-performance compression of visual information - A tutorial review - Part I : Still Pictures

    Get PDF
    Digital images have become an important source of information in the modern world of communication systems. In their raw form, digital images require a tremendous amount of memory. Many research efforts have been devoted to the problem of image compression in the last two decades. Two different compression categories must be distinguished: lossless and lossy. Lossless compression is achieved if no distortion is introduced in the coded image. Applications requiring this type of compression include medical imaging and satellite photography. For applications such as video telephony or multimedia applications, some loss of information is usually tolerated in exchange for a high compression ratio. In this two-part paper, the major building blocks of image coding schemes are overviewed. Part I covers still image coding, and Part II covers motion picture sequences. In this first part, still image coding schemes have been classified into predictive, block transform, and multiresolution approaches. Predictive methods are suited to lossless and low-compression applications. Transform-based coding schemes achieve higher compression ratios for lossy compression but suffer from blocking artifacts at high-compression ratios. Multiresolution approaches are suited for lossy as well for lossless compression. At lossy high-compression ratios, the typical artifact visible in the reconstructed images is the ringing effect. New applications in a multimedia environment drove the need for new functionalities of the image coding schemes. For that purpose, second-generation coding techniques segment the image into semantically meaningful parts. Therefore, parts of these methods have been adapted to work for arbitrarily shaped regions. In order to add another functionality, such as progressive transmission of the information, specific quantization algorithms must be defined. A final step in the compression scheme is achieved by the codeword assignment. Finally, coding results are presented which compare stateof- the-art techniques for lossy and lossless compression. The different artifacts of each technique are highlighted and discussed. Also, the possibility of progressive transmission is illustrated

    The 1993 Space and Earth Science Data Compression Workshop

    Get PDF
    The Earth Observing System Data and Information System (EOSDIS) is described in terms of its data volume, data rate, and data distribution requirements. Opportunities for data compression in EOSDIS are discussed

    Iris Identification using Keypoint Descriptors and Geometric Hashing

    Get PDF
    Iris is one of the most reliable biometric trait due to its stability and randomness. Conventional recognition systems transform the iris to polar coordinates and perform well for co-operative databases. However, the problem aggravates to manifold for recognizing non-cooperative irises. In addition, the transformation of iris to polar domain introduces aliasing effect. In this thesis, the aforementioned issues are addressed by considering Noise Independent Annular Iris for feature extraction. Global feature extraction approaches are rendered as unsuitable for annular iris due to change in scale as they could not achieve invariance to ransformation and illumination. On the contrary, local features are invariant to image scaling, rotation and partially invariant to change in illumination and viewpoint. To extract local features, Harris Corner Points are detected from iris and matched using novel Dual stage approach. Harris corner improves accuracy but fails to achieve scale invariance. Further, Scale Invariant Feature Transform (SIFT) has been applied to annular iris and results are found to be very promising. However, SIFT is computationally expensive for recognition due to higher dimensional descriptor. Thus, a recently evolved keypoint descriptor called Speeded Up Robust Features (SURF) is applied to mark performance improvement in terms of time as well as accuracy. For identification, retrieval time plays a significant role in addition to accuracy. Traditional indexing approaches cannot be applied to biometrics as data are unstructured. In this thesis, two novel approaches has been developed for indexing iris database. In the first approach, Energy Histogram of DCT coefficients is used to form a B-tree. This approach performs well for cooperative databases. In the second approach, indexing is done using Geometric Hashing of SIFT keypoints. The latter indexing approach achieves invariance to similarity transformations, illumination and occlusion and performs with an accuracy of more than 98% for cooperative as well as non-cooperative databases

    Automatic stereoscopic video object-based watermarking using qualified significant wavelet trees

    Get PDF
    In this paper a fully automatic scheme for embedding visually recognizable watermark patterns to video objects is proposed. The architecture consists of 3 main modules. During the first module unsupervised video object extraction is performed, by analyzing stereoscopic pairs of frames. In the second module each video object is decomposed into three levels with ten subbands, using the Shape Adaptive Discrete Wavelet Transform (SA-DWT) and three pairs of subbands are formed (HL3 , HL2), (LH3, LH2) and (HH3, HH2). Next Qualified Significant Wavelet Trees (QSWTs) are estimated for the specific pair of subbands with the highest energy content. QSWTs are derived from the Embedded Zerotree Wavelet (EZW) algorithm and they are high-energy paths of wavelet coefficients. Finally during the third module, visually recognizable watermark patterns are redundantly embedded to the coefficients of the highest energy QSWTs and the inverse SA-DWT is applied to provide the watermarked video object. Performance of the proposed video object watermarking system is tested under various signal distortions such as JPEG lossy compression, sharpening, blurring and adding different types of noise. Furthermore the case of transmission losses for the watermarked video objects is also investigated. Experimental results on real life video objects indicate the efficiency and robustness of the proposed schemeFacultad de Informátic
    corecore