7,741 research outputs found

    Band-pass filtering of the time sequences of spectral parameters for robust wireless speech recognition

    Get PDF
    In this paper we address the problem of automatic speech recognition when wireless speech communication systems are involved. In this context, three main sources of distortion should be considered: acoustic environment, speech coding and transmission errors. Whilst the first one has already received a lot of attention, the last two deserve further investigation in our opinion. We have found out that band-pass filtering of the recognition features improves ASR performance when distortions due to these particular communication systems are present. Furthermore, we have evaluated two alternative configurations at different bit error rates (BER) typical of these channels: band-pass filtering the LP-MFCC parameters or a modification of the RASTA-PLP using a sharper low-pass section perform consistently better than LP-MFCC and RASTA-PLP, respectively.Publicad

    Fingerprinting Smart Devices Through Embedded Acoustic Components

    Full text link
    The widespread use of smart devices gives rise to both security and privacy concerns. Fingerprinting smart devices can assist in authenticating physical devices, but it can also jeopardize privacy by allowing remote identification without user awareness. We propose a novel fingerprinting approach that uses the microphones and speakers of smart phones to uniquely identify an individual device. During fabrication, subtle imperfections arise in device microphones and speakers which induce anomalies in produced and received sounds. We exploit this observation to fingerprint smart devices through playback and recording of audio samples. We use audio-metric tools to analyze and explore different acoustic features and analyze their ability to successfully fingerprint smart devices. Our experiments show that it is even possible to fingerprint devices that have the same vendor and model; we were able to accurately distinguish over 93% of all recorded audio clips from 15 different units of the same model. Our study identifies the prominent acoustic features capable of fingerprinting devices with high success rate and examines the effect of background noise and other variables on fingerprinting accuracy

    Comparison of speech intelligibility over the telephone using a hearing aid micorphone and telecoil

    Get PDF
    The purpose of the current study is to determine if the smaller, modern hearing aid has affected the speech intelligibility over the telephone using a telecoil and hearing aid microphone. Six hearing impaired listeners were situated in a quiet office and were asked to repeat aloud Connected Speech Sentences (CST) they heard through the telephone while wearing their hearing aid in telecoil only mode, microphone only mode, and without their hearing aid. The CST sentences were presented in three different signal-to-noise ratios (Quiet, +10dB, & +5dB) with the recorded speech babble of the CST test. It was discovered that the listeners performed, on average, better in all listening conditions without their hearing aid in. However, no statistical significance was seen between any of the test conditions. Implications of this will be discussed

    Articulation rate in Swedish child-directed speech increases as a function of the age of the child even when surprisal is controlled for

    Get PDF
    In earlier work, we have shown that articulation rate in Swedish child-directed speech (CDS) increases as a function of the age of the child, even when utterance length and differences in articulation rate between subjects are controlled for. In this paper we show on utterance level in spontaneous Swedish speech that i) for the youngest children, articulation rate in CDS is lower than in adult-directed speech (ADS), ii) there is a significant negative correlation between articulation rate and surprisal (the negative log probability) in ADS, and iii) the increase in articulation rate in Swedish CDS as a function of the age of the child holds, even when surprisal along with utterance length and differences in articulation rate between speakers are controlled for. These results indicate that adults adjust their articulation rate to make it fit the linguistic capacity of the child.Comment: 5 pages, Interspeech 201

    Wireless sensor systems in indoor situation modeling II (WISM II)

    Get PDF
    fi=vertaisarvioimaton|en=nonPeerReviewed

    Assistive Listening Devices: A Guide

    Full text link
    Objective: The purpose of this research was to develop a guide on assistive listening devices (ALDs) describing the various types of ALDs, the basic underlying concepts, their advantages and disadvantages, the instrumentation and its components, and the setup and procedures for specification/evaluation of ALDs in accordance with national standards or guidelines issued by professional organizations in our field. This guide is intended for audiologists, hearing scientists, and audiology and hearing science students. Method: A thorough review of the previous ALD literature including national and international standards for set-up and installation, specification/evaluation and verification of ALDs; guidelines from professional audiology and acoustic and hearing sciences organizations for ALD set-up and installation, specification/evaluation and verification; peer-reviewed studies on ALDs; text-book chapters and books on ALDs; and ALD websites from professional organizations. Results: This guide was organized by ALD type, and was subcategorized by the basic underlying concepts, their advantages and disadvantages, the instrumentation and components, and the setup/installation and procedures for specification/evaluation and verification. A comparative analysis was also performed on the relative benefits of various ALDs in a real-word application setting. Discussion: This guide demonstrates that ALDs facilitate communicative efficiency in persons with hearing loss in adverse listening environments. Selection of an appropriate ALD should be based on the intended system use and the intended listening environment. Appropriately selected and fitted ALDs help individuals detect environmental sounds or improve their speech recognition in specific listening settings. Also, ALDs can enable higher levels of communicative performance would be obtained with just the use of individual hearing technology alone. Conclusion: The research findings demonstrate that ALDs improve audibility and overall listening benefit for individuals with hearing loss, especially those with compatible hearing technology. The guide can help one ensure optimal ALD performance to maximize communicative benefit; it serves as a resource for audiologists, hearing scientists, and audiology and hearing science students to develop a better understanding of topics related to ALDs; appropriate ALDs to recommend to persons with hearing loss for various listening situations; set-up and installation of ALDs; and evaluation and verification of ALD performanc

    A model and architecture for situation determination

    Get PDF
    Automatically determining the situation of an ad-hoc group of people and devices within a smart environment is a significant challenge in pervasive computing systems. Current approaches often rely on an environment expert to correlate the situations that occur with the available sensor data, while other machine learning based approaches require long training periods before the system can be used. Furthermore, situations are commonly recognised at a low-level of granularity, which limits the scope of situation-aware applications. This paper presents a novel approach to situation determination that attempts to overcome these issues by providing a reusable library of general situation specifications that can be easily extended to create new specific situations, and immediately deployed without the need of an environment expert. A proposed architecture of an accompanying situation determination middleware is provided, as well as an analysis of a prototype implementation

    Wireless Audio Interactive Knot

    Get PDF
    Thesis (S.M.)--Massachusetts Institute of Technology, School of Architecture and Planning, Program in Media Arts and Sciences, 2001.Includes bibliographical references (leaves 44-45).The Sound Transformer is a new type of musical instrument. It looks a little like a saxophone, but when you sing or "kazoo" into it, astonishing transforms and mutations come out. What actually happens is that the input sound is sent via 802.11 wireless link to a net server that transforms the sound and sends it back to the instrument's speaker. In other words, instead of a resonant acoustic body, or a local computer synthesizer, this architecture allows sound to be sourced or transformed by an infinite array of online services, and channeled through a gesturally expressive handheld. Emerging infrastructures (802.11, Bluetooth, 3G and 4G, etc) seem to aim at this new class of instrument. But can such an architecture really work? In particular, given the delays incurred by decoupling the sound transformation from the instrument over a wireless network, are interactive music applications feasible? My thesis is that they are. To prove this, I built a platform called WAI-KNOT (for Wireless Audio Interactive Knot) in order to examine the latency issues as well as other design elements, and test their viability and impact on real music making. The Sound Transformer is a WAI-KNOT application.Adam Douglas Smith.S.M
    • …
    corecore