22,991 research outputs found

    Bilateral cochlear implantation or bimodal listening in the paediatric population : retrospective analysis of decisive criteria

    Get PDF
    Introduction: In children with bilateral severe to profound hearing loss, bilateral hearing can be achieved by either bimodal stimulation (CIHA) or bilateral cochlear implantation (BICI). The aim of this study was to analyse the audiologic test protocol that is currently applied to make decisions regarding the bilateral hearing modality in the paediatric population. Methods: Pre- and postoperative audiologic test results of 21 CIHA, 19 sequential BICI and 12 simultaneous BICI children were examined retrospectively. Results: Deciding between either simultaneous BICI or unilateral implantation was mainly based on the infant's preoperative Auditory Brainstem Response thresholds. Evolution from CIHA to sequential BICI was mainly based on the audiometric test results in the contralateral (hearing aid) ear after unilateral cochlear implantation. Preoperative audiometric thresholds in the hearing aid ear were significantly better in CIHA versus sequential BICI children (p < 0.001 and p = 0.001 in unaided and aided condition, respectively). Decisive values obtained in the hearing aid ear in favour of BICI were: An average hearing threshold measured at 0.5, 1, 2 and 4 kHz of at least 93 dB HL without, and at least 52 dB HL with hearing aid together with a 40% aided speech recognition score and a 70% aided score on the phoneme discrimination subtest of the Auditory Speech Sounds Evaluation test battery. Conclusions: Although pure tone audiometry offers no information about bimodal benefit, it remains the most obvious audiometric evaluation in the decision process on the mode of bilateral stimulation in the paediatric population. A theoretical test protocol for adequate evaluation of bimodal benefit in the paediatric population is proposed

    Audio Source Separation Using Sparse Representations

    Get PDF
    This is the author's final version of the article, first published as A. Nesbit, M. G. Jafari, E. Vincent and M. D. Plumbley. Audio Source Separation Using Sparse Representations. In W. Wang (Ed), Machine Audition: Principles, Algorithms and Systems. Chapter 10, pp. 246-264. IGI Global, 2011. ISBN 978-1-61520-919-4. DOI: 10.4018/978-1-61520-919-4.ch010file: NesbitJafariVincentP11-audio.pdf:n\NesbitJafariVincentP11-audio.pdf:PDF owner: markp timestamp: 2011.02.04file: NesbitJafariVincentP11-audio.pdf:n\NesbitJafariVincentP11-audio.pdf:PDF owner: markp timestamp: 2011.02.04The authors address the problem of audio source separation, namely, the recovery of audio signals from recordings of mixtures of those signals. The sparse component analysis framework is a powerful method for achieving this. Sparse orthogonal transforms, in which only few transform coefficients differ significantly from zero, are developed; once the signal has been transformed, energy is apportioned from each transform coefficient to each estimated source, and, finally, the signal is reconstructed using the inverse transform. The overriding aim of this chapter is to demonstrate how this framework, as exemplified here by two different decomposition methods which adapt to the signal to represent it sparsely, can be used to solve different problems in different mixing scenarios. To address the instantaneous (neither delays nor echoes) and underdetermined (more sources than mixtures) mixing model, a lapped orthogonal transform is adapted to the signal by selecting a basis from a library of predetermined bases. This method is highly related to the windowing methods used in the MPEG audio coding framework. In considering the anechoic (delays but no echoes) and determined (equal number of sources and mixtures) mixing case, a greedy adaptive transform is used based on orthogonal basis functions that are learned from the observed data, instead of being selected from a predetermined library of bases. This is found to encode the signal characteristics, by introducing a feedback system between the bases and the observed data. Experiments on mixtures of speech and music signals demonstrate that these methods give good signal approximations and separation performance, and indicate promising directions for future research

    Structured Sparsity Models for Multiparty Speech Recovery from Reverberant Recordings

    Get PDF
    We tackle the multi-party speech recovery problem through modeling the acoustic of the reverberant chambers. Our approach exploits structured sparsity models to perform room modeling and speech recovery. We propose a scheme for characterizing the room acoustic from the unknown competing speech sources relying on localization of the early images of the speakers by sparse approximation of the spatial spectra of the virtual sources in a free-space model. The images are then clustered exploiting the low-rank structure of the spectro-temporal components belonging to each source. This enables us to identify the early support of the room impulse response function and its unique map to the room geometry. To further tackle the ambiguity of the reflection ratios, we propose a novel formulation of the reverberation model and estimate the absorption coefficients through a convex optimization exploiting joint sparsity model formulated upon spatio-spectral sparsity of concurrent speech representation. The acoustic parameters are then incorporated for separating individual speech signals through either structured sparse recovery or inverse filtering the acoustic channels. The experiments conducted on real data recordings demonstrate the effectiveness of the proposed approach for multi-party speech recovery and recognition.Comment: 31 page
    • …
    corecore