138 research outputs found

    Neuronal representation of sound source location in the auditory cortex during active navigation

    Get PDF
    The ability to localize sounds is crucial for the survival of both predators as well as prey. The former rely on their senses to lead them to the latter, which in turn also benefit from locating a predator in the vicinity to escape accordingly. In such cases, the sound localization process typically takes place while the animals are in motion. Since the cues that the brain uses to localize sounds are head-centered (egocentric), they can change very rapidly when an animal moves and rotates. This constitutes an even bigger challenge than sound localization in a static environment. Up to now, however, both aspects have mostly been studied separately in neuroscience, thus limiting our understanding of active sound localization during navigation. This thesis reports on the development of a novel behavioral paradigm – the Sensory Island Task (SIT) – to promote sound localization during unrestricted motion. By attributing a different behavioral meaning (associated to different outcomes) to two spatially separated sound sources, Mongolian gerbils (Meriones unguiculatus) were trained to forage for an area (target island) in the arena that triggered a change in the active sound source to the target loudspeaker and to report its detection by remaining within the island for a duration of 6 s. Importantly, the two loudspeakers played identical sounds and the location of the target island in the arena was changed randomly every trial. When the probability of successfully identifying the target island exceeded the chance level, a tetrode bundle was implanted in the primary auditory cortex of the gerbils to record neuronal responses during task performance. Canonically, the auditory cortex (AC) is described as possessing neurons with a broad hemispheric tuning. Nonetheless, context and behavioral state have been shown to modulate the neuronal responses in the AC. The experiments described in this thesis demonstrate the existence of a large variety of additional, previously unreported (or underreported) spatial tuning types. In particular, neurons that were sensitive to the midline and, most intriguingly, neurons that were sensitive to the task identity of the active loudspeaker were observed. The latter comprise neurons that were spatially tuned to only one of the two loudspeakers, neurons that exhibited a large difference in the preferred egocentric sound-source location for the two loudspeakers as well as spatially untuned neurons whose firing rate changed depending on the active loudspeaker. Additionally, temporal complexity in the neuronal responses was observed, with neurons changing their preferred egocentric sound-source location throughout their response to a sound. Corroborating earlier studies, also here it was found that the task-specific choice of the animal was reflected in the neuronal responses. Specifically, the neuronal firing rate decreased before the animal successfully finished a trial in comparison to situations in which the gerbil incorrectly left the target island before trial completion. Furthermore, the differential behavioral meaning between the two loudspeakers was found to be represented in the neuronal tuning acuity, with neurons being more sharply tuned to sounds coming from the target than from the background loudspeaker. Lastly, by implementing an artificial neural network, all of the observed phenomena could be studied in a common framework, enabling a better and more comprehensive understanding of the computational relevance of the diversity of observed neuronal responses. Strikingly, the algorithm was capable of predicting not only the egocentric sound-source location but also which sound source was active – both with high accuracy. Taken together, the results presented in this thesis suggest the existence of an interlaced coding of egocentric and allocentric information in the neurons of the primary auditory cortex. These novel findings thus contribute towards a better understanding of how sound sources are perceptually stable during self-motion, an effect that could be advantageous for selective hearing

    Classification and Separation Techniques based on Fundamental Frequency for Speech Enhancement

    Get PDF
    [ES] En esta tesis se desarrollan nuevos algoritmos de clasificación y mejora de voz basados en las propiedades de la frecuencia fundamental (F0) de la señal vocal. Estas propiedades permiten su discriminación respecto al resto de señales de la escena acústica, ya sea mediante la definición de características (para clasificación) o la definición de modelos de señal (para separación). Tres contribuciones se aportan en esta tesis: 1) un algoritmo de clasificación de entorno acústico basado en F0 para audífonos digitales, capaz de clasificar la señal en las clases voz y no-voz; 2) un algoritmo de detección de voz sonora basado en la aperiodicidad, capaz de funcionar en ruido no estacionario y con aplicación a mejora de voz; 3) un algoritmo de separación de voz y ruido basado en descomposición NMF, donde el ruido se modela de una forma genérica mediante restricciones matemáticas.[EN]This thesis is focused on the development of new classification and speech enhancement algorithms based, explicitly or implicitly, on the fundamental frequency (F0). The F0 of speech has a number of properties that enable speech discrimination from the remaining signals in the acoustic scene, either by defining F0-based signal features (for classification) or F0-based signal models (for separation). Three main contributions are included in this work: 1) an acoustic environment classification algorithm for hearing aids based on F0 to classify the input signal into speech and nonspeech classes; 2) a frame-by-frame basis voiced speech detection algorithm based on the aperiodicity measure, able to work under non-stationary noise and applicable to speech enhancement; 3) a speech denoising algorithm based on a regularized NMF decomposition, in which the background noise is described in a generic way with mathematical constraints.Tesis Univ. Jaén. Departamento de Ingeniería de Telecomunición. Leída el 11 de enero de 201

    Models and analysis of vocal emissions for biomedical applications

    Get PDF
    This book of Proceedings collects the papers presented at the 3rd International Workshop on Models and Analysis of Vocal Emissions for Biomedical Applications, MAVEBA 2003, held 10-12 December 2003, Firenze, Italy. The workshop is organised every two years, and aims to stimulate contacts between specialists active in research and industrial developments, in the area of voice analysis for biomedical applications. The scope of the Workshop includes all aspects of voice modelling and analysis, ranging from fundamental research to all kinds of biomedical applications and related established and advanced technologies

    Predicting room acoustical behavior with the ODEON computer model

    Get PDF

    Proceedings of the 7th Sound and Music Computing Conference

    Get PDF
    Proceedings of the SMC2010 - 7th Sound and Music Computing Conference, July 21st - July 24th 2010
    • …
    corecore