561 research outputs found

    Effects of reverberation on perceptual segregation of competing voices

    Get PDF
    Two experiments investigated the effect of reverberation on listeners’ ability to perceptually segregate two competing voices. Culling et al. [Speech Commun. 14, 71–96 (1994)] found that for competing synthetic vowels, masked identification thresholds were increased by reverberation only when combined with modulation of fundamental frequency (F0). The present investigation extended this finding to running speech. Speech reception thresholds (SRTs) were measured for a male voice against a single interfering female voice within a virtual room with controlled reverberation. The two voices were either (1) co-located in virtual space at 0° azimuth or (2) separately located at ±60° azimuth. In experiment 1, target and interfering voices were either normally intonated or resynthesized with a fixed F0. In anechoic conditions, SRTs were lower for normally intonated and for spatially separated sources, while, in reverberant conditions, the SRTs were all the same. In experiment 2, additional conditions employed inverted F0 contours. Inverted F0 contours yielded higher SRTs in all conditions, regardless of reverberation. The results suggest that reverberation can seriously impair listeners’ ability to exploit differences in F0 and spatial location between competing voices. The levels of reverberation employed had no effect on speech intelligibility in quiet. © 2003 Acoustical Society of America

    Reverberation impairs brainstem temporal representations of voiced vowel sounds: challenging "periodicity-tagged" segregation of competing speech in rooms.

    Get PDF
    The auditory system typically processes information from concurrently active sound sources (e.g., two voices speaking at once), in the presence of multiple delayed, attenuated and distorted sound-wave reflections (reverberation). Brainstem circuits help segregate these complex acoustic mixtures into "auditory objects." Psychophysical studies demonstrate a strong interaction between reverberation and fundamental-frequency (F0) modulation, leading to impaired segregation of competing vowels when segregation is on the basis of F0 differences. Neurophysiological studies of complex-sound segregation have concentrated on sounds with steady F0s, in anechoic environments. However, F0 modulation and reverberation are quasi-ubiquitous. We examine the ability of 129 single units in the ventral cochlear nucleus (VCN) of the anesthetized guinea pig to segregate the concurrent synthetic vowel sounds /a/ and /i/, based on temporal discharge patterns under closed-field conditions. We address the effects of added real-room reverberation, F0 modulation, and the interaction of these two factors, on brainstem neural segregation of voiced speech sounds. A firing-rate representation of single-vowels' spectral envelopes is robust to the combination of F0 modulation and reverberation: local firing-rate maxima and minima across the tonotopic array code vowel-formant structure. However, single-vowel F0-related periodicity information in shuffled inter-spike interval distributions is significantly degraded in the combined presence of reverberation and F0 modulation. Hence, segregation of double-vowels' spectral energy into two streams (corresponding to the two vowels), on the basis of temporal discharge patterns, is impaired by reverberation; specifically when F0 is modulated. All unit types (primary-like, chopper, onset) are similarly affected. These results offer neurophysiological insights to perceptual organization of complex acoustic scenes under realistically challenging listening conditions.This work was supported by a grant from the BBSRC to Ian M. Winter. Mark Sayles received a University of Cambridge MB/PhD studentship. Tony Watkins (University of Reading, UK) provided the real-room impulse responses. Portions of the data analysis and manuscript preparation were performed by Mark Sayles during the course of an Action on Hearing Loss funded UK–US Fulbright Commission professional scholarship held in the Auditory Neurophysiology and Modeling Laboratory at Purdue University, USA. Mark Sayles is currently supported by a post-doctoral fellowship from Fonds Wetenschappelijk Onderzoek—Vlaanderen, held in the Laboratory of Auditory Neurophysiology at KU Leuven, Belgium.This paper was originally published in Frontiers in Systems Neuroscience (Sayles M, Stasiak A, Winter IM, Frontiers in Systems Neuroscience 2015, 8, 248, doi:10.3389/fnsys.2014.00248)

    Reverberation limits the release from informational masking obtained in the harmonic and binaural domains

    Get PDF
    A difference in fundamental frequency (ΔF0) and a difference in spatial location (ΔSL) are two cues known to provide masking releases when multiple speakers talk at once in a room. Situations were examined in which reverberation should have no effect on the mechanisms underlying the release from energetic masking produced by these two cues. Speech reception thresholds using both unpredictable target sentences and the coordinate response measure followed a similar pattern. Both ΔF0s and ΔSLs provided masking releases in the presence of non-speech maskers (matched in excitation pattern and temporal envelope to speech maskers) which, as intended, were robust to reverberation. Larger masking releases were obtained for speech maskers, but critically, they were affected by reverberation. The results suggest that reverberation either limits the amount of informational masking there is to begin with, or affects its release by ΔF0s or ΔSLs

    Role of differences in fundamental frequency between competing voices in a reverberant room

    Get PDF
    In noisy conversations, listeners can segregate competing voices on the basis of their fundamental frequency (FO). The aim of this thesis was to investigate which mechanisms underlie this FO-segregation ability and whether this ability is affected by reverberation. This work provided evidence for a mechanism, which cancels interfering voices on the basis of their harmonic structure a process termed harmonic cancellation. We developed a paradigm in which listeners had to detect a band of noise masked by a harmonic or inharmonic complex masker (Chapter II). Harmonic cancellation was found to be beneficial up to about 3 kHz, sensitive to a degree of inharmonicity reflected by a peak autocorrelation of 0.9 or less, and to integrate harmonic information over very large bands. In addition to harmonic cancellation, listeners may also use FO as a sequential cue, provided that AFO is sufficiently large (Chapter III), in order to organise the auditory scene in the presence of several talkers a process termed sequential FO-grouping. By manipulating the FO of competing sources heard in anechoic or in reverberant environments, the Speech Reception Threshold (SRT) of a target voice masked by buzz (Chapter IV) or speech (Chapter V) interferers, was elevated when the interferer but not the target, was FO-modulated and especially in reverberation for the buzz interferer. These results were explained in terms of disruption of harmonic cancellation. Moreover, the benefit of an 8 semitone AFO was disrupted by reverberation even for monotonized sources, suggesting that reverberation is also detrimental to sequential FO-grouping. To conclude, the listener's ability to segregate voices by FO relies on the mechanisms of harmonic cancellation and sequential FO-grouping. Both these mechanisms are likely to be disrupted in realistic situations of conversation, i.e. real speech in reverberant rooms

    Effect of Reverberation Context on Spatial Hearing Performance of Normally Hearing Listeners

    Get PDF
    Previous studies provide evidence that listening experience in a particular reverberant environment improves speech intelligibility and localization performance in that environment. Such studies, however, are few, and there is little knowledge of the underlying mechanisms. The experiments presented in this thesis explored the effect of reverberation context, in particular, the similarity in interaural coherence within a context, on listeners\u27 performance in sound localization, speech perception in a spatially separated noise, spatial release from speech-on-speech masking, and target location identification in a multi-talker configuration. All experiments were conducted in simulated reverberant environments created with a loudspeaker array in an anechoic chamber. The reflections comprising the reverberation in each environment had the same temporal and relative amplitude patterns, but varied in their lateral spread, which affected the interaural coherence of reverberated stimuli. The effect of reverberation context was examined by comparing performance in two reverberation contexts, mixed and fixed. In the mixed context, the reverberation environment applied to each stimulus varied trial-by-trial, whereas in the fixed context, the reverberation environment was held constant within a block of trials. In Experiment I (absolute judgement of sound location), variability in azimuth judgments was lower in the fixed than in the mixed context, suggesting that sound localization depended not only on the cues presented in isolated trials. In Experiment II, the intelligibility of speech in a spatially separated noise was found to be similar in both reverberation contexts. That result contrasts with other studies, and suggests that the fixed context did not assist listeners in compensating for degraded interaural coherence. In Experiment III, speech intelligibility in multi-talker configurations was found to be better in the fixed context, but only when the talkers were separated. That is, the fixed context improved spatial release from masking. However, in the presence of speech maskers, consistent reverberation did not improve the localizability of the target talker in a three-alternative location-identification task. Those results suggest that in multi-talker situations, consistent coherence may not improve target localizability, but rather that consistent context may facilitate the buildup of spatial selective attention
    • …
    corecore