29 research outputs found

    Measure and model of vocal-tract length discrimination in cochlear implants

    Get PDF
    Voice discrimination is crucial to selectively listen to a particular talker in a crowded environment. In normalhearing listeners, it strongly relies on the perception of two dimensions: the fundamental frequency and the vocal-tract length. Yet, very little is known about the perception of the latter in cochlear implants. The present study reports discrimination thresholds for vocal-tract length in normal-hearing listeners and cochlear-implant users. The behavioral results were then used to determine the effective spectral resolution in a model of electric hearing: effective resolution in the implant was found to be poorer than previously suggested by psychophysical measurements. Such a model could be used for clinical purposes, or to facilitate the development of new strategies.</p

    Progressive associative phonagnosia: A neuropsychological analysis

    Get PDF
    AbstractThere are few detailed studies of impaired voice recognition, or phonagnosia. Here we describe two patients with progressive phonagnosia in the context of frontotemporal lobar degeneration. Patient QR presented with behavioural decline and increasing difficulty recognising familiar voices, while patient KL presented with progressive prosopagnosia. In a series of neuropsychological experiments we assessed the ability of QR and KL to recognise and judge the familiarity of voices, faces and proper names, to recognise vocal emotions, to perceive and discriminate voices, and to recognise environmental sounds and musical instruments. The patients were assessed in relation to a group of healthy age-matched control subjects. QR exhibited severe impairments of voice identification and familiarity judgments with relatively preserved recognition of difficulty-matched faces and environmental sounds; recognition of musical instruments was impaired, though better than recognition of voices. In contrast, patient KL exhibited severe impairments of both voice and face recognition, with relatively preserved recognition of musical instruments and environmental sounds. Both patients demonstrated preserved ability to analyse perceptual properties of voices and to recognise vocal emotions. The voice processing deficit in both patients could be characterised as associative phonagnosia: in the case of QR, this was relatively selective for voices, while in the case of KL, there was evidence for a multimodal impairment of person knowledge. The findings have implications for current cognitive models of voice recognition

    Does knowing speaker sex facilitate vowel recognition at short durations?

    Get PDF
    A man, woman or child saying the same vowel do so with very different voices. The auditory system solves the complex problem of extracting what the man, woman or child has said despite substantial differences in the acoustic properties of their voices. Much of the acoustic variation between the voices of men and woman is due to changes in the underlying anatomical mechanisms for producing speech. If the auditory system knew the sex of the speaker then it could potentially correct for speaker sex related acoustic variation thus facilitating vowel recognition. This study measured the minimum stimulus duration necessary to accurately discriminate whether a brief vowel segment was spoken by a man or woman, and the minimum stimulus duration necessary to accurately recognise what vowel was spoken. Results showed that reliable vowel recognition precedes reliable speaker sex discrimination, thus questioning the use of speaker sex information in compensating for speaker sex related acoustic variation in the voice. Furthermore, the pattern of performance across experiments where the fundamental frequency and formant frequency information of speaker’s voices were systematically varied, was markedly different depending on whether the task was speaker-sex discrimination or vowel recognition. This argues for there being little relationship between perception of speaker sex (indexical information) and perception of what has been said (linguistic information) at short durations

    Auditory object cognition in dementia

    Get PDF
    The cognition of nonverbal sounds in dementia has been relatively little explored. Here we undertook a systematic study of nonverbal sound processing in patient groups with canonical dementia syndromes comprising clinically diagnosed typical amnestic Alzheimer's disease (AD; n = 21), progressive nonfluent aphasia (PNFA; n = 5), logopenic progressive aphasia (LPA; n = 7) and aphasia in association with a progranulin gene mutation (GAA; n = 1), and in healthy age-matched controls (n = 20). Based on a cognitive framework treating complex sounds as 'auditory objects', we designed a novel neuropsychological battery to probe auditory object cognition at early perceptual (sub-object), object representational (apperceptive) and semantic levels. All patients had assessments of peripheral hearing and general neuropsychological functions in addition to the experimental auditory battery. While a number of aspects of auditory object analysis were impaired across patient groups and were influenced by general executive (working memory) capacity, certain auditory deficits had some specificity for particular dementia syndromes. Patients with AD had a disproportionate deficit of auditory apperception but preserved timbre processing. Patients with PNFA had salient deficits of timbre and auditory semantic processing, but intact auditory size and apperceptive processing. Patients with LPA had a generalised auditory deficit that was influenced by working memory function. In contrast, the patient with GAA showed substantial preservation of auditory function, but a mild deficit of pitch direction processing and a more severe deficit of auditory apperception. The findings provide evidence for separable stages of auditory object analysis and separable profiles of impaired auditory object cognition in different dementia syndromes. (C) 2011 Elsevier Ltd. All rights reserved

    Highly Accurate and Robust Identity Perception From Personally Familiar Voices

    Get PDF
    Previous research suggests that familiarity with a voice can afford benefits for voice and speech perception. However, even familiar voice perception has been reported to be error-prone in previous research, especially in the face of challenges such as reduced verbal cues and acoustic distortions. It has been hypothesised that such findings may arise due to listeners not being “familiar enough” with the voices used in laboratory studies, and thus being inexperienced with their full vocal repertoire. By extension, voice perception based on highly familiar voices – acquired via substantial, naturalistic experience – should therefore be more robust than voice perception from less familiar voices. We investigated this proposal by contrasting voice perception of personally-familiar voices (participants’ romantic partners) versus lab-trained voices in challenging experimental tasks. Specifically, we tested how differences in familiarity may affect voice identity perception from non-verbal vocalisations and acoustically-modulated speech. Large benefits for the personally-familiar voice over less familiar, lab-trained voice were found for identity recognition, with listeners displaying both highly accurate yet more conservative recognition of personally familiar voices. However, no familiar-voice benefits were found for speech comprehension against background noise. Our findings suggest that listeners have fine-tuned representations of highly familiar voices that result in more robust and accurate voice recognition despite challenging listening contexts, yet these advantages may not always extend to speech perception. Our study therefore highlights that familiarity is indeed a continuum, with identity perception for personally-familiar voices being highly accurate

    Highly accurate and robust identity perception from personally familiar voices

    Get PDF
    Previous research suggests that familiarity with a voice can afford benefits for voice and speech perception. However, even familiar voice perception has been reported to be error-prone in previous research, especially in the face of challenges such as reduced verbal cues and acoustic distortions. It has been hypothesised that such findings may arise due to listeners not being “familiar enough” with the voices used in laboratory studies, and thus being inexperienced with their full vocal repertoire. By extension, voice perception based on highly familiar voices – acquired via substantial, naturalistic experience – should therefore be more robust than voice perception from less familiar voices. We investigated this proposal by contrasting voice perception of personally-familiar voices (participants’ romantic partners) versus lab-trained voices in challenging experimental tasks. Specifically, we tested how differences in familiarity may affect voice identity perception from non-verbal vocalisations and acoustically-modulated speech. Large benefits for the personally-familiar voice over less familiar, lab-trained voice were found for identity recognition, with listeners displaying both highly accurate yet more conservative recognition of personally familiar voices. However, no familiar-voice benefits were found for speech comprehension against background noise. Our findings suggest that listeners have fine-tuned representations of highly familiar voices that result in more robust and accurate voice recognition despite challenging listening contexts, yet these advantages may not always extend to speech perception. Our study therefore highlights that familiarity is indeed a continuum, with identity perception for personally-familiar voices being highly accurate

    On the color of voices:the relationship between cochlear implant users’ voice cue perception and speech intelligibility in cocktail-party scenarios

    Get PDF
    Cochlear implants (CIs) are neuroprosthetic devices that are surgically implanted to restore functional hearing in deaf and hard-of-hearing individuals. Most CI users can understand speech well in quiet situations, yet, it becomes quite challenging for them to understand speech in crowded environments, especially when multiple people are speaking simultaneously. This dissertation investigated whether such difficulties are related to the poor representation of voice cues in the implant arising from degraded spectral and temporal resolution from signal processing strategies. Human voices are characterized by their pitch (F0), in addition to a second dimension called the vocal-tract length (VTL). This dimension directly scales with the size of the speaker and, therefore, plays a crucial role in the distinction between male and female talkers, or between adults and children. The research questions were: whether CI users’ speech intelligibility in the presence of a competing talker (speech-on-speech; SoS) is related to their sensitivity to the F0 and VTL differences between the speakers, whether this relationship is influenced by the spectral resolution in the implant, and whether optimizing signal processing algorithms could improve the perception of such voice cues. The data showed that CI users’ SoS intelligibility was related to how sensitive they were to both F0 and VTL differences, and that this relationship was influenced by the spectral resolution in the implant. The data also provided evidence that CI users can draw a benefit from voice differences between male and female speakers, but not between female speakers and children. In addition, spectral enhancement techniques and optimization of some implant parameters were both shown to contribute to an improvement in SoS intelligibility and VTL sensitivity, respectively. These findings lay the foundations for future optimizations of the implant to improve CI users’ speech intelligibility in noisy settings

    Development of voice perception is dissociated across gender cues in school-age children

    Get PDF
    Funder: Center for Language and Cognition Groningen (CLCG)Funder: LabEx CeLyA (“Centre Lyonnais d’Acoustique”, ANR-10-LABX-0060/ANR-11-IDEX-0007)Abstract: Children’s ability to distinguish speakers’ voices continues to develop throughout childhood, yet it remains unclear how children’s sensitivity to voice cues, such as differences in speakers’ gender, develops over time. This so-called voice gender is primarily characterized by speakers’ mean fundamental frequency (F0), related to glottal pulse rate, and vocal-tract length (VTL), related to speakers’ size. Here we show that children’s acquisition of adult-like performance for discrimination, a lower-order perceptual task, and categorization, a higher-order cognitive task, differs across voice gender cues. Children’s discrimination was adult-like around the age of 8 for VTL but still differed from adults at the age of 12 for F0. Children’s perceptual weight attributed to F0 for gender categorization was adult-like around the age of 6 but around the age of 10 for VTL. Children’s discrimination and weighting of F0 and VTL were only correlated for 4- to 6-year-olds. Hence, children’s development of discrimination and weighting of voice gender cues are dissociated, i.e., adult-like performance for F0 and VTL is acquired at different rates and does not seem to be closely related. The different developmental patterns for auditory discrimination and categorization highlight the complexity of the relationship between perceptual and cognitive mechanisms of voice perception
    corecore