428 research outputs found
Recommended from our members
Relationships between human auditory cortical structure and function
The human auditory cortex comprises multiple areas, largely distributed across the supratemporal plane, but the precise number and configuration of auditory areas and their functional significance have not yet been clearly established. In this paper, we discuss recent research concerning architectonic and functional organisation within the human auditory cortex, as well as architectonic and neurophysiological studies in non-human species, which can provide a broad conceptual framework for interpreting functional specialisation in humans. We review the pattern in human auditory cortex of the functional responses to various acoustic cues, such as frequency, pitch, sound level, temporal variation, motion and spatial location, and we discuss their correspondence to what is known about the organisation of the auditory cortex in other primates. There is some neuroimaging evidence of multiple tonotopically organised fields in humans and of functional specialisations of the fields in the processing of different sound features. It is thought that the primary area, on Heschl's gyrus, may have a larger involvement in processing basic sound features, such as frequency and level, and that posterior non-primary areas on the planum temporale may play a larger role in processing more spectrotemporally complex sounds. Ways in which current knowledge of auditory cortical organisation and different data analysis approaches may benefit future functional neuroimaging studies which seek to link auditory cortical structure and function are discussed
Recommended from our members
Can meaningful effective connectivities be obtained between auditory cortical regions?
Structural equation modelling (SEM) of neuroimaging data can be evaluated both for the goodness of fit of the model and for the strength of path coefficients (as an index of effective connectivity). SEM of auditory fMRI data is made difficult by the necessary sparse temporal sampling of the time series (to avoid contamination of auditory activation by the response to scanner noise), and by the paucity of well-defined anatomical information to constrain the functional model. We used SEM (i.e. a model incorporating latent variables) to investigate how well fMRI data in four adjacent cortical fields can be described as an auditory network. Seven out of 14 models (2 hemispheres x (6 subjects and 1 group)) produced a plausible description of the measured data. Since the auditory model to be tested is not fully validated by anatomical data, our approach requires that goodness of fit must be confirmed to assure generalisability of connectivity patterns. For good-fitting models, connectivity patterns varied significantly across subjects and were not replicable across stimulus conditions. SEM of central auditory function therefore appears to be highly sensitive to the voxel-selection procedure and/or the sampling of the time series
Can training improve speech intelligibility and voice recognition?
We found that people learn voices very rapidly. We are able to recognize a new voice (and distinguish it from other voices) accurately after as little as 10 minutes of training. While recognition of a voice seems to plateau quite quickly (our recognition doesn\u27t improve with more training), intelligibility does keep improving as training continues up to one hour. We think the benefits of voice familiarity (such as improved intelligibility in everyday settings, helping people with hearing loss or jobs in noisy environments) can be achieved through deliberate training.https://ir.lib.uwo.ca/brainscanresearchsummaries/1007/thumbnail.jp
Speech-evoked brain activity is more robust to competing speech when it is spoken by someone familiar: Speech representations for familiar voices
When speech is masked by competing sound, people are better at understanding what is said if the talker is familiar compared to unfamiliar. The benefit is robust, but how does processing of familiar voices facilitate intelligibility? We combined high-resolution fMRI with representational similarity analysis to quantify the difference in distributed activity between clear and masked speech. We demonstrate that brain representations of spoken sentences are less affected by a competing sentence when they are spoken by a friend or partner than by someone unfamiliarâeffectively, showing a cortical signal-to-noise ratio (SNR) enhancement for familiar voices. This effect correlated with the familiar-voice intelligibility benefit. We functionally parcellated auditory cortex, and found that the most prominent familiar-voice advantage was manifest along the posterior superior and middle temporal gyri. Overall, our results demonstrate that experience-driven improvements in intelligibility are associated with enhanced multivariate pattern activity in posterior temporal cortex
Recommended from our members
Spectral and temporal processing in human auditory cortex
Hierarchical processing suggests that spectrally and temporally complex stimuli will evoke more activation than do simple stimuli, particularly in non-primary auditory fields. This hypothesis was tested using two tones, a single frequency tone and a harmonic tone, that were either static or frequency modulated to create four stimuli. We interpret the location of differences in activation by drawing comparisons between fMRI and human cytoarchitectonic data, reported in the same brain space. Harmonic tones produced more activation than single tones in right Heschl's gyrus (HG) and bilaterally in the lateral supratemporal plane (STP). Activation was also greater to frequency-modulated tones than to static tones in these areas, plus in left HG and bilaterally in an anterolateral part of the STP and the superior temporal sulcus. An elevated response magnitude to both frequency-modulated tones was found in the lateral portion of the primary area, and putatively in three surrounding non-primary regions on the lateral STP (one anterior and two posterior to HG). A focal site on the posterolateral STP showed an especially high response to the frequency-modulated harmonic tone. Our data highlight the involvement of both primary and lateral non-primary auditory regions
Intelligibility benefit for familiar voices is not accompanied by better discrimination of fundamental frequency or vocal tract length
Speech is more intelligible when it is spoken by familiar than unfamiliar people. If this benefit arises because key voice characteristics like perceptual correlates of fundamental frequency or vocal tract length (VTL) are more accurately represented for familiar voices, listeners may be able to discriminate smaller manipulations to such characteristics for familiar than unfamiliar voices. We measured participantsâ (N = 17) thresholds for discriminating pitch (correlate of fundamental frequency, or glottal pulse rate) and formant spacing (correlate of VTL; âVTL-timbreâ) for voices that were familiar (participantsâ friends) and unfamiliar (other participantsâ friends). As expected, familiar voices were more intelligible. However, discrimination thresholds were no smaller for the same familiar voices. The size of the intelligibility benefit for a familiar over an unfamiliar voice did not relate to the difference in discrimination thresholds for the same voices. Also, the familiar-voice intelligibility benefit was just as large following perceptible manipulations to pitch and VTL-timbre. These results are more consistent with cognitive accounts of speech perception than traditional accounts that predict better discrimination
Intelligibility benefit for familiar voices does not depend on better discrimination of fundamental frequency or vocal tract length
Speech is more intelligible when it is spoken by familiar than unfamiliar people. Two cues to voice identity are glottal pulse rate (GPR) and vocal tract length (VTL): perhaps these features are more accurately represented for familiar voices in a listenerâs brain. If so, listeners should be able to discriminate smaller manipulations to perceptual correlates of these vocal parameters for familiar than unfamiliar voices. We recruited pairs of friends who had known each other for 0.5â22.5 years. We measured thresholds for discriminating pitch (correlate of GPR) and formant spacing (correlate of VTL; âVTL-timbreâ) for voices that were familiar (friends) and unfamiliar (friends of other participants). When a competing talker was present, speech was substantially more intelligible when it was spoken in a familiar voice. Discrimination thresholds were not systematically smaller for familiar compared to unfamiliar talkers. Although, participants detected smaller deviations to VTL-timbre than pitch uniquely for familiar talkers, suggesting a different balance of characteristics contribute to discrimination of familiar and unfamiliar voices. Across participants, we found no relationship between the size of the intelligibility benefit for a familiar over an unfamiliar voice and the difference in discrimination thresholds for the same voices. Also, the intelligibility benefit was not affected by the acoustic manipulations we imposed on voices to assess discrimination thresholds. Overall, these results provide no evidence that two important cues to voice identityâpitch and VTL-timbreâare more accurately represented when voices are familiar, or are necessarily responsible for the large intelligibility benefit derived from familiar voices
Speech-evoked brain activity is more robust to competing speech when it is spoken by someone familiar
The representation of spoken-sentence information in specific regions of the brain is more resistant to interference by competing speech if the target talker is familiar. The posterior temporal cortex represents information about target speech more robustly in the presence of competing speech when the target talker is a friend or partner. We have also shown that the relative robustness of the representations for a familiar, compared to an unfamiliar, voice aligns with the intelligibility benefit that the listener gains from that familiar voice.https://ir.lib.uwo.ca/brainscanresearchsummaries/1006/thumbnail.jp
Speech-evoked brain activity is more robust to competing speech when it is spoken by someone familiar
When speech is masked by competing sound, people are better at understanding what is said if the talker is familiar compared to unfamiliar. The benefit is robust, but how does processing of familiar voices facilitate intelligibility? We combined high-resolution fMRI with representational similarity analysis to quantify the difference in distributed activity between clear and masked speech. We demonstrate that brain representations of spoken sentences are less affected by a competing sentence when they are spoken by a friend or partner than by someone unfamiliarâeffectively, showing a cortical signal-to-noise ratio (SNR) enhancement for familiar voices. This effect correlated with the familiar-voice intelligibility benefit. We functionally parcellated auditory cortex, and found that the most prominent familiar-voice advantage was manifest along the posterior superior and middle temporal gyri. Overall, our results demonstrate that experience-driven improvements in intelligibility are associated with enhanced multivariate pattern activity in posterior temporal cortex
Effects of a consistent target or masker voice on target speech intelligibility in two- and three-talker mixtures.
When the spatial location or identity of a sound is held constant, it is not masked as effectively by competing sounds. This suggests that experience with a particular voice over time might facilitate perceptual organization in multitalker environments. The current study examines whether listeners benefit from experience with a voice only when it is the target, or also when it is a masker, using diotic presentation and a closed-set task (coordinate response measure). A reliable interaction was observed such that, in two-talker mixtures, consistency of masker or target voice over 3-7 trials significantly benefited target recognition performance, whereas in three-talker mixtures, target, but not masker, consistency was beneficial. Overall, this work suggests that voice consistency improves intelligibility, although somewhat differently when two talkers, compared to three talkers, are present, suggesting that consistent-voice information facilitates intelligibility in at least two different ways. Listeners can use a template-matching strategy to extract a known voice from a mixture when it is the target. However, consistent-voice information facilitates segregation only when two, but not three, talkers are present
- âŠ