151 research outputs found

    Irregular speech rate dissociates auditory cortical entrainment, evoked responses, and frontal alpha

    Get PDF
    The entrainment of slow rhythmic auditory cortical activity to the temporal regularities in speech is considered to be a central mechanism underlying auditory perception. Previous work has shown that entrainment is reduced when the quality of the acoustic input is degraded, but has also linked rhythmic activity at similar time scales to the encoding of temporal expectations. To understand these bottom-up and top-down contributions to rhythmic entrainment, we manipulated the temporal predictive structure of speech by parametrically altering the distribution of pauses between syllables or words, thereby rendering the local speech rate irregular while preserving intelligibility and the envelope fluctuations of the acoustic signal. Recording EEG activity in human participants, we found that this manipulation did not alter neural processes reflecting the encoding of individual sound transients, such as evoked potentials. However, the manipulation significantly reduced the fidelity of auditory delta (but not theta) band entrainment to the speech envelope. It also reduced left frontal alpha power and this alpha reduction was predictive of the reduced delta entrainment across participants. Our results show that rhythmic auditory entrainment in delta and theta bands reflect functionally distinct processes. Furthermore, they reveal that delta entrainment is under top-down control and likely reflects prefrontal processes that are sensitive to acoustical regularities rather than the bottom-up encoding of acoustic features

    Local Temporal Regularities in Child-Directed Speech in Spanish

    Get PDF
    Published online: Oct 4, 2022Purpose: The purpose of this study is to characterize the local (utterance-level) temporal regularities of child-directed speech (CDS) that might facilitate phonological development in Spanish, classically termed a syllable-timed language. Method: Eighteen female adults addressed their 4-year-old children versus other adults spontaneously and also read aloud (CDS vs. adult-directed speech [ADS]). We compared CDS and ADS speech productions using a spectrotemporal model (Leong & Goswami, 2015), obtaining three temporal metrics: (a) distribution of modulation energy, (b) temporal regularity of stressed syllables, and (c) syllable rate. Results: CDS was characterized by (a) significantly greater modulation energy in the lower frequencies (0.5–4 Hz), (b) more regular rhythmic occurrence of stressed syllables, and (c) a slower syllable rate than ADS, across both spontaneous and read conditions. Discussion: CDS is characterized by a robust local temporal organization (i.e., within utterances) with amplitude modulation bands aligning with delta and theta electrophysiological frequency bands, respectively, showing greater phase synchronization than in ADS, facilitating parsing of stress units and syllables. These temporal regularities, together with the slower rate of production of CDS, might support the automatic extraction of phonological units in speech and hence support the phonological development of children. Supplemental Material: https://doi.org/10.23641/asha.21210893This study was supported by the Formación de Personal Investigado Grant BES-2016-078125 by Ministerio Español de Economía, Industria y Competitividad and Fondo Social Europeo awarded to Jose Pérez-Navarro; through Project RTI2018-096242-B-I00 (Ministerio de Ciencia, Innovación y Universidades [MCIU]/Agencia Estatal de Investigación [AEI]/Fondo Europeo de Desarrollo Regional [FEDER], Unión Europea) funded by MCIU, the AEI, and FEDER awarded to Marie Lallier; by the Basque Government through the Basque Excellence Research Centre 2018-2021 Program; and by the Spanish State Research Agency through Basque Center on Cognition, Brain and Language Severo Ochoa Excellence Accreditation SEV- 2015-0490. We want to thank the participants and their children for their volunteer contribution to our study

    Enhanced amplitude modulations contribute to the Lombard intelligibility benefit: Evidence from the Nijmegen Corpus of Lombard Speech

    No full text
    Speakers adjust their voice when talking in noise, which is known as Lombard speech. These acoustic adjustments facilitate speech comprehension in noise relative to plain speech (i.e., speech produced in quiet). However, exactly which characteristics of Lombard speech drive this intelligibility benefit in noise remains unclear. This study assessed the contribution of enhanced amplitude modulations to the Lombard speech intelligibility benefit by demonstrating that (1) native speakers of Dutch in the Nijmegen Corpus of Lombard Speech (NiCLS) produce more pronounced amplitude modulations in noise vs. in quiet; (2) more enhanced amplitude modulations correlate positively with intelligibility in a speech-in-noise perception experiment; (3) transplanting the amplitude modulations from Lombard speech onto plain speech leads to an intelligibility improvement, suggesting that enhanced amplitude modulations in Lombard speech contribute towards intelligibility in noise. Results are discussed in light of recent neurobiological models of speech perception with reference to neural oscillators phase-locking to the amplitude modulations in speech, guiding the processing of speech

    One Way or Another: Cortical Language Areas Flexibly Adapt Processing Strategies to Perceptual And Contextual Properties of Speech

    Get PDF
    Published:07 April 2021Cortical circuits rely on the temporal regularities of speech to optimize signal parsing for sound-to-meaning mapping. Bottom-up speech analysis is accelerated by top–down predictions about upcoming words. In everyday communications, however, listeners are regularly presented with challenging input—fluctuations of speech rate or semantic content. In this study, we asked how reducing speech temporal regularity affects its processing—parsing, phonological analysis, and ability to generate context-based predictions. To ensure that spoken sentences were natural and approximated semantic constraints of spontaneous speech we built a neural network to select stimuli from large corpora. We analyzed brain activity recorded with magnetoencephalography during sentence listening using evoked responses, speech-to-brain synchronization and representational similarity analysis. For normal speech theta band (6.5–8 Hz) speech-to-brain synchronization was increased and the left fronto-temporal areas generated stronger contextual predictions. The reverse was true for temporally irregular speech—weaker theta synchronization and reduced top–down effects. Interestingly, delta-band (0.5 Hz) speech tracking was greater when contextual/semantic predictions were lower or if speech was temporally jittered. We conclude that speech temporal regularity is relevant for (theta) syllabic tracking and robust semantic predictions while the joint support of temporal and contextual predictability reduces word and phrase-level cortical tracking (delta).The European Union’s Horizon 2020 research and innovation programme (under the Marie Sklodowska-Curie grant agreement No 798971 awarded to A.K.G.); the Spanish Ministry of Science, Innovation and Universities (grant RTI2018-096311-B-I00 to N.M.); the Agencia Estatal de Investigación (AEI), the Fondo Europeo de Desarrollo Regional (FEDER); the Basque Government (through the BERC 2018-2021 program), the Spanish State Research Agency through BCBL Severo Ochoa excellence accreditation (SEV-2015-0490), DeepText project (KK-2020/00088) and Ixa excellence research group (IT1343-19). the UPV/EHU (a postdoctoral grant ESPDOC18/101 to A.B.); the NVIDIA Corporation (to A.B. with the donation of a Titan V GPU used for this research)

    Synchronous, but not entrained: Exogenous and endogenous cortical rhythms of speech and language processing

    Get PDF
    Research into speech processing is often focused on a phenomenon termed ‘entrainment’, whereby the cortex shadows rhythmic acoustic information with oscillatory activity. Entrainment has been observed to a range of rhythms present in speech; in addition, synchronicity with abstract information (e.g., syntactic structures) has been observed. Entrainment accounts face two challenges: First, speech is not exactly rhythmic; second, synchronicity with representations that lack a clear acoustic counterpart has been described. We propose that apparent entrainment does not always result from acoustic information. Rather, internal rhythms may have functionalities in the generation of abstract representations and predictions. While acoustics may often provide punctate opportunities for entrainment, internal rhythms may also live a life of their own to infer and predict information, leading to intrinsic synchronicity—not to be counted as entrainment. This possibility may open up new research avenues in the psycho– and neurolinguistic study of language processing and language development

    Speech rhythm and language acquisition: an amplitude modulation phase hierarchy perspective.

    Get PDF
    Language lies at the heart of our experience as humans and disorders of language acquisition carry severe developmental costs. Rhythmic processing lies at the heart of language acquisition. Here, I review our understanding of the perceptual and neural mechanisms that support language acquisition, from a novel amplitude modulation perspective. Amplitude modulation patterns in infant- and child-directed speech support the perceptual experience of rhythm, and the brain encodes these rhythm patterns in part via neuroelectric oscillations. When brain rhythms align themselves with (entrain to) acoustic rhythms, speech intelligibility improves. Recent advances in the auditory neuroscience of speech processing enable studies of neuronal oscillatory entrainment in children and infants. The "amplitude modulation phase hierarchy" theoretical perspective on language acquisition is applicable across languages, and cross-language investigations adopting this novel perspective would be valuable for the field

    Speech rhythms and multiplexed oscillatory sensory coding in the human brain

    Get PDF
    Cortical oscillations are likely candidates for segmentation and coding of continuous speech. Here, we monitored continuous speech processing with magnetoencephalography (MEG) to unravel the principles of speech segmentation and coding. We demonstrate that speech entrains the phase of low-frequency (delta, theta) and the amplitude of high-frequency (gamma) oscillations in the auditory cortex. Phase entrainment is stronger in the right and amplitude entrainment is stronger in the left auditory cortex. Furthermore, edges in the speech envelope phase reset auditory cortex oscillations thereby enhancing their entrainment to speech. This mechanism adapts to the changing physical features of the speech envelope and enables efficient, stimulus-specific speech sampling. Finally, we show that within the auditory cortex, coupling between delta, theta, and gamma oscillations increases following speech edges. Importantly, all couplings (i.e., brain-speech and also within the cortex) attenuate for backward-presented speech, suggesting top-down control. We conclude that segmentation and coding of speech relies on a nested hierarchy of entrained cortical oscillations

    Auditory cortical delta-entrainment interacts with oscillatory power in multiple fronto-parietal networks

    Get PDF
    The timing of slow auditory cortical activity aligns to the rhythmic fluctuations in speech. This entrainment is considered to be a marker of the prosodic and syllabic encoding of speech, and has been shown to correlate with intelligibility. Yet, whether and how auditory cortical entrainment is influenced by the activity in other speech–relevant areas remains unknown. Using source-localized MEG data, we quantified the dependency of auditory entrainment on the state of oscillatory activity in fronto-parietal regions. We found that delta band entrainment interacted with the oscillatory activity in three distinct networks. First, entrainment in the left anterior superior temporal gyrus (STG) was modulated by beta power in orbitofrontal areas, possibly reflecting predictive top-down modulations of auditory encoding. Second, entrainment in the left Heschl's Gyrus and anterior STG was dependent on alpha power in central areas, in line with the importance of motor structures for phonological analysis. And third, entrainment in the right posterior STG modulated theta power in parietal areas, consistent with the engagement of semantic memory. These results illustrate the topographical network interactions of auditory delta entrainment and reveal distinct cross-frequency mechanisms by which entrainment can interact with different cognitive processes underlying speech perception

    Acoustically driven cortical delta oscillations underpin prosodic chunking

    Get PDF
    Oscillation-based models of speech perception postulate a cortical computational principle by which decoding is performed within a window structure derived by a segmentation process. Segmentation of syllable-size chunks is realized by a theta oscillator. We provide evidence for an analogous role of a delta oscillator in the segmentation of phrase-sized chunks. We recorded Magnetoencephalography (MEG) in humans, while participants performed a target identification task. Random-digit strings, with phrase-long chunks of two digits, were presented at chunk rates of 1.8 Hz or 2.6 Hz, inside or outside the delta frequency band (defined here to be 0.5 - 2 Hz). Strong periodicities were elicited by chunk rates inside of delta in superior, middle temporal areas and speech-motor integration areas. Periodicities were diminished or absent for chunk rates outside delta, in line with behavioral performance. Our findings show that prosodic chunking of phrase-sized acoustic segments is correlated with acoustic-driven delta oscillations, expressing anatomically specific patterns of neuronal periodicities
    corecore