13 research outputs found

    Envelope regularity discrimination

    Get PDF
    © 2019 Acoustical Society of America. The ability to discriminate irregular from regular amplitude modulation was assessed using the "envelope regularity discrimination" test. The amount of irregularity was parametrically varied and quantified by an "irregularity index." Normative data were gathered for young subjects with normal audiometric thresholds. Parameters varied were the carrier and modulation frequencies, f c and f m , and the baseline modulation index, m. All tests were performed using a background threshold-equalizing noise. The main findings were (1) using f c = 4000 Hz, f m = 8 Hz, and m = 0.3, performance improved over the first two threshold runs and then remained roughly stable, and there was a high correlation between thresholds obtained at 80 dB sound pressure level (SPL) and at 20 dB sensation level; (2) using f m = 8 Hz and m = 0.3 with a level of 80 dB SPL, thresholds did not vary significantly across f c = 1000, 2000, and 4000 Hz; (3) using f m = 8 Hz and f c = 4000 Hz with a level of 80 dB SPL, thresholds did not vary significantly for m from 0.2 to 0.5; and (4) using m = 0.3 and f c = 4000 Hz with a level of 80 dB SPL, thresholds improved with increasing f m from 2 to 16 Hz. For all conditions, there was substantial individual variability, probably resulting from differences in "processing efficiency.

    Does training with amplitude modulated tones affect tone-vocoded speech perception?

    Get PDF
    Temporal-envelope cues are essential for successful speech perception. We asked here whether training on stimuli containing temporal-envelope cues without speech content can improve the perception of spectrally-degraded (vocoded) speech in which the temporal-envelope (but not the temporal fine structure) is mainly preserved. Two groups of listeners were trained on different amplitude-modulation (AM) based tasks, either AM detection or AM-rate discrimination (21 blocks of 60 trials during two days, 1260 trials; frequency range: 4Hz, 8Hz, and 16Hz), while an additional control group did not undertake any training. Consonant identification in vocoded vowel-consonant-vowel stimuli was tested before and after training on the AM tasks (or at an equivalent time interval for the control group). Following training, only the trained groups showed a significant improvement in the perception of vocoded speech, but the improvement did not significantly differ from that observed for controls. Thus, we do not find convincing evidence that this amount of training with temporal-envelope cues without speech content provide significant benefit for vocoded speech intelligibility. Alternative training regimens using vocoded speech along the linguistic hierarchy should be explored

    Investigating the effect of one year of learning to play a musical instrument on speech-in-noise perception and phonological short-term memory in 5-to-7-year-old children

    Get PDF
    The benefits in speech-in-noise perception, language and cognition brought about by extensive musical training in adults and children have been demonstrated in a number of cross-sectional studies. Therefore, this study aimed to investigate whether one year of school-delivered musical training, consisting of individual and group instrumental classes, was capable of producing advantages for speech-in-noise perception and phonological short-term memory in children tested in a simulated classroom environment. Forty-one children aged 5-7 years at the first measurement point participated in the study and either went to a music-focused or a sport-focused private school with an otherwise equivalent school curriculum. The children's ability to detect number and color words in noise was measured under a number of conditions including different masker types (speech-shaped noise, single-talker background) and under varying spatial combinations of target and masker (spatially collocated, spatially separated). Additionally, a cognitive factor essential to speech perception, namely phonological short-term memory, was assessed. Findings were unable to confirm that musical training of the frequency and duration administered was associated with a musicians' advantage for either speech in noise, under any of the masker or spatial conditions tested, or phonological short-term memory

    The contribution of individual differences in memory span and language ability to spatial release from masking in young children

    Get PDF
    Purpose Working memory capacity and language ability modulate speech reception; however, the respective roles of peripheral and cognitive processing are unclear. The contribution of individual differences in these abilities to utilization of spatial cues when separating speech from informational and energetic masking backgrounds in children has not yet been determined. Therefore, this study explored whether speech reception in children is modulated by environmental factors, such as the type of background noise and spatial configuration of target and noise sources, and individual differences in the cognitive and linguistic abilities of listeners. Method Speech reception thresholds were assessed in 39 children aged 5-7 years in simulated school listening environments. Speech reception thresholds of target sentences spoken by an adult male consisting of number and color combinations were measured using an adaptive procedure, with speech-shaped white noise and single-talker backgrounds that were either collocated (target and back-ground at 0°) or spatially separated (target at 0°, background noise at 90° to the right). Spatial release from masking was assessed alongside memory span and expressive language. Results and Conclusion Significant main effect results showed that speech reception thresholds were highest for informational maskers and collocated conditions. Significant interactions indicated that individual differences in memory span and language ability were related to spatial release from masking advantages. Specifically, individual differences in memory span and language were related to the utilization of spatial cues in separated conditions. Language differences were related to auditory stream segregation abilities in collocated conditions that lack helpful spatial cues, pointing to the utilization of language processes to make up for losses in spatial information

    When hearing loss masquerades as cognitive decline

    No full text
    Editorial commentary: When hearing loss masquerades as cognitive decline

    On the possible overestimation of cognitive decline: the impact of age-related hearing loss on cognitive-test performance

    No full text
    Individual differences and age-related normal and pathological changes in mental abilities require the use of cognitive screening and assessment tools. However, simultaneously occurring deficits in sensory processing, whose prevalence increases especially in old age, may negatively impact cognitive-test performance and thus result in an overestimation of cognitive decline. This hypothesis was tested using an impairment-simulation approach. Young normal-hearing university students performed three memory tasks, using auditorily presented speech stimuli that were either unprocessed or processed to mimic some of the perceptual consequences of age-related hearing loss (ARHL). Both short-term-memory and working-memory capacities were significantly lower in the simulated-hearing-loss condition, despite good intelligibility of the test stimuli. The findings are consistent with the notion that, in case of ARHL, the perceptual processing of auditory stimuli used in cognitive assessments requires additional (cognitive) resources that cannot be used toward the execution of the cognitive task itself. Researchers and clinicians would be well advised to consider sensory impairments as a confounding variable when administering cognitive tasks and interpreting their results

    The association between the processing of binaural temporal-fine-structure information and audiometric threshold and age: a meta-analysis

    No full text
    The ability to process binaural temporal fine structure (TFS) information, which influences the perception of speech in spatially distributed soundscapes, declines with increasing hearing loss and age. Because of the relatively small sample sizes used in previous studies, and the population-unrepresentative distribution of hearing loss and ages within study samples, it has been difficult to determine the relative and combined contributions of hearing loss and age. The aim of this study was to survey published and unpublished studies that assessed binaural TFS sensitivity using the TFS-low frequency (LF) test. Results from 19 studies were collated, yielding sample sizes of 147 to 648, depending on the test frequency. At least for the test frequency of 500 Hz, there were at least 67 listeners in each of four adult age groups and the distribution of audiometric thresholds at the test frequency within each group was similar to that for the population as a whole. Binaural TFS sensitivity declined with increasing age across the adult lifespan and with increasing hearing loss in old adulthood. For all test frequencies, both audiometric threshold and age were significantly negatively correlated with TFS-LF sensitivity (r ranging from −0.19 to −0.64) but the correlation was always significantly higher for age than for audiometric threshold. Regression analyses showed that the standardized regression coefficient was greater for age than for audiometric threshold, and that there was a significant interaction; the effect of increasing age among older listeners was greater when the hearing loss was ≥30 dB than when it was < 30 dB

    Forward masking of amplitude modulation across ears and its tuning in the modulation domain

    No full text
    Frequency selectivity in the amplitude modulation (AM) domain has been demonstrated using both simultaneous AM masking and forward AM masking. This has been explained using the concept of a modulation filter bank (MFB). Here, we assessed whether the MFB occurs before or after the point of binaural interaction in the auditory pathway by using forward masking in the AM domain in an ipsilateral condition (masker AM and signal AM applied to the left ear with an unmodulated carrier in the right ear) and a contralateral condition (masker AM applied to the right ear and signal AM applied to the left ear). The carrier frequency was 8 kHz, the signal AM frequency, fs, was 40 or 80 Hz, and the masker AM frequency ranged from 0.25 to 4 times fs. Contralateral forward AM masking did occur, but it was smaller than ipsilateral AM masking. Tuning in the AM domain was slightly sharper for ipsilateral than for contralateral masking, perhaps reflecting confusion of the signal and masker AM in the ipsilateral condition when their AM frequencies were the same. The results suggest that there might be an MFB both before and after the point in the auditory pathway where binaural interaction occurs

    No evidence for a link between noise exposure and auditory temporal processing for young adults with normal audiograms

    No full text
    The link between lifetime noise exposure and temporal processing abilities was investigated for 45 normal-hearing participants, recruited from a population of undergraduate students, aged 18 to 23 years. A self-report instrument was employed to assess the amount of neuropathic noise (here defined as sounds with levels exceeding approximately 80 dBA) each participant had been exposed to and sensitivity to temporal-fine-structure and temporal-envelope information was determined using frequency discrimination and envelope irregularity detection tasks, respectively. Despite sizable individual variability in all measures, correlations between noise exposure and the ability to process temporal cues were small and non-significan

    Predicting speech perception in older listeners with sensorineural hearing loss using automatic speech recognition

    No full text
    The objective of this study was to provide proof of concept that the speech intelligibility in quiet of unaided older hearing-impaired (OHI) listeners can be predicted by automatic speech recognition (ASR). Twenty-four OHI listeners completed three speech-identification tasks using speech materials of varying linguistic complexity and predictability (i.e., logatoms, words, and sentences). An ASR system was first trained on different speech materials and then used to recognize the same speech stimuli presented to the listeners but processed to mimic some of the perceptual consequences of age-related hearing loss experienced by each of the listeners: the elevation of hearing thresholds (by linear filtering), the loss of frequency selectivity (by spectrally smearing), and loudness recruitment (by raising the amplitude envelope to a power). Independently of the size of the lexicon used in the ASR system, strong to very strong correlations were observed between human and machine intelligibility scores. However, large root-mean-square errors (RMSEs) were observed for all conditions. The simulation of frequency selectivity loss had a negative impact on the strength of the correlation and the RMSE. Highest correlations and smallest RMSEs were found for logatoms, suggesting that the prediction system reflects mostly the functioning of the peripheral part of the auditory system. In the case of sentences, the prediction of human intelligibility was significantly improved by taking into account cognitive performance. This study demonstrates for the first time that ASR, even when trained on intact independent speech material, can be used to estimate trends in speech intelligibility of OHI listeners
    corecore