10,726 research outputs found

    Acoustic cues for the korean stop contrast-dialectal variation

    Get PDF
    In this study, cross-dialectal variation in the use of the acoustic cues of VOT and F0 to mark the laryngeal contrast in Korean stops is examined with Chonnam Korean and Seoul Korean. Prior experimental results (Han & Weitzman, 1970; Hardcastle, 1973; Jun, 1993 &1998; Kim, C., 1965) show that pitch values in the vowel onset following the target stop consonants play a supplementary role to VOT in designating the three contrastive laryngeal categories. F0 contours are determined in part by the intonational system of a language, which raises the question of how the intonational system interacts with phonological contrasts. Intonational difference might be linked to dissimilar patterns in using the complementary acoustic cues of VOT and F0. This hypothesis is tested with 6 Korean speakers, three Seoul Korean and three Chonnam Korean speakers. The results show that Chonnam Korean involves more 3-way VOT and a 2-way distinction in F0 distribution in comparison to Seoul Korean that shows more 3-way F0 distribution and a 2-way VOT distinction. The two acoustic cues are complementary in that one cue is rather faithful in marking 3-way contrast, while the other cue marks the contrast less distinctively. It also seems that these variations are not completely arbitrary, but linked to the phonological characteristics in dialects. Chonnam Korean, in which the initial tonal realization in the accentual phrase is expected to be more salient, tends to minimize the F0 perturbation effect from the preceding consonants by taking more overlaps in F0 distribution. And a 3-way distribution of VOT in Chonnam Korean, as compensation, can be also understood as a durational sensitivity. Without these characteristics, Seoul Korean shows relatively more overlapping distribution in VOT and more 3-way separation in F0 distribution

    Context-related acoustic variation in male fallow deer (Dama dama) groans

    Get PDF
    While social and behavioural contexts are known to affect the acoustic structure of vocal signals in several mammal species, few studies have investigated context-related acoustic variation during inter-sexual advertisement and/or intra-sexual competition. Here we recorded male fallow deer groans during the breeding season and investigated how key acoustic parameters (fundamental frequency and formant frequencies) vary as a function of the social context in which they are produced. We found that in the presence of females, male fallow deer produced groans with higher mean fundamental frequency when vocal males were also present than they did when no vocal males were in close vicinity. We attribute this to the increased arousal state typically associated with this context. In addition, groan minimum formant frequency spacing was slightly, but significantly lower (indicating marginally more extended vocal tracts) when males were alone than when potential mates and/or competitors were nearby. This indicates that, contrary to our predictions, male fallow deer do not exaggerate the acoustic impression of their body size by further lowering their formant frequencies in the presence of potential mating partners and competitors. Furthermore, since the magnitude of the variation in groan minimum formant frequency spacing remains small compared to documented inter-individual differences, our findings are consistent with the hypothesis that formants are reliable static cues to body size during intra- and inter-sexual advertisement that do not concurrently encode dynamic motivation-related informatio

    Effect of formant frequency spacing on perceived gender in pre-pubertal children's voices

    Get PDF
    <div><p>Background</p><p>It is usually possible to identify the sex of a pre-pubertal child from their voice, despite the absence of sex differences in fundamental frequency at these ages. While it has been suggested that the overall spacing between formants (formant frequency spacing - ΔF) is a key component of the expression and perception of sex in children's voices, the effect of its continuous variation on sex and gender attribution has not yet been investigated.</p><p>Methodology/Principal findings</p><p>In the present study we manipulated voice ΔF of eight year olds (two boys and two girls) along continua covering the observed variation of this parameter in pre-pubertal voices, and assessed the effect of this variation on adult ratings of speakers' sex and gender in two separate experiments. In the first experiment (sex identification) adults were asked to categorise the voice as either male or female. The resulting identification function exhibited a gradual slope from male to female voice categories. In the second experiment (gender rating), adults rated the voices on a continuum from “masculine boy” to “feminine girl”, gradually decreasing their masculinity ratings as ΔF increased.</p><p>Conclusions/Significance</p><p>These results indicate that the role of ΔF in voice gender perception, which has been reported in adult voices, extends to pre-pubertal children's voices: variation in ΔF not only affects the perceived sex, but also the perceived masculinity or femininity of the speaker. We discuss the implications of these observations for the expression and perception of gender in children's voices given the absence of anatomical dimorphism in overall vocal tract length before puberty.</p></div

    Auditory communication in domestic dogs: vocal signalling in the extended social environment of a companion animal

    Get PDF
    Domestic dogs produce a range of vocalisations, including barks, growls, and whimpers, which are shared with other canid species. The source–filter model of vocal production can be used as a theoretical and applied framework to explain how and why the acoustic properties of some vocalisations are constrained by physical characteristics of the caller, whereas others are more dynamic, influenced by transient states such as arousal or motivation. This chapter thus reviews how and why particular call types are produced to transmit specific types of information, and how such information may be perceived by receivers. As domestication is thought to have caused a divergence in the vocal behaviour of dogs as compared to the ancestral wolf, evidence of both dog–human and human–dog communication is considered. Overall, it is clear that domestic dogs have the potential to acoustically broadcast a range of information, which is available to conspecific and human receivers. Moreover, dogs are highly attentive to human speech and are able to extract speaker identity, emotional state, and even some types of semantic information

    Production and perception of English Word Final Stops By Malay Speakers

    Get PDF
    A few influential speech studies have been carried out using established speech learning models, which confirmed that the analysis of first language (L1) and second language (L2) at a phonemic level provides only a partial view of deeper relationships between languages in contact. Therefore, studies focusing on cross-language phonetic differences as a causative factor in L2 learner difficulties have been proposed to understand second language learners’ (L2) speech production and how listeners respond perceptually to the phonetic properties of L2. This paper presents a study of the production and perception of the final stops by English learners (L2) whose first language is Malay (L1). A total of 23 students, comprising 16 male and 7 female Malay subjects (L1 as Malay and their L2 as English) with normal hearing and speech development participated in this study. A short interview was conducted in order to gain background information about information about each subject, to introduce them to the study, to inform them about the process of recording, the materials to be used in the recording session, and how the materials should be managed during recording time. Acoustic measurements of selected segments occurring in word final positions (via spectrographic analysis, syllable rhyme duration and phonation) were taken. Results of the voicing contrast realisation in Malay accented English and Malaysian listeners' perceptual identification/discrimination abilities with final voiced/voiceless stops in Malay and English are presented and discussed. The findings revealed that the Malay students’ realisation of final stops in L2 is largely identical to their L1. In addition, the results also showed that accurate ‘perception’ may not always lead to accurate ‘production’

    Acoustic cues to tonal contrasts in Mandarin: Implications for cochlear implants

    Get PDF
    The present study systematically manipulated three acoustic cues-fundamental frequency (f0), amplitude envelope, and duration-to investigate their contributions to tonal contrasts in Mandarin. Simplified stimuli with all possible combinations of these three cues were presented for identification to eight normal-hearing listeners, all native speakers of Mandarin from Taiwan. The f0 information was conveyed either by an f0-controlled sawtooth carrier or a modulated noise so as to compare the performance achievable by a clear indication of voice f0 and what is possible with purely temporal coding of f0. Tone recognition performance with explicit f0 was much better than that with any combination of other acoustic cues (consistently greater than 90% correct compared to 33%-65%; chance is 25%). In the absence of explicit f0, the temporal coding of f0 and amplitude envelope both contributed somewhat to tone recognition, while duration had only a marginal effect. Performance based on these secondary cues varied greatly across listeners. These results explain the relatively poor perception of tone in cochlear implant users, given that cochlear implants currently provide only weak cues to f0, so that users must rely upon the purely temporal (and secondary) features for the perception of tone. (c) 2008 Acoustical Society of America

    Listeners normalize speech for contextual speech rate even without an explicit recognition task

    No full text
    Speech can be produced at different rates. Listeners take this rate variation into account by normalizing vowel duration for contextual speech rate: An ambiguous Dutch word /m?t/ is perceived as short /mAt/ when embedded in a slow context, but long /ma:t/ in a fast context. Whilst some have argued that this rate normalization involves low-level automatic perceptual processing, there is also evidence that it arises at higher-level cognitive processing stages, such as decision making. Prior research on rate-dependent speech perception has only used explicit recognition tasks to investigate the phenomenon, involving both perceptual processing and decision making. This study tested whether speech rate normalization can be observed without explicit decision making, using a cross-modal repetition priming paradigm. Results show that a fast precursor sentence makes an embedded ambiguous prime (/m?t/) sound (implicitly) more /a:/-like, facilitating lexical access to the long target word "maat" in a (explicit) lexical decision task. This result suggests that rate normalization is automatic, taking place even in the absence of an explicit recognition task. Thus, rate normalization is placed within the realm of everyday spoken conversation, where explicit categorization of ambiguous sounds is rare
    corecore