3,126 research outputs found

    The interaction of syllabification and voicing perception in american english

    Get PDF
    The current paper explores these two sorts of phonetic explanations of the relationship between syllabic position and the voicing contrast in American English. It has long been observed that the contrast between, for example, /p/ and /b/ is expressed differently, depending on the position of the stop with respect to the vowel. Preceding a vowel within a syllable, the contrast is largely one of aspiration. /p/ is aspirated, while /b/ is voiceless, or in some dialects voiced or even an implosive. Following a vowel within a syllable, both /p/ and /b/ both tend to lack voicing in the closure and the contrast is expressed largely by dynamic differences in the transition between the previous vowel and the stop. Here, vowel and closure duration are negatively correlated such that the /p/ has a shorter vowel and longer closure duration. This difference is often enhanced by the addition of glottalization to /p/. In addition to these differences, there are additional differences connected to higher-level organization involving stress and feet edges. To make the current discussion more tractable, we will restrict ourselves to the two conditions (CV and VC) laid out above

    Brittany Bernal - Sensorimotor Adaptation of Vowel Production in Stop Consonant Contexts

    Get PDF
    The purpose of this research is to measure the compensatory and adaptive articulatory response to shifted formants in auditory feedback to compare the resulting amount of sensorimotor learning that takes place in speakers upon saying the words /pep/ and /tet/. These words were chosen in order to analyze the coarticulatory effects of voiceless consonants /p/ and /t/ on sensorimotor adaptation of the vowel /e/. The formant perturbations were done using the Audapt software, which takes an input speech sample and plays it back to the speaker in real-time via headphones. Formants are high-energy acoustic resonance patterns measured in hertz that reflect positions of articulators during the production of speech sounds. The two lowest frequency formants (F1 and F2) can uniquely distinguish among the vowels of American English. For this experiment, Audapt shifted F1 down and F2 up, and those who adapt were expected to shift in the opposite direction of the perturbation. The formant patterns and vowel boundaries were analyzed using TF32 and S+ software, which led to conclusions about the adaptive responses. Manipulating auditory feedback by shifting formant values is hypothesized to elicit sensorimotor adaptation, a form of short-term motor learning. The amount of adaptation is expected to be greater for the word /pep/ rather than /tet/ because there is less competition for articulatory placement of the tongue during production of bilabial consonants. This methodology could be further developed to help those with motor speech disorders remedy their speech errors with much less conscious effort than traditional therapy techniques.https://epublications.marquette.edu/mcnair_2013/1008/thumbnail.jp

    Changes in the McGurk Effect Across Phonetic Contexts

    Full text link
    To investigate the process underlying audiovisual speech perception, the McGurk illusion was examined across a range of phonetic contexts. Two major changes were found. First, the frequency of illusory /g/ fusion percepts increased relative to the frequency of illusory /d/ fusion percepts as vowel context was shifted from /i/ to /a/ to /u/. This trend could not be explained by biases present in perception of the unimodal visual stimuli. However, the change found in the McGurk fusion effect across vowel environments did correspond systematically with changes in second format frequency patterns across contexts. Second, the order of consonants in illusory combination percepts was found to depend on syllable type. This may be due to differences occuring across syllable contexts in the timecourses of inputs from the two modalities as delaying the auditory track of a vowel-consonant stimulus resulted in a change in the order of consonants perceived. Taken together, these results suggest that the speech perception system either fuses audiovisual inputs into a visually compatible percept with a similar second formant pattern to that of the acoustic stimulus or interleaves the information from different modalities, at a phonemic or subphonemic level, based on their relative arrival times.National Institutes of Health (R01 DC02852

    Lexical Effects in Perception of Tamil Geminates

    Get PDF
    Lexical status effects are a phenomenon in which listeners use their prior lexical knowledge of a language to identify ambiguous speech sounds in a word based on its word or nonword status. This phenomenon has been demonstrated for ambiguous initial English consonants (one example being the Ganong Effect, a phenomenon in which listeners perceive an ambiguous speech sound as a phoneme that would complete a real word rather than a nonsense word) as a supporting factor for top-down lexical processing affecting listeners' subsequent acoustic judgement, but not for ambiguous mid-word consonants in non-English languages. In this experiment, we attempt to look at ambiguous mid-word consonants with Tamil, a South Asian language in order to see if the same top-down lexical effect was applicable outside of English. These Tamil consonants can present as either singletons (single speech sounds) or geminates (doubled speech sounds).We hypothesized that by creating ambiguous stimuli between a geminate word kuppam and a singleton non-word like kubam, participants would be more likely to perceive the ambiguous sound as a phoneme that completes the real word rather than the nonword (in this case, perceiving the ambiguous sound as a /p/ for kuppam instead of kubam). Participants listened to the ambiguous stimuli in two separate sets of continua (kuppam/suppam and nakkam/pakkam) and then indicated which word they heard in a four-alternative forced choice word identification task. Results showed that participants identified the ambiguous sounds as the sound that completed the actual word, but only for one set of continua (kuppam/suppam). These data suggest that there may be strong top-down lexical effects for ambiguous sounds in certain stimuli in Tamil, but not others.No embargoAcademic Major: LinguisticsAcademic Major: Psycholog

    On the avoidance of voiced sibilant affricates

    Get PDF
    In this paper it is argued that several typologically unrelated languages share the tendency to avoid voiced sibilant affricates. This tendency is explained by appealing to the phonetic properties of the sounds, and in particular to their aerodynamic characteristics. On the basis of experimental evidence it is shown that conflicting air pressure requirements for maintaining voicing and frication are responsible for the avoidance of voiced affricates. In particular, the air pressure released from the stop phase of the affricate is too high to maintain voicing which in consequence leads to a devoicing of the frication part

    A silent speech system based on permanent magnet articulography and direct synthesis

    Get PDF
    In this paper we present a silent speech interface (SSI) system aimed at restoring speech communication for individuals who have lost their voice due to laryngectomy or diseases affecting the vocal folds. In the proposed system, articulatory data captured from the lips and tongue using permanent magnet articulography (PMA) are converted into audible speech using a speaker-dependent transformation learned from simultaneous recordings of PMA and audio signals acquired before laryngectomy. The transformation is represented using a mixture of factor analysers, which is a generative model that allows us to efficiently model non-linear behaviour and perform dimensionality reduction at the same time. The learned transformation is then deployed during normal usage of the SSI to restore the acoustic speech signal associated with the captured PMA data. The proposed system is evaluated using objective quality measures and listening tests on two databases containing PMA and audio recordings for normal speakers. Results show that it is possible to reconstruct speech from articulator movements captured by an unobtrusive technique without an intermediate recognition step. The SSI is capable of producing speech of sufficient intelligibility and naturalness that the speaker is clearly identifiable, but problems remain in scaling up the process to function consistently for phonetically rich vocabularies

    Postalveolar fricatives in Slavic languages as retroflexes

    Get PDF
    The present study poses the question on what phonetic and phonological grounds postalveolar fricatives in Polish can be analyzed as retroflex and whether postalveolar fricatives in other Slavic languages are retroflex as well. Velarization and incompatibility with front vowels are introduced as articulatory criteria for retroflexion, based on crosslinguistic data. According to these criteria, Polish and Russian have retroflex fricatives, whereas Bulgarian and Czech do not. In a phonological representation of these Slavic retroflexes, the necessity of perceptual features is shown. Lastly, it is illustrated that palatalization of retroflex fricatives both in Slavic languages and more generally causes a phonetic and phonological change to a non-retroflex sound

    The weight of phonetic substance in the structure of sound inventories

    Get PDF
    In the research field initiated by Lindblom & Liljencrants in 1972, we illustrate the possibility of giving substance to phonology, predicting the structure of phonological systems with nonphonological principles, be they listener-oriented (perceptual contrast and stability) or speaker-oriented (articulatory contrast and economy). We proposed for vowel systems the Dispersion-Focalisation Theory (Schwartz et al., 1997b). With the DFT, we can predict vowel systems using two competing perceptual constraints weighted with two parameters, respectively λ and α. The first one aims at increasing auditory distances between vowel spectra (dispersion), the second one aims at increasing the perceptual salience of each spectrum through formant proximities (focalisation). We also introduced new variants based on research in physics - namely, phase space (λ,α) and polymorphism of a given phase, or superstructures in phonological organisations (VallĂ©e et al., 1999) which allow us to generate 85.6% of 342 UPSID systems from 3- to 7-vowel qualities. No similar theory for consonants seems to exist yet. Therefore we present in detail a typology of consonants, and then suggest ways to explain plosive vs. fricative and voiceless vs. voiced consonants predominances by i) comparing them with language acquisition data at the babbling stage and looking at the capacity to acquire relatively different linguistic systems in relation with the main degrees of freedom of the articulators; ii) showing that the places “preferred” for each manner are at least partly conditioned by the morphological constraints that facilitate or complicate, make possible or impossible the needed articulatory gestures, e.g. the complexity of the articulatory control for voicing and the aerodynamics of fricatives. A rather strict coordination between the glottis and the oral constriction is needed to produce acceptable voiced fricatives (Mawass et al., 2000). We determine that the region where the combinations of Ag (glottal area) and Ac (constriction area) values results in a balance between the voice and noise components is indeed very narrow. We thus demonstrate that some of the main tendencies in the phonological vowel and consonant structures of the world’s languages can be explained partly by sensorimotor constraints, and argue that actually phonology can take part in a theory of Perception-for-Action-Control
    • 

    corecore