1,023 research outputs found
Teaching the Prosody of Emotive Communication in a Second Language
The expression and the perception of emotional states in a foreign language represent a difficult task for the learners. One of the reasons is the fact that, more than other aspects related to speech, the expression of emotional states in second language requires full control of the prosodic resources that contribute to their realization. The aim of this chapter is to give an overview of the main tenets of the interface between prosody and pragmatic competence in L2 and in particular the expression and perception of emotions. The chapter will also outline some of the outcomes of the research in the field, focusing on experimental studies that have been conducted with learners of Italian as L2. The second part of the chapter will be devoted on the instructional practice aimed at developing the awareness of pragmatic-prosodic aspects of emotive communication in speech. Teaching practices such as a training focused on the expression of emotions (anger, joy, sadness, disgust, fear, and surprise) and video dubbing projects have proven to be useful tools to improve the performance of learners both in production and in perception of prosodic patterns of emotional communication
Lexical and Prosodic Pitch Modifications in Cantonese Infant-directed Speech
Published online 03 February 2021The functions of acoustic-phonetic modifications in infant-directed speech (IDS) remain a
question: do they specifically serve to facilitate language learning via enhanced phonemic
contrasts (the hyperarticulation hypothesis) or primarily to improve communication via
prosodic exaggeration (the prosodic hypothesis)? The study of lexical tones provides a
unique opportunity to shed light on this, as lexical tones are phonemically contrastive,
yet their primary cue, pitch, is also a prosodic cue. This study investigated Cantonese
IDS and found increased intra-talker variation of lexical tones, which more likely posed
a challenge to rather than facilitated phonetic learning. Although tonal space was
expanded which could facilitate phonetic learning, its expansion was a function of
overall intonational modifications. Similar findings were observed in speech to pets
who should not benefit from larger phonemic distinction. We conclude that lexicaltone
adjustments in IDS mainly serve to broadly enhance communication rather than
specifically increase phonemic contrast for learners.This work was supported by the University Grants Committee (HKSAR) (RGC34000118), the Innovation and
Technology Fund (HKSAR) (ITS/067/18), Dr. Stanley Ho Medical Development Foundation, and the
Global Parent Child Resource Centre Limited. The second author’s work is supported by the Basque
Government through the BERC 2018-2021 program and by the Spanish Ministry of Science and
Innovation through the Ramon y Cajal Research Fellowship, PID2019-105528GA-I00
Children\u27s Sensitivity to Pitch Variation in Language
Children acquire consonant and vowel categories by 12 months, but take much longer to learn to interpret perceptible variation. This dissertation considers children’s interpretation of pitch variation. Pitch operates, often simultaneously, at different levels of linguistic structure. English-learning children must disregard pitch at the lexical level—since English is not a tone language—while still attending to pitch for its other functions. Chapters 1 and 5 outline the learning problem and suggest ways children might solve it. Chapter 2 demonstrates that 2.5-year-olds know pitch cannot differentiate words in English. Chapter 3 finds that not until age 4–5 do children correctly interpret pitch cues to emotions. Chapter 4 demonstrates some sensitivity between 2.5 and 5 years to the pitch cue to lexical stress, but continuing difficulties at the older ages. These findings suggest a late trajectory for interpretation of prosodic variation; throughout, I propose explanations for this protracted time-course
The listening talker: A review of human and algorithmic context-induced modifications of speech
International audienceSpeech output technology is finding widespread application, including in scenarios where intelligibility might be compromised - at least for some listeners - by adverse conditions. Unlike most current algorithms, talkers continually adapt their speech patterns as a response to the immediate context of spoken communication, where the type of interlocutor and the environment are the dominant situational factors influencing speech production. Observations of talker behaviour can motivate the design of more robust speech output algorithms. Starting with a listener-oriented categorisation of possible goals for speech modification, this review article summarises the extensive set of behavioural findings related to human speech modification, identifies which factors appear to be beneficial, and goes on to examine previous computational attempts to improve intelligibility in noise. The review concludes by tabulating 46 speech modifications, many of which have yet to be perceptually or algorithmically evaluated. Consequently, the review provides a roadmap for future work in improving the robustness of speech output
Acoustic-phonetic realisation of Polish syllable prominence: a corpus study.
Malisz Z, Wagner P. Acoustic-phonetic realisation of Polish syllable prominence: a corpus study. In: Gibbon D, Hirst D, Campbell N, eds. Rhythm, melody and harmony in speech. Studies in honour of Wiktor Jassem. Speech and Language Technology. Vol 14/15. Poznań, Poland; 2012: 105-114
Cracking the social code of speech prosody using reverse correlation
Human listeners excel at forming high-level social representations about each other, even from the briefest of utterances. In particular, pitch is widely recognized as the auditory dimension that conveys most of the information about a speaker's traits, emotional states, and attitudes. While past research has primarily looked at the influence of mean pitch, almost nothing is known about how intonation patterns, i.e., finely tuned pitch trajectories around the mean, may determine social judgments in speech. Here, we introduce an experimental paradigm that combines state-of-the-art voice transformation algorithms with psychophysical reverse correlation and show that two of the most important dimensions of social judgments, a speaker's perceived dominance and trustworthiness, are driven by robust and distinguishing pitch trajectories in short utterances like the word "Hello," which remained remarkably stable whether male or female listeners judged male or female speakers. These findings reveal a unique communicative adaptation that enables listeners to infer social traits regardless of speakers' physical characteristics, such as sex and mean pitch. By characterizing how any given individual's mental representations may differ from this generic code, the method introduced here opens avenues to explore dysprosody and social-cognitive deficits in disorders like autism spectrum and schizophrenia. In addition, once derived experimentally, these prototypes can be applied to novel utterances, thus providing a principled way to modulate personality impressions in arbitrary speech signals
Recommended from our members
Prosodic Predictors of Upcoming Positive or Negative Content in Spoken Messages
This article examines potential prosodic predictors of emotional speech in utterances perceived as conveying that good or bad news is about to be delivered. Speakers were asked to call an experimental confederate to inform her about whether or not she had been given a job she had applied for. A perception study was then performed in which initial fragments of the recorded utterances, not containing any explicit lexical cues to emotional content, were presented to listeners who had to rate whether good or bad news would follow the utterance. The utterances were then examined to discover acoustic and prosodic features that distinguished between good and bad news. It was found that speakers in the production study were not simply reflecting their own positive or negative mood during the experiment, but rather appeared to be influenced by the valence of the positive or negative message they were preparing to deliver. Positive and negative utterances appeared to be judged differently with respect to a number of perceived attributes of the speakers’ voices (like sounding hesitant or nervous). These attributes correlated with a number of automatically obtained acoustic features
Eyebrow raising in dialogue: discourse structure, utterance function, and pitch accents
Some studies have suggested a relationship between eyebrow raising and different aspects of the verbal message, but our knowledge about this link is still
very limited. If we could establish and characterise a relation between eyebrow
raises and the linguistic signal we could better understand human multimodal
communication behaviour. We could also improve the credibility and efficiency
of computer animated conversational agents in multimodal communication systems.This thesis investigated eyebrow raising in a corpus of task-oriented English
dialogues. Applying a standard dialogue coding scheme (Conversational Game
Analysis, Carletta et al., 1997), eyebrow raises were studied in connection with
discourse structure and utterance function. Supporting the prediction, more
frequent and longer eyebrow raising occurred in the initial utterance of highlevel discourse segments than anywhere else in the dialogue (where 'high-level
discourse segment' = transaction, and 'utterance' = move, following Carletta et al.).
Additionally, eyebrow raises were more frequent in instructions than in requests
for or acknowledgements of information. Instructions also had longer eyebrow
raising than any other type of utterance. Contrary to the prediction, the start of a
lower-level discourse segment (conversational game) did not have more eyebrow
raising than any other position in the dialogue, and queries did not have more
eyebrow raising than any other type of utterance.Eyebrow raises were also studied in relation to intonational events, namely pitch
accents. Results showed evidence of alignment between the brow raise start and
the start of a pitch accent. Most pitch accents were not associated with brow raising, but when brow raises occurred they tended to immediately precede a pitch
accent on the speech signal. To investigate what could explain the alignment
between the two events, pitch accents aligned with eyebrow raises were compared to all other pitch accents in terms of: phonological characteristics (primary
vs. secondary pitch accents, and downstep-initial vs. non-initial pitch accents), information structure (given vs. new information in referring expressions, and the
last quarter vs. earlier parts of the utterance length) and type of utterance in which
they occurred (instruction vs. non-instruction). Those comparisons suggested that
brow raises may be aligned more frequently with pitch accents in downstepinitial position and in instructions. No differences were found in terms of information structure or between primary/secondary accents.The results provide evidence of a link between eyebrow raising and spoken language. Eyebrow raises may signal the start of linguistic units such as discourse
segments and some prosodic phenomena, they may be related to utterance function, and they are aligned with pitch accents. Possible linguistic functions are
proposed, such as structuring and emphasising information in the verbal message
- …