9,237 research outputs found

    How do you say ‘hello’? Personality impressions from brief novel voices

    Get PDF
    On hearing a novel voice, listeners readily form personality impressions of that speaker. Accurate or not, these impressions are known to affect subsequent interactions; yet the underlying psychological and acoustical bases remain poorly understood. Furthermore, hitherto studies have focussed on extended speech as opposed to analysing the instantaneous impressions we obtain from first experience. In this paper, through a mass online rating experiment, 320 participants rated 64 sub-second vocal utterances of the word ‘hello’ on one of 10 personality traits. We show that: (1) personality judgements of brief utterances from unfamiliar speakers are consistent across listeners; (2) a two-dimensional ‘social voice space’ with axes mapping Valence (Trust, Likeability) and Dominance, each driven by differing combinations of vocal acoustics, adequately summarises ratings in both male and female voices; and (3) a positive combination of Valence and Dominance results in increased perceived male vocal Attractiveness, whereas perceived female vocal Attractiveness is largely controlled by increasing Valence. Results are discussed in relation to the rapid evaluation of personality and, in turn, the intent of others, as being driven by survival mechanisms via approach or avoidance behaviours. These findings provide empirical bases for predicting personality impressions from acoustical analyses of short utterances and for generating desired personality impressions in artificial voices

    Auditory communication in domestic dogs: vocal signalling in the extended social environment of a companion animal

    Get PDF
    Domestic dogs produce a range of vocalisations, including barks, growls, and whimpers, which are shared with other canid species. The source–filter model of vocal production can be used as a theoretical and applied framework to explain how and why the acoustic properties of some vocalisations are constrained by physical characteristics of the caller, whereas others are more dynamic, influenced by transient states such as arousal or motivation. This chapter thus reviews how and why particular call types are produced to transmit specific types of information, and how such information may be perceived by receivers. As domestication is thought to have caused a divergence in the vocal behaviour of dogs as compared to the ancestral wolf, evidence of both dog–human and human–dog communication is considered. Overall, it is clear that domestic dogs have the potential to acoustically broadcast a range of information, which is available to conspecific and human receivers. Moreover, dogs are highly attentive to human speech and are able to extract speaker identity, emotional state, and even some types of semantic information

    Norm-based coding of voice identity in human auditory cortex

    Get PDF
    Listeners exploit small interindividual variations around a generic acoustical structure to discriminate and identify individuals from their voice—a key requirement for social interactions. The human brain contains temporal voice areas (TVA) [1] involved in an acoustic-based representation of voice identity [2, 3, 4, 5 and 6], but the underlying coding mechanisms remain unknown. Indirect evidence suggests that identity representation in these areas could rely on a norm-based coding mechanism [4, 7, 8, 9, 10 and 11]. Here, we show by using fMRI that voice identity is coded in the TVA as a function of acoustical distance to two internal voice prototypes (one male, one female)—approximated here by averaging a large number of same-gender voices by using morphing [12]. Voices more distant from their prototype are perceived as more distinctive and elicit greater neuronal activity in voice-sensitive cortex than closer voices—a phenomenon not merely explained by neuronal adaptation [13 and 14]. Moreover, explicit manipulations of distance-to-mean by morphing voices toward (or away from) their prototype elicit reduced (or enhanced) neuronal activity. These results indicate that voice-sensitive cortex integrates relevant acoustical features into a complex representation referenced to idealized male and female voice prototypes. More generally, they shed light on remarkable similarities in cerebral representations of facial and vocal identity

    BigEAR: Inferring the Ambient and Emotional Correlates from Smartphone-based Acoustic Big Data

    Get PDF
    This paper presents a novel BigEAR big data framework that employs psychological audio processing chain (PAPC) to process smartphone-based acoustic big data collected when the user performs social conversations in naturalistic scenarios. The overarching goal of BigEAR is to identify moods of the wearer from various activities such as laughing, singing, crying, arguing, and sighing. These annotations are based on ground truth relevant for psychologists who intend to monitor/infer the social context of individuals coping with breast cancer. We pursued a case study on couples coping with breast cancer to know how the conversations affect emotional and social well being. In the state-of-the-art methods, psychologists and their team have to hear the audio recordings for making these inferences by subjective evaluations that not only are time-consuming and costly, but also demand manual data coding for thousands of audio files. The BigEAR framework automates the audio analysis. We computed the accuracy of BigEAR with respect to the ground truth obtained from a human rater. Our approach yielded overall average accuracy of 88.76% on real-world data from couples coping with breast cancer.Comment: 6 pages, 10 equations, 1 Table, 5 Figures, IEEE International Workshop on Big Data Analytics for Smart and Connected Health 2016, June 27, 2016, Washington DC, US

    Speech characteristics of monozygotic twins and a same-sex sibling: an acoustic case study of coarticulation patterns in read speech

    Get PDF
    This case study reports on an acoustic investigation of the motor speech characteristics of a set of young adult male monozygotic (MZ) twins and compares them to those of an age- and sex-matched sibling who participated in the study 2 years later to match for demographic factors. Coarticulation patterns were investigated from read samples of consonant-vowel sequences in monosyllabic words containing a variety of consonants and vowels. This was done by examining F2 vowel onsets and F2 vowel targets, plotted as F2 locus equations. Data were processed for between-sibling differences using a number of statistical tests. Results indicated that the MZ twins displayed F2 parameters and coarticulation patterns which were more similar than those of their age- and sex-matched sibling. The results of this case study therefore suggest that acoustic phonetic parameters used to index coarticulation patterns have the potential to profile some of the similarities and differences in the speech characteristics of genetically related individuals

    Voice parameters predict sex-specific body morphology in men and women

    Get PDF
    Studies of several mammalian species confirm that formant frequencies (vocal tract resonances) predict height and weight better than does fundamental frequency (F0, perceived as pitch) in same-sex adults due to differential anatomical constraints. However, our recent meta-analysis (Pisanski et al., 2014, Animal Behaviour, 95, 89–99) indicated that formants and F0 could explain no more than 10% and 2% of the variance in human height, respectively, controlling for sex and age. Here, we examined whether other voice parameters, many of which are affected by sex hormones, can indicate additional variance in human body size or shape, and whether these relationships differ between the sexes. Using a cross-cultural sample of 700 men and women, we examined relationships among 19 voice parameters (minimum–maximum F0, mean F0, F0 variability, formant-based vocal tract length estimates, shimmer, jitter, harmonics-to-noise ratio) and eight indices of body size or shape (height, weight, body mass index, hip, waist and chest circumferences, waist-to-hip ratio, chest-to-hip ratio). Our results confirm that formant measures explain the most variance in heights and weights of men and women, whereas shimmer, jitter and harmonics-to-noise ratio do not indicate height, weight or body mass index in either sex. In contrast, these perturbation and noise parameters, in addition to F0 range and variability, explained more variance in body shape than did formants or mean F0, particularly among men. Shimmer or jitter explained the most variance in men's hip circumferences (12%) and chest-to-hip ratios (6%), whereas harmonics-to-noise ratio and formants explained the most variance in women's waist-to-hip ratios (11%), and significantly more than in men's waist-to-hip ratios. Our study represents the most comprehensive analysis of vocal indicators of human body size to date and offers a foundation for future research examining the hormonal mechanisms of voice production in humans and perceptual playback experiments

    The sound of concepts: The link between auditory and conceptual brain systems

    Get PDF
    Concepts in long-term memory are important building blocks of human cognition and are the basis for object recognition, language and thought. While it is well accepted that concepts are comprised of features related to sensory object attributes, it is still unclear how these features are represented in the brain. Of central interest is whether concepts are essentially grounded in perception. This would imply a common neuroanatomical substrate for perceptual and conceptual processing. Here we show using functional magnetic resonance imaging and recordings of event-related potentials that acoustic conceptual features rapidly recruit auditory areas even when implicitly presented through visual words. Recognizing words denoting objects for which acoustic features are highly relevant (e.g. "telephone") suffices to ignite cell assemblies in the posterior superior and middle temporal gyrus (pSTG/MTG) that were also activated by listening to real sounds. Activity in pSTG/MTG had an onset of 150 ms and increased parametrically as a function of acoustic feature relevance. Both findings suggest a conceptual origin of this effect rather than post-conceptual strategies such as imagery. The presently demonstrated link between auditory and conceptual brain systems parallels observations in other memory systems suggesting that modality-specificity represents a general organizational principle in cortical memory representation. The understanding of concepts as a partial reinstatement of brain activity during perception stresses the necessity of rich sensory experiences for concept acquisition. The modality-specific nature of concepts could also explain the difficulties in achieving a consensus about overall definitions of abstract concepts such as freedom or justice unless embedded in a concrete, experienced situation

    Jaw Rotation in Dysarthria Measured With a Single Electromagnetic Articulography Sensor

    Get PDF
    Purpose This study evaluated a novel method for characterizing jaw rotation using orientation data from a single electromagnetic articulography sensor. This method was optimized for clinical application, and a preliminary examination of clinical feasibility and value was undertaken. Method The computational adequacy of the single-sensor orientation method was evaluated through comparisons of jaw-rotation histories calculated from dual-sensor positional data for 16 typical talkers. The clinical feasibility and potential value of single-sensor jaw rotation were assessed through comparisons of 7 talkers with dysarthria and 19 typical talkers in connected speech. Results The single-sensor orientation method allowed faster and safer participant preparation, required lower data-acquisition costs, and generated less high-frequency artifact than the dual-sensor positional approach. All talkers with dysarthria, regardless of severity, demonstrated jaw-rotation histories with more numerous changes in movement direction and reduced smoothness compared with typical talkers. Conclusions Results suggest that the single-sensor orientation method for calculating jaw rotation during speech is clinically feasible. Given the preliminary nature of this study and the small participant pool, the clinical value of such measures remains an open question. Further work must address the potential confound of reduced speaking rate on movement smoothness
    corecore