699 research outputs found
Sex-specific fundamental and formant frequency patterns in a cross-sectional study
An extensive developmental acoustic study of the speech patterns of children and adults was reported by Lee and colleagues [Lee et al., J. Acoust. Soc. Am. 105, 1455-1468 (1999)]. This paper presents a reexamination of selected fundamental frequency and formant frequency data presented in their report for 10 monophthongs by investigating sex-specific and developmental patterns using two different approaches. The first of these includes the investigation of age- and sex-specific formant frequency patterns in the monophthongs. The second, the investigation of fundamental frequency and formant frequency data using the critical band rate (bark) scale and a number of acoustic-phonetic dimensions of the monophthongs from an age- and sex-specific perspective. These acoustic-phonetic dimensions include: vowel spaces and distances from speaker centroids; frequency differences between the formant frequencies of males and females; vowel openness/closeness and frontness/backness; the degree of vocal effort; and formant frequency ranges. Both approaches reveal both age- and sex-specific development patterns which also appear to be dependent on whether vowels are peripheral or non-peripheral. The developmental emergence of these sex-specific differences are discussed with reference to anatomical, physiological, sociophonetic and culturally determined factors. Some directions for further investigation into the age-linked sex differences in speech across the lifespan are also proposed
Potential of the Cone Pressuremeter Test for obtaining stiffness degradation for offshore wind turbine monopile foundations
publishedVersio
A typology of laterals in twelve English dialects
Allophonic patterns of variation in English laterals have been well studied in phonetics and phonology for decades, but establishing broad generalizations across varieties has proven challenging. In this study, we advance a typology of onset/coda lateral distinctions in English, using crowdsourced recordings from 95 speakers across twelve dialects of Anglo (UK) English. Results confirm the existence of dialects with and without onset/coda distinctions, and conditional inference trees are used to identity three main patterns in the data: (1) clear onsets and dark codas; (2) intermediate/dark onsets and dark codas, but with a positional distinction intact; (3) dark onsets and dark codas, with minimal or no distinctions between positions
Automatic recognition of schwa variants in spontaneous Hungarian speech
This paper analyzes the nature of the process involved in optional vowel reduction in Hungarian, and the acoustic structure of schwa variants in spontaneous speech. The study focuses on the acoustic patterns of both the basic realizations of Hungarian vowels and their realizations as neutral vowels (schwas), as well as on the design, implementation, and evaluation of a set of algorithms for the recognition of both types of realizations from the speech waveform. The authors address the question whether schwas form a unified group of vowels or they show some dependence on the originally intended articulation of the vowel they stand for. The acoustic study uses a database consisting of over 4,000 utterances extracted from continuous speech, and recorded from 19 speakers. The authors propose methods for the recognition of neutral vowels depending on the various vowels they replace in spontaneous speech. Mel-Frequency Cepstral Coefficients are calculated and used for the training of Hidden Markov Models. The recognition system was trained on 2,500 utterances and then tested on 1,500 utterances. The results show that a neutral vowel can be detected in 72% of all occurrences. Stressed and unstressed syllables can be distinguished in 92% of all cases. Neutralized vowels do not form a unified group of phoneme realizations. The pronunciation of schwa heavily depends on the original articulation configuration of the intended vowel
The Role of Native Language and the Fundamental Design of the Auditory System in Detecting Rhythm Changes
Accepted December 13, 2018Purpose: We investigated whether rhythm discrimination
is mainly driven by the native language of the listener or
by the fundamental design of the human auditory system
and universal cognitive mechanisms shared by all people
irrespective of rhythmic patterns in their native language.
Method: In multiple experiments, we asked participants to
listen to 2 continuous acoustic sequences and to determine
whether their rhythms were the same or different (AX
discrimination). Participants were native speakers of
4 languages with different rhythmic properties (Spanish,
French, English, and German) to understand whether the
predominant rhythmic patterns of a native language affect
sensitivity, bias, and reaction time in detecting rhythmic
changes in linguistic (Experiment 2) and in nonlinguistic
(Experiments 1 and 2) acoustic sequences. We examined
sensitivity and bias measures, as well as reaction times.
We also computed Bayes factors in order to assess the
effect of native language.
Results: All listeners performed better (i.e., responded
faster and manifested higher sensitivity and accuracy)
when detecting the presence or absence of a rhythm
change when the 1st stimulus in an AX test pair exhibited
regular rhythm (i.e., a syllable-timed rhythmic pattern)
than when the 1st stimulus exhibited irregular rhythm (i.e.,
stress-timed rhythmic pattern). This result pattern was
observed both on linguistic and nonlinguistic stimuli
and was not modulated by the native language of the
participant.
Conclusion: We conclude that rhythm change detection is
a fundamental function of a processing system that relies
on general auditory mechanisms and is not modulated by
linguistic experience.The authors acknowledge support from Spanish Ministry of Economy and Competitiveness Grant PSI2017-82563-P (awarded to A. G. S.), the “Severo Ochoa” Programme for Centres/Units of Excellence in R&D Grant SEV-2015-490 (BCBL), and the Basque Foundation for Science Grant IKERBASQUE (awarded to A. G. S. and M. O.). D. M. G. was supported by Grant PIA/Basal FB0003 from the Chilean Research Council. L. P. was supported by the Spanish Ministry of Economy and Competitiveness via Juan de la Cierva fellowship
Comparing phoneme frequency, age of acquisition, and loss in aphasia:Implications for phonological universals
Phonological complexity may be central to the nature of human language. It may shape the distribution of phonemes and phoneme sequences within languages, but also determine age of acquisition and susceptibility to loss in aphasia. We evaluated this claim using frequency statistics derived from a corpus of phonologically transcribed Italian words (phonitalia, available at phonitalia,org), rankings of phoneme age of acquisition (AoA) and rate of phoneme errors in patients with apraxia of speech (AoS) as an indication of articulatory complexity. These measures were related to cross-linguistically derived markedness rankings. We found strong correspondences. AoA, however, was predicted by both apraxic errors and frequency, suggesting independent contributions of these variables. Our results support the reality of universal principles of complexity. In addition they suggest that these complexity principles have articulatory underpinnings since they modulate the production of patients with AoS, but not the production of patients with more central phonological difficulties
Anticipatory coarticulation in Hungarian VnC sequences
The duration of the vowel and the nasal was analyzed in the casual pronunciation of Hungarian words containing the sequence V
n
.C, where ‘.’ is a syllable boundary and C is a stop, affricate, fricative, or approximant. It was found that due to anticipatory coarticulation the duration of
n
is significantly shorter before fricatives and approximants than before stops and affricates.A teaching algorithm was used to distinguish between stops/affricates and fricatives/approximants in V
n
C sequences. We used an approach to the classification of C by means of the support vector machine (SVM) and the properties of Radial basis function (RBF) kernel (using MATLAB, version 7.0). The results show close to 95% correct responses for the stop/affricate vs. fricative/approximant distinction of C, as opposed to about 60% correct responses for the classification of the voicing feature of C
Zebra finches and Dutch adults exhibit the same cue weighting bias in vowel perception
Vocal tract resonances, called formants, are the most important parameters in human speech production and perception. They encode linguistic meaning and have been shown to be perceived by a wide range of species. Songbirds are also sensitive to different formant patterns in human speech. They can categorize words differing only in their vowels based on the formant patterns independent of speaker identity in a way comparable to humans. These results indicate that speech perception mechanisms are more similar between songbirds and humans than realized before. One of the major questions regarding formant perception concerns the weighting of different formants in the speech signal (“acoustic cue weighting”) and whether this process is unique to humans. Using an operant Go/NoGo design, we trained zebra finches to discriminate syllables, whose vowels differed in their first three formants. When subsequently tested with novel vowels, similar in either their first formant or their second and third formants to the familiar vowels, similarity in the higher formants was weighted much more strongly than similarity in the lower formant. Thus, zebra finches indeed exhibit a cue weighting bias. Interestingly, we also found that Dutch speakers when tested with the same paradigm exhibit the same cue weighting bias. This, together with earlier findings, supports the hypothesis that human speech evolution might have exploited general properties of the vertebrate auditory system
- …