9,136 research outputs found
Automatic Estimation of Intelligibility Measure for Consonants in Speech
In this article, we provide a model to estimate a real-valued measure of the
intelligibility of individual speech segments. We trained regression models
based on Convolutional Neural Networks (CNN) for stop consonants
\textipa{/p,t,k,b,d,g/} associated with vowel \textipa{/A/}, to estimate the
corresponding Signal to Noise Ratio (SNR) at which the Consonant-Vowel (CV)
sound becomes intelligible for Normal Hearing (NH) ears. The intelligibility
measure for each sound is called SNR, and is defined to be the SNR level
at which human participants are able to recognize the consonant at least 90\%
correctly, on average, as determined in prior experiments with NH subjects.
Performance of the CNN is compared to a baseline prediction based on automatic
speech recognition (ASR), specifically, a constant offset subtracted from the
SNR at which the ASR becomes capable of correctly labeling the consonant.
Compared to baseline, our models were able to accurately estimate the
SNR~intelligibility measure with less than 2 [dB] Mean Squared Error
(MSE) on average, while the baseline ASR-defined measure computes
SNR~with a variance of 5.2 to 26.6 [dB], depending on the consonant.Comment: 5 pages, 1 figure, 7 tables, submitted to Inter Speech 2020
Conferenc
Verification of feature regions for stops and fricatives in natural speech
The presence of acoustic cues and their importance in speech perception have
long remained debatable topics. In spite of several studies that exist in this
eld, very little is known about what exactly humans perceive in speech. This
research takes a novel approach towards understanding speech perception. A
new method, named three-dimensional deep search (3DDS), was developed
to explore the perceptual cues of 16 consonant-vowel (CV) syllables, namely
/pa/, /ta/, /ka/, /ba/, /da/, /ga/, /fa/, /Ta/, /sa/, /Sa/, /va/, /Da/, /za/,
/Za/, from naturally produced speech. A veri cation experiment was then
conducted to further verify the ndings of the 3DDS method. For this pur-
pose, the time-frequency coordinate that de nes each CV was ltered out
using the short-time Fourier transform (STFT), and perceptual tests were
then conducted. A comparison between unmodi ed speech sounds and those
without the acoustic cues was made. In most of the cases, the scores dropped
from 100% to chance levels even at 12 dB SNR. This clearly emphasizes the
importance of features in identifying each CV. The results con rm earlier
ndings that stops are characterized by a short-duration burst preceding the
vowel by 10 cs in the unvoiced case, and appearing almost coincident
with the vowel in the voiced case. As has been previously hypothesized,
we con rmed that the F2 transition plays no signi cant role in consonant
identi cation. 3DDS analysis labels the /sa/ and /za/ perceptual features
as an intense frication noise around 4 kHz, preceding the vowel by 15{20
cs, with the /za/ feature being around 5 cs shorter in duration than that
of /sa/; the /Sa/ and /Za/ events are found to be frication energy near 2
kHz, preceding the vowel by 17{20 cs. /fa/ has a relatively weak burst and
frication energy over a wide-band including 2{6 kHz, while /va/ has a cue
in the 1.5 kHz mid-frequency region preceding the vowel by 7{10 cs. New
information is established regarding /Da/ and /Ta/, especially with regards
to the nature of their signi cant confusions
Variable Rate Working Memories for Phonetic Categorization and Invariant Speech Perception
Speech can be understood at widely varying production rates. A working memory is described for short-term storage of temporal lists of input items. The working memory is a cooperative-competitive neural network that automatically adjusts its integration rate, or gain, to generate a short-term memory code for a list that is independent of item presentation rate. Such an invariant working memory model is used to simulate data of Repp (1980) concerning the changes of phonetic category boundaries as a function of their presentation rate. Thus the variability of categorical boundaries can be traced to the temporal in variance of the working memory code.Air Force Office of Scientific Research (F49620-92-J-0225, 90-0128); Defense Advanced Research Projects Agency (ONR N00014-92-J-4015); Office of Naval Research (N00014-91-J-4100
A silent speech system based on permanent magnet articulography and direct synthesis
In this paper we present a silent speech interface (SSI) system aimed at restoring speech communication for individuals who have lost their voice due to laryngectomy or diseases affecting the vocal folds. In the proposed system, articulatory data captured from the lips and tongue using permanent magnet articulography (PMA) are converted into audible speech using a speaker-dependent transformation learned from simultaneous recordings of PMA and audio signals acquired before laryngectomy. The transformation is represented using a mixture of factor analysers, which is a generative model that allows us to efficiently model non-linear behaviour and perform dimensionality reduction at the same time. The learned transformation is then deployed during normal usage of the SSI to restore the acoustic speech signal associated with the captured PMA data. The proposed system is evaluated using objective quality measures and listening tests on two databases containing PMA and audio recordings for normal speakers. Results show that it is possible to reconstruct speech from articulator movements captured by an unobtrusive technique without an intermediate recognition step. The SSI is capable of producing speech of sufficient intelligibility and naturalness that the speaker is clearly identifiable, but problems remain in scaling up the process to function consistently for phonetically rich vocabularies
Changes in the McGurk Effect Across Phonetic Contexts
To investigate the process underlying audiovisual speech perception, the McGurk illusion was examined across a range of phonetic contexts. Two major changes were found. First, the frequency of illusory /g/ fusion percepts increased relative to the frequency of illusory /d/ fusion percepts as vowel context was shifted from /i/ to /a/ to /u/. This trend could not be explained by biases present in perception of the unimodal visual stimuli. However, the change found in the McGurk fusion effect across vowel environments did correspond systematically with changes in second format frequency patterns across contexts. Second, the order of consonants in illusory combination percepts was found to depend on syllable type. This may be due to differences occuring across syllable contexts in the timecourses of inputs from the two modalities as delaying the auditory track of a vowel-consonant stimulus resulted in a change in the order of consonants perceived. Taken together, these results suggest that the speech perception system either fuses audiovisual inputs into a visually compatible percept with a similar second formant pattern to that of the acoustic stimulus or interleaves the information from different modalities, at a phonemic or subphonemic level, based on their relative arrival times.National Institutes of Health (R01 DC02852
Turkish /h/ deletion : evidence for the interplay of speech perception and phonology
It has been hypothesized that sounds which are less perceptible are more likely to be altered than more salient sounds, the rationale being that the loss of information resulting from a change in a sound which is difficult to perceive is not as great as the loss resulting from a change in a more salient sound. Kohler (1990) suggested that the tendency to reduce articulatory movements is countered by perceptual and social constraints, finding that fricatives are relatively resistant to reduction in colloquial German. Kohler hypothesized that this is due to the perceptual salience of fricatives, a hypothesis which was supported by the results of a perception experiment by Hura, Lindblom, and Diehl (1992). These studies showed that the relative salience of speech sounds is relevant to explaining phonological behavior. An additional factor is the impact of different acoustic environments on the perceptibility of speech sounds. Steriade (1997) found that voicing contrasts are more common in positions where more cues to voicing are available. The P-map, proposed by Steriade (2001a, b), allows the representation of varying salience of segments in different contexts. Many researchers have posited a relationship between speech perception and phonology. The purpose of this paper is to provide experimental evidence for this relationship, drawing on the case of Turkish /h/ deletion
- …