93 research outputs found
An Empirical Method for Comparing Pitch Patterns in Spoken and Musical Melodies: A Comment on J.G.S. Pearl's "Eavesdropping with a Master: Leos JanĆ”Äek and the Music of Speech."
Music and speech both feature structured melodic patterns, yet these patterns are rarely compared using empirical methods. One reason for this has been a lack of tools which allow quantitative comparisons of spoken and musical pitch sequences. Recently, a new model of speech intonation perception has been proposed based on principles of pitch perception in speech. The āprosogramā model converts a sentence's fundamental frequency contour into a sequence of discrete tones and glides. This sequence is meant to represent a listener's perception of pitch in connected speech. This article briefly describes the prosogram and suggests a few ways in which it can be used to compare the structure of spoken and musical melodies
Why would Musical Training Benefit the Neural Encoding of Speech? The OPERA Hypothesis
Mounting evidence suggests that musical training benefits the neural encoding of speech. This paper offers a hypothesis specifying why such benefits occur. The āOPERAā hypothesis proposes that such benefits are driven by adaptive plasticity in speech-processing networks, and that this plasticity occurs when five conditions are met. These are: (1) Overlap: there is anatomical overlap in the brain networks that process an acoustic feature used in both music and speech (e.g., waveform periodicity, amplitude envelope), (2) Precision: music places higher demands on these shared networks than does speech, in terms of the precision of processing, (3) Emotion: the musical activities that engage this network elicit strong positive emotion, (4) Repetition: the musical activities that engage this network are frequently repeated, and (5) Attention: the musical activities that engage this network are associated with focused attention. According to the OPERA hypothesis, when these conditions are met neural plasticity drives the networks in question to function with higher precision than needed for ordinary speech communication. Yet since speech shares these networks with music, speech processing benefits. The OPERA hypothesis is used to account for the observed superior subcortical encoding of speech in musically trained individuals, and to suggest mechanisms by which musical training might improve linguistic reading abilities
Temporal Generalization of Synchronized Saccades Beyond the Trained Range in Monkeys
Synchronized movements with external periodic rhythms, such as dancing to a beat, are commonly observed in daily life. Although it has been well established that some vocal learning species (including parrots and humans) spontaneously develop this ability, it has only recently been shown that monkeys are also capable of predictive and tempo-flexible synchronization to periodic stimuli. In our previous study, monkeys were trained to make predictive saccades for alternately presented visual stimuli at fixed stimulus onset asynchronies (SOAs) to obtain a liquid reward. The monkeys generalized predictive synchronization to novel SOAs in the middle of trained range, suggesting a capacity for tempo-flexible synchronization. However, it is possible that when encountering a novel tempo, the monkeys might sample learned saccade sequences from those for the short and long SOAs so that the mean saccade interval matched the untrained SOA. To eliminate this possibility, in the current study we tested monkeys on novel SOAs outside the trained range. Animals were trained to generate synchronized eye movements for 600 and 900-ms SOAs for a few weeks, and then were tested for longer SOAs. The accuracy and precision of predictive saccades for one untrained SOA (1200 ms) were comparable to those for the trained conditions. On the other hand, the variance of predictive saccade latency and the proportion of reactive saccades increased significantly in the longer SOA conditions (1800 and 2400 ms), indicating that temporal prediction of periodic stimuli was difficult in this range, similar to previous results on synchronized tapping in humans. Our results suggest that monkeys might share similar synchronization mechanisms with humans, which can be subject to physiological examination in future studies
Testing beat perception without sensory cues to the beat: the Beat-Drop Alignment Test (BDAT)
Beat perception can serve as a window into internal time-keeping mechanisms, auditoryāmotor interactions, and aspects of cognition. One aspect of beat perception is the covert continuation of an internal pulse. Of the several popular tests of beat perception, none provide a satisfying test of this faculty of covert continuation. The current study proposes a new beat-perception test focused on covert pulse continuation: The Beat-Drop Alignment Test (BDAT). In this test, participants must identify the beat in musical excerpts and then judge whether a single probe falls on or off the beat. The probe occurs during a short break in the rhythmic components of the music when no rhythmic events are present, forcing participants to judge beat alignment relative to an internal pulse maintained in the absence of local acoustic timing cues. Here, we present two large (N > 100) tests of the BDAT. In the first, we explore the effect of test item parameters (e.g., probe displacement) on performance. In the second, we correlate scores on an adaptive version of the BDAT with the computerized adaptive Beat Alignment Test (CA-BAT) scores and indices of musical experience. Musical experience indices outperform CA-BAT score as a predictor of BDAT score, suggesting that the BDAT measures a distinct aspect of beat perception that is more experience-dependent and may draw on cognitive resources such as working memory and musical imagery differently than the BAT. The BDAT may prove useful in future behavioral and neural research on beat perception, and all stimuli and code are freely available for download
Recommended from our members
Weighting of cues to categorization of song versus speech in tone-language and non-tone-language speakers
One of the most important auditory categorization tasks a listener faces is determining a sound's domain, a process which is a prerequisite for successful within-domain categorization tasks such as recognizing different speech sounds or musical tones. Speech and song are universal in human cultures: how do listeners categorize a sequence of words as belonging to one or the other of these domains? There is growing interest in the acoustic cues that distinguish speech and song, but it remains unclear whether there are cross-cultural differences in the evidence upon which listeners rely when making this fundamental perceptual categorization. Here we use the speech-to-song illusion, in which some spoken phrases perceptually transform into song when repeated, to investigate cues to this domain-level categorization in native speakers of tone languages (Mandarin and Cantonese speakers residing in the United Kingdom and China) and in native speakers of a non-tone language (English). We find that native tone-language and non-tone-language listeners largely agree on which spoken phrases sound like song after repetition, and we also find that the strength of this transformation is not significantly different across language backgrounds or countries of residence. Furthermore, we find a striking similarity in the cues upon which listeners rely when perceiving word sequences as singing versus speech, including small pitch intervals, flat within-syllable pitch contours, and steady beats. These findings support the view that there are certain widespread cross-cultural similarities in the mechanisms by which listeners judge if a word sequence is spoken or sung
Mapping Specific Mental Content during Musical Imagery
Humans can mentally represent auditory information without an external stimulus, but the specificity of these internal representations remains unclear. Here, we asked how similar the temporally unfolding neural representations of imagined music are compared to those during the original perceived experience. We also tested whether rhythmic motion can influence the neural representation of music during imagery as during perception. Participants first memorized six 1-min-long instrumental musical pieces with high accuracy. Functional MRI data were collected during: 1) silent imagery of melodies to the beat of a visual metronome; 2) same but while tapping to the beat; and 3) passive listening. During imagery, inter-subject correlation analysis showed that melody-specific temporal response patterns were reinstated in right associative auditory cortices. When tapping accompanied imagery, the melody-specific neural patterns were reinstated in more extensive temporal-lobe regions bilaterally. These results indicate that the specific contents of conscious experience are encoded similarly during imagery and perception in the dynamic activity of auditory cortices. Furthermore, rhythmic motion can enhance the reinstatement of neural patterns associated with the experience of complex sounds, in keeping with models of motor to sensory influences in auditory processing
The impact of basal ganglia lesions on sensorimotor synchronization, spontaneous motor tempo, and the detection of tempo changes
a b s t r a c t The basal ganglia (BG) are part of extensive subcortico-cortical circuits that are involved in a variety of motor and non-motor cognitive functions. Accumulating evidence suggests that one specific function that engages the BG and associated cortico-striato-thalamo-cortical circuitry is temporal processing, i.e., the mechanisms that underlie the encoding, decoding and evaluation of temporal relations or temporal structure. In the current study we investigated the interplay of two processes that require precise representations of temporal structure, namely the perception of an auditory pacing signal and manual motor production by means of finger tapping in a sensorimotor synchronization task. Patients with focal lesions of the BG and healthy control participants were asked to align finger taps to tone sequences that either did or did not contain a tempo acceleration or tempo deceleration at a predefined position, and to continue tapping at the final tempo after the pacing sequence had ceased. Performance in this adaptive synchronization-continuation paradigm differed between the two groups. Selective damage to the BG affected the abilities to detect tempo changes and to perform attention-dependent error correction, particularly in response to tempo decelerations. An additional assessment of preferred spontaneous, i.e., unpaced but regular, production rates yielded more heterogeneous results in the patient group. Together these findings provide evidence for less efficient processing in the perception and the production of temporal structure in patients with focal BG lesions. The results also support the functional role of the BG system in attention-dependent temporal processing
Intonation processing in congenital amusia: discrimination, identification and imitation
This study investigated whether congenital amusia, a neuro-developmental disorder of musical perception, also has implications for speech intonation processing. In total, 16 British amusics and 16 matched controls completed five intonation perception tasks and two pitch threshold tasks. Compared with controls, amusics showed impaired performance on discrimination, identification and imitation of statements and questions that were characterized primarily by pitch direction differences in the final word. This intonation-processing deficit in amusia was largely associated with a psychophysical pitch direction discrimination deficit. These findings suggest that amusia impacts upon oneās language abilities in subtle ways, and support previous evidence that pitch processing in language and music involves shared mechanisms
- ā¦