1,624 research outputs found

    Parallel Auditory Filtering By Sustained and Transient Channels Separates Coarticulated Vowels and Consonants

    Full text link
    A neural model of peripheral auditory processing is described and used to separate features of coarticulated vowels and consonants. After preprocessing of speech via a filterbank, the model splits into two parallel channels, a sustained channel and a transient channel. The sustained channel is sensitive to relatively stable parts of the speech waveform, notably synchronous properties of the vocalic portion of the stimulus it extends the dynamic range of eighth nerve filters using coincidence deteectors that combine operations of raising to a power, rectification, delay, multiplication, time averaging, and preemphasis. The transient channel is sensitive to critical features at the onsets and offsets of speech segments. It is built up from fast excitatory neurons that are modulated by slow inhibitory interneurons. These units are combined over high frequency and low frequency ranges using operations of rectification, normalization, multiplicative gating, and opponent processing. Detectors sensitive to frication and to onset or offset of stop consonants and vowels are described. Model properties are characterized by mathematical analysis and computer simulations. Neural analogs of model cells in the cochlear nucleus and inferior colliculus are noted, as are psychophysical data about perception of CV syllables that may be explained by the sustained transient channel hypothesis. The proposed sustained and transient processing seems to be an auditory analog of the sustained and transient processing that is known to occur in vision.Air Force Office of Scientific Research (F49620-92-J-0225); Advanced Research Projects Agency (AFOSR 90-0083, ONR N00014-92-J-4015); Office of Naval Research (N00014-95-I-0409

    Variable Rate Working Memories for Phonetic Categorization and Invariant Speech Perception

    Full text link
    Speech can be understood at widely varying production rates. A working memory is described for short-term storage of temporal lists of input items. The working memory is a cooperative-competitive neural network that automatically adjusts its integration rate, or gain, to generate a short-term memory code for a list that is independent of item presentation rate. Such an invariant working memory model is used to simulate data of Repp (1980) concerning the changes of phonetic category boundaries as a function of their presentation rate. Thus the variability of categorical boundaries can be traced to the temporal in variance of the working memory code.Air Force Office of Scientific Research (F49620-92-J-0225, 90-0128); Defense Advanced Research Projects Agency (ONR N00014-92-J-4015); Office of Naval Research (N00014-91-J-4100

    Neural Dynamics of Phonetic Trading Relations for Variable-Rate CV Syllables

    Full text link
    The perception of CV syllables exhibits a trading relationship between voice onset time (VOT) of a consonant and duration of a vowel. Percepts of [ba] and [wa] can, for example, depend on the durations of the consonant and vowel segments, with an increase in the duration of the subsequent vowel switching the percept of the preceding consonant from [w] to [b]. A neural model, called PHONET, is proposed to account for these findings. In the model, C and V inputs are filtered by parallel auditory streams that respond preferentially to transient and sustained properties of the acoustic signal, as in vision. These streams are represented by working memories that adjust their processing rates to cope with variable acoustic input rates. More rapid transient inputs can cause greater activation of the transient stream which, in turn, can automatically gain control the processing rate in the sustained stream. An invariant percept obtains when the relative activations of C and V representations in the two streams remain uncha.nged. The trading relation may be simulated as a result of how different experimental manipulations affect this ratio. It is suggested that the brain can use duration of a subsequent vowel to make the [b]/[w] distinction because the speech code is a resonant event that emerges between working mernory activation patterns and the nodes that categorize them.Advanced Research Projects Agency (90-0083); Air Force Office of Scientific Reseearch (F19620-92-J-0225); Pacific Sierra Research Corporation (91-6075-2

    A Neural Network for Synthesizing the Pitch of an Acoustic Source

    Full text link
    This article describes a neural network model capable of generating a spatial representation of the pitch of an acoustic source. Pitch is one of several auditory percepts used by humans to separate multiple sound sources in the environment from each other. The model provides a neural instantiation of a type of "harmonic sieve". It is capable of quantitatively simulating a large body of psychoacoustical data, including new data on octave shift perception.Air Force Office of Scientific Research (90-0128, 90-0175); Defense Advanced Research Projects Agency (90-0083); National Science Foundation (IRI 90-24877); American Society for Engineering Educatio

    Neural Control of Interlimb Oscillations I: Human Bimanual Coordination

    Full text link
    How do humans and other animals accomplish coordinated movements? How are novel combinations of limb joints rapidly assembled into new behavioral units that rnove together in in-phase or anti-phase movement patterns during complex movement tasks? A neural central pattern generator (CPG) model simulates data from human bimanual coordination tasks. As in the data, anti-phase oscillations at low frequencies switch to in-phase oscillations at high frequencies, in-phase oscillation occur both at low and high frequencies, phase fluctuations occur at the anti-phase in-phase transition, a "seagull effect" of larger errors occurs at intermediate phases, and oscillations slip toward in-phase and anti-phase when driven at intermediate phases. These oscillations and bifurcations are emergent properties of the CPG model in response to volitional inputs. The CPC model is a version of the Ellias-Grossberg oscillator. Its neurons obey Hodgkin-Huxley type equations whose excitatory signals operate on a faster time scale than their inhibitory signals in a recurrent on-center off-surround anatomy. When an equal cornmand or GO signal activates both model channels the model CPC: can generate both in-phase and anti-phase oscillations at different GO amplitudes. Phase transitions frorn either in-phase to anti-phase oscillations, or from anti-phase to in- phase oscillations, can occur in different pararncter ranges, as the GO signal increases.Air Force Office of Scientific Research (F49620-92-J-0499, 90-0083, F49620-92-J-0225, 90-0128); Office of Naval Research (N00014-92-J-1309); Army Research Office (DAAL03-0088); National Science Foundation (IRI-90-24877

    Neural Control of Rhythmic Coordinated Movements

    Full text link
    How do humans and other animals accomplish coordinated movements? How are novel combinations of limb joints rapidly assembled into new behavioral units that move together in in-phase or anti-phase movement patterns during complex movement tasks? A neural model simulates data from human bimanual coordination tasks. As in the data, anti-phase oscillations at low frequencies switch to in-phase oscillations at high frequencies, in-phase oscillations occur both at low and high frequencies, phase fluctuations occur at the anti-phase in-phase transition, a "seagull effect" of larger errors occurs at intermediate phases, and oscillations slip toward in-phase and anti-phase when driven at intermediate phases.Air Force Office of Scientific Research (90-0128, F49620-92-J-0225, F49620-92-J-0499, 90-0083); Office of Naval Research (N00014-92-J-1309, N00014-92-J-1309); National Science Foundation (IIU-90-24877); Army Research Office (DAAL03-88-K-0088

    A Neural Pattern Generator that Exhibits Frequency-Dependent In-Phase and Anti-Phase Oscillations

    Full text link
    This article describes a. neural pattern generator based on a cooperative-competitive feedback neural network. The two-channel version of the generator supports both in-phase and anti-phase oscillations. A scalar arousal level controls both the oscillation phase and frequency. As arousal increases, oscillation frequency increases and bifurcations from in-phase to anti-phase, or anti-phase to in-phase oscillations can occur. Coupled versions of the model exhibit oscillatory patterns which correspond to the gaits used in locomotion and other oscillatory movements by various animals.Air Force Office of Scientific Research (90-0128, 90-0175); National Science Foundation (IRI-90-24877); Army Research Office (DAAL03-88-k-0088

    Neural Control of Interlimb Oscillations I: Human Bimanual Coordination

    Full text link
    How do humans and other animals accomplish coordinated movements? How are novel combinations of limb joints rapidly assembled into new behavioral units that rnove together in in-phase or anti-phase movement patterns during complex movement tasks? A neural central pattern generator (CPG) model simulates data from human bimanual coordination tasks. As in the data, anti-phase oscillations at low frequencies switch to in-phase oscillations at high frequencies, in-phase oscillation occur both at low and high frequencies, phase fluctuations occur at the anti-phase in-phase transition, a "seagull effect" of larger errors occurs at intermediate phases, and oscillations slip toward in-phase and anti-phase when driven at intermediate phases. These oscillations and bifurcations are emergent properties of the CPG model in response to volitional inputs. The CPC model is a version of the Ellias-Grossberg oscillator. Its neurons obey Hodgkin-Huxley type equations whose excitatory signals operate on a faster time scale than their inhibitory signals in a recurrent on-center off-surround anatomy. When an equal cornmand or GO signal activates both model channels the model CPC: can generate both in-phase and anti-phase oscillations at different GO amplitudes. Phase transitions frorn either in-phase to anti-phase oscillations, or from anti-phase to in- phase oscillations, can occur in different pararncter ranges, as the GO signal increases.Air Force Office of Scientific Research (F49620-92-J-0499, 90-0083, F49620-92-J-0225, 90-0128); Office of Naval Research (N00014-92-J-1309); Army Research Office (DAAL03-0088); National Science Foundation (IRI-90-24877

    Neural Control of Interlimb Oscillations II. Biped and Quadruped Gaits and Bifurications

    Full text link
    Behavioral data concerning animal and human gaits and gait transitions are simulated as emergent properties of a central pattern generator (CPG) model. The CPG model is a version of the Ellias-Grossberg oscillator. Its neurons obey Hodgkin-Huxley type equations whose excitatory signals operate on a faster time scale than their inhibitory signals in a recurrent on-center off-surround anatomy. A descending command or GO signal activates the gaits and triggers gait transitions as its amplitude increases. A single model CPG can generate both in-phase and anti-phase oscillations at different GO amplitudes. Phase transition from either in-phase to anti-phase oscillations, or from anti-phase to in-phase oscillations, can occur in different parameter ranges, as the GO signal increases. Quadruped vertebrate gaits, including the amble, the walk, all three pairwise gaits (trot, pace, and gallop), and the pronk are simulated using this property. Rapid gait transitions are simulated in the order walk, trot, pace, and gallop that occurs in the cat, along with the observed increase in oscillation frequency. Precise control of quadruped gait switching uses GO-dependent. modulation of inhibitory interactions, which generates a different functional anatomy at different arousal levels. The primary human gaits (the walk and the run) and elephant gaits (the amble and the walk) are simulated, without modulation, by oscillations with the same phase relationships but different waveform shapes at different GO signal levels, much as the duty cycles of the feet are longer in the walk than in the run. Relevant neural data from spinal cord, globus palliclus, and motor cortex, among other structures, are discussedArmy Research Office (DAAL03-88-K-0088); Advanced Research Projects Agency (90-0083); National Science Foundation (IRI-90-24877); Office of Naval Research (N00014-92-J-1309); Air Force Office of Scientific Research (F49620-92-J-0499, F49620-92-J-0225, 90-0128

    ARSTREAM: A Neural Network Model of Auditory Scene Analysis and Source Segregation

    Full text link
    Multiple sound sources often contain harmonics that overlap and may be degraded by environmental noise. The auditory system is capable of teasing apart these sources into distinct mental objects, or streams. Such an "auditory scene analysis" enables the brain to solve the cocktail party problem. A neural network model of auditory scene analysis, called the AIRSTREAM model, is presented to propose how the brain accomplishes this feat. The model clarifies how the frequency components that correspond to a give acoustic source may be coherently grouped together into distinct streams based on pitch and spatial cues. The model also clarifies how multiple streams may be distinguishes and seperated by the brain. Streams are formed as spectral-pitch resonances that emerge through feedback interactions between frequency-specific spectral representaion of a sound source and its pitch. First, the model transforms a sound into a spatial pattern of frequency-specific activation across a spectral stream layer. The sound has multiple parallel representations at this layer. A sound's spectral representation activates a bottom-up filter that is sensitive to harmonics of the sound's pitch. The filter activates a pitch category which, in turn, activate a top-down expectation that allows one voice or instrument to be tracked through a noisy multiple source environment. Spectral components are suppressed if they do not match harmonics of the top-down expectation that is read-out by the selected pitch, thereby allowing another stream to capture these components, as in the "old-plus-new-heuristic" of Bregman. Multiple simultaneously occuring spectral-pitch resonances can hereby emerge. These resonance and matching mechanisms are specialized versions of Adaptive Resonance Theory, or ART, which clarifies how pitch representations can self-organize durin learning of harmonic bottom-up filters and top-down expectations. The model also clarifies how spatial location cues can help to disambiguate two sources with similar spectral cures. Data are simulated from psychophysical grouping experiments, such as how a tone sweeping upwards in frequency creates a bounce percept by grouping with a downward sweeping tone due to proximity in frequency, even if noise replaces the tones at their interection point. Illusory auditory percepts are also simulated, such as the auditory continuity illusion of a tone continuing through a noise burst even if the tone is not present during the noise, and the scale illusion of Deutsch whereby downward and upward scales presented alternately to the two ears are regrouped based on frequency proximity, leading to a bounce percept. Since related sorts of resonances have been used to quantitatively simulate psychophysical data about speech perception, the model strengthens the hypothesis the ART-like mechanisms are used at multiple levels of the auditory system. Proposals for developing the model to explain more complex streaming data are also provided.Air Force Office of Scientific Research (F49620-01-1-0397, F49620-92-J-0225); Office of Naval Research (N00014-01-1-0624); Advanced Research Projects Agency (N00014-92-J-4015); British Petroleum (89A-1204); National Science Foundation (IRI-90-00530); American Society of Engineering Educatio
    • …
    corecore