21 research outputs found
Cross-Modal Prediction in Speech Perception
Speech perception often benefits from vision of the speaker's lip movements when they are available. One potential mechanism underlying this reported gain in perception arising from audio-visual integration is on-line prediction. In this study we address whether the preceding speech context in a single modality can improve audiovisual processing and whether this improvement is based on on-line information-transfer across sensory modalities. In the experiments presented here, during each trial, a speech fragment (context) presented in a single sensory modality (voice or lips) was immediately continued by an audiovisual target fragment. Participants made speeded judgments about whether voice and lips were in agreement in the target fragment. The leading single sensory context and the subsequent audiovisual target fragment could be continuous in either one modality only, both (context in one modality continues into both modalities in the target fragment) or neither modalities (i.e., discontinuous). The results showed quicker audiovisual matching responses when context was continuous with the target within either the visual or auditory channel (Experiment 1). Critically, prior visual context also provided an advantage when it was cross-modally continuous (with the auditory channel in the target), but auditory to visual cross-modal continuity resulted in no advantage (Experiment 2). This suggests that visual speech information can provide an on-line benefit for processing the upcoming auditory input through the use of predictive mechanisms. We hypothesize that this benefit is expressed at an early level of speech analysis
World Health Organization/International Society for Biomedical Research on Alcoholism study on state and trait markers of alcohol use and dependence: Back to the future
This article summarizes content proceedings of a symposium held at the 2004 International Society for Biomedical Research on Alcoholism Congress in Mannheim, Germany. The chairs were Boris Tabakoff and Friedrich M. Wurst. The presentations were (1) Genetic associations with alcoholism and affective disorders, by Paula Hoffman; (2) Proteomic analysis of blood constituents in alcoholism, by Boris Tabakoff; (3) Contrasts between the responses of GGT and CDT to high alcohol intake, and a test of their combined use, by John Whitfield; (4) Direct ethanol metabolites such as ethyl glucuronide, fatty acid ethyl esters, phosphatidylethanol and ethyl sulfate: a new line of sensitive and specific biomarkers, by Friedrich Martin Wurst; and (5) Genetic studies of alcoholism subtypes in a Han Taiwanese population, by Ru-Band Lu
Cross-modal prediction in speech depends on prior linguistic experience
The sight of a speaker’s facial movements during the perception of a spoken message can benefit speech processing through online predictive mechanisms. Recent evidence suggests that these predictive mechanisms can operate across sensory modalities, that is, vision and audition. However, to date, behavioral and electrophysiological demonstrations of cross-modal prediction in speech have considered only the speaker’s native language. Here, we address a question of current debate, namely whether the level of representation involved in cross-modal prediction is phonological or pre-phonological. We do this by testing participants in an unfamiliar language. If cross-modal prediction is predominantly based on phonological representations tuned to the phonemic categories of the native language of the listener, then it should be more effective in the listener’s native language than in an unfamiliar one. We tested Spanish and English native speakers in an audiovisual matching paradigm that allowed us to evaluate visual-to-auditory prediction, using sentences in the participant’s native language and in an unfamiliar language. The benefits of cross-modal prediction were only seen in the native language, regardless of the particular language or participant’s linguistic background. This pattern of results implies that cross-modal visual-to-auditory prediction during speech processing makes strong use of phonological representations, rather than low-level spatiotemporal correlations across facial movements and sounds