550 research outputs found

    Meta-analyses support a taxonomic model for representations of different categories of audio-visual interaction events in the human brain

    Get PDF
    Our ability to perceive meaningful action events involving objects, people and other animate agents is characterized in part by an interplay of visual and auditory sensory processing and their cross-modal interactions. However, this multisensory ability can be altered or dysfunctional in some hearing and sighted individuals, and in some clinical populations. The present meta-analysis sought to test current hypotheses regarding neurobiological architectures that may mediate audio-visual multisensory processing. Reported coordinates from 82 neuroimaging studies (137 experiments) that revealed some form of audio-visual interaction in discrete brain regions were compiled, converted to a common coordinate space, and then organized along specific categorical dimensions to generate activation likelihood estimate (ALE) brain maps and various contrasts of those derived maps. The results revealed brain regions (cortical “hubs”) preferentially involved in multisensory processing along different stimulus category dimensions, including (1) living versus non-living audio-visual events, (2) audio-visual events involving vocalizations versus actions by living sources, (3) emotionally valent events, and (4) dynamic-visual versus static-visual audio-visual stimuli. These meta-analysis results are discussed in the context of neurocomputational theories of semantic knowledge representations and perception, and the brain volumes of interest are available for download to facilitate data interpretation for future neuroimaging studies

    Decoding the consumer’s brain: Neural representations of consumer experience

    Get PDF
    Understanding consumer experience – what consumers think about brands, how they feel about services, whether they like certain products – is crucial to marketing practitioners. ‘Neuromarketing’, as the application of neuroscience in marketing research is called, has generated excitement with the promise of understanding consumers’ minds by probing their brains directly. Recent advances in neuroimaging analysis leverage machine learning and pattern classification techniques to uncover patterns from neuroimaging data that can be associated with thoughts and feelings. In this dissertation, I measure brain responses of consumers by functional magnetic resonance imaging (fMRI) in order to ‘decode’ their mind. In three different studies, I have demonstrated how different aspects of consumer experience can be studied with fMRI recordings. First, I study how consumers think about brand image by comparing their brain responses during passive viewing of visual templates (photos depicting various social scenarios) to those during active visualizing of a brand’s image. Second, I use brain responses during viewing of affective pictures to decode emotional responses during watching of movie-trailers. Lastly, I examine whether marketing videos that evoke s

    Neocortical substrates of feelings evoked with music in the ACC, insula, and somatosensory cortex

    Get PDF
    Neurobiological models of emotion focus traditionally on limbic/paralimbic regions as neural substrates of emotion generation, and insular cortex (in conjunction with isocortical anterior cingulate cortex, ACC) as the neural substrate of feelings. An emerging view, however, highlights the importance of isocortical regions beyond insula and ACC for the subjective feeling of emotions. We used music to evoke feelings of joy and fear, and multivariate pattern analysis (MVPA) to decode representations of feeling states in functional magnetic resonance (fMRI) data of n = 24 participants. Most of the brain regions providing information about feeling representations were neocortical regions. These included, in addition to granular insula and cingulate cortex, primary and secondary somatosensory cortex, premotor cortex, frontal operculum, and auditory cortex. The multivoxel activity patterns corresponding to feeling representations emerged within a few seconds, gained in strength with increasing stimulus duration, and replicated results of a hypothesis-generating decoding analysis from an independent experiment. Our results indicate that several neocortical regions (including insula, cingulate, somatosensory and premotor cortices) are important for the generation and modulation of feeling states. We propose that secondary somatosensory cortex, which covers the parietal operculum and encroaches on the posterior insula, is of particular importance for the encoding of emotion percepts, i.e., preverbal representations of subjective feeling.publishedVersio

    Effect of Olfaction on the Perception of Movie Clips

    Get PDF
    Odours are important to many species but their effect on human perception in the context of concurrent auditory and visual stimulation has received little investigation. Here we examined how the experience of viewing audio-visual movie clips changes when accompanied by congruent or incongruent odours. Using an olfactometer to control odourant delivery, thirty-five undergraduate students from Western University were randomly presented 36 different odour-video pairs twice. Following each presentation, participants completed three Likert scales to assess multisensory interaction in terms of engagement, pleasantness, and emotional arousal. Comparison of congruent and incongruent odours to the no odour control condition revealed that incongruent odours had a greater effect than congruent odours on participant ratings, and that this effect acted to negatively influence experience, reducing engagement, pleasantness, and emotional arousal. There was little difference between congruent odours and no odour on ratings of engagement and emotional arousal; however, even congruent odours reduced pleasantness ratings, suggesting all odours used were, to an extent, unpleasant. An interaction suggested that certain movies were more strongly modulated by odour than others. We interpret our results as evidence of crossmodal competition, in which the presence of an odour leads to suppression of the auditory and visual modalities. This was confirmed using functional magnetic resonance imaging in a single participant. Future research should continue to investigate the surprising role odour plays in multisensory interaction
    corecore