3,359 research outputs found
PRIMING OF OBJECT CATEGORIZATION WITHIN AND ACROSS LEVELS OF SPECIFICITY
Identification of objects can occur at different levels of specificity. Dependingon task and context, an object can be classified at the superordinate level (as ananimal), at the basic level (a bird) or at the subordinate level (a sparrow). Whatare the interactions between these representational levels and do they rely onthe same sequential processes that lead to successful object identification? Inthis electroencephalogram study, a task-switching paradigm (covert naming orliving/non-living judgment) was used. Images of objects were repeated eitherwithin the same task, or with a switch from a covert naming task to a livingor non-living judgment and vice versa. While covert naming accesses entrylevel(basic or subordinate), living/non-living judgments rely on superordinateclassification. Our beha-vioural results demonstrated clear priming effectswithin both tasks. However, asymmetries were found when task-switching hadoccurred, with facilitation for covert naming but not for categorization. Wealso found lower accuracy and early-starting and persistent enhancements ofevent-related potentials (ERPs) for covert naming, indicating that this task wasmore difficult and involved more intense perceptual and semantic processing.Perceptual priming was marked by consistent reductions of the ERP componentL1 for repeated presentations, both with and without task switching. Additionalrepetition effects were found in early event-related activity between 150-190 ms(N1) when a repeated image had been named at initial presentation. We conclude that differences in N1 indicate task-related changes in the identification processitself. Such enhancements for covert naming again emerge in a later timewindow associated with depth of semantic processing. Meanwhile, L1 reflectsmodulations due to implicit memory of objects. In conclusion, evidence wasfound for representational overlap; changes in ERP markers started early andrevealed cross-task priming at the level of object structure analysis and moreintense perceptual and semantic processing for covert naming
Recommended from our members
Extracting Spatiotemporal Word and Semantic Representations from Multiscale Neurophysiological Recordings in Humans
With the recent advent of neuroimaging techniques, the majority of the research studying the neural basis of language processing has focused on the localization of various lexical and semantic functions. Unfortunately, the limited time resolution of functional neuroimaging prevents a detailed analysis of the dynamics involved in word recognition, and the hemodynamic basis of these techniques prevents the study of the underlying neurophysiology. Compounding this problem, current techniques for the analysis of high-dimensional neural data are mainly sensitive to large effects in a small area, preventing a thorough study of the distributed processing involved for representing semantic knowledge. This thesis demonstrates the use of multivariate machine-learning techniques for the study of the neural representation of semantic and speech information in electro/magneto-physiological recordings with high temporal resolution. Support vector machines (SVMs) allow for the decoding of semantic category and word-specific information from non-invasive electroencephalography (EEG) and magnetoenecephalography (MEG) and demonstrate the consistent, but spatially and temporally distributed nature of such information. Moreover, the anteroventral temporal lobe (avTL) may be important for coordinating these distributed representations, as supported by the presence of supramodal category-specific information in intracranial recordings from the avTL as early as 150ms after auditory or visual word presentation. Finally, to study the inputs to this lexico-semantic system, recordings from a high density microelectrode array in anterior superior temporal gyrus (aSTG) are obtained, and the recorded spiking activity demonstrates the presence of single neurons that respond specifically to speech sounds. The successful decoding of word identity from this firing rate information suggests that the aSTG may be involved in the population coding of acousto-phonetic speech information that is likely on the pathway for mapping speech-sounds to meaning in the avTL. The feasibility of extracting semantic and phonological information from multichannel neural recordings using machine learning techniques provides a powerful method for studying language using large datasets and has potential implications for the development of fast and intuitive communication prostheses.Engineering and Applied Science
Understanding everyday experiences of reminiscence for people living with blindness: Practices, tensions and probing new design possibilities
There is growing attention in the HCI community on how technology could be designed to support experiences of reminiscence on past life experiences. Yet, this research has largely overlooked people living with blindness. I present a study that aims to understand everyday experiences of reminiscence for people living with blindness. I conducted a qualitative study with 9 participants living with blindness to understand their personal routines, wishes and desires, and challenges and tensions regarding the experience of reminiscence. Findings are interpreted to discuss new possibilities that offer starting points for future design initiatives and openings for collaboration aimed at creating technology to better support the practices of capturing, sharing, and reflecting on significant memories of the past
Recommended from our members
Dynamic activity patterns in the anterior temporal lobe represents object semantics.
The anterior temporal lobe (ATL) is considered a crucial area for the representation of transmodal concepts. Recent evidence suggests that specific regions within the ATL support the representation of individual object concepts, as shown by studies combining multivariate analysis methods and explicit measures of semantic knowledge. This research looks to further our understanding by probing conceptual representations at a spatially and temporally resolved neural scale. Representational similarity analysis was applied to human intracranial recordings from anatomically defined lateral to medial ATL sub-regions. Neural similarity patterns were tested against semantic similarity measures, where semantic similarity was defined by a hybrid corpus-based and feature-based approach. Analyses show that the perirhinal cortex, in the medial ATL, significantly related to semantic effects around 200 to 400 ms, and were greater than more lateral ATL regions. Further, semantic effects were present in low frequency (theta and alpha) oscillatory phase signals. These results provide converging support that more medial regions of the ATL support the representation of basic-level visual object concepts within the first 400 ms, and provide a bridge between prior fMRI and MEG work by offering detailed evidence for the presence of conceptual representations within the ATL
A humanness dimension to visual object coding in the brain
Neuroimaging studies investigating human object recognition have primarily focused on a relatively small number of object categories, in particular, faces, bodies, scenes, and vehicles. More recent studies have taken a broader focus, investigating hypothesized dichotomies, for example, animate versus inanimate, and continuous feature dimensions, such as biologically similarity. These studies typically have used stimuli that are identified as animate or inanimate, neglecting objects that may not fit into this dichotomy. We generated a novel stimulus set including standard objects and objects that blur the animate-inanimate dichotomy, for example, robots and toy animals. We used MEG time-series decoding to study the brain's emerging representation of these objects. Our analysis examined contemporary models of object coding such as dichotomous animacy, as well as several new higher order models that take into account an object's capacity for agency (i.e. its ability to move voluntarily) and capacity to experience the world. We show that early (0–200 ms) responses are predicted by the stimulus shape, assessed using a retinotopic model and shape similarity computed from human judgments. Thereafter, higher order models of agency/experience provided a better explanation of the brain's representation of the stimuli. Strikingly, a model of human similarity provided the best account for the brain's representation after an initial perceptual processing phase. Our findings provide evidence for a new dimension of object coding in the human brain – one that has a “human-centric” focus
Combining brain-computer interfaces and assistive technologies: state-of-the-art and challenges
In recent years, new research has brought the field of EEG-based Brain-Computer Interfacing (BCI) out of its infancy and into a phase of relative maturity through many demonstrated prototypes such as brain-controlled wheelchairs, keyboards, and computer games. With this proof-of-concept phase in the past, the time is now ripe to focus on the development of practical BCI technologies that can be brought out of the lab and into real-world applications. In particular, we focus on the prospect of improving the lives of countless disabled individuals through a combination of BCI technology with existing assistive technologies (AT). In pursuit of more practical BCIs for use outside of the lab, in this paper, we identify four application areas where disabled individuals could greatly benefit from advancements in BCI technology, namely,“Communication and Control”, “Motor Substitution”, “Entertainment”, and “Motor Recovery”. We review the current state of the art and possible future developments, while discussing the main research issues in these four areas. In particular, we expect the most progress in the development of technologies such as hybrid BCI architectures, user-machine adaptation algorithms, the exploitation of users’ mental states for BCI reliability and confidence measures, the incorporation of principles in human-computer interaction (HCI) to improve BCI usability, and the development of novel BCI technology including better EEG devices
Issues and recommendations from the OHBM COBIDAS MEEG committee for reproducible EEG and MEG research
The Organization for Human Brain Mapping (OHBM) has been active in advocating for the instantiation of best practices in neuroimaging data acquisition, analysis, reporting and sharing of both data and analysis code to deal with issues in science related to reproducibility and replicability. Here we summarize recommendations for such practices in magnetoencephalographic (MEG) and electroencephalographic (EEG) research, recently developed by the OHBM neuroimaging community known by the abbreviated name of COBIDAS MEEG. We discuss the rationale for the guidelines and their general content, which encompass many topics under active discussion in the field. We highlight future opportunities and challenges to maximizing the sharing and exploitation of MEG and EEG data, and we also discuss how this ‘living’ set of guidelines will evolve to continually address new developments in neurophysiological assessment methods and multimodal integration of neurophysiological data with other data types.Peer reviewe
- …