3 research outputs found
Multi-head attention-based masked sequence model for mapping functional brain networks
The investigation of functional brain networks (FBNs) using task-based functional magnetic resonance imaging (tfMRI) has gained significant attention in the field of neuroimaging. Despite the availability of several methods for constructing FBNs, including traditional methods like GLM and deep learning methods such as spatiotemporal self-attention mechanism (STAAE), these methods have design and training limitations. Specifically, they do not consider the intrinsic characteristics of fMRI data, such as the possibility that the same signal value at different time points could represent different brain states and meanings. Furthermore, they overlook prior knowledge, such as task designs, during training. This study aims to overcome these limitations and develop a more efficient model by drawing inspiration from techniques in the field of natural language processing (NLP). The proposed model, called the Multi-head Attention-based Masked Sequence Model (MAMSM), uses a multi-headed attention mechanism and mask training approach to learn different states corresponding to the same voxel values. Additionally, it combines cosine similarity and task design curves to construct a novel loss function. The MAMSM was applied to seven task state datasets from the Human Connectome Project (HCP) tfMRI dataset. Experimental results showed that the features acquired by the MAMSM model exhibit a Pearson correlation coefficient with the task design curves above 0.95 on average. Moreover, the model can extract more meaningful networks beyond the known task-related brain networks. The experimental results demonstrated that MAMSM has great potential in advancing the understanding of functional brain networks
Summary of ChatGPT/GPT-4 Research and Perspective Towards the Future of Large Language Models
This paper presents a comprehensive survey of ChatGPT and GPT-4,
state-of-the-art large language models (LLM) from the GPT series, and their
prospective applications across diverse domains. Indeed, key innovations such
as large-scale pre-training that captures knowledge across the entire world
wide web, instruction fine-tuning and Reinforcement Learning from Human
Feedback (RLHF) have played significant roles in enhancing LLMs' adaptability
and performance. We performed an in-depth analysis of 194 relevant papers on
arXiv, encompassing trend analysis, word cloud representation, and distribution
analysis across various application domains. The findings reveal a significant
and increasing interest in ChatGPT/GPT-4 research, predominantly centered on
direct natural language processing applications, while also demonstrating
considerable potential in areas ranging from education and history to
mathematics, medicine, and physics. This study endeavors to furnish insights
into ChatGPT's capabilities, potential implications, ethical concerns, and
offer direction for future advancements in this field.Comment: 35 pages, 3 figure
Discovering Dynamic Functional Brain Networks via Spatial and Channel-wise Attention
Using deep learning models to recognize functional brain networks (FBNs) in
functional magnetic resonance imaging (fMRI) has been attracting increasing
interest recently. However, most existing work focuses on detecting static FBNs
from entire fMRI signals, such as correlation-based functional connectivity.
Sliding-window is a widely used strategy to capture the dynamics of FBNs, but
it is still limited in representing intrinsic functional interactive dynamics
at each time step. And the number of FBNs usually need to be set manually. More
over, due to the complexity of dynamic interactions in brain, traditional
linear and shallow models are insufficient in identifying complex and spatially
overlapped FBNs across each time step. In this paper, we propose a novel
Spatial and Channel-wise Attention Autoencoder (SCAAE) for discovering FBNs
dynamically. The core idea of SCAAE is to apply attention mechanism to FBNs
construction. Specifically, we designed two attention modules: 1) spatial-wise
attention (SA) module to discover FBNs in the spatial domain and 2) a
channel-wise attention (CA) module to weigh the channels for selecting the FBNs
automatically. We evaluated our approach on ADHD200 dataset and our results
indicate that the proposed SCAAE method can effectively recover the dynamic
changes of the FBNs at each fMRI time step, without using sliding windows. More
importantly, our proposed hybrid attention modules (SA and CA) do not enforce
assumptions of linearity and independence as previous methods, and thus provide
a novel approach to better understanding dynamic functional brain networks.Comment: 12 pages,6 figures, submitted to 36th Conference on Neural
Information Processing Systems (NeurIPS 2022