29,211 research outputs found
Brain explorer for connectomic analysis
Visualization plays a vital role in the analysis of multimodal neuroimaging data. A major challenge in neuroimaging visualization is how to integrate structural, functional, and connectivity data to form a comprehensive visual context for data exploration, quality control, and hypothesis discovery. We develop a new integrated visualization solution for brain imaging data by combining scientific and information visualization techniques within the context of the same anatomical structure. In this paper, new surface texture techniques are developed to map non-spatial attributes onto both 3D brain surfaces and a planar volume map which is generated by the proposed volume rendering technique, spherical volume rendering. Two types of non-spatial information are represented: (1) time series data from resting-state functional MRI measuring brain activation; (2) network properties derived from structural connectivity data for different groups of subjects, which may help guide the detection of differentiation features. Through visual exploration, this integrated solution can help identify brain regions with highly correlated functional activations as well as their activation patterns. Visual detection of differentiation features can also potentially discover image-based phenotypic biomarkers for brain diseases
Rendering techniques for multimodal data
Many different direct volume rendering methods have been developed to visualize 3D scalar fields on uniform rectilinear grids. However, little work has been done on rendering simultaneously various properties of the same 3D region measured with different registration devices or at different instants of time. The demand for this type of visualization is rapidly increasing in scientific applications such as medicine in which the visual integration of multiple modalities allows a better comprehension of the anatomy and a perception of its relationships with activity. This paper presents different strategies of Direct Multimodal Volume Rendering (DMVR). It is restricted to voxel models with a known 3D rigid alignment transformation. The paper evaluates at which steps of the render-ing pipeline must the data fusion be realized in order to accomplish the desired visual integration and to provide fast re-renders when some fusion parameters are modified. In addition, it analyzes how existing monomodal visualization al-gorithms can be extended to multiple datasets and it compares their efficiency and their computational cost.Postprint (published version
Integration of multimodal data based on surface registration
The paper proposes and evaluates a strategy for the alignment of
anatomical and functional data of the brain. The method takes as an
input two different sets of images of a same patient: MR data and
SPECT. It proceeds in four steps: first, it constructs two voxel
models from the two image sets; next, it extracts from the two voxel
models the surfaces of regions of interest; in the third step, the
surfaces are interactively aligned by corresponding pairs; finally a
unique volume model is constructed by selectively applying the
geometrical transformations associated to the regions and weighting
their contributions. The main advantages of this strategy are (i) that
it can be applied retrospectively, (ii) that it is tri-dimensional,
and (iii) that it is local. Its main disadvantage with regard to
previously published methods it that it requires the extraction of
surfaces. However, this step is often required for other stages of the
multimodal analysis such as the visualization and therefore its cost
can be accounted in the global cost of the process.Postprint (published version
Embodiment, sound and visualization : a multimodal perspective in music education
Recently, many studies have emphasized the role of body movements in
processing, sharing and giving meaning to music. At the same time, neuroscience
studies, suggest that different parts of the brain are integrated and activated by the same
stimuli: sounds, for example, can be perceived by touch and can evoke imagery, energy,
fluency and periodicity. This interaction of auditory, visual and motor senses can be
found in the verbal descriptions of music and among children during their spontaneous
games. The question to be asked is, if a more multisensory and embodied approach
could redefine some of our assumptions regarding musical education. Recent research
on embodiment and multimodal perception in instrumental teaching could suggest new
directions in musical education. Can we consider the integration between the activities
of body movement, listening, metaphor visualization, and singing, as more effective
than a disembodied and fragmented approach for the process of musical understanding
Design of a multimodal rendering system
This paper addresses the rendering of aligned regular multimodal
datasets. It presents a general framework of multimodal data fusion
that includes several data merging methods. We also analyze the
requirements of a rendering system able to provide these different
fusion methods. On the basis of these requirements, we propose a novel
design for a multimodal rendering system. The design has been
implemented and proved showing to be efficient and flexible.Postprint (published version
A Fast hierarchical traversal strategy for multimodal visualization
In the last years there is a growing demand of multimodal medical rendering systems able to visualize simultaneously data coming from different sources. This paper addresses the Direct Volume Rendering (DVR) of aligned multimodal data in medical applications. Specifically, it proposes a hierarchical representation of the multimodal data set based on the construction of a Fusion Decision Tree (FDT) that, together with a run-length encoding of the non-empty data, provides means of efficiently accessing to the data. Three different implementations of these structures are proposed. The simulations results show that the traversal of the data is fast and that the method is suitable when interactive modifications of the fusion parameters are required.Postprint (published version
Neuroimaging of structural pathology and connectomics in traumatic brain injury: Toward personalized outcome prediction.
Recent contributions to the body of knowledge on traumatic brain injury (TBI) favor the view that multimodal neuroimaging using structural and functional magnetic resonance imaging (MRI and fMRI, respectively) as well as diffusion tensor imaging (DTI) has excellent potential to identify novel biomarkers and predictors of TBI outcome. This is particularly the case when such methods are appropriately combined with volumetric/morphometric analysis of brain structures and with the exploration of TBI-related changes in brain network properties at the level of the connectome. In this context, our present review summarizes recent developments on the roles of these two techniques in the search for novel structural neuroimaging biomarkers that have TBI outcome prognostication value. The themes being explored cover notable trends in this area of research, including (1) the role of advanced MRI processing methods in the analysis of structural pathology, (2) the use of brain connectomics and network analysis to identify outcome biomarkers, and (3) the application of multivariate statistics to predict outcome using neuroimaging metrics. The goal of the review is to draw the community's attention to these recent advances on TBI outcome prediction methods and to encourage the development of new methodologies whereby structural neuroimaging can be used to identify biomarkers of TBI outcome
Multimodal Grounding for Language Processing
This survey discusses how recent developments in multimodal processing
facilitate conceptual grounding of language. We categorize the information flow
in multimodal processing with respect to cognitive models of human information
processing and analyze different methods for combining multimodal
representations. Based on this methodological inventory, we discuss the benefit
of multimodal grounding for a variety of language processing tasks and the
challenges that arise. We particularly focus on multimodal grounding of verbs
which play a crucial role for the compositional power of language.Comment: The paper has been published in the Proceedings of the 27 Conference
of Computational Linguistics. Please refer to this version for citations:
https://www.aclweb.org/anthology/papers/C/C18/C18-1197
- …