2,921 research outputs found

    Simultaneous Coherent Structure Coloring facilitates interpretable clustering of scientific data by amplifying dissimilarity

    Get PDF
    The clustering of data into physically meaningful subsets often requires assumptions regarding the number, size, or shape of the subgroups. Here, we present a new method, simultaneous coherent structure coloring (sCSC), which accomplishes the task of unsupervised clustering without a priori guidance regarding the underlying structure of the data. sCSC performs a sequence of binary splittings on the dataset such that the most dissimilar data points are required to be in separate clusters. To achieve this, we obtain a set of orthogonal coordinates along which dissimilarity in the dataset is maximized from a generalized eigenvalue problem based on the pairwise dissimilarity between the data points to be clustered. This sequence of bifurcations produces a binary tree representation of the system, from which the number of clusters in the data and their interrelationships naturally emerge. To illustrate the effectiveness of the method in the absence of a priori assumptions, we apply it to three exemplary problems in fluid dynamics. Then, we illustrate its capacity for interpretability using a high-dimensional protein folding simulation dataset. While we restrict our examples to dynamical physical systems in this work, we anticipate straightforward translation to other fields where existing analysis tools require ad hoc assumptions on the data structure, lack the interpretability of the present method, or in which the underlying processes are less accessible, such as genomics and neuroscience

    Mapping Topographic Structure in White Matter Pathways with Level Set Trees

    Full text link
    Fiber tractography on diffusion imaging data offers rich potential for describing white matter pathways in the human brain, but characterizing the spatial organization in these large and complex data sets remains a challenge. We show that level set trees---which provide a concise representation of the hierarchical mode structure of probability density functions---offer a statistically-principled framework for visualizing and analyzing topography in fiber streamlines. Using diffusion spectrum imaging data collected on neurologically healthy controls (N=30), we mapped white matter pathways from the cortex into the striatum using a deterministic tractography algorithm that estimates fiber bundles as dimensionless streamlines. Level set trees were used for interactive exploration of patterns in the endpoint distributions of the mapped fiber tracks and an efficient segmentation of the tracks that has empirical accuracy comparable to standard nonparametric clustering methods. We show that level set trees can also be generalized to model pseudo-density functions in order to analyze a broader array of data types, including entire fiber streamlines. Finally, resampling methods show the reliability of the level set tree as a descriptive measure of topographic structure, illustrating its potential as a statistical descriptor in brain imaging analysis. These results highlight the broad applicability of level set trees for visualizing and analyzing high-dimensional data like fiber tractography output

    Integration schemes and decomposition modes in complex problem solving

    Get PDF
    Integrated product development requires that decomposition and integration schemes be congruent and in harmony with each other. This harmony marks a complex and large scale product development project with success. In domain of problem solving one needs to resort to a priori knowledge about problem structure indicating the problem’s underlying couplings/complexity. Here, we refer to the problem structure as the self map of the system. Usually for large scale problems the self needs to be decomposed for tractability purposes. The structure of the problem after decomposition is the real structure to be dealt with. In this paper we measure the complexity of systems before and after decomposition. We refer to complexity of the system/problem before decomposition as self complexity and that of system after decomposition as real complexity. It is shown that real complexity cannot be less than self complexity regardless of the decomposition type, and therefore decomposition does not reduce the overall complexity of problem. Also it is shown that by using real complexity measure of decomposed system, weak and strong emergence can be detected in the system. This would have important implications in problem classification and choosing the right design process that is in congruence with complexity of the problem

    Contributions in computational intelligence with results in functional neuroimaging

    Get PDF
    This thesis applies computational intelligence methodologies to study functional brain images. It is a state-of-the-art application relative to unsupervised learning domain to functional neuroimaging. There are also contributions related to computational intelligence on topics relative to clustering validation and spatio-temporal clustering analysis. Speci_cally, there are the presentation of a new separation measure based on fuzzy sets theory to establish the validity of the fuzzy clustering outcomes and the presentation of a framework to approach the parcellation of functional neuroimages taking in account both spatial and temporal patterns. These contributions have been applied to neuroimages obtained with functional Magnetic Resonance Imaging, using both active and passive paradigm and using both in-house data and fMRI repository. The results obtained shown, globally, an improvement on the quality of the neuroimaging analysis using the methodological contributions proposed

    A novel Big Data analytics and intelligent technique to predict driver's intent

    Get PDF
    Modern age offers a great potential for automatically predicting the driver's intent through the increasing miniaturization of computing technologies, rapid advancements in communication technologies and continuous connectivity of heterogeneous smart objects. Inside the cabin and engine of modern cars, dedicated computer systems need to possess the ability to exploit the wealth of information generated by heterogeneous data sources with different contextual and conceptual representations. Processing and utilizing this diverse and voluminous data, involves many challenges concerning the design of the computational technique used to perform this task. In this paper, we investigate the various data sources available in the car and the surrounding environment, which can be utilized as inputs in order to predict driver's intent and behavior. As part of investigating these potential data sources, we conducted experiments on e-calendars for a large number of employees, and have reviewed a number of available geo referencing systems. Through the results of a statistical analysis and by computing location recognition accuracy results, we explored in detail the potential utilization of calendar location data to detect the driver's intentions. In order to exploit the numerous diverse data inputs available in modern vehicles, we investigate the suitability of different Computational Intelligence (CI) techniques, and propose a novel fuzzy computational modelling methodology. Finally, we outline the impact of applying advanced CI and Big Data analytics techniques in modern vehicles on the driver and society in general, and discuss ethical and legal issues arising from the deployment of intelligent self-learning cars
    • …
    corecore