15 research outputs found

    Hierarchies and Ranks for Persistence Pairs

    Full text link
    We develop a novel hierarchy for zero-dimensional persistence pairs, i.e., connected components, which is capable of capturing more fine-grained spatial relations between persistence pairs. Our work is motivated by a lack of spatial relationships between features in persistence diagrams, leading to a limited expressive power. We build upon a recently-introduced hierarchy of pairs in persistence diagrams that augments the pairing stored in persistence diagrams with information about which components merge. Our proposed hierarchy captures differences in branching structure. Moreover, we show how to use our hierarchy to measure the spatial stability of a pairing and we define a rank function for persistence pairs and demonstrate different applications.Comment: Topology-based Methods in Visualization 201

    Natural Stratifications of Reeb Spaces and Higher Morse Functions

    Full text link
    Both Reeb spaces and higher Morse functions induce natural stratifications. In the former, we show that the data of the Jacobi set of a function f:XRkf:X \to \mathbb{R}^k induces stratifications on X,RkX,\mathbb{R}^k, and the associated Reeb space, and give conditions under which maps between these three spaces are stratified maps. We then extend this type of construction to the codomain of higher Morse functions, using the singular locus to induce a stratification of which sub-posets are equivalent to multi-parameter filtrations.Comment: v2: Additional examples/figures, Appendix on notions of criticality for PL ma

    Task-based Augmented Reeb Graphs with Dynamic ST-Trees

    Get PDF
    International audienceThis paper presents, to the best of our knowledge, the first parallel algorithm for the computation of the augmented Reeb graph of piecewise linear scalar data. Such augmented Reeb graphs have a wide range of applications , including contour seeding and feature based segmentation. Our approach targets shared-memory multi-core workstations. For this, it completely revisits the optimal, but sequential, Reeb graph algorithm, which is capable of handing data in arbitrary dimension and with optimal time complexity. We take advantage of Fibonacci heaps to exploit the ST-Tree data structure through independent local propagations, while maintaining the optimal, linearithmic time complexity of the sequential reference algorithm. These independent propagations can be expressed using OpenMP tasks, hence benefiting in parallel from the dynamic load balancing of the task runtime while enabling us to increase the parallelism degree thanks to a dual sweep. We present performance results on triangulated surfaces and tetrahedral meshes. We provide comparisons to related work and show that our new algorithm results in superior time performance in practice, both in sequential and in parallel. An open-source C++ implementation is provided for reproducibility

    Avoiding the Global Sort: A Faster Contour Tree Algorithm

    Get PDF
    We revisit the classical problem of computing the \emph{contour tree} of a scalar field f:MRf:\mathbb{M} \to \mathbb{R}, where M\mathbb{M} is a triangulated simplicial mesh in Rd\mathbb{R}^d. The contour tree is a fundamental topological structure that tracks the evolution of level sets of ff and has numerous applications in data analysis and visualization. All existing algorithms begin with a global sort of at least all critical values of ff, which can require (roughly) Ω(nlogn)\Omega(n\log n) time. Existing lower bounds show that there are pathological instances where this sort is required. We present the first algorithm whose time complexity depends on the contour tree structure, and avoids the global sort for non-pathological inputs. If CC denotes the set of critical points in M\mathbb{M}, the running time is roughly O(vClogv)O(\sum_{v \in C} \log \ell_v), where v\ell_v is the depth of vv in the contour tree. This matches all existing upper bounds, but is a significant improvement when the contour tree is short and fat. Specifically, our approach ensures that any comparison made is between nodes in the same descending path in the contour tree, allowing us to argue strong optimality properties of our algorithm. Our algorithm requires several novel ideas: partitioning M\mathbb{M} in well-behaved portions, a local growing procedure to iteratively build contour trees, and the use of heavy path decompositions for the time complexity analysis

    Thermodynamic Tree: The Space of Admissible Paths

    Full text link
    Is a spontaneous transition from a state x to a state y allowed by thermodynamics? Such a question arises often in chemical thermodynamics and kinetics. We ask the more formal question: is there a continuous path between these states, along which the conservation laws hold, the concentrations remain non-negative and the relevant thermodynamic potential G (Gibbs energy, for example) monotonically decreases? The obvious necessary condition, G(x)\geq G(y), is not sufficient, and we construct the necessary and sufficient conditions. For example, it is impossible to overstep the equilibrium in 1-dimensional (1D) systems (with n components and n-1 conservation laws). The system cannot come from a state x to a state y if they are on the opposite sides of the equilibrium even if G(x) > G(y). We find the general multidimensional analogue of this 1D rule and constructively solve the problem of the thermodynamically admissible transitions. We study dynamical systems, which are given in a positively invariant convex polyhedron D and have a convex Lyapunov function G. An admissible path is a continuous curve along which GG does not increase. For x,y from D, x\geq y (x precedes y) if there exists an admissible path from x to y and x \sim y if x\geq y and y\geq x. The tree of G in D is a quotient space D/~. We provide an algorithm for the construction of this tree. In this algorithm, the restriction of G onto the 1-skeleton of DD (the union of edges) is used. The problem of existence of admissible paths between states is solved constructively. The regions attainable by the admissible paths are described.Comment: Extended version, 31 page, 9 figures, 69 cited references, many minor correction

    Constructing Desirable Scalar Fields for Morse Analysis on Meshes

    Get PDF
    Morse theory is a powerful mathematical tool that uses the local differential properties of a manifold to make conclusions about global topological aspects of the manifold. Morse theory has been proven to be a very useful tool in computer graphics, geometric data processing and understanding. This work is divided into two parts. The first part is concerned with constructing geometry and symmetry aware scalar functions on a triangulated 22-manifold. To effectively apply Morse theory to discrete manifolds, one needs to design scalar functions on them with certain properties such as respecting the symmetry and the geometry of the surface and having the critical points of the scalar function coincide with feature or symmetry points on the surface. In this work, we study multiple methods that were suggested in the literature to construct such functions such as isometry invariant scalar functions, Poisson fields and discrete conformal factors. We suggest multiple refinements to each family of these functions and we propose new methods to construct geometry and symmetry-aware scalar functions using mainly the theory of the Laplace-Beltrami operator. Our proposed methods are general and can be applied in many areas such as parametrization, shape analysis, symmetry detection and segmentation. In the second part of the thesis we utilize Morse theory to give topologically consistent segmentation algorithms

    Semi-automatic transfer function generation for volumetric data visualization using contour tree analyses

    Get PDF

    Collaborative Motion Planning

    Get PDF
    Planning motion is an essential component for any autonomous robotic system. An intelligent agent must be able to efficiently plan collision-free paths in order to move through its world. Despite its importance, this problem is PSPACE-Hard which means that even planning motions for simple robots is computationally difficult. State-of-the-art approaches trade completeness (always able to provide a solution if one exists or report none exists) for probabilistic completeness (probabilistically guaranteed to find a solution if one exists but cannot report if none exists) and improved efficiency. These methods use sampling-based techniques to design a sequence of motions for the robot. However, as these methods are random in nature, the probability of their success is directly related to the expansiveness, or openness, of the underlying planning space. In other words, narrow passages, complex systems, and various constraints make planning with these methods difficult. On the other hand, humans can often determine approximate solutions for these difficult solutions quickly. In this research, we explore user-guided planning in which a human operator works together with a sampling-based motion planner. By having a human-in-the-loop, a human can steer a sampling-based planner towards a solution. This strategy can provide benefits to many applications such as computer-aided design and virtual prototyping, to name a few. We begin by classifying and creating simple models of common user-guided and heuristic-guided motion planning methods. Our models encompass three forms of user input: configuration-based, path-based, and region-based input. We compare and contrast these approaches and motivate our choice of a region-based collaborative framework. Through this analysis, we gain insight into user-guided planning and further motivate methods that harness low interface complexity and work entirely in workspace, which is most natural to a human operator. Further, we extend the theory of expansiveness to analyze the various types of user inputs. Our novel region-based collaboration framework takes advantage of human intuition by allowing a user to define regions in the workspace to bias and/or constrain the search space of a sampling-based motion planner. This approach allows a user to bias a high dimensional search with low dimensional input, supports intermittent user hints, and empowers a user to customize motion solutions. Finally, we extend region steering to both non-holonomic robotic systems and a human-inspired approach to motion planning. Our results show that this region-based framework can aid many variants of sampling-based planning, reduce computation time, support solution customization, and can be used to develop advanced heuristic methods for solving motion planning problems. We provide experiments exemplifying our approach in planning motions for complex robotic applications such as mobile manipulators, car-like, and free-flying robots

    Doctor of Philosophy

    Get PDF
    dissertationA broad range of applications capture dynamic data at an unprecedented scale. Independent of the application area, finding intuitive ways to understand the dynamic aspects of these increasingly large data sets remains an interesting and, to some extent, unsolved research problem. Generically, dynamic data sets can be described by some, often hierarchical, notion of feature of interest that exists at each moment in time, and those features evolve across time. Consequently, exploring the evolution of these features is considered to be one natural way of studying these data sets. Usually, this process entails the ability to: 1) define and extract features from each time step in the data set; 2) find their correspondences over time; and 3) analyze their evolution across time. However, due to the large data sizes, visualizing the evolution of features in a comprehensible manner and performing interactive changes are challenging. Furthermore, feature evolution details are often unmanageably large and complex, making it difficult to identify the temporal trends in the underlying data. Additionally, many existing approaches develop these components in a specialized and standalone manner, thus failing to address the general task of understanding feature evolution across time. This dissertation demonstrates that interactive exploration of feature evolution can be achieved in a non-domain-specific manner so that it can be applied across a wide variety of application domains. In particular, a novel generic visualization and analysis environment that couples a multiresolution unified spatiotemporal representation of features with progressive layout and visualization strategies for studying the feature evolution across time is introduced. This flexible framework enables on-the-fly changes to feature definitions, their correspondences, and other arbitrary attributes while providing an interactive view of the resulting feature evolution details. Furthermore, to reduce the visual complexity within the feature evolution details, several subselection-based and localized, per-feature parameter value-based strategies are also enabled. The utility and generality of this framework is demonstrated by using several large-scale dynamic data sets

    Shape analysis of the human brain.

    Get PDF
    Autism is a complex developmental disability that has dramatically increased in prevalence, having a decisive impact on the health and behavior of children. Methods used to detect and recommend therapies have been much debated in the medical community because of the subjective nature of diagnosing autism. In order to provide an alternative method for understanding autism, the current work has developed a 3-dimensional state-of-the-art shape based analysis of the human brain to aid in creating more accurate diagnostic assessments and guided risk analyses for individuals with neurological conditions, such as autism. Methods: The aim of this work was to assess whether the shape of the human brain can be used as a reliable source of information for determining whether an individual will be diagnosed with autism. The study was conducted using multi-center databases of magnetic resonance images of the human brain. The subjects in the databases were analyzed using a series of algorithms consisting of bias correction, skull stripping, multi-label brain segmentation, 3-dimensional mesh construction, spherical harmonic decomposition, registration, and classification. The software algorithms were developed as an original contribution of this dissertation in collaboration with the BioImaging Laboratory at the University of Louisville Speed School of Engineering. The classification of each subject was used to construct diagnoses and therapeutic risk assessments for each patient. Results: A reliable metric for making neurological diagnoses and constructing therapeutic risk assessment for individuals has been identified. The metric was explored in populations of individuals having autism spectrum disorders, dyslexia, Alzheimers disease, and lung cancer. Conclusion: Currently, the clinical applicability and benefits of the proposed software approach are being discussed by the broader community of doctors, therapists, and parents for use in improving current methods by which autism spectrum disorders are diagnosed and understood
    corecore