201 research outputs found

    A 3d geoscience information system framework

    Get PDF
    Two-dimensional geographical information systems are extensively used in the geosciences to create and analyse maps. However, these systems are unable to represent the Earth's subsurface in three spatial dimensions. The objective of this thesis is to overcome this deficiency, to provide a general framework for a 3d geoscience information system (GIS), and to contribute to the public discussion about the development of an infrastructure for geological observation data, geomodels, and geoservices. Following the objective, the requirements for a 3d GIS are analysed. According to the requirements, new geologically sensible query functionality for geometrical, topological and geological properties has been developed and the integration of 3d geological modeling and data management system components in a generic framework has been accomplished. The 3d geoscience information system framework presented here is characterized by the following features: - Storage of geological observation data and geomodels in a XML-database server. According to a new data model, geological observation data can be referenced by a set of geomodels. - Functionality for querying observation data and 3d geomodels based on their 3d geometrical, topological, material, and geological properties were developed and implemented as plug-in for a 3d geomodeling user application. - For database queries, the standard XML query language has been extended with 3d spatial operators. The spatial database query operations are computed using a XML application server which has been developed for this specific purpose. This technology allows sophisticated 3d spatial and geological database queries. Using the developed methods, queries can be answered like: "Select all sandstone horizons which are intersected by the set of faults F". This request contains a topological and a geological material parameter. The combination of queries with other GIS methods, like visual and statistical analysis, allows geoscience investigations in a novel 3d GIS environment. More generally, a 3d GIS enables geologists to read and understand a 3d digital geomodel analogously as they read a conventional 2d geological map

    Point set signature and algorithm of classifications on its basis

    Get PDF
    На Π΄Π°Π½Π½Ρ‹ΠΉ ΠΌΠΎΠΌΠ΅Π½Ρ‚ сущСствуСт большоС количСство Π·Π°Π΄Π°Ρ‡ ΠΏΠΎ Π°Π²Ρ‚ΠΎΠΌΠ°Ρ‚ΠΈΠ·ΠΈΡ€ΠΎΠ²Π°Π½Π½ΠΎΠΉ ΠΎΠ±Ρ€Π°Π±ΠΎΡ‚ΠΊΠ΅ ΠΌΠ½ΠΎΠ³ΠΎΠΌΠ΅Ρ€Π½Ρ‹Ρ… Π΄Π°Π½Π½Ρ‹Ρ…, Π½Π°ΠΏΡ€ΠΈΠΌΠ΅Ρ€, классификация, кластСризация, ΠΏΡ€ΠΎΠ³Π½ΠΎΠ·ΠΈΡ€ΠΎΠ²Π°Π½ΠΈΠ΅, Π·Π°Π΄Π°Ρ‡ΠΈ управлСния слоТными ΠΎΠ±ΡŠΠ΅ΠΊΡ‚Π°ΠΌΠΈ. БоотвСтствСнно, Π²ΠΎΠ·Π½ΠΈΠΊΠ°Π΅Ρ‚ Π½Π΅ΠΎΠ±Ρ…ΠΎΠ΄ΠΈΠΌΠΎΡΡ‚ΡŒ Π² Ρ€Π°Π·Π²ΠΈΡ‚ΠΈΠΈ матСматичСского ΠΈ алгоритмичСского обСспСчСния для Ρ€Π΅ΡˆΠ΅Π½ΠΈΡ Π²ΠΎΠ·Π½ΠΈΠΊΠ°ΡŽΡ‰ΠΈΡ… Π·Π°Π΄Π°Ρ‡. ЦСлью исслСдования являСтся Ρ€Π°Π·Π²ΠΈΡ‚ΠΈΠ΅ Π°Π»Π³ΠΎΡ€ΠΈΡ‚ΠΌΠΎΠ² классификации Ρ‚ΠΎΡ‡Π΅Ρ‡Π½Ρ‹Ρ… мноТСств Π½Π° основС ΠΈΡ… пространствСнного распрСдСлСния. Π’ Ρ€Π°Π±ΠΎΡ‚Π΅ прСдлагаСтся Ρ€Π°ΡΡΠΌΠ°Ρ‚Ρ€ΠΈΠ²Π°Ρ‚ΡŒ Π΄Π°Π½Π½Ρ‹Π΅ ΠΊΠ°ΠΊ Ρ‚ΠΎΡ‡ΠΊΠΈ Π² ΠΌΠ½ΠΎΠ³ΠΎΠΌΠ΅Ρ€Π½ΠΎΠΌ мСтричСском пространствС. Π’ Ρ€Π°Π±ΠΎΡ‚Π΅ рассмотрСны ΠΏΠΎΠ΄Ρ…ΠΎΠ΄Ρ‹ ΠΊ описанию характСристик Ρ‚ΠΎΡ‡Π΅Ρ‡Π½Ρ‹Ρ… мноТСств Π² пространствах высокой размСрности ΠΈ прСдлагаСтся ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ ΠΊ описанию Ρ‚ΠΎΡ‡Π΅Ρ‡Π½ΠΎΠ³ΠΎ мноТСства Π½Π° основС сигнатур, ΠΊΠΎΡ‚ΠΎΡ€Ρ‹Π΅ ΠΏΡ€Π΅Π΄ΡΡ‚Π°Π²Π»ΡΡŽΡ‚ собой характСристику заполнСнности Ρ‚ΠΎΡ‡Π΅Ρ‡Π½ΠΎΠ³ΠΎ мноТСства Π½Π° основС Ρ€Π°ΡΡˆΠΈΡ€Π΅Π½ΠΈΡ понятия пространствСнного Ρ…Π΅ΡˆΠΈΡ€ΠΎΠ²Π°Π½ΠΈΡ. ΠžΠ±ΠΎΠ±Ρ‰Π΅Π½Π½Ρ‹ΠΉ ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ ΠΊ Π²Ρ‹Ρ‡ΠΈΡΠ»Π΅Π½ΠΈΡŽ сигнатур Ρ‚ΠΎΡ‡Π΅Ρ‡Π½Ρ‹Ρ… мноТСств Π·Π°ΠΊΠ»ΡŽΡ‡Π°Π΅Ρ‚ΡΡ Π² Ρ€Π°Π·Π±ΠΈΠ΅Π½ΠΈΠΈ пространства, Π·Π°Π½ΠΈΠΌΠ°Π΅ΠΌΠΎΠ³ΠΎ мноТСством Π½Π° Ρ€Π΅Π³ΡƒΠ»ΡΡ€Π½ΡƒΡŽ сСтку с ΠΏΠΎΠΌΠΎΡ‰ΡŒΡŽ ΠΌΠ΅Ρ‚ΠΎΠ΄Π° пространствСнного Ρ…Π΅ΡˆΠΈΡ€ΠΎΠ²Π°Π½ΠΈΡ, вычислСния гСомСтричСских характСристик мноТСства Π² ΠΏΠΎΠ»ΡƒΡ‡Π΅Π½Π½Ρ‹Ρ… ячСйках ΠΈ опрСдСлСния Π½Π°ΠΈΠ±ΠΎΠ»Π΅Π΅ Π·Π°ΠΏΠΎΠ»Π½Π΅Π½Π½Ρ‹Ρ… ячССк ΠΏΠΎ ΠΊΠ°ΠΆΠ΄ΠΎΠΌΡƒ ΠΈΠ· пространствСнных ΠΈΠ·ΠΌΠ΅Ρ€Π΅Π½ΠΈΠΉ. ΠŸΡ€Π΅Π΄Π»Π°Π³Π°Π΅Ρ‚ΡΡ Π½ΠΎΠ²Ρ‹ΠΉ ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ ΠΊ классификации Π½Π° основС сигнатур мноТСства, ΠΊΠΎΡ‚ΠΎΡ€Ρ‹ΠΉ Π·Π°ΠΊΠ»ΡŽΡ‡Π°Π΅Ρ‚ΡΡ Π² Π½Π°Ρ…ΠΎΠΆΠ΄Π΅Π½ΠΈΠΈ сигнатур для Ρ‚ΠΎΡ‡Π΅ΠΊ с извСстным Π·Π½Π°Ρ‡Π΅Π½ΠΈΠ΅ΠΌ принадлСТности ΠΊ Π½Π΅ΠΊΠΎΡ‚ΠΎΡ€Ρ‹ΠΌ классам, Π° для Π½ΠΎΠ²Ρ‹Ρ… Ρ‚ΠΎΡ‡Π΅ΠΊ вычисляСтся расстояниС ΠΎΡ‚ Ρ…Π΅ΡˆΠ° Ρ‚ΠΎΡ‡ΠΊΠΈ Π΄ΠΎ сигнатуры ΠΊΠ°ΠΆΠ΄ΠΎΠ³ΠΎ ΠΈΠ· извСстных мноТСств, Π½Π° основС Ρ‡Π΅Π³ΠΎ опрСдСляСтся Π½Π°ΠΈΠ±ΠΎΠ»Π΅Π΅ вСроятный класс Ρ‚ΠΎΡ‡ΠΊΠΈ. Π’ качСствС ΠΈΡΠΏΠΎΠ»ΡŒΠ·ΡƒΠ΅ΠΌΡ‹Ρ… ΠΌΠ΅Ρ‚Ρ€ΠΈΠΊ ΠΏΡ€Π΅Π΄Π»Π°Π³Π°ΡŽΡ‚ΡΡ Π•Π²ΠΊΠ»ΠΈΠ΄ΠΎΠ²ΠΎ расстояниС ΠΈ ΠΌΠ΅Ρ‚Ρ€ΠΈΠΊΠ° городских ΠΊΠ²Π°Ρ€Ρ‚Π°Π»ΠΎΠ². Π’ Ρ€Π°Π±ΠΎΡ‚Π΅ ΠΏΡ€ΠΎΠ²Π΅Π΄Ρ‘Π½ ΡΡ€Π°Π²Π½ΠΈΡ‚Π΅Π»ΡŒΠ½Ρ‹ΠΉ Π°Π½Π°Π»ΠΈΠ· ΠΈΡΠΏΠΎΠ»ΡŒΠ·ΡƒΠ΅ΠΌΡ‹Ρ… ΠΌΠ΅Ρ‚Ρ€ΠΈΠΊ с Ρ‚ΠΎΡ‡ΠΊΠΈ зрСния точности классификации. ΠŸΡ€Π΅ΠΈΠΌΡƒΡ‰Π΅ΡΡ‚Π²Π°ΠΌΠΈ ΠΏΡ€Π΅Π΄Π»ΠΎΠΆΠ΅Π½Π½ΠΎΠ³ΠΎ ΠΏΠΎΠ΄Ρ…ΠΎΠ΄Π° ΡΠ²Π»ΡΡŽΡ‚ΡΡ простота вычислСний ΠΈ высокая ΡΡ‚Π΅ΠΏΠ΅Π½ΡŒ точности классификации для Ρ€Π°Π²Π½ΠΎΠΌΠ΅Ρ€Π½ΠΎ распрСдСлСнных Ρ‚ΠΎΡ‡Π΅ΠΊ. ΠŸΡ€Π΅Π΄ΡΡ‚Π°Π²Π»Π΅Π½Π½Ρ‹ΠΉ Π°Π»Π³ΠΎΡ€ΠΈΡ‚ΠΌ Ρ€Π΅Π°Π»ΠΈΠ·ΠΎΠ²Π°Π½ Π² Π²ΠΈΠ΄Π΅ ΠΏΡ€ΠΎΠ³Ρ€Π°ΠΌΠΌΠ½ΠΎΠ³ΠΎ прилоТСния Π½Π° языкС Python с использованиСм Π±ΠΈΠ±Π»ΠΈΠΎΡ‚Π΅ΠΊΠΈ NumPy. Π’Π°ΠΊΠΆΠ΅ рассмотрСны Π²Π°Ρ€ΠΈΠ°Π½Ρ‚Ρ‹ использования ΠΏΡ€Π΅Π΄Π»ΠΎΠΆΠ΅Π½Π½ΠΎΠ³ΠΎ ΠΏΠΎΠ΄Ρ…ΠΎΠ΄Π° для Π·Π°Π΄Π°Ρ‡ с Π½Π΅ числовыми Π΄Π°Π½Π½Ρ‹ΠΌΠΈ, Ρ‚Π°ΠΊΠΈΠΌΠΈ ΠΊΠ°ΠΊ строковыС ΠΈ Π±ΡƒΠ»Π΅Π²Ρ‹ значСния. Для Ρ‚Π°ΠΊΠΈΡ… Π΄Π°Π½Π½Ρ‹Ρ… ΠΏΡ€Π΅Π΄Π»ΠΎΠΆΠ΅Π½ΠΎ ΠΈΡΠΏΠΎΠ»ΡŒΠ·ΠΎΠ²Π°Ρ‚ΡŒ ΠΌΠ΅Ρ‚Ρ€ΠΈΠΊΡƒ Π₯эмминга, ΠΏΡ€ΠΎΠ²Π΅Π΄Ρ‘Π½Π½Ρ‹Π΅ экспСримСнты ΠΏΠΎΠΊΠ°Π·Π°Π»ΠΈ Ρ€Π°Π±ΠΎΡ‚ΠΎΡΠΏΠΎΡΠΎΠ±Π½ΠΎΡΡ‚ΡŒ Π°Π»Π³ΠΎΡ€ΠΈΡ‚ΠΌΠ° для Ρ‚Π°ΠΊΠΈΡ… Ρ‚ΠΈΠΏΠΎΠ² Π΄Π°Π½Π½Ρ‹Ρ….There are many unsolved problems in the field of automatic multi-dimensional data processing, for example, classification, clustering, regression, and control of complex objects. This leads to the need of development of mathematical and algorithmical background for such problems. In our research we aim to development of classification algorithms of point sets based on their spatial distribution. We propose to consider data as points in multi-dimensional metric space. The approaches to describe point set features in high dimensional spaces are viewed. The algorithm of describing of point set based on their signatures, that are spatial distribution of point set is considered. In our approach we extend spatial hashing technique. The generalized method of computation of point set signatures is to split space, occupied by point set into regular grid by the spatial hashing algorithm, then we evaluate geometrical characteristics of the set in cells of the grid and define cells, that contain most of the points for the all of coordinate axis. The new approach to classification by means of point set signatures is developed that is to find signatures of known points with the classes defined and then we compute spatial hashes for unknown points and their distance to the signatures of classes. The probable class of the tested point is defined by the minimal distance among all distances to each signature. To define distance in our approach we use Manhattan and Euclidean metric. The comparative study of impact of metrics used to the classification error is provided. The main advantage of our method is computation simplicity and low classification error for evenly distributed points. Prototype implementation of our algorithm was written in order to test this algorithm for practical classification applications. The implementation was coded in Python with use NumPy library. The use of our algorithm to the classification of non-numerical data such as texts and booleans is viewed. For such data types we propose use of Hamming distance and experiments done show practical viability for such data types

    Representing Edge Flows on Graphs via Sparse Cell Complexes

    Full text link
    Obtaining sparse, interpretable representations of observable data is crucial in many machine learning and signal processing tasks. For data representing flows along the edges of a graph, an intuitively interpretable way to obtain such representations is to lift the graph structure to a simplicial complex: The eigenvectors of the associated Hodge-Laplacian, respectively the incidence matrices of the corresponding simplicial complex then induce a Hodge decomposition, which can be used to represent the observed data in terms of gradient, curl, and harmonic flows. In this paper, we generalize this approach to cellular complexes and introduce the cell inference optimization problem, i.e., the problem of augmenting the observed graph by a set of cells, such that the eigenvectors of the associated Hodge Laplacian provide a sparse, interpretable representation of the observed edge flows on the graph. We show that this problem is NP-hard and introduce an efficient approximation algorithm for its solution. Experiments on real-world and synthetic data demonstrate that our algorithm outperforms current state-of-the-art methods while being computationally efficient.Comment: 9 pages, 6 figures (plus appendix). For evaluation code, see https://anonymous.4open.science/r/edge-flow-repr-cell-complexes-11C

    A New Oscillating-Error Technique for Classifiers

    Full text link
    This paper describes a new method for reducing the error in a classifier. It uses an error correction update that includes the very simple rule of either adding or subtracting the error adjustment, based on whether the variable value is currently larger or smaller than the desired value. While a traditional neuron would sum the inputs together and then apply a function to the total, this new method can change the function decision for each input value. This gives added flexibility to the convergence procedure, where through a series of transpositions, variables that are far away can continue towards the desired value, whereas variables that are originally much closer can oscillate from one side to the other. Tests show that the method can successfully classify some benchmark datasets. It can also work in a batch mode, with reduced training times and can be used as part of a neural network architecture. Some comparisons with an earlier wave shape paper are also made

    Dist2Cycle: A Simplicial Neural Network for Homology Localization

    Get PDF
    Simplicial complexes can be viewed as high dimensional generalizations of graphs that explicitly encode multi-way ordered relations between vertices at different resolutions, all at once. This concept is central towards detection of higher dimensional topological features of data, features to which graphs, encoding only pairwise relationships, remain oblivious. While attempts have been made to extend Graph Neural Networks (GNNs) to a simplicial complex setting, the methods do not inherently exploit, or reason about, the underlying topological structure of the network. We propose a graph convolutional model for learning functions parametrized by the kk-homological features of simplicial complexes. By spectrally manipulating their combinatorial kk-dimensional Hodge Laplacians, the proposed model enables learning topological features of the underlying simplicial complexes, specifically, the distance of each kk-simplex from the nearest "optimal" kk-th homology generator, effectively providing an alternative to homology localization.Comment: 9 pages, 5 figure

    Graph-based Semi-Supervised & Active Learning for Edge Flows

    Full text link
    We present a graph-based semi-supervised learning (SSL) method for learning edge flows defined on a graph. Specifically, given flow measurements on a subset of edges, we want to predict the flows on the remaining edges. To this end, we develop a computational framework that imposes certain constraints on the overall flows, such as (approximate) flow conservation. These constraints render our approach different from classical graph-based SSL for vertex labels, which posits that tightly connected nodes share similar labels and leverages the graph structure accordingly to extrapolate from a few vertex labels to the unlabeled vertices. We derive bounds for our method's reconstruction error and demonstrate its strong performance on synthetic and real-world flow networks from transportation, physical infrastructure, and the Web. Furthermore, we provide two active learning algorithms for selecting informative edges on which to measure flow, which has applications for optimal sensor deployment. The first strategy selects edges to minimize the reconstruction error bound and works well on flows that are approximately divergence-free. The second approach clusters the graph and selects bottleneck edges that cross cluster-boundaries, which works well on flows with global trends

    Local Dirac Synchronization on Networks

    Get PDF
    We propose Local Dirac Synchronization which uses the Dirac operator to capture the dynamics of coupled nodes and link signals on an arbitrary network. In Local Dirac Synchronization, the harmonic modes of the dynamics oscillate freely while the other modes are interacting non-linearly, leading to a collectively synchronized state when the coupling constant of the model is increased. Local Dirac Synchronization is characterized by discontinuous transitions and the emergence of a rhythmic coherent phase. In this rhythmic phase, one of the two complex order parameters oscillates in the complex plane at a slow frequency (called emergent frequency) in the frame in which the intrinsic frequencies have zero average. Our theoretical results obtained within the annealed approximation are validated by extensive numerical results on fully connected networks and sparse Poisson and scale-free networks. Local Dirac Synchronization on both random and real networks, such as the connectome of Caenorhabditis Elegans, reveals the interplay between topology (Betti numbers and harmonic modes) and non-linear dynamics. This unveils how topology might play a role in the onset of brain rhythms.Comment: 17 pages, 16 figures + appendice

    Modelling and recognition of protein contact networks by multiple kernel learning and dissimilarity representations

    Get PDF
    Multiple kernel learning is a paradigm which employs a properly constructed chain of kernel functions able to simultaneously analyse different data or different representations of the same data. In this paper, we propose an hybrid classification system based on a linear combination of multiple kernels defined over multiple dissimilarity spaces. The core of the training procedure is the joint optimisation of kernel weights and representatives selection in the dissimilarity spaces. This equips the system with a two-fold knowledge discovery phase: by analysing the weights, it is possible to check which representations are more suitable for solving the classification problem, whereas the pivotal patterns selected as representatives can give further insights on the modelled system, possibly with the help of field-experts. The proposed classification system is tested on real proteomic data in order to predict proteins' functional role starting from their folded structure: specifically, a set of eight representations are drawn from the graph-based protein folded description. The proposed multiple kernel-based system has also been benchmarked against a clustering-based classification system also able to exploit multiple dissimilarities simultaneously. Computational results show remarkable classification capabilities and the knowledge discovery analysis is in line with current biological knowledge, suggesting the reliability of the proposed system

    A statistical approach for fracture property realization and macroscopic failure analysis of brittle materials

    Get PDF
    Lacking the energy dissipative mechanics such as plastic deformation to rebalance localized stresses, similar to their ductile counterparts, brittle material fracture mechanics is associated with catastrophic failure of purely brittle and quasi-brittle materials at immeasurable and measurable deformation scales respectively. This failure, in the form macroscale sharp cracks, is highly dependent on the composition of the material microstructure. Further, the complexity of this relationship and the resulting crack patterns is exacerbated under highly dynamic loading conditions. A robust brittle material model must account for the multiscale inhomogeneity as well as the probabilistic distribution of the constituents which cause material heterogeneity and influence the complex mechanisms of dynamic fracture responses of the material. Continuum-based homogenization is carried out via finite element-based micromechanical analysis of a material neighbor which gives is geometrically described as a sampling windows (i.e., statistical volume elements). These volume elements are well-defined such that they are representative of the material while propagating material randomness from the inherent microscale defects. Homogenization yields spatially defined elastic and fracture related effective properties, utilized to statistically characterize the material in terms of these properties. This spatial characterization is made possible by performing homogenization at prescribed spatial locations which collectively comprise a non-uniform spatial grid which allows the mapping of each effective material properties to an associated spatial location. Through stochastic decomposition of the derived empirical covariance of the sampled effective material property, the Karhunen-Loeve method is used to generate realizations of a continuous and spatially-correlated random field approximation that preserve the statistics of the material from which it is derived. Aspects of modeling both isotropic and anisotropic brittle materials, from a statistical viewpoint, are investigated to determine how each influences the macroscale fracture response of these materials under highly dynamic conditions. The effects of modeling a material both explicitly by representations of discrete multiscale constituents and/or implicitly by continuum representation of material properties is studies to determine how each model influences the resulting material fracture response. For the implicit material representations, both a statistical white noise (i.e., Weibull-based spatially-uncorrelated) and colored noise (i.e., Karhunen-Loeve spatially-correlated model) random fields are employed herein
    • …
    corecore