124 research outputs found

    Big Data Computing for Geospatial Applications

    Get PDF
    The convergence of big data and geospatial computing has brought forth challenges and opportunities to Geographic Information Science with regard to geospatial data management, processing, analysis, modeling, and visualization. This book highlights recent advancements in integrating new computing approaches, spatial methods, and data management strategies to tackle geospatial big data challenges and meanwhile demonstrates opportunities for using big data for geospatial applications. Crucial to the advancements highlighted in this book is the integration of computational thinking and spatial thinking and the transformation of abstract ideas and models to concrete data structures and algorithms

    Semantic Search and Discovery for Earth Observation Products using Ontology Services

    Get PDF
    Η πρόσβαση σε δεδομένα που έχουν προέλθει από την παρατήρηση της Γης παραμένει δύσκολη για τους περισσότερους απλούς χρήστες μέχρι και σήμερα. Οι υπάρχουσες μηχανές αναζήτησης απευθύνονται σε ειδικούς του πεδίου παρατήρησης της Γης, αδυνατώντας να καλύψουν τις ανάγκες επιστημονικών κοινοτήτων από άλλα πεδία, καθώς και απλών χρηστών που δεν είναι εξοικειωμένοι με τα δεδομένα παρατήρησης της Γης. Στα πλαίσια αυτής της διπλωματικής αναπτύχθηκαν σημασιολογικές τεχνολογίες οι οποίες ενσωματώθηκαν σε μια πλατφόρμα αναζήτησης ΕΟ-netCDF δεδομένων. Οι τεχνολογίες αυτές με τη χρήση οντολογιών επιτρέπουν την εύκολη αναζήτηση και πρόσβαση σε δεδομένα που έχουν προέλθει από την παρατήρηση της Γης.Access to Earth Observation products remains difficult for end-users in most domains. Although various search engines have been developed, these are targeted for advanced Earth Observation users, and fail to support scientific communities from other domains, as well as casual users not familiar with the concepts of Earth Observation.In the context of this thesis, we developed semantic technologies that were used to semantically enhance a search engine for EO-netCDF product. We present how these technologies utilize ontology services to substantially improve the ability of end-users to explore, understand and exploit the vast amount of Earth Observation data that is available nowadays

    Raspberry Pi Based Intelligent Wireless Sensor Node for Localized Torrential Rain Monitoring

    Get PDF
    Wireless sensor networks are proved to be effective in long-time localized torrential rain monitoring. However, the existing widely used architecture of wireless sensor networks for rain monitoring relies on network transportation and back-end calculation, which causes delay in response to heavy rain in localized areas. Our work improves the architecture by applying logistic regression and support vector machine classification to an intelligent wireless sensor node which is created by Raspberry Pi. The sensor nodes in front-end not only obtain data from sensors, but also can analyze the probabilities of upcoming heavy rain independently and give early warnings to local clients in time. When the sensor nodes send the probability to back-end server, the burdens of network transport are released. We demonstrate by simulation results that our sensor system architecture has potentiality to increase the local response to heavy rain. The monitoring capacity is also raised

    INTELLIGENT CYBERINFRASTRUCTURE FOR BIG DATA ENABLED HYDROLOGICAL MODELING, PREDICTION, AND EVALUATION

    Get PDF
    Most hydrologic data are associated with spatiotemporal information, which is capable of presenting patterns and changes in both spatial and temporal aspects. The demands of retrieving, managing, analyzing, visualizing, and sharing these data have been continuously increasing. However, spatiotemporal hydrologic data are generally complex, which can be difficult to work with knowledge from hydrology alone. With the assistance of geographic information systems (GIS) and web-based technologies, a solution of establishing a cyberinfrastructure as the backbone to support such demands has emerged. This interdisciplinary dissertation described the advancement of traditional approaches for organizing and managing spatiotemporal hydrologic data, integrating and executing hydrologic models, analyzing and evaluating the results, and sharing the entire process. A pilot study was conducted in Chapter 2, in which a globally shared flood cyberinfrastructure was created to collect, organize, and manage flood databases that visually provide useful information to authorities and the public in real-time. The cyberinfrastructure used public cloud services provided by Google Fusion Table and crowdsourcing data collection methods to provide location-based visualization as well as statistical analysis and graphing capabilities. This study intended to engage citizen-scientists and presented an opportunity to modernize the existing paradigm used to collect, manage, analyze, and visualize water-related disasters eventually. An observationally based monthly evapotranspiration (ET) product was produced in Chapter 3, using the simple water balance equation across the conterminous United States (CONUS). The best quality ground- and satellite-based observations of the water budget components, i.e., precipitation, runoff, and water storage change were adopted, while ET is computed as the residual. A land surface model-based downscaling approach to disaggregate the monthly GRACE equivalent water thickness (EWT) data to daily, 0.125º values was developed. The derived ET was evaluated against three sets of existing ET products and showed reliable results. The new ET product and the disaggregated GRACE data could be used as a benchmark dataset for researches in hydrological and climatological changes and terrestrial water and energy cycle dynamics over the CONUS. The study in Chapter 4 developed an automated hydrological modeling framework for any non-hydrologists with internet access, who can organize hydrologic data, execute hydrologic models, and visualize results graphically and statistically for further analysis in real-time. By adopting Hadoop distributed file system (HDFS) and Apache Hive, the efficiency of data processing and query were significantly increased. Two lumped hydrologic models, lumped Coupled Routing and Excess STorage (CREST) model and HyMOD model, were integrated as a proof of concept in this web framework. Evaluation of selected basins over the CONUS were performed as a demonstration. Our vision is to simplify the processes of using hydrologic models for researchers and modelers, as well as to unlock the potential and educate the less experienced public on hydrologic models

    The role of semantic web technologies for IoT data in underpinning environmental science

    Get PDF
    The advent of Internet of Things (IoT) technology has the potential to generate a huge amount of heterogeneous data at different geographical locations and with various temporal resolutions in environmental science. In many other areas of IoT deployment, volume and velocity dominate, however in environmental science, the more general pattern is quite distinct and often variety dominates. There exists a large number of small, heterogeneous and potentially complex datasets and the key challenge is to understand the interdependencies between these disparate datasets representing different environmental facets. These characteristics pose several data challenges including data interpretation, interoperability and integration, to name but a few, and there is a pressing need to address these challenges. The author postulates that Semantic Web technologies and associated techniques have the potential to address the aforementioned data challenges and support environmental science. The main goal of this thesis is to examine the potential role of Semantic Web technologies in making sense of such complex and heterogeneous environmental data in all its complexity. The thesis explores the state-of-the-art in the use of such technologies in the context of environmental science. After an in-depth assessment of related work, the thesis further examined the characteristics of environmental data through semi-structured interviews with leading experts. Through this, three key research challenges emerge: discovering interdependencies between disparate datasets, geospatial data integration and reasoning, and data heterogeneity. In response to these challenges, an ontology was developed that semantically enriches all sensor measurements stemmed from an experimental Environmental IoT infrastructure. The resultant ontology was evaluated through three real-world use-cases derived from the interviews. This led to a number of major contributions from this work including: the development of an ontology tailored for streaming environmental data offering semantic enrichment of IoT data, support for spatio-temporal data integration and reasoning, and the analysis of unique characteristics of environmental science around data

    A Two-Level Information Modelling Translation Methodology and Framework to Achieve Semantic Interoperability in Constrained GeoObservational Sensor Systems

    Get PDF
    As geographical observational data capture, storage and sharing technologies such as in situ remote monitoring systems and spatial data infrastructures evolve, the vision of a Digital Earth, first articulated by Al Gore in 1998 is getting ever closer. However, there are still many challenges and open research questions. For example, data quality, provenance and heterogeneity remain an issue due to the complexity of geo-spatial data and information representation. Observational data are often inadequately semantically enriched by geo-observational information systems or spatial data infrastructures and so they often do not fully capture the true meaning of the associated datasets. Furthermore, data models underpinning these information systems are typically too rigid in their data representation to allow for the ever-changing and evolving nature of geo-spatial domain concepts. This impoverished approach to observational data representation reduces the ability of multi-disciplinary practitioners to share information in an interoperable and computable way. The health domain experiences similar challenges with representing complex and evolving domain information concepts. Within any complex domain (such as Earth system science or health) two categories or levels of domain concepts exist. Those concepts that remain stable over a long period of time, and those concepts that are prone to change, as the domain knowledge evolves, and new discoveries are made. Health informaticians have developed a sophisticated two-level modelling systems design approach for electronic health documentation over many years, and with the use of archetypes, have shown how data, information, and knowledge interoperability among heterogenous systems can be achieved. This research investigates whether two-level modelling can be translated from the health domain to the geo-spatial domain and applied to observing scenarios to achieve semantic interoperability within and between spatial data infrastructures, beyond what is possible with current state-of-the-art approaches. A detailed review of state-of-the-art SDIs, geo-spatial standards and the two-level modelling methodology was performed. A cross-domain translation methodology was developed, and a proof-of-concept geo-spatial two-level modelling framework was defined and implemented. The Open Geospatial Consortium’s (OGC) Observations & Measurements (O&M) standard was re-profiled to aid investigation of the two-level information modelling approach. An evaluation of the method was undertaken using II specific use-case scenarios. Information modelling was performed using the two-level modelling method to show how existing historical ocean observing datasets can be expressed semantically and harmonized using two-level modelling. Also, the flexibility of the approach was investigated by applying the method to an air quality monitoring scenario using a technologically constrained monitoring sensor system. This work has demonstrated that two-level modelling can be translated to the geospatial domain and then further developed to be used within a constrained technological sensor system; using traditional wireless sensor networks, semantic web technologies and Internet of Things based technologies. Domain specific evaluation results show that twolevel modelling presents a viable approach to achieve semantic interoperability between constrained geo-observational sensor systems and spatial data infrastructures for ocean observing and city based air quality observing scenarios. This has been demonstrated through the re-purposing of selected, existing geospatial data models and standards. However, it was found that re-using existing standards requires careful ontological analysis per domain concept and so caution is recommended in assuming the wider applicability of the approach. While the benefits of adopting a two-level information modelling approach to geospatial information modelling are potentially great, it was found that translation to a new domain is complex. The complexity of the approach was found to be a barrier to adoption, especially in commercial based projects where standards implementation is low on implementation road maps and the perceived benefits of standards adherence are low. Arising from this work, a novel set of base software components, methods and fundamental geo-archetypes have been developed. However, during this work it was not possible to form the required rich community of supporters to fully validate geoarchetypes. Therefore, the findings of this work are not exhaustive, and the archetype models produced are only indicative. The findings of this work can be used as the basis to encourage further investigation and uptake of two-level modelling within the Earth system science and geo-spatial domain. Ultimately, the outcomes of this work are to recommend further development and evaluation of the approach, building on the positive results thus far, and the base software artefacts developed to support the approach

    Helmholtz Portfolio Theme Large-Scale Data Management and Analysis (LSDMA)

    Get PDF
    The Helmholtz Association funded the "Large-Scale Data Management and Analysis" portfolio theme from 2012-2016. Four Helmholtz centres, six universities and another research institution in Germany joined to enable data-intensive science by optimising data life cycles in selected scientific communities. In our Data Life cycle Labs, data experts performed joint R&D together with scientific communities. The Data Services Integration Team focused on generic solutions applied by several communities

    24th International Conference on Information Modelling and Knowledge Bases

    Get PDF
    In the last three decades information modelling and knowledge bases have become essentially important subjects not only in academic communities related to information systems and computer science but also in the business area where information technology is applied. The series of European – Japanese Conference on Information Modelling and Knowledge Bases (EJC) originally started as a co-operation initiative between Japan and Finland in 1982. The practical operations were then organised by professor Ohsuga in Japan and professors Hannu Kangassalo and Hannu Jaakkola in Finland (Nordic countries). Geographical scope has expanded to cover Europe and also other countries. Workshop characteristic - discussion, enough time for presentations and limited number of participants (50) / papers (30) - is typical for the conference. Suggested topics include, but are not limited to: 1. Conceptual modelling: Modelling and specification languages; Domain-specific conceptual modelling; Concepts, concept theories and ontologies; Conceptual modelling of large and heterogeneous systems; Conceptual modelling of spatial, temporal and biological data; Methods for developing, validating and communicating conceptual models. 2. Knowledge and information modelling and discovery: Knowledge discovery, knowledge representation and knowledge management; Advanced data mining and analysis methods; Conceptions of knowledge and information; Modelling information requirements; Intelligent information systems; Information recognition and information modelling. 3. Linguistic modelling: Models of HCI; Information delivery to users; Intelligent informal querying; Linguistic foundation of information and knowledge; Fuzzy linguistic models; Philosophical and linguistic foundations of conceptual models. 4. Cross-cultural communication and social computing: Cross-cultural support systems; Integration, evolution and migration of systems; Collaborative societies; Multicultural web-based software systems; Intercultural collaboration and support systems; Social computing, behavioral modeling and prediction. 5. Environmental modelling and engineering: Environmental information systems (architecture); Spatial, temporal and observational information systems; Large-scale environmental systems; Collaborative knowledge base systems; Agent concepts and conceptualisation; Hazard prediction, prevention and steering systems. 6. Multimedia data modelling and systems: Modelling multimedia information and knowledge; Contentbased multimedia data management; Content-based multimedia retrieval; Privacy and context enhancing technologies; Semantics and pragmatics of multimedia data; Metadata for multimedia information systems. Overall we received 56 submissions. After careful evaluation, 16 papers have been selected as long paper, 17 papers as short papers, 5 papers as position papers, and 3 papers for presentation of perspective challenges. We thank all colleagues for their support of this issue of the EJC conference, especially the program committee, the organising committee, and the programme coordination team. The long and the short papers presented in the conference are revised after the conference and published in the Series of “Frontiers in Artificial Intelligence” by IOS Press (Amsterdam). The books “Information Modelling and Knowledge Bases” are edited by the Editing Committee of the conference. We believe that the conference will be productive and fruitful in the advance of research and application of information modelling and knowledge bases. Bernhard Thalheim Hannu Jaakkola Yasushi Kiyok

    Smart breeding driven by big data, artificial intelligence, and integrated genomic-enviromic prediction

    Get PDF
    The first paradigm of plant breeding involves direct selection-based phenotypic observation, followed by predictive breeding using statistical models for quantitative traits constructed based on genetic experimental design and, more recently, by incorporation of molecular marker genotypes. However, plant performance or phenotype (P) is determined by the combined effects of genotype (G), envirotype (E), and genotype by environment interaction (GEI). Phenotypes can be predicted more precisely by training a model using data collected from multiple sources, including spatiotemporal omics (genomics, phenomics, and enviromics across time and space). Integration of 3D information profiles (G-P-E), each with multidimensionality, provides predictive breeding with both tremendous opportunities and great challenges. Here, we first review innovative technologies for predictive breeding. We then evaluate multidimensional information profiles that can be integrated with a predictive breeding strategy, particularly envirotypic data, which have largely been neglected in data collection and are nearly untouched in model construction. We propose a smart breeding scheme, integrated genomic-enviromic prediction (iGEP), as an extension of genomic prediction, using integrated multiomics information, big data technology, and artificial intelligence (mainly focused on machine and deep learning). We discuss how to implement iGEP, including spatiotemporal models, environmental indices, factorial and spatiotemporal structure of plant breeding data, and cross-species prediction. A strategy is then proposed for prediction-based crop redesign at both the macro (individual, population, and species) and micro (gene, metabolism, and network) scales. Finally, we provide perspectives on translating smart breeding into genetic gain through integrative breeding platforms and open-source breeding initiatives. We call for coordinated efforts in smart breeding through iGEP, institutional partnerships, and innovative technological support
    corecore