2,642 research outputs found

    Internet of things

    Get PDF
    Manual of Digital Earth / Editors: Huadong Guo, Michael F. Goodchild, Alessandro Annoni .- Springer, 2020 .- ISBN: 978-981-32-9915-3Digital Earth was born with the aim of replicating the real world within the digital world. Many efforts have been made to observe and sense the Earth, both from space (remote sensing) and by using in situ sensors. Focusing on the latter, advances in Digital Earth have established vital bridges to exploit these sensors and their networks by taking location as a key element. The current era of connectivity envisions that everything is connected to everything. The concept of the Internet of Things(IoT)emergedasaholisticproposaltoenableanecosystemofvaried,heterogeneous networked objects and devices to speak to and interact with each other. To make the IoT ecosystem a reality, it is necessary to understand the electronic components, communication protocols, real-time analysis techniques, and the location of the objects and devices. The IoT ecosystem and the Digital Earth (DE) jointly form interrelated infrastructures for addressing today’s pressing issues and complex challenges. In this chapter, we explore the synergies and frictions in establishing an efïŹcient and permanent collaboration between the two infrastructures, in order to adequately address multidisciplinary and increasingly complex real-world problems. Although there are still some pending issues, the identiïŹed synergies generate optimism for a true collaboration between the Internet of Things and the Digital Earth

    UK utility data integration: overcoming schematic heterogeneity

    Get PDF
    In this paper we discuss syntactic, semantic and schematic issues which inhibit the integration of utility data in the UK. We then focus on the techniques employed within the VISTA project to overcome schematic heterogeneity. A Global Schema based architecture is employed. Although automated approaches to Global Schema definition were attempted the heterogeneities of the sector were too great. A manual approach to Global Schema definition was employed. The techniques used to define and subsequently map source utility data models to this schema are discussed in detail. In order to ensure a coherent integrated model, sub and cross domain validation issues are then highlighted. Finally the proposed framework and data flow for schematic integration is introduced

    Training of Crisis Mappers and Map Production from Multi-sensor Data: Vernazza Case Study (Cinque Terre National Park, Italy)

    Get PDF
    This aim of paper is to presents the development of a multidisciplinary project carried out by the cooperation between Politecnico di Torino and ITHACA (Information Technology for Humanitarian Assistance, Cooperation and Action). The goal of the project was the training in geospatial data acquiring and processing for students attending Architecture and Engineering Courses, in order to start up a team of "volunteer mappers". Indeed, the project is aimed to document the environmental and built heritage subject to disaster; the purpose is to improve the capabilities of the actors involved in the activities connected in geospatial data collection, integration and sharing. The proposed area for testing the training activities is the Cinque Terre National Park, registered in the World Heritage List since 1997. The area was affected by flood on the 25th of October 2011. According to other international experiences, the group is expected to be active after emergencies in order to upgrade maps, using data acquired by typical geomatic methods and techniques such as terrestrial and aerial Lidar, close-range and aerial photogrammetry, topographic and GNSS instruments etc.; or by non conventional systems and instruments such us UAV, mobile mapping etc. The ultimate goal is to implement a WebGIS platform to share all the data collected with local authorities and the Civil Protectio

    A Two-Level Information Modelling Translation Methodology and Framework to Achieve Semantic Interoperability in Constrained GeoObservational Sensor Systems

    Get PDF
    As geographical observational data capture, storage and sharing technologies such as in situ remote monitoring systems and spatial data infrastructures evolve, the vision of a Digital Earth, first articulated by Al Gore in 1998 is getting ever closer. However, there are still many challenges and open research questions. For example, data quality, provenance and heterogeneity remain an issue due to the complexity of geo-spatial data and information representation. Observational data are often inadequately semantically enriched by geo-observational information systems or spatial data infrastructures and so they often do not fully capture the true meaning of the associated datasets. Furthermore, data models underpinning these information systems are typically too rigid in their data representation to allow for the ever-changing and evolving nature of geo-spatial domain concepts. This impoverished approach to observational data representation reduces the ability of multi-disciplinary practitioners to share information in an interoperable and computable way. The health domain experiences similar challenges with representing complex and evolving domain information concepts. Within any complex domain (such as Earth system science or health) two categories or levels of domain concepts exist. Those concepts that remain stable over a long period of time, and those concepts that are prone to change, as the domain knowledge evolves, and new discoveries are made. Health informaticians have developed a sophisticated two-level modelling systems design approach for electronic health documentation over many years, and with the use of archetypes, have shown how data, information, and knowledge interoperability among heterogenous systems can be achieved. This research investigates whether two-level modelling can be translated from the health domain to the geo-spatial domain and applied to observing scenarios to achieve semantic interoperability within and between spatial data infrastructures, beyond what is possible with current state-of-the-art approaches. A detailed review of state-of-the-art SDIs, geo-spatial standards and the two-level modelling methodology was performed. A cross-domain translation methodology was developed, and a proof-of-concept geo-spatial two-level modelling framework was defined and implemented. The Open Geospatial Consortium’s (OGC) Observations & Measurements (O&M) standard was re-profiled to aid investigation of the two-level information modelling approach. An evaluation of the method was undertaken using II specific use-case scenarios. Information modelling was performed using the two-level modelling method to show how existing historical ocean observing datasets can be expressed semantically and harmonized using two-level modelling. Also, the flexibility of the approach was investigated by applying the method to an air quality monitoring scenario using a technologically constrained monitoring sensor system. This work has demonstrated that two-level modelling can be translated to the geospatial domain and then further developed to be used within a constrained technological sensor system; using traditional wireless sensor networks, semantic web technologies and Internet of Things based technologies. Domain specific evaluation results show that twolevel modelling presents a viable approach to achieve semantic interoperability between constrained geo-observational sensor systems and spatial data infrastructures for ocean observing and city based air quality observing scenarios. This has been demonstrated through the re-purposing of selected, existing geospatial data models and standards. However, it was found that re-using existing standards requires careful ontological analysis per domain concept and so caution is recommended in assuming the wider applicability of the approach. While the benefits of adopting a two-level information modelling approach to geospatial information modelling are potentially great, it was found that translation to a new domain is complex. The complexity of the approach was found to be a barrier to adoption, especially in commercial based projects where standards implementation is low on implementation road maps and the perceived benefits of standards adherence are low. Arising from this work, a novel set of base software components, methods and fundamental geo-archetypes have been developed. However, during this work it was not possible to form the required rich community of supporters to fully validate geoarchetypes. Therefore, the findings of this work are not exhaustive, and the archetype models produced are only indicative. The findings of this work can be used as the basis to encourage further investigation and uptake of two-level modelling within the Earth system science and geo-spatial domain. Ultimately, the outcomes of this work are to recommend further development and evaluation of the approach, building on the positive results thus far, and the base software artefacts developed to support the approach

    The Inclusive City of Johannesburg and the Challenge of Affordable Housing

    Get PDF
    Johannesburg, as South Africa’s largest city and the most important economic center, still suffers under the patterns of inequality of the past. Although the spatial change has been rapid since 1994 after the post-apartheid era, the spatial divisions along class lines still remain. Strategic spatial planning strategies of the last two and a half decades were targeting the reduction of spatial patterns of inequality and exclusion by fostering compact and integrated spatial development on national, regional and urban level and currently the local provision of inclusionary housing. This study, based on literature review and census data evaluation, investigates the current challenges of Johannesburg, such as growing population, existing inequalities and the affordable housing backlog. Moreover, the current affordable housing programmes and new policy requirements in Johannesburg are described and analysed. Despite the release of numerous city-wide policies, such as the Spatial Development Framework 2040 (2016) or the Inclusionary Housing Policy (2019) the research shows that the situation has not significantly improved yet, which is indicated by a lack of restrictions in the requirements, implementation deficiencies and limited policy uptake in planning practice

    StudMap 3.0 : an interoperable web-based platform for geospatial data offers in academic life

    Get PDF
    Dissertation submitted in partial fulfilment of the requirements for the Degree of Master of Science in Geospatial TechnologiesGeographic Information Systems has now entered the realm of web and yields for feasible solutions to balance the technology offers with the users’ needs to share, access and explore the massive amounts of geodata available. Challenges occur when moving forward from old 2D platforms towards innovative and integrated webGIS systems that align functionality with the necessity to grant a complete understanding of the surrounding reality. 3D space responds to this but, however, stands only at the beginning of its era and cannot yet reach the development of 2D web integration. Research is now aiming at possible webGIS solutions to adapt to the special structure imposed by 3D data. In this context, this thesis focuses on designing an architecture for 2D and 3D geospatial data integration on a student-oriented web platform. This concept was further delivered and validated through a real case scenario – Studmap 3.0, a webGIS platform to serve the students of the University of Muenster in their academical life. The portal currently grants availability of geospatial data and web services of regional interest in a smart GIS environment that allows access and comparison of official services with own data. The implementation of Studmap 3.0 aided in the continuous improvement of the proposed architecture model and developed under a design science research cycle that reached its end once the final approval of its users was attained via a usability evaluation. Final strengths and drawbacks of the proposed architecture were ultimately identified together with an expert usability evaluation and a lab-based usability test of the resulting portal interface suitability for academic use. The results fall under the acceptable range with an 83.75 score for the System Usability Scale standardized questions when addressed to experts and a score of 83.87 when addressed to students. For the open-ended questions, the interface received an overall positive critique. A summary of future participants’ opinion on the benefits, drawbacks and proposed improvements was also delivered. Peers interested in similar concepts can use both this model and its final remarks as a reference for their work

    WEB MAPPING ARCHITECTURES BASED ON OPEN SPECIFICATIONS AND FREE AND OPEN SOURCE SOFTWARE IN THE WATER DOMAIN

    Get PDF
    The availability of water-related data and information across different geographical and jurisdictional scales is of critical importance for the conservation and management of water resources in the 21st century. Today information assets are often found fragmented across multiple agencies that use incompatible data formats and procedures for data collection, storage, maintenance, analysis, and distribution. The growing adoption of Web mapping systems in the water domain is reducing the gap between data availability and its practical use and accessibility. Nevertheless, more attention must be given to the design and development of these systems to achieve high levels of interoperability and usability while fulfilling different end user informational needs. This paper first presents a brief overview of technologies used in the water domain, and then presents three examples of Web mapping architectures based on free and open source software (FOSS) and the use of open specifications (OS) that address different users' needs for data sharing, visualization, manipulation, scenario simulations, and map production. The purpose of the paper is to illustrate how the latest developments in OS for geospatial and water-related data collection, storage, and sharing, combined with the use of mature FOSS projects facilitate the creation of sophisticated interoperable Web-based information systems in the water domain

    The PREVIEW Global Risk Data Platform: a geoportal to serve and share global data on risk to natural hazards

    Get PDF
    With growing world population and concentration in urban and coastal areas, the exposure to natural hazards is increasing and results in higher risk of human and economic losses. Improving the identification of areas, population and assets potentially exposed to natural hazards is essential to reduce the consequences of such events. Disaster risk is a function of hazard, exposure and vulnerability. Modelling risk at the global level requires accessing and processing a large number of data, from numerous collaborating centres. <br><br> These data need to be easily updated, and there is a need for centralizing access to this information as well as simplifying its use for non GIS specialists. The Hyogo Framework for Action provides the mandate for data sharing, so that governments and international development agencies can take appropriate decision for disaster risk reduction. <br><br> Timely access and easy integration of geospatial data are essential to support efforts in Disaster Risk Reduction. However various issues in data availability, accessibility and integration limit the use of such data. In consequence, a framework that facilitate sharing and exchange of geospatial data on natural hazards should improve decision-making process. The PREVIEW Global Risk Data Platform is a highly interactive web-based GIS portal supported by a Spatial Data Infrastructure that offers free and interoperable access to more than 60 global data sets on nine types of natural hazards (tropical cyclones and related storm surges, drought, earthquakes, biomass fires, floods, landslides, tsunamis and volcanic eruptions) and related exposure and risk. This application portrays an easy-to-use online interactive mapping interface so that users can easily work with it and seamlessly integrate data in their own data flow using fully compliant OGC Web Services (OWS)

    Designing a Framework for Exchanging Partial Sets of BIM Information on a Cloud-Based Service

    Get PDF
    The rationale behind this research study was based on the recognised difficulty of exchanging data at element or object level due to the inefficiencies of compatible hardware and software. Interoperability depicts the need to pass data between applications, allowing multiple types of experts and applications to contribute to the work at hand. The only way that software file exchanges between two applications can produce consistent data and change management results for large projects is through a building model repository. The overall aim of this thesis was to design and develop an integrated process that would advance key decisions at an early design stage through faster information exchanges during collaborative work. In the construction industry, Building Information Modeling is the most integrated shared model between all disciplines. It is based on a manufacturing-like process where standardised deliverables are used throughout the life cycle with effective collaboration as its main driving force. However, the dilemma is how to share these properties of BIM applications on one single platform asynchronously. Cloud Computing is a centralized heterogeneous network that enables different applications to be connected to each other. The methodology used in the research was based on triangulation of data which incorporated many techniques featuring a mixture of both quantitative and qualitative analysis. The results identified the need to re-engineer Simplified Markup Language, in order to exchange partial data sets of intelligent object architecture on an integrated platform. The designed and tested prototype produced findings that enhanced project decisions at a relatively early design stage, improved communication and collaboration techniques and cross disciple co-ordination

    Geospatial data harmonization from regional level to european level: a usa case in forest fire data

    Get PDF
    Dissertation submitted in partial fulfillment of the requirements for the Degree of Master of Science in Geospatial Technologies.Geospatial data harmonization is becoming more and more important to increase interoperability of heterogeneous data derived from various sources in spatial data infrastructures. To address this harmonization issue we present the current status of data availability among different communities, languages, and administrative scales from regional to national and European levels. With a use case in forest data models in Europe, interoperability of burned area data derived from Europe and Valencia Community in Spain were tested and analyzed on the syntactic, schematic and semantic level. We suggest approaches for achieving a higher chance of data interoperability to guide forest domain experts in forest fire analysis. For testing syntactic interoperability, a common platform in the context of formats and web services was examined. We found that establishing OGC standard web services in a combination with GIS software applications that support various formats and web services can increase the chance of achieving syntactic interoperability between multiple geospatial data derived from different sources. For testing schematic and semantic interoperability, the ontology-based schema mapping approach was taken to transform a regional data model to a European data model on the conceptual level. The Feature Manipulation Engine enabled various types of data transformation from source to target attributes to achieve schematic interoperability. Ontological modelling in Protégé helped identify a common concept between the source and target data models, especially in cases where matching attributes were not found at the schematic level. Establishment of the domain ontology was explored to reach common ground between application ontologies and achieve a higher level of semantic interoperability
    • 

    corecore