49 research outputs found

    Efficient, decentralized detection of qualitative spatial events in a dynamic scalar field

    Get PDF
    This paper describes an efficient, decentralized algorithm to monitor qualitative spatial events in a dynamic scalar field. The events of interest involve changes to the critical points (i.e., peak, pits and passes) and edges of the surface network derived from the field. Four fundamental types of event (appearance, disappearance, movement and switch) are defined. Our algorithm is designed to rely purely on qualitative information about the neighborhoods of nodes in the sensor network and does not require information about nodes' coordinate positions. Experimental investigations confirm that our algorithm is efficient, with O(n) overall communication complexity (where n is the number of nodes in the sensor network), an even load balance and low operational latency. The accuracy of event detection is comparable to established centralized algorithms for the identification of critical points of a surface network. Our algorithm is relevant to a broad range of environmental monitoring applications of sensor networks

    Support Vector Methods for Higher-Level Event Extraction in Point Data

    Get PDF
    Phenomena occur both in space and time. Correspondingly, ability to model spatiotemporal behavior translates into ability to model phenomena as they occur in reality. Given the complexity inherent when integrating spatial and temporal dimensions, however, the establishment of computational methods for spatiotemporal analysis has proven relatively elusive. Nonetheless, one method, the spatiotemporal helix, has emerged from the field of video processing. Designed to efficiently summarize and query the deformation and movement of spatiotemporal events, the spatiotemporal helix has been demonstrated as capable of describing and differentiating the evolution of hurricanes from sequences of images. Being derived from image data, the representations of events for which the spatiotemporal helix was originally created appear in areal form (e.g., a hurricane covering several square miles is represented by groups of pixels). ii Many sources of spatiotemporal data, however, are not in areal form and instead appear as points. Examples of spatiotemporal point data include those from an epidemiologist recording the time and location of cases of disease and environmental observations collected by a geosensor at the point of its location. As points, these data cannot be directly incorporated into the spatiotemporal helix for analysis. However, with the analytic potential for clouds of point data limited, phenomena represented by point data are often described in terms of events. Defined as change units localized in space and time, the concept of events allows for analysis at multiple levels. For instance lower-level events refer to occurrences of interest described by single data streams at point locations (e.g., an individual case of a certain disease or a significant change in chemical concentration in the environment) while higher-level events describe occurrences of interest derived from aggregations of lower-level events and are frequently described in areal form (e.g., a disease cluster or a pollution cloud). Considering that these higher-level events appear in areal form, they could potentially be incorporated into the spatiotemporal helix. With deformation being an important element of spatiotemporal analysis, however, at the crux of a process for spatiotemporal analysis based on point data would be accurate translation of lower-level event points into representations of higher-level areal events. A limitation of current techniques for the derivation of higher-level events is that they imply bias a priori regarding the shape of higher-level events (e.g., elliptical, convex, linear) which could limit the description of the deformation of higher-level events over time. The objective of this research is to propose two newly developed kernel methods, support vector clustering (SVC) and support vector machines (SVMs), as means for iii translating lower-level event points into higher-level event areas that follow the distribution of lower-level points. SVC is suggested for the derivation of higher-level events arising in point process data while SVMs are explored for their potential with scalar field data (i.e., spatially continuous real-valued data). Developed in the field of machine learning to solve complex non-linear problems, both of these methods are capable of producing highly non-linear representations of higher-level events that may be more suitable than existing methods for spatiotemporal analysis of deformation. To introduce these methods, this thesis is organized so that a context for these methods is first established through a description of existing techniques. This discussion leads to a technical explanation of the mechanics of SVC and SVMs and to the implementation of each of the kernel methods on simulated datasets. Results from these simulations inform discussion regarding the application potential of SVC and SVMs

    Development of a GIS-based method for sensor network deployment and coverage optimization

    Get PDF
    Au cours des dernières années, les réseaux de capteurs ont été de plus en plus utilisés dans différents contextes d’application allant de la surveillance de l’environnement au suivi des objets en mouvement, au développement des villes intelligentes et aux systèmes de transport intelligent, etc. Un réseau de capteurs est généralement constitué de nombreux dispositifs sans fil déployés dans une région d'intérêt. Une question fondamentale dans un réseau de capteurs est l'optimisation de sa couverture spatiale. La complexité de l'environnement de détection avec la présence de divers obstacles empêche la couverture optimale de plusieurs zones. Par conséquent, la position du capteur affecte la façon dont une région est couverte ainsi que le coût de construction du réseau. Pour un déploiement efficace d'un réseau de capteurs, plusieurs algorithmes d'optimisation ont été développés et appliqués au cours des dernières années. La plupart de ces algorithmes reposent souvent sur des modèles de capteurs et de réseaux simplifiés. En outre, ils ne considèrent pas certaines informations spatiales de l'environnement comme les modèles numériques de terrain, les infrastructures construites humaines et la présence de divers obstacles dans le processus d'optimisation. L'objectif global de cette thèse est d'améliorer les processus de déploiement des capteurs en intégrant des informations et des connaissances géospatiales dans les algorithmes d'optimisation. Pour ce faire, trois objectifs spécifiques sont définis. Tout d'abord, un cadre conceptuel est développé pour l'intégration de l'information contextuelle dans les processus de déploiement des réseaux de capteurs. Ensuite, sur la base du cadre proposé, un algorithme d'optimisation sensible au contexte local est développé. L'approche élargie est un algorithme local générique pour le déploiement du capteur qui a la capacité de prendre en considération de l'information spatiale, temporelle et thématique dans différents contextes d'applications. Ensuite, l'analyse de l'évaluation de la précision et de la propagation d'erreurs est effectuée afin de déterminer l'impact de l'exactitude des informations contextuelles sur la méthode d'optimisation du réseau de capteurs proposée. Dans cette thèse, l'information contextuelle a été intégrée aux méthodes d'optimisation locales pour le déploiement de réseaux de capteurs. L'algorithme développé est basé sur le diagramme de Voronoï pour la modélisation et la représentation de la structure géométrique des réseaux de capteurs. Dans l'approche proposée, les capteurs change leur emplacement en fonction des informations contextuelles locales (l'environnement physique, les informations de réseau et les caractéristiques des capteurs) visant à améliorer la couverture du réseau. La méthode proposée est implémentée dans MATLAB et est testée avec plusieurs jeux de données obtenus à partir des bases de données spatiales de la ville de Québec. Les résultats obtenus à partir de différentes études de cas montrent l'efficacité de notre approche.In recent years, sensor networks have been increasingly used for different applications ranging from environmental monitoring, tracking of moving objects, development of smart cities and smart transportation system, etc. A sensor network usually consists of numerous wireless devices deployed in a region of interest. A fundamental issue in a sensor network is the optimization of its spatial coverage. The complexity of the sensing environment with the presence of diverse obstacles results in several uncovered areas. Consequently, sensor placement affects how well a region is covered by sensors as well as the cost for constructing the network. For efficient deployment of a sensor network, several optimization algorithms are developed and applied in recent years. Most of these algorithms often rely on oversimplified sensor and network models. In addition, they do not consider spatial environmental information such as terrain models, human built infrastructures, and the presence of diverse obstacles in the optimization process. The global objective of this thesis is to improve sensor deployment processes by integrating geospatial information and knowledge in optimization algorithms. To achieve this objective three specific objectives are defined. First, a conceptual framework is developed for the integration of contextual information in sensor network deployment processes. Then, a local context-aware optimization algorithm is developed based on the proposed framework. The extended approach is a generic local algorithm for sensor deployment, which accepts spatial, temporal, and thematic contextual information in different situations. Next, an accuracy assessment and error propagation analysis is conducted to determine the impact of the accuracy of contextual information on the proposed sensor network optimization method. In this thesis, the contextual information has been integrated in to the local optimization methods for sensor network deployment. The extended algorithm is developed based on point Voronoi diagram in order to represent geometrical structure of sensor networks. In the proposed approach sensors change their location based on local contextual information (physical environment, network information and sensor characteristics) aiming to enhance the network coverage. The proposed method is implemented in MATLAB and tested with several data sets obtained from Quebec City spatial database. Obtained results from different case studies show the effectiveness of our approach

    Towards Spatial Queries over Phenomena in Sensor Networks

    Get PDF
    Today, technology developments enable inexpensive production and deployment of tiny sensing and computing nodes. Networked through wireless radio, such senor nodes form a new platform, wireless sensor networks, which provide novel ability to monitor spatiotemporally continuous phenomena. By treating a wireless sensor network as a database system, users can pose SQL-based queries over phenomena without needing to program detailed sensor node operations. DBMS-internally, intelligent and energyefficient data collection and processing algorithms have to be implemented to support spatial query processing over sensor networks. This dissertation proposes spatial query support for two views of continuous phenomena: field-based and object-based. A field-based view of continuous phenomena depicts them as a value distribution over a geographical area. However, due to the discrete and comparatively sparse distribution of sensor nodes, estimation methods are necessary to generate a field-based query result, and it has to be computed collaboratively ‘in-the-network’ due to energy constraints. This dissertation proposes SWOP, an in-network algorithm using Gaussian Kernel estimation. The key contribution is the use of a small number of Hermite coefficients to approximate the Gaussian Kernel function for sub-clustered sensor nodes, and processes the estimation result efficiently. An object-based view of continuous phenomena is interested in aspects such as the boundary of an ‘interesting region’ (e.g. toxic plume). This dissertation presents NED, which provides object boundary detection in sensor networks. NED encodes partial event estimation results based on confidence levels into optimized, variable length messages exchanged locally among neighboring sensor nodes to save communication cost. Therefore, sensor nodes detect objects and boundaries based on moving averages to eliminate noise effects and enhance detection quality. Furthermore, the dissertation proposes the SNAKE-based approach, which uses deformable curves to track the spatiotemporal changes of such objects incrementally in sensor networks. In the proposed algorithm, only neighboring nodes exchange messages to maintain the curve structures. Based on in-network tracking of deformable curves, other types of spatial and spatiotemporal properties of objects, such as area, can be provided by the sensor network. The experimental results proved that our approaches are resource friendly within the constrained sensor networks, while providing high quality query results

    SST: Integrated Fluorocarbon Microsensor System Using Catalytic Modification

    Get PDF
    Selective, sensitive, and reliable sensors are urgently needed to detect air-borne halogenated volatile organic compounds (VOCs). This broad class of compounds includes chlorine, fluorine, bromine, and iodine containing hydrocarbons used as solvents, refrigerants, herbicides, and more recently as chemical warfare agents (CWAs). It is important to be able to detect very low concentrations of halocarbon solvents and insecticides because of their acute health effects even in very low concentrations. For instance, the nerve agent sarin (isopropyl methylphosphonofluoridate), first developed as an insecticide by German chemists in 1938, is so toxic that a ten minute exposure at an airborne concentration of only 65 parts per billion (ppb) can be fatal. Sarin became a household term when religious cult members on Tokyo subway trains poisoned over 5,500 people, killing 12. Sarin and other CWAs remain a significant threat to the health and safety of the general public. The goal of this project is to design a sensor system to detect and identify the composition and concentration of fluorinated VOCs. The system should be small, robust, compatible with metal oxide semiconductor (MOS) technology, cheap, if produced in large scale, and has the potential to be versatile in terms of low power consumption, detection of other gases, and integration in a portable system. The proposed VOC sensor system has three major elements that will be integrated into a microreactor flow cell: a temperature-programmable microhotplate array/reactor system which serves as the basic sensor platform; an innovative acoustic wave sensor, which detects material removal (instead of deposition) to verify and quantify the presence of fluorine; and an intelligent method, support vector machines, that will analyze the complex and high dimensional data furnished by the sensor system. The superior and complementary aspects of the three elements will be carefully integrated to create a system which is more sensitive and selective than other CWA detection systems that are commercially available or described in the research literature. While our sensor system will be developed to detect fluorinated VOCs, it can be adapted for other applications in which a target analyte can be catalytically converted for selective detection. Therefore, this investigation will examine the relationships between individual sensor element performance and joint sensor platform performance, integrated with state-of-the-art data analysis techniques. During development of the sensor system, the investigators will consider traditional reactor design concepts such as mass transfer and residence time effects, and will apply them to the emerging field of microsystems. The proposed research will provide the fundamental basis and understanding for examining multifunctional sensor platforms designed to provide extreme selectivity to targeted molecules. The project will involve interdisciplinary researchers and students and will connect to K-12 and RET programs for underrepresented students from rural areas

    Swarm Intelligence

    Get PDF
    Swarm Intelligence has emerged as one of the most studied artificial intelligence branches during the last decade, constituting the fastest growing stream in the bio-inspired computation community. A clear trend can be deduced analyzing some of the most renowned scientific databases available, showing that the interest aroused by this branch has increased at a notable pace in the last years. This book describes the prominent theories and recent developments of Swarm Intelligence methods, and their application in all fields covered by engineering. This book unleashes a great opportunity for researchers, lecturers, and practitioners interested in Swarm Intelligence, optimization problems, and artificial intelligence

    Spatiotemporal Wireless Sensor Network Field Approximation with Multilayer Perceptron Artificial Neural Network Models

    Get PDF
    As sensors become increasingly compact and dependable in natural environments, spatially-distributed heterogeneous sensor network systems steadily become more pervasive. However, any environmental monitoring system must account for potential data loss due to a variety of natural and technological causes. Modeling a natural spatial region can be problematic due to spatial nonstationarities in environmental variables, and as particular regions may be subject to specific influences at different spatial scales. Relationships between processes within these regions are often ephemeral, so models designed to represent them cannot remain static. Integrating temporal factors into this model engenders further complexity. This dissertation evaluates the use of multilayer perceptron neural network models in the context of sensor networks as a possible solution to many of these problems given their data-driven nature, their representational flexibility and straightforward fitting process. The relative importance of parameters is determined via an adaptive backpropagation training process, which converges to a best-fit model for sensing platforms to validate collected data or approximate missing readings. As conditions evolve over time such that the model can no longer adapt to changes, new models are trained to replace the old. We demonstrate accuracy results for the MLP generally on par with those of spatial kriging, but able to integrate additional physical and temporal parameters, enabling its application to any region with a collection of available data streams. Potential uses of this model might be not only to approximate missing data in the sensor field, but also to flag potentially incorrect, unusual or atypical data returned by the sensor network. Given the potential for spatial heterogeneity in a monitored phenomenon, this dissertation further explores the benefits of partitioning a space and applying individual MLP models to these partitions. A system of neural models using both spatial and temporal parameters can be envisioned such that a spatiotemporal space partitioned by k-means is modeled by k neural models with internal weightings varying individually according to the dominant processes within the assigned region of each. Evaluated on simulated and real data on surface currents of theGulf ofMaine, partitioned models show significant improved results over single global models

    Situation awareness in environmental monitoring

    Get PDF

    A Two-Level Information Modelling Translation Methodology and Framework to Achieve Semantic Interoperability in Constrained GeoObservational Sensor Systems

    Get PDF
    As geographical observational data capture, storage and sharing technologies such as in situ remote monitoring systems and spatial data infrastructures evolve, the vision of a Digital Earth, first articulated by Al Gore in 1998 is getting ever closer. However, there are still many challenges and open research questions. For example, data quality, provenance and heterogeneity remain an issue due to the complexity of geo-spatial data and information representation. Observational data are often inadequately semantically enriched by geo-observational information systems or spatial data infrastructures and so they often do not fully capture the true meaning of the associated datasets. Furthermore, data models underpinning these information systems are typically too rigid in their data representation to allow for the ever-changing and evolving nature of geo-spatial domain concepts. This impoverished approach to observational data representation reduces the ability of multi-disciplinary practitioners to share information in an interoperable and computable way. The health domain experiences similar challenges with representing complex and evolving domain information concepts. Within any complex domain (such as Earth system science or health) two categories or levels of domain concepts exist. Those concepts that remain stable over a long period of time, and those concepts that are prone to change, as the domain knowledge evolves, and new discoveries are made. Health informaticians have developed a sophisticated two-level modelling systems design approach for electronic health documentation over many years, and with the use of archetypes, have shown how data, information, and knowledge interoperability among heterogenous systems can be achieved. This research investigates whether two-level modelling can be translated from the health domain to the geo-spatial domain and applied to observing scenarios to achieve semantic interoperability within and between spatial data infrastructures, beyond what is possible with current state-of-the-art approaches. A detailed review of state-of-the-art SDIs, geo-spatial standards and the two-level modelling methodology was performed. A cross-domain translation methodology was developed, and a proof-of-concept geo-spatial two-level modelling framework was defined and implemented. The Open Geospatial Consortium’s (OGC) Observations & Measurements (O&M) standard was re-profiled to aid investigation of the two-level information modelling approach. An evaluation of the method was undertaken using II specific use-case scenarios. Information modelling was performed using the two-level modelling method to show how existing historical ocean observing datasets can be expressed semantically and harmonized using two-level modelling. Also, the flexibility of the approach was investigated by applying the method to an air quality monitoring scenario using a technologically constrained monitoring sensor system. This work has demonstrated that two-level modelling can be translated to the geospatial domain and then further developed to be used within a constrained technological sensor system; using traditional wireless sensor networks, semantic web technologies and Internet of Things based technologies. Domain specific evaluation results show that twolevel modelling presents a viable approach to achieve semantic interoperability between constrained geo-observational sensor systems and spatial data infrastructures for ocean observing and city based air quality observing scenarios. This has been demonstrated through the re-purposing of selected, existing geospatial data models and standards. However, it was found that re-using existing standards requires careful ontological analysis per domain concept and so caution is recommended in assuming the wider applicability of the approach. While the benefits of adopting a two-level information modelling approach to geospatial information modelling are potentially great, it was found that translation to a new domain is complex. The complexity of the approach was found to be a barrier to adoption, especially in commercial based projects where standards implementation is low on implementation road maps and the perceived benefits of standards adherence are low. Arising from this work, a novel set of base software components, methods and fundamental geo-archetypes have been developed. However, during this work it was not possible to form the required rich community of supporters to fully validate geoarchetypes. Therefore, the findings of this work are not exhaustive, and the archetype models produced are only indicative. The findings of this work can be used as the basis to encourage further investigation and uptake of two-level modelling within the Earth system science and geo-spatial domain. Ultimately, the outcomes of this work are to recommend further development and evaluation of the approach, building on the positive results thus far, and the base software artefacts developed to support the approach

    Generating Contour Maps for Dynamic Fields Monitored by Sensor Networks

    Get PDF
    Wireless sensor networks provide a new tool that enables researchers and scientists to efficiently monitor dynamic fields. In order to extend the lifetime of the network, it is important for us to minimize network data transmission as much as possible. Previous work proposed many useful aggregation techniques to answer max, min and average questions, and some of them have been employed in real applications. But we cannot get spatial information from these aggregation techniques. This thesis presents an efficient aggregation technique for continuous generation of contour maps for a dynamic field monitored by a wireless sensor network. A contour map is a useful data representation schema that provides an efficient way to visualize an approximation to the monitored field. In this thesis, we discuss an energy-efficient technique, which we call Isovector Aggregation, for generating such contours using an in-network approach. Our technique achieves energy efficiency in two principal ways. Firstly, only a selection of nodes close to contours are chosen to report, and each reported message contains information about a part or all of the contours, rather than any single node’s ID and value pair. Secondly, contours are progressively merged and simplified along the data routing tree, which eliminates many unnecessary contour points from contour vectors before they are transmitted back to the base station. Using Isovector Aggregation, the base station receives a complete representation of the contours that requires no further processing. Analysis shows that for region-related monitoring, Isovector Aggregation is the only technique that has O( n) traffic generation and that considers in-network traffic reduction at the same time. These two factors make Isovector Aggregation highly scalable. Experimental results using simulations also show that Isovector Aggregation involves considerably less data transmission compared to other approaches, such as the no-aggregation approach and the Isolines Aggregation technique, without compromising the accuracy of representations of the baseline maps
    corecore