9 research outputs found

    Leveraging spatio-temporal redundancy for RFID data cleansing

    Full text link

    FP-Growth Tree Based Algorithms Analysis: CP-Tree and K Map

    Get PDF
    We propose a novel frequent-pattern tree (FP-tree) structure; our performance study shows that the FP-growth method is efficient and scalable for mining both long and short frequent patterns, and is about an order of magnitude faster than the Apriori algorithm and also faster than some recently reported new frequent-pattern mining methods. FP-tree method is efficient algorithm in association mining to mine frequent patterns in data mining, in spite of long or short frequent data patterns. By using compact best tree structure and partitioning-based and divide-and-conquer data mining searching method, it can be reduces the costs searchsubstantially .it just as the analysis multi-CPU or reduce computer memory to solve problem. But this approach can be apparently decrease the costs for exchanging and combining control information and the algorithm complexity is also greatly decreased, solve this problem efficiently. Even if main adopting multi-CPU technique, raising the requirement is basically hardware, best performanceimprovement is still to be limited. Is there any other way that most one may it can reduce these costs in FP-tree construction, performance best improvement is still limited

    RFID-Based Indoor Spatial Query Evaluation with Bayesian Filtering Techniques

    Full text link
    People spend a significant amount of time in indoor spaces (e.g., office buildings, subway systems, etc.) in their daily lives. Therefore, it is important to develop efficient indoor spatial query algorithms for supporting various location-based applications. However, indoor spaces differ from outdoor spaces because users have to follow the indoor floor plan for their movements. In addition, positioning in indoor environments is mainly based on sensing devices (e.g., RFID readers) rather than GPS devices. Consequently, we cannot apply existing spatial query evaluation techniques devised for outdoor environments for this new challenge. Because Bayesian filtering techniques can be employed to estimate the state of a system that changes over time using a sequence of noisy measurements made on the system, in this research, we propose the Bayesian filtering-based location inference methods as the basis for evaluating indoor spatial queries with noisy RFID raw data. Furthermore, two novel models, indoor walking graph model and anchor point indexing model, are created for tracking object locations in indoor environments. Based on the inference method and tracking models, we develop innovative indoor range and k nearest neighbor (kNN) query algorithms. We validate our solution through use of both synthetic data and real-world data. Our experimental results show that the proposed algorithms can evaluate indoor spatial queries effectively and efficiently. We open-source the code, data, and floor plan at https://github.com/DataScienceLab18/IndoorToolKit

    Distributed Inference and Query Processing for RFID Tracking and Monitoring

    Get PDF
    In this paper, we present the design of a scalable, distributed stream processing system for RFID tracking and monitoring. Since RFID data lacks containment and location information that is key to query processing, we propose to combine location and containment inference with stream query processing in a single architecture, with inference as an enabling mechanism for high-level query processing. We further consider challenges in instantiating such a system in large distributed settings and design techniques for distributed inference and query processing. Our experimental results, using both real-world data and large synthetic traces, demonstrate the accuracy, efficiency, and scalability of our proposed techniques.Comment: VLDB201

    Cleansing Indoor RFID Tracking Data

    Get PDF

    Hybrid approach on multi- spatiotemporal data framework towards analysis of long-lead upstream flood: a case of Niger State, Nigeria

    Get PDF
    Floods have become a global concern because of the vast economic and ecological havoc that ensue. Thus, a flood risk mitigation strategy is used to reduce flood-related consequences by a long-lead identification of its occurrence. A wide range of causative factors, including the adoption of hybrid multi-spatiotemporal data framework is considered in implementing the strategy. Besides the structural or homogenous non-structural factors, the adoption of various Information Systems-based tools are also required to accurately analyse the multiple natural causative factors. Essentially, this was needed to address the inaccurate flood vulnerability classifications and short time of flood prediction. Thus, this study proposes a framework named: Hybrid Multi-spatiotemporal data Framework for Long-lead Upstream Flood Analysis (HyM-SLUFA) to provide a new dimension on flood vulnerability studies by uncovering the influence of multiple factors derived from topography, hydrology, vegetal and precipitation features towards regional flood vulnerability classification and long-lead analysis. In developing the proposed framework, the spatial images were geometrically and radiometrically corrected with the aid of Quantum Geographic Information System (QGIS). The temporal data were cleaned by means of winsorization methods using STATA statistical tool. The hybrid segment of the framework classifies flood vulnerability and performs long-lead analysis. The classification and analysis were conducted using the corrected spatial images to acquire better understanding on the interaction between the extracted features and rainfall in inducing flood as well as producing various regional flood vulnerabilities within the study area. Additionally, with the aid of regression technique, precipitation and water level data were used to perform long-lead flood analysis to provide a foresight of any potential flooding event in order to take proactive measures. As to confirm the reliability and validity of the proposed framework, an accuracy assessment was conducted on the outputs of the data. This study found the influence of various Flood Causative Factors (FCFs) used in the developed HyM-SLUFA framework, by revealing the spatial disparity indicating that the slope of a region shows a more accurate level of flood vulnerability compared to other FCFs, which generally causes severe upstream floods when there is low volume of precipitation within regions of low slope degree. Theoretically, the HyM-SLUFA will serve as a guide that can be adopted or adapted for similar studies. Especially, by considering the trend of precipitation and the pattern of flood vulnerability classifications depicted by various FCFs. These classifications will determine the kind(s) of policies that will be implemented in town planning, and the Flood Inducible Precipitation Volumes can provide a foresight of any potential flooding event in order to take practical proactive measures by the local authority

    The design and development of multi-agent based RFID middleware system for data and devices management

    Get PDF
    Thesis (D. Tech. (Electrical Engineering)) - Central University of technology, Free State, 2012Radio frequency identification technology (RFID) has emerged as a key technology for automatic identification and promises to revolutionize business processes. While RFID technology adoption is improving rapidly, reliable and widespread deployment of this technology still faces many significant challenges. The key deployment challenges include how to use the simple, unreliable raw data generated by RFID deployments to make business decisions; and how to manage a large number of deployed RFID devices. In this thesis, a multi-agent based RFID middleware which addresses some of the RFID data and device management challenges was developed. The middleware developed abstracts the auto-identification applications from physical RFID device specific details and provides necessary services such as device management, data cleaning, event generation, query capabilities and event persistence. The use of software agent technology offers a more scalable and distributed system architecture for the proposed middleware. As part of a multi-agent system, application-independent domain ontology for RFID devices was developed. This ontology can be used or extended in any application interested with RFID domain ontology. In order to address the event processing tasks within the proposed middleware system, a temporal-based RFID data model which considers both applications’ temporal and spatial granules in the data model itself for efficient event processing was developed. The developed data model extends the conventional Entity-Relationship constructs by adding a time attribute to the model. By maintaining the history of events and state changes, the data model captures the fundamental RFID application logic within the data model. Hence, this new data model supports efficient generation of application level events, updating, querying and analysis of both recent and historical events. As part of the RFID middleware, an adaptive sliding-window based data cleaning scheme for reducing missed readings from RFID data streams (called WSTD) was also developed. The WSTD scheme models the unreliability of the RFID readings by viewing RFID streams as a statistical sample of tags in the physical world, and exploits techniques grounded in sampling theory to drive its cleaning processes. The WSTD scheme is capable of efficiently coping with both environmental variations and tag dynamics by automatically and continuously adapting its cleaning window size, based on observed readings

    Leveraging Spatio-Temporal Redundancy for RFID Data Cleansing Figure 1: Spatial overlapping of detection regions

    No full text
    ABSTRACT Radio Frequency Identification (RFID) technologies are used in many applications for data collection. However, raw RFID readings are usually of low quality and may contain many anomalies. An ideal solution for RFID data cleansing should address the following issues. First, in many applications, duplicate readings (by multiple readers simultaneously or by a single reader over a period of time) of the same object are very common. The solution should take advantage of the resulting data redundancy for data cleaning. Second, prior knowledge about the readers and the environment (e.g., prior data distribution, false negative rates of readers) may help improve data quality and remove data anomalies, and a desired solution must be able to quantify the degree of uncertainty based on such knowledge. Third, the solution should take advantage of given constraints in target applications (e.g., the number of objects in a same location cannot exceed a given value) to elevate the accuracy of data cleansing. There are a number of existing RFID data cleansing techniques. However, none of them support all the aforementioned features. In this paper we propose a Bayesian inference based approach for cleaning RFID raw data. Our approach takes full advantage of data redundancy. To capture the likelihood, we design an n-state detection model and formally prove that the 3-state model can maximize the system performance. Moreover, in order to sample from the posterior, we devise a Metropolis-Hastings sampler with Constraints (MH-C), which incorporates constraint management to clean RFID raw data with high efficiency and accuracy. We validate our solution with a common RFID application and demonstrate the advantages of our approach through extensive simulations
    corecore