25,994 research outputs found
Event detection in location-based social networks
With the advent of social networks and the rise of mobile technologies, users have become ubiquitous sensors capable of monitoring various real-world events in a crowd-sourced manner. Location-based social networks have proven to be faster than traditional media channels in reporting and geo-locating breaking news, i.e. Osama Bin Laden’s death was first confirmed on Twitter even before the announcement from the communication department at the White House. However, the deluge of user-generated data on these networks requires intelligent systems capable of identifying and characterizing such events in a comprehensive manner. The data mining community coined the term, event detection , to refer to the task of uncovering emerging patterns in data streams . Nonetheless, most data mining techniques do not reproduce the underlying data generation process, hampering to self-adapt in fast-changing scenarios. Because of this, we propose a probabilistic machine learning approach to event detection which explicitly models the data generation process and enables reasoning about the discovered events. With the aim to set forth the differences between both approaches, we present two techniques for the problem of event detection in Twitter : a data mining technique called Tweet-SCAN and a machine learning technique called Warble. We assess and compare both techniques in a dataset of tweets geo-located in the city of Barcelona during its annual festivities. Last but not least, we present the algorithmic changes and data processing frameworks to scale up the proposed techniques to big data workloads.This work is partially supported by Obra Social “la Caixa”, by the Spanish Ministry of Science and Innovation under contract (TIN2015-65316), by the Severo Ochoa Program (SEV2015-0493), by SGR programs of the Catalan Government (2014-SGR-1051, 2014-SGR-118), Collectiveware (TIN2015-66863-C2-1-R) and BSC/UPC NVIDIA GPU Center of Excellence.We would also like to thank the reviewers for their constructive feedback.Peer ReviewedPostprint (author's final draft
A Novel Framework for Online Amnesic Trajectory Compression in Resource-constrained Environments
State-of-the-art trajectory compression methods usually involve high
space-time complexity or yield unsatisfactory compression rates, leading to
rapid exhaustion of memory, computation, storage and energy resources. Their
ability is commonly limited when operating in a resource-constrained
environment especially when the data volume (even when compressed) far exceeds
the storage limit. Hence we propose a novel online framework for error-bounded
trajectory compression and ageing called the Amnesic Bounded Quadrant System
(ABQS), whose core is the Bounded Quadrant System (BQS) algorithm family that
includes a normal version (BQS), Fast version (FBQS), and a Progressive version
(PBQS). ABQS intelligently manages a given storage and compresses the
trajectories with different error tolerances subject to their ages. In the
experiments, we conduct comprehensive evaluations for the BQS algorithm family
and the ABQS framework. Using empirical GPS traces from flying foxes and cars,
and synthetic data from simulation, we demonstrate the effectiveness of the
standalone BQS algorithms in significantly reducing the time and space
complexity of trajectory compression, while greatly improving the compression
rates of the state-of-the-art algorithms (up to 45%). We also show that the
operational time of the target resource-constrained hardware platform can be
prolonged by up to 41%. We then verify that with ABQS, given data volumes that
are far greater than storage space, ABQS is able to achieve 15 to 400 times
smaller errors than the baselines. We also show that the algorithm is robust to
extreme trajectory shapes.Comment: arXiv admin note: substantial text overlap with arXiv:1412.032
When Things Matter: A Data-Centric View of the Internet of Things
With the recent advances in radio-frequency identification (RFID), low-cost
wireless sensor devices, and Web technologies, the Internet of Things (IoT)
approach has gained momentum in connecting everyday objects to the Internet and
facilitating machine-to-human and machine-to-machine communication with the
physical world. While IoT offers the capability to connect and integrate both
digital and physical entities, enabling a whole new class of applications and
services, several significant challenges need to be addressed before these
applications and services can be fully realized. A fundamental challenge
centers around managing IoT data, typically produced in dynamic and volatile
environments, which is not only extremely large in scale and volume, but also
noisy, and continuous. This article surveys the main techniques and
state-of-the-art research efforts in IoT from data-centric perspectives,
including data stream processing, data storage models, complex event
processing, and searching in IoT. Open research issues for IoT data management
are also discussed
Spatial Data Quality in the IoT Era:Management and Exploitation
Within the rapidly expanding Internet of Things (IoT), growing amounts of spatially referenced data are being generated. Due to the dynamic, decentralized, and heterogeneous nature of the IoT, spatial IoT data (SID) quality has attracted considerable attention in academia and industry. How to invent and use technologies for managing spatial data quality and exploiting low-quality spatial data are key challenges in the IoT. In this tutorial, we highlight the SID consumption requirements in applications and offer an overview of spatial data quality in the IoT setting. In addition, we review pertinent technologies for quality management and low-quality data exploitation, and we identify trends and future directions for quality-aware SID management and utilization. The tutorial aims to not only help researchers and practitioners to better comprehend SID quality challenges and solutions, but also offer insights that may enable innovative research and applications
Exploiting Deep Features for Remote Sensing Image Retrieval: A Systematic Investigation
Remote sensing (RS) image retrieval is of great significant for geological
information mining. Over the past two decades, a large amount of research on
this task has been carried out, which mainly focuses on the following three
core issues: feature extraction, similarity metric and relevance feedback. Due
to the complexity and multiformity of ground objects in high-resolution remote
sensing (HRRS) images, there is still room for improvement in the current
retrieval approaches. In this paper, we analyze the three core issues of RS
image retrieval and provide a comprehensive review on existing methods.
Furthermore, for the goal to advance the state-of-the-art in HRRS image
retrieval, we focus on the feature extraction issue and delve how to use
powerful deep representations to address this task. We conduct systematic
investigation on evaluating correlative factors that may affect the performance
of deep features. By optimizing each factor, we acquire remarkable retrieval
results on publicly available HRRS datasets. Finally, we explain the
experimental phenomenon in detail and draw conclusions according to our
analysis. Our work can serve as a guiding role for the research of
content-based RS image retrieval
Recommended from our members
A continuously updated, geospatially rectified database of utility-scale wind turbines in the United States.
Over 60,000 utility-scale wind turbines are installed in the United States as of October, 2019, representing over 97 gigawatts of electric power capacity; US wind turbine installations continue to grow at a rapid pace. Yet, until April 2018, no publicly-available, regularly updated data source existed to describe those turbines and their locations. Under a cooperative research and development agreement, analysts from three organizations collaborated to develop and release the United States Wind Turbine Database (USWTDB) - a publicly available, continuously updated, spatially rectified data source of locations and attributes of utility-scale wind turbines in the United States. Technical specifications and wind facility data, incorporated from five sources, undergo rigorous quality control. The location of each turbine is visually verified using high-resolution aerial imagery. The quarterly-updated data are available in a variety of formats, including an interactive web application, comma-separated values (CSV), shapefile, and application programming interface (API). The data are used widely by academic researchers, engineers and developers from wind energy companies, government agencies, planners, educators, and the general public
- …