38 research outputs found
On efficient and scalable time-continuous spatial crowdsourcing
The proliferation of advanced mobile terminals opened up a new crowdsourcing
avenue, spatial crowdsourcing, to utilize the crowd potential to perform
real-world tasks. In this work, we study a new type of spatial crowdsourcing,
called time-continuous spatial crowdsourcing (TCSC in short). It supports broad
applications for long-term continuous spatial data acquisition, ranging from
environmental monitoring to traffic surveillance in citizen science and
crowdsourcing projects. However, due to limited budgets and limited
availability of workers in practice, the data collected is often incomplete,
incurring data deficiency problem. To tackle that, in this work, we first
propose an entropy-based quality metric, which captures the joint effects of
incompletion in data acquisition and the imprecision in data interpolation.
Based on that, we investigate quality-aware task assignment methods for both
single- and multi-task scenarios. We show the NP-hardness of the single-task
case, and design polynomial-time algorithms with guaranteed approximation
ratios. We study novel indexing and pruning techniques for further enhancing
the performance in practice. Then, we extend the solution to multi-task
scenarios and devise a parallel framework for speeding up the process of
optimization. We conduct extensive experiments on both real and synthetic
datasets to show the effectiveness of our proposals
Incentive Mechanisms for Participatory Sensing: Survey and Research Challenges
Participatory sensing is a powerful paradigm which takes advantage of
smartphones to collect and analyze data beyond the scale of what was previously
possible. Given that participatory sensing systems rely completely on the
users' willingness to submit up-to-date and accurate information, it is
paramount to effectively incentivize users' active and reliable participation.
In this paper, we survey existing literature on incentive mechanisms for
participatory sensing systems. In particular, we present a taxonomy of existing
incentive mechanisms for participatory sensing systems, which are subsequently
discussed in depth by comparing and contrasting different approaches. Finally,
we discuss an agenda of open research challenges in incentivizing users in
participatory sensing.Comment: Updated version, 4/25/201
Mechanisms for improving information quality in smartphone crowdsensing systems
Given its potential for a large variety of real-life applications, smartphone crowdsensing has recently gained tremendous attention from the research community. Smartphone crowdsensing is a paradigm that allows ordinary citizens to participate in large-scale sensing surveys by using user-friendly applications installed in their smartphones. In this way, fine-grained sensing information is obtained from smartphone users without employing fixed and expensive infrastructure, and with negligible maintenance costs.
Existing smartphone sensing systems depend completely on the participants\u27 willingness to submit up-to-date and accurate information regarding the events being monitored. Therefore, it becomes paramount to scalably and effectively determine, enforce, and optimize the information quality of the sensing reports submitted by the participants. To this end, mechanisms to improve information quality in smartphone crowdsensing systems were designed in this work. Firstly, the FIRST framework is presented, which is a reputation-based mechanism that leverages the concept of mobile trusted participants to determine and improve the information quality of collected data. Secondly, it is mathematically modeled and studied the problem of maximizing the likelihood of successful execution of sensing tasks when participants having uncertain mobility execute sensing tasks. Two incentive mechanisms based on game and auction theory are then proposed to efficiently and scalably solve such problem. Experimental results demonstrate that the mechanisms developed in this thesis outperform existing state of the art in improving information quality in smartphone crowdsensing systems --Abstract, page iii
Halo - A Personal IoT Air Monitor Powered by Harvested Energy
Urban air pollution leads to widespread respiratory illness and millions of deaths annually. PM2.5, particulate matter with a diameter less than 2.5 micrometers, is the product of many common combustion reactions and poses a particularly serious health risk. Its small size allows it to penetrate deep into the lungs and enter the bloodstream. Existing air quality monitors are aimed at scientific research, di↵erentiating between pollutants and providing high accuracy in measurement. These devices are prohibitively expensive and cannot easily be carried around. Due to the highly localized nature of air pollution, and in order to allow individuals and institutions to easily monitor their real-time exposure to PM2.5, we propose Halo, an air quality monitor costing less than $100. Halo is powered by a 500 mW solar panel and equipped with a 1500 mAh Lithium-Ion battery in order to handle 150 mW peak power consumption and operate continuously for over 24 hours without power input. The device is small enough to be clipped to a backpack or bag for easy portability, and it can be used in personal or public settings. Using an IR emitter and detector, Halo measures reflected IR light to determine the particulate concentration in the air with an error less than 10%. It uses Bluetooth Low Energy (BLE) to communicate these values to a user’s phone. From the phone, air data can be time-stamped, stored in a cloud database, and visualized in an app for easy monitoring of pollution trends and pollution exposure. Additionally, the cloud database allows for the aggregation of data from multiple devices to create crowdsourced pollution maps. These maps can be used to pinpoint areas with particularly bad air quality in order to try to make changes to these areas or to help users to know to avoid these areas in possible
A Semi-supervised Sensing Rate Learning based CMAB Scheme to Combat COVID-19 by Trustful Data Collection in the Crowd
Mobile CrowdSensing (MCS), through employing considerable workers to sense
and collect data in a participatory manner, has been recognized as a promising
paradigm for building many large-scale applications in a cost-effective way,
such as combating COVID-19. The recruitment of trustworthy and high-quality
workers is an important research issue for MCS. Previous studies assume that
the qualities of workers are known in advance, or the platform knows the
qualities of workers once it receives their collected data. In reality, to
reduce their costs and thus maximize revenue, many strategic workers do not
perform their sensing tasks honestly and report fake data to the platform. So,
it is very hard for the platform to evaluate the authenticity of the received
data. In this paper, an incentive mechanism named Semi-supervision based
Combinatorial Multi-Armed Bandit reverse Auction (SCMABA) is proposed to solve
the recruitment problem of multiple unknown and strategic workers in MCS.
First, we model the worker recruitment as a multi-armed bandit reverse auction
problem, and design an UCB-based algorithm to separate the exploration and
exploitation, considering the Sensing Rates (SRs) of recruited workers as the
gain of the bandit. Next, a Semi-supervised Sensing Rate Learning (SSRL)
approach is proposed to quickly and accurately obtain the workers' SRs, which
consists of two phases, supervision and self-supervision. Last, SCMABA is
designed organically combining the SRs acquisition mechanism with multi-armed
bandit reverse auction, where supervised SR learning is used in the
exploration, and the self-supervised one is used in the exploitation. We prove
that our SCMABA achieves truthfulness and individual rationality. Additionally,
we exhibit outstanding performances of the SCMABA mechanism through in-depth
simulations of real-world data traces.Comment: 18 pages, 14 figure
Report on the 2015 NSF Workshop on Unified Annotation Tooling
On March 30 & 31, 2015, an international group of twenty-three researchers with expertise in linguistic annotation convened in Sunny Isles Beach, Florida to discuss problems with and potential solutions for the state of linguistic annotation tooling. The participants comprised 14 researchers from the U.S. and 9 from outside the U.S., with 7 countries and 4 continents represented, and hailed from fields and specialties including computational linguistics, artificial intelligence, speech processing, multi-modal data processing, clinical & medical natural language processing, linguistics, documentary linguistics, sign-language linguistics, corpus linguistics, and the digital humanities. The motivating problem of the workshop was the balkanization of annotation tooling, namely, that even though linguistic annotation requires sophisticated tool support to efficiently generate high-quality data, the landscape of tools for the field is fractured, incompatible, inconsistent, and lacks key capabilities. The overall goal of the workshop was to chart the way forward, centering on five key questions: (1) What are the problems with current tool landscape? (2) What are the possible benefits of solving some or all of these problems? (3) What capabilities are most needed? (4) How should we go about implementing these capabilities? And, (5) How should we ensure longevity and sustainability of the solution? I surveyed the participants before their arrival, which provided significant raw material for ideas, and the workshop discussion itself resulted in identification of ten specific classes of problems, five sets of most-needed capabilities. Importantly, we identified annotation project managers in computational linguistics as the key recipients and users of any solution, thereby succinctly addressing questions about the scope and audience of potential solutions. We discussed management and sustainability of potential solutions at length. The participants agreed on sixteen recommendations for future work. This technical report contains a detailed discussion of all these topics, a point-by-point review of the discussion in the workshop as it unfolded, detailed information on the participants and their expertise, and the summarized data from the surveys
A Survey on Mobile Crowdsensing Systems: Challenges, Solutions, and Opportunities
Mobile crowdsensing (MCS) has gained significant attention in recent years and has become an appealing paradigm for urban sensing. For data collection, MCS systems rely on contribution from mobile devices of a large number of participants or a crowd. Smartphones, tablets, and wearable devices are deployed widely and already equipped with a rich set of sensors, making them an excellent source of information. Mobility and intelligence of humans guarantee higher coverage and better context awareness if compared to traditional sensor networks. At the same time, individuals may be reluctant to share data for privacy concerns. For this reason, MCS frameworks are specifically designed to include incentive mechanisms and address privacy concerns.
Despite the growing interest in the research community, MCS solutions need a deeper investigation and categorization on many aspects that span from sensing and communication to system management and data storage. In this paper, we take the research on MCS a step further by presenting a survey on existing works in the domain and propose a detailed taxonomy to shed light on the current landscape and classify applications, methodologies, and architectures. Our objective is not only to analyze and consolidate past research but also to outline potential future research directions and synergies with other research areas
Exploring Diversity and Fairness in Machine Learning
With algorithms, artificial intelligence, and machine learning becoming ubiquitous in our society, we need to start thinking about the implications and ethical concerns of new machine learning models. In fact, two types of biases that impact machine learning models are social injustice bias (bias created by society) and measurement bias (bias created by unbalanced sampling). Biases against groups of individuals found in machine learning models can be mitigated through the use of diversity and fairness constraints. This dissertation introduces models to help humans make decisions by enforcing diversity and fairness constraints.
This work starts with a call to action. Bias is rife in hiring, and since algorithms are being used in multiple companies to filter applicants, we need to pay special attention to this application. Inspired by this hiring application, I introduce new multi-armed bandit frameworks to help assign human resources in the hiring process while enforcing diversity through a submodular utility function. These frameworks increase diversity while using less resources compared to original admission decisions of the Computer Science graduate program at the University of Maryland. Moving outside of hiring I present a contextual multi-armed bandit algorithm that enforces group fairness by learning a societal bias term and correcting for it. This algorithm is tested on two real world datasets and shows marked improvement over other in-use algorithms. Additionally I take a look at fairness in traditional machine learning domain adaptation. I provide the first theoretical analysis of this setting and test the resulting model on two deal world datasets. Finally I explore extensions to my core work, delving into suicidality, comprehension of fairness definitions, and student evaluations