100 research outputs found
Incentive mechanism design for mobile crowd sensing systems
The recent proliferation of increasingly capable and affordable mobile devices with a plethora of on-board and portable sensors that pervade every corner of the world has given rise to the fast development and wide deployment of mobile crowd sensing (MCS) systems. Nowadays, applications of MCS systems have covered almost every aspect of people's everyday living and working, such as ambient environment monitoring, healthcare, floor plan reconstruction, smart transportation, indoor localization, and many others.
Despite their tremendous benefits, MCS systems pose great new research challenges, of which, this thesis targets one important facet, that is, to effectively incentivize (crowd) workers to achieve maximum participation in MCS systems. Participating in crowd sensing tasks is usually a costly procedure for individual workers. On one hand, it consumes workers' resources, such as computing power, battery, and so forth. On the other hand, a considerable portion of sensing tasks require the submission of workers' sensitive and private information, which causes privacy leakage for participants. Clearly, the power of crowd sensing could not be fully unleashed, unless workers are properly incentivized to participate via satisfactory rewards that effectively compensate their participation costs.
Targeting the above challenge, in this thesis, I present a series of novel incentive mechanisms, which can be utilized to effectively incentivize worker participation in MCS systems. The proposed mechanisms not only incorporate workers' quality of information in order to selectively recruit relatively more reliable workers for sensing, but also preserve workers' privacy so as to prevent workers from being disincentivized by excessive privacy leakage. I demonstrate through rigorous theoretical analyses and extensive simulations that the proposed incentive mechanisms bear many desirable properties theoretically, and have great potential to be practically applied
Anonymizing and Trading Person-specific Data with Trust
In the past decade, data privacy, security, and trustworthiness have gained tremendous attention from research communities, and these are still active areas of research with the proliferation of cloud services and social media applications. The data is growing at a rapid pace. It has become an integral part of almost every industry and business, including commercial and non-profit organizations. It often contains person-specific information and a data custodian who holds it must be responsible for managing its use, disclosure, accuracy and privacy protection. In this thesis, we present three research problems. The first two problems address the concerns of stakeholders on privacy protection, data trustworthiness, and profit distribution in the online market for trading person-specific data. The third problem addresses the health information custodians (HICs) concern on privacy-preserving healthcare network data publishing.
Our first research problem is identified in cloud-based data integration service where data providers collaborate with their trading partners in order to deliver quality data mining services. Data-as-a-Service (DaaS) enables data integration to serve the demands of data consumers. Data providers face challenges not only to protect private data over the cloud but also to legally adhere to privacy compliance rules when trading person-specific data. We propose a model that allows the collaboration of multiple data providers for integrating their data and derives the contribution of each data provider by valuating the incorporated cost factors. This model serves as a guide for business decision-making, such as estimating the potential privacy risk and finding the sub-optimal value for publishing mashup data. Experiments on real-life data demonstrate that our approach can identify the sub-optimal value in data mashup for different privacy models, including K-anonymity, LKC-privacy, and ϵ-differential privacy, with various anonymization algorithms and privacy parameters.
Second, consumers demand a good quality of data for accurate analysis and effective decision- making while the data providers intend to maximize their profits by competing with peer providers. In addition, the data providers or custodians must conform to privacy policies to avoid potential penalties for privacy breaches. To address these challenges, we propose a two-fold solution: (1) we present the first information entropy-based trust computation algorithm, IEB_Trust, that allows a semi-trusted arbitrator to detect the covert behavior of a dishonest data provider and chooses the qualified providers for a data mashup, and (2) we incorporate the Vickrey-Clarke-Groves (VCG) auction mechanism for the valuation of data providers’ attributes into the data mashup process. Experiments on real-life data demonstrate the robustness of our approach in restricting dishonest providers from participation in the data mashup and improving the efficiency in comparison to provenance-based approaches. Furthermore, we derive the monetary shares for the chosen providers from their information utility and trust scores over the differentially private release of the integrated dataset under their joint privacy requirements.
Finally, we address the concerns of HICs of exchanging healthcare data to provide better and more timely services while mitigating the risk of exposing patients’ sensitive information to privacy threats. We first model a complex healthcare dataset using a heterogeneous information network that consists of multi-type entities and their relationships. We then propose DiffHetNet, an edge-based differentially private algorithm, to protect the sensitive links of patients from inbound and outbound attacks in the heterogeneous health network. We evaluate the performance of our proposed method in terms of information utility and efficiency on different types of real-life datasets that can be modeled as networks. Experimental results suggest that DiffHetNet generally yields less information loss and is significantly more efficient in terms of runtime in comparison with existing network anonymization methods. Furthermore, DiffHetNet is scalable to large network datasets
Ethical research in public policy.
Public policy research is research for a purpose, guided by a distinctive range of normative considerations. The values are the values of public service; the work is generally done in the public domain; and the research is an intrinsic part of the democratic process, which depends on deliberation and accountability. Conventional representations of ethical research typically focus on ‘human subjects’ research, which raises different kinds of ethical issues to public policy research. Existing research ethics advice does not address the issues surrounding public policy research. Such research is typically concerned with collective action and the work of institutions, and the central guiding principles are not about responsibility to research participants, but duties to the public, as seen in principles of beneficence, citizenship, empowerment and the democratic process
Ethical Evidence and Policymaking
EPDF and EPUB available Open Access under CC-BY-NC-ND licence.
This important book offers practical advice for using evidence and research in policymaking. The book has two aims. First, it builds a case for ethics and global values in research and knowledge exchange, and second, it examines specific policy areas and how evidence can guide practice.
The book covers important policy areas including the GM debate, the environment, Black Lives Matter and COVID-19. Each chapter assesses the ethical challenges, the status of evidence in explaining or describing the issue and possible solutions to the problem. The book will enable policymakers and their advisors to seek evidence for their decisions from research that has been conducted ethically and with integrity
Process Mining Workshops
This open access book constitutes revised selected papers from the International Workshops held at the Third International Conference on Process Mining, ICPM 2021, which took place in Eindhoven, The Netherlands, during October 31–November 4, 2021. The conference focuses on the area of process mining research and practice, including theory, algorithmic challenges, and applications. The co-located workshops provided a forum for novel research ideas. The 28 papers included in this volume were carefully reviewed and selected from 65 submissions. They stem from the following workshops: 2nd International Workshop on Event Data and Behavioral Analytics (EDBA) 2nd International Workshop on Leveraging Machine Learning in Process Mining (ML4PM) 2nd International Workshop on Streaming Analytics for Process Mining (SA4PM) 6th International Workshop on Process Querying, Manipulation, and Intelligence (PQMI) 4th International Workshop on Process-Oriented Data Science for Healthcare (PODS4H) 2nd International Workshop on Trust, Privacy, and Security in Process Analytics (TPSA) One survey paper on the results of the XES 2.0 Workshop is included
LIPIcs, Volume 251, ITCS 2023, Complete Volume
LIPIcs, Volume 251, ITCS 2023, Complete Volum
Process Mining Workshops
This open access book constitutes revised selected papers from the International Workshops held at the Third International Conference on Process Mining, ICPM 2021, which took place in Eindhoven, The Netherlands, during October 31–November 4, 2021. The conference focuses on the area of process mining research and practice, including theory, algorithmic challenges, and applications. The co-located workshops provided a forum for novel research ideas. The 28 papers included in this volume were carefully reviewed and selected from 65 submissions. They stem from the following workshops: 2nd International Workshop on Event Data and Behavioral Analytics (EDBA) 2nd International Workshop on Leveraging Machine Learning in Process Mining (ML4PM) 2nd International Workshop on Streaming Analytics for Process Mining (SA4PM) 6th International Workshop on Process Querying, Manipulation, and Intelligence (PQMI) 4th International Workshop on Process-Oriented Data Science for Healthcare (PODS4H) 2nd International Workshop on Trust, Privacy, and Security in Process Analytics (TPSA) One survey paper on the results of the XES 2.0 Workshop is included
Generalized asset integrity games
Generalized assets represent a class of multi-scale adaptive state-transition systems with domain-oblivious performance criteria. The governance of such assets must proceed without exact specifications, objectives, or constraints. Decision making must rapidly scale in the presence of uncertainty, complexity, and intelligent adversaries.
This thesis formulates an architecture for generalized asset planning. Assets are modelled as dynamical graph structures which admit topological performance indicators, such as dependability, resilience, and efficiency. These metrics are used to construct robust model configurations. A normalized compression distance (NCD) is computed between a given active/live asset model and a reference configuration to produce an integrity score. The utility derived from the asset is monotonically proportional to this integrity score, which represents the proximity to ideal conditions. The present work considers the situation between an asset manager and an intelligent adversary, who act within a stochastic environment to control the integrity state of the asset. A generalized asset integrity game engine (GAIGE) is developed, which implements anytime algorithms to solve a stochastically perturbed two-player zero-sum game. The resulting planning strategies seek to stabilize deviations from minimax trajectories of the integrity score.
Results demonstrate the performance and scalability of the GAIGE. This approach represents a first-step towards domain-oblivious architectures for complex asset governance and anytime planning
Modeling, Quantifying, and Limiting Adversary Knowledge
Users participating in online services are required to relinquish
control over potentially sensitive personal information, exposing
them to intentional or unintentional miss-use of said information by
the service providers.
Users wishing to avoid this must either abstain from often extremely
useful services, or provide false information which is usually
contrary to the terms of service they must abide by.
An attractive middle-ground alternative is to maintain control in
the hands of the users and provide a mechanism with which
information that is necessary for useful services can be queried.
Users need not trust any external party in the management of their
information but are now faced with the problem of judging when
queries by service providers should be answered or when they should
be refused due to revealing too much sensitive information.
Judging query safety is difficult.
Two queries may be benign in isolation but might reveal more than a
user is comfortable with in combination.
Additionally malicious adversaries who wish to learn more than
allowed might query in a manner that attempts to hide the flows of
sensitive information.
Finally, users cannot rely on human inspection of queries due to its
volume and the general lack of expertise.
This thesis tackles the automation of query judgment, giving the
self-reliant user a means with which to discern benign queries from
dangerous or exploitive ones.
The approach is based on explicit modeling and tracking of the
knowledge of adversaries as they learn about a user through the
queries they are allowed to observe.
The approach quantifies the absolute risk a user is exposed, taking
into account all the information that has been revealed already when
determining to answer a query.
Proposed techniques for approximate but sound probabilistic
inference are used to tackle the tractability of the approach,
letting the user tradeoff utility (in terms of the queries judged
safe) and efficiency (in terms of the expense of knowledge
tracking), while maintaining the guarantee that risk to the user is
never underestimated.
We apply the approach to settings where user data changes over time
and settings where multiple users wish to pool their data to perform
useful collaborative computations without revealing too much
information.
By addressing one of the major obstacles preventing the viability of
personal information control, this work brings the attractive
proposition closer to reality
- …