971 research outputs found
Towards a Theory of Systems Engineering Processes: A Principal-Agent Model of a One-Shot, Shallow Process
Systems engineering processes coordinate the effort of different individuals
to generate a product satisfying certain requirements. As the involved
engineers are self-interested agents, the goals at different levels of the
systems engineering hierarchy may deviate from the system-level goals which may
cause budget and schedule overruns. Therefore, there is a need of a systems
engineering theory that accounts for the human behavior in systems design. To
this end, the objective of this paper is to develop and analyze a
principal-agent model of a one-shot (single iteration), shallow (one level of
hierarchy) systems engineering process. We assume that the systems engineer
maximizes the expected utility of the system, while the subsystem engineers
seek to maximize their expected utilities. Furthermore, the systems engineer is
unable to monitor the effort of the subsystem engineer and may not have a
complete information about their types or the complexity of the design task.
However, the systems engineer can incentivize the subsystem engineers by
proposing specific contracts. To obtain an optimal incentive, we pose and solve
numerically a bi-level optimization problem. Through extensive simulations, we
study the optimal incentives arising from different system-level value
functions under various combinations of effort costs, problem-solving skills,
and task complexities
Optimum Statistical Estimation with Strategic Data Sources
We propose an optimum mechanism for providing monetary incentives to the data
sources of a statistical estimator such as linear regression, so that high
quality data is provided at low cost, in the sense that the sum of payments and
estimation error is minimized. The mechanism applies to a broad range of
estimators, including linear and polynomial regression, kernel regression, and,
under some additional assumptions, ridge regression. It also generalizes to
several objectives, including minimizing estimation error subject to budget
constraints. Besides our concrete results for regression problems, we
contribute a mechanism design framework through which to design and analyze
statistical estimators whose examples are supplied by workers with cost for
labeling said examples
A Framework for Quality Assurance in Crowdsourcing
The emergence of online paid micro-crowdsourcing platforms, such as Amazon Mechanical Turk (AMT), allows on-demand and at scale distribution of tasks to human workers around the world. In such settings, online workers come and complete small tasks posted by a company, working for as long or as little as they wish. Such temporary employer-employee relationships give rise to adverse selection, moral hazard, and many other challenges. How can we ensure that the submitted work is accurate, especially when the verification cost is comparable to the cost of performing the task? How can we estimate the exhibited quality of the workers? What pricing strategies should be used to induce the effort of workers with varying ability levels? We develop a comprehensive framework for managing the quality in such micro crowdsourcing settings: First, we describe an algorithm for estimating the error rates of the participating workers, and show how to separate systematic worker biases from unrecoverable errors and generate an unbiased “worker quality” measurement. Next, we present a selective repeated-labeling algorithm that acquires labels in a way so that quality requirements can be met at minimum cost. Then, we propose a quality-adjusted pricing scheme that adjusts the payment level according to the contributed value by each worker. We test our compensation scheme in a principal-agent setting in which workers respond to incentives by varying their effort. Our simulation results demonstrate that the proposed pricing scheme is able to induce workers to exert higher levels of effort and yield larger profits for employers compared to the commonly adopted uniform pricing schemes. We also describe strategies that build on our quality control and pricing framework, to tackle crowdsourced tasks of increasingly higher complexity, while still maintaining a tight quality control of the process
Revolutionizing Crowdworking Campaigns: Conquering Adverse Selection and Moral Hazard with the Help of Smart Contracts
Crowdworking is increasingly being applied by companies to outsource tasks beyond their core competencies flexibly and cost-effectively to an unknown group. However, the anonymous and financially incentivized nature of crowdworkers creates information asymmetries and conflicts of interest, leading to inefficiencies and intensifying the principal-agent problem. Our paper offers a solution to the widespread problem of inefficient Crowdworking campaigns. We first derive the currently applied Crowdworking campaign process based on a qualitative study. Subsequently, we identify the broadest adverse selection and moral hazard problems in the process. We then analyze how the blockchain application of smart contracts can counteract those challenges and develop a process model that maps a Crowdworking campaign using smart contracts. We explain how our developed process significantly reduces adverse selection and moral hazard at each stage. Thus, our research provides approaches to make online labor more attractive and transparent for companies and online workers
Incentive Mechanisms for Participatory Sensing: Survey and Research Challenges
Participatory sensing is a powerful paradigm which takes advantage of
smartphones to collect and analyze data beyond the scale of what was previously
possible. Given that participatory sensing systems rely completely on the
users' willingness to submit up-to-date and accurate information, it is
paramount to effectively incentivize users' active and reliable participation.
In this paper, we survey existing literature on incentive mechanisms for
participatory sensing systems. In particular, we present a taxonomy of existing
incentive mechanisms for participatory sensing systems, which are subsequently
discussed in depth by comparing and contrasting different approaches. Finally,
we discuss an agenda of open research challenges in incentivizing users in
participatory sensing.Comment: Updated version, 4/25/201
Information Gathering with Peers: Submodular Optimization with Peer-Prediction Constraints
We study a problem of optimal information gathering from multiple data
providers that need to be incentivized to provide accurate information. This
problem arises in many real world applications that rely on crowdsourced data
sets, but where the process of obtaining data is costly. A notable example of
such a scenario is crowd sensing. To this end, we formulate the problem of
optimal information gathering as maximization of a submodular function under a
budget constraint, where the budget represents the total expected payment to
data providers. Contrary to the existing approaches, we base our payments on
incentives for accuracy and truthfulness, in particular, {\em peer-prediction}
methods that score each of the selected data providers against its best peer,
while ensuring that the minimum expected payment is above a given threshold. We
first show that the problem at hand is hard to approximate within a constant
factor that is not dependent on the properties of the payment function.
However, for given topological and analytical properties of the instance, we
construct two greedy algorithms, respectively called PPCGreedy and
PPCGreedyIter, and establish theoretical bounds on their performance w.r.t. the
optimal solution. Finally, we evaluate our methods using a realistic crowd
sensing testbed.Comment: Longer version of AAAI'18 pape
- …