21,354 research outputs found
Considering Human Aspects on Strategies for Designing and Managing Distributed Human Computation
A human computation system can be viewed as a distributed system in which the
processors are humans, called workers. Such systems harness the cognitive power
of a group of workers connected to the Internet to execute relatively simple
tasks, whose solutions, once grouped, solve a problem that systems equipped
with only machines could not solve satisfactorily. Examples of such systems are
Amazon Mechanical Turk and the Zooniverse platform. A human computation
application comprises a group of tasks, each of them can be performed by one
worker. Tasks might have dependencies among each other. In this study, we
propose a theoretical framework to analyze such type of application from a
distributed systems point of view. Our framework is established on three
dimensions that represent different perspectives in which human computation
applications can be approached: quality-of-service requirements, design and
management strategies, and human aspects. By using this framework, we review
human computation in the perspective of programmers seeking to improve the
design of human computation applications and managers seeking to increase the
effectiveness of human computation infrastructures in running such
applications. In doing so, besides integrating and organizing what has been
done in this direction, we also put into perspective the fact that the human
aspects of the workers in such systems introduce new challenges in terms of,
for example, task assignment, dependency management, and fault prevention and
tolerance. We discuss how they are related to distributed systems and other
areas of knowledge.Comment: 3 figures, 1 tabl
Human Computation and Economics
This article is devoted to economical aspects of Human Computation (HC) and
to perspectives of HC in economics. As of economical aspects of HC, it is first
observed that much of what makes HC systems effective is economical in nature
suggesting that complexity being reconsidered as a âHC complexityâ and the conception
of efficient HC systems as a âHC economicsâ. This article also points to the
relevance of HC in the development of standard software and to the importance of
competition in HC systems. As of HC in economics, it is first argued that markets
can be seen as HC systems avant la lettre. Looking more closely at financial markets,
the article then points to a speed differential between transactions and credit
risk awareness that compromises the efficiency of financial markets. Finally, a HCbased
credit risk rating is proposed that, overcoming the afore mentioned speed
differential, holds promise for better functioning financial markets
Incentive Mechanisms for Participatory Sensing: Survey and Research Challenges
Participatory sensing is a powerful paradigm which takes advantage of
smartphones to collect and analyze data beyond the scale of what was previously
possible. Given that participatory sensing systems rely completely on the
users' willingness to submit up-to-date and accurate information, it is
paramount to effectively incentivize users' active and reliable participation.
In this paper, we survey existing literature on incentive mechanisms for
participatory sensing systems. In particular, we present a taxonomy of existing
incentive mechanisms for participatory sensing systems, which are subsequently
discussed in depth by comparing and contrasting different approaches. Finally,
we discuss an agenda of open research challenges in incentivizing users in
participatory sensing.Comment: Updated version, 4/25/201
Human-agent collectives
We live in a world where a host of computer systems, distributed throughout our physical and information environments, are increasingly implicated in our everyday actions. Computer technologies impact all aspects of our lives and our relationship with the digital has fundamentally altered as computers have moved out of the workplace and away from the desktop. Networked computers, tablets, phones and personal devices are now commonplace, as are an increasingly diverse set of digital devices built into the world around us. Data and information is generated at unprecedented speeds and volumes from an increasingly diverse range of sources. It is then combined in unforeseen ways, limited only by human imagination. Peopleâs activities and collaborations are becoming ever more dependent upon and intertwined with this ubiquitous information substrate. As these trends continue apace, it is becoming apparent that many endeavours involve the symbiotic interleaving of humans and computers. Moreover, the emergence of these close-knit partnerships is inducing profound change. Rather than issuing instructions to passive machines that wait until they are asked before doing anything, we will work in tandem with highly inter-connected computational components that act autonomously and intelligently (aka agents). As a consequence, greater attention needs to be given to the balance of control between people and machines. In many situations, humans will be in charge and agents will predominantly act in a supporting role. In other cases, however, the agents will be in control and humans will play the supporting role. We term this emerging class of systems human-agent collectives (HACs) to reflect the close partnership and the flexible social interactions between the humans and the computers. As well as exhibiting increased autonomy, such systems will be inherently open and social. This means the participants will need to continually and flexibly establish and manage a range of social relationships. Thus, depending on the task at hand, different constellations of people, resources, and information will need to come together, operate in a coordinated fashion, and then disband. The openness and presence of many distinct stakeholders means participation will be motivated by a broad range of incentives rather than diktat. This article outlines the key research challenges involved in developing a comprehensive understanding of HACs. To illuminate this agenda, a nascent application in the domain of disaster response is presented
Group Minds and the Case of Wikipedia
Group-level cognitive states are widely observed in human social systems, but
their discussion is often ruled out a priori in quantitative approaches. In
this paper, we show how reference to the irreducible mental states and
psychological dynamics of a group is necessary to make sense of large scale
social phenomena. We introduce the problem of mental boundaries by reference to
a classic problem in the evolution of cooperation. We then provide an explicit
quantitative example drawn from ongoing work on cooperation and conflict among
Wikipedia editors, showing how some, but not all, effects of individual
experience persist in the aggregate. We show the limitations of methodological
individualism, and the substantial benefits that come from being able to refer
to collective intentions, and attributions of cognitive states of the form
"what the group believes" and "what the group values".Comment: 21 pages, 6 figures; matches published versio
Trust in social machines: the challenges
The World Wide Web has ushered in a new generation of applications constructively linking people and computers to create what have been called âsocial machines.â The âcomponentsâ of these machines are people and technologies. It has long been recognised that for people to participate in social machines, they have to trust the processes. However, the notions of trust often used tend to be imported from agent-based computing, and may be too formal, objective and selective to describe human trust accurately. This paper applies a theory of human trust to social machines research, and sets out some of the challenges to system designers
- âŚ