19,397 research outputs found
Recommended from our members
"But you Promised": Methods to Improve Crowd Engagement In Non-Ground Truth Tasks
Crowdsourcing platforms were initially designed to recruit people to perform tasks that were simple cognitively but difficult for computers. One challenge in these settings is to identify an incentive mechanism for motivating workers to complete tasks and do high-quality work. Previous research has studied the use of financial incentive mechanisms and social comparison as motivators. These mechanisms can only be applied to ground truth tasks, tasks for which there is an objective performance scale. In this paper, we define and compare three innovative methods for improving worker engagement on non-ground truth tasks drawing on a psychological theory of commitment. The three methods are similar in asking participants to promise they will complete a task, but they differ in terms of how the commitment is made. In the first method, participants commit by signing a contract; in the second, by listening to a recording; in the third, by recording a personal commitment. The last two methods significantly improved the task completion rate when compared to two baseline conditions. The methods we propose can be implemented simply, can be used for any task, and do not affect participants' behavior other than by improving their engagement.Engineering and Applied Science
A Review on the Applications of Crowdsourcing in Human Pathology
The advent of the digital pathology has introduced new avenues of diagnostic
medicine. Among them, crowdsourcing has attracted researchers' attention in the
recent years, allowing them to engage thousands of untrained individuals in
research and diagnosis. While there exist several articles in this regard,
prior works have not collectively documented them. We, therefore, aim to review
the applications of crowdsourcing in human pathology in a semi-systematic
manner. We firstly, introduce a novel method to do a systematic search of the
literature. Utilizing this method, we, then, collect hundreds of articles and
screen them against a pre-defined set of criteria. Furthermore, we crowdsource
part of the screening process, to examine another potential application of
crowdsourcing. Finally, we review the selected articles and characterize the
prior uses of crowdsourcing in pathology
Target-oriented Proactive Dialogue Systems with Personalization: Problem Formulation and Dataset Curation
Target-oriented dialogue systems, designed to proactively steer conversations
toward predefined targets or accomplish specific system-side goals, are an
exciting area in conversational AI. In this work, by formulating a <dialogue
act, topic> pair as the conversation target, we explore a novel problem of
personalized target-oriented dialogue by considering personalization during the
target accomplishment process. However, there remains an emergent need for
high-quality datasets, and building one from scratch requires tremendous human
effort. To address this, we propose an automatic dataset curation framework
using a role-playing approach. Based on this framework, we construct a
large-scale personalized target-oriented dialogue dataset, TopDial, which
comprises about 18K multi-turn dialogues. The experimental results show that
this dataset is of high quality and could contribute to exploring personalized
target-oriented dialogue.Comment: Accepted to EMNLP-2023 main conferenc
CommuniSense: Crowdsourcing Road Hazards in Nairobi
Nairobi is one of the fastest growing metropolitan cities and a major
business and technology powerhouse in Africa. However, Nairobi currently lacks
monitoring technologies to obtain reliable data on traffic and road
infrastructure conditions. In this paper, we investigate the use of mobile
crowdsourcing as means to gather and document Nairobi's road quality
information. We first present the key findings of a city-wide road quality
survey about the perception of existing road quality conditions in Nairobi.
Based on the survey's findings, we then developed a mobile crowdsourcing
application, called CommuniSense, to collect road quality data. The application
serves as a tool for users to locate, describe, and photograph road hazards. We
tested our application through a two-week field study amongst 30 participants
to document various forms of road hazards from different areas in Nairobi. To
verify the authenticity of user-contributed reports from our field study, we
proposed to use online crowdsourcing using Amazon's Mechanical Turk (MTurk) to
verify whether submitted reports indeed depict road hazards. We found 92% of
user-submitted reports to match the MTurkers judgements. While our prototype
was designed and tested on a specific city, our methodology is applicable to
other developing cities.Comment: In Proceedings of 17th International Conference on Human-Computer
Interaction with Mobile Devices and Services (MobileHCI 2015
Toward the Optimized Crowdsourcing Strategy for OCR Post-Correction
Digitization of historical documents is a challenging task in many digital
humanities projects. A popular approach for digitization is to scan the
documents into images, and then convert images into text using Optical
Character Recognition (OCR) algorithms. However, the outcome of OCR processing
of historical documents is usually inaccurate and requires post-processing
error correction. This study investigates how crowdsourcing can be utilized to
correct OCR errors in historical text collections, and which crowdsourcing
methodology is the most effective in different scenarios and for various
research objectives. A series of experiments with different micro-task's
structures and text lengths was conducted with 753 workers on the Amazon's
Mechanical Turk platform. The workers had to fix OCR errors in a selected
historical text. To analyze the results, new accuracy and efficiency measures
have been devised. The analysis suggests that in terms of accuracy, the optimal
text length is medium (paragraph-size) and the optimal structure of the
experiment is two-phase with a scanned image. In terms of efficiency, the best
results were obtained when using longer text in the single-stage structure with
no image. The study provides practical recommendations to researchers on how to
build the optimal crowdsourcing task for OCR post-correction. The developed
methodology can also be utilized to create golden standard historical texts for
automatic OCR post-correction. This is the first attempt to systematically
investigate the influence of various factors on crowdsourcing-based OCR
post-correction and propose an optimal strategy for this process.Comment: 25 pages, 12 figures, 1 tabl
Ariel - Volume 3 Number 4
Editors
Richard J. Bonanno
Robin A. Edwards
Associate Editors
Steven Ager
Tom Williams
Lay-out Editor
Eugenia Miller
Contributing Editors
Paul Bialas
Robert Breckenridge
Lynne Porter
David Jacoby
Terry Burt
Mark Pearlman
Michael Leo
Mike LeWitt
Editors Emeritus
Delvyn C. Case, Jr.
Paul M. Fernhof
- …