1,137 research outputs found

    Location Privacy in Spatial Crowdsourcing

    Full text link
    Spatial crowdsourcing (SC) is a new platform that engages individuals in collecting and analyzing environmental, social and other spatiotemporal information. With SC, requesters outsource their spatiotemporal tasks to a set of workers, who will perform the tasks by physically traveling to the tasks' locations. This chapter identifies privacy threats toward both workers and requesters during the two main phases of spatial crowdsourcing, tasking and reporting. Tasking is the process of identifying which tasks should be assigned to which workers. This process is handled by a spatial crowdsourcing server (SC-server). The latter phase is reporting, in which workers travel to the tasks' locations, complete the tasks and upload their reports to the SC-server. The challenge is to enable effective and efficient tasking as well as reporting in SC without disclosing the actual locations of workers (at least until they agree to perform a task) and the tasks themselves (at least to workers who are not assigned to those tasks). This chapter aims to provide an overview of the state-of-the-art in protecting users' location privacy in spatial crowdsourcing. We provide a comparative study of a diverse set of solutions in terms of task publishing modes (push vs. pull), problem focuses (tasking and reporting), threats (server, requester and worker), and underlying technical approaches (from pseudonymity, cloaking, and perturbation to exchange-based and encryption-based techniques). The strengths and drawbacks of the techniques are highlighted, leading to a discussion of open problems and future work

    Answering a calling: medical professionals' digital careers in crowdsourcing

    Full text link
    One of the most striking trends in individuals’ careers over the last decade has been the dramatic increase in the proportion of the labor force working beyond their employers’ physical boundaries because of the digital revolution in the gig economy. This trend has drawn much attention in the changing nature of work, workplace and careers. However, little empirical research has explored how and why individuals behave in the interface between online platforms and traditional organizations. In my dissertation, I explore these questions by studying medical professionals’ digital careers in the Chinese healthcare crowdsourcing industry, also known as “mobile doctors.” First, by analyzing approximately 240-hour observations and 43 interviews with Chinese physicians, I identify a key issue in this new career – time conflict between crowdsourcing and traditional work. The findings show that physicians respond to time conflict in a variety of ways, including time theft, an essential yet under-researched construct in the crowdsourcing literature which reflects the tension between traditional work and crowdsourcing. Second, by analyzing archival data of 4,034 doctors’ 3.1 million time records on a Chinese healthcare platform across half a year, I show that time theft for crowdsourcing is related to the traditional work context, including hospitals’ boundary control and offline crowd worker social groups. Finally, I further explore, via interview data, why such seemingly costly and deviant time theft is adopted by mobile doctors. The findings reveal that medical professionals assume the extra burden of working for crowdsourcing with the hope of answering unfulfilled occupational callings in traditional work and adding meaning to their work. Overall, these findings contribute to a better understanding of the shifting nature of work and careers in the digital economy by documenting and explaining mobile doctors’ participation in this new world of work

    Accurate and budget-efficient text, image, and video analysis systems powered by the crowd

    Full text link
    Crowdsourcing systems empower individuals and companies to outsource labor-intensive tasks that cannot currently be solved by automated methods and are expensive to tackle by domain experts. Crowdsourcing platforms are traditionally used to provide training labels for supervised machine learning algorithms. Crowdsourced tasks are distributed among internet workers who typically have a range of skills and knowledge, differing previous exposure to the task at hand, and biases that may influence their work. This inhomogeneity of the workforce makes the design of accurate and efficient crowdsourcing systems challenging. This dissertation presents solutions to improve existing crowdsourcing systems in terms of accuracy and efficiency. It explores crowdsourcing tasks in two application areas, political discourse and annotation of biomedical and everyday images. The first part of the dissertation investigates how workers' behavioral factors and their unfamiliarity with data can be leveraged by crowdsourcing systems to control quality. Through studies that involve familiar and unfamiliar image content, the thesis demonstrates the benefit of explicitly accounting for a worker's familiarity with the data when designing annotation systems powered by the crowd. The thesis next presents Crowd-O-Meter, a system that automatically predicts the vulnerability of crowd workers to believe \enquote{fake news} in text and video. The second part of the dissertation explores the reversed relationship between machine learning and crowdsourcing by incorporating machine learning techniques for quality control of crowdsourced end products. In particular, it investigates if machine learning can be used to improve the quality of crowdsourced results and also consider budget constraints. The thesis proposes an image analysis system called ICORD that utilizes behavioral cues of the crowd worker, augmented by automated evaluation of image features, to infer the quality of a worker-drawn outline of a cell in a microscope image dynamically. ICORD determines the need to seek additional annotations from other workers in a budget-efficient manner. Next, the thesis proposes a budget-efficient machine learning system that uses fewer workers to analyze easy-to-label data and more workers for data that require extra scrutiny. The system learns a mapping from data features to number of allocated crowd workers for two case studies, sentiment analysis of twitter messages and segmentation of biomedical images. Finally, the thesis uncovers the potential for design of hybrid crowd-algorithm methods by describing an interactive system for cell tracking in time-lapse microscopy videos, based on a prediction model that determines when automated cell tracking algorithms fail and human interaction is needed to ensure accurate tracking

    Enhancing Automation and Interoperability in Enterprise Crowdsourcing Environments

    Get PDF
    The last couple of years have seen a fascinating evolution. While the early Web predominantly focused on human consumption of Web content, the widespread dissemination of social software and Web 2.0 technologies enabled new forms of collaborative content creation and problem solving. These new forms often utilize the principles of collective intelligence, a phenomenon that emerges from a group of people who either cooperate or compete with each other to create a result that is better or more intelligent than any individual result (Leimeister, 2010; Malone, Laubacher, & Dellarocas, 2010). Crowdsourcing has recently gained attention as one of the mechanisms that taps into the power of web-enabled collective intelligence (Howe, 2008). Brabham (2013) defines it as “an online, distributed problem-solving and production model that leverages the collective intelligence of online communities to serve specific organizational goals” (p. xix). Well-known examples of crowdsourcing platforms are Wikipedia, Amazon Mechanical Turk, or InnoCentive. Since the emergence of the term crowdsourcing in 2006, one popular misconception is that crowdsourcing relies largely on an amateur crowd rather than a pool of professional skilled workers (Brabham, 2013). As this might be true for low cognitive tasks, such as tagging a picture or rating a product, it is often not true for complex problem-solving and creative tasks, such as developing a new computer algorithm or creating an impressive product design. This raises the question of how to efficiently allocate an enterprise crowdsourcing task to appropriate members of the crowd. The sheer number of crowdsourcing tasks available at crowdsourcing intermediaries makes it especially challenging for workers to identify a task that matches their skills, experiences, and knowledge (Schall, 2012, p. 2). An explanation why the identification of appropriate expert knowledge plays a major role in crowdsourcing is partly given in Condorcet’s jury theorem (Sunstein, 2008, p. 25). The theorem states that if the average participant in a binary decision process is more likely to be correct than incorrect, then as the number of participants increases, the higher the probability is that the aggregate arrives at the right answer. When assuming that a suitable participant for a task is more likely to give a correct answer or solution than an improper one, efficient task recommendation becomes crucial to improve the aggregated results in crowdsourcing processes. Although some assumptions of the theorem, such as independent votes, binary decisions, and homogenous groups, are often unrealistic in practice, it illustrates the importance of an optimized task allocation and group formation that consider the task requirements and workers’ characteristics. Ontologies are widely applied to support semantic search and recommendation mechanisms (Middleton, De Roure, & Shadbolt, 2009). However, little research has investigated the potentials and the design of an ontology for the domain of enterprise crowdsourcing. The author of this thesis argues in favor of enhancing the automation and interoperability of an enterprise crowdsourcing environment with the introduction of a semantic vocabulary in form of an expressive but easy-to-use ontology. The deployment of a semantic vocabulary for enterprise crowdsourcing is likely to provide several technical and economic benefits for an enterprise. These benefits were the main drivers in efforts made during the research project of this thesis: 1. Task allocation: With the utilization of the semantics, requesters are able to form smaller task-specific crowds that perform tasks at lower costs and in less time than larger crowds. A standardized and controlled vocabulary allows requesters to communicate specific details about a crowdsourcing activity within a web page along with other existing displayed information. This has advantages for both contributors and requesters. On the one hand, contributors can easily and precisely search for tasks that correspond to their interests, experiences, skills, knowledge, and availability. On the other hand, crowdsourcing systems and intermediaries can proactively recommend crowdsourcing tasks to potential contributors (e.g., based on their social network profiles). 2. Quality control: Capturing and storing crowdsourcing data increases the overall transparency of the entire crowdsourcing activity and thus allows for a more sophisticated quality control. Requesters are able to check the consistency and receive appropriate support to verify and validate crowdsourcing data according to defined data types and value ranges. Before involving potential workers in a crowdsourcing task, requesters can also judge their trustworthiness based on previous accomplished tasks and hence improve the recruitment process. 3. Task definition: A standardized set of semantic entities supports the configuration of a crowdsourcing task. Requesters can evaluate historical crowdsourcing data to get suggestions for equal or similar crowdsourcing tasks, for example, which incentive or evaluation mechanism to use. They may also decrease their time to configure a crowdsourcing task by reusing well-established task specifications of a particular type. 4. Data integration and exchange: Applying a semantic vocabulary as a standard format for describing enterprise crowdsourcing activities allows not only crowdsourcing systems inside but also crowdsourcing intermediaries outside the company to extract crowdsourcing data from other business applications, such as project management, enterprise resource planning, or social software, and use it for further processing without retyping and copying the data. Additionally, enterprise or web search engines may exploit the structured data and provide enhanced search, browsing, and navigation capabilities, for example, clustering similar crowdsourcing tasks according to the required qualifications or the offered incentives.:Summary: Hetmank, L. (2014). Enhancing Automation and Interoperability in Enterprise Crowdsourcing Environments (Summary). Article 1: Hetmank, L. (2013). Components and Functions of Crowdsourcing Systems – A Systematic Literature Review. In 11th International Conference on Wirtschaftsinformatik (WI). Leipzig. Article 2: Hetmank, L. (2014). A Synopsis of Enterprise Crowdsourcing Literature. In 22nd European Conference on Information Systems (ECIS). Tel Aviv. Article 3: Hetmank, L. (2013). Towards a Semantic Standard for Enterprise Crowdsourcing – A Scenario-based Evaluation of a Conceptual Prototype. In 21st European Conference on Information Systems (ECIS). Utrecht. Article 4: Hetmank, L. (2014). Developing an Ontology for Enterprise Crowdsourcing. In Multikonferenz Wirtschaftsinformatik (MKWI). Paderborn. Article 5: Hetmank, L. (2014). An Ontology for Enhancing Automation and Interoperability in Enterprise Crowdsourcing Environments (Technical Report). Retrieved from http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-155187

    Towards Computational Assessment of Idea Novelty

    Get PDF
    In crowdsourcing ideation websites, companies can easily collect large amount of ideas. Screening through such volume of ideas is very costly and challenging, necessitating automatic approaches. It would be particularly useful to automatically evaluate idea novelty since companies commonly seek novel ideas. Three computational approaches were tested, based on Latent Semantic Analysis (LSA), Latent Dirichlet Allocation (LDA) and term frequency–inverse document frequency (TF-IDF), respectively. These three approaches were used on three set of ideas and the computed idea novelty was compared with human expert evaluation. TF-IDF based measure correlated better with expert evaluation than the other two measures. However, our results show that these approaches do not match human judgement well enough to replace it

    Proximity as a Service via Cellular Network-Assisted Mobile Device-to-Device

    Get PDF
    PhD ThesisThe research progress of communication has brought a lot of novel technologies to meet the multi-dimensional demands such as pervasive connection, low delay and high bandwidth. Device-to-Device (D2D) communication is a way to no longer treat the User Equipment (UEs) as a terminal, but rather as a part of the network for service provisioning. This thesis decouples UEs into service providers (helpers) and service requesters. By collaboration among proximal devices, with the coordination of cellular networks, some local tasks can be achieved, such as coverage extension, computation o oading, mobile crowdsourcing and mobile crowdsensing. This thesis proposes a generic framework Proximity as a Service (PaaS) for increasing the coverage with demands of service continuity. As one of the use cases, the optimal helper selection algorithm of PaaS for increasing the service coverage with demands of service continuity is called ContAct based Proximity (CAP). Mainly, fruitful contact information (e.g., contact duration, frequency, and interval) is captured, and is used to handle ubiquitous proximal services through the optimal selection of helpers. The nature of PaaS is evaluated under the Helsinki city scenario, with movement model of Points Of Interest (POI) and with critical factors in uencing the service demands (e.g., success ratio, disruption duration and frequency). Simulation results show the advantage of CAP, in both success ratio and continuity of the service (outputs). Based on this perspective, metrics such as service success ratio and continuity as a service evaluation of the PaaS are evaluated using the statistical theory of the Design Of Experiments (DOE). DOE is used as there are many dimensions to the state space (access tolerance, selected helper number, helper access limit, and transmit range) that can in uence the results. A key contribution of this work is that it brings rigorous statistical experiment design methods into the research into mobile computing. Results further reveal the influence of four factors (inputs), e.g., service tolerance, number of helpers allocated, the number of concurrent devices supported by each helper and transmit range. Based on this perspective, metrics such as service success ratio and continuity are evaluated using DOE. The results show that transmit range is the most dominant factor. The number of selected helpers is the second most dominant factor. Since di erent factors have di erent regression levels, a uni ed 4 level full factorial experiment and a cubic multiple regression analysis have been carried out. All the interactions and the corresponding coe cients have been found. This work is the rst one to evaluate LTE-Direct and WiFi-Direct in an opportunistic proximity service. The contribution of the results for industry is to guide how many users need to cooperate to enable mobile computing and for academia. This reveals the facts that: 1, in some cases, the improvement of spectrum e ciency brought by D2D is not important; 2, nodal density and the resources used in D2D air-interfaces are important in the eld of mobile computing. This work built a methodology to study the D2D networks with a di erent perspective (PaaS)

    Modeling and Simulation Study of Designer’s Bidirectional Behavior of Task Selection in Open Source Design Process

    Get PDF
    Open source design (OSD) is an emerging mode of product design. In OSD process, how to select right tasks directly influences the efficiency and quality of task completion, hence impacting the whole evolution process of OSD. In this paper, designer’s bidirectional behavior of task selection integrating passive selection based on website recommendation and autonomous selection is modeled. First, the model of passive selection behavior by website recommendation is proposed with application of collaborative filtering algorithm, based on a three-dimensional matrix including information of design agents, tasks, and skills; second, the model of autonomous selection behavior is described in consideration of factors such as skill and incentive; third, the model of bidirectional selection behavior is described integrating the aforementioned two selection algorithms. At last, contrast simulation analysis of bidirectional selection, passive selection based on website recommendation, and autonomous selection is proposed with ANOVA, and results show that task selection behavior has significant effect on OSD evolution process and that bidirectional selection behavior is more effective to shorten evolution cycle according to the experiment settings. In addition, the simulation study testifies the model of bidirectional selection by describing the task selection process of OSD in microperspective
    • 

    corecore