9,048 research outputs found

    A Labeling Task Design for Supporting Algorithmic Needs: Facilitating Worker Diversity and Reducing AI Bias

    Full text link
    Studies on supervised machine learning (ML) recommend involving workers from various backgrounds in training dataset labeling to reduce algorithmic bias. Moreover, sophisticated tasks for categorizing objects in images are necessary to improve ML performance, further complicating micro-tasks. This study aims to develop a task design incorporating the fair participation of people, regardless of their specific backgrounds or task's difficulty. By collaborating with 75 labelers from diverse backgrounds for 3 months, we analyzed workers' log-data and relevant narratives to identify the task's hurdles and helpers. The findings revealed that workers' decision-making tendencies varied depending on their backgrounds. We found that the community that positively helps workers and the machine's feedback perceived by workers could make people easily engaged in works. Hence, ML's bias could be expectedly mitigated. Based on these findings, we suggest an extended human-in-the-loop approach that connects labelers, machines, and communities rather than isolating individual workers.Comment: 45 pages, 4 figure

    The challenges of blended learning using a media annotation tool

    Get PDF
    Blended learning has been evolving as an important approach to learning and teaching in tertiary education. This approach incorporates learning in both online and face-to-face modes and promotes deep learning by incorporating the best of both approaches. An innovation in blended learning is the use of an online media annotation tool (MAT) in combination with face-to-face classes. This tool allows students to annotate their own or teacher-uploaded video adding to their understanding of professional skills in various disciplines in tertiary education. Examination of MAT occurred in 2011 and included nine cohorts of students using the tool. This article canvasses selected data relating to MAT including insights into the use of blended learning focussing on the challenges of combining face-to-face and online learning using a relatively new online tool

    The Impact of Automatic Pre-annotation in Clinical Note Data Element Extraction - the CLEAN Tool

    Full text link
    Objective. Annotation is expensive but essential for clinical note review and clinical natural language processing (cNLP). However, the extent to which computer-generated pre-annotation is beneficial to human annotation is still an open question. Our study introduces CLEAN (CLinical note rEview and ANnotation), a pre-annotation-based cNLP annotation system to improve clinical note annotation of data elements, and comprehensively compares CLEAN with the widely-used annotation system Brat Rapid Annotation Tool (BRAT). Materials and Methods. CLEAN includes an ensemble pipeline (CLEAN-EP) with a newly developed annotation tool (CLEAN-AT). A domain expert and a novice user/annotator participated in a comparative usability test by tagging 87 data elements related to Congestive Heart Failure (CHF) and Kawasaki Disease (KD) cohorts in 84 public notes. Results. CLEAN achieved higher note-level F1-score (0.896) over BRAT (0.820), with significant difference in correctness (P-value < 0.001), and the mostly related factor being system/software (P-value < 0.001). No significant difference (P-value 0.188) in annotation time was observed between CLEAN (7.262 minutes/note) and BRAT (8.286 minutes/note). The difference was mostly associated with note length (P-value < 0.001) and system/software (P-value 0.013). The expert reported CLEAN to be useful/satisfactory, while the novice reported slight improvements. Discussion. CLEAN improves the correctness of annotation and increases usefulness/satisfaction with the same level of efficiency. Limitations include untested impact of pre-annotation correctness rate, small sample size, small user size, and restrictedly validated gold standard. Conclusion. CLEAN with pre-annotation can be beneficial for an expert to deal with complex annotation tasks involving numerous and diverse target data elements

    ForDigitStress: A multi-modal stress dataset employing a digital job interview scenario

    Full text link
    We present a multi-modal stress dataset that uses digital job interviews to induce stress. The dataset provides multi-modal data of 40 participants including audio, video (motion capturing, facial recognition, eye tracking) as well as physiological information (photoplethysmography, electrodermal activity). In addition to that, the dataset contains time-continuous annotations for stress and occurred emotions (e.g. shame, anger, anxiety, surprise). In order to establish a baseline, five different machine learning classifiers (Support Vector Machine, K-Nearest Neighbors, Random Forest, Long-Short-Term Memory Network) have been trained and evaluated on the proposed dataset for a binary stress classification task. The best-performing classifier achieved an accuracy of 88.3% and an F1-score of 87.5%

    Full Issue

    Get PDF

    MAGiC: A multimodal framework for analysing gaze in dyadic communication

    Get PDF
    The analysis of dynamic scenes has been a challenging domain in eye tracking research. This study presents a framework, named MAGiC, for analyzing gaze contact and gaze aversion in face-to-face communication. MAGiC provides an environment that is able to detect and track the conversation partner’s face automatically, overlay gaze data on top of the face video, and incorporate speech by means of speech-act annotation. Specifically, MAGiC integrates eye tracking data for gaze, audio data for speech segmentation, and video data for face tracking. MAGiC is an open source framework and its usage is demonstrated via publicly available video content and wiki pages. We explored the capabilities of MAGiC through a pilot study and showed that it facilitates the analysis of dynamic gaze data by reducing the annotation effort and the time spent for manual analysis of video data

    Using Hypervideo to support undergraduate students' reflection on work practices: a qualitative study

    Get PDF
    Abstract According to several exploratory studies, the HyperVideo seems to be particularly useful in highlighting the existing connections between the school-based and the work-based contexts, between authentic work situations and theoretical underpinnings. This tool and its features, in particular, the video annotation, seems to constitute an instrument which facilitates the students' reflection on work-practices. Even though several researchers have already studied the efficacy of HyperVideo, studies concerning the qualitative differences between a reflection process activated with or without its use are still missing. Therefore, the present contribution is focused on the reflective processes activated by two groups of students engaged in a higher education course while they carry out a reflective activity on work practices using the HyperVideo or not. The aim is to investigate wether the HyperVideo can be useful for students to foster the connection between theoretical concepts and work practices. Through multi-step qualitative analysis which combined Thematic Qualitative Text Analysis and Grounded Theory, a sample of reflective reports drafted by a group of students who employed HiperVideo to make a video-interview on a work-practice and to reflect on it (Group A) was compared with a sample of reflective reports drafted by a group who did not use it to complete the same task (Group B). The results emerging from the comparison of the coding frequencies between the two groups show that HyperVideo can support the reflective processes of students, better connecting theory and professional practice
    corecore