4,231 research outputs found
Toward Theoretical Techniques for Measuring the Use of Human Effort in Visual Analytic Systems
Visual analytic systems have long relied on user studies and standard datasets to demonstrate advances to the state of the art, as well as to illustrate the efficiency of solutions to domain-specific challenges. This approach has enabled some important comparisons between systems, but unfortunately the narrow scope required to facilitate these comparisons has prevented many of these lessons from being generalized to new areas. At the same time, advanced visual analytic systems have made increasing use of human-machine collaboration to solve problems not tractable by machine computation alone. To continue to make progress in modeling user tasks in these hybrid visual analytic systems, we must strive to gain insight into what makes certain tasks more complex than others. This will require the development of mechanisms for describing the balance to be struck between machine and human strengths with respect to analytical tasks and workload. In this paper, we argue for the necessity of theoretical tools for reasoning about such balance in visual analytic systems and demonstrate the utility of the Human Oracle Model for this purpose in the context of sensemaking in visual analytics. Additionally, we make use of the Human Oracle Model to guide the development of a new system through a case study in the domain of cybersecurity
A review of rapid serial visual presentation-based brain-computer interfaces
International audienceRapid serial visual presentation (RSVP) combined with the detection of event related brain responses facilitates the selection of relevant information contained in a stream of images presented rapidly to a human. Event related potentials (ERPs) measured non-invasively with electroencephalography (EEG) can be associated with infrequent targets amongst a stream of images. Human-machine symbiosis may be augmented by enabling human interaction with a computer, without overt movement, and/or enable optimization of image/information sorting processes involving humans. Features of the human visual system impact on the success of the RSVP paradigm, but pre-attentive processing supports the identification of target information post presentation of the information by assessing the co-occurrence or time-locked EEG potentials. This paper presents a comprehensive review and evaluation of the limited but significant literature on research in RSVP-based brain-computer interfaces (BCIs). Applications that use RSVP-based BCIs are categorized based on display mode and protocol design, whilst a range of factors influencing ERP evocation and detection are analyzed. Guidelines for using the RSVP-based BCI paradigms are recommended, with a view to further standardizing methods and enhancing the inter-relatability of experimental design to support future research and the use of RSVP-based BCIs in practice
Recent Advances in Artificial Intelligence-Assisted Ultrasound Scanning
Funded by the Spanish Ministry of Economic Affairs and Digital Transformation (Project MIA.2021.M02.0005 TARTAGLIA, from the Recovery, Resilience, and Transformation Plan financed by the European Union through Next Generation EU funds). TARTAGLIA takes place under the R&D Missions in Artificial Intelligence program, which is part of the Spain Digital 2025 Agenda and the Spanish National Artificial Intelligence Strategy.Ultrasound (US) is a flexible imaging modality used globally as a first-line medical exam procedure in many different clinical cases. It benefits from the continued evolution of ultrasonic technologies and a well-established US-based digital health system. Nevertheless, its diagnostic performance still presents challenges due to the inherent characteristics of US imaging, such as manual operation and significant operator dependence. Artificial intelligence (AI) has proven to recognize complicated scan patterns and provide quantitative assessments for imaging data. Therefore, AI technology has the potential to help physicians get more accurate and repeatable outcomes in the US. In this article, we review the recent advances in AI-assisted US scanning. We have identified the main areas where AI is being used to facilitate US scanning, such as standard plane recognition and organ identification, the extraction of standard clinical planes from 3D US volumes, and the scanning guidance of US acquisitions performed by humans or robots. In general, the lack of standardization and reference datasets in this field makes it difficult to perform comparative studies among the different proposed methods. More open-access repositories of large US datasets with detailed information about the acquisition are needed to facilitate the development of this very active research field, which is expected to have a very positive impact on US imaging.Depto. de Estructura de la Materia, FĂsica TĂ©rmica y ElectrĂłnicaFac. de Ciencias FĂsicasTRUEMinistry of Economic Affairs and Digital Transformation from the Recovery, Resilience, and Transformation PlanNext Generation EU fundspu
Human-computer collaboration for skin cancer recognition
The rapid increase in telemedicine coupled with recent advances in diagnostic artificial intelligence (AI) create the imperative to consider the opportunities and risks of inserting AI-based support into new paradigms of care. Here we build on recent achievements in the accuracy of image-based AI for skin cancer diagnosis to address the effects of varied representations of AI-based support across different levels of clinical expertise and multiple clinical workflows. We find that good quality AI-based support of clinical decision-making improves diagnostic accuracy over that of either AI or physicians alone, and that the least experienced clinicians gain the most from AI-based support. We further find that AI-based multiclass probabilities outperformed content-based image retrieval (CBIR) representations of AI in the mobile technology environment, and AI-based support had utility in simulations of second opinions and of telemedicine triage. In addition to demonstrating the potential benefits associated with good quality AI in the hands of non-expert clinicians, we find that faulty AI can mislead the entire spectrum of clinicians, including experts. Lastly, we show that insights derived from AI class-activation maps can inform improvements in human diagnosis. Together, our approach and findings offer a framework for future studies across the spectrum of image-based diagnostics to improve human-computer collaboration in clinical practice
Understanding Questions that Arise When Working with Business Documents
While digital assistants are increasingly used to help with various
productivity tasks, less attention has been paid to employing them in the
domain of business documents. To build an agent that can handle users'
information needs in this domain, we must first understand the types of
assistance that users desire when working on their documents. In this work, we
present results from two user studies that characterize the information needs
and queries of authors, reviewers, and readers of business documents. In the
first study, we used experience sampling to collect users' questions in-situ as
they were working with their documents, and in the second, we built a
human-in-the-loop document Q&A system which rendered assistance with a variety
of users' questions. Our results have implications for the design of document
assistants that complement AI with human intelligence including whether
particular skillsets or roles within the document are needed from human
respondents, as well as the challenges around such systems.Comment: This paper will appear in CSCW'2
Speed of Rapid Serial Visual Presentation of Pictures, Numbers and Words Affects Event-Related Potential-Based Detection Accuracy
Rapid serial visual presentation (RSVP) based brain-computer interfaces (BCIs) can detect target images among a continuous stream of rapidly presented images, by classifying a viewer’s event related potentials (ERPs) associated with the target and non-targets images. Whilst the majority of RSVP-BCI studies to date have concentrated on the identification of a single type of image, namely pictures , here we study the capability of RSVP-BCI to detect three different target image types: pictures, numbers and words . The impact of presentation duration (speed) i.e., 100–200ms (5–10Hz), 200–300ms (3.3–5Hz) or 300–400ms (2.5–3.3Hz), is also investigated. 2-way repeated measure ANOVA on accuracies of detecting targets from non-target stimuli (ratio 1:9) measured via area under the receiver operator characteristics curve (AUC) for subjects revealed a significant effect of factor Stimulus-Type ( pictures, numbers, words ) (F (2,28) = 7.243, ) and for Stimulus-Duration (F (2,28) = 5.591, p = 0.011). Furthermore, there is an interaction between stimulus type and duration: F (4,56) = 4.419, ). The results indicate that when designing RSVP-BCI paradigms, the content of the images and the rate at which images are presented impact on the accuracy of detection and hence these parameters are key experimental variables in protocol design and applications, which apply RSVP for multimodal image datasets
Deep Learning Body Region Classification of MRI and CT examinations
Standardized body region labelling of individual images provides data that
can improve human and computer use of medical images. A CNN-based classifier
was developed to identify body regions in CT and MRI. 17 CT (18 MRI) body
regions covering the entire human body were defined for the classification
task. Three retrospective databases were built for the AI model training,
validation, and testing, with a balanced distribution of studies per body
region. The test databases originated from a different healthcare network.
Accuracy, recall and precision of the classifier was evaluated for patient age,
patient gender, institution, scanner manufacturer, contrast, slice thickness,
MRI sequence, and CT kernel. The data included a retrospective cohort of 2,934
anonymized CT cases (training: 1,804 studies, validation: 602 studies, test:
528 studies) and 3,185 anonymized MRI cases (training: 1,911 studies,
validation: 636 studies, test: 638 studies). 27 institutions from primary care
hospitals, community hospitals and imaging centers contributed to the test
datasets. The data included cases of all genders in equal proportions and
subjects aged from a few months old to +90 years old. An image-level prediction
accuracy of 91.9% (90.2 - 92.1) for CT, and 94.2% (92.0 - 95.6) for MRI was
achieved. The classification results were robust across all body regions and
confounding factors. Due to limited data, performance results for subjects
under 10 years-old could not be reliably evaluated. We show that deep learning
models can classify CT and MRI images by body region including lower and upper
extremities with high accuracy.Comment: 21 pages, 2 figures, 4 table
- …