2,665 research outputs found
Rethinking Human-AI Collaboration in Complex Medical Decision Making: A Case Study in Sepsis Diagnosis
Today's AI systems for medical decision support often succeed on benchmark
datasets in research papers but fail in real-world deployment. This work
focuses on the decision making of sepsis, an acute life-threatening systematic
infection that requires an early diagnosis with high uncertainty from the
clinician. Our aim is to explore the design requirements for AI systems that
can support clinical experts in making better decisions for the early diagnosis
of sepsis. The study begins with a formative study investigating why clinical
experts abandon an existing AI-powered Sepsis predictive module in their
electrical health record (EHR) system. We argue that a human-centered AI system
needs to support human experts in the intermediate stages of a medical
decision-making process (e.g., generating hypotheses or gathering data),
instead of focusing only on the final decision. Therefore, we build SepsisLab
based on a state-of-the-art AI algorithm and extend it to predict the future
projection of sepsis development, visualize the prediction uncertainty, and
propose actionable suggestions (i.e., which additional laboratory tests can be
collected) to reduce such uncertainty. Through heuristic evaluation with six
clinicians using our prototype system, we demonstrate that SepsisLab enables a
promising human-AI collaboration paradigm for the future of AI-assisted sepsis
diagnosis and other high-stakes medical decision making.Comment: Under submission to CHI202
Assessing the Accuracy of Task Time Prediction of an Emerging Human Performance Modeling Software - CogTool
There is a need for a human performance modeling tool which not only has the ability to accurately estimate skilled user task time for any interface design, but can be used by modelers with little or no programming knowledge and at a minimal cost. To fulfill this need, this research investigated the accuracy of task time prediction of a modeling tool – CogTool - on two versions of an interface design used extensively in the petrochemical industry – DeltaV. CogTool uses the KeyStroke Level Model (KLM) to calculate and generate time predictions based on specified operators. The data collected from a previous study (Koffskey, Ikuma, & Harvey, 2013) that investigated how human participants (24 students and 4 operators) performed on these interfaces (in terms of mean speed in seconds) were compared to CogTool’s numeric time estimate. Three tasks (pump I, pump II and cascade system failures) on each interface for both participant groups were tested on both interfaces (improved and poor), on the general hypothesis that CogTool will make task time predictions for each of the modeled tasks, within a certain range of what actual human participants had demonstrated. The 95% confidence interval (CI) tests of the means were used to determine if the predictions fall within the intervals. The estimated task time from CogTool did not fall within the 95% CI in 9 of 12 cases. Of the 3 that were contained in the acceptable interval, two belonged to the experienced operator group for tasks performed on the improved interface, implying that CogTool was better in predicting the operators’ performance than the students’. A control room monitoring task, by its nature, places great demand on an operator’s mental capacity. This also includes the fact that operators work on multiple screens and/or consoles, sometimes requiring them to commit information to memory that they have to revisit a screen to check on some vital information. In this regard, it is suggested that the one user mental operator for “think time” (estimated as 1.2sec), should be revised in CogTool to accommodate the demand on the operator. For this reason, the present CogTool prediction did not meet expectations in estimating control room operator task time, but it however succeeded in showing where the poor interface could be improved by comparing the detailed steps to the improved interface
Trends in Smart City Development
This report examines the meanings and practices associated with the term 'smart cities.' Smart city initiatives involve three components: information and communication technologies (ICTs) that generate and aggregate data; analytical tools which convert that data into usable information; and organizational structures that encourage collaboration, innovation, and the application of that information to solve public problems
Practical, appropriate, empirically-validated guidelines for designing educational games
There has recently been a great deal of interest in the
potential of computer games to function as innovative
educational tools. However, there is very little evidence of
games fulfilling that potential. Indeed, the process of
merging the disparate goals of education and games design
appears problematic, and there are currently no practical
guidelines for how to do so in a coherent manner. In this
paper, we describe the successful, empirically validated
teaching methods developed by behavioural psychologists
and point out how they are uniquely suited to take
advantage of the benefits that games offer to education. We
conclude by proposing some practical steps for designing
educational games, based on the techniques of Applied
Behaviour Analysis. It is intended that this paper can both
focus educational games designers on the features of games
that are genuinely useful for education, and also introduce a
successful form of teaching that this audience may not yet
be familiar with
Analyzing Granger causality in climate data with time series classification methods
Attribution studies in climate science aim for scientifically ascertaining the influence of climatic variations on natural or anthropogenic factors. Many of those studies adopt the concept of Granger causality to infer statistical cause-effect relationships, while utilizing traditional autoregressive models. In this article, we investigate the potential of state-of-the-art time series classification techniques to enhance causal inference in climate science. We conduct a comparative experimental study of different types of algorithms on a large test suite that comprises a unique collection of datasets from the area of climate-vegetation dynamics. The results indicate that specialized time series classification methods are able to improve existing inference procedures. Substantial differences are observed among the methods that were tested
- …