2 research outputs found
When Eye-Tracking Meets Machine Learning: A Systematic Review on Applications in Medical Image Analysis
Eye-gaze tracking research offers significant promise in enhancing various
healthcare-related tasks, above all in medical image analysis and
interpretation. Eye tracking, a technology that monitors and records the
movement of the eyes, provides valuable insights into human visual attention
patterns. This technology can transform how healthcare professionals and
medical specialists engage with and analyze diagnostic images, offering a more
insightful and efficient approach to medical diagnostics. Hence, extracting
meaningful features and insights from medical images by leveraging eye-gaze
data improves our understanding of how radiologists and other medical experts
monitor, interpret, and understand images for diagnostic purposes. Eye-tracking
data, with intricate human visual attention patterns embedded, provides a
bridge to integrating artificial intelligence (AI) development and human
cognition. This integration allows novel methods to incorporate domain
knowledge into machine learning (ML) and deep learning (DL) approaches to
enhance their alignment with human-like perception and decision-making.
Moreover, extensive collections of eye-tracking data have also enabled novel
ML/DL methods to analyze human visual patterns, paving the way to a better
understanding of human vision, attention, and cognition. This systematic review
investigates eye-gaze tracking applications and methodologies for enhancing
ML/DL algorithms for medical image analysis in depth
Spatial and time domain analysis of eye-tracking data during screening of brain magnetic resonance images
Introduction: Eye-tracking research has been widely used in radiology applications. Prior studies exclusively analysed either temporal or spatial eye-tracking features, both of which alone do not completely characterise the spatiotemporal dynamics of radiologists’ gaze features.
Purpose: Our research aims to quantify human visual search dynamics in both domains during brain stimuli screening to explore the relationship between reader characteristics and stimuli complexity. The methodology can be used to discover strategies to aid trainee radiologists in identifying pathology, and to select regions of interest for machine vision applications.
Method: The study was performed using eye-tracking data 5 seconds in duration from 57 readers (15 Brain-experts, 11 Other-experts, 5 Registrars and 26 Naïves) for 40 neuroradiological images as stimuli (i.e., 20 normal and 20 pathological brain MRIs). The visual scanning patterns were analysed by calculating the fractal dimension (FD) and Hurst exponent (HE) using re-scaled range (R/S) and detrended fluctuation analysis (DFA) methods. The FD was used to measure the spatial geometrical complexity of the gaze patterns, and the HE analysis was used to measure participants’ focusing skill. The focusing skill is referred to persistence/anti-persistence of the participants’ gaze on the stimulus over time. Pathological and normal stimuli were analysed separately both at the “First Second” and full “Five Seconds” viewing duration.
Results: All experts were more focused and a had higher visual search complexity compared to Registrars and Naïves. This was seen in both the pathological and normal stimuli in the first and five second analyses. The Brain-experts subgroup was shown to achieve better focusing skill than Other-experts due to their domain specific expertise. Indeed, the FDs found when viewing pathological stimuli were higher than those in normal ones. Viewing normal stimuli resulted in an increase of FD found in five second data, unlike pathological stimuli, which did not change. In contrast to the FDs, the scanpath HEs of pathological and normal stimuli were similar. However, participants’ gaze was more focused for “Five Seconds” than “First Second” data.
Conclusions: The HE analysis of the scanpaths belonging to all experts showed that they have greater focus than Registrars and NaĂŻves. This may be related to their higher visual search complexity than non-experts due to their training and expertise