20 research outputs found
Recommended from our members
Machine learning based small bowel video capsule endoscopy analysis: Challenges and opportunities
YesVideo capsule endoscopy (VCE) is a revolutionary technology for the early diagnosis of gastric disorders. However, owing to the high redundancy and subtle manifestation of anomalies among thousands of frames, the manual construal of VCE videos requires considerable patience, focus, and time. The automatic analysis of these videos using computational methods is a challenge as the capsule is untamed in motion and captures frames inaptly. Several machine learning (ML) methods, including recent deep convolutional neural networks approaches, have been adopted after evaluating their potential of improving the VCE analysis. However, the clinical impact of these methods is yet to be investigated. This survey aimed to highlight the gaps between existing ML-based research methodologies and clinically significant rules recently established by gastroenterologists based on VCE. A framework for interpreting raw frames into contextually relevant frame-level findings and subsequently merging these findings with meta-data to obtain a disease-level diagnosis was formulated. Frame-level findings can be more intelligible for discriminative learning when organized in a taxonomical hierarchy. The proposed taxonomical hierarchy, which is formulated based on pathological and visual similarities, may yield better classification metrics by setting inference classes at a higher level than training classes. Mapping from the frame level to the disease level was structured in the form of a graph based on clinical relevance inspired by the recent international consensus developed by domain experts. Furthermore, existing methods for VCE summarization, classification, segmentation, detection, and localization were critically evaluated and compared based on aspects deemed significant by clinicians. Numerous studies pertain to single anomaly detection instead of a pragmatic approach in a clinical setting. The challenges and opportunities associated with VCE analysis were delineated. A focus on maximizing the discriminative power of features corresponding to various subtle lesions and anomalies may help cope with the diverse and mimicking nature of different VCE frames. Large multicenter datasets must be created to cope with data sparsity, bias, and class imbalance. Explainability, reliability, traceability, and transparency are important for an ML-based diagnostics system in a VCE. Existing ethical and legal bindings narrow the scope of possibilities where ML can potentially be leveraged in healthcare. Despite these limitations, ML based video capsule endoscopy will revolutionize clinical practice, aiding clinicians in rapid and accurate diagnosis
ADVANCED INTRAVASCULAR MAGNETIC RESONANCE IMAGING WITH INTERACTION
Intravascular (IV) Magnetic Resonance Imaging (MRI) is a specialized class of interventional MRI (iMRI) techniques that acquire MRI images through blood vessels to guide, identify and/or treat pathologies inside the human body which are otherwise difficult to locate and treat precisely. Here, interactions based on real-time computations and feedback are explored to improve the accuracy and efficiency of IVMRI procedures.
First, an IV MRI-guided high-intensity focused ultrasound (HIFU) ablation method is developed for targeting perivascular pathology with minimal injury to the vessel wall. To take advantage of real-time feedback, a software interface is developed for monitoring thermal dose with real-time MRI thermometry, and an MRI-guided ablation protocol developed and tested on muscle and liver tissue ex vivo. It is shown that, with cumulative thermal dose monitored with MRI thermometry, lesion location and dimensions can be estimated consistently, and desirable thermal lesions can be achieved in animals in vivo.
Second, to achieve fully interactive IV MRI, high-resolution real-time 10 frames-per-second (fps) MRI endoscopy is developed as an advance over prior methods of MRI endoscopy. Intravascular transmit-receive MRI endoscopes are fabricated for highly under-sampled radial-projection MRI in a clinical 3Tesla MRI scanner. Iterative nonlinear reconstruction is accelerated using graphics processor units (GPU) to achieve true real-time endoscopy visualization at the scanner. The results of high-speed MRI endoscopy at 6-10 fps are consistent with fully-sampled MRI endoscopy and histology, with feasibility demonstrated in vivo in a large animal model.
Last, a general framework for automatic imaging contrast tuning over MRI protocol parameters is explored. The framework reveals typical signal patterns over different protocol parameters from calibration imaging data and applies this knowledge to design efficient acquisition strategies and predicts contrasts under unacquired protocols. An external computer in real-time communication with the MRI console is utilized for online processing and controlling MRI acquisitions. This workflow enables machine learning for optimizing acquisition strategies in general, and provides a foundation for efficiently tuning MRI protocol parameters to perform interventional MRI in the highly varying and interactive environments commonly in play. This work is loosely inspired by prior research on extremely accelerated MRI relaxometry using the minimal-acquisition linear algebraic modeling (SLAM) method
A review of silhouette extraction algorithms for use within visual hull pipelines
© 2020, © 2020 Informa UK Limited, trading as Taylor & Francis Group. Markerless motion capture would permit the study of human biomechanics in environments where marker-based systems are impractical, e.g. outdoors or underwater. The visual hull tool may enable such data to be recorded, but it requires the accurate detection of the silhouette of the object in multiple camera views. This paper reviews the top-performing algorithms available to date for silhouette extraction, with the visual hull in mind as the downstream application; the rationale is that higher-quality silhouettes would lead to higher-quality visual hulls, and consequently better measurement of movement. This paper is the first attempt in the literature to compare silhouette extraction algorithms that belong to different fields of Computer Vision, namely background subtraction, semantic segmentation, and multi-view segmentation. It was found that several algorithms exist that would be substantial improvements over the silhouette extraction algorithms traditionally used in visual hull pipelines. In particular, FgSegNet v2 (a background subtraction algorithm), DeepLabv3+ JFT (a semantic segmentation algorithm), and Djelouah 2013 (a multi-view segmentation algorithm) are the most accurate and promising methods for the extraction of silhouettes from 2D images to date, and could seamlessly be integrated within a visual hull pipeline for studies of human movement or biomechanics