26 research outputs found
CoronARe: A Coronary Artery Reconstruction Challenge
CoronARe ranks state-of-the-art methods in symbolic and tomographic coronary artery reconstruction from interventional C-arm rotational angiography. Specifically, we benchmark the performance of the methods using accurately pre-processed data, and study the effects of imperfect pre-processing conditions (segmentation and background subtraction errors). In this first iteration of the challenge, evaluation is performed in a controlled environment using digital phantom images, where accurate 3D ground truth is known
AutoSNAP: Automatically Learning Neural Architectures for Instrument Pose Estimation
Despite recent successes, the advances in Deep Learning have not yet been
fully translated to Computer Assisted Intervention (CAI) problems such as pose
estimation of surgical instruments. Currently, neural architectures for
classification and segmentation tasks are adopted ignoring significant
discrepancies between CAI and these tasks. We propose an automatic framework
(AutoSNAP) for instrument pose estimation problems, which discovers and learns
the architectures for neural networks. We introduce 1)~an efficient testing
environment for pose estimation, 2)~a powerful architecture representation
based on novel Symbolic Neural Architecture Patterns (SNAPs), and 3)~an
optimization of the architecture using an efficient search scheme. Using
AutoSNAP, we discover an improved architecture (SNAPNet) which outperforms both
the hand-engineered i3PosNet and the state-of-the-art architecture search
method DARTS.Comment: Accepted at MICCAI 2020 Preparing code for release at
https://github.com/MECLabTUDA/AutoSNA
2018 Robotic Scene Segmentation Challenge
In 2015 we began a sub-challenge at the EndoVis workshop at MICCAI in Munich using endoscope images of ex-vivo tissue with automatically generated annotations from robot forward kinematics and instrument CAD models. However, the limited background variation and simple motion rendered the dataset uninformative in learning about which techniques would be suitable for segmentation in real surgery. In 2017, at the same workshop in Quebec we introduced the robotic instrument segmentation dataset with 10 teams participating in the challenge to perform binary, articulating parts and type segmentation of da Vinci instruments. This challenge included realistic instrument motion and more complex porcine tissue as background and was widely addressed with modifications on U-Nets and other popular CNN architectures. In 2018 we added to the complexity by introducing a set of anatomical objects and medical devices to the segmented classes. To avoid over-complicating the challenge, we continued with porcine data which is dramatically simpler than human tissue due to the lack of fatty tissue occluding many organs
Direct Gene Expression Profile Prediction for Uveal Melanoma from Digital Cytopathology Images via Deep Learning and Salient Image Region Identification
To demonstrate that deep learning (DL) methods can produce robust prediction of gene expression profile (GEP) in uveal melanoma (UM) based on digital cytopathology images.
Evaluation of a diagnostic test or technology.
Deidentified smeared cytology slides stained with hematoxylin and eosin obtained from a fine needle aspirated from UM.
Digital whole-slide images were generated by fine-needle aspiration biopsies of UM tumors that underwent GEP testing. A multistage DL system was developed with automatic region-of-interest (ROI) extraction from digital cytopathology images, an attention-based neural network, ROI feature aggregation, and slide-level data augmentation.
The ability of our DL system in predicting GEP on a slide (patient) level. Data were partitioned at the patient level (73% training; 27% testing).
In total, our study included 89 whole-slide images from 82 patients and 121 388 unique ROIs. The testing set included 24 slides from 24 patients (12 class 1 tumors; 12 class 2 tumors; 1 slide per patient). Our DL system for GEP prediction achieved an area under the receiver operating characteristic curve of 0.944, an accuracy of 91.7%, a sensitivity of 91.7%, and a specificity of 91.7% on a slide-level analysis. The incorporation of slide-level feature aggregation and data augmentation produced a more predictive DL model (PÂ = 0.0031).
Our current work established a complete pipeline for GEP prediction in UM tumors: from automatic ROI extraction from digital cytopathology whole-slide images to slide-level predictions. Our DL system demonstrated robust performance and, if validated prospectively, could serve as an image-based alternative to GEP testing
PĂ„lagt delt fast bosted: En analyse av lagmannsrettspraksis - hvilke momenter vektlegges og hvilken vekt har de?
Labelling data is expensive and time consuming especially for domains such as
medical imaging that contain volumetric imaging data and require expert
knowledge. Exploiting a larger pool of labeled data available across multiple
centers, such as in federated learning, has also seen limited success since
current deep learning approaches do not generalize well to images acquired with
scanners from different manufacturers. We aim to address these problems in a
common, learning-based image simulation framework which we refer to as
Federated Simulation. We introduce a physics-driven generative approach that
consists of two learnable neural modules: 1) a module that synthesizes 3D
cardiac shapes along with their materials, and 2) a CT simulator that renders
these into realistic 3D CT Volumes, with annotations. Since the model of
geometry and material is disentangled from the imaging sensor, it can
effectively be trained across multiple medical centers. We show that our data
synthesis framework improves the downstream segmentation performance on several
datasets. Project Page: https://nv-tlabs.github.io/fed-sim/ .Comment: MICCAI 2020 (Early Accept
Direct Gene Expression Profile Prediction for Uveal Melanoma from Digital Cytopathology Images via Deep Learning and Salient Image Region Identification
Objective: To demonstrate that deep learning (DL) methods can produce robust prediction of gene expression profile (GEP) in uveal melanoma (UM) based on digital cytopathology images. Design: Evaluation of a diagnostic test or technology. Subjects, Participants, and Controls: Deidentified smeared cytology slides stained with hematoxylin and eosin obtained from a fine needle aspirated from UM. Methods: Digital whole-slide images were generated by fine-needle aspiration biopsies of UM tumors that underwent GEP testing. A multistage DL system was developed with automatic region-of-interest (ROI) extraction from digital cytopathology images, an attention-based neural network, ROI feature aggregation, and slide-level data augmentation. Main Outcome Measures: The ability of our DL system in predicting GEP on a slide (patient) level. Data were partitioned at the patient level (73% training; 27% testing). Results: In total, our study included 89 whole-slide images from 82 patients and 121 388 unique ROIs. The testing set included 24 slides from 24 patients (12 class 1 tumors; 12 class 2 tumors; 1 slide per patient). Our DL system for GEP prediction achieved an area under the receiver operating characteristic curve of 0.944, an accuracy of 91.7%, a sensitivity of 91.7%, and a specificity of 91.7% on a slide-level analysis. The incorporation of slide-level feature aggregation and data augmentation produced a more predictive DL model (PÂ = 0.0031). Conclusions: Our current work established a complete pipeline for GEP prediction in UM tumors: from automatic ROI extraction from digital cytopathology whole-slide images to slide-level predictions. Our DL system demonstrated robust performance and, if validated prospectively, could serve as an image-based alternative to GEP testing