3 research outputs found
Federated Benchmarking of Medical Artificial Intelligence With MedPerf
Medical artificial intelligence (AI) has tremendous potential to advance healthcare by supporting and contributing to the evidence-based practice of medicine, personalizing patient treatment, reducing costs, and improving both healthcare provider and patient experience. Unlocking this potential requires systematic, quantitative evaluation of the performance of medical AI models on large-scale, heterogeneous data capturing diverse patient populations. Here, to meet this need, we introduce MedPerf, an open platform for benchmarking AI models in the medical domain. MedPerf focuses on enabling federated evaluation of AI models, by securely distributing them to different facilities, such as healthcare organizations. This process of bringing the model to the data empowers each facility to assess and verify the performance of AI models in an efficient and human-supervised process, while prioritizing privacy. We describe the current challenges healthcare and AI communities face, the need for an open platform, the design philosophy of MedPerf, its current implementation status and real-world deployment, our roadmap and, importantly, the use of MedPerf with multiple international institutions within cloud-based technology and on-premises scenarios. Finally, we welcome new contributions by researchers and organizations to further strengthen MedPerf as an open benchmarking platform
Federated Cycling (FedCy): Semi-supervised Federated Learning of Surgical Phases
Recent advancements in deep learning methods bring computer-assistance a step
closer to fulfilling promises of safer surgical procedures. However, the
generalizability of such methods is often dependent on training on diverse
datasets from multiple medical institutions, which is a restrictive requirement
considering the sensitive nature of medical data. Recently proposed
collaborative learning methods such as Federated Learning (FL) allow for
training on remote datasets without the need to explicitly share data. Even so,
data annotation still represents a bottleneck, particularly in medicine and
surgery where clinical expertise is often required. With these constraints in
mind, we propose FedCy, a federated semi-supervised learning (FSSL) method that
combines FL and self-supervised learning to exploit a decentralized dataset of
both labeled and unlabeled videos, thereby improving performance on the task of
surgical phase recognition. By leveraging temporal patterns in the labeled
data, FedCy helps guide unsupervised training on unlabeled data towards
learning task-specific features for phase recognition. We demonstrate
significant performance gains over state-of-the-art FSSL methods on the task of
automatic recognition of surgical phases using a newly collected
multi-institutional dataset of laparoscopic cholecystectomy videos.
Furthermore, we demonstrate that our approach also learns more generalizable
features when tested on data from an unseen domain.Comment: 11 pages, 4 figure
Preserving privacy in surgical video analysis using a deep learning classifier to identify out-of-body scenes in endoscopic videos
Abstract Surgical video analysis facilitates education and research. However, video recordings of endoscopic surgeries can contain privacy-sensitive information, especially if the endoscopic camera is moved out of the body of patients and out-of-body scenes are recorded. Therefore, identification of out-of-body scenes in endoscopic videos is of major importance to preserve the privacy of patients and operating room staff. This study developed and validated a deep learning model for the identification of out-of-body images in endoscopic videos. The model was trained and evaluated on an internal dataset of 12 different types of laparoscopic and robotic surgeries and was externally validated on two independent multicentric test datasets of laparoscopic gastric bypass and cholecystectomy surgeries. Model performance was evaluated compared to human ground truth annotations measuring the receiver operating characteristic area under the curve (ROC AUC). The internal dataset consisting of 356,267 images from 48 videos and the two multicentric test datasets consisting of 54,385 and 58,349 images from 10 and 20 videos, respectively, were annotated. The model identified out-of-body images with 99.97% ROC AUC on the internal test dataset. Mean ± standard deviation ROC AUC on the multicentric gastric bypass dataset was 99.94 ± 0.07% and 99.71 ± 0.40% on the multicentric cholecystectomy dataset, respectively. The model can reliably identify out-of-body images in endoscopic videos and is publicly shared. This facilitates privacy preservation in surgical video analysis