33 research outputs found

    The natural language processing of radiology requests and reports of chest imaging:Comparing five transformer models’ multilabel classification and a proof-of-concept study

    Get PDF
    Background: Radiology requests and reports contain valuable information about diagnostic findings and indications, and transformer-based language models are promising for more accurate text classification. Methods: In a retrospective study, 2256 radiologist-annotated radiology requests (8 classes) and reports (10 classes) were divided into training and testing datasets (90% and 10%, respectively) and used to train 32 models. Performance metrics were compared by model type (LSTM, Bertje, RobBERT, BERT-clinical, BERT-multilingual, BERT-base), text length, data prevalence, and training strategy. The best models were used to predict the remaining 40,873 cases’ categories of the datasets of requests and reports. Results: The RobBERT model performed the best after 4000 training iterations, resulting in AUC values ranging from 0.808 [95% CI (0.757–0.859)] to 0.976 [95% CI (0.956–0.996)] for the requests and 0.746 [95% CI (0.689–0.802)] to 1.0 [95% CI (1.0–1.0)] for the reports. The AUC for the classification of normal reports was 0.95 [95% CI (0.922–0.979)]. The predicted data demonstrated variability of both diagnostic yield for various request classes and request patterns related to COVID-19 hospital admission data. Conclusion: Transformer-based natural language processing is feasible for the multilabel classification of chest imaging request and report items. Diagnostic yield varies with the information in the requests

    VeLight:A 3D virtual reality tool for CT-based anatomy teaching and training

    Get PDF
    Abstract: For doctors and other medical professionals, the human body is the focus of their daily practice. A solid understanding of how it is built up, that is, the anatomy of the human body, is essential to ensure safe medical practice. Current anatomy education takes place either using text books or via dissecting human cadavers, with text books being the most traditional way to learn anatomy due to the cost of the alternatives. However, printed media offer only a 2D perception of a part of the human body. Although dissection of human cadavers can give a more direct observation and interaction with human bodies, it is extremely costly because of the need of preserving human bodies and maintaining dissection rooms. To solve this issue, we developed VeLight, a system with which students can learn anatomy based on CT datasets using a 3D Virtual Reality display (zSpace). VeLight offers simple and intuitive interactions, and allows teachers to design their own courses using their own material. The system offers an interactive, depth-perceptive learning experience and improves the learning process. We conducted an informal user study to validate the effectiveness of VeLight. The results show that participants were able to learn and remember how to work with VeLight very quickly. All participants reported enthusiasm for the potential of VeLight in the domain of medical education. Graphic Abstract: [Figure not available: see fulltext.

    Comparison of computed tomography image features extracted by radiomics, self-supervised learning and end-to-end deep learning for outcome prediction of oropharyngeal cancer

    Get PDF
    Background and purpose: To compare the prediction performance of image features of computed tomography (CT) images extracted by radiomics, self-supervised learning and end-to-end deep learning for local control (LC), regional control (RC), locoregional control (LRC), distant metastasis-free survival (DMFS), tumor-specific survival (TSS), overall survival (OS) and disease-free survival (DFS) of oropharyngeal squamous cell carcinoma (OPSCC) patients after (chemo)radiotherapy.Methods and materials: The OPC-Radiomics dataset was used for model development and independent internal testing and the UMCG-OPC set for external testing. Image features were extracted from the Gross Tumor Volume contours of the primary tumor (GTVt) regions in CT scans when using radiomics or a self-supervised learning-based method (autoencoder). Clinical and combined (radiomics, autoencoder or end-to-end) models were built using multivariable Cox proportional-hazard analysis with clinical features only and both clinical and image features for LC, RC, LRC, DMFS, TSS, OS and DFS prediction, respectively.Results: In the internal test set, combined autoencoder models performed better than clinical models and combined radiomics models for LC, RC, LRC, DMFS, TSS and DFS prediction (largest improvements in C-index: 0.91 vs. 0.76 in RC and 0.74 vs. 0.60 in DMFS). In the external test set, combined radiomics models performed better than clinical and combined autoencoder models for all endpoints (largest improvements in LC, 0.82 vs. 0.71). Furthermore, combined models performed better in risk stratification than clinical models and showed good calibration for most endpoints.Conclusions: Image features extracted using self-supervised learning showed best internal prediction performance while radiomics features have better external generalizability.</p

    CT-based deep multi-label learning prediction model for outcome in patients with oropharyngeal squamous cell carcinoma

    Get PDF
    Background: Personalized treatment is increasingly required for oropharyngeal squamous cell carcinoma (OPSCC) patients due to emerging new cancer subtypes and treatment options. Outcome prediction model can help identify low or high-risk patients who may be suitable to receive de-escalation or intensified treatment approaches.Purpose: To develop a deep learning (DL)-based model for predicting multiple and associated efficacy endpoints in OPSCC patients based on computed tomography (CT).Methods: Two patient cohorts were used in this study: a development cohort consisting of 524 OPSCC patients (70% for training and 30% for independent testing) and an external test cohort of 396 patients. Pre-treatment CT-scans with the gross primary tumor volume contours (GTVt) and clinical parameters were available to predict endpoints, including 2-year local control (LC), regional control (RC), locoregional control (LRC), distant metastasis-free survival (DMFS), disease-specific survival (DSS), overall survival (OS), and disease-free survival (DFS). We proposed DL outcome prediction models with the multi-label learning (MLL) strategy that integrates the associations of different endpoints based on clinical factors and CT-scans.Results: The multi-label learning models outperformed the models that were developed based on a single endpoint for all endpoints especially with high AUCs ≥ 0.80 for 2-year RC, DMFS, DSS, OS, and DFS in the internal independent test set and for all endpoints except 2-year LRC in the external test set. Furthermore, with the models developed, patients could be stratified into high and low-risk groups that were significantly different for all endpoints in the internal test set and for all endpoints except DMFS in the external test set. Conclusion: MLL models demonstrated better discriminative ability for all 2-year efficacy endpoints than single outcome models in the internal test and for all endpoints except LRC in the external set.</p

    PHP48 COST SENSITIVENESS AND PHYSICIAN TREATMENT CHOICES

    Get PDF
    Objectives To explore the relationship between nodule count and lung cancer probability in baseline low-dose CT lung cancer screening. Materials and Methods Included were participants from the NELSON trial with at least one baseline nodule (3392 participants [45% of screen-group], 7258 nodules). We determined nodule count per participant. Malignancy was confirmed by histology. Nodules not diagnosed as screen-detected or interval cancer until the end of the fourth screening round were regarded as benign. We compared lung cancer probability per nodule count category. Results 1746 (51.5%) participants had one nodule, 800 (23.6%) had two nodules, 354 (10.4%) had three nodules, 191 (5.6%) had four nodules, and 301 (8.9%) had > 4 nodules. Lung cancer in a baseline nodule was diagnosed in 134 participants (139 cancers; 4.0%). Median nodule count in participants with only benign nodules was 1 (Inter-quartile range [IQR]: 1–2), and 2 (IQR 1–3) in participants with lung cancer (p = NS). At baseline, malignancy was detected mostly in the largest nodule (64/66 cancers). Lung cancer probability was 62/1746 (3.6%) in case a participant had one nodule, 33/800 (4.1%) for two nodules, 17/354 (4.8%) for three nodules, 12/191 (6.3%) for four nodules and 10/301 (3.3%) for > 4 nodules (p = NS). Conclusion In baseline lung cancer CT screening, half of participants with lung nodules have more than one nodule. Lung cancer probability does not significantly change with the number of nodules. Baseline nodule count will not help to differentiate between benign and malignant nodules. Each nodule found in lung cancer screening should be assessed separately independent of the presence of other nodules

    Mobile health apps to facilitate self-care: a qualitative study of user experiences

    Get PDF
    Objective: Consumers are living longer, creating more pressure on the health system and increasing their requirement for self-care of chronic conditions. Despite rapidly-increasing numbers of mobile health applications (‘apps’) for consumers’ self-care, there is a paucity of research into consumer engagement with electronic self-monitoring. This paper presents a qualitative exploration of how health consumers use apps for health monitoring, their perceived benefits from use of health apps, and suggestions for improvement of health apps. Materials and Methods: ‘Health app’ was defined as any commercially-available health or fitness app with capacity for self-monitoring. English-speaking consumers aged 18 years and older using any health app for self-monitoring were recruited for interview from the metropolitan area of Perth, Australia. The semi-structured interview guide comprised questions based on the Technology Acceptance Model, Health Information Technology Acceptance Model, and the Mobile Application Rating Scale, and is the only study to do so. These models also facilitated deductive thematic analysis of interview transcripts. Implicit and explicit responses not aligned to these models were analyzed inductively.Results: Twenty-two consumers (15 female, seven male) participated, 13 of whom were aged 26–35 years. Eighteen participants reported on apps used on iPhones. Apps were used to monitor diabetes, asthma, depression, celiac disease, blood pressure, chronic migraine, pain management, menstrual cycle irregularity, and fitness. Most were used approximately weekly for several minutes per session, and prior to meeting initial milestones, with significantly decreased usage thereafter. Deductive and inductive thematic analysis reduced the data to four dominant themes: engagement in use of the app; technical functionality of the app; ease of use and design features; and management of consumers’ data. Conclusions: The semi-structured interviews provided insight into usage, benefits and challenges of health monitoring using apps. Understanding the range of consumer experiences and expectations can inform design of health apps to encourage persistence in self-monitoring

    From Physical Film to Picture Archiving and Communication Systems.

    No full text
    corecore