81 research outputs found

    FDA-cleared artificial intelligence and machine learning-based medical devices and their 510(k) predicate networks

    Full text link
    The US Food and Drug Administration is clearing an increasing number of artificial intelligence and machine learning (AI/ML)-based medical devices through the 510(k) pathway. This pathway allows clearance if the device is substantially equivalent to a former cleared device (ie, predicate). We analysed the predicate networks of cleared AI/ML-based medical devices (cleared between 2019 and 2021), their underlying tasks, and recalls. More than a third of cleared AI/ML-based medical devices originated from non-AI/ML-based medical devices in the first generation. Devices with the longest time since the last predicate device with an AI/ML component were haematology (2001), radiology (2001), and cardiovascular devices (2008). Especially for devices in radiology, the AI/ML tasks changed frequently along the device's predicate network, raising safety concerns. To date, only a few recalls might have affected the AI/ML components. To improve patient care, a stronger focus should be placed on the distinctive characteristics of AI/ML when defining substantial equivalence between a new AI/ML-based medical device and predicate devices

    Fully automated breast segmentation on spiral breast computed tomography images

    Full text link
    INTRODUCTION The quantification of the amount of the glandular tissue and breast density is important to assess breast cancer risk. Novel photon-counting breast computed tomography (CT) technology has the potential to quantify them. For accurate analysis, a dedicated method to segment the breast components-the adipose and glandular tissue, skin, pectoralis muscle, skinfold section, rib, and implant-is required. We propose a fully automated breast segmentation method for breast CT images. METHODS The framework consists of four parts: (1) investigate, (2) segment the components excluding adipose and glandular tissue, (3) assess the breast density, and (4) iteratively segment the glandular tissue according to the estimated density. For the method, adapted seeded watershed and region growing algorithm were dedicatedly developed for the breast CT images and optimized on 68 breast images. The segmentation performance was qualitatively (five-point Likert scale) and quantitatively (Dice similarity coefficient [DSC] and difference coefficient [DC]) demonstrated according to human reading by experienced radiologists. RESULTS The performance evaluation on each component and overall segmentation for 17 breast CT images resulted in DSCs ranging 0.90-0.97 and in DCs 0.01-0.08. The readers rated 4.5-4.8 (5 highest score) with an excellent inter-reader agreement. The breast density varied by 3.7%-7.1% when including mis-segmented muscle or skin. CONCLUSION The automatic segmentation results coincided with the human expert's reading. The accurate segmentation is important to avoid the significant bias in breast density analysis. Our method enables accurate quantification of the breast density and amount of the glandular tissue that is directly related to breast cancer risk

    RadAdapt: Radiology Report Summarization via Lightweight Domain Adaptation of Large Language Models

    Full text link
    We systematically investigate lightweight strategies to adapt large language models (LLMs) for the task of radiology report summarization (RRS). Specifically, we focus on domain adaptation via pretraining (on natural language, biomedical text, and clinical text) and via prompting (zero-shot, in-context learning) or parameter-efficient fine-tuning (prefix tuning, LoRA). Our results on the MIMIC-III dataset consistently demonstrate best performance by maximally adapting to the task via pretraining on clinical text and parameter-efficient fine-tuning on RRS examples. Importantly, this method fine-tunes a mere 0.32% of parameters throughout the model, in contrast to end-to-end fine-tuning (100% of parameters). Additionally, we study the effect of in-context examples and out-of-distribution (OOD) training before concluding with a radiologist reader study and qualitative analysis. Our findings highlight the importance of domain adaptation in RRS and provide valuable insights toward developing effective natural language processing solutions for clinical tasks.Comment: 12 pages, 9 figure

    Bacterial computing with engineered populations

    Get PDF
    We describe strategies for the construction of bacterial computing platforms by describing a number of results from the recently completed bacterial computing with engineered populations project. In general, the implementation of such systems requires a framework containing various components such as intracellular circuits, single cell input/output and cell–cell interfacing, as well as extensive analysis. In this overview paper, we describe our approach to each of these, and suggest possible areas for future research

    Clinical Text Summarization: Adapting Large Language Models Can Outperform Human Experts

    Full text link
    Sifting through vast textual data and summarizing key information imposes a substantial burden on how clinicians allocate their time. Although large language models (LLMs) have shown immense promise in natural language processing (NLP) tasks, their efficacy across diverse clinical summarization tasks has not yet been rigorously examined. In this work, we employ domain adaptation methods on eight LLMs, spanning six datasets and four distinct summarization tasks: radiology reports, patient questions, progress notes, and doctor-patient dialogue. Our thorough quantitative assessment reveals trade-offs between models and adaptation methods in addition to instances where recent advances in LLMs may not lead to improved results. Further, in a clinical reader study with six physicians, we depict that summaries from the best adapted LLM are preferable to human summaries in terms of completeness and correctness. Our ensuing qualitative analysis delineates mutual challenges faced by both LLMs and human experts. Lastly, we correlate traditional quantitative NLP metrics with reader study scores to enhance our understanding of how these metrics align with physician preferences. Our research marks the first evidence of LLMs outperforming human experts in clinical text summarization across multiple tasks. This implies that integrating LLMs into clinical workflows could alleviate documentation burden, empowering clinicians to focus more on personalized patient care and other irreplaceable human aspects of medicine.Comment: 23 pages, 22 figure

    Reflection of neuroblastoma intratumor heterogeneity in the new OHC-NB1 disease model

    Get PDF
    Accurate modeling of intratumor heterogeneity presents a bottleneck against drug testing. Flexibility in a preclinical platform is also desirable to support assessment of different endpoints. We established the model system, OHC-NB1, from a bone marrow metastasis from a patient diagnosed with MYCN-amplified neuroblastoma and performed whole-exome sequencing on the source metastasis and the different models and passages during model development (monolayer cell line, 3D spheroid culture and subcutaneous xenograft tumors propagated in mice). OHC-NB1 harbors a MYCN amplification in double minutes, 1p deletion, 17q gain and diploid karyotype, which persisted in all models. A total of 80-540 single-nucleotide variants (SNVs) was detected in each sample, and comparisons between the source metastasis and models identified 34 of 80 somatic SNVs to be propagated in the models. Clonal reconstruction using the combined copy number and SNV data revealed marked clonal heterogeneity in the originating metastasis, with 4 clones being reflected in the model systems. The set of OHC-NB1 models represents 43% of somatic SNVs and 23% of the cellularity in the originating metastasis with varying clonal compositions, indicating that heterogeneity is partially preserved in our model system
    • …
    corecore