9,880 research outputs found

    Combined information from Raman spectroscopy and optical coherence tomography for enhanced diagnostic accuracy in tissue discrimination

    Get PDF
    We thank the UK EPSRC for funding, the CR-UK/EPSRC/MRC/DoH (England) imaging programme, the European Union project FAMOS (FP7 ICT, contract no. 317744) and the European Union project IIIOS (FP7/2007-2013, contract no. 238802). We thank Tayside Tissue Bank for providing us with the tissue samples under request number TR000289. K.D. is a Royal Society-Wolfson Merit Award Holder.Optical spectroscopy and imaging methods have proved to have potential to discriminate between normal and abnormal tissue types through minimally invasive procedures. Raman spectroscopy and Optical Coherence Tomography (OCT) provides chemical and morphological information of tissues respectively, which are complementary to each other. When used individually they might not be able to obtain high enough sensitivity and specificity that is clinically relevant. In this study we combined Raman spectroscopy information with information obtained from OCT to enhance the sensitivity and specificity in discriminating between Colonic Adenocarcinoma from Normal Colon. OCT being an imaging technique, the information from this technique is conventionally analyzed qualitatively. To combine with Raman spectroscopy information, it was essential to quantify the morphological information obtained from OCT. Texture analysis was used to extract information from OCT images, which in-turn was combined with the information obtained from Raman spectroscopy. The sensitivity and specificity of the classifier was estimated using leave one out cross validation (LOOCV) method where support vector machine (SVM) was used for binary classification of the tissues. The sensitivity obtained using Raman spectroscopy and OCT individually was 89% and 78% respectively and the specificity was 77% and 74% respectively. Combining the information derived using the two techniques increased both sensitivity and specificity to 94% demonstrating that combining complementary optical information enhances diagnostic accuracy. These results demonstrate that a multimodal approach using Raman-OCT would be able to enhance the diagnostic accuracy for identifying normal and cancerous tissue types.Publisher PD

    MELINDA: A Multimodal Dataset for Biomedical Experiment Method Classification

    Full text link
    We introduce a new dataset, MELINDA, for Multimodal biomEdicaL experImeNt methoD clAssification. The dataset is collected in a fully automated distant supervision manner, where the labels are obtained from an existing curated database, and the actual contents are extracted from papers associated with each of the records in the database. We benchmark various state-of-the-art NLP and computer vision models, including unimodal models which only take either caption texts or images as inputs, and multimodal models. Extensive experiments and analysis show that multimodal models, despite outperforming unimodal ones, still need improvements especially on a less-supervised way of grounding visual concepts with languages, and better transferability to low resource domains. We release our dataset and the benchmarks to facilitate future research in multimodal learning, especially to motivate targeted improvements for applications in scientific domains.Comment: In The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21), 202

    On Interpretability of Deep Learning based Skin Lesion Classifiers using Concept Activation Vectors

    Full text link
    Deep learning based medical image classifiers have shown remarkable prowess in various application areas like ophthalmology, dermatology, pathology, and radiology. However, the acceptance of these Computer-Aided Diagnosis (CAD) systems in real clinical setups is severely limited primarily because their decision-making process remains largely obscure. This work aims at elucidating a deep learning based medical image classifier by verifying that the model learns and utilizes similar disease-related concepts as described and employed by dermatologists. We used a well-trained and high performing neural network developed by REasoning for COmplex Data (RECOD) Lab for classification of three skin tumours, i.e. Melanocytic Naevi, Melanoma and Seborrheic Keratosis and performed a detailed analysis on its latent space. Two well established and publicly available skin disease datasets, PH2 and derm7pt, are used for experimentation. Human understandable concepts are mapped to RECOD image classification model with the help of Concept Activation Vectors (CAVs), introducing a novel training and significance testing paradigm for CAVs. Our results on an independent evaluation set clearly shows that the classifier learns and encodes human understandable concepts in its latent representation. Additionally, TCAV scores (Testing with CAVs) suggest that the neural network indeed makes use of disease-related concepts in the correct way when making predictions. We anticipate that this work can not only increase confidence of medical practitioners on CAD but also serve as a stepping stone for further development of CAV-based neural network interpretation methods.Comment: Accepted for the IEEE International Joint Conference on Neural Networks (IJCNN) 202

    Eye-CU: Sleep Pose Classification for Healthcare using Multimodal Multiview Data

    Full text link
    Manual analysis of body poses of bed-ridden patients requires staff to continuously track and record patient poses. Two limitations in the dissemination of pose-related therapies are scarce human resources and unreliable automated systems. This work addresses these issues by introducing a new method and a new system for robust automated classification of sleep poses in an Intensive Care Unit (ICU) environment. The new method, coupled-constrained Least-Squares (cc-LS), uses multimodal and multiview (MM) data and finds the set of modality trust values that minimizes the difference between expected and estimated labels. The new system, Eye-CU, is an affordable multi-sensor modular system for unobtrusive data collection and analysis in healthcare. Experimental results indicate that the performance of cc-LS matches the performance of existing methods in ideal scenarios. This method outperforms the latest techniques in challenging scenarios by 13% for those with poor illumination and by 70% for those with both poor illumination and occlusions. Results also show that a reduced Eye-CU configuration can classify poses without pressure information with only a slight drop in its performance.Comment: Ten-page manuscript including references and ten figure
    • …
    corecore