559 research outputs found

    Ultrasound-Augmented Laparoscopy

    Get PDF
    Laparoscopic surgery is perhaps the most common minimally invasive procedure for many diseases in the abdomen. Since the laparoscopic camera provides only the surface view of the internal organs, in many procedures, surgeons use laparoscopic ultrasound (LUS) to visualize deep-seated surgical targets. Conventionally, the 2D LUS image is visualized in a display spatially separate from that displays the laparoscopic video. Therefore, reasoning about the geometry of hidden targets requires mentally solving the spatial alignment, and resolving the modality differences, which is cognitively very challenging. Moreover, the mental representation of hidden targets in space acquired through such cognitive medication may be error prone, and cause incorrect actions to be performed. To remedy this, advanced visualization strategies are required where the US information is visualized in the context of the laparoscopic video. To this end, efficient computational methods are required to accurately align the US image coordinate system with that centred in the camera, and to render the registered image information in the context of the camera such that surgeons perceive the geometry of hidden targets accurately. In this thesis, such a visualization pipeline is described. A novel method to register US images with a camera centric coordinate system is detailed with an experimental investigation into its accuracy bounds. An improved method to blend US information with the surface view is also presented with an experimental investigation into the accuracy of perception of the target locations in space

    A novel NMF-based DWI CAD framework for prostate cancer.

    Get PDF
    In this thesis, a computer aided diagnostic (CAD) framework for detecting prostate cancer in DWI data is proposed. The proposed CAD method consists of two frameworks that use nonnegative matrix factorization (NMF) to learn meaningful features from sets of high-dimensional data. The first technique, is a three dimensional (3D) level-set DWI prostate segmentation algorithm guided by a novel probabilistic speed function. This speed function is driven by the features learned by NMF from 3D appearance, shape, and spatial data. The second technique, is a probabilistic classifier that seeks to label a prostate segmented from DWI data as either alignat, contain cancer, or benign, containing no cancer. This approach uses a NMF-based feature fusion to create a feature space where data classes are clustered. In addition, using DWI data acquired at a wide range of b-values (i.e. magnetic field strengths) is investigated. Experimental analysis indicates that for both of these frameworks, using NMF producing more accurate segmentation and classification results, respectively, and that combining the information from DWI data at several b-values can assist in detecting prostate cancer

    Exploring variability in medical imaging

    Get PDF
    Although recent successes of deep learning and novel machine learning techniques improved the perfor- mance of classification and (anomaly) detection in computer vision problems, the application of these methods in medical imaging pipeline remains a very challenging task. One of the main reasons for this is the amount of variability that is encountered and encapsulated in human anatomy and subsequently reflected in medical images. This fundamental factor impacts most stages in modern medical imaging processing pipelines. Variability of human anatomy makes it virtually impossible to build large datasets for each disease with labels and annotation for fully supervised machine learning. An efficient way to cope with this is to try and learn only from normal samples. Such data is much easier to collect. A case study of such an automatic anomaly detection system based on normative learning is presented in this work. We present a framework for detecting fetal cardiac anomalies during ultrasound screening using generative models, which are trained only utilising normal/healthy subjects. However, despite the significant improvement in automatic abnormality detection systems, clinical routine continues to rely exclusively on the contribution of overburdened medical experts to diagnosis and localise abnormalities. Integrating human expert knowledge into the medical imaging processing pipeline entails uncertainty which is mainly correlated with inter-observer variability. From the per- spective of building an automated medical imaging system, it is still an open issue, to what extent this kind of variability and the resulting uncertainty are introduced during the training of a model and how it affects the final performance of the task. Consequently, it is very important to explore the effect of inter-observer variability both, on the reliable estimation of model’s uncertainty, as well as on the model’s performance in a specific machine learning task. A thorough investigation of this issue is presented in this work by leveraging automated estimates for machine learning model uncertainty, inter-observer variability and segmentation task performance in lung CT scan images. Finally, a presentation of an overview of the existing anomaly detection methods in medical imaging was attempted. This state-of-the-art survey includes both conventional pattern recognition methods and deep learning based methods. It is one of the first literature surveys attempted in the specific research area.Open Acces

    A non-invasive image based system for early diagnosis of prostate cancer.

    Get PDF
    Prostate cancer is the second most fatal cancer experienced by American males. The average American male has a 16.15% chance of developing prostate cancer, which is 8.38% higher than lung cancer, the second most likely cancer. The current in-vitro techniques that are based on analyzing a patients blood and urine have several limitations concerning their accuracy. In addition, the prostate Specific Antigen (PSA) blood-based test, has a high chance of false positive diagnosis, ranging from 28%-58%. Yet, biopsy remains the gold standard for the assessment of prostate cancer, but only as the last resort because of its invasive nature, high cost, and potential morbidity rates. The major limitation of the relatively small needle biopsy samples is the higher possibility of producing false positive diagnosis. Moreover, the visual inspection system (e.g., Gleason grading system) is not quantitative technique and different observers may classify a sample differently, leading to discrepancies in the diagnosis. As reported in the literature that the early detection of prostate cancer is a crucial step for decreasing prostate cancer related deaths. Thus, there is an urgent need for developing objective, non-invasive image based technology for early detection of prostate cancer. The objective of this dissertation is to develop a computer vision methodology, later translated into a clinically usable software tool, which can improve sensitivity and specificity of early prostate cancer diagnosis based on the well-known hypothesis that malignant tumors are will connected with the blood vessels than the benign tumors. Therefore, using either Diffusion Weighted Magnetic Resonance imaging (DW-MRI) or Dynamic Contrast Enhanced Magnetic Resonance Imaging (DCE-MRI), we will be able to interrelate the amount of blood in the detected prostate tumors by estimating either the Apparent Diffusion Coefficient (ADC) in the prostate with the malignancy of the prostate tumor or perfusion parameters. We intend to validate this hypothesis by demonstrating that automatic segmentation of the prostate from either DW-MRI or DCE-MRI after handling its local motion, provides discriminatory features for early prostate cancer diagnosis. The proposed CAD system consists of three majors components, the first two of which constitute new research contributions to a challenging computer vision problem. The three main components are: (1) A novel Shape-based segmentation approach to segment the prostate from either low contrast DW-MRI or DCE-MRI data; (2) A novel iso-contours-based non-rigid registration approach to ensure that we have voxel-on-voxel matches of all data which may be more difficult due to gross patient motion, transmitted respiratory effects, and intrinsic and transmitted pulsatile effects; and (3) Probabilistic models for the estimated diffusion and perfusion features for both malignant and benign tumors. Our results showed a 98% classification accuracy using Leave-One-Subject-Out (LOSO) approach based on the estimated ADC for 30 patients (12 patients diagnosed as malignant; 18 diagnosed as benign). These results show the promise of the proposed image-based diagnostic technique as a supplement to current technologies for diagnosing prostate cancer

    Sparse feature learning for image analysis in segmentation, classification, and disease diagnosis.

    Get PDF
    The success of machine learning algorithms generally depends on intermediate data representation, called features that disentangle the hidden factors of variation in data. Moreover, machine learning models are required to be generalized, in order to reduce the specificity or bias toward the training dataset. Unsupervised feature learning is useful in taking advantage of large amount of unlabeled data, which is available to capture these variations. However, learned features are required to capture variational patterns in data space. In this dissertation, unsupervised feature learning with sparsity is investigated for sparse and local feature extraction with application to lung segmentation, interpretable deep models, and Alzheimer\u27s disease classification. Nonnegative Matrix Factorization, Autoencoder and 3D Convolutional Autoencoder are used as architectures or models for unsupervised feature learning. They are investigated along with nonnegativity, sparsity and part-based representation constraints for generalized and transferable feature extraction

    A Variational Approach to Multi-Modal Image Matching

    Get PDF
    We address the problem of non-parametric multi-modal image matching. We propose a generic framework which relies on a global variational formulation and show its versatility through three different multi-modal registration methods : supervised registration by joint intensity learning, maximization of the mutual information and maximization of the correlation ratio. Regulariz- ation is performed by using a functional borrowed from linear elasticity theory. We also consider a geometry-driven regularization method. Experiments on synthetic images and preliminary results on the realignment of MRI datasets are presented

    Cancer diagnosis using deep learning: A bibliographic review

    Get PDF
    In this paper, we first describe the basics of the field of cancer diagnosis, which includes steps of cancer diagnosis followed by the typical classification methods used by doctors, providing a historical idea of cancer classification techniques to the readers. These methods include Asymmetry, Border, Color and Diameter (ABCD) method, seven-point detection method, Menzies method, and pattern analysis. They are used regularly by doctors for cancer diagnosis, although they are not considered very efficient for obtaining better performance. Moreover, considering all types of audience, the basic evaluation criteria are also discussed. The criteria include the receiver operating characteristic curve (ROC curve), Area under the ROC curve (AUC), F1 score, accuracy, specificity, sensitivity, precision, dice-coefficient, average accuracy, and Jaccard index. Previously used methods are considered inefficient, asking for better and smarter methods for cancer diagnosis. Artificial intelligence and cancer diagnosis are gaining attention as a way to define better diagnostic tools. In particular, deep neural networks can be successfully used for intelligent image analysis. The basic framework of how this machine learning works on medical imaging is provided in this study, i.e., pre-processing, image segmentation and post-processing. The second part of this manuscript describes the different deep learning techniques, such as convolutional neural networks (CNNs), generative adversarial models (GANs), deep autoencoders (DANs), restricted Boltzmann’s machine (RBM), stacked autoencoders (SAE), convolutional autoencoders (CAE), recurrent neural networks (RNNs), long short-term memory (LTSM), multi-scale convolutional neural network (M-CNN), multi-instance learning convolutional neural network (MIL-CNN). For each technique, we provide Python codes, to allow interested readers to experiment with the cited algorithms on their own diagnostic problems. The third part of this manuscript compiles the successfully applied deep learning models for different types of cancers. Considering the length of the manuscript, we restrict ourselves to the discussion of breast cancer, lung cancer, brain cancer, and skin cancer. The purpose of this bibliographic review is to provide researchers opting to work in implementing deep learning and artificial neural networks for cancer diagnosis a knowledge from scratch of the state-of-the-art achievements

    Automation of Hessian-Based Tubularity Measure Response Function in 3D Biomedical Images

    Get PDF
    The blood vessels and nerve trees consist of tubular objects interconnected into a complex tree- or web-like structure that has a range of structural scale 5 μm diameter capillaries to 3 cm aorta. This large-scale range presents two major problems; one is just making the measurements, and the other is the exponential increase of component numbers with decreasing scale. With the remarkable increase in the volume imaged by, and resolution of, modern day 3D imagers, it is almost impossible to make manual tracking of the complex multiscale parameters from those large image data sets. In addition, the manual tracking is quite subjective and unreliable. We propose a solution for automation of an adaptive nonsupervised system for tracking tubular objects based on multiscale framework and use of Hessian-based object shape detector incorporating National Library of Medicine Insight Segmentation and Registration Toolkit (ITK) image processing libraries
    corecore