57 research outputs found

    Multi-scale and multi-spectral shape analysis: from 2d to 3d

    Get PDF
    Shape analysis is a fundamental aspect of many problems in computer graphics and computer vision, including shape matching, shape registration, object recognition and classification. Since the SIFT achieves excellent matching results in 2D image domain, it inspires us to convert the 3D shape analysis to 2D image analysis using geometric maps. However, the major disadvantage of geometric maps is that it introduces inevitable, large distortions when mapping large, complex and topologically complicated surfaces to a canonical domain. It is demanded for the researchers to construct the scale space directly on the 3D shape. To address these research issues, in this dissertation, in order to find the multiscale processing for the 3D shape, we start with shape vector image diffusion framework using the geometric mapping. Subsequently, we investigate the shape spectrum field by introducing the implementation and application of Laplacian shape spectrum. In order to construct the scale space on 3D shape directly, we present a novel idea to solve the diffusion equation using the manifold harmonics in the spectral point of view. Not only confined on the mesh, by using the point-based manifold harmonics, we rigorously derive our solution from the diffusion equation which is the essential of the scale space processing on the manifold. Built upon the point-based manifold harmonics transform, we generalize the diffusion function directly on the point clouds to create the scale space. In virtue of the multiscale structure from the scale space, we can detect the feature points and construct the descriptor based on the local neighborhood. As a result, multiscale shape analysis directly on the 3D shape can be achieved

    Multidimensional image analysis of cardiac function in MRI

    Get PDF
    Cardiac morphology is a key indicator of cardiac health. Important metrics that are currently in clinical use are left-ventricle cardiac ejection fraction, cardiac muscle (myocardium) mass, myocardium thickness and myocardium thickening over the cardiac cycle. Advances in imaging technologies have led to an increase in temporal and spatial resolution. Such an increase in data presents a laborious task for medical practitioners to analyse. In this thesis, measurement of the cardiac left-ventricle function is achieved by developing novel methods for the automatic segmentation of the left-ventricle blood-pool and the left ventricle myocardium boundaries. A preliminary challenge faced in this task is the removal of noise from Magnetic Resonance Imaging (MRI) data, which is addressed by using advanced data filtering procedures. Two mechanisms for left-ventricle segmentation are employed. Firstly segmentation of the left ventricle blood-pool for the measurement of ejection fraction is undertaken in the signal intensity domain. Utilising the high discrimination between blood and tissue, a novel methodology based on a statistical partitioning method offers success in localising and segmenting the blood pool of the left ventricle. From this initialisation, the estimation of the outer wall (epi-cardium) of the left ventricle can be achieved using gradient information and prior knowledge. Secondly, a more involved method for extracting the myocardium of the leftventricle is developed, that can better perform segmentation in higher dimensions. Spatial information is incorporated in the segmentation by employing a gradient-based boundary evolution. A level-set scheme is implemented and a novel formulation for the extraction of the cardiac muscle is introduced. Two surfaces, representing the inner and the outer boundaries of the left-ventricle, are simultaneously evolved using a coupling function and supervised with a probabilistic model of expertly assisted manual segmentations

    Application of diffusion techniques to the segmentation of Mr 3D images for virtual colonoscopy

    Get PDF
    Master'sMASTER OF ENGINEERIN

    Feature extraction to aid disease detection and assessment of disease progression in CT and MR colonography

    Get PDF
    Computed tomographic colonography (CTC) is a technique employed to examine the whole colon for cancers and premalignant adenomas (polyps). Oral preparation is taken to fully cleanse the colon, and gas insufflation maximises the attenuation contrast between the enoluminal colon surface and the lumen. The procedure is performed routinely with the patient both prone and supine to redistribute gas and residue. This helps to differentiate fixed colonic pathology from mobile faecal residue and also helps discover pathology occluded by retained fluid or luminal collapse. Matching corresponding endoluminal surface locations with the patient in the prone and supine positions is therefore an essential aspect of interpretation by radiologists; however, interpretation can be difficult and time consuming due to the considerable colonic deformations that occur during repositioning. Hence, a method for automated registration has the potential to improve efficiency and diagnostic accuracy. I propose a novel method to establish correspondence between prone and supine CT colonography acquisitions automatically. The problem is first simplified by detecting haustral folds which are elongated ridgelike endoluminal structures and can be identified by curvature based measurements. These are subsequently matched using appearance based features, and their relative geometric relationships. It is shown that these matches can be used to find correspondence along the full length of the colon, but may also be used in conjunction with other registration methods to achieve a more robust and accurate result, explicitly addressing the problem of colonic collapse. The potential clinical value of this method has been assessed in an external clinical validation, and the application to follow-up CTC surveillance has been investigated. MRI has recently been applied as a tool to quantitatively evaluate the therapeutic response to therapy in patients with Crohn's disease, and is the preferred choice for repeated imaging. A primary biomarker for this evaluation is the measurement of variations of bowel wall thickness on changing from the active phase of the disease to remission; however, a poor level of interobserver agreement of measured thickness is reported and therefore a system for accurate, robust and reproducible measurements is desirable. I propose a novel method which will automatically track sections of colon, by estimating the positions of elliptical cross sections. Subsequently, estimation of the positions of the inner and outer bowel walls are made based on image gradient information and therefore a thickness measurement value can be extracted

    Surgical Subtask Automation for Intraluminal Procedures using Deep Reinforcement Learning

    Get PDF
    Intraluminal procedures have opened up a new sub-field of minimally invasive surgery that use flexible instruments to navigate through complex luminal structures of the body, resulting in reduced invasiveness and improved patient benefits. One of the major challenges in this field is the accurate and precise control of the instrument inside the human body. Robotics has emerged as a promising solution to this problem. However, to achieve successful robotic intraluminal interventions, the control of the instrument needs to be automated to a large extent. The thesis first examines the state-of-the-art in intraluminal surgical robotics and identifies the key challenges in this field, which include the need for safe and effective tool manipulation, and the ability to adapt to unexpected changes in the luminal environment. To address these challenges, the thesis proposes several levels of autonomy that enable the robotic system to perform individual subtasks autonomously, while still allowing the surgeon to retain overall control of the procedure. The approach facilitates the development of specialized algorithms such as Deep Reinforcement Learning (DRL) for subtasks like navigation and tissue manipulation to produce robust surgical gestures. Additionally, the thesis proposes a safety framework that provides formal guarantees to prevent risky actions. The presented approaches are evaluated through a series of experiments using simulation and robotic platforms. The experiments demonstrate that subtask automation can improve the accuracy and efficiency of tool positioning and tissue manipulation, while also reducing the cognitive load on the surgeon. The results of this research have the potential to improve the reliability and safety of intraluminal surgical interventions, ultimately leading to better outcomes for patients and surgeons

    Registration of prone and supine CT colonography images and its clinical application

    Get PDF
    Computed tomographic (CT) colonography is a technique for detecting bowel cancer and potentially precancerous polyps. CT imaging is performed on the cleansed and insufflated bowel in order to produce a virtual endoluminal representation similar to optical colonoscopy. Because fluids and stool can mimic pathology, images are acquired with the patient in both prone and supine positions. Radiologists then match endoluminal locations visually between the two acquisitions in order to determine whether pathology is real or not. This process is hindered by the fact that the colon can undergo considerable deformation between acquisitions. Robust and accurate automated registration between prone and supine data acquisitions is therefore pivotal for medical interpretation, but a challenging problem. The method proposed in this thesis reduces the complexity of the registration task of aligning the prone and supine CT colonography acquisitions. This is done by utilising cylindrical representations of the colonic surface which reflect the colon's specific anatomy. Automated alignment in the cylindrical domain is achieved by non-rigid image registration using surface curvatures, applicable even when cases exhibit local luminal collapses. It is furthermore shown that landmark matches for initialisation improve the registration's accuracy and robustness. Additional performance improvements are achieved by symmetric and inverse-consistent registration and iteratively deforming the surface in order to compensate for differences in distension and bowel preparation. Manually identified reference points in human data and fiducial markers in a porcine phantom are used to validate the registration accuracy. The potential clinical impact of the method has been evaluated using data that reflects clinical practise. Furthermore, correspondence between follow-up CT colonography acquisitions is established in order to facilitate the clinical need to investigate polyp growth over time. Accurate registration has the potential to both improve the diagnostic process and decrease the radiologist's interpretation time. Furthermore, its result could be integrated into algorithms for improved computer-aided detection of colonic polyps

    A framework for tumor segmentation and interactive immersive visualization of medical image data for surgical planning

    Get PDF
    This dissertation presents the framework for analyzing and visualizing digital medical images. Two new segmentation methods have been developed: a probability based segmentation algorithm, and a segmentation algorithm that uses a fuzzy rule based system to generate similarity values for segmentation. A visualization software application has also been developed to effectively view and manipulate digital medical images on a desktop computer as well as in an immersive environment.;For the probabilistic segmentation algorithm, image data are first enhanced by manually setting the appropriate window center and width, and if needed a sharpening or noise removal filter is applied. To initialize the segmentation process, a user places a seed point within the object of interest and defines a search region for segmentation. Based on the pixels\u27 spatial and intensity properties, a probabilistic selection criterion is used to extract pixels with a high probability of belonging to the object. To facilitate the segmentation of multiple slices, an automatic seed selection algorithm was developed to keep the seeds in the object as its shape and/or location changes between consecutive slices.;The second segmentation method, a new segmentation method using a fuzzy rule based system to segment tumors in a three-dimensional CT data was also developed. To initialize the segmentation process, the user selects a region of interest (ROI) within the tumor in the first image of the CT study set. Using the ROI\u27s spatial and intensity properties, fuzzy inputs are generated for use in the fuzzy rules inference system. Using a set of predefined fuzzy rules, the system generates a defuzzified output for every pixel in terms of similarity to the object. Pixels with the highest similarity values are selected as tumor. This process is automatically repeated for every subsequent slice in the CT set without further user input, as the segmented region from the previous slice is used as the ROI for the current slice. This creates a propagation of information from the previous slices, used to segment the current slice. The membership functions used during the fuzzification and defuzzification processes are adaptive to the changes in the size and pixel intensities of the current ROI. The proposed method is highly customizable to suit different needs of a user, requiring information from only a single two-dimensional image.;Segmentation results from both algorithms showed success in segmenting the tumor from seven of the ten CT datasets with less than 10% false positive errors and five test cases with less than 10% false negative errors. The consistency of the segmentation results statistics also showed a high repeatability factor, with low values of inter- and intra-user variability for both methods.;The visualization software developed is designed to load and display any DICOM/PACS compatible three-dimensional image data for visualization and interaction in an immersive virtual environment. The software uses the open-source libraries DCMTK: DICOM Toolkit for parsing of digital medical images, Coin3D and SimVoleon for scenegraph management and volume rendering, and VRJuggler for virtual reality display and interaction. A user can apply pseudo-coloring in real time with multiple interactive clipping planes to slice into the volume for an interior view. A windowing feature controls the tissue density ranges to display. A wireless gamepad controller as well as a simple and intuitive menu interface control user interactions. The software is highly scalable as it can be used on a single desktop computer to a cluster of computers for an immersive multi-projection virtual environment. By wearing a pair of stereo goggles, the surgeon is immersed within the model itself, thus providing a sense of realism as if the surgeon is inside the patient.;The tools developed in this framework are designed to improve patient care by fostering the widespread use of advanced visualization and computational intelligence in preoperative planning, surgical training, and diagnostic assistance. Future work includes further improvements to both segmentation methods with plans to incorporate the use of deformable models and level set techniques to include tumor shape features as part of the segmentation criteria. For the surgical planning components, additional controls and interactions with the simulated endoscopic camera and the ability to segment the colon or a selected region of the airway for a fixed-path navigation as a full virtual endoscopy tool will also be implemented. (Abstract shortened by UMI.

    Vision-based retargeting for endoscopic navigation

    Get PDF
    Endoscopy is a standard procedure for visualising the human gastrointestinal tract. With the advances in biophotonics, imaging techniques such as narrow band imaging, confocal laser endomicroscopy, and optical coherence tomography can be combined with normal endoscopy for assisting the early diagnosis of diseases, such as cancer. In the past decade, optical biopsy has emerged to be an effective tool for tissue analysis, allowing in vivo and in situ assessment of pathological sites with real-time feature-enhanced microscopic images. However, the non-invasive nature of optical biopsy leads to an intra-examination retargeting problem, which is associated with the difficulty of re-localising a biopsied site consistently throughout the whole examination. In addition to intra-examination retargeting, retargeting of a pathological site is even more challenging across examinations, due to tissue deformation and changing tissue morphologies and appearances. The purpose of this thesis is to address both the intra- and inter-examination retargeting problems associated with optical biopsy. We propose a novel vision-based framework for intra-examination retargeting. The proposed framework is based on combining visual tracking and detection with online learning of the appearance of the biopsied site. Furthermore, a novel cascaded detection approach based on random forests and structured support vector machines is developed to achieve efficient retargeting. To cater for reliable inter-examination retargeting, the solution provided in this thesis is achieved by solving an image retrieval problem, for which an online scene association approach is proposed to summarise an endoscopic video collected in the first examination into distinctive scenes. A hashing-based approach is then used to learn the intrinsic representations of these scenes, such that retargeting can be achieved in subsequent examinations by retrieving the relevant images using the learnt representations. For performance evaluation of the proposed frameworks, extensive phantom, ex vivo and in vivo experiments have been conducted, with results demonstrating the robustness and potential clinical values of the methods proposed.Open Acces
    corecore