30 research outputs found

    Registration of retinal images from Public Health by minimising an error between vessels using an affine model with radial distortions

    Get PDF
    In order to estimate a registration model of eye fundus images made of an affinity and two radial distortions, we introduce an estimation criterion based on an error between the vessels. In [1], we estimated this model by minimising the error between characteristics points. In this paper, the detected vessels are selected using the circle and ellipse equations of the overlap area boundaries deduced from our model. Our method successfully registers 96 % of the 271 pairs in a Public Health dataset acquired mostly with different cameras. This is better than our previous method [1] and better than three other state-of-the-art methods. On a publicly available dataset, ours still better register the images than the reference method

    Superimposition of eye fundus images for longitudinal analysis from large public health databases

    Get PDF
    In this paper, a method is presented for superimposition (i.e. registration) of eye fundus images from persons with diabetes screened over many years for diabetic retinopathy. The method is fully automatic and robust to camera changes and colour variations across the images both in space and time. All the stages of the process are designed for longitudinal analysis of cohort public health databases where retinal examinations are made at approximately yearly intervals. The method relies on a model correcting two radial distortions and an affine transformation between pairs of images which is robustly fitted on salient points. Each stage involves linear estimators followed by non-linear optimisation. The model of image warping is also invertible for fast computation. The method has been validated (1) on a simulated montage and (2) on public health databases with 69 patients with high quality images (271 pairs acquired mostly with different types of camera and 268 pairs acquired mostly with the same type of camera) with success rates of 92% and 98%, and five patients (20 pairs) with low quality images with a success rate of 100%. Compared to two state-of-the-art methods, ours gives better results.Comment: This is an author-created, un-copyedited version of an article published in Biomedical Physics \& Engineering Express. IOP Publishing Ltd is not responsible for any errors or omissions in this version of the manuscript or any version derived from it. The Version of Record is available online at https://doi.org/10.1088/2057-1976/aa7d1

    Characterising pattern asymmetry in pigmented skin lesions

    Get PDF
    Abstract. In clinical diagnosis of pigmented skin lesions asymmetric pigmentation is often indicative of melanoma. This paper describes a method and measures for characterizing lesion symmetry. The estimate of mirror symmetry is computed first for a number of axes at different degrees of rotation with respect to the lesion centre. The statistics of these estimates are the used to assess the overall symmetry. The method is applied to three different lesion representations showing the overall pigmentation, the pigmentation pattern, and the pattern of dermal melanin. The best measure is a 100% sensitive and 96% specific indicator of melanoma on a test set of 33 lesions, with a separate training set consisting of 66 lesions

    Handbook of Vascular Biometrics

    Get PDF

    Handbook of Vascular Biometrics

    Get PDF
    This open access handbook provides the first comprehensive overview of biometrics exploiting the shape of human blood vessels for biometric recognition, i.e. vascular biometrics, including finger vein recognition, hand/palm vein recognition, retina recognition, and sclera recognition. After an introductory chapter summarizing the state of the art in and availability of commercial systems and open datasets/open source software, individual chapters focus on specific aspects of one of the biometric modalities, including questions of usability, security, and privacy. The book features contributions from both academia and major industrial manufacturers

    3D-3D Deformable Registration and Deep Learning Segmentation based Neck Diseases Analysis in MRI

    Full text link
    Whiplash, cervical dystonia (CD), neck pain and work-related upper limb disorder (WRULD) are the most common diseases in the cervical region. Headaches, stiffness, sensory disturbance to the legs and arms, optical problems, aching in the back and shoulder, and auditory and visual problems are common symptoms seen in patients with these diseases. CD patients may also suffer tormenting spasticity in some neck muscles, with the symptoms possibly being acute and persisting for a long time, sometimes a lifetime. Whiplash-associated disorders (WADs) may occur due to sudden forward and backward movements of the head and neck occurring during a sporting activity or vehicle or domestic accident. These diseases affect private industries, insurance companies and governments, with the socio-economic costs significantly related to work absences, long-term sick leave, early disability and disability support pensions, health care expenses, reduced productivity and insurance claims. Therefore, diagnosing and treating neck-related diseases are important issues in clinical practice. The reason for these afflictions resulting from accident is the impairment of the cervical muscles which undergo atrophy or pseudo-hypertrophy due to fat infiltrating into them. These morphological changes have to be determined by identifying and quantifying their bio-markers before applying any medical intervention. Volumetric studies of neck muscles are reliable indicators of the proper treatments to apply. Radiation therapy, chemotherapy, injection of a toxin or surgery could be possible ways of treating these diseases. However, the dosages required should be precise because the neck region contains some sensitive organs, such as nerves, blood vessels and the trachea and spinal cord. Image registration and deep learning-based segmentation can help to determine appropriate treatments by analyzing the neck muscles. However, this is a challenging task for medical images due to complexities such as many muscles crossing multiple joints and attaching to many bones. Also, their shapes and sizes vary greatly across populations whereas their cross-sectional areas (CSAs) do not change in proportion to the heights and weights of individuals, with their sizes varying more significantly between males and females than ages. Therefore, the neck's anatomical variabilities are much greater than those of other parts of the human body. Some other challenges which make analyzing neck muscles very difficult are their compactness, similar gray-level appearances, intra-muscular fat, sliding due to cardiac and respiratory motions, false boundaries created by intramuscular fat, low resolution and contrast in medical images, noise, inhomogeneity and background clutter with the same composition and intensity. Furthermore, a patient's mode, position and neck movements during the capture of an image create variability. However, very little significant research work has been conducted on analyzing neck muscles. Although previous image registration efforts form a strong basis for many medical applications, none can satisfy the requirements of all of them because of the challenges associated with their implementation and low accuracy which could be due to anatomical complexities and variabilities or the artefacts of imaging devices. In existing methods, multi-resolution- and heuristic-based methods are popular. However, the above issues cause conventional multi-resolution-based registration methods to be trapped in local minima due to their low degrees of freedom in their geometrical transforms. Although heuristic-based methods are good at handling large mismatches, they require pre-segmentation and are computationally expensive. Also, current deformable methods often face statistical instability problems and many local optima when dealing with small mismatches. On the other hand, deep learning-based methods have achieved significant success over the last few years. Although a deeper network can learn more complex features and yields better performances, its depth cannot be increased as this would cause the gradient to vanish during training and result in training difficulties. Recently, researchers have focused on attention mechanisms for deep learning but current attention models face a challenge in the case of an application with compact and similar small multiple classes, large variability, low contrast and noise. The focus of this dissertation is on the design of 3D-3D image registration approaches as well as deep learning-based semantic segmentation methods for analyzing neck muscles. In the first part of this thesis, a novel object-constrained hierarchical registration framework for aligning inter-subject neck muscles is proposed. Firstly, to handle large-scale local minima, it uses a coarse registration technique which optimizes a new edge position difference (EPD) similarity measure to align large mismatches. Also, a new transformation based on the discrete periodic spline wavelet (DPSW), affine and free-form-deformation (FFD) transformations are exploited. Secondly, to avoid the monotonous nature of using transformations in multiple stages, affine registration technique, which uses a double-pushing system by changing the edges in the EPD and switching the transformation's resolutions, is designed to align small mismatches. The EPD helps in both the coarse and fine techniques to implement object-constrained registration via controlling edges which is not possible using traditional similarity measures. Experiments are performed on clinical 3D magnetic resonance imaging (MRI) scans of the neck, with the results showing that the EPD is more effective than the mutual information (MI) and the sum of squared difference (SSD) measures in terms of the volumetric dice similarity coefficient (DSC). Also, the proposed method is compared with two state-of-the-art approaches with ablation studies of inter-subject deformable registration and achieves better accuracy, robustness and consistency. However, as this method is computationally complex and has a problem handling large-scale anatomical variabilities, another 3D-3D registration framework with two novel contributions is proposed in the second part of this thesis. Firstly, a two-stage heuristic search optimization technique for handling large mismatches,which uses a minimal user hypothesis regarding these mismatches and is computationally fast, is introduced. It brings a moving image hierarchically closer to a fixed one using MI and EPD similarity measures in the coarse and fine stages, respectively, while the images do not require pre-segmentation as is necessary in traditional heuristic optimization-based techniques. Secondly, a region of interest (ROI) EPD-based registration framework for handling small mismatches using salient anatomical information (AI), in which a convex objective function is formed through a unique shape created from the desired objects in the ROI, is proposed. It is compared with two state-of-the-art methods on a neck dataset, with the results showing that it is superior in terms of accuracy and is computationally fast. In the last part of this thesis, an evaluation study of recent U-Net-based convolutional neural networks (CNNs) is performed on a neck dataset. It comprises 6 recent models, the U-Net, U-Net with a conditional random field (CRF-Unet), attention U-Net (A-Unet), nested U-Net or U-Net++, multi-feature pyramid (MFP)-Unet and recurrent residual U-Net (R2Unet) and 4 with more comprehensive modifications, the multi-scale U-Net (MS-Unet), parallel multi-scale U-Net (PMSUnet), recurrent residual attention U-Net (R2A-Unet) and R2A-Unet++ in neck muscles segmentation, with analyses of the numerical results indicating that the R2Unet architecture achieves the best accuracy. Also, two deep learning-based semantic segmentation approaches are proposed. In the first, a new two-stage U-Net++ (TS-UNet++) uses two different types of deep CNNs (DCNNs) rather than one similar to the traditional multi-stage method, with the U-Net++ in the first stage and the U-Net in the second. More convolutional blocks are added after the input and before the output layers of the multi-stage approach to better extract the low- and high-level features. A new concatenation-based fusion structure, which is incorporated in the architecture to allow deep supervision, helps to increase the depth of the network without accelerating the gradient-vanishing problem. Then, more convolutional layers are added after each concatenation of the fusion structure to extract more representative features. The proposed network is compared with the U-Net, U-Net++ and two-stage U-Net (TS-UNet) on the neck dataset, with the results indicating that it outperforms the others. In the second approach, an explicit attention method, in which the attention is performed through a ROI evolved from ground truth via dilation, is proposed. It does not require any additional CNN, as does a cascaded approach, to localize the ROI. Attention in a CNN is sensitive with respect to the area of the ROI. This dilated ROI is more capable of capturing relevant regions and suppressing irrelevant ones than a bounding box and region-level coarse annotation, and is used during training of any CNN. Coarse annotation, which does not require any detailed pixel wise delineation that can be performed by any novice person, is used during testing. This proposed ROI-based attention method, which can handle compact and similar small multiple classes with objects with large variabilities, is compared with the automatic A-Unet and U-Net, and performs best

    Automatic segmentation of anterior segment optical coherence tomography images

    Get PDF
    Automatic segmentation of anterior segment optical coherence tomography (AS OCT) images provides an important tool to aid management of ocular diseases. Having precise details about the topography and thickness of an individual eye enables treatments to be tailored to a specific problem. OCT is an imaging technique that can be used to acquire volumetric data of the anterior segment of the human eye. Fast automatic segmentation of this data, which is not available, means clinically useful information can be obtained without the need for time consuming error-prone manual analysis of the images. This thesis presents newly developed automatic segmentation techniques of OCT images. Segmentation of 2D OCT images is first performed. One of the main challenges segmenting 2D OCT images is the presence of regions of the image that generally have a low signal to noise ratio. This is overcome by the use of shape based terms. A number of different methods, such as level set, graph cut, and graph theory, are developed to do this. The segmentation techniques are validated by comparison to expert manual segmentation and previously published segmentation techniques. The best method, graph theory with shape, was able to achieve segmentation comparable to manual segmentation. Good agreement is found with manual segmentation for the best 2D segmentation method, graph theory with shape, achieving a Dice similarity coefficient of 0.96, which is comparable to inter-observer agreement. It performed significantly better than previously published techniques. The 2D segmentation techniques are then extended to 3D segmentation of OCT images. The challenge here is motion artefact or poor alignment between each 2D images comprising the 3D images. Different segmentation strategies are investigated including direct segmentation by level set or graph cut approaches, and segmentation with registration. In particular the latter requires the introduction of a registration step to align multiple 2D images to produce a 3D representation to overcome the presence of involuntary motion artefacts. This method produces the best performance. In particular, it uses graph theory and dynamic programming to segment the anterior and posterior surfaces in individual 2D images with shape constraint. Genetic algorithms are then used to align 2D images to produce a full 3D representation of the anterior segment based on landmarks or geometric constraints. For the 3D segmentation, a data set of 17 eyes is used for validation. These have each been imaged twice so a repeatability measurement can be made. Good repeatability of results is demonstrated with the 3D alignment method. A mean difference of 1.77 pixels is found between the same surfaces of the repeated scans of the same eye. Overall, a new automation method is developed that can produce maps of the anterior and posterior surfaces of the cornea from a 3D images of the anterior segment of a human eye. This will be a valuable tool that can be used for patient specific biomechanical modelling of the human eye

    Image Registration Workshop Proceedings

    Get PDF
    Automatic image registration has often been considered as a preliminary step for higher-level processing, such as object recognition or data fusion. But with the unprecedented amounts of data which are being and will continue to be generated by newly developed sensors, the very topic of automatic image registration has become and important research topic. This workshop presents a collection of very high quality work which has been grouped in four main areas: (1) theoretical aspects of image registration; (2) applications to satellite imagery; (3) applications to medical imagery; and (4) image registration for computer vision research

    Neuroimaging Biomarkers in Paediatric Sickle Cell Disease

    Get PDF
    Sickle Cell Disease (SCD) is a collection of genetic haemoglobinopathies, the most common and severe being homozygous sickle cell anaemia. In the UK, it has been estimated that 1 in 2000 children are born with SCD. The disease is characterised by chronic anaemia, recurrent pain crises and vascular occlusion. Neurologically, there is a high incidence of stroke in childhood, as well as cognitive dysfunction. Newborn screening programmes and preventative treatments have allowed a much longer lifespan; however recently, neurological research has shifted to characterising subtler aspects of brain development and functioning that may be critically important to the individual’s quality of life. This thesis overviews the neurological and neurocognitive complications of SCD, and how magnetic resonance imaging (MRI) can provide biomarkers for severity of disease. During the PhD, retrospective and prospective cognitive and MRI data were collected and analysed. Diagnostic clinical MRI sequences and advanced MRI sequences were applied, as well as a neuropsychological test battery aimed at intelligence and executive function. First, this thesis reviews the intelligence literature in SCD and includes previously unreported data, finding patients, regardless of abnormality seen on conventional MRI, have lowered full-scale intelligence quotient than controls. Then, to determine imaging biomarkers, volumetric differences and diffusion characteristics were identified. Patients were found to have decreased volumes of subcortical structures compared to controls, in groups corresponding to disease severity. Results from a three-year longitudinal clinical trial suggest evidence of atrophy in paediatric patients, with no apparent protective effect of treatment. Diffusion tensor imaging revealed reduced white matter integrity across the brain, correlating with recognised markers of disease severity (i.e. oxygen saturation and haemoglobin from a full blood count). Overall, the four experiments bridge a gap in the cognitive and neuroimaging literature of the extent of neurological injury in children with SCD
    corecore