49 research outputs found

    Graph-based Data Modeling and Analysis for Data Fusion in Remote Sensing

    Get PDF
    Hyperspectral imaging provides the capability of increased sensitivity and discrimination over traditional imaging methods by combining standard digital imaging with spectroscopic methods. For each individual pixel in a hyperspectral image (HSI), a continuous spectrum is sampled as the spectral reflectance/radiance signature to facilitate identification of ground cover and surface material. The abundant spectrum knowledge allows all available information from the data to be mined. The superior qualities within hyperspectral imaging allow wide applications such as mineral exploration, agriculture monitoring, and ecological surveillance, etc. The processing of massive high-dimensional HSI datasets is a challenge since many data processing techniques have a computational complexity that grows exponentially with the dimension. Besides, a HSI dataset may contain a limited number of degrees of freedom due to the high correlations between data points and among the spectra. On the other hand, merely taking advantage of the sampled spectrum of individual HSI data point may produce inaccurate results due to the mixed nature of raw HSI data, such as mixed pixels, optical interferences and etc. Fusion strategies are widely adopted in data processing to achieve better performance, especially in the field of classification and clustering. There are mainly three types of fusion strategies, namely low-level data fusion, intermediate-level feature fusion, and high-level decision fusion. Low-level data fusion combines multi-source data that is expected to be complementary or cooperative. Intermediate-level feature fusion aims at selection and combination of features to remove redundant information. Decision level fusion exploits a set of classifiers to provide more accurate results. The fusion strategies have wide applications including HSI data processing. With the fast development of multiple remote sensing modalities, e.g. Very High Resolution (VHR) optical sensors, LiDAR, etc., fusion of multi-source data can in principal produce more detailed information than each single source. On the other hand, besides the abundant spectral information contained in HSI data, features such as texture and shape may be employed to represent data points from a spatial perspective. Furthermore, feature fusion also includes the strategy of removing redundant and noisy features in the dataset. One of the major problems in machine learning and pattern recognition is to develop appropriate representations for complex nonlinear data. In HSI processing, a particular data point is usually described as a vector with coordinates corresponding to the intensities measured in the spectral bands. This vector representation permits the application of linear and nonlinear transformations with linear algebra to find an alternative representation of the data. More generally, HSI is multi-dimensional in nature and the vector representation may lose the contextual correlations. Tensor representation provides a more sophisticated modeling technique and a higher-order generalization to linear subspace analysis. In graph theory, data points can be generalized as nodes with connectivities measured from the proximity of a local neighborhood. The graph-based framework efficiently characterizes the relationships among the data and allows for convenient mathematical manipulation in many applications, such as data clustering, feature extraction, feature selection and data alignment. In this thesis, graph-based approaches applied in the field of multi-source feature and data fusion in remote sensing area are explored. We will mainly investigate the fusion of spatial, spectral and LiDAR information with linear and multilinear algebra under graph-based framework for data clustering and classification problems

    Predicting and explaining nonlinear material response using deep Physically Guided Neural Networks with Internal Variables

    Full text link
    Nonlinear materials are often difficult to model with classical state model theory because they have a complex and sometimes inaccurate physical and mathematical description or we simply do not know how to describe such materials in terms of relations between external and internal variables. In many disciplines, Neural Network methods have arisen as powerful tools to identify very complex and non-linear correlations. In this work, we use the very recently developed concept of Physically Guided Neural Networks with Internal Variables (PGNNIV) to discover constitutive laws using a model-free approach and training solely with measured force-displacement data. PGNNIVs make a particular use of the physics of the problem to enforce constraints on specific hidden layers and are able to make predictions without internal variable data. We demonstrate that PGNNIVs are capable of predicting both internal and external variables under unseen load scenarios, regardless of the nature of the material considered (linear, with hardening or softening behavior and hyperelastic), unravelling the constitutive law of the material hence explaining its nature altogether, placing the method in what is known as eXplainable Artificial Intelligence (XAI).Comment: Main text: 25 pages, 6 figures. Appendices: 13 pages, 12 figure

    Efficient sampling strategies for x-ray micro computed tomography with an intensity-modulated beam

    Get PDF
    The term "cycloidal CT" refers to a family of efficient sampling strategies that can be applied to x-ray micro-computed tomography (CT) systems which operate with an intensity-modulated beam. Such a beam can be employed to provide access to a phase contrast channel and high spatial resolutions (a few um). Phase contrast can offer better image contrast of samples which have traditionally been "invisible” to x-rays due to their weak attenuation, and high resolutions help view crucial details in samples. Cycloidal sampling strategies provide images more quickly than the gold standard in the field ("dithering”). I conceived and compared four practical implementation strategies for cycloidal CT, three of which are "flyscans” (the sample moves continuously). Flyscans acquire images of similar resolution to dithering with no overheads, reducing acquisition time to exposure time. I also developed a "knife-edge” position tracking method which tracks subpixel motions of the sample stage. This information can be used to facilitate, automate, and improve the reconstruction of cycloidal data. I analysed the effects of different levels of dose on the signal-to-noise ratio (SNR) of an image acquired with cycloidal CT. The results show that cycloidal images yield the same SNR as dithered images with less dose, although a more extensive study is required. Finally, I explored the potential of using cycloidal CT for intraoperative specimen imaging and tissue engineering. My results are encouraging for tissue engineering; for intraoperative imaging, the cycloidal images did not show comparable resolution to the dithered images, although that is possibly linked to issues with the dataset. Overall, my work has provided a benchmark for the implementation and application of cycloidal CT for the first time. Besides a summary of my research, this thesis is meant to be a comprehensive guide for facilitating uptake of cycloidal CT within the scientific community and beyond

    Algorithms, applications and systems towards interpretable pattern mining from multi-aspect data

    Get PDF
    How do humans move around in the urban space and how do they differ when the city undergoes terrorist attacks? How do users behave in Massive Open Online courses~(MOOCs) and how do they differ if some of them achieve certificates while some of them not? What areas in the court elite players, such as Stephen Curry, LeBron James, like to make their shots in the course of the game? How can we uncover the hidden habits that govern our online purchases? Are there unspoken agendas in how different states pass legislation of certain kinds? At the heart of these seemingly unconnected puzzles is this same mystery of multi-aspect mining, i.g., how can we mine and interpret the hidden pattern from a dataset that simultaneously reveals the associations, or changes of the associations, among various aspects of the data (e.g., a shot could be described with three aspects, player, time of the game, and area in the court)? Solving this problem could open gates to a deep understanding of underlying mechanisms for many real-world phenomena. While much of the research in multi-aspect mining contribute broad scope of innovations in the mining part, interpretation of patterns from the perspective of users (or domain experts) is often overlooked. Questions like what do they require for patterns, how good are the patterns, or how to read them, have barely been addressed. Without efficient and effective ways of involving users in the process of multi-aspect mining, the results are likely to lead to something difficult for them to comprehend. This dissertation proposes the M^3 framework, which consists of multiplex pattern discovery, multifaceted pattern evaluation, and multipurpose pattern presentation, to tackle the challenges of multi-aspect pattern discovery. Based on this framework, we develop algorithms, applications, and analytic systems to enable interpretable pattern discovery from multi-aspect data. Following the concept of meaningful multiplex pattern discovery, we propose PairFac to close the gap between human information needs and naive mining optimization. We demonstrate its effectiveness in the context of impact discovery in the aftermath of urban disasters. We develop iDisc to target the crossing of multiplex pattern discovery with multifaceted pattern evaluation. iDisc meets the specific information need in understanding multi-level, contrastive behavior patterns. As an example, we use iDisc to predict student performance outcomes in Massive Open Online Courses given users' latent behaviors. FacIt is an interactive visual analytic system that sits at the intersection of all three components and enables for interpretable, fine-tunable, and scrutinizable pattern discovery from multi-aspect data. We demonstrate each work's significance and implications in its respective problem context. As a whole, this series of studies is an effort to instantiate the M^3 framework and push the field of multi-aspect mining towards a more human-centric process in real-world applications

    Toward sparse and geometry adapted video approximations

    Get PDF
    Video signals are sequences of natural images, where images are often modeled as piecewise-smooth signals. Hence, video can be seen as a 3D piecewise-smooth signal made of piecewise-smooth regions that move through time. Based on the piecewise-smooth model and on related theoretical work on rate-distortion performance of wavelet and oracle based coding schemes, one can better analyze the appropriate coding strategies that adaptive video codecs need to implement in order to be efficient. Efficient video representations for coding purposes require the use of adaptive signal decompositions able to capture appropriately the structure and redundancy appearing in video signals. Adaptivity needs to be such that it allows for proper modeling of signals in order to represent these with the lowest possible coding cost. Video is a very structured signal with high geometric content. This includes temporal geometry (normally represented by motion information) as well as spatial geometry. Clearly, most of past and present strategies used to represent video signals do not exploit properly its spatial geometry. Similarly to the case of images, a very interesting approach seems to be the decomposition of video using large over-complete libraries of basis functions able to represent salient geometric features of the signal. In the framework of video, these features should model 2D geometric video components as well as their temporal evolution, forming spatio-temporal 3D geometric primitives. Through this PhD dissertation, different aspects on the use of adaptivity in video representation are studied looking toward exploiting both aspects of video: its piecewise nature and the geometry. The first part of this work studies the use of localized temporal adaptivity in subband video coding. This is done considering two transformation schemes used for video coding: 3D wavelet representations and motion compensated temporal filtering. A theoretical R-D analysis as well as empirical results demonstrate how temporal adaptivity improves coding performance of moving edges in 3D transform (without motion compensation) based video coding. Adaptivity allows, at the same time, to equally exploit redundancy in non-moving video areas. The analogy between motion compensated video and 1D piecewise-smooth signals is studied as well. This motivates the introduction of local length adaptivity within frame-adaptive motion compensated lifted wavelet decompositions. This allows an optimal rate-distortion performance when video motion trajectories are shorter than the transformation "Group Of Pictures", or when efficient motion compensation can not be ensured. After studying temporal adaptivity, the second part of this thesis is dedicated to understand the fundamentals of how can temporal and spatial geometry be jointly exploited. This work builds on some previous results that considered the representation of spatial geometry in video (but not temporal, i.e, without motion). In order to obtain flexible and efficient (sparse) signal representations, using redundant dictionaries, the use of highly non-linear decomposition algorithms, like Matching Pursuit, is required. General signal representation using these techniques is still quite unexplored. For this reason, previous to the study of video representation, some aspects of non-linear decomposition algorithms and the efficient decomposition of images using Matching Pursuits and a geometric dictionary are investigated. A part of this investigation concerns the study on the influence of using a priori models within approximation non-linear algorithms. Dictionaries with a high internal coherence have some problems to obtain optimally sparse signal representations when used with Matching Pursuits. It is proved, theoretically and empirically, that inserting in this algorithm a priori models allows to improve the capacity to obtain sparse signal approximations, mainly when coherent dictionaries are used. Another point discussed in this preliminary study, on the use of Matching Pursuits, concerns the approach used in this work for the decompositions of video frames and images. The technique proposed in this thesis improves a previous work, where authors had to recur to sub-optimal Matching Pursuit strategies (using Genetic Algorithms), given the size of the functions library. In this work the use of full search strategies is made possible, at the same time that approximation efficiency is significantly improved and computational complexity is reduced. Finally, a priori based Matching Pursuit geometric decompositions are investigated for geometric video representations. Regularity constraints are taken into account to recover the temporal evolution of spatial geometric signal components. The results obtained for coding and multi-modal (audio-visual) signal analysis, clarify many unknowns and show to be promising, encouraging to prosecute research on the subject

    Electronic Imaging & the Visual Arts. EVA 2013 Florence

    Get PDF
    Important Information Technology topics are presented: multimedia systems, data-bases, protection of data, access to the content. Particular reference is reserved to digital images (2D, 3D) regarding Cultural Institutions (Museums, Libraries, Palace – Monuments, Archaeological Sites). The main parts of the Conference Proceedings regard: Strategic Issues, EC Projects and Related Networks & Initiatives, International Forum on “Culture & Technology”, 2D – 3D Technologies & Applications, Virtual Galleries – Museums and Related Initiatives, Access to the Culture Information. Three Workshops are related to: International Cooperation, Innovation and Enterprise, Creative Industries and Cultural Tourism

    Face age estimation using wrinkle patterns

    Get PDF
    Face age estimation is a challenging problem due to the variation of craniofacial growth, skin texture, gender and race. With recent growth in face age estimation research, wrinkles received attention from a number of research, as it is generally perceived as aging feature and soft biometric for person identification. In a face image, wrinkle is a discontinuous and arbitrary line pattern that varies in different face regions and subjects. Existing wrinkle detection algorithms and wrinkle-based features are not robust for face age estimation. They are either weakly represented or not validated against the ground truth. The primary aim of this thesis is to develop a robust wrinkle detection method and construct novel wrinkle-based methods for face age estimation. First, Hybrid Hessian Filter (HHF) is proposed to segment the wrinkles using the directional gradient and a ridge-valley Gaussian kernel. Second, Hessian Line Tracking (HLT) is proposed for wrinkle detection by exploring the wrinkle connectivity of surrounding pixels using a cross-sectional profile. Experimental results showed that HLT outperforms other wrinkle detection algorithms with an accuracy of 84% and 79% on the datasets of FORERUS and FORERET while HHF achieves 77% and 49%, respectively. Third, Multi-scale Wrinkle Patterns (MWP) is proposed as a novel feature representation for face age estimation using the wrinkle location, intensity and density. Fourth, Hybrid Aging Patterns (HAP) is proposed as a hybrid pattern for face age estimation using Facial Appearance Model (FAM) and MWP. Fifth, Multi-layer Age Regression (MAR) is proposed as a hierarchical model in complementary of FAM and MWP for face age estimation. For performance assessment of age estimation, four datasets namely FGNET, MORPH, FERET and PAL with different age ranges and sample sizes are used as benchmarks. Results showed that MAR achieves the lowest Mean Absolute Error (MAE) of 3.00 ( 4.14) on FERET and HAP scores a comparable MAE of 3.02 ( 2.92) as state of the art. In conclusion, wrinkles are important features and the uniqueness of this pattern should be considered in developing a robust model for face age estimation

    Representation and manipulation of images based on linear functionals

    Get PDF

    Segmentation and Characterization of Small Retinal Vessels in Fundus Images Using the Tensor Voting Approach

    Get PDF
    RÉSUMÉ La rétine permet de visualiser facilement une partie du réseau vasculaire humain. Elle offre ainsi un aperçu direct sur le développement et le résultat de certaines maladies liées au réseau vasculaire dans son entier. Chaque complication visible sur la rétine peut avoir un impact sur la capacité visuelle du patient. Les plus petits vaisseaux sanguins sont parmi les premières structures anatomiques affectées par la progression d’une maladie, être capable de les analyser est donc crucial. Les changements dans l’état, l’aspect, la morphologie, la fonctionnalité, ou même la croissance des petits vaisseaux indiquent la gravité des maladies. Le diabète est une maladie métabolique qui affecte des millions de personnes autour du monde. Cette maladie affecte le taux de glucose dans le sang et cause des changements pathologiques dans différents organes du corps humain. La rétinopathie diabétique décrit l’en- semble des conditions et conséquences du diabète au niveau de la rétine. Les petits vaisseaux jouent un rôle dans le déclenchement, le développement et les conséquences de la rétinopa- thie. Dans les dernières étapes de cette maladie, la croissance des nouveaux petits vaisseaux, appelée néovascularisation, présente un risque important de provoquer la cécité. Il est donc crucial de détecter tous les changements qui ont lieu dans les petits vaisseaux de la rétine dans le but de caractériser les vaisseaux sains et les vaisseaux anormaux. La caractérisation en elle-même peut faciliter la détection locale d’une rétinopathie spécifique. La segmentation automatique des structures anatomiques comme le réseau vasculaire est une étape cruciale. Ces informations peuvent être fournies à un médecin pour qu’elles soient considérées lors de son diagnostic. Dans les systèmes automatiques d’aide au diagnostic, le rôle des petits vaisseaux est significatif. Ne pas réussir à les détecter automatiquement peut conduire à une sur-segmentation du taux de faux positifs des lésions rouges dans les étapes ultérieures. Les efforts de recherche se sont concentrés jusqu’à présent sur la localisation précise des vaisseaux de taille moyenne. Les modèles existants ont beaucoup plus de difficultés à extraire les petits vaisseaux sanguins. Les modèles existants ne sont pas robustes à la grande variance d’apparence des vaisseaux ainsi qu’à l’interférence avec l’arrière-plan. Les modèles de la littérature existante supposent une forme générale qui n’est pas suffisante pour s’adapter à la largeur étroite et la courbure qui caractérisent les petits vaisseaux sanguins. De plus, le contraste avec l’arrière-plan dans les régions des petits vaisseaux est très faible. Les méthodes de segmentation ou de suivi produisent des résultats fragmentés ou discontinus. Par ailleurs, la segmentation des petits vaisseaux est généralement faite aux dépends de l’amplification du bruit. Les modèles déformables sont inadéquats pour segmenter les petits vaisseaux. Les forces utilisées ne sont pas assez flexibles pour compenser le faible contraste, la largeur, et vii la variance des vaisseaux. Enfin, les approches de type apprentissage machine nécessitent un entraînement avec une base de données étiquetée. Il est très difficile d’obtenir ces bases de données dans le cas des petits vaisseaux. Cette thèse étend les travaux de recherche antérieurs en fournissant une nouvelle mé- thode de segmentation des petits vaisseaux rétiniens. La détection de ligne à échelles multiples (MSLD) est une méthode récente qui démontre une bonne performance de segmentation dans les images de la rétine, tandis que le vote tensoriel est une méthode proposée pour reconnecter les pixels. Une approche combinant un algorithme de détection de ligne et de vote tensoriel est proposée. L’application des détecteurs de lignes a prouvé son efficacité à segmenter les vais- seaux de tailles moyennes. De plus, les approches d’organisation perceptuelle comme le vote tensoriel ont démontré une meilleure robustesse en combinant les informations voisines d’une manière hiérarchique. La méthode de vote tensoriel est plus proche de la perception humain que d’autres modèles standards. Comme démontré dans ce manuscrit, c’est un outil pour segmenter les petits vaisseaux plus puissant que les méthodes existantes. Cette combinaison spécifique nous permet de surmonter les défis de fragmentation éprouvés par les méthodes de type modèle déformable au niveau des petits vaisseaux. Nous proposons également d’utiliser un seuil adaptatif sur la réponse de l’algorithme de détection de ligne pour être plus robuste aux images non-uniformes. Nous illustrons également comment une combinaison des deux méthodes individuelles, à plusieurs échelles, est capable de reconnecter les vaisseaux sur des distances variables. Un algorithme de reconstruction des vaisseaux est également proposé. Cette dernière étape est nécessaire car l’information géométrique complète est requise pour pouvoir utiliser la segmentation dans un système d’aide au diagnostic. La segmentation a été validée sur une base de données d’images de fond d’oeil à haute résolution. Cette base contient des images manifestant une rétinopathie diabétique. La seg- mentation emploie des mesures de désaccord standards et aussi des mesures basées sur la perception. En considérant juste les petits vaisseaux dans les images de la base de données, l’amélioration dans le taux de sensibilité que notre méthode apporte par rapport à la méthode standard de détection multi-niveaux de lignes est de 6.47%. En utilisant les mesures basées sur la perception, l’amélioration est de 7.8%. Dans une seconde partie du manuscrit, nous proposons également une méthode pour caractériser les rétines saines ou anormales. Certaines images contiennent de la néovascula- risation. La caractérisation des vaisseaux en bonne santé ou anormale constitue une étape essentielle pour le développement d’un système d’aide au diagnostic. En plus des défis que posent les petits vaisseaux sains, les néovaisseaux démontrent eux un degré de complexité encore plus élevé. Ceux-ci forment en effet des réseaux de vaisseaux à la morphologie com- plexe et inhabituelle, souvent minces et à fortes courbures. Les travaux existants se limitent viii à l’utilisation de caractéristiques de premier ordre extraites des petits vaisseaux segmentés. Notre contribution est d’utiliser le vote tensoriel pour isoler les jonctions vasculaires et d’uti- liser ces jonctions comme points d’intérêts. Nous utilisons ensuite une statistique spatiale de second ordre calculée sur les jonctions pour caractériser les vaisseaux comme étant sains ou pathologiques. Notre méthode améliore la sensibilité de la caractérisation de 9.09% par rapport à une méthode de l’état de l’art. La méthode développée s’est révélée efficace pour la segmentation des vaisseaux réti- niens. Des tenseurs d’ordre supérieur ainsi que la mise en œuvre d’un vote par tenseur via un filtrage orientable pourraient être étudiés pour réduire davantage le temps d’exécution et résoudre les défis encore présents au niveau des jonctions vasculaires. De plus, la caractéri- sation pourrait être améliorée pour la détection de la rétinopathie proliférative en utilisant un apprentissage supervisé incluant des cas de rétinopathie diabétique non proliférative ou d’autres pathologies. Finalement, l’incorporation des méthodes proposées dans des systèmes d’aide au diagnostic pourrait favoriser le dépistage régulier pour une détection précoce des rétinopathies et d’autres pathologies oculaires dans le but de réduire la cessité au sein de la population.----------ABSTRACT As an easily accessible site for the direct observation of the circulation system, human retina can offer a unique insight into diseases development or outcome. Retinal vessels are repre- sentative of the general condition of the whole systematic circulation, and thus can act as a "window" to the status of the vascular network in the whole body. Each complication on the retina can have an adverse impact on the patient’s sight. In this direction, small vessels’ relevance is very high as they are among the first anatomical structures that get affected as diseases progress. Moreover, changes in the small vessels’ state, appearance, morphology, functionality, or even growth indicate the severity of the diseases. This thesis will focus on the retinal lesions due to diabetes, a serious metabolic disease affecting millions of people around the world. This disorder disturbs the natural blood glucose levels causing various pathophysiological changes in different systems across the human body. Diabetic retinopathy is the medical term that describes the condition when the fundus and the retinal vessels are affected by diabetes. As in other diseases, small vessels play a crucial role in the onset, the development, and the outcome of the retinopathy. More importantly, at the latest stage, new small vessels, or neovascularizations, growth constitutes a factor of significant risk for blindness. Therefore, there is a need to detect all the changes that occur in the small retinal vessels with the aim of characterizing the vessels to healthy or abnormal. The characterization, in turn, can facilitate the detection of a specific retinopathy locally, like the sight-threatening proliferative diabetic retinopathy. Segmentation techniques can automatically isolate important anatomical structures like the vessels, and provide this information to the physician to assist him in the final decision. In comprehensive systems for the automatization of DR detection, small vessels role is significant as missing them early in a CAD pipeline might lead to an increase in the false positive rate of red lesions in subsequent steps. So far, the efforts have been concentrated mostly on the accurate localization of the medium range vessels. In contrast, the existing models are weak in case of the small vessels. The required generalization to adapt an existing model does not allow the approaches to be flexible, yet robust to compensate for the increased variability in the appearance as well as the interference with the background. So far, the current template models (matched filtering, line detection, and morphological processing) assume a general shape for the vessels that is not enough to approximate the narrow, curved, characteristics of the small vessels. Additionally, due to the weak contrast in the small vessel regions, the current segmentation and the tracking methods produce fragmented or discontinued results. Alternatively, the small vessel segmentation can be accomplished at the expense of x background noise magnification, in the case of using thresholding or the image derivatives methods. Furthermore, the proposed deformable models are not able to propagate a contour to the full extent of the vasculature in order to enclose all the small vessels. The deformable model external forces are ineffective to compensate for the low contrast, the low width, the high variability in the small vessel appearance, as well as the discontinuities. Internal forces, also, are not able to impose a global shape constraint to the contour that could be able to approximate the variability in the appearance of the vasculature in different categories of vessels. Finally, machine learning approaches require the training of a classifier on a labelled set. Those sets are difficult to be obtained, especially in the case of the smallest vessels. In the case of the unsupervised methods, the user has to predefine the number of clusters and perform an effective initialization of the cluster centers in order to converge to the global minimum. This dissertation expanded the previous research work and provides a new segmentation method for the smallest retinal vessels. Multi-scale line detection (MSLD) is a recent method that demonstrates good segmentation performance in the retinal images, while tensor voting is a method first proposed for reconnecting pixels. For the first time, we combined the line detection with the tensor voting framework. The application of the line detectors has been proved an effective way to segment medium-sized vessels. Additionally, perceptual organization approaches like tensor voting, demonstrate increased robustness by combining information coming from the neighborhood in a hierarchical way. Tensor voting is closer than standard models to the way human perception functions. As we show, it is a more powerful tool to segment small vessels than the existing methods. This specific combination allows us to overcome the apparent fragmentation challenge of the template methods at the smallest vessels. Moreover, we thresholded the line detection response adaptively to compensate for non-uniform images. We also combined the two individual methods in a multi-scale scheme in order to reconnect vessels at variable distances. Finally, we reconstructed the vessels from their extracted centerlines based on pixel painting as complete geometric information is required to be able to utilize the segmentation in a CAD system. The segmentation was validated on a high-resolution fundus image database that in- cludes diabetic retinopathy images of varying stages, using standard discrepancy as well as perceptual-based measures. When only the smallest vessels are considered, the improve- ments in the sensitivity rate for the database against the standard multi-scale line detection method is 6.47%. For the perceptual-based measure, the improvement is 7.8% against the basic method. The second objective of the thesis was to implement a method for the characterization of isolated retinal areas into healthy or abnormal cases. Some of the original images, from which xi these patches are extracted, contain neovascularizations. Investigation of image features for the vessels characterization to healthy or abnormal constitutes an essential step in the direction of developing CAD system for the automatization of DR screening. Given that the amount of data will significantly increase under CAD systems, the focus on this category of vessels can facilitate the referral of sight-threatening cases to early treatment. In addition to the challenges that small healthy vessels pose, neovessels demonstrate an even higher degree of complexity as they form networks of convolved, twisted, looped thin vessels. The existing work is limited to the use of first-order characteristics extracted from the small segmented vessels that limits the study of patterns. Our contribution is in using the tensor voting framework to isolate the retinal vascular junctions and in turn using those junctions as points of interests. Second, we exploited second-order statistics computed on the junction spatial distribution to characterize the vessels as healthy or neovascularizations. In fact, the second-order spatial statistics extracted from the junction distribution are combined with widely used features to improve the characterization sensitivity by 9.09% over the state of art. The developed method proved effective for the segmentation of the retinal vessels. Higher order tensors along with the implementation of tensor voting via steerable filtering could be employed to further reduce the execution time, and resolve the challenges at vascular junctions. Moreover, the characterization could be advanced to the detection of prolifera- tive retinopathy by extending the supervised learning to include non-proliferative diabetic retinopathy cases or other pathologies. Ultimately, the incorporation of the methods into CAD systems could facilitate screening for the effective reduction of the vision-threatening diabetic retinopathy rates, or the early detection of other than ocular pathologies
    corecore