86 research outputs found

    A Bottom-Up Review of Image Analysis Methods for Suspicious Region Detection in Mammograms.

    Get PDF
    Breast cancer is one of the most common death causes amongst women all over the world. Early detection of breast cancer plays a critical role in increasing the survival rate. Various imaging modalities, such as mammography, breast MRI, ultrasound and thermography, are used to detect breast cancer. Though there is a considerable success with mammography in biomedical imaging, detecting suspicious areas remains a challenge because, due to the manual examination and variations in shape, size, other mass morphological features, mammography accuracy changes with the density of the breast. Furthermore, going through the analysis of many mammograms per day can be a tedious task for radiologists and practitioners. One of the main objectives of biomedical imaging is to provide radiologists and practitioners with tools to help them identify all suspicious regions in a given image. Computer-aided mass detection in mammograms can serve as a second opinion tool to help radiologists avoid running into oversight errors. The scientific community has made much progress in this topic, and several approaches have been proposed along the way. Following a bottom-up narrative, this paper surveys different scientific methodologies and techniques to detect suspicious regions in mammograms spanning from methods based on low-level image features to the most recent novelties in AI-based approaches. Both theoretical and practical grounds are provided across the paper sections to highlight the pros and cons of different methodologies. The paper's main scope is to let readers embark on a journey through a fully comprehensive description of techniques, strategies and datasets on the topic

    Learning Credal Sum-Product Networks

    Get PDF
    Probabilistic representations, such as Bayesian and Markov networks, are fundamental to much of statistical machine learning. Thus, learning probabilistic representations directly from data is a deep challenge, the main computational bottleneck being inference that is intractable. Tractable learning is a powerful new paradigm that attempts to learn distributions that support efficient probabilistic querying. By leveraging local structure, representations such as sum-product networks (SPNs) can capture high tree-width models with many hidden layers, essentially a deep architecture, while still admitting a range of probabilistic queries to be computable in time polynomial in the network size. While the progress is impressive, numerous data sources are incomplete, and in the presence of missing data, structure learning methods nonetheless revert to single distributions without characterizing the loss in confidence. In recent work, credal sum-product networks, an imprecise extension of sum-product networks, were proposed to capture this robustness angle. In this work, we are interested in how such representations can be learnt and thus study how the computational machinery underlying tractable learning and inference can be generalized for imprecise probabilities.Comment: Accepted to AKBC 202

    Normative model for the diagnosis of neuropsychiatric disorders using deep learning methods

    Get PDF
    Tese de mestrado integrado, Engenharia BiomĂ©dica e BiofĂ­sica (Engenharia ClĂ­nica e Instrumentação MĂ©dica) Universidade de Lisboa, Faculdade de CiĂȘncias, 2021The diagnosis of neuropsychiatric disorders (NPDs) is still exclusively dependent on the analysis of the signs and symptoms of the patients since there are no biomarkers useful for clinical practice. Considering that several signs and symptoms are shared among different NPDs, the diagnosis is sometimes incorrect. Therefore, therapeutic approaches do not always succeed, which has an impact on the quality of life of neuropsychiatric patients. Furthermore, NPDs have a global economic and demographic impact. For this reason, technological solutions, such as DL, have been researched for the optimization of diagnosis, in the non-technological field of neuropsychiatry. However, the most promising studies on the diagnosis of NPDs with deep learning (DL) are based on binary classification, which may not be the most adequate approach to deal with the continuous spectrum of NPDs. Here, a DL-based normative model was developed to investigate functional connectivity abnormalities, that may contribute to the development of a novel diagnostic procedure. This method is here used to evaluate how patients deviate from a normal pattern learned by a group of healthy people. To create and evaluate the normative model, resting-state functional magnetic resonance imaging (rs-fMRI) data from three different databases were used. In order to maximise the balance between the amount and the quality of the data, conditions were defined to restrict the variability of the scan parameters. Subsequently, rs-fMRI data were trimmed to the lowest number of time points presented in the sample (150). Then, standard preprocessing steps were performed, including removal of the first 4 volumes of functional data, motion correction, spatial smoothing, and high pass filtering. Single-session independent component analysis (ICA) was run, and the FSL-FIX tool was used to clean noise and artefacts. The functional images were then registered to the T1-weighted brain extracted structural images, and finally to the Montreal Neurosciences Institute 152 standard space. Dual regression was applied using fourteen resting-state functional brain networks (FBN) previously identified in the literature. The Pearson’s correlation coefficient between the extracted blood oxygen level-dependent (BOLD) time series of each FBN was calculated, and a 14x14 network connectivity matrix was generated for each subject. The second part of the project consisted of the creation and optimization of a normative model. The normative model consisted of an autoencoder (AE) with three hidden layers. The AE was trained only in healthy subjects and was tested in both healthy subjects and neuropsychiatric patients, including schizophrenia (SCZ), bipolar disorder (BD), and attention deficit hyperactivity disorder (ADHD) patients. The hypothesis was that the model would “fail” on reconstructing data from neuropsychiatric patients. To evaluate the model performance, graph theory metrics were applied. Besides, the mean squared error was calculated for each feature (correlation between pairs of FBN) to evaluate which regions were worse reconstructed for each group of subjects. The pipeline for NPDs was tested for a SCZ case study, with the addition of a clustering algorithm. The results of this dissertation revealed that the proposed pipeline was able to identify patterns of functional connectivity abnormality that characterize different NPDs. Moreover, the results found for the two SCZ groups of patients were similar, which demonstrated that the normative model here presented was also generalizable. However, the group-level differences did not withstand individual-level analysis, implying that NPDs are highly heterogeneous. These findings support the idea that a precision-based medical approach, focusing on the specific functional network changes of individual patients, may be more beneficial than the traditional group-based diagnostic classification. A personalised diagnosis would allow for personalised therapy, improving the quality of life of neuropsychiatric patients

    Review of constraints on vision-based gesture recognition for human–computer interaction

    Get PDF
    The ability of computers to recognise hand gestures visually is essential for progress in human-computer interaction. Gesture recognition has applications ranging from sign language to medical assistance to virtual reality. However, gesture recognition is extremely challenging not only because of its diverse contexts, multiple interpretations, and spatio-temporal variations but also because of the complex non-rigid properties of the hand. This study surveys major constraints on vision-based gesture recognition occurring in detection and pre-processing, representation and feature extraction, and recognition. Current challenges are explored in detail

    Online Multi-Stage Deep Architectures for Feature Extraction and Object Recognition

    Get PDF
    Multi-stage visual architectures have recently found success in achieving high classification accuracies over image datasets with large variations in pose, lighting, and scale. Inspired by techniques currently at the forefront of deep learning, such architectures are typically composed of one or more layers of preprocessing, feature encoding, and pooling to extract features from raw images. Training these components traditionally relies on large sets of patches that are extracted from a potentially large image dataset. In this context, high-dimensional feature space representations are often helpful for obtaining the best classification performances and providing a higher degree of invariance to object transformations. Large datasets with high-dimensional features complicate the implementation of visual architectures in memory constrained environments. This dissertation constructs online learning replacements for the components within a multi-stage architecture and demonstrates that the proposed replacements (namely fuzzy competitive clustering, an incremental covariance estimator, and multi-layer neural network) can offer performance competitive with their offline batch counterparts while providing a reduced memory footprint. The online nature of this solution allows for the development of a method for adjusting parameters within the architecture via stochastic gradient descent. Testing over multiple datasets shows the potential benefits of this methodology when appropriate priors on the initial parameters are unknown. Alternatives to batch based decompositions for a whitening preprocessing stage which take advantage of natural image statistics and allow simple dictionary learners to work well in the problem domain are also explored. Expansions of the architecture using additional pooling statistics and multiple layers are presented and indicate that larger codebook sizes are not the only step forward to higher classification accuracies. Experimental results from these expansions further indicate the important role of sparsity and appropriate encodings within multi-stage visual feature extraction architectures

    Multiple instance fuzzy inference.

    Get PDF
    A novel fuzzy learning framework that employs fuzzy inference to solve the problem of multiple instance learning (MIL) is presented. The framework introduces a new class of fuzzy inference systems called Multiple Instance Fuzzy Inference Systems (MI-FIS). Fuzzy inference is a powerful modeling framework that can handle computing with knowledge uncertainty and measurement imprecision effectively. Fuzzy Inference performs a non-linear mapping from an input space to an output space by deriving conclusions from a set of fuzzy if-then rules and known facts. Rules can be identified from expert knowledge, or learned from data. In multiple instance problems, the training data is ambiguously labeled. Instances are grouped into bags, labels of bags are known but not those of individual instances. MIL deals with learning a classifier at the bag level. Over the years, many solutions to this problem have been proposed. However, no MIL formulation employing fuzzy inference exists in the literature. In this dissertation, we introduce multiple instance fuzzy logic that enables fuzzy reasoning with bags of instances. Accordingly, different multiple instance fuzzy inference styles are proposed. The Multiple Instance Mamdani style fuzzy inference (MI-Mamdani) extends the standard Mamdani style inference to compute with multiple instances. The Multiple Instance Sugeno style fuzzy inference (MI-Sugeno) is an extension of the standard Sugeno style inference to handle reasoning with multiple instances. In addition to the MI-FIS inference styles, one of the main contributions of this work is an adaptive neuro-fuzzy architecture designed to handle bags of instances as input and capable of learning from ambiguously labeled data. The proposed architecture, called Multiple Instance-ANFIS (MI-ANFIS), extends the standard Adaptive Neuro Fuzzy Inference System (ANFIS). We also propose different methods to identify and learn fuzzy if-then rules in the context of MIL. In particular, a novel learning algorithm for MI-ANFIS is derived. The learning is achieved by using the backpropagation algorithm to identify the premise parameters and consequent parameters of the network. The proposed framework is tested and validated using synthetic and benchmark datasets suitable for MIL problems. Additionally, we apply the proposed Multiple Instance Inference to the problem of region-based image categorization as well as to fuse the output of multiple discrimination algorithms for the purpose of landmine detection using Ground Penetrating Radar
    • 

    corecore