246 research outputs found
EXPLAINABLE FEATURE- AND DECISION-LEVEL FUSION
Information fusion is the process of aggregating knowledge from multiple data sources to produce more consistent, accurate, and useful information than any one individual source can provide. In general, there are three primary sources of data/information: humans, algorithms, and sensors. Typically, objective data---e.g., measurements---arise from sensors. Using these data sources, applications such as computer vision and remote sensing have long been applying fusion at different levels (signal, feature, decision, etc.). Furthermore, the daily advancement in engineering technologies like smart cars, which operate in complex and dynamic environments using multiple sensors, are raising both the demand for and complexity of fusion. There is a great need to discover new theories to combine and analyze heterogeneous data arising from one or more sources.
The work collected in this dissertation addresses the problem of feature- and decision-level fusion. Specifically, this work focuses on fuzzy choquet integral (ChI)-based data fusion methods. Most mathematical approaches for data fusion have focused on combining inputs relative to the assumption of independence between them. However, often there are rich interactions (e.g., correlations) between inputs that should be exploited. The ChI is a powerful aggregation tool that is capable modeling these interactions. Consider the fusion of m sources, where there are 2m unique subsets (interactions); the ChI is capable of learning the worth of each of these possible source subsets. However, the complexity of fuzzy integral-based methods grows quickly, as the number of trainable parameters for the fusion of m sources scales as 2m. Hence, we require a large amount of training data to avoid the problem of over-fitting. This work addresses the over-fitting problem of ChI-based data fusion with novel regularization strategies. These regularization strategies alleviate the issue of over-fitting while training with limited data and also enable the user to consciously push the learned methods to take a predefined, or perhaps known, structure. Also, the existing methods for training the ChI for decision- and feature-level data fusion involve quadratic programming (QP). The QP-based learning approach for learning ChI-based data fusion solutions has a high space complexity. This has limited the practical application of ChI-based data fusion methods to six or fewer input sources. To address the space complexity issue, this work introduces an online training algorithm for learning ChI. The online method is an iterative gradient descent approach that processes one observation at a time, enabling the applicability of ChI-based data fusion on higher dimensional data sets.
In many real-world data fusion applications, it is imperative to have an explanation or interpretation. This may include providing information on what was learned, what is the worth of individual sources, why a decision was reached, what evidence process(es) were used, and what confidence does the system have on its decision. However, most existing machine learning solutions for data fusion are black boxes, e.g., deep learning. In this work, we designed methods and metrics that help with answering these questions of interpretation, and we also developed visualization methods that help users better understand the machine learning solution and its behavior for different instances of data
Development of Machine Learning Techniques for Diabetic Retinopathy Risk Estimation
La retinopatia diabètica (DR) és una malaltia crònica. És una de les principals complicacions de
diabetis i una causa essencial de pèrdua de visió entre les persones que pateixen diabetis.
Els pacients diabètics han de ser analitzats periòdicament per tal de detectar signes de
desenvolupament de la retinopatia en una fase inicial. El cribratge precoç i freqüent disminueix
el risc de pèrdua de visió i minimitza la càrrega als centres assistencials. El nombre
dels pacients diabètics està en augment i creixements ràpids, de manera que el fa difícil
que consumeix recursos per realitzar un cribatge anual a tots ells.
L’objectiu principal d’aquest doctorat. la tesi consisteix en construir un sistema de suport de decisions clíniques
(CDSS) basat en dades de registre de salut electrònic (EHR). S'utilitzarà aquest CDSS per estimar el risc de desenvolupar RD.
En aquesta tesi doctoral s'estudien mètodes d'aprenentatge automàtic per constuir un CDSS basat en regles lingüístiques difuses. El coneixement expressat en aquest tipus de regles facilita que el metge sàpiga quines combindacions de les condicions són les poden provocar el risc de desenvolupar RD.
En aquest treball, proposo un mètode per reduir la incertesa en la classificació dels
pacients que utilitzen arbres de decisió difusos (FDT). A continuació es combinen diferents arbres, usant la tècnica de
Fuzzy Random Forest per millorar la qualitat de la predicció.
A continuació es proposen diverses tècniques d'agregació que millorin la fusió dels resultats que ens dóna
cadascun dels arbres FDT. Per millorar la decisió final dels nostres models, proposo tres mesures difuses que
s'utilitzen amb integrals de Choquet i Sugeno. La definició d’aquestes mesures difuses es basa en els valors de confiança de les regles. En particular, una d'elles és una mesura difusa que es troba en la qual
l'estructura jeràrquica de la FDT és explotada per trobar els valors de la mesura difusa.
El resultat final de la recerca feta ha donat lloc a un programari que es pot instal·lar en centres d’assistència primària i hospitals, i pot ser usat pels metges de capçalera per fer l'avaluació preventiva i el cribatge de la Retinopatia Diabètica.La retinopatía diabética (RD) es una enfermedad crónica. Es una de las principales complicaciones de
diabetes y una causa esencial de pérdida de visión entre las personas que padecen diabetes.
Los pacientes diabéticos deben ser examinados periódicamente para detectar signos de diabetes.
desarrollo de retinopatía en una etapa temprana. La detección temprana y frecuente disminuye
el riesgo de pérdida de visión y minimiza la carga en los centros de salud. El número
de pacientes diabéticos es enorme y está aumentando rápidamente, lo que lo hace difícil y
Consume recursos para realizar una evaluación anual para todos ellos.
El objetivo principal de esta tesis es construir un sistema de apoyo a la decisión clínica
(CDSS) basado en datos de registros de salud electrónicos (EHR). Este CDSS será utilizado
para estimar el riesgo de desarrollar RD.
En este tesis doctoral se estudian métodos de aprendizaje automático para construir un CDSS basado
en reglas lingüísticas difusas. El conocimiento expresado en este tipo de reglas facilita que el médico
pueda saber que combinaciones de las condiciones son las que pueden provocar el riesgo de desarrollar RD.
En este trabajo propongo un método para reducir la incertidumbre en la clasificación de los
pacientes que usan árboles de decisión difusos (FDT). A continuación se combinan diferentes árboles usando
la técnica de Fuzzy Random Forest para mejorar la calidad de la predicción.
Se proponen también varias políticas para fusionar los resultados de que nos da cada uno de los árboles (FDT).
Para mejorar la decisión final propongo tres medidas difusas que se usan con las integrales Choquet y Sugeno.
La definición de estas medidas difusas se basa en los valores de confianza de
las reglas. En particular, uno de ellos es una medida difusa descomponible en la que se usa
la estructura jerárquica del FDT para encontrar los valores de la medida difusa.
Como resultado final de la investigación se ha construido un software que puede instalarse en centros de atención médica y hospitales, i que puede ser usado por los médicos de cabecera para hacer la evaluación preventiva y
el cribado de la Retinopatía Diabética.Diabetic retinopathy (DR) is a chronic illness. It is one of the main complications of
diabetes, and an essential cause of vision loss among people suffering from diabetes.
Diabetic patients must be periodically screened in order to detect signs of diabetic
retinopathy development in an early stage. Early and frequent screening decreases
the risk of vision loss and minimizes the load on the health care centres. The number
of the diabetic patients is huge and rapidly increasing so that makes it hard and
resource-consuming to perform a yearly screening to all of them.
The main goal of this Ph.D. thesis is to build a clinical decision support system
(CDSS) based on electronic health record (EHR) data. This CDSS will be utilised
to estimate the risk of developing RD.
In this Ph.D. thesis, I focus on developing novel interpretable machine learning
systems. Fuzzy based systems with linguistic terms are going to be proposed. The
output of such systems makes the physician know what combinations of the features
that can cause the risk of developing DR.
In this work, I propose a method to reduce the uncertainty in classifying diabetic
patients using fuzzy decision trees. A Fuzzy Random forest (FRF) approach is
proposed as well to estimate the risk for developing DR.
Several policies are going to be proposed to merge the classification results
achieved by different Fuzzy Decision Trees (FDT) models to improve the quality of
the final decision of our models, I propose three fuzzy measures that are used with Choquet and Sugeno integrals.
The definition of these fuzzy measures is based on the confidence values of
the rules. In particular, one of them is a decomposable fuzzy measure in which the
hierarchical structure of the FDT is exploited to find the values of the fuzzy measure.
Out of this Ph.D. work, we have built a CDSS software that may be installed in the health care centres and hospitals
in order to evaluate and detect Diabetic Retinopathy at early stages
Learning nonlinear monotone classifiers using the Choquet Integral
In der jüngeren Vergangenheit hat das Lernen von Vorhersagemodellen, die eine monotone Beziehung zwischen Ein- und Ausgabevariablen garantieren, wachsende Aufmerksamkeit im Bereich des maschinellen Lernens erlangt. Besonders für flexible nichtlineare Modelle stellt die Gewährleistung der Monotonie eine große Herausforderung für die Umsetzung dar. Die vorgelegte Arbeit nutzt das Choquet Integral als mathematische Grundlage für die Entwicklung neuer Modelle für nichtlineare Klassifikationsaufgaben. Neben den bekannten Einsatzgebieten des Choquet-Integrals als flexible Aggregationsfunktion in multi-kriteriellen Entscheidungsverfahren, findet der Formalismus damit Eingang als wichtiges Werkzeug für Modelle des maschinellen Lernens. Neben dem Vorteil, Monotonie und Flexibilität auf elegante Weise mathematisch vereinbar zu machen, bietet das Choquet-Integral Möglichkeiten zur Quantifizierung von Wechselwirkungen zwischen Gruppen von Attributen der Eingabedaten, wodurch interpretierbare Modelle gewonnen werden können. In der Arbeit werden konkrete Methoden für das Lernen mit dem Choquet Integral entwickelt, welche zwei unterschiedliche Ansätze nutzen, die Maximum-Likelihood-Schätzung und die strukturelle Risikominimierung. Während der erste Ansatz zu einer Verallgemeinerung der logistischen Regression führt, wird der zweite mit Hilfe von Support-Vektor-Maschinen realisiert. In beiden Fällen wird das Lernproblem imWesentlichen auf die Parameter-Identifikation von Fuzzy-Maßen für das Choquet Integral zurückgeführt. Die exponentielle Anzahl von Freiheitsgraden zur Modellierung aller Attribut-Teilmengen stellt dabei besondere Herausforderungen im Hinblick auf Laufzeitkomplexität und Generalisierungsleistung. Vor deren Hintergrund werden die beiden Ansätze praktisch bewertet und auch theoretisch analysiert. Zudem werden auch geeignete Verfahren zur Komplexitätsreduktion und Modellregularisierung vorgeschlagen und untersucht. Die experimentellen Ergebnisse sind auch für anspruchsvolle Referenzprobleme im Vergleich mit aktuellen Verfahren sehr gut und heben die Nützlichkeit der Kombination aus Monotonie und Flexibilität des Choquet Integrals in verschiedenen Ansätzen des maschinellen Lernens hervor
Feature and Decision Level Fusion Using Multiple Kernel Learning and Fuzzy Integrals
The work collected in this dissertation addresses the problem of data fusion. In other words, this is the problem of making decisions (also known as the problem of classification in the machine learning and statistics communities) when data from multiple sources are available, or when decisions/confidence levels from a panel of decision-makers are accessible. This problem has become increasingly important in recent years, especially with the ever-increasing popularity of autonomous systems outfitted with suites of sensors and the dawn of the ``age of big data.\u27\u27 While data fusion is a very broad topic, the work in this dissertation considers two very specific techniques: feature-level fusion and decision-level fusion. In general, the fusion methods proposed throughout this dissertation rely on kernel methods and fuzzy integrals. Both are very powerful tools, however, they also come with challenges, some of which are summarized below. I address these challenges in this dissertation.
Kernel methods for classification is a well-studied area in which data are implicitly mapped from a lower-dimensional space to a higher-dimensional space to improve classification accuracy. However, for most kernel methods, one must still choose a kernel to use for the problem. Since there is, in general, no way of knowing which kernel is the best, multiple kernel learning (MKL) is a technique used to learn the aggregation of a set of valid kernels into a single (ideally) superior kernel. The aggregation can be done using weighted sums of the pre-computed kernels, but determining the summation weights is not a trivial task. Furthermore, MKL does not work well with large datasets because of limited storage space and prediction speed. These challenges are tackled by the introduction of many new algorithms in the following chapters. I also address MKL\u27s storage and speed drawbacks, allowing MKL-based techniques to be applied to big data efficiently.
Some algorithms in this work are based on the Choquet fuzzy integral, a powerful nonlinear aggregation operator parameterized by the fuzzy measure (FM). These decision-level fusion algorithms learn a fuzzy measure by minimizing a sum of squared error (SSE) criterion based on a set of training data. The flexibility of the Choquet integral comes with a cost, however---given a set of N decision makers, the size of the FM the algorithm must learn is 2N. This means that the training data must be diverse enough to include 2N independent observations, though this is rarely encountered in practice. I address this in the following chapters via many different regularization functions, a popular technique in machine learning and statistics used to prevent overfitting and increase model generalization. Finally, it is worth noting that the aggregation behavior of the Choquet integral is not intuitive. I tackle this by proposing a quantitative visualization strategy allowing the FM and Choquet integral behavior to be shown simultaneously
- …