8 research outputs found

    Learning Pretopological Spaces for Lexical Taxonomy Acquisition

    Full text link
    International audienceIn this paper, we propose a new methodology for semi-supervised acquisition of lexical taxonomies from a list of existing terms. Our approach is based on the theory of pretopology that offers a powerful formalism to model semantic relations and transform a list of terms into a structured term space by combining different discriminant criteria. In order to learn a parameterized pretopological space, we define the Learning Pretopological Spaces strategy based on genetic algorithms. The rare but accurate pieces of knowledge given by an expert (semi-supervision) or automatically extracted with existing linguistic patterns (auto-supervision) are used to parameterize the different features defining the pretopological term space. Then, a structuring algorithm is used to transform the pretopological space into a lexical taxonomy, i.e. a direct acyclic graph. Results over three standard datasets (two from WordNet and one from UMLS) evidence improved performances against existing associative and pattern-based state-of-the-art approaches

    Semantic Systems. The Power of AI and Knowledge Graphs

    Get PDF
    This open access book constitutes the refereed proceedings of the 15th International Conference on Semantic Systems, SEMANTiCS 2019, held in Karlsruhe, Germany, in September 2019. The 20 full papers and 8 short papers presented in this volume were carefully reviewed and selected from 88 submissions. They cover topics such as: web semantics and linked (open) data; machine learning and deep learning techniques; semantic information management and knowledge integration; terminology, thesaurus and ontology management; data mining and knowledge discovery; semantics in blockchain and distributed ledger technologies

    Explaining deep neural networks through knowledge extraction and graph analysis

    Get PDF
    Explainable Artificial Intelligence (XAI) has recently become an active research field due to the need for transparency and accountability when deploying AI models for high-stake decision making. Despite the success of Deep Neural Networks (DNNs), understanding their decision processes is still a known challenge. The research direction presented in this thesis stems from the idea that combining knowledge with deep representations can be the key to more transparent decision making. Specifically, we have focused on Computer Vision tasks and Convolutional Neural Networks (CNNs) and we have proposed a graph representation, called co-activation graph, that serves as an intermediate representation between knowledge encoded within a CNN and the semantics contained in external knowledge bases. Given a trained CNN, we first show how a co-activation graph can be created and exploited to generate global insights for the model’s inner-workings. Then, we propose a taxonomy extraction method that captures how symbolic class concepts and their hypernyms in a given domain are hierarchically organised in the model’s subsymbolic representation. We then illustrate how background knowledge can be connected to the graph in order to generate textual local factual and counterfactual explanations. Our results indicate that graph analysis approaches applied to co-activation graphs can reveal important insights into how CNNs work and enable both global and local semantic explanations. Despite focusing on CNN architectures, we believe that our approach can be adapted to other architectures which would make it possible to apply the same methodology in other domains such as Natural Language Processing. At the end of the thesis we will discuss interesting research directions that are being investigated in the area of using knowledge graphs and graph analysis for explainability of deep learning models, and we outline opportunities for the development of this research are

    Feature Learning for RGB-D Data

    Get PDF
    RGB-D data has turned out to be a very useful representation for solving fundamental computer vision problems. It takes the advantages of the color images that provide appearance information of an object and also the depth image that is immune to the variations in color, illumination, rotation angle and scale. With the invention of the low-cost Microsoft Kinect sensor, which was initially used for gaming and later became a popular device for computer vision, high quality RGB-D data can be acquired easily. RGB-D image/video can facilitate a wide range of application areas, such as computer vision, robotics, construction and medical imaging. Furthermore, how to fuse RGB information and depth information is still a problem in computer vision. It is not enough to simply concatenate RGB data and depth data together. A new fusion method could better fuse RGB images and depth images. It still needs more powerful algorithms on this. In this thesis, to explore more advantages of RGB-D data, we use some popular RGB-D datasets for deep feature learning algorithms evaluation, hyper-parameter optimization, local multi-modal feature learning, RGB-D data fusion and recognizing RGB information from RGB-D images: i)With the success of Deep Neural Network in computer vision, deep features from fused RGB-D data can be proved to gain better results than RGB data only. However, different deep learning algorithms show different performance on different RGB-D datasets. Through large-scale experiments to comprehensively evaluate the performance of deep feature learning models for RGB-D image/ video classification, we obtain the conclusion that RGB-D fusion methods using CNNs always outperform other selected methods (DBNs, SDAE and LSTM). On the other side, since LSTM can learn from experience to classify, process and predict time series, it achieved better performances than DBN and SDAE in video classification tasks. ii) Hyper-parameter optimization can help researchers quickly choose an initial set of hyper-parameters for a new coming classification task, thus reducing the number of trials in terms of hyper-parameter space. We present a simple and efficient framework for improving the efficiency and accuracy of hyper-parameter optimization by considering the classification complexity of a particular dataset. We verify this framework on three real-world RGB-D datasets. After the analysis of experiments, we confirm that our framework can provide deeper insights into the relationship between dataset classification tasks and hyperparameters optimization, thus quickly choosing an accurate initial set of hyper-parameters for a new coming classification task. iii) We propose a new Convolutional Neural Networks (CNNs)-based local multi-modal feature learning framework for RGB-D scene classification. This method can effectively capture much of the local structure from the RGB-D scene images and automatically learn a fusion strategy for the object-level recognition step instead of simply training a classifier on top of features extracted from both modalities. Experiments are conducted on two popular datasets to thoroughly test the performance of our method, which show that our method with local multi-modal CNNs greatly outperforms state-of-the-art approaches. Our method has the potential to improve RGB-D scene understanding. Some extended evaluation shows that CNNs trained using a scene-centric dataset is able to achieve an improvement on scene benchmarks compared to a network trained using an object-centric dataset. iv) We propose a novel method for RGB-D data fusion. We project raw RGB-D data into a complex space and then jointly extract features from the fused RGB-D images. Besides three observations about the fusion methods, the experimental results also show that our method achieves competing performance against the classical SIFT. v) We propose a novel method called adaptive Visual-Depth Embedding (aVDE) which learns the compact shared latent space between two representations of labeled RGB and depth modalities in the source domain first. Then the shared latent space can help the transfer of the depth information to the unlabeled target dataset. At last, aVDE matches features and reweights instances jointly across the shared latent space and the projected target domain for an adaptive classifier. This method can utilize the additional depth information in the source domain and simultaneously reduce the domain mismatch between the source and target domains. On two real-world image datasets, the experimental results illustrate that the proposed method significantly outperforms the state-of-the-art methods

    An analysis of target recipient groups for monovalent 2009 pandemic influenza vaccine and trivalent seasonal influenza vaccines in 2009-10 and 2010-11

    Get PDF
    Poster Presentation: SPA5 - How to Evaluate Vaccine Effectiveness and Efficacy?: abstract no. A513PINTRODUCTION: Vaccination is generally considered to be the best primary prevention measure against influenza virus infection. Many countries encourage specific target groups of people to undertake vaccination, often with financial subsidies or a list of priority. To understand differential patterns of national target groups for influenza vaccination before, during and after the 2009 influenza pandemic, we reviewed and identified changes in national target groups for trivalent seasonal influenza and the monovalent 2009 pandemic influenza vaccines dur...postprin

    Efficacy of live attenuated seasonal and pandemic influenza vaccine in school-age children: a randomized controlled trial

    Get PDF
    Poster Presentation: SPA5 - How to Evaluate Vaccine Effectiveness and Efficacy?: abstract no. A508PBACKGROUND: A novel pandemic influenza A(H1N1) virus emerged in North America in early 2009 and rapidly spread worldwide. Monovalent pH1N1 vaccines were licensed later in 2009 based on preliminary studies demonstrating their immunogenicity and safety. In this study we report the efficacy of live attenuated monovalent pH1N1 vacc...postprin

    A modular genetic programming system

    Get PDF
    Genetic Programming (GP) is an evolutionary algorithm for the automatic discovery of symbolic expressions, e.g. computer programs or mathematical formulae, that encode solutions to a user-defined task. Recent advances in GP systems and computer performance made it possible to successfully apply this algorithm to real-world applications. This work offers three main contributions to the state-of-the art in GP systems: (I) The documentation of RGP, a state-of-the art GP software implemented as an extension package to the popular R environment for statistical computation and graphics. GP and RPG are introduced both formally and with a series of tutorial examples. As R itself, RGP is available under an open source license. (II) A comprehensive empirical analysis of modern GP heuristics based on the methodology of Sequential Parameter Optimization. The effects and interactions of the most important GP algorithm parameters are analyzed and recommendations for good parameter settings are given. (III) Two extensive case studies based on real-world industrial applications. The first application involves process control models in steel production, while the second is about meta-model-based optimization of cyclone dust separators. A comparison with traditional and modern regression methods reveals that GP offers equal or superior performance in both applications, with the additional benefit of understandable and easy to deploy models. Main motivation of this work is the advancement of GP in real-world application areas. The focus lies on a subset of application areas that are known to be practical for GP, first of all symbolic regression and classification. It has been written with practitioners from academia and industry in mind
    corecore