195 research outputs found

    Automatic Discovery of Word Semantic Relations

    Get PDF
    In this paper, we propose an unsupervised methodology to automatically discover pairs of semantically related words by highlighting their local environment and evaluating their semantic similarity in local and global semantic spaces. This proposal di®ers from previous research as it tries to take the best of two different methodologies i.e. semantic space models and information extraction models. It can be applied to extract close semantic relations, it limits the search space and it is unsupervised

    Advancing duplicate question detection with deep learning

    Get PDF

    Lightweight and Efficient Neural Natural Language Processing with Quaternion Networks

    Full text link
    Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quaternion algebra and hypercomplex spaces, enabling not only expressive inter-component interactions but also significantly (75%75\%) reduced parameter size due to lesser degrees of freedom in the Hamilton product. We propose Quaternion variants of models, giving rise to new architectures such as the Quaternion attention Model and Quaternion Transformer. Extensive experiments on a battery of NLP tasks demonstrates the utility of proposed Quaternion-inspired models, enabling up to 75%75\% reduction in parameter size without significant loss in performance.Comment: ACL 201

    Semantic Parsing in Limited Resource Conditions

    Full text link
    This thesis explores challenges in semantic parsing, specifically focusing on scenarios with limited data and computational resources. It offers solutions using techniques like automatic data curation, knowledge transfer, active learning, and continual learning. For tasks with no parallel training data, the thesis proposes generating synthetic training examples from structured database schemas. When there is abundant data in a source domain but limited parallel data in a target domain, knowledge from the source is leveraged to improve parsing in the target domain. For multilingual situations with limited data in the target languages, the thesis introduces a method to adapt parsers using a limited human translation budget. Active learning is applied to select source-language samples for manual translation, maximizing parser performance in the target language. In addition, an alternative method is also proposed to utilize machine translation services, supplemented by human-translated data, to train a more effective parser. When computational resources are limited, a continual learning approach is introduced to minimize training time and computational memory. This maintains the parser's efficiency in previously learned tasks while adapting it to new tasks, mitigating the problem of catastrophic forgetting. Overall, the thesis provides a comprehensive set of methods to improve semantic parsing in resource-constrained conditions.Comment: PhD thesis, year of award 2023, 172 page

    From Language Comprehension Towards General AI

    Full text link
    Language comprehension or more formally, natural language understanding is one of the major undertakings in Artificial Intelligence. In this work, we explore a few of the problems in language understanding using fixed deep learning models. Specifically, first, we look into question generation. Asking questions relates to the cognitive ability of language comprehension and context understanding. For that reason, making progress in question generation is significant. We introduce a novel task called “question generation with masked target answer” and propose various models and present the baseline result for the task. Next, we extend on the question generation task and develop a large-scale dataset for our task and for question generation in general. Next, we explore the problem of paraphrase identification, in which the task is to decide whether a pair of sentences is a paraphrase of each other. We present various machine learning models and discuss their performance. Moving on from the fixed architecture of deep learning models, we then explore the area of neuroevolution where the models constantly change based on some evolutionary operators and learn until an optimal architecture is found. This direction promises to create a more general form of intelligence. In particular, we formulate a recombination algorithm called Highest Varying k-Features Recombination(HVk-FR) and use it on top of various mutation operators to evolve the models. We show how our proposed algorithm can actually go in the direction of optimal network structure starting from a basic one-layer deep network

    Computational Understanding, Generation and Evaluation of Creative Expressions

    Get PDF
    Computational creativity has received a good amount of research interest in generating creative artefacts programmatically. At the same time, research has been conducted in computational aesthetics, which essentially tries to analyse creativity exhibited in art. This thesis aims to unite these two distinct lines of research in the context of natural language generation by building, from models for interpretation and generation, a cohesive whole that can assess its own generations. I present a novel method for interpreting one of the most difficult rhetoric devices in the figurative use of language: metaphors. The method does not rely on hand-annotated data and it is purely data-driven. It obtains the state of the art results and is comparable to the interpretations given by humans. We show how a metaphor interpretation model can be used in generating metaphors and metaphorical expressions. Furthermore, as a creative natural language generation task, we demonstrate assigning creative names to colours using an algorithmic approach that leverages a knowledge base of stereotypical associations for colours. Colour names produced by the approach were favoured by human judges to names given by humans 70% of the time. A genetic algorithm-based method is elaborated for slogan generation. The use of a genetic algorithm makes it possible to model the generation of text while optimising multiple fitness functions, as part of the evolutionary process, to assess the aesthetic quality of the output. Our evaluation indicates that having multiple balanced aesthetics outperforms a single maximised aesthetic. From an interplay of neural networks and the traditional AI approach of genetic algorithms, we present a symbiotic framework. This is called the master-apprentice framework. This makes it possible for the system to produce more diverse output as the neural network can learn from both the genetic algorithm and real people. The master-apprentice framework emphasises a strong theoretical foundation for the creative problem one seeks to solve. From this theoretical foundation, a reasoned evaluation method can be derived. This thesis presents two different evaluation practices based on two different theories on computational creativity. This research is conducted in two distinct practical tasks: pun generation in English and poetry generation in Finnish.Laskennallista luovuutta on tutkittu paljon puhtaan tuottamisen näkökulmasta ja saman aikaan tutkimusta on tehty laskennallisen estetiikan saralla. Väitöskirjani yhdistää näitä kahta eri koulukuntaa, sillä kehittämäni laskennallisesti luovat järjestelmät käyttävät tuottamisessa apuna estetiikkaa; järjestelmät siis tulkitsevat teoksiaan samaan aikaan, kun ne niitä tuottavat. Käsittelen väitöskirjassani metaforien automaattista tulkintaa, värien nimien tuottamista, sloganien tuottamista sekä suomenkielisen runouden tuottamista. Metodeina käytän perinteistä koneoppimisalgoritmia, eli niin kutsuttua geneettistä algoritmia, sekä neuroverkkoja. Niiden yhdistelmää nimitän mestari ja oppipoika -malliksi, jossa geneettinen algoritmi opettaa neuroverkkoja

    PRIOR: Prototype Representation Joint Learning from Medical Images and Reports

    Full text link
    Contrastive learning based vision-language joint pre-training has emerged as a successful representation learning strategy. In this paper, we present a prototype representation learning framework incorporating both global and local alignment between medical images and reports. In contrast to standard global multi-modality alignment methods, we employ a local alignment module for fine-grained representation. Furthermore, a cross-modality conditional reconstruction module is designed to interchange information across modalities in the training phase by reconstructing masked images and reports. For reconstructing long reports, a sentence-wise prototype memory bank is constructed, enabling the network to focus on low-level localized visual and high-level clinical linguistic features. Additionally, a non-auto-regressive generation paradigm is proposed for reconstructing non-sequential reports. Experimental results on five downstream tasks, including supervised classification, zero-shot classification, image-to-text retrieval, semantic segmentation, and object detection, show the proposed method outperforms other state-of-the-art methods across multiple datasets and under different dataset size settings. The code is available at https://github.com/QtacierP/PRIOR.Comment: Accepted by ICCV 202
    corecore