6 research outputs found
Towards A Question Answering System over Temporal Knowledge Graph Embedding
Question Answering (QA) over knowledge graphs is a vital topic within information retrieval. Questions with temporal intent are a special case of questions for QA systems that have received only limited attention so far. In this paper, we study using temporal knowledge graph embeddings (TKGEs) for temporal QA. Firstly, we propose a microservice-based architecture for building temporal QA systems on pre-trained TKGE models. Secondly, we present a Bayesian model average (BMA) ensemble method, where results of several link prediction tasks on separated TKGE models are combined to find better answers. Within the system built using the microservice-based architecture, the experiments on two benchmark datasets show that BMA provides better results than the individual models.</p
Core Building Blocks: Next Gen Geo Spatial GPT Application
This paper proposes MapGPT which is a novel approach that integrates the
capabilities of language models, specifically large language models (LLMs),
with spatial data processing techniques. This paper introduces MapGPT, which
aims to bridge the gap between natural language understanding and spatial data
analysis by highlighting the relevant core building blocks. By combining the
strengths of LLMs and geospatial analysis, MapGPT enables more accurate and
contextually aware responses to location-based queries. The proposed
methodology highlights building LLMs on spatial and textual data, utilizing
tokenization and vector representations specific to spatial information. The
paper also explores the challenges associated with generating spatial vector
representations. Furthermore, the study discusses the potential of
computational capabilities within MapGPT, allowing users to perform geospatial
computations and obtain visualized outputs. Overall, this research paper
presents the building blocks and methodology of MapGPT, highlighting its
potential to enhance spatial data understanding and generation in natural
language processing applications
Systematic review of question answering over knowledge bases
Over the years, a growing number of semantic data repositories have been made available on the web. However, this has created new challenges in exploiting these resources efficiently. Querying services require knowledge beyond the typical user’s expertise, which is a critical issue in adopting semantic information solutions. Several proposals to overcome this dif- ficulty have suggested using question answering (QA) systems to provide user‐friendly interfaces and allow natural language use. Because question answering over knowledge bases (KBQAs) is a very active research topic, a comprehensive view of the field is essential. The purpose of this study was to conduct a systematic review of methods and systems for KBQAs to identify their main advantages and limitations. The inclusion criteria rationale was English full‐text articles published since 2015 on methods and systems for KBQAs.info:eu-repo/semantics/publishedVersio
Recommended from our members
Chatbot Interaction with Artificial Intelligence:human data augmentation with T5 and language transformer ensemble for text classification
In this work we present the Chatbot Interaction with Artificial Intelligence (CI-AI) framework as an approach to the training of a transformer based chatbot-like architecture for task classification with a focus on natural human interaction with a machine as opposed to interfaces, code, or formal commands. The intelligent system augments human-sourced data via artificial paraphrasing in order to generate a large set of training data for further classical, attention, and language transformation-based learning approaches for Natural Language Processing (NLP). Human beings are asked to paraphrase commands and questions for task identification for further execution of algorithms as skills. The commands and questions are split into training and validation sets. A total of 483 responses were recorded. Secondly, the training set is paraphrased by the T5 model in order to augment it with further data. Seven state-of-the-art transformer-based text classification algorithms (BERT, DistilBERT, RoBERTa, DistilRoBERTa, XLM, XLM-RoBERTa, and XLNet) are benchmarked for both sets after fine-tuning on the training data for two epochs. We find that all models are improved when training data is augmented by the T5 model, with an average increase of classification accuracy by 4.01%. The best result was the RoBERTa model trained on T5 augmented data which achieved 98.96% classification accuracy. Finally, we found that an ensemble of the five best-performing transformer models via Logistic Regression of output label predictions led to an accuracy of 99.59% on the dataset of human responses. A highly-performing model allows the intelligent system to interpret human commands at the social-interaction level through a chatbot-like interface (e.g. “Robot, can we have a conversation?”) and allows for better accessibility to AI by non-technical users