555 research outputs found
Modelling discourse in contested domains: A semiotic and cognitive framework
This paper examines the representational requirements for interactive, collaborative systems intended to support sensemaking and argumentation over contested issues. We argue that a perspective supported by semiotic and cognitively oriented discourse analyses offers both theoretical insights and motivates representational requirements for the semantics of tools for contesting meaning. We introduce our semiotic approach, highlighting its implications for discourse representation, before describing a research system (ClaiMaker) designed to support the construction of scholarly argumentation by allowing analysts to publish and contest 'claims' about scientific contributions. We show how ClaiMaker's representational scheme is grounded in specific assumptions concerning the nature of explicit modelling, and the evolution of meaning within a discourse community. These characteristics allow the system to represent scholarly discourse as a dynamic process, in the form of continuously evolving structures. A cognitively oriented discourse analysis then shows how the use of a small set of cognitive relational primitives in the underlying ontology opens possibilities for offering users advanced forms of computational service for analysing collectively constructed argumentation networks
Modelling naturalistic argumentation in research literatures: representation and interaction design issues
This paper characterises key weaknesses in the ability of current digital libraries to support scholarly inquiry, and as a way to address these, proposes computational services grounded in semiformal models of the naturalistic argumentation commonly found in research lteratures. It is argued that a design priority is to balance formal expressiveness with usability, making it critical to co-evolve the modelling scheme with appropriate user interfaces for argument construction and analysis. We specify the requirements for an argument modelling scheme for use by untrained researchers, describe the resulting ontology, contrasting it with other domain modelling and semantic web approaches, before discussing passive and intelligent user interfaces designed to support analysts in the construction, navigation and analysis of scholarly argument structures in a Web-based environment
Towards Foundational Semantics - Ontological Semantics Revisited -
Cimiano P, Reyle U. Towards Foundational Semantics - Ontological Semantics Revisited -. In: Bennett B, Fellbaum C, eds. Formal Ontology in Information Systems, Proceedings of the Fourth International Conference, FOIS 2006. Frontiers in Artificial Intelligence and Applications, 150. IOS Press; 2006: 51-62
The VERBMOBIL domain model version 1.0
This report describes the domain model used in the German Machine Translation project VERBMOBIL. In order make the design principles underlying the modeling explicit, we begin with a brief sketch of the VERBMOBIL demonstrator architecture from the perspective of the domain model. We then present some rather general considerations on the nature of domain modeling and its relationship to semantics. We claim that the semantic information contained in the model mainly serves two tasks. For one thing, it provides the basis for a conceptual transfer from German to English; on the other hand, it provides information needed for disambiguation. We argue that these tasks pose different requirements, and that domain modeling in general is highly task-dependent. A brief overview of domain models or ontologies used in existing NLP systems confirms this position. We finally describe the different parts of the domain model, explain our design decisions, and present examples of how the information contained in the model can be actually used in the VERBMOBIL demonstrator. In doing so, we also point out the main functionality of FLEX, the Description Logic system used for the modeling
Corpus-driven Semantics of Concession: Where do Expectations Come from?
Concession is one of the trickiest semantic discourse relations appearing in natural language. Many have tried to sub-categorize Concession and to define formal criteria to both distinguish its subtypes as well as for distinguishing Concession from the (similar) semantic relation of Contrast. But there is still a lack of consensus among the different proposals. In this paper, we focus on those approaches, e.g. (Lagerwerf 1998), (Winter & Rimon 1994), and (Korbayova & Webber 2007), assuming that Concession features two primary interpretations, "direct" and "indirect". We argue that this two way classification falls short of accounting for the full range of variants identified in naturally occurring data. Our investigation of one thousand Concession tokens in the Penn Discourse Treebank (PDTB) reveals that the interpretation of concessive relations varies according to the source of expectation. Four sources of expectation are identified. Each is characterized by a different relation holding between the eventuality that raises the expectation and the eventuality describing the expectation. We report a) a reliable inter-annotator agreement on the four types of sources identified in the PDTB data, b) a significant improvement on the annotation of previous disagreements on Concession-Contrast in the PDTB and c) a novel logical account of Concession using basic constructs from Hobbs' (1998) logic. Our proposal offers a uniform framework for the interpretation of Concession while accounting for the different sources of expectation by modifying a single predicate in the proposed formulae
Pathway toward prior knowledge-integrated machine learning in engineering
Despite the digitalization trend and data volume surge, first-principles
models (also known as logic-driven, physics-based, rule-based, or
knowledge-based models) and data-driven approaches have existed in parallel,
mirroring the ongoing AI debate on symbolism versus connectionism. Research for
process development to integrate both sides to transfer and utilize domain
knowledge in the data-driven process is rare. This study emphasizes efforts and
prevailing trends to integrate multidisciplinary domain professions into
machine acknowledgeable, data-driven processes in a two-fold organization:
examining information uncertainty sources in knowledge representation and
exploring knowledge decomposition with a three-tier knowledge-integrated
machine learning paradigm. This approach balances holist and reductionist
perspectives in the engineering domain.Comment: 8 pages, 4 figure
Retrieving Evidence from EHRs with LLMs: Possibilities and Challenges
Unstructured Electronic Health Record (EHR) data often contains critical
information complementary to imaging data that would inform radiologists'
diagnoses. However, time constraints and the large volume of notes frequently
associated with individual patients renders manual perusal of such data to
identify relevant evidence infeasible in practice. Modern Large Language Models
(LLMs) provide a flexible means of interacting with unstructured EHR data, and
may provide a mechanism to efficiently retrieve and summarize unstructured
evidence relevant to a given query. In this work, we propose and evaluate an
LLM (Flan-T5 XXL) for this purpose. Specifically, in a zero-shot setting we
task the LLM to infer whether a patient has or is at risk of a particular
condition; if so, we prompt the model to summarize the supporting evidence.
Enlisting radiologists for manual evaluation, we find that this LLM-based
approach provides outputs consistently preferred to a standard information
retrieval baseline, but we also highlight the key outstanding challenge: LLMs
are prone to hallucinating evidence. However, we provide results indicating
that model confidence in outputs might indicate when LLMs are hallucinating,
potentially providing a means to address this
Knowledge graph embedding enhancement using ontological knowledge in the biomedical domain
The biomedical field is a critical area for natural language processing (NLP) applications because it involves a vast amount of unstructured data, including clinical notes, medical publications, and electronic health records. NLP techniques can help extract valuable information from these documents, such as disease symptoms, medication usage, and treatment outcomes, which can improve patient care and clinical decision-making. MAPS S.p.A. currently produces Clinika, a software that extracts knowledge from clinical corpora. Clinika performs the task of Named Entity Recognition (NER) by linking entities to medical concepts from an established knowledge base, in this case, the Unified Medical Language System (UMLS). This dissertation details how we approached designing and implementing a component for the new version of Clinika, specifically a mention embedder that uses embeddings to perform entity linking with UMLS concepts. We focused on enhancing existing dense contextual embeddings by injecting ontological knowledge, using two parallel approaches: (1) taking the embeddings as a by-product of an entity alignment model aided by an ontology, and (2) fine-tuning a contextual language model with custom sampling strategies. We evaluated both approaches with suitable experiments from the relevant literature. After testing, we discontinued the first approach but found more significant results using the second. The results on the tasks chosen to evaluate the embeddings were not promising, we address the causes in the Error Analysis section, and discuss further work on this topic
Explainable Artificial Intelligence in Data Science: From Foundational Issues Towards Socio-technical Considerations
A widespread need to explain the behavior and outcomes of AI-based systems has
emerged, due to their ubiquitous presence. Thus, providing renewed momentum to
the relatively new research area of eXplainable AI (XAI). Nowadays, the importance
of XAI lies in the fact that the increasing control transference to this kind of system
for decision making -or, at least, its use for assisting executive stakeholders- already
afects many sensitive realms (as in Politics, Social Sciences, or Law). The decision making power handover to opaque AI systems makes mandatory explaining those,
primarily in application scenarios where the stakeholders are unaware of both the
high technology applied and the basic principles governing the technological solu tions. The issue should not be reduced to a merely technical problem; the explainer
would be compelled to transmit richer knowledge about the system (including its
role within the informational ecosystem where he/she works). To achieve such an
aim, the explainer could exploit, if necessary, practices from other scientifc and
humanistic areas. The frst aim of the paper is to emphasize and justify the need
for a multidisciplinary approach that is benefciated from part of the scientifc and
philosophical corpus on Explaining, underscoring the particular nuances of the issue
within the feld of Data Science. The second objective is to develop some arguments
justifying the authors’ bet by a more relevant role of ideas inspired by, on the one
hand, formal techniques from Knowledge Representation and Reasoning, and on
the other hand, the modeling of human reasoning when facing the explanation. This
way, explaining modeling practices would seek a sound balance between the pure
technical justifcation and the explainer-explainee agreement.Agencia Estatal de Investigación PID2019-109152GB-I00/AEI/10.13039/50110001103
Large Language Models and Knowledge Graphs: Opportunities and Challenges
Large Language Models (LLMs) have taken Knowledge Representation -- and the
world -- by storm. This inflection point marks a shift from explicit knowledge
representation to a renewed focus on the hybrid representation of both explicit
knowledge and parametric knowledge. In this position paper, we will discuss
some of the common debate points within the community on LLMs (parametric
knowledge) and Knowledge Graphs (explicit knowledge) and speculate on
opportunities and visions that the renewed focus brings, as well as related
research topics and challenges.Comment: 30 page
- …