258 research outputs found

    Knowledge Graph semantic enhancement of input data for improving AI

    Full text link
    Intelligent systems designed using machine learning algorithms require a large number of labeled data. Background knowledge provides complementary, real world factual information that can augment the limited labeled data to train a machine learning algorithm. The term Knowledge Graph (KG) is in vogue as for many practical applications, it is convenient and useful to organize this background knowledge in the form of a graph. Recent academic research and implemented industrial intelligent systems have shown promising performance for machine learning algorithms that combine training data with a knowledge graph. In this article, we discuss the use of relevant KGs to enhance input data for two applications that use machine learning -- recommendation and community detection. The KG improves both accuracy and explainability

    Evaluating the Deductive Competence of Large Language Models

    Full text link
    The development of highly fluent large language models (LLMs) has prompted increased interest in assessing their reasoning and problem-solving capabilities. We investigate whether several LLMs can solve a classic type of deductive reasoning problem from the cognitive science literature. The tested LLMs have limited abilities to solve these problems in their conventional form. We performed follow up experiments to investigate if changes to the presentation format and content improve model performance. We do find performance differences between conditions; however, they do not improve overall performance. Moreover, we find that performance interacts with presentation format and content in unexpected ways that differ from human performance. Overall, our results suggest that LLMs have unique reasoning biases that are only partially predicted from human reasoning performance.Comment: 7 pages, 7 figures, under revie

    Expanding the Set of Pragmatic Considerations in Conversational AI

    Full text link
    Despite considerable performance improvements, current conversational AI systems often fail to meet user expectations. We discuss several pragmatic limitations of current conversational AI systems. We illustrate pragmatic limitations with examples that are syntactically appropriate, but have clear pragmatic deficiencies. We label our complaints as "Turing Test Triggers" (TTTs) as they indicate where current conversational AI systems fall short compared to human behavior. We develop a taxonomy of pragmatic considerations intended to identify what pragmatic competencies a conversational AI system requires and discuss implications for the design and evaluation of conversational AI systems.Comment: Pre-print version of paper that appeared at Multidisciplinary Perspectives on COntext-aware embodied Spoken Interactions (MP-COSIN) workshop at IEEE RO-MAN 202

    Defining and Detecting Toxicity on Social Media: Context and Knowledge are Key

    Get PDF
    As the role of online platforms has become increasingly prominent for communication, toxic behaviors, such as cyberbullying and harassment, have been rampant in the last decade. On the other hand, online toxicity is multi-dimensional and sensitive in nature, which makes its detection challenging. As the impact of exposure to online toxicity can lead to serious implications for individuals and communities, reliable models and algorithms are required for detecting and understanding such communications. In this paper We define toxicity to provide a foundation drawing social theories. Then, we provide an approach that identifies multiple dimensions of toxicity and incorporates explicit knowledge in a statistical learning algorithm to resolve ambiguity across such dimensions

    Creating a common trajectory: Shared decision making and distributed cognition in medical consultations

    Get PDF
    The growing literature on shared decision making and patient centered care emphasizes the patient’s role in clinical care, but research on clinical reasoning almost exclusively addresses physician cognition. In this article, we suggest clinical cognition is distributed between physicians and patients and assess how distributed clinical cognition functions during interactions between medical professionals and patients with Multiple Sclerosis (MS). A combination of cognitive task analysis and discourse analysis reveals the distribution of clinical reasoning between 24 patients and 3 medical professionals engaged in MS management. Findings suggest that cognition was distributed between patients and physicians in all major tasks except for the interpretation of MRI results. Otherwise, patients and physicians collaborated through discourse to develop a common trajectory to guide clinical reasoning. The patients’ role in clinical cognition expands the concept of patient-centered care and suggests the need to optimize physician-patient distributed cognition rather than physician cognition in isolation

    Requesting Distant Robotic Action: An Ontology for Naming and Action Identification for Planning on the Mars Exploration Rover Mission.

    Get PDF
    This paper focuses on the development of a naming convention and the use of abbreviated names and a related ontology for science work and distant robotic action that comprise requests for a robotic rover during the NASA Mars Exploration Rover (MER) mission, run by the Jet Propulsion Laboratory (JPL). We demonstrate how abbreviated names and an associated ontology support sharing and identifying information among teams and software tools. An ontology of distant action must take into account a dynamic environment, changing in response to physical events and intentional actions, and reflect the influence of context on the meaning of action. The nascent domain of Martian tele-robotic science, in which specialists request work from a rover moving through a distant landscape, as well as the need to consider the interdisciplinary teams involved in completing that work, required an empirical approach. The formulation of this ontology used ethnographic methods and grounded theory to study human behavior and work practice with software tools

    Towards Efficient Scoring of Student-generated Long-form Analogies in STEM

    Get PDF
    Switching from an analogy pedagogy based on comprehension to analogy pedagogy based on production raises an impractical manual analogy scoring problem. Conventional symbol-matching approaches to computational analogy evaluation focus on positive cases, and challenge computational feasibility. This work presents the Discriminative Analogy Features (DAF) pipeline to identify the discriminative features of strong and weak long-form text analogies. We introduce four feature categories (semantic, syntactic, sentiment, and statistical) used with supervised vector-based learning methods to discriminate between strong and weak analogies. Using a modestly sized vector of engineered features with SVM attains a 0.67 macro F1 score. While a semantic feature is the most discriminative, out of the top 15 discriminative features, most are syntactic. Combining these engineered features with an ELMo-generated embedding still improves classification relative to an embedding alone. While an unsupervised K-Means clustering-based approach falls short, similar hints of improvement appear when inputs include the engineered features used in supervised learning
    • …
    corecore