123 research outputs found

    Schema-aware Reference as Prompt Improves Data-Efficient Relational Triple and Event Extraction

    Full text link
    Information Extraction, which aims to extract structural relational triple or event from unstructured texts, often suffers from data scarcity issues. With the development of pre-trained language models, many prompt-based approaches to data-efficient information extraction have been proposed and achieved impressive performance. However, existing prompt learning methods for information extraction are still susceptible to several potential limitations: (i) semantic gap between natural language and output structure knowledge with pre-defined schema; (ii) representation learning with locally individual instances limits the performance given the insufficient features. In this paper, we propose a novel approach of schema-aware Reference As Prompt (RAP), which dynamically leverage schema and knowledge inherited from global (few-shot) training data for each sample. Specifically, we propose a schema-aware reference store, which unifies symbolic schema and relevant textual instances. Then, we employ a dynamic reference integration module to retrieve pertinent knowledge from the datastore as prompts during training and inference. Experimental results demonstrate that RAP can be plugged into various existing models and outperforms baselines in low-resource settings on four datasets of relational triple extraction and event extraction. In addition, we provide comprehensive empirical ablations and case analysis regarding different types and scales of knowledge in order to better understand the mechanisms of RAP. Code is available in https://github.com/zjunlp/RAP.Comment: Work in progres

    Editing Large Language Models: Problems, Methods, and Opportunities

    Full text link
    Despite the ability to train capable LLMs, the methodology for maintaining their relevancy and rectifying errors remains elusive. To this end, the past few years have witnessed a surge in techniques for editing LLMs, the objective of which is to efficiently alter the behavior of LLMs within a specific domain without negatively impacting performance across other inputs. This paper embarks on a deep exploration of the problems, methods, and opportunities related to model editing for LLMs. In particular, we provide an exhaustive overview of the task definition and challenges associated with model editing, along with an in-depth empirical analysis of the most progressive methods currently at our disposal. We also build a new benchmark dataset to facilitate a more robust evaluation and pinpoint enduring issues intrinsic to existing techniques. Our objective is to provide valuable insights into the effectiveness and feasibility of each editing technique, thereby assisting the community in making informed decisions on the selection of the most appropriate method for a specific task or context. Code and datasets are available at https://github.com/zjunlp/EasyEdit.Comment: EMNLP 2023. Updated with new experiment

    Editing Conceptual Knowledge for Large Language Models

    Full text link
    Recently, there has been a growing interest in knowledge editing for Large Language Models (LLMs). Current approaches and evaluations merely explore the instance-level editing, while whether LLMs possess the capability to modify concepts remains unclear. This paper pioneers the investigation of editing conceptual knowledge for LLMs, by constructing a novel benchmark dataset ConceptEdit and establishing a suite of new metrics for evaluation. The experimental results reveal that, although existing editing methods can efficiently modify concept-level definition to some extent, they also have the potential to distort the related instantial knowledge in LLMs, leading to poor performance. We anticipate this can inspire further progress in better understanding LLMs. Our project homepage is available at https://zjunlp.github.io/project/ConceptEdit.Comment: Work in progress. Code: https://github.com/zjunlp/EasyEdit Dataset: https://huggingface.co/datasets/zjunlp/ConceptEdi

    EasyEdit: An Easy-to-use Knowledge Editing Framework for Large Language Models

    Full text link
    Large Language Models (LLMs) usually suffer from knowledge cutoff or fallacy issues, which means they are unaware of unseen events or generate text with incorrect facts owing to the outdated/noisy data. To this end, many knowledge editing approaches for LLMs have emerged -- aiming to subtly inject/edit updated knowledge or adjust undesired behavior while minimizing the impact on unrelated inputs. Nevertheless, due to significant differences among various knowledge editing methods and the variations in task setups, there is no standard implementation framework available for the community, which hinders practitioners to apply knowledge editing to applications. To address these issues, we propose EasyEdit, an easy-to-use knowledge editing framework for LLMs. It supports various cutting-edge knowledge editing approaches and can be readily apply to many well-known LLMs such as T5, GPT-J, LlaMA, etc. Empirically, we report the knowledge editing results on LlaMA-2 with EasyEdit, demonstrating that knowledge editing surpasses traditional fine-tuning in terms of reliability and generalization. We have released the source code on GitHub at https://github.com/zjunlp/EasyEdit, along with Google Colab tutorials and comprehensive documentation for beginners to get started. Besides, we present an online system for real-time knowledge editing, and a demo video at http://knowlm.zjukg.cn/easyedit.mp4.Comment: The project website is https://github.com/zjunlp/EasyEdi

    Survey on Factuality in Large Language Models: Knowledge, Retrieval and Domain-Specificity

    Full text link
    This survey addresses the crucial issue of factuality in Large Language Models (LLMs). As LLMs find applications across diverse domains, the reliability and accuracy of their outputs become vital. We define the Factuality Issue as the probability of LLMs to produce content inconsistent with established facts. We first delve into the implications of these inaccuracies, highlighting the potential consequences and challenges posed by factual errors in LLM outputs. Subsequently, we analyze the mechanisms through which LLMs store and process facts, seeking the primary causes of factual errors. Our discussion then transitions to methodologies for evaluating LLM factuality, emphasizing key metrics, benchmarks, and studies. We further explore strategies for enhancing LLM factuality, including approaches tailored for specific domains. We focus two primary LLM configurations standalone LLMs and Retrieval-Augmented LLMs that utilizes external data, we detail their unique challenges and potential enhancements. Our survey offers a structured guide for researchers aiming to fortify the factual reliability of LLMs.Comment: 62 pages; 300+ reference

    A Comprehensive Study of Knowledge Editing for Large Language Models

    Full text link
    Large Language Models (LLMs) have shown extraordinary capabilities in understanding and generating text that closely mirrors human communication. However, a primary limitation lies in the significant computational demands during training, arising from their extensive parameterization. This challenge is further intensified by the dynamic nature of the world, necessitating frequent updates to LLMs to correct outdated information or integrate new knowledge, thereby ensuring their continued relevance. Note that many applications demand continual model adjustments post-training to address deficiencies or undesirable behaviors. There is an increasing interest in efficient, lightweight methods for on-the-fly model modifications. To this end, recent years have seen a burgeoning in the techniques of knowledge editing for LLMs, which aim to efficiently modify LLMs' behaviors within specific domains while preserving overall performance across various inputs. In this paper, we first define the knowledge editing problem and then provide a comprehensive review of cutting-edge approaches. Drawing inspiration from educational and cognitive research theories, we propose a unified categorization criterion that classifies knowledge editing methods into three groups: resorting to external knowledge, merging knowledge into the model, and editing intrinsic knowledge. Furthermore, we introduce a new benchmark, KnowEdit, for a comprehensive empirical evaluation of representative knowledge editing approaches. Additionally, we provide an in-depth analysis of knowledge location, which can give a deeper understanding of the knowledge structures inherent within LLMs. Finally, we discuss several potential applications of knowledge editing, outlining its broad and impactful implications.Comment: Ongoing work; 52 pages, 282 citations; benchmark is available at https://huggingface.co/datasets/zjunlp/KnowEdit code is available at https://github.com/zjunlp/EasyEdit paper list is available at https://github.com/zjunlp/KnowledgeEditingPaper

    Phylogeny of the Infraorder Pentatomomorpha Based on Fossil and Extant Morphology, with Description of a New Fossil Family from China

    Get PDF
    <div><h3>Background</h3><p>An extinct new family of Pentatomomorpha, Venicoridae Yao, Ren & Cai <b>fam. nov.</b>, with 2 new genera and 2 new species (<em>Venicoris solaris</em> Yao, Ren & Rider <b>gen. & sp. nov.</b> and <em>Clavaticoris zhengi</em> Yao, Ren & Cai <b>gen. & sp. nov.</b>) are described from the Early Cretaceous Yixian Formation in Northeast China.</p> <h3>Methodology/Principal Findings</h3><p>A cladistic analysis based on a combination of fossil and extant morphological characters clarified the phylogenetic status of the new family and has allowed the reconstruction of intersuperfamily and interfamily relationships within the Infraorder Pentatomomorpha. The fossil record and diversity of Pentatomomorpha during the Mesozoic is discussed.</p> <h3>Conclusions/Significance</h3><p>Pentatomomorpha is a monophyletic group; Aradoidea and the Trichophora are sister groups; these fossils belong to new family, treated as the sister group of remainder of Trichophora; Pentatomoidea is a monophyletic group; Piesmatidae should be separated as a superfamily, Piesmatoidea. Origin time of Pentatomomorpha should be tracked back to the Middle or Early Triassic.</p> </div

    Prevotella genus and its related NOD-like receptor signaling pathway in young males with stage III periodontitis

    Get PDF
    BackgroundAs periodontitis progresses, the oral microbiota community changes dynamically. In this study, we evaluated the dominant bacteria and their roles in the potential pathway in young males with stage III periodontitis.Methods16S rRNA sequencing was performed to evaluate variations in the composition of oral bacteria between males with stage I and III periodontitis and identify the dominant bacteria of each group. Function prediction was obtained based on 16S rRNA sequencing data. The inhibitor of the predominant pathway for stage III periodontitis was used to investigate the role of the dominant bacteria in periodontitis in vivo and in vitro.ResultsChao1 index, Observed Species and Phylogenetic Diversity (PD) whole tree values were significantly higher in the stage III periodontitis group. β-diversity suggested that samples could be divided according to the stages of periodontitis. The dominant bacteria in stage III periodontitis were Prevotella, Prevotella_7, and Dialister, whereas that in stage I periodontitis was Cardiobacterium. KEGG analysis predicted that variations in the oral microbiome may be related to the NOD-like receptor signaling pathway. The inhibitor of this pathway, NOD-IN-1, decreased P. intermedia -induced Tnf-α mRNA expression and increased P. intermedia -induced Il-6 mRNA expression, consistent with the ELISA results. Immunohistochemistry confirmed the down-regulation of TNF-α and IL-6 expressions by NOD-IN-1 in P. intermedia–induced periodontitis.ConclusionThe composition of the oral bacteria in young males varied according to the stage of periodontitis. The species richness of oral microtia was greater in young males with stage III periodontitis than those with stage I periodontitis. Prevotella was the dominant bacteria in young males with stage III periodontitis, and inhibition of the NOD-like receptor signaling pathway can decrease the periodontal inflammation induced by P. intermedia

    SPTAN1/Numb Axis Senses Cell Density To Restrain Cell Growth and Oncogenesis Through Hippo Signaling

    Get PDF
    The loss of contact inhibition is a key step during carcinogenesis. The Hippo-Yes-associated protein (Hippo/YAP) pathway is an important regulator of cell growth in a cell density-dependent manner. However, how Hippo signaling senses cell density in this context remains elusive. Here, we report that high cell density induced the phosphorylation of spectrin α chain, nonerythrocytic 1 (SPTAN1), a plasma membrane-stabilizing protein, to recruit NUMB endocytic adaptor protein isoforms 1 and 2 (NUMB1/2), which further sequestered microtubule affinity-regulating kinases (MARKs) in the plasma membrane and rendered them inaccessible for phosphorylation and inhibition of the Hippo kinases sterile 20-like kinases MST1 and MST2 (MST1/2). WW45 interaction with MST1/2 was thereby enhanced, resulting in the activation of Hippo signaling to block YAP activity for cell contact inhibition. Importantly, low cell density led to SPTAN1 dephosphorylation and NUMB cytoplasmic location, along with MST1/2 inhibition and, consequently, YAP activation. Moreover, double KO of NUMB and WW45 in the liver led to appreciable organ enlargement and rapid tumorigenesis. Interestingly, NUMB isoforms 3 and 4, which have a truncated phosphotyrosine-binding (PTB) domain and are thus unable to interact with phosphorylated SPTAN1 and activate MST1/2, were selectively upregulated in liver cancer, which correlated with YAP activation. We have thus revealed a SPTAN1/NUMB1/2 axis that acts as a cell density sensor to restrain cell growth and oncogenesis by coupling external cell-cell contact signals to intracellular Hippo signaling
    • …
    corecore