521 research outputs found

    Virtual Node Tuning for Few-shot Node Classification

    Full text link
    Few-shot Node Classification (FSNC) is a challenge in graph representation learning where only a few labeled nodes per class are available for training. To tackle this issue, meta-learning has been proposed to transfer structural knowledge from base classes with abundant labels to target novel classes. However, existing solutions become ineffective or inapplicable when base classes have no or limited labeled nodes. To address this challenge, we propose an innovative method dubbed Virtual Node Tuning (VNT). Our approach utilizes a pretrained graph transformer as the encoder and injects virtual nodes as soft prompts in the embedding space, which can be optimized with few-shot labels in novel classes to modulate node embeddings for each specific FSNC task. A unique feature of VNT is that, by incorporating a Graph-based Pseudo Prompt Evolution (GPPE) module, VNT-GPPE can handle scenarios with sparse labels in base classes. Experimental results on four datasets demonstrate the superiority of the proposed approach in addressing FSNC with unlabeled or sparsely labeled base classes, outperforming existing state-of-the-art methods and even fully supervised baselines.Comment: Accepted to KDD 202

    Contextualization Distillation from Large Language Model for Knowledge Graph Completion

    Full text link
    While textual information significantly enhances the performance of pre-trained language models (PLMs) in knowledge graph completion (KGC), the static and noisy nature of existing corpora collected from Wikipedia articles or synsets definitions often limits the potential of PLM-based KGC models. To surmount these challenges, we introduce the Contextualization Distillation strategy, a versatile plug-in-and-play approach compatible with both discriminative and generative KGC frameworks. Our method begins by instructing large language models (LLMs) to transform compact, structural triplets into context-rich segments. Subsequently, we introduce two tailored auxiliary tasks, reconstruction and contextualization, allowing smaller KGC models to assimilate insights from these enriched triplets. Comprehensive evaluations across diverse datasets and KGC techniques highlight the efficacy and adaptability of our approach, revealing consistent performance enhancements irrespective of underlying pipelines or architectures. Moreover, our analysis makes our method more explainable and provides insight into generating path selection, as well as the choosing of suitable distillation tasks. All the code and data in this work will be released at https://github.com/David-Li0406/Contextulization-DistillationComment: Accepted by EACL 2024 findings v3: add missing citation

    Tuning-Free Accountable Intervention for LLM Deployment -- A Metacognitive Approach

    Full text link
    Large Language Models (LLMs) have catalyzed transformative advances across a spectrum of natural language processing tasks through few-shot or zero-shot prompting, bypassing the need for parameter tuning. While convenient, this modus operandi aggravates ``hallucination'' concerns, particularly given the enigmatic ``black-box'' nature behind their gigantic model sizes. Such concerns are exacerbated in high-stakes applications (e.g., healthcare), where unaccountable decision errors can lead to devastating consequences. In contrast, human decision-making relies on nuanced cognitive processes, such as the ability to sense and adaptively correct misjudgments through conceptual understanding. Drawing inspiration from human cognition, we propose an innovative \textit{metacognitive} approach, dubbed \textbf{CLEAR}, to equip LLMs with capabilities for self-aware error identification and correction. Our framework facilitates the construction of concept-specific sparse subnetworks that illuminate transparent decision pathways. This provides a novel interface for model \textit{intervention} after deployment. Our intervention offers compelling advantages: (\textit{i})~at deployment or inference time, our metacognitive LLMs can self-consciously identify potential mispredictions with minimum human involvement, (\textit{ii})~the model has the capability to self-correct its errors efficiently, obviating the need for additional tuning, and (\textit{iii})~the rectification procedure is not only self-explanatory but also user-friendly, enhancing the interpretability and accessibility of the model. By integrating these metacognitive features, our approach pioneers a new path toward engendering greater trustworthiness and accountability in the deployment of LLMs

    Contrastive Meta-Learning for Few-shot Node Classification

    Full text link
    Few-shot node classification, which aims to predict labels for nodes on graphs with only limited labeled nodes as references, is of great significance in real-world graph mining tasks. Particularly, in this paper, we refer to the task of classifying nodes in classes with a few labeled nodes as the few-shot node classification problem. To tackle such a label shortage issue, existing works generally leverage the meta-learning framework, which utilizes a number of episodes to extract transferable knowledge from classes with abundant labeled nodes and generalizes the knowledge to other classes with limited labeled nodes. In essence, the primary aim of few-shot node classification is to learn node embeddings that are generalizable across different classes. To accomplish this, the GNN encoder must be able to distinguish node embeddings between different classes, while also aligning embeddings for nodes in the same class. Thus, in this work, we propose to consider both the intra-class and inter-class generalizability of the model. We create a novel contrastive meta-learning framework on graphs, named COSMIC, with two key designs. First, we propose to enhance the intra-class generalizability by involving a contrastive two-step optimization in each episode to explicitly align node embeddings in the same classes. Second, we strengthen the inter-class generalizability by generating hard node classes via a novel similarity-sensitive mix-up strategy. Extensive experiments on few-shot node classification datasets verify the superiority of our framework over state-of-the-art baselines. Our code is provided at https://github.com/SongW-SW/COSMIC.Comment: SIGKDD 202

    Media Bias Matters: Understanding the Impact of Politically Biased News on Vaccine Attitudes in Social Media

    Full text link
    News media has been utilized as a political tool to stray from facts, presenting biased claims without evidence. Amid the COVID-19 pandemic, politically biased news (PBN) has significantly undermined public trust in vaccines, despite strong medical evidence supporting their efficacy. In this paper, we analyze: (i) how inherent vaccine stances subtly influence individuals' selection of news sources and participation in social media discussions; and (ii) the impact of exposure to PBN on users' attitudes toward vaccines. In doing so, we first curate a comprehensive dataset that connects PBN with related social media discourse. Utilizing advanced deep learning and causal inference techniques, we reveal distinct user behaviors between social media groups with various vaccine stances. Moreover, we observe that individuals with moderate stances, particularly the vaccine-hesitant majority, are more vulnerable to the influence of PBN compared to those with extreme views. Our findings provide critical insights to foster this line of research.Comment: 9 pages, 6 figures, 3 table

    Transductive Linear Probing: A Novel Framework for Few-Shot Node Classification

    Full text link
    Few-shot node classification is tasked to provide accurate predictions for nodes from novel classes with only few representative labeled nodes. This problem has drawn tremendous attention for its projection to prevailing real-world applications, such as product categorization for newly added commodity categories on an E-commerce platform with scarce records or diagnoses for rare diseases on a patient similarity graph. To tackle such challenging label scarcity issues in the non-Euclidean graph domain, meta-learning has become a successful and predominant paradigm. More recently, inspired by the development of graph self-supervised learning, transferring pretrained node embeddings for few-shot node classification could be a promising alternative to meta-learning but remains unexposed. In this work, we empirically demonstrate the potential of an alternative framework, \textit{Transductive Linear Probing}, that transfers pretrained node embeddings, which are learned from graph contrastive learning methods. We further extend the setting of few-shot node classification from standard fully supervised to a more realistic self-supervised setting, where meta-learning methods cannot be easily deployed due to the shortage of supervision from training classes. Surprisingly, even without any ground-truth labels, transductive linear probing with self-supervised graph contrastive pretraining can outperform the state-of-the-art fully supervised meta-learning based methods under the same protocol. We hope this work can shed new light on few-shot node classification problems and foster future research on learning from scarcely labeled instances on graphs.Comment: Accepted to the First Learning on Graph Conference (LoG 2022

    Discontinuous Galerkin Immersed Finite Volume Element Method for Anisotropic Flow Models in Porous Medium

    Get PDF
    By choosing the trial function space to the immersed finite element space and the test function space to be piecewise constant function space, we develop a discontinuous Galerkin immersed finite volume element method to solve numerically a kind of anisotropic diffusion models governed by the elliptic interface problems with discontinuous tensor-conductivity. The existence and uniqueness of the discrete scheme are proved, and an optimal-order energy-norm estimate and L2-norm estimate for the numerical solution are derived

    A comparison of posterior lumbar interbody fusion and transforaminal lumbar interbody fusion: a literature review and meta-analysis

    Get PDF
    BACKGROUND: We compared the perioperative results and complications associated with PLIF and TLIF, and collected evidence for choosing the better fusion method. METHODS: A literature survey of the MEDLINE and EMBASE databases identified 7 comparative observational studies that met our inclusion criteria. Checklists by Cowley were used to evaluate the risk of bias of the included studies. A database including patient demographic information, perioperative results, and complications was established. The summary odds ratio and weighed mean difference with 95% confidence interval were calculated with a random-effects model. RESULTS: We found that PLIF had a higher complication rate (P <0.00001), and TLIF reduced the rate of durotomy (P = 0.01). No statistical difference was found between the two groups with regard to clinical satisfaction (P = 0.54), blood loss (P = 0.14), vertebral root injury (P = 0.08), graft malposition (P = 0.06), infection (P = 0.36), or rate of radiographic fusion (P = 0.27). The evidence indicated that PLIF required longer operative time (P = 0.03). CONCLUSIONS: The evidence indicated that TLIF could reduce the complication rate and durotomy. Neither TLIP nor PLIF was found superior in terms of clinical satisfaction or radiographic fusion rate. PLIF might result in longer time in surgery. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/1471-2474-15-367) contains supplementary material, which is available to authorized users

    Evolution characteristics on coal fractures induced with CO2 phase transition fracturing based on CT scanning

    Get PDF
    This study conducted the CO2-PTF coal experiment to further reveal the fracturing transformation mechanism of CO2 phase transition fracturing(CO2-PTF)coal. According to the CT scanning and 3D fracture reconstruction, we analyzed the fracture structure parameters of coal before and after CO2-PTF, and clarified the evolution characteristics of the three-dimensional fracture structure of coal induced by CO2-PTF. The research results indicated that after CO2-PTF, the total number of fractures in the coal sample decreased, while the total volume and surface area of fractures increased. The CO2-PTF generated fracture expansion and transformation effects where the small-scale fractures were expanded and transformed into larger scale fractures under the CO2-PTF pressure. The number, volume, and surface area of fractures of less than 1 000 μm in length were significantly reduced, while the volume and surface area of fractures of longer than 1 000 μm in length were significantly increased. The expansion and connection between fractures caused a decrease in their quantity. CO2-PTF improves the connectivity of the three-dimensional fracture in coal and is conducive to gas migration and production. This study offers new insights into and evaluation method for the effect of CO2-PTF, and could provide references for the research on fracture evolution characteristics in other unconventional natural gas reservoirs and their modifications
    • …
    corecore