81 research outputs found
Curriculum Graph Machine Learning: A Survey
Graph machine learning has been extensively studied in both academia and
industry. However, in the literature, most existing graph machine learning
models are designed to conduct training with data samples in a random order,
which may suffer from suboptimal performance due to ignoring the importance of
different graph data samples and their training orders for the model
optimization status. To tackle this critical problem, curriculum graph machine
learning (Graph CL), which integrates the strength of graph machine learning
and curriculum learning, arises and attracts an increasing amount of attention
from the research community. Therefore, in this paper, we comprehensively
overview approaches on Graph CL and present a detailed survey of recent
advances in this direction. Specifically, we first discuss the key challenges
of Graph CL and provide its formal problem definition. Then, we categorize and
summarize existing methods into three classes based on three kinds of graph
machine learning tasks, i.e., node-level, link-level, and graph-level tasks.
Finally, we share our thoughts on future research directions. To the best of
our knowledge, this paper is the first survey for curriculum graph machine
learning.Comment: IJCAI 2023 Survey Trac
Automated Graph Machine Learning: Approaches, Libraries, Benchmarks and Directions
Graph machine learning has been extensively studied in both academic and
industry. However, as the literature on graph learning booms with a vast number
of emerging methods and techniques, it becomes increasingly difficult to
manually design the optimal machine learning algorithm for different
graph-related tasks. To tackle the challenge, automated graph machine learning,
which aims at discovering the best hyper-parameter and neural architecture
configuration for different graph tasks/data without manual design, is gaining
an increasing number of attentions from the research community. In this paper,
we extensively discuss automated graph machine learning approaches, covering
hyper-parameter optimization (HPO) and neural architecture search (NAS) for
graph machine learning. We briefly overview existing libraries designed for
either graph machine learning or automated machine learning respectively, and
further in depth introduce AutoGL, our dedicated and the world's first
open-source library for automated graph machine learning. Also, we describe a
tailored benchmark that supports unified, reproducible, and efficient
evaluations. Last but not least, we share our insights on future research
directions for automated graph machine learning. This paper is the first
systematic and comprehensive discussion of approaches, libraries as well as
directions for automated graph machine learning.Comment: 20 pages, 4 figures. arXiv admin note: text overlap with
arXiv:2103.0074
Graph Meets LLMs: Towards Large Graph Models
Large models have emerged as the most recent groundbreaking achievements in
artificial intelligence, and particularly machine learning. However, when it
comes to graphs, large models have not achieved the same level of success as in
other fields, such as natural language processing and computer vision. In order
to promote applying large models for graphs forward, we present a perspective
paper to discuss the challenges and opportunities associated with developing
large graph models. First, we discuss the desired characteristics of large
graph models. Then, we present detailed discussions from three key
perspectives: representation basis, graph data, and graph models. In each
category, we provide a brief overview of recent advances and highlight the
remaining challenges together with our visions. Finally, we discuss valuable
applications of large graph models. We believe this perspective can encourage
further investigations into large graph models, ultimately pushing us one step
closer towards artificial general intelligence (AGI). We are the first to
comprehensively study large graph models, to the best of our knowledge.Comment: Accepted by NeurIPS 2023 New Frontiers in Graph Learning Workshop.
Comments are welcom
LLM4DyG: Can Large Language Models Solve Problems on Dynamic Graphs?
In an era marked by the increasing adoption of Large Language Models (LLMs)
for various tasks, there is a growing focus on exploring LLMs' capabilities in
handling web data, particularly graph data. Dynamic graphs, which capture
temporal network evolution patterns, are ubiquitous in real-world web data.
Evaluating LLMs' competence in understanding spatial-temporal information on
dynamic graphs is essential for their adoption in web applications, which
remains unexplored in the literature. In this paper, we bridge the gap via
proposing to evaluate LLMs' spatial-temporal understanding abilities on dynamic
graphs, to the best of our knowledge, for the first time. Specifically, we
propose the LLM4DyG benchmark, which includes nine specially designed tasks
considering the capability evaluation of LLMs from both temporal and spatial
dimensions. Then, we conduct extensive experiments to analyze the impacts of
different data generators, data statistics, prompting techniques, and LLMs on
the model performance. Finally, we propose Disentangled Spatial-Temporal
Thoughts (DST2) for LLMs on dynamic graphs to enhance LLMs' spatial-temporal
understanding abilities. Our main observations are: 1) LLMs have preliminary
spatial-temporal understanding abilities on dynamic graphs, 2) Dynamic graph
tasks show increasing difficulties for LLMs as the graph size and density
increase, while not sensitive to the time span and data generation mechanism,
3) the proposed DST2 prompting method can help to improve LLMs'
spatial-temporal understanding abilities on dynamic graphs for most tasks. The
data and codes will be open-sourced at publication time
Multitask Learning for Citation Purpose Classification
We present our entry into the 2021 3C Shared Task Citation Context
Classification based on Purpose competition. The goal of the competition is to
classify a citation in a scientific article based on its purpose. This task is
important because it could potentially lead to more comprehensive ways of
summarizing the purpose and uses of scientific articles, but it is also
difficult, mainly due to the limited amount of available training data in which
the purposes of each citation have been hand-labeled, along with the
subjectivity of these labels. Our entry in the competition is a multi-task
model that combines multiple modules designed to handle the problem from
different perspectives, including hand-generated linguistic features, TF-IDF
features, and an LSTM-with-attention model. We also provide an ablation study
and feature analysis whose insights could lead to future work.Comment: Second Workshop on Scholarly Document Processin
Personalized and privacy-preserving federated heterogeneous medical image analysis with PPPML-HMI
Heterogeneous data is endemic due to the use of diverse models and settings
of devices by hospitals in the field of medical imaging. However, there are few
open-source frameworks for federated heterogeneous medical image analysis with
personalization and privacy protection simultaneously without the demand to
modify the existing model structures or to share any private data. In this
paper, we proposed PPPML-HMI, an open-source learning paradigm for personalized
and privacy-preserving federated heterogeneous medical image analysis. To our
best knowledge, personalization and privacy protection were achieved
simultaneously for the first time under the federated scenario by integrating
the PerFedAvg algorithm and designing our novel cyclic secure aggregation with
the homomorphic encryption algorithm. To show the utility of PPPML-HMI, we
applied it to a simulated classification task namely the classification of
healthy people and patients from the RAD-ChestCT Dataset, and one real-world
segmentation task namely the segmentation of lung infections from COVID-19 CT
scans. For the real-world task, PPPML-HMI achieved 5\% higher Dice score
on average compared to conventional FL under the heterogeneous scenario.
Meanwhile, we applied the improved deep leakage from gradients to simulate
adversarial attacks and showed the solid privacy-preserving capability of
PPPML-HMI. By applying PPPML-HMI to both tasks with different neural networks,
a varied number of users, and sample sizes, we further demonstrated the strong
robustness of PPPML-HMI
- …