17 research outputs found
Efficient Large Language Models Fine-Tuning On Graphs
Learning from Text-Attributed Graphs (TAGs) has attracted significant
attention due to its wide range of real-world applications. The rapid evolution
of large language models (LLMs) has revolutionized the way we process textual
data, which indicates a strong potential to replace shallow text embedding
generally used in Graph Neural Networks (GNNs). However, we find that existing
LLM approaches that exploit text information in graphs suffer from inferior
computation and data efficiency. In this work, we introduce a novel and
efficient approach for the end-to-end fine-tuning of Large Language Models
(LLMs) on TAGs, named LEADING. The proposed approach maintains computation cost
and memory overhead comparable to the graph-less fine-tuning of LLMs. Moreover,
it transfers the rick knowledge in LLMs to downstream graph learning tasks
effectively with limited labeled data in semi-supervised learning. Its superior
computation and data efficiency are demonstrated through comprehensive
experiments, offering a promising solution for a wide range of LLMs and graph
learning tasks on TAGs
A Vehicular Trust Blockchain Framework with Scalable Byzantine Consensus
The maturing blockchain technology has gradually promoted decentralized data storage from cryptocurrencies to other applications, such as trust management, resulting in new challenges based on specific scenarios. Taking the mobile trust blockchain within a vehicular network as an example, many users require the system to process massive traffic information for accurate trust assessment, preserve data reliably, and respond quickly. While existing vehicular blockchain systems ensure immutability, transparency, and traceability, they are limited in terms of scalability, performance, and security. To address these issues, this paper proposes a novel decentralized vehicle trust management solution and a well-matched blockchain framework that provides both security and performance. The paper primarily addresses two issues: i) To provide accurate trust evaluation, the trust model adopts a decentralized and peer-review-based trust computation method secured by trusted execution environments (TEEs). ii) To ensure reliable trust management, a multi-shard blockchain framework is developed with a novel hierarchical Byzantine consensus protocol, improving efficiency and security while providing high scalability and performance. The proposed scheme combines the decentralized trust model with a multi-shard blockchain, preserving trust information through a hierarchical consensus protocol. Finally, real-world experiments are conducted by developing a testbed deployed on both local and cloud servers for performance measurements
FENDI: High-Fidelity Entanglement Distribution in the Quantum Internet
A quantum network distributes quantum entanglements between remote nodes,
which is key to many quantum applications. However, unavoidable noise in
quantum operations could lead to both low throughput and low quality of
entanglement distribution. This paper aims to address the simultaneous
exponential degradation in throughput and quality in a buffered multi-hop
quantum network. Based on an end-to-end fidelity model with worst-case
(isotropic) noise, we formulate the high-fidelity remote entanglement
distribution problem for a single source-destination pair, and prove its
NP-hardness. To address the problem, we develop a fully polynomial-time
approximation scheme for the control plane of the quantum network, and a
distributed data plane protocol that achieves the desired long-term throughput
and worst-case fidelity based on control plane outputs. To evaluate our
algorithm and protocol, we develop a discrete-time quantum network simulator.
Simulation results show the superior performance of our approach compared to
existing fidelity-agnostic and fidelity-aware solutions