93,507 research outputs found
TPGNN: Learning High-order Information in Dynamic Graphs via Temporal Propagation
Temporal graph is an abstraction for modeling dynamic systems that consist of
evolving interaction elements. In this paper, we aim to solve an important yet
neglected problem -- how to learn information from high-order neighbors in
temporal graphs? -- to enhance the informativeness and discriminativeness for
the learned node representations. We argue that when learning high-order
information from temporal graphs, we encounter two challenges, i.e.,
computational inefficiency and over-smoothing, that cannot be solved by
conventional techniques applied on static graphs. To remedy these deficiencies,
we propose a temporal propagation-based graph neural network, namely TPGNN. To
be specific, the model consists of two distinct components, i.e., propagator
and node-wise encoder. The propagator is leveraged to propagate messages from
the anchor node to its temporal neighbors within -hop, and then
simultaneously update the state of neighborhoods, which enables efficient
computation, especially for a deep model. In addition, to prevent
over-smoothing, the model compels the messages from -hop neighbors to update
the -hop memory vector preserved on the anchor. The node-wise encoder adopts
transformer architecture to learn node representations by explicitly learning
the importance of memory vectors preserved on the node itself, that is,
implicitly modeling the importance of messages from neighbors at different
layers, thus mitigating the over-smoothing. Since the encoding process will not
query temporal neighbors, we can dramatically save time consumption in
inference. Extensive experiments on temporal link prediction and node
classification demonstrate the superiority of TPGNN over state-of-the-art
baselines in efficiency and robustness.Comment: Under revie
Network classification in temporal networks using motifs
Network classification has a variety of applications, such as detecting communities within networks and finding similarities between those representing different aspects of the real world. However, most existing work in this area focus on examining static undirected networks without considering directed edges or temporality. In this paper, we propose a new methodology that utilizes feature representation for network classification based on the temporal motif distribution of the network and a null model for comparing against random graphs. Experimental results show that our method improves accuracy by up 10% compared to the state-of-the-art embedding method in network classification, for tasks such as classifying network type, identifying communities in email exchange network, and identifying users given their app-switching behaviors
STWalk: Learning Trajectory Representations in Temporal Graphs
Analyzing the temporal behavior of nodes in time-varying graphs is useful for
many applications such as targeted advertising, community evolution and outlier
detection. In this paper, we present a novel approach, STWalk, for learning
trajectory representations of nodes in temporal graphs. The proposed framework
makes use of structural properties of graphs at current and previous time-steps
to learn effective node trajectory representations. STWalk performs random
walks on a graph at a given time step (called space-walk) as well as on graphs
from past time-steps (called time-walk) to capture the spatio-temporal behavior
of nodes. We propose two variants of STWalk to learn trajectory
representations. In one algorithm, we perform space-walk and time-walk as part
of a single step. In the other variant, we perform space-walk and time-walk
separately and combine the learned representations to get the final trajectory
embedding. Extensive experiments on three real-world temporal graph datasets
validate the effectiveness of the learned representations when compared to
three baseline methods. We also show the goodness of the learned trajectory
embeddings for change point detection, as well as demonstrate that arithmetic
operations on these trajectory representations yield interesting and
interpretable results.Comment: 10 pages, 5 figures, 2 table
- …