173 research outputs found
Recommended from our members
GRAPH REPRESENTATION LEARNING WITH BOX EMBEDDINGS
Graphs are ubiquitous data structures, present in many machine-learning tasks, such as link prediction of products and node classification of scientific papers. As gradient descent drives the training of most modern machine learning architectures, the ability to encode graph-structured data using a differentiable representation is essential to make use of this data. Most approaches encode graph structure in Euclidean space, however, it is non-trivial to model directed edges. The naive solution is to represent each node using a separate source and target vector, however, this can decouple the representation, making it harder for the model to capture information within longer paths in the graph.
In this dissertation, we propose to model graphs by representing each node as a \textit{box} (a Cartesian product of intervals) where directed edges are captured by the relative containment of one box in another. Theoretical proof shows that our proposed box embeddings have the expressiveness to represent any \emph{directed acyclic graph}. We also perform rigorous empirical evaluations of vector, hyperbolic, and region-based geometric representations on several families of synthetic and real-world directed graphs. Extensive experimental results suggest that the box containment can allow for transitive relationships to be modeled easily. We further propose t-Box, a variant of box embeddings that learns the temperature together during training. t-Box uses a learned smoothing parameter to achieve better representational capacity than vector models in low dimensions, while also avoiding performance saturation common to other geometric models in high dimensions.
Though promising, modeling directed graphs that both contain cycles and some element of transitivity, two properties common in real-world settings, is challenging. Box embeddings, which can be thought of as representing the graph as an intersection over some learned super-graphs, have a natural inductive bias toward modeling transitivity, but (as we prove) cannot model cycles. To address this issue, we propose binary code box embeddings, where a learned binary code selects a subset of graphs for intersection. We explore several variants, including global binary codes (amounting to a union over intersections) and per-vertex binary codes (allowing greater flexibility) as well as methods of regularization. Theoretical and empirical results show that the proposed models not only preserve a useful inductive bias of transitivity but also have sufficient representational capacity to model arbitrary graphs, including graphs with cycles.
Lastly, we discuss the use case where box embeddings are not free parameters but are produced by functions. In particular, we explore whether neural networks can map node features into the box space. This is critical in many real-world scenarios. On the one hand, graphs are sparse and the majority of vertices only have few connections or are completely isolated. On the other hand, there may exist rich node features such as attributes and descriptions, that could be useful for prediction tasks. The experimental analysis points out both the effectiveness and insufficiency of multi-layer perceptron-based encoders under different circumstances
Height in splittings of hyperbolic groups
Suppose is a hyperbolic subgroup of a hyperbolic group . Assume there
exists such that the intersection of essentially distinct
conjugates of is always finite. Further assume splits over with
hyperbolic vertex and edge groups and the two inclusions of are
quasi-isometric embeddings. Then is quasiconvex in . This answers a
question of Swarup and provides a partial converse to the main theorem of
\cite{GMRS}.Comment: 16 pages, no figures, no table
Topologically Trivial Closed Walks in Directed Surface Graphs
Let G be a directed graph with n vertices and m edges, embedded on a surface S, possibly with boundary, with first Betti number beta. We consider the complexity of finding closed directed walks in G that are either contractible (trivial in homotopy) or bounding (trivial in integer homology) in S. Specifically, we describe algorithms to determine whether G contains a simple contractible cycle in O(n+m) time, or a contractible closed walk in O(n+m) time, or a bounding closed walk in O(beta (n+m)) time. Our algorithms rely on subtle relationships between strong connectivity in G and in the dual graph G^*; our contractible-closed-walk algorithm also relies on a seminal topological result of Hass and Scott. We also prove that detecting simple bounding cycles is NP-hard.
We also describe three polynomial-time algorithms to compute shortest contractible closed walks, depending on whether the fundamental group of the surface is free, abelian, or hyperbolic. A key step in our algorithm for hyperbolic surfaces is the construction of a context-free grammar with O(g^2L^2) non-terminals that generates all contractible closed walks of length at most L, and only contractible closed walks, in a system of quads of genus g >= 2. Finally, we show that computing shortest simple contractible cycles, shortest simple bounding cycles, and shortest bounding closed walks are all NP-hard
Pseudo-Riemannian Graph Convolutional Networks
Graph convolutional networks (GCNs) are powerful frameworks for learning
embeddings of graph-structured data. GCNs are traditionally studied through the
lens of Euclidean geometry. Recent works find that non-Euclidean Riemannian
manifolds provide specific inductive biases for embedding hierarchical or
spherical data. However, they cannot align well with data of mixed graph
topologies. We consider a larger class of pseudo-Riemannian manifolds that
generalize hyperboloid and sphere. We develop new geodesic tools that allow for
extending neural network operations into geodesically disconnected
pseudo-Riemannian manifolds. As a consequence, we derive a pseudo-Riemannian
GCN that models data in pseudo-Riemannian manifolds of constant nonzero
curvature in the context of graph neural networks. Our method provides a
geometric inductive bias that is sufficiently flexible to model mixed
heterogeneous topologies like hierarchical graphs with cycles. We demonstrate
the representational capabilities of this method by applying it to the tasks of
graph reconstruction, node classification and link prediction on a series of
standard graphs with mixed topologies. Empirical results demonstrate that our
method outperforms Riemannian counterparts when embedding graphs of complex
topologies.Comment: 20 page
- …