36,814 research outputs found

    Efficient Computation of Distance Labeling for Decremental Updates in Large Dynamic Graphs

    Get PDF
    Since today's real-world graphs, such as social network graphs, are evolving all the time, it is of great importance to perform graph computations and analysis in these dynamic graphs. Due to the fact that many applications such as social network link analysis with the existence of inactive users need to handle failed links or nodes, decremental computation and maintenance for graphs is considered a challenging problem. Shortest path computation is one of the most fundamental operations for managing and analyzing large graphs. A number of indexing methods have been proposed to answer distance queries in static graphs. Unfortunately, there is little work on answering such queries for dynamic graphs. In this paper, we focus on the problem of computing the shortest path distance in dynamic graphs, particularly on decremental updates (i.e., edge deletions). We propose maintenance algorithms based on distance labeling, which can handle decremental updates efficiently. By exploiting properties of distance labeling in original graphs, we are able to efficiently maintain distance labeling for new graphs. We experimentally evaluate our algorithms using eleven real-world large graphs and confirm the effectiveness and efficiency of our approach. More specifically, our method can speed up index re-computation by up to an order of magnitude compared with the state-of-the-art method, Pruned Landmark Labeling (PLL)

    Using Light Spanning Graphs for Passenger Assignment in Public Transport

    Get PDF
    In a public transport network a passenger’s preferred route from a point x to another point y is usually the shortest path from x to y. However, it is simply impossible to provide all the shortest paths of a network via public transport. Hence, it is a natural question how a lighter sub-network should be designed in order to satisfy both the operator as well as the passengers.We provide a detailed analysis of the interplay of the following three quality measures of lighter public transport networks: - building cost: the sum of the costs of all edges remaining in the lighter network, - routing costs: the sum of all shortest paths costs weighted by the demands, - fairness: compared to the original network, for each two points the shortest path in the new network should cost at most a given multiple of the shortest path in the original network. We study the problem by generalizing the concepts of optimum communication spanning trees (Hu, 1974) and optimum requirement graphs (Wu, Chao, and Tang, 2002) to generalized optimum requirement graphs (GORGs), which are graphs achieving the social optimum amongst all subgraphs satisfying a given upper bound on the building cost. We prove that the corresponding decision problem is NP-complete, even on orb-webs, a variant of grids which serves as an important model of cities with a center. For the case that the given network is a parametric city (cf. Fielbaum et. al., 2017) with a heavy vertex we provide a polynomial-time algorithm solving the GORG-problem. Concerning the fairness-aspect, we prove that light spanners are a strong concept for public transport optimization. We underpin our theoretical considerations with integer programming-based experiments that allow us to compare the fairness-approach with the routing cost-approach as well as passenger assignment approaches from the literature.Peer reviewe

    Using Causality-Aware Graph Neural Networks to Predict Temporal Centralities in Dynamic Graphs

    Full text link
    Node centralities play a pivotal role in network science, social network analysis, and recommender systems. In temporal data, static path-based centralities like closeness or betweenness can give misleading results about the true importance of nodes in a temporal graph. To address this issue, temporal generalizations of betweenness and closeness have been defined that are based on the shortest time-respecting paths between pairs of nodes. However, a major issue of those generalizations is that the calculation of such paths is computationally expensive. Addressing this issue, we study the application of De Bruijn Graph Neural Networks (DBGNN), a causality-aware graph neural network architecture, to predict temporal path-based centralities in time series data. We experimentally evaluate our approach in 13 temporal graphs from biological and social systems and show that it considerably improves the prediction of both betweenness and closeness centrality compared to a static Graph Convolutional Neural Network

    Randomized shortest paths and their applications

    Get PDF
    In graph analysis, the Shortest Path problem identifies the optimal, most cost effective, path between two nodes. This problem has been the object of many studies and extensions in heterogeneous domains such as: speech recognition, social network analysis, biological sequence alignment, path planning, or zero-sum games among others. Although the shortest path focuses on the optimal cost of reaching a destination node, it does not take into account other useful information contained on the graph, such as the degree of connectivity of two nodes. On the other hand, measures taking connectivity information into account have their own drawbacks, specially when graphs become large. A new family of distances which interpolates between both extremes is introduced by the Randomized Shortest Path (RSP) framework. By spreading randomization through a graph, the RSP leads to applications where some degree of randomness would be desired. Through this work, we try to investigate whether the RSP framework can be applied to different domains in which randomization is useful, and either solve an existing problem with a new approach, or prove to outperform existing methods.(FSA - Sciences de l'ingénieur) -- UCL, 201

    Topology Discovery of Sparse Random Graphs With Few Participants

    Get PDF
    We consider the task of topology discovery of sparse random graphs using end-to-end random measurements (e.g., delay) between a subset of nodes, referred to as the participants. The rest of the nodes are hidden, and do not provide any information for topology discovery. We consider topology discovery under two routing models: (a) the participants exchange messages along the shortest paths and obtain end-to-end measurements, and (b) additionally, the participants exchange messages along the second shortest path. For scenario (a), our proposed algorithm results in a sub-linear edit-distance guarantee using a sub-linear number of uniformly selected participants. For scenario (b), we obtain a much stronger result, and show that we can achieve consistent reconstruction when a sub-linear number of uniformly selected nodes participate. This implies that accurate discovery of sparse random graphs is tractable using an extremely small number of participants. We finally obtain a lower bound on the number of participants required by any algorithm to reconstruct the original random graph up to a given edit distance. We also demonstrate that while consistent discovery is tractable for sparse random graphs using a small number of participants, in general, there are graphs which cannot be discovered by any algorithm even with a significant number of participants, and with the availability of end-to-end information along all the paths between the participants.Comment: A shorter version appears in ACM SIGMETRICS 2011. This version is scheduled to appear in J. on Random Structures and Algorithm

    Methods to Determine Node Centrality and Clustering in Graphs with Uncertain Structure

    Full text link
    Much of the past work in network analysis has focused on analyzing discrete graphs, where binary edges represent the "presence" or "absence" of a relationship. Since traditional network measures (e.g., betweenness centrality) utilize a discrete link structure, complex systems must be transformed to this representation in order to investigate network properties. However, in many domains there may be uncertainty about the relationship structure and any uncertainty information would be lost in translation to a discrete representation. Uncertainty may arise in domains where there is moderating link information that cannot be easily observed, i.e., links become inactive over time but may not be dropped or observed links may not always corresponds to a valid relationship. In order to represent and reason with these types of uncertainty, we move beyond the discrete graph framework and develop social network measures based on a probabilistic graph representation. More specifically, we develop measures of path length, betweenness centrality, and clustering coefficient---one set based on sampling and one based on probabilistic paths. We evaluate our methods on three real-world networks from Enron, Facebook, and DBLP, showing that our proposed methods more accurately capture salient effects without being susceptible to local noise, and that the resulting analysis produces a better understanding of the graph structure and the uncertainty resulting from its change over time.Comment: Longer version of paper appearing in Fifth International AAAI Conference on Weblogs and Social Media. 9 pages, 4 Figure

    Fast Shortest Path Distance Estimation in Large Networks

    Full text link
    We study the problem of preprocessing a large graph so that point-to-point shortest-path queries can be answered very fast. Computing shortest paths is a well studied problem, but exact algorithms do not scale to huge graphs encountered on the web, social networks, and other applications. In this paper we focus on approximate methods for distance estimation, in particular using landmark-based distance indexing. This approach involves selecting a subset of nodes as landmarks and computing (offline) the distances from each node in the graph to those landmarks. At runtime, when the distance between a pair of nodes is needed, we can estimate it quickly by combining the precomputed distances of the two nodes to the landmarks. We prove that selecting the optimal set of landmarks is an NP-hard problem, and thus heuristic solutions need to be employed. Given a budget of memory for the index, which translates directly into a budget of landmarks, different landmark selection strategies can yield dramatically different results in terms of accuracy. A number of simple methods that scale well to large graphs are therefore developed and experimentally compared. The simplest methods choose central nodes of the graph, while the more elaborate ones select central nodes that are also far away from one another. The efficiency of the suggested techniques is tested experimentally using five different real world graphs with millions of edges; for a given accuracy, they require as much as 250 times less space than the current approach in the literature which considers selecting landmarks at random. Finally, we study applications of our method in two problems arising naturally in large-scale networks, namely, social search and community detection.Yahoo! Research (internship
    • …
    corecore