7 research outputs found

    Vector Centrality in Hypergraphs

    Full text link
    Identifying the most influential nodes in networked systems is of vital importance to optimize their function and control. Several scalar metrics have been proposed to that effect, but the recent shift in focus towards network structures which go beyond a simple collection of dyadic interactions has rendered them void of performance guarantees. We here introduce a new measure of node's centrality, which is no longer a scalar value, but a vector with dimension one lower than the highest order of interaction in a hypergraph. Such a vectorial measure is linked to the eigenvector centrality for networks containing only dyadic interactions, but it has a significant added value in all other situations where interactions occur at higher-orders. In particular, it is able to unveil different roles which may be played by the same node at different orders of interactions -- information that is otherwise impossible to retrieve by single scalar measures. We demonstrate the efficacy of our measure with applications to synthetic networks and to three real world hypergraphs, and compare our results with those obtained by applying other scalar measures of centrality proposed in the literature.Comment: 10 pages, 9 figure

    On the stability of the Erdős-Ko-Rado theorem

    No full text
    Delete the edges of a Kneser graph independently of each other with some probability: for what probabilities is the independence number of this random graph equal to the independence number of the Kneser graph itself? We prove a sharp threshold result for this question in certain regimes. Since an independent set in the Kneser graph is the same as an intersecting (uniform) family, this gives us a random analogue of the Erdős-Ko-Rado theorem

    Learning Supervised PageRank with Gradient-Based and Gradient-Free Optimization Methods

    No full text
    In this paper, we consider a non-convex loss-minimization problem of learning Supervised PageRank models, which can account for features of nodes and edges. We propose gradient-based and random gradient-free methods to solve this problem. Our algorithms are based on the concept of an inexact oracle and unlike the state-of-the-art gradient-based method we manage to provide theoretically the convergence rate guarantees for both of them. Finally, we compare the performance of the proposed optimization methods with the state of the art applied to a ranking task

    Learning Supervised PageRank with Gradient-Based and Gradient-Free Optimization Methods

    No full text
    In this paper, we consider a non-convex loss-minimization problem of learning Supervised PageRank models, which can account for features of nodes and edges. We propose gradient-based and random gradient-free methods to solve this problem. Our algorithms are based on the concept of an inexact oracle and unlike the state-of-the-art gradient-based method we manage to provide theoretically the convergence rate guarantees for both of them. Finally, we compare the performance of the proposed optimization methods with the state of the art applied to a ranking task
    corecore