4,754 research outputs found

    Demystification of Graph and Information Entropy

    Get PDF
    Shannon entropy is an information-theoretic measure of unpredictability in probabilistic models. Recently, it has been used to form a tool, called the von Neumann entropy, to study quantum mechanics and network flows by appealing to algebraic properties of graph matrices. But still, little is known about what the von Neumann entropy says about the combinatorial structure of the graphs themselves. This paper gives a new formulation of the von Neumann entropy that describes it as a rate at which random movement settles down in a graph. At the same time, this new perspective gives rise to a generalization of von Neumann entropy to directed graphs, thus opening a new branch of research. Finally, it is conjectured that a directed cycle maximizes von Neumann entropy for directed graphs on a fixed number of vertices

    On the Von Neumann Entropy of Graphs

    Full text link
    The von Neumann entropy of a graph is a spectral complexity measure that has recently found applications in complex networks analysis and pattern recognition. Two variants of the von Neumann entropy exist based on the graph Laplacian and normalized graph Laplacian, respectively. Due to its computational complexity, previous works have proposed to approximate the von Neumann entropy, effectively reducing it to the computation of simple node degree statistics. Unfortunately, a number of issues surrounding the von Neumann entropy remain unsolved to date, including the interpretation of this spectral measure in terms of structural patterns, understanding the relation between its two variants, and evaluating the quality of the corresponding approximations. In this paper we aim to answer these questions by first analysing and comparing the quadratic approximations of the two variants and then performing an extensive set of experiments on both synthetic and real-world graphs. We find that 1) the two entropies lead to the emergence of similar structures, but with some significant differences; 2) the correlation between them ranges from weakly positive to strongly negative, depending on the topology of the underlying graph; 3) the quadratic approximations fail to capture the presence of non-trivial structural patterns that seem to influence the value of the exact entropies; 4) the quality of the approximations, as well as which variant of the von Neumann entropy is better approximated, depends on the topology of the underlying graph

    The laplacian of a graph as a density matrix: a basic combinatorial approach to separability of mixed states

    Full text link
    We study entanglement properties of mixed density matrices obtained from combinatorial Laplacians. This is done by introducing the notion of the density matrix of a graph. We characterize the graphs with pure density matrices and show that the density matrix of a graph can be always written as a uniform mixture of pure density matrices of graphs. We consider the von Neumann entropy of these matrices and we characterize the graphs for which the minimum and maximum values are attained. We then discuss the problem of separability by pointing out that separability of density matrices of graphs does not always depend on the labelling of the vertices. We consider graphs with a tensor product structure and simple cases for which combinatorial properties are linked to the entanglement of the state. We calculate the concurrence of all graph on four vertices representing entangled states. It turns out that for some of these graphs the value of the concurrence is exactly fractional.Comment: 20 pages, 11 figure

    The von Neumann entropy of networks

    Full text link
    We normalize the combinatorial Laplacian of a graph by the degree sum, look at its eigenvalues as a probability distribution and then study its Shannon entropy. Equivalently, we represent a graph with a quantum mechanical state and study its von Neumann entropy. At the graph-theoretic level, this quantity may be interpreted as a measure of regularity; it tends to be larger in relation to the number of connected components, long paths and nontrivial symmetries. When the set of vertices is asymptotically large, we prove that regular graphs and the complete graph have equal entropy, and specifically it turns out to be maximum. On the other hand, when the number of edges is fixed, graphs with large cliques appear to minimize the entropy.Comment: 6 pages, 3 figure
    corecore