69,589 research outputs found

    P-matrix recognition is co-NP-complete

    Get PDF
    This is a summary of the proof by G.E. Coxson that P-matrix recognition is co-NP-complete. The result follows by a reduction from the MAX CUT problem using results of S. Poljak and J. Rohn.Comment: 9 page

    Computational complexity of μ calculation

    Get PDF
    The structured singular value μ measures the robustness of uncertain systems. Numerous researchers over the last decade have worked on developing efficient methods for computing μ. This paper considers the complexity of calculating μ with general mixed real/complex uncertainty in the framework of combinatorial complexity theory. In particular, it is proved that the μ recognition problem with either pure real or mixed real/complex uncertainty is NP-hard. This strongly suggests that it is futile to pursue exact methods for calculating μ of general systems with pure real or mixed uncertainty for other than small problems

    On the Hardness of Entropy Minimization and Related Problems

    Full text link
    We investigate certain optimization problems for Shannon information measures, namely, minimization of joint and conditional entropies H(X,Y)H(X,Y), H(XY)H(X|Y), H(YX)H(Y|X), and maximization of mutual information I(X;Y)I(X;Y), over convex regions. When restricted to the so-called transportation polytopes (sets of distributions with fixed marginals), very simple proofs of NP-hardness are obtained for these problems because in that case they are all equivalent, and their connection to the well-known \textsc{Subset sum} and \textsc{Partition} problems is revealed. The computational intractability of the more general problems over arbitrary polytopes is then a simple consequence. Further, a simple class of polytopes is shown over which the above problems are not equivalent and their complexity differs sharply, namely, minimization of H(X,Y)H(X,Y) and H(YX)H(Y|X) is trivial, while minimization of H(XY)H(X|Y) and maximization of I(X;Y)I(X;Y) are strongly NP-hard problems. Finally, two new (pseudo)metrics on the space of discrete probability distributions are introduced, based on the so-called variation of information quantity, and NP-hardness of their computation is shown.Comment: IEEE Information Theory Workshop (ITW) 201

    Filtering graphs to check isomorphism and extracting mapping by using the Conductance Electrical Model

    Get PDF
    © 2016. This manuscript version is made available under the CC-BY-NC-ND 4.0 license http://creativecommons.org/licenses/by-nc-nd/4.0/This paper presents a new method of filtering graphs to check exact graph isomorphism and extracting their mapping. Each graph is modeled by a resistive electrical circuit using the Conductance Electrical Model (CEM). By using this model, a necessary condition to check the isomorphism of two graphs is that their equivalent resistances have the same values, but this is not enough, and we have to look for their mapping to find the sufficient condition. We can compute the isomorphism between two graphs in O(N-3), where N is the order of the graph, if their star resistance values are different, otherwise the computational time is exponential, but only with respect to the number of repeated star resistance values, which usually is very small. We can use this technique to filter graphs that are not isomorphic and in case that they are, we can obtain their node mapping. A distinguishing feature over other methods is that, even if there exists repeated star resistance values, we can extract a partial node mapping (of all the nodes except the repeated ones and their neighbors) in O(N-3). The paper presents the method and its application to detect isomorphic graphs in two well know graph databases, where some graphs have more than 600 nodes. (C) 2016 Elsevier Ltd. All rights reserved.Postprint (author's draft

    Unit Grid Intersection Graphs: Recognition and Properties

    Full text link
    It has been known since 1991 that the problem of recognizing grid intersection graphs is NP-complete. Here we use a modified argument of the above result to show that even if we restrict to the class of unit grid intersection graphs (UGIGs), the recognition remains hard, as well as for all graph classes contained inbetween. The result holds even when considering only graphs with arbitrarily large girth. Furthermore, we ask the question of representing UGIGs on grids of minimal size. We show that the UGIGs that can be represented in a square of side length 1+epsilon, for a positive epsilon no greater than 1, are exactly the orthogonal ray graphs, and that there exist families of trees that need an arbitrarily large grid

    Bounded Search Tree Algorithms for Parameterized Cograph Deletion: Efficient Branching Rules by Exploiting Structures of Special Graph Classes

    Full text link
    Many fixed-parameter tractable algorithms using a bounded search tree have been repeatedly improved, often by describing a larger number of branching rules involving an increasingly complex case analysis. We introduce a novel and general search strategy that branches on the forbidden subgraphs of a graph class relaxation. By using the class of P4P_4-sparse graphs as the relaxed graph class, we obtain efficient bounded search tree algorithms for several parameterized deletion problems. We give the first non-trivial bounded search tree algorithms for the cograph edge-deletion problem and the trivially perfect edge-deletion problems. For the cograph vertex deletion problem, a refined analysis of the runtime of our simple bounded search algorithm gives a faster exponential factor than those algorithms designed with the help of complicated case distinctions and non-trivial running time analysis [21] and computer-aided branching rules [11].Comment: 23 pages. Accepted in Discrete Mathematics, Algorithms and Applications (DMAA

    Compressive PCA for Low-Rank Matrices on Graphs

    Get PDF
    We introduce a novel framework for an approxi- mate recovery of data matrices which are low-rank on graphs, from sampled measurements. The rows and columns of such matrices belong to the span of the first few eigenvectors of the graphs constructed between their rows and columns. We leverage this property to recover the non-linear low-rank structures efficiently from sampled data measurements, with a low cost (linear in n). First, a Resrtricted Isometry Property (RIP) condition is introduced for efficient uniform sampling of the rows and columns of such matrices based on the cumulative coherence of graph eigenvectors. Secondly, a state-of-the-art fast low-rank recovery method is suggested for the sampled data. Finally, several efficient, parallel and parameter-free decoders are presented along with their theoretical analysis for decoding the low-rank and cluster indicators for the full data matrix. Thus, we overcome the computational limitations of the standard linear low-rank recovery methods for big datasets. Our method can also be seen as a major step towards efficient recovery of non- linear low-rank structures. For a matrix of size n X p, on a single core machine, our method gains a speed up of p2/kp^2/k over Robust Principal Component Analysis (RPCA), where k << p is the subspace dimension. Numerically, we can recover a low-rank matrix of size 10304 X 1000, 100 times faster than Robust PCA
    corecore