13,701 research outputs found

    Fast Algorithms for Parameterized Problems with Relaxed Disjointness Constraints

    Full text link
    In parameterized complexity, it is a natural idea to consider different generalizations of classic problems. Usually, such generalization are obtained by introducing a "relaxation" variable, where the original problem corresponds to setting this variable to a constant value. For instance, the problem of packing sets of size at most pp into a given universe generalizes the Maximum Matching problem, which is recovered by taking p=2p=2. Most often, the complexity of the problem increases with the relaxation variable, but very recently Abasi et al. have given a surprising example of a problem --- rr-Simple kk-Path --- that can be solved by a randomized algorithm with running time O(2O(klogrr))O^*(2^{O(k \frac{\log r}{r})}). That is, the complexity of the problem decreases with rr. In this paper we pursue further the direction sketched by Abasi et al. Our main contribution is a derandomization tool that provides a deterministic counterpart of the main technical result of Abasi et al.: the O(2O(klogrr))O^*(2^{O(k \frac{\log r}{r})}) algorithm for (r,k)(r,k)-Monomial Detection, which is the problem of finding a monomial of total degree kk and individual degrees at most rr in a polynomial given as an arithmetic circuit. Our technique works for a large class of circuits, and in particular it can be used to derandomize the result of Abasi et al. for rr-Simple kk-Path. On our way to this result we introduce the notion of representative sets for multisets, which may be of independent interest. Finally, we give two more examples of problems that were already studied in the literature, where the same relaxation phenomenon happens. The first one is a natural relaxation of the Set Packing problem, where we allow the packed sets to overlap at each element at most rr times. The second one is Degree Bounded Spanning Tree, where we seek for a spanning tree of the graph with a small maximum degree

    More Applications of the d-Neighbor Equivalence: Connectivity and Acyclicity Constraints

    Get PDF
    In this paper, we design a framework to obtain efficient algorithms for several problems with a global constraint (acyclicity or connectivity) such as Connected Dominating Set, Node Weighted Steiner Tree, Maximum Induced Tree, Longest Induced Path, and Feedback Vertex Set. For all these problems, we obtain 2^O(k)* n^O(1), 2^O(k log(k))* n^O(1), 2^O(k^2) * n^O(1) and n^O(k) time algorithms parameterized respectively by clique-width, Q-rank-width, rank-width and maximum induced matching width. Our approach simplifies and unifies the known algorithms for each of the parameters and match asymptotically also the running time of the best algorithms for basic NP-hard problems such as Vertex Cover and Dominating Set. Our framework is based on the d-neighbor equivalence defined in [Bui-Xuan, Telle and Vatshelle, TCS 2013]. The results we obtain highlight the importance and the generalizing power of this equivalence relation on width measures. We also prove that this equivalence relation could be useful for Max Cut: a W[1]-hard problem parameterized by clique-width. For this latter problem, we obtain n^O(k), n^O(k) and n^(2^O(k)) time algorithm parameterized by clique-width, Q-rank-width and rank-width

    Patching Colors with Tensors

    Get PDF

    Parameterized Algorithms for Graph Partitioning Problems

    Full text link
    We study a broad class of graph partitioning problems, where each problem is specified by a graph G=(V,E)G=(V,E), and parameters kk and pp. We seek a subset UVU\subseteq V of size kk, such that α1m1+α2m2\alpha_1m_1 + \alpha_2m_2 is at most (or at least) pp, where α1,α2R\alpha_1,\alpha_2\in\mathbb{R} are constants defining the problem, and m1,m2m_1, m_2 are the cardinalities of the edge sets having both endpoints, and exactly one endpoint, in UU, respectively. This class of fixed cardinality graph partitioning problems (FGPP) encompasses Max (k,nk)(k,n-k)-Cut, Min kk-Vertex Cover, kk-Densest Subgraph, and kk-Sparsest Subgraph. Our main result is an O(4k+o(k)Δk)O^*(4^{k+o(k)}\Delta^k) algorithm for any problem in this class, where Δ1\Delta \geq 1 is the maximum degree in the input graph. This resolves an open question posed by Bonnet et al. [IPEC 2013]. We obtain faster algorithms for certain subclasses of FGPPs, parameterized by pp, or by (k+p)(k+p). In particular, we give an O(4p+o(p))O^*(4^{p+o(p)}) time algorithm for Max (k,nk)(k,n-k)-Cut, thus improving significantly the best known O(pp)O^*(p^p) time algorithm

    Speeding-up Dynamic Programming with Representative Sets - An Experimental Evaluation of Algorithms for Steiner Tree on Tree Decompositions

    Full text link
    Dynamic programming on tree decompositions is a frequently used approach to solve otherwise intractable problems on instances of small treewidth. In recent work by Bodlaender et al., it was shown that for many connectivity problems, there exist algorithms that use time, linear in the number of vertices, and single exponential in the width of the tree decomposition that is used. The central idea is that it suffices to compute representative sets, and these can be computed efficiently with help of Gaussian elimination. In this paper, we give an experimental evaluation of this technique for the Steiner Tree problem. A comparison of the classic dynamic programming algorithm and the improved dynamic programming algorithm that employs the table reduction shows that the new approach gives significant improvements on the running time of the algorithm and the size of the tables computed by the dynamic programming algorithm, and thus that the rank based approach from Bodlaender et al. does not only give significant theoretical improvements but also is a viable approach in a practical setting, and showcases the potential of exploiting the idea of representative sets for speeding up dynamic programming algorithms

    More applications of the d-neighbor equivalence: acyclicity and connectivity constraints

    Full text link
    In this paper, we design a framework to obtain efficient algorithms for several problems with a global constraint (acyclicity or connectivity) such as Connected Dominating Set, Node Weighted Steiner Tree, Maximum Induced Tree, Longest Induced Path, and Feedback Vertex Set. We design a meta-algorithm that solves all these problems and whose running time is upper bounded by 2O(k)nO(1)2^{O(k)}\cdot n^{O(1)}, 2O(klog(k))nO(1)2^{O(k \log(k))}\cdot n^{O(1)}, 2O(k2)nO(1)2^{O(k^2)}\cdot n^{O(1)} and nO(k)n^{O(k)} where kk is respectively the clique-width, Q\mathbb{Q}-rank-width, rank-width and maximum induced matching width of a given decomposition. Our meta-algorithm simplifies and unifies the known algorithms for each of the parameters and its running time matches asymptotically also the running times of the best known algorithms for basic NP-hard problems such as Vertex Cover and Dominating Set. Our framework is based on the dd-neighbor equivalence defined in [Bui-Xuan, Telle and Vatshelle, TCS 2013]. The results we obtain highlight the importance of this equivalence relation on the algorithmic applications of width measures. We also prove that our framework could be useful for W[1]W[1]-hard problems parameterized by clique-width such as Max Cut and Maximum Minimal Cut. For these latter problems, we obtain nO(k)n^{O(k)}, nO(k)n^{O(k)} and n2O(k)n^{2^{O(k)}} time algorithms where kk is respectively the clique-width, the Q\mathbb{Q}-rank-width and the rank-width of the input graph

    Fair Knapsack

    Full text link
    We study the following multiagent variant of the knapsack problem. We are given a set of items, a set of voters, and a value of the budget; each item is endowed with a cost and each voter assigns to each item a certain value. The goal is to select a subset of items with the total cost not exceeding the budget, in a way that is consistent with the voters' preferences. Since the preferences of the voters over the items can vary significantly, we need a way of aggregating these preferences, in order to select the socially best valid knapsack. We study three approaches to aggregating voters' preferences, which are motivated by the literature on multiwinner elections and fair allocation. This way we introduce the concepts of individually best, diverse, and fair knapsack. We study the computational complexity (including parameterized complexity, and complexity under restricted domains) of the aforementioned multiagent variants of knapsack.Comment: Extended abstract will appear in Proc. of 33rd AAAI 201
    corecore