30,376 research outputs found

    A local 2-approximation algorithm for the vertex cover problem

    Get PDF
    We present a distributed 2-approximation algorithm for the minimum vertex cover problem. The algorithm is deterministic, and it runs in (Δ + 1)2 synchronous communication rounds, where Δ is the maximum degree of the graph. For Δ = 3, we give a 2-approximation algorithm also for the weighted version of the problem.Peer reviewe

    Distributed CONGEST Approximation of Weighted Vertex Covers and Matchings

    Get PDF
    We provide CONGEST model algorithms for approximating minimum weighted vertex cover and the maximum weighted matching. For bipartite graphs, we show that a (1+ε)(1+\varepsilon)-approximate weighted vertex cover can be computed deterministically in polylogarithmic time. This generalizes a corresponding result for the unweighted vertex cover problem shown in [Faour, Kuhn; OPODIS '20]. Moreover, we show that in general weighted graph families that are closed under taking subgraphs and in which we can compute an independent set of weight at least a λ\lambda-fraction of the total weight, one can compute a (22λ+ε)(2-2\lambda +\varepsilon)-approximate weighted vertex cover in polylogarithmic time in the CONGEST model. Our result in particular implies that in graphs of arboricity aa, one can compute a (21/a+ε)(2-1/a+\varepsilon)-approximate weighted vertex cover. For maximum weighted matchings, we show that a (1ε)(1-\varepsilon)-approximate solution can be computed deterministically in polylogarithmic CONGEST rounds (for constant ε\varepsilon). We also provide a more efficient randomized algorithm. Our algorithm generalizes results of [Lotker, Patt-Shamir, Pettie; SPAA '08] and [Bar-Yehuda, Hillel, Ghaffari, Schwartzman; PODC '17] for the unweighted case. Finally, we show that even in the LOCAL model and in bipartite graphs of degree 3\leq 3, if ε<ε0\varepsilon<\varepsilon_0 for some constant ε0>0\varepsilon_0>0, then computing a (1+ε)(1+\varepsilon)-approximation for the unweighted minimum vertex cover problem requires Ω(lognε)\Omega\big(\frac{\log n}{\varepsilon}\big) rounds. This generalizes aresult of [G\"o\"os, Suomela; DISC '12], who showed that computing a (1+ε0)(1+\varepsilon_0)-approximation in such graphs requires Ω(logn)\Omega(\log n) rounds

    Faster Exponential-Time Approximation Algorithms Using Approximate Monotone Local Search

    Get PDF
    We generalize the monotone local search approach of Fomin, Gaspers, Lokshtanov and Saurabh [J.ACM 2019], by establishing a connection between parameterized approximation and exponential-time approximation algorithms for monotone subset minimization problems. In a monotone subset minimization problem the input implicitly describes a non-empty set family over a universe of size n which is closed under taking supersets. The task is to find a minimum cardinality set in this family. Broadly speaking, we use approximate monotone local search to show that a parameterized ?-approximation algorithm that runs in c^k?n^?(1) time, where k is the solution size, can be used to derive an ?-approximation randomized algorithm that runs in d??n^?(1) time, where d is the unique value in (1, 1+{c-1}/?) such that ?(1/??{d-1}/{c-1}) = {ln c}/? and ?(a?b) is the Kullback-Leibler divergence. This running time matches that of Fomin et al. for ? = 1, and is strictly better when ? > 1, for any c > 1. Furthermore, we also show that this result can be derandomized at the expense of a sub-exponential multiplicative factor in the running time. We use an approximate variant of the exhaustive search as a benchmark for our algorithm. We show that the classic 2??n^?(1) exhaustive search can be adapted to an ?-approximate exhaustive search that runs in time (1+exp(-???(1/(?))))??n^?(1), where ? is the entropy function. Furthermore, we provide a lower bound stating that the running time of this ?-approximate exhaustive search is the best achievable running time in an oracle model. When compared to approximate exhaustive search, and to other techniques, the running times obtained by approximate monotone local search are strictly better for any ? ? 1, c > 1. We demonstrate the potential of approximate monotone local search by deriving new and faster exponential approximation algorithms for Vertex Cover, 3-Hitting Set, Directed Feedback Vertex Set, Directed Subset Feedback Vertex Set, Directed Odd Cycle Transversal and Undirected Multicut. For instance, we get a 1.1-approximation algorithm for Vertex Cover with running time 1.114??n^?(1), improving upon the previously best known 1.1-approximation running in time 1.127??n^?(1) by Bourgeois et al. [DAM 2011]

    Dynamic set cover : improved amortized and worst-case update time

    Get PDF
    In the dynamic minimum set cover problem, a challenge is to minimize the update time while guaranteeing close to the optimal min(O(log n), f) approximation factor. (Throughout, m, n, f, and C are parameters denoting the maximum number of sets, number of elements, frequency, and the cost range.) In the high-frequency range, when f = Ω(log n), this was achieved by a deterministic O(log n)-approximation algorithm with O(f log n) amortized update time [Gupta et al. STOC'17]. In the low-frequency range, the line of work by Gupta et al. [STOC'17], Abboud et al. [STOC'19], and Bhattacharya et al. [ICALP'15, IPCO'17, FOCS'19] led to a deterministic (1 + ∊) f-approximation algorithm with O(f log(Cn)/∊2) amortized update time. In this paper we improve the latter update time and provide the first bounds that subsume (and sometimes improve) the state-of-the-art dynamic vertex cover algorithms. We obtain: (1) (1 + ∊) f-approximation ratio in O(f log2(Cn)/∊3) worst-case update time: No non-trivial worst-case update time was previously known for dynamic set cover. Our bound subsumes and improves by a logarithmic factor the O(log3 n/poly(∊)) worst-case update time for unweighted dynamic vertex cover (i.e., when f = 2 and C = 1) by Bhattacharya et al. [SODA'17]. (2) (1 + ∊) f-approximation ratio in O ((f2/∊3) + (f/∊2) log C) amortized update time: This result improves the previous O(f log (Cn)/∊2) update time bound for most values of f in the low-frequency range, i.e. whenever f = o(log n). It is the first that is independent of m and n. It subsumes the constant amortized update time of Bhattacharya and Kulkarni [SODA'19] for unweighted dynamic vertex cover (i.e., when f = 2 and C = 1). These results are achieved by leveraging the approximate complementary slackness and background schedulers techniques. These techniques were used in the local update scheme for dynamic vertex cover. Our main technical contribution is to adapt these techniques within the global update scheme of Bhattacharya et al. [FOCS'19] for the dynamic set cover problem

    Parameterized Distributed Algorithms

    Get PDF
    In this work, we initiate a thorough study of graph optimization problems parameterized by the output size in the distributed setting. In such a problem, an algorithm decides whether a solution of size bounded by k exists and if so, it finds one. We study fundamental problems, including Minimum Vertex Cover (MVC), Maximum Independent Set (MaxIS), Maximum Matching (MaxM), and many others, in both the LOCAL and CONGEST distributed computation models. We present lower bounds for the round complexity of solving parameterized problems in both models, together with optimal and near-optimal upper bounds. Our results extend beyond the scope of parameterized problems. We show that any LOCAL (1+epsilon)-approximation algorithm for the above problems must take Omega(epsilon^{-1}) rounds. Joined with the (epsilon^{-1}log n)^{O(1)} rounds algorithm of [Ghaffari et al., 2017] and the Omega (sqrt{(log n)/(log log n)}) lower bound of [Fabian Kuhn et al., 2016], the lower bounds match the upper bound up to polynomial factors in both parameters. We also show that our parameterized approach reduces the runtime of exact and approximate CONGEST algorithms for MVC and MaxM if the optimal solution is small, without knowing its size beforehand. Finally, we propose the first o(n^2) rounds CONGEST algorithms that approximate MVC within a factor strictly smaller than 2

    Optimally Repurposing Existing Algorithms to Obtain Exponential-Time Approximations

    Full text link
    The goal of this paper is to understand how exponential-time approximation algorithms can be obtained from existing polynomial-time approximation algorithms, existing parameterized exact algorithms, and existing parameterized approximation algorithms. More formally, we consider a monotone subset minimization problem over a universe of size nn (e.g., Vertex Cover or Feedback Vertex Set). We have access to an algorithm that finds an α\alpha-approximate solution in time cknO(1)c^k \cdot n^{O(1)} if a solution of size kk exists (and more generally, an extension algorithm that can approximate in a similar way if a set can be extended to a solution with kk further elements). Our goal is to obtain a dnnO(1)d^n \cdot n^{O(1)} time β\beta-approximation algorithm for the problem with dd as small as possible. That is, for every fixed α,c,β1\alpha,c,\beta \geq 1, we would like to determine the smallest possible dd that can be achieved in a model where our problem-specific knowledge is limited to checking the feasibility of a solution and invoking the α\alpha-approximate extension algorithm. Our results completely resolve this question: (1) For every fixed α,c,β1\alpha,c,\beta \geq 1, a simple algorithm (``approximate monotone local search'') achieves the optimum value of dd. (2) Given α,c,β1\alpha,c,\beta \geq 1, we can efficiently compute the optimum dd up to any precision ε>0\varepsilon > 0. Earlier work presented algorithms (but no lower bounds) for the special case α=β=1\alpha = \beta = 1 [Fomin et al., J. ACM 2019] and for the special case α=β>1\alpha = \beta > 1 [Esmer et al., ESA 2022]. Our work generalizes these results and in particular confirms that the earlier algorithms are optimal in these special cases.Comment: 80 pages, 5 figure

    Approximation Algorithms for Distributionally Robust Stochastic Optimization

    Get PDF
    Two-stage stochastic optimization is a widely used framework for modeling uncertainty, where we have a probability distribution over possible realizations of the data, called scenarios, and decisions are taken in two stages: we take first-stage actions knowing only the underlying distribution and before a scenario is realized, and may take additional second-stage recourse actions after a scenario is realized. The goal is typically to minimize the total expected cost. A common criticism levied at this model is that the underlying probability distribution is itself often imprecise. To address this, an approach that is quite versatile and has gained popularity in the stochastic-optimization literature is the two-stage distributionally robust stochastic model: given a collection D of probability distributions, our goal now is to minimize the maximum expected total cost with respect to a distribution in D. There has been almost no prior work however on developing approximation algorithms for distributionally robust problems where the underlying scenario collection is discrete, as is the case with discrete-optimization problems. We provide frameworks for designing approximation algorithms in such settings when the collection D is a ball around a central distribution, defined relative to two notions of distance between probability distributions: Wasserstein metrics (which include the L_1 metric) and the L_infinity metric. Our frameworks yield efficient algorithms even in settings with an exponential number of scenarios, where the central distribution may only be accessed via a sampling oracle. For distributionally robust optimization under a Wasserstein ball, we first show that one can utilize the sample average approximation (SAA) method (solve the distributionally robust problem with an empirical estimate of the central distribution) to reduce the problem to the case where the central distribution has a polynomial-size support, and is represented explicitly. This follows because we argue that a distributionally robust problem can be reduced in a novel way to a standard two-stage stochastic problem with bounded inflation factor, which enables one to use the SAA machinery developed for two-stage stochastic problems. Complementing this, we show how to approximately solve a fractional relaxation of the SAA problem (i.e., the distributionally robust problem obtained by replacing the original central distribution with its empirical estimate). Unlike in two-stage {stochastic, robust} optimization with polynomially many scenarios, this turns out to be quite challenging. We utilize a variant of the ellipsoid method for convex optimization in conjunction with several new ideas to show that the SAA problem can be approximately solved provided that we have an (approximation) algorithm for a certain max-min problem that is akin to, and generalizes, the k-max-min problem (find the worst-case scenario consisting of at most k elements) encountered in two-stage robust optimization. We obtain such an algorithm for various discrete-optimization problems; by complementing this via rounding algorithms that provide local (i.e., per-scenario) approximation guarantees, we obtain the first approximation algorithms for the distributionally robust versions of a variety of discrete-optimization problems including set cover, vertex cover, edge cover, facility location, and Steiner tree, with guarantees that are, except for set cover, within O(1)-factors of the guarantees known for the deterministic version of the problem. For distributionally robust optimization under an L_infinity ball, we consider a fractional relaxation of the problem, and replace its objective function with a proxy function that is pointwise close to the true objective function (within a factor of 2). We then show that we can efficiently compute approximate subgradients of the proxy function, provided that we have an algorithm for the problem of computing the t worst scenarios under a given first-stage decision, given an integer t. We can then approximately minimize the proxy function via a variant of the ellipsoid method, and thus obtain an approximate solution for the fractional relaxation of the distributionally robust problem. Complementing this via rounding algorithms with local guarantees, we obtain approximation algorithms for distributionally robust versions of various covering problems, including set cover, vertex cover, edge cover, and facility location, with guarantees that are within O(1)-factors of the guarantees known for their deterministic versions

    Faster Exponential-Time Approximation Algorithms Using Approximate Monotone Local Search

    Get PDF
    We generalize the monotone local search approach of Fomin, Gaspers, Lokshtanov and Saurabh [J.ACM 2019], by establishing a connection between parameterized approximation and exponential-time approximation algorithms for {\em monotone subset minimization} problems. In a {\em monotone subset minimization} problem the input implicitly describes a non-empty set family over a universe of size nn which is closed under taking supersets. The task is to find a minimum cardinality set in this family. Broadly speaking, we use {\em approximate monotone local search} to show that a parameterized α\alpha-approximation algorithm that runs in c^k \cdot n^{\OO(1)} time, where kk is the solution size, can be used to derive an α\alpha-approximation randomized algorithm that runs in d^n \cdot n^{\OO(1)} time, where dd is the unique value in d(1,1+c1α)d\in \left (1, 1+\frac{c-1}{\alpha} \right) such that \D{\frac{1}{\alpha}}{\frac{d-1}{c-1}} =\frac{\ln c }{\alpha} and \D{a}{b} is the Kullback-Leibler divergence. This running time matches that of Fomin et al.\ for α=1\alpha=1, and is strictly better when α>1\alpha >1, for any c>1c >1. Furthermore, we also show that this result can be derandomized at the expense of a sub-exponential multiplicative factor in the running time. We use an approximate variant of the exhaustive search as a benchmark for our algorithm. We show that the classic 2^n \cdot n^{\OO(1)} exhaustive search can be adapted to an α\alpha-approximate exhaustive search that runs in time \left ( 1+ \exp\left (-\alpha \cdot \entropy\left (\frac{1}{\alpha}\right)\right)\right)^n \cdot n^{\OO(1)}, where \entropy is the entropy function. Furthermore, we provide a lower bound stating that the running time of this α\alpha-approximate exhaustive search is the best achievable running time in an oracle model. When compared to approximate exhaustive search, and to other techniques, the running times obtained by approximate monotone local search are strictly better for any α1, c>1\alpha \geq 1,~c >1. We demonstrate the potential of approximate monotone local search by deriving new and faster exponential approximation algorithms for {\sc Vertex Cover}, {\sc 33-Hitting Set}, {\sc Directed Feedback Vertex Set}, {\sc Directed Subset Feedback Vertex Set}, {\sc Directed Odd Cycle Transversal} and {\sc Undirected Multicut}. For instance, we get a 1.11.1-approximation algorithm for {\sc Vertex Cover} with running time 1.114^n \cdot n^{\OO(1)}, improving upon the previously best known 1.11.1-approximation running in time 1.127^n \cdot n^{\OO(1)} by Bourgeois et al.\ [DAM 2011]
    corecore