120 research outputs found

    Directed Subset Feedback Vertex Set Is Fixed-Parameter Tractable

    Get PDF
    Given a graph GG and an integer kk, the Feedback Vertex Set (FVS) problem asks if there is a vertex set TT of size at most kk that hits all cycles in the graph. The fixed-parameter tractability status of FVS in directed graphs was a long-standing open problem until Chen et al. (STOC '08) showed that it is FPT by giving a 4kk!nO(1)4^{k}k!n^{O(1)} time algorithm. In the subset versions of this problems, we are given an additional subset SS of vertices (resp., edges) and we want to hit all cycles passing through a vertex of SS (resp. an edge of SS). Recently, the Subset Feedback Vertex Set in undirected graphs was shown to be FPT by Cygan et al. (ICALP '11) and independently by Kakimura et al. (SODA '12). We generalize the result of Chen et al. (STOC '08) by showing that Subset Feedback Vertex Set in directed graphs can be solved in time 2O(k3)nO(1)2^{O(k^3)}n^{O(1)}. By our result, we complete the picture for feedback vertex set problems and their subset versions in undirected and directed graphs. Besides proving the fixed-parameter tractability of Directed Subset Feedback Vertex Set, we reformulate the random sampling of important separators technique in an abstract way that can be used for a general family of transversal problems. Moreover, we modify the probability distribution used in the technique to achieve better running time; in particular, this gives an improvement from 22O(k)2^{2^{O(k)}} to 2O(k2)2^{O(k^2)} in the parameter dependence of the Directed Multiway Cut algorithm of Chitnis et al. (SODA '12).Comment: To appear in ACM Transactions on Algorithms. A preliminary version appeared in ICALP '12. We would like to thank Marcin Pilipczuk for pointing out a missing case in the conference version which has been considered in this version. Also, we give an single exponential FPT algorithm improving on the double exponential algorithm from the conference versio

    Electric routing and concurrent flow cutting

    Get PDF
    We investigate an oblivious routing scheme, amenable to distributed computation and resilient to graph changes, based on electrical flow. Our main technical contribution is a new rounding method which we use to obtain a bound on the L1->L1 operator norm of the inverse graph Laplacian. We show how this norm reflects both latency and congestion of electric routing.Comment: 21 pages, 0 figures. To be published in Springer LNCS Book No. 5878, Proceedings of The 20th International Symposium on Algorithms and Computation (ISAAC'09

    Constrained Non-Monotone Submodular Maximization: Offline and Secretary Algorithms

    Full text link
    Constrained submodular maximization problems have long been studied, with near-optimal results known under a variety of constraints when the submodular function is monotone. The case of non-monotone submodular maximization is less understood: the first approximation algorithms even for the unconstrainted setting were given by Feige et al. (FOCS '07). More recently, Lee et al. (STOC '09, APPROX '09) show how to approximately maximize non-monotone submodular functions when the constraints are given by the intersection of p matroid constraints; their algorithm is based on local-search procedures that consider p-swaps, and hence the running time may be n^Omega(p), implying their algorithm is polynomial-time only for constantly many matroids. In this paper, we give algorithms that work for p-independence systems (which generalize constraints given by the intersection of p matroids), where the running time is poly(n,p). Our algorithm essentially reduces the non-monotone maximization problem to multiple runs of the greedy algorithm previously used in the monotone case. Our idea of using existing algorithms for monotone functions to solve the non-monotone case also works for maximizing a submodular function with respect to a knapsack constraint: we get a simple greedy-based constant-factor approximation for this problem. With these simpler algorithms, we are able to adapt our approach to constrained non-monotone submodular maximization to the (online) secretary setting, where elements arrive one at a time in random order, and the algorithm must make irrevocable decisions about whether or not to select each element as it arrives. We give constant approximations in this secretary setting when the algorithm is constrained subject to a uniform matroid or a partition matroid, and give an O(log k) approximation when it is constrained by a general matroid of rank k.Comment: In the Proceedings of WINE 201

    Beating ratio 0.5 for weighted oblivious matching problems

    Get PDF
    ESA 2016 is organized in collaboration with the European Association for Theoretical Computer Science (EATCS) and is a part of ALGO 2016We prove the first non-trivial performance ratios strictly above 0.5 for weighted versions of the oblivious matching problem. Even for the unweighted version, since Aronson, Dyer, Frieze, and Suen first proved a non-trivial ratio above 0.5 in the mid-1990s, during the next twenty years several attempts have been made to improve this ratio, until Chan, Chen, Wu and Zhao successfully achieved a significant ratio of 0.523 very recently (SODA 2014). To the best of our knowledge, our work is the first in the literature that considers the node-weighted and edge-weighted versions of the problem in arbitrary graphs (as opposed to bipartite graphs). (1) For arbitrary node weights, we prove that a weighted version of the Ranking algorithm has ratio strictly above 0.5. We have discovered a new structural property of the ranking algorithm: if a node has two unmatched neighbors at the end of algorithm, then it will still be matched even when its rank is demoted to the bottom. This property allows us to form LP constraints for both the node-weighted and the unweighted oblivious matching problems. As a result, we prove that the ratio for the node-weighted case is at least 0.501512. Interestingly via the structural property, we can also improve slightly the ratio for the unweighted case to 0.526823 (from the previous best 0.523166 in SODA 2014). (2) For a bounded number of distinct edge weights, we show that ratio strictly above 0.5 can be achieved by partitioning edges carefully according to the weights, and running the (unweighted) Ranking algorithm on each part. Our analysis is based on a new primal-dual framework known as matching coverage, in which dual feasibility is bypassed. Instead, only dual constraints corresponding to edges in an optimal matching are satisfied. Using this framework we also design and analyze an algorithm for the edge-weighted online bipartite matching problem with free disposal. We prove that for the case of bounded online degrees, the ratio is strictly above 0.5.published_or_final_versio

    Packing Returning Secretaries

    Full text link
    We study online secretary problems with returns in combinatorial packing domains with nn candidates that arrive sequentially over time in random order. The goal is to accept a feasible packing of candidates of maximum total value. In the first variant, each candidate arrives exactly twice. All 2n2n arrivals occur in random order. We propose a simple 0.5-competitive algorithm that can be combined with arbitrary approximation algorithms for the packing domain, even when the total value of candidates is a subadditive function. For bipartite matching, we obtain an algorithm with competitive ratio at least 0.5721o(1)0.5721 - o(1) for growing nn, and an algorithm with ratio at least 0.54590.5459 for all n1n \ge 1. We extend all algorithms and ratios to k2k \ge 2 arrivals per candidate. In the second variant, there is a pool of undecided candidates. In each round, a random candidate from the pool arrives. Upon arrival a candidate can be either decided (accept/reject) or postponed (returned into the pool). We mainly focus on minimizing the expected number of postponements when computing an optimal solution. An expected number of Θ(nlogn)\Theta(n \log n) is always sufficient. For matroids, we show that the expected number can be reduced to O(rlog(n/r))O(r \log (n/r)), where rn/2r \le n/2 is the minimum of the ranks of matroid and dual matroid. For bipartite matching, we show a bound of O(rlogn)O(r \log n), where rr is the size of the optimum matching. For general packing, we show a lower bound of Ω(nloglogn)\Omega(n \log \log n), even when the size of the optimum is r=Θ(logn)r = \Theta(\log n).Comment: 23 pages, 5 figure

    Experimental demonstration of near-infrared negative-index metamaterials

    Get PDF
    Metal-based negative refractive-index materials have been extensively studied in the microwave region. However, negative-index metamaterials have not been realized at near-IR or visible frequencies due to difficulties of fabrication and to the generally poor optical properties of metals at these wavelengths. In this Letter, we report the first fabrication and experimental verification of a transversely structured metal-dielectricmetal multilayer exhibiting a negative refractive index around 2 mu m. Both the amplitude and the phase of the transmission and reflection were measured experimentally, and are in good agreement with a rigorous coupled wave analysis

    Matroid and Knapsack Center Problems

    Full text link
    In the classic kk-center problem, we are given a metric graph, and the objective is to open kk nodes as centers such that the maximum distance from any vertex to its closest center is minimized. In this paper, we consider two important generalizations of kk-center, the matroid center problem and the knapsack center problem. Both problems are motivated by recent content distribution network applications. Our contributions can be summarized as follows: 1. We consider the matroid center problem in which the centers are required to form an independent set of a given matroid. We show this problem is NP-hard even on a line. We present a 3-approximation algorithm for the problem on general metrics. We also consider the outlier version of the problem where a given number of vertices can be excluded as the outliers from the solution. We present a 7-approximation for the outlier version. 2. We consider the (multi-)knapsack center problem in which the centers are required to satisfy one (or more) knapsack constraint(s). It is known that the knapsack center problem with a single knapsack constraint admits a 3-approximation. However, when there are at least two knapsack constraints, we show this problem is not approximable at all. To complement the hardness result, we present a polynomial time algorithm that gives a 3-approximate solution such that one knapsack constraint is satisfied and the others may be violated by at most a factor of 1+ϵ1+\epsilon. We also obtain a 3-approximation for the outlier version that may violate the knapsack constraint by 1+ϵ1+\epsilon.Comment: A preliminary version of this paper is accepted to IPCO 201

    Prophet Inequalities for IID Random Variables from an Unknown Distribution

    Get PDF
    A central object in optimal stopping theory is the single-choice prophet inequality for independent, identically distributed random variables: given a sequence of random variables X1, . . . , Xn drawn independently from a distribution F , the goal is to choose a stopping time τ so as to maximize α such that for all distributions F we have E[Xτ ] ≥ α · E[maxt Xt ]. What makes this problem challenging is that the decision whether τ = t may only depend on the values of the random variables X1, . . . , Xt and on the distribution F . For a long time the best known bound for the problem had been α ≥ 1 − 1/e ≈ 0.632, but quite recently a tight bound of α ≈ 0.745 was obtained. The case where F is unknown, such that the decision whether τ = t may depend only on the values of the random variables X1, . . . , Xt , is equally well motivated but has received much less attention. A straightforward guarantee for this case of α ≥ 1/e ≈ 0.368 can be derived from the solution to the secretary problem, where an arbitrary set of values arrive in random order and the goal is to maximize the probability of selecting the largest value. We show that this bound is in fact tight. We then investigate the case where the stopping time may additionally depend on a limited number of samples from F , and show that even with o(n) samples α ≤ 1/e. On the other hand, n samples allow for a significant improvement, while O(n2) samples are equivalent to knowledge of the distribution: specifically, with n samples α ≥ 1 − 1/e ≈ 0.632 and α ≤ ln(2) ≈ 0.693, and with O(n2) samples α ≥ 0.745 − ε for any ε > 0

    Revenue Guarantees in the Generalized Second Price Auction

    Get PDF
    Sponsored search auctions are the main source of revenue for search engines. In such an auction, a set of utility maximizing advertisers competes for a set of ad slots. The assignment of advertisers to slots depends on the bids they submit; these bids may be different than the true valuations of the advertisers for the slots. Variants of the celebrated VCG auction mechanism guarantee that advertisers act truthfully and, under some assumptions, lead to revenue or social welfare maximization. Still, the sponsored search industry mostly uses generalized second price (GSP) auctions; these auctions are known to be nontruthful and suboptimal in terms of social welfare and revenue. In an attempt to explain this tradition, we study a Bayesian setting wherein the valuations of advertisers are drawn independently from a common regular probability distribution. In this setting, it is well known from the work of Myerson [1981] that the optimal revenue is obtained by the VCG mechanism with a particular reserve price that depends on the probability distribution. We show that, by appropriately setting the reserve price, the revenue over any Bayes-Nash equilibrium of the game induced by the GSP auction is at most a small constant factor away from the optimal revenue, improving previous results of Lucier et al. [2012]. Our analysis is based on the Bayes-Nash equilibrium conditions and the improved results are obtained by bounding the utility of each player at equilibrium using infinitely many deviating bids and also by developing novel prophet-like inequalities
    corecore