120 research outputs found
Directed Subset Feedback Vertex Set Is Fixed-Parameter Tractable
Given a graph and an integer , the Feedback Vertex Set (FVS) problem
asks if there is a vertex set of size at most that hits all cycles in
the graph. The fixed-parameter tractability status of FVS in directed graphs
was a long-standing open problem until Chen et al. (STOC '08) showed that it is
FPT by giving a time algorithm. In the subset versions of
this problems, we are given an additional subset of vertices (resp., edges)
and we want to hit all cycles passing through a vertex of (resp. an edge of
). Recently, the Subset Feedback Vertex Set in undirected graphs was shown
to be FPT by Cygan et al. (ICALP '11) and independently by Kakimura et al.
(SODA '12). We generalize the result of Chen et al. (STOC '08) by showing that
Subset Feedback Vertex Set in directed graphs can be solved in time
. By our result, we complete the picture for feedback
vertex set problems and their subset versions in undirected and directed
graphs. Besides proving the fixed-parameter tractability of Directed Subset
Feedback Vertex Set, we reformulate the random sampling of important separators
technique in an abstract way that can be used for a general family of
transversal problems. Moreover, we modify the probability distribution used in
the technique to achieve better running time; in particular, this gives an
improvement from to in the parameter dependence of
the Directed Multiway Cut algorithm of Chitnis et al. (SODA '12).Comment: To appear in ACM Transactions on Algorithms. A preliminary version
appeared in ICALP '12. We would like to thank Marcin Pilipczuk for pointing
out a missing case in the conference version which has been considered in
this version. Also, we give an single exponential FPT algorithm improving on
the double exponential algorithm from the conference versio
Electric routing and concurrent flow cutting
We investigate an oblivious routing scheme, amenable to distributed
computation and resilient to graph changes, based on electrical flow. Our main
technical contribution is a new rounding method which we use to obtain a bound
on the L1->L1 operator norm of the inverse graph Laplacian. We show how this
norm reflects both latency and congestion of electric routing.Comment: 21 pages, 0 figures. To be published in Springer LNCS Book No. 5878,
Proceedings of The 20th International Symposium on Algorithms and Computation
(ISAAC'09
Constrained Non-Monotone Submodular Maximization: Offline and Secretary Algorithms
Constrained submodular maximization problems have long been studied, with
near-optimal results known under a variety of constraints when the submodular
function is monotone. The case of non-monotone submodular maximization is less
understood: the first approximation algorithms even for the unconstrainted
setting were given by Feige et al. (FOCS '07). More recently, Lee et al. (STOC
'09, APPROX '09) show how to approximately maximize non-monotone submodular
functions when the constraints are given by the intersection of p matroid
constraints; their algorithm is based on local-search procedures that consider
p-swaps, and hence the running time may be n^Omega(p), implying their algorithm
is polynomial-time only for constantly many matroids. In this paper, we give
algorithms that work for p-independence systems (which generalize constraints
given by the intersection of p matroids), where the running time is poly(n,p).
Our algorithm essentially reduces the non-monotone maximization problem to
multiple runs of the greedy algorithm previously used in the monotone case.
Our idea of using existing algorithms for monotone functions to solve the
non-monotone case also works for maximizing a submodular function with respect
to a knapsack constraint: we get a simple greedy-based constant-factor
approximation for this problem.
With these simpler algorithms, we are able to adapt our approach to
constrained non-monotone submodular maximization to the (online) secretary
setting, where elements arrive one at a time in random order, and the algorithm
must make irrevocable decisions about whether or not to select each element as
it arrives. We give constant approximations in this secretary setting when the
algorithm is constrained subject to a uniform matroid or a partition matroid,
and give an O(log k) approximation when it is constrained by a general matroid
of rank k.Comment: In the Proceedings of WINE 201
Beating ratio 0.5 for weighted oblivious matching problems
ESA 2016 is organized in collaboration with the European Association for Theoretical Computer Science (EATCS) and is a part of ALGO 2016We prove the first non-trivial performance ratios strictly above 0.5 for weighted versions of the oblivious matching problem. Even for the unweighted version, since Aronson, Dyer, Frieze, and Suen first proved a non-trivial ratio above 0.5 in the mid-1990s, during the next twenty years several attempts have been made to improve this ratio, until Chan, Chen, Wu and Zhao successfully achieved a significant ratio of 0.523 very recently (SODA 2014). To the best of our knowledge, our work is the first in the literature that considers the node-weighted and edge-weighted versions of the problem in arbitrary graphs (as opposed to bipartite graphs). (1) For arbitrary node weights, we prove that a weighted version of the Ranking algorithm has ratio strictly above 0.5. We have discovered a new structural property of the ranking algorithm: if a node has two unmatched neighbors at the end of algorithm, then it will still be matched even when its rank is demoted to the bottom. This property allows us to form LP constraints for both the node-weighted and the unweighted oblivious matching problems. As a result, we prove that the ratio for the node-weighted case is at least 0.501512. Interestingly via the structural property, we can also improve slightly the ratio for the unweighted case to 0.526823 (from the previous best 0.523166 in SODA 2014). (2) For a bounded number of distinct edge weights, we show that ratio strictly above 0.5 can be achieved by partitioning edges carefully according to the weights, and running the (unweighted) Ranking algorithm on each part. Our analysis is based on a new primal-dual framework known as matching coverage, in which dual feasibility is bypassed. Instead, only dual constraints corresponding to edges in an optimal matching are satisfied. Using this framework we also design and analyze an algorithm for the edge-weighted online bipartite matching problem with free disposal. We prove that for the case of bounded online degrees, the ratio is strictly above 0.5.published_or_final_versio
Packing Returning Secretaries
We study online secretary problems with returns in combinatorial packing
domains with candidates that arrive sequentially over time in random order.
The goal is to accept a feasible packing of candidates of maximum total value.
In the first variant, each candidate arrives exactly twice. All arrivals
occur in random order. We propose a simple 0.5-competitive algorithm that can
be combined with arbitrary approximation algorithms for the packing domain,
even when the total value of candidates is a subadditive function. For
bipartite matching, we obtain an algorithm with competitive ratio at least
for growing , and an algorithm with ratio at least
for all . We extend all algorithms and ratios to arrivals
per candidate.
In the second variant, there is a pool of undecided candidates. In each
round, a random candidate from the pool arrives. Upon arrival a candidate can
be either decided (accept/reject) or postponed (returned into the pool). We
mainly focus on minimizing the expected number of postponements when computing
an optimal solution. An expected number of is always
sufficient. For matroids, we show that the expected number can be reduced to
, where is the minimum of the ranks of matroid and
dual matroid. For bipartite matching, we show a bound of , where
is the size of the optimum matching. For general packing, we show a lower
bound of , even when the size of the optimum is .Comment: 23 pages, 5 figure
Experimental demonstration of near-infrared negative-index metamaterials
Metal-based negative refractive-index materials have been extensively studied in the microwave region. However, negative-index metamaterials have not been realized at near-IR or visible frequencies due to difficulties of fabrication and to the generally poor optical properties of metals at these wavelengths. In this Letter, we report the first fabrication and experimental verification of a transversely structured metal-dielectricmetal multilayer exhibiting a negative refractive index around 2 mu m. Both the amplitude and the phase of the transmission and reflection were measured experimentally, and are in good agreement with a rigorous coupled wave analysis
Matroid and Knapsack Center Problems
In the classic -center problem, we are given a metric graph, and the
objective is to open nodes as centers such that the maximum distance from
any vertex to its closest center is minimized. In this paper, we consider two
important generalizations of -center, the matroid center problem and the
knapsack center problem. Both problems are motivated by recent content
distribution network applications. Our contributions can be summarized as
follows:
1. We consider the matroid center problem in which the centers are required
to form an independent set of a given matroid. We show this problem is NP-hard
even on a line. We present a 3-approximation algorithm for the problem on
general metrics. We also consider the outlier version of the problem where a
given number of vertices can be excluded as the outliers from the solution. We
present a 7-approximation for the outlier version.
2. We consider the (multi-)knapsack center problem in which the centers are
required to satisfy one (or more) knapsack constraint(s). It is known that the
knapsack center problem with a single knapsack constraint admits a
3-approximation. However, when there are at least two knapsack constraints, we
show this problem is not approximable at all. To complement the hardness
result, we present a polynomial time algorithm that gives a 3-approximate
solution such that one knapsack constraint is satisfied and the others may be
violated by at most a factor of . We also obtain a 3-approximation
for the outlier version that may violate the knapsack constraint by
.Comment: A preliminary version of this paper is accepted to IPCO 201
Prophet Inequalities for IID Random Variables from an Unknown Distribution
A central object in optimal stopping theory is the single-choice prophet inequality for independent, identically distributed random variables: given a sequence of random variables X1, . . . , Xn drawn independently from a distribution F , the goal is to choose a stopping time τ so as to maximize α such that for all distributions F we have E[Xτ ] ≥ α · E[maxt Xt ]. What makes this problem challenging is that the decision whether τ = t may only depend on the values of the random variables X1, . . . , Xt and on the distribution F . For a long time the best known bound for the problem had been α ≥ 1 − 1/e ≈ 0.632, but quite recently a tight bound of α ≈ 0.745 was obtained. The case where F is unknown, such that the decision whether τ = t may depend only on the values of the random variables X1, . . . , Xt , is equally well motivated but has received much less attention. A straightforward guarantee for this case of α ≥ 1/e ≈ 0.368 can be derived from the solution to the secretary problem, where an arbitrary set of values arrive in random order and the goal is to maximize the probability of selecting the largest value. We show that this bound is in fact tight. We then investigate the case where the stopping time may additionally depend on a limited number of samples from F , and show that even with o(n) samples α ≤ 1/e. On the other hand, n samples allow for a significant improvement, while O(n2) samples are equivalent to knowledge of the distribution: specifically, with n samples α ≥ 1 − 1/e ≈ 0.632 and α ≤ ln(2) ≈ 0.693, and with O(n2) samples α ≥ 0.745 − ε for any ε > 0
Revenue Guarantees in the Generalized Second Price Auction
Sponsored search auctions are the main source of revenue for search engines. In such an auction, a set of utility maximizing advertisers competes for a set of ad slots. The assignment of advertisers to slots depends on the bids they submit; these bids may be different than the true valuations of the advertisers for the slots. Variants of the celebrated VCG auction mechanism guarantee that advertisers act truthfully and, under some assumptions, lead to revenue or social welfare maximization. Still, the sponsored search industry mostly uses generalized second price (GSP) auctions; these auctions are known to be nontruthful and suboptimal in terms of social welfare and revenue. In an attempt to explain this tradition, we study a Bayesian setting wherein the valuations of advertisers are drawn independently from a common regular probability distribution. In this setting, it is well known from the work of Myerson [1981] that the optimal revenue is obtained by the VCG mechanism with a particular reserve price that depends on the probability distribution. We show that, by appropriately setting the reserve price, the revenue over any Bayes-Nash equilibrium of the game induced by the GSP auction is at most a small constant factor away from the optimal revenue, improving previous results of Lucier et al. [2012]. Our analysis is based on the Bayes-Nash equilibrium conditions and the improved results are obtained by bounding the utility of each player at equilibrium using infinitely many deviating bids and also by developing novel prophet-like inequalities
- …