70,173 research outputs found
LP-Based Algorithms for Capacitated Facility Location
Linear programming has played a key role in the study of algorithms for
combinatorial optimization problems. In the field of approximation algorithms,
this is well illustrated by the uncapacitated facility location problem. A
variety of algorithmic methodologies, such as LP-rounding and primal-dual
method, have been applied to and evolved from algorithms for this problem.
Unfortunately, this collection of powerful algorithmic techniques had not yet
been applicable to the more general capacitated facility location problem. In
fact, all of the known algorithms with good performance guarantees were based
on a single technique, local search, and no linear programming relaxation was
known to efficiently approximate the problem.
In this paper, we present a linear programming relaxation with constant
integrality gap for capacitated facility location. We demonstrate that the
fundamental theories of multi-commodity flows and matchings provide key
insights that lead to the strong relaxation. Our algorithmic proof of
integrality gap is obtained by finally accessing the rich toolbox of LP-based
methodologies: we present a constant factor approximation algorithm based on
LP-rounding.Comment: 25 pages, 6 figures; minor revision
Constrained Non-Monotone Submodular Maximization: Offline and Secretary Algorithms
Constrained submodular maximization problems have long been studied, with
near-optimal results known under a variety of constraints when the submodular
function is monotone. The case of non-monotone submodular maximization is less
understood: the first approximation algorithms even for the unconstrainted
setting were given by Feige et al. (FOCS '07). More recently, Lee et al. (STOC
'09, APPROX '09) show how to approximately maximize non-monotone submodular
functions when the constraints are given by the intersection of p matroid
constraints; their algorithm is based on local-search procedures that consider
p-swaps, and hence the running time may be n^Omega(p), implying their algorithm
is polynomial-time only for constantly many matroids. In this paper, we give
algorithms that work for p-independence systems (which generalize constraints
given by the intersection of p matroids), where the running time is poly(n,p).
Our algorithm essentially reduces the non-monotone maximization problem to
multiple runs of the greedy algorithm previously used in the monotone case.
Our idea of using existing algorithms for monotone functions to solve the
non-monotone case also works for maximizing a submodular function with respect
to a knapsack constraint: we get a simple greedy-based constant-factor
approximation for this problem.
With these simpler algorithms, we are able to adapt our approach to
constrained non-monotone submodular maximization to the (online) secretary
setting, where elements arrive one at a time in random order, and the algorithm
must make irrevocable decisions about whether or not to select each element as
it arrives. We give constant approximations in this secretary setting when the
algorithm is constrained subject to a uniform matroid or a partition matroid,
and give an O(log k) approximation when it is constrained by a general matroid
of rank k.Comment: In the Proceedings of WINE 201
On Quasi-Newton Forward--Backward Splitting: Proximal Calculus and Convergence
We introduce a framework for quasi-Newton forward--backward splitting
algorithms (proximal quasi-Newton methods) with a metric induced by diagonal
rank- symmetric positive definite matrices. This special type of
metric allows for a highly efficient evaluation of the proximal mapping. The
key to this efficiency is a general proximal calculus in the new metric. By
using duality, formulas are derived that relate the proximal mapping in a
rank- modified metric to the original metric. We also describe efficient
implementations of the proximity calculation for a large class of functions;
the implementations exploit the piece-wise linear nature of the dual problem.
Then, we apply these results to acceleration of composite convex minimization
problems, which leads to elegant quasi-Newton methods for which we prove
convergence. The algorithm is tested on several numerical examples and compared
to a comprehensive list of alternatives in the literature. Our quasi-Newton
splitting algorithm with the prescribed metric compares favorably against
state-of-the-art. The algorithm has extensive applications including signal
processing, sparse recovery, machine learning and classification to name a few.Comment: arXiv admin note: text overlap with arXiv:1206.115
Coresets Meet EDCS: Algorithms for Matching and Vertex Cover on Massive Graphs
As massive graphs become more prevalent, there is a rapidly growing need for
scalable algorithms that solve classical graph problems, such as maximum
matching and minimum vertex cover, on large datasets. For massive inputs,
several different computational models have been introduced, including the
streaming model, the distributed communication model, and the massively
parallel computation (MPC) model that is a common abstraction of
MapReduce-style computation. In each model, algorithms are analyzed in terms of
resources such as space used or rounds of communication needed, in addition to
the more traditional approximation ratio.
In this paper, we give a single unified approach that yields better
approximation algorithms for matching and vertex cover in all these models. The
highlights include:
* The first one pass, significantly-better-than-2-approximation for matching
in random arrival streams that uses subquadratic space, namely a
-approximation streaming algorithm that uses space
for constant .
* The first 2-round, better-than-2-approximation for matching in the MPC
model that uses subquadratic space per machine, namely a
-approximation algorithm with memory per
machine for constant .
By building on our unified approach, we further develop parallel algorithms
in the MPC model that give a -approximation to matching and an
-approximation to vertex cover in only MPC rounds and
memory per machine. These results settle multiple open
questions posed in the recent paper of Czumaj~et.al. [STOC 2018]
Knowledge Refinement via Rule Selection
In several different applications, including data transformation and entity
resolution, rules are used to capture aspects of knowledge about the
application at hand. Often, a large set of such rules is generated
automatically or semi-automatically, and the challenge is to refine the
encapsulated knowledge by selecting a subset of rules based on the expected
operational behavior of the rules on available data. In this paper, we carry
out a systematic complexity-theoretic investigation of the following rule
selection problem: given a set of rules specified by Horn formulas, and a pair
of an input database and an output database, find a subset of the rules that
minimizes the total error, that is, the number of false positive and false
negative errors arising from the selected rules. We first establish
computational hardness results for the decision problems underlying this
minimization problem, as well as upper and lower bounds for its
approximability. We then investigate a bi-objective optimization version of the
rule selection problem in which both the total error and the size of the
selected rules are taken into account. We show that testing for membership in
the Pareto front of this bi-objective optimization problem is DP-complete.
Finally, we show that a similar DP-completeness result holds for a bi-level
optimization version of the rule selection problem, where one minimizes first
the total error and then the size
- …