19 research outputs found
Algorithms as Mechanisms: The Price of Anarchy of Relax-and-Round
Many algorithms that are originally designed without explicitly considering
incentive properties are later combined with simple pricing rules and used as
mechanisms. The resulting mechanisms are often natural and simple to
understand. But how good are these algorithms as mechanisms? Truthful reporting
of valuations is typically not a dominant strategy (certainly not with a
pay-your-bid, first-price rule, but it is likely not a good strategy even with
a critical value, or second-price style rule either). Our goal is to show that
a wide class of approximation algorithms yields this way mechanisms with low
Price of Anarchy.
The seminal result of Lucier and Borodin [SODA 2010] shows that combining a
greedy algorithm that is an -approximation algorithm with a
pay-your-bid payment rule yields a mechanism whose Price of Anarchy is
. In this paper we significantly extend the class of algorithms for
which such a result is available by showing that this close connection between
approximation ratio on the one hand and Price of Anarchy on the other also
holds for the design principle of relaxation and rounding provided that the
relaxation is smooth and the rounding is oblivious.
We demonstrate the far-reaching consequences of our result by showing its
implications for sparse packing integer programs, such as multi-unit auctions
and generalized matching, for the maximum traveling salesman problem, for
combinatorial auctions, and for single source unsplittable flow problems. In
all these problems our approach leads to novel simple, near-optimal mechanisms
whose Price of Anarchy either matches or beats the performance guarantees of
known mechanisms.Comment: Extended abstract appeared in Proc. of 16th ACM Conference on
Economics and Computation (EC'15
Fast Convex Decomposition for Truthful Social Welfare Approximation
Approximating the optimal social welfare while preserving truthfulness is a
well studied problem in algorithmic mechanism design. Assuming that the social
welfare of a given mechanism design problem can be optimized by an integer
program whose integrality gap is at most , Lavi and Swamy~\cite{Lavi11}
propose a general approach to designing a randomized -approximation
mechanism which is truthful in expectation. Their method is based on
decomposing an optimal solution for the relaxed linear program into a convex
combination of integer solutions. Unfortunately, Lavi and Swamy's decomposition
technique relies heavily on the ellipsoid method, which is notorious for its
poor practical performance. To overcome this problem, we present an alternative
decomposition technique which yields an approximation
and only requires a quadratic number of calls to an integrality gap verifier
Distributed Connectivity Decomposition
We present time-efficient distributed algorithms for decomposing graphs with
large edge or vertex connectivity into multiple spanning or dominating trees,
respectively. As their primary applications, these decompositions allow us to
achieve information flow with size close to the connectivity by parallelizing
it along the trees. More specifically, our distributed decomposition algorithms
are as follows:
(I) A decomposition of each undirected graph with vertex-connectivity
into (fractionally) vertex-disjoint weighted dominating trees with total weight
, in rounds.
(II) A decomposition of each undirected graph with edge-connectivity
into (fractionally) edge-disjoint weighted spanning trees with total
weight , in
rounds.
We also show round complexity lower bounds of
and
for the above two decompositions,
using techniques of [Das Sarma et al., STOC'11]. Moreover, our
vertex-connectivity decomposition extends to centralized algorithms and
improves the time complexity of [Censor-Hillel et al., SODA'14] from
to near-optimal .
As corollaries, we also get distributed oblivious routing broadcast with
-competitive edge-congestion and -competitive
vertex-congestion. Furthermore, the vertex connectivity decomposition leads to
near-time-optimal -approximation of vertex connectivity: centralized
and distributed . The former moves
toward the 1974 conjecture of Aho, Hopcroft, and Ullman postulating an
centralized exact algorithm while the latter is the first distributed vertex
connectivity approximation
Algorithms to Approximate Column-Sparse Packing Problems
Column-sparse packing problems arise in several contexts in both
deterministic and stochastic discrete optimization. We present two unifying
ideas, (non-uniform) attenuation and multiple-chance algorithms, to obtain
improved approximation algorithms for some well-known families of such
problems. As three main examples, we attain the integrality gap, up to
lower-order terms, for known LP relaxations for k-column sparse packing integer
programs (Bansal et al., Theory of Computing, 2012) and stochastic k-set
packing (Bansal et al., Algorithmica, 2012), and go "half the remaining
distance" to optimal for a major integrality-gap conjecture of Furedi, Kahn and
Seymour on hypergraph matching (Combinatorica, 1993).Comment: Extended abstract appeared in SODA 2018. Full version in ACM
Transactions of Algorithm