301 research outputs found
Computing Equilibria in Markets with Budget-Additive Utilities
We present the first analysis of Fisher markets with buyers that have
budget-additive utility functions. Budget-additive utilities are elementary
concave functions with numerous applications in online adword markets and
revenue optimization problems. They extend the standard case of linear
utilities and have been studied in a variety of other market models. In
contrast to the frequently studied CES utilities, they have a global satiation
point which can imply multiple market equilibria with quite different
characteristics. Our main result is an efficient combinatorial algorithm to
compute a market equilibrium with a Pareto-optimal allocation of goods. It
relies on a new descending-price approach and, as a special case, also implies
a novel combinatorial algorithm for computing a market equilibrium in linear
Fisher markets. We complement these positive results with a number of hardness
results for related computational questions. We prove that it is NP-hard to
compute a market equilibrium that maximizes social welfare, and it is PPAD-hard
to find any market equilibrium with utility functions with separate satiation
points for each buyer and each good.Comment: 21 page
Average-case Approximation Ratio of Scheduling without Payments
Apart from the principles and methodologies inherited from Economics and Game
Theory, the studies in Algorithmic Mechanism Design typically employ the
worst-case analysis and approximation schemes of Theoretical Computer Science.
For instance, the approximation ratio, which is the canonical measure of
evaluating how well an incentive-compatible mechanism approximately optimizes
the objective, is defined in the worst-case sense. It compares the performance
of the optimal mechanism against the performance of a truthful mechanism, for
all possible inputs.
In this paper, we take the average-case analysis approach, and tackle one of
the primary motivating problems in Algorithmic Mechanism Design -- the
scheduling problem [Nisan and Ronen 1999]. One version of this problem which
includes a verification component is studied by [Koutsoupias 2014]. It was
shown that the problem has a tight approximation ratio bound of (n+1)/2 for the
single-task setting, where n is the number of machines. We show, however, when
the costs of the machines to executing the task follow any independent and
identical distribution, the average-case approximation ratio of the mechanism
given in [Koutsoupias 2014] is upper bounded by a constant. This positive
result asymptotically separates the average-case ratio from the worst-case
ratio, and indicates that the optimal mechanism for the problem actually works
well on average, although in the worst-case the expected cost of the mechanism
is Theta(n) times that of the optimal cost
Approximating Generalized Network Design under (Dis)economies of Scale with Applications to Energy Efficiency
In a generalized network design (GND) problem, a set of resources are
assigned to multiple communication requests. Each request contributes its
weight to the resources it uses and the total load on a resource is then
translated to the cost it incurs via a resource specific cost function. For
example, a request may be to establish a virtual circuit, thus contributing to
the load on each edge in the circuit. Motivated by energy efficiency
applications, recently, there is a growing interest in GND using cost functions
that exhibit (dis)economies of scale ((D)oS), namely, cost functions that
appear subadditive for small loads and superadditive for larger loads.
The current paper advances the existing literature on approximation
algorithms for GND problems with (D)oS cost functions in various aspects: (1)
we present a generic approximation framework that yields approximation results
for a much wider family of requests in both directed and undirected graphs; (2)
our framework allows for unrelated weights, thus providing the first
non-trivial approximation for the problem of scheduling unrelated parallel
machines with (D)oS cost functions; (3) our framework is fully combinatorial
and runs in strongly polynomial time; (4) the family of (D)oS cost functions
considered in the current paper is more general than the one considered in the
existing literature, providing a more accurate abstraction for practical energy
conservation scenarios; and (5) we obtain the first approximation ratio for GND
with (D)oS cost functions that depends only on the parameters of the resources'
technology and does not grow with the number of resources, the number of
requests, or their weights. The design of our framework relies heavily on
Roughgarden's smoothness toolbox (JACM 2015), thus demonstrating the possible
usefulness of this toolbox in the area of approximation algorithms.Comment: 39 pages, 1 figure. An extended abstract of this paper is to appear
in the 50th Annual ACM Symposium on the Theory of Computing (STOC 2018
Linear Coupling: An Ultimate Unification of Gradient and Mirror Descent
First-order methods play a central role in large-scale machine learning. Even
though many variations exist, each suited to a particular problem, almost all
such methods fundamentally rely on two types of algorithmic steps: gradient
descent, which yields primal progress, and mirror descent, which yields dual
progress.
We observe that the performances of gradient and mirror descent are
complementary, so that faster algorithms can be designed by LINEARLY COUPLING
the two. We show how to reconstruct Nesterov's accelerated gradient methods
using linear coupling, which gives a cleaner interpretation than Nesterov's
original proofs. We also discuss the power of linear coupling by extending it
to many other settings that Nesterov's methods cannot apply to.Comment: A new section added; polished writin
Complexity Theory
Computational Complexity Theory is the mathematical study of the intrinsic power and limitations of computational resources like time, space, or randomness. The current workshop focused on recent developments in various sub-areas including arithmetic complexity, Boolean complexity, communication complexity, cryptography, probabilistic proof systems, pseudorandomness, and quantum computation. Many of the developments are related to diverse mathematical ïŹelds such as algebraic geometry, combinatorial number theory, probability theory, representation theory, and the theory of error-correcting codes
- âŠ