17,427 research outputs found
Communication Complexity of Cake Cutting
We study classic cake-cutting problems, but in discrete models rather than
using infinite-precision real values, specifically, focusing on their
communication complexity. Using general discrete simulations of classical
infinite-precision protocols (Robertson-Webb and moving-knife), we roughly
partition the various fair-allocation problems into 3 classes: "easy" (constant
number of rounds of logarithmic many bits), "medium" (poly-logarithmic total
communication), and "hard". Our main technical result concerns two of the
"medium" problems (perfect allocation for 2 players and equitable allocation
for any number of players) which we prove are not in the "easy" class. Our main
open problem is to separate the "hard" from the "medium" classes.Comment: Added efficient communication protocol for the monotone crossing
proble
Cake Cutting Algorithms for Piecewise Constant and Piecewise Uniform Valuations
Cake cutting is one of the most fundamental settings in fair division and
mechanism design without money. In this paper, we consider different levels of
three fundamental goals in cake cutting: fairness, Pareto optimality, and
strategyproofness. In particular, we present robust versions of envy-freeness
and proportionality that are not only stronger than their standard
counter-parts but also have less information requirements. We then focus on
cake cutting with piecewise constant valuations and present three desirable
algorithms: CCEA (Controlled Cake Eating Algorithm), MEA (Market Equilibrium
Algorithm) and CSD (Constrained Serial Dictatorship). CCEA is polynomial-time,
robust envy-free, and non-wasteful. It relies on parametric network flows and
recent generalizations of the probabilistic serial algorithm. For the subdomain
of piecewise uniform valuations, we show that it is also group-strategyproof.
Then, we show that there exists an algorithm (MEA) that is polynomial-time,
envy-free, proportional, and Pareto optimal. MEA is based on computing a
market-based equilibrium via a convex program and relies on the results of
Reijnierse and Potters [24] and Devanur et al. [15]. Moreover, we show that MEA
and CCEA are equivalent to mechanism 1 of Chen et. al. [12] for piecewise
uniform valuations. We then present an algorithm CSD and a way to implement it
via randomization that satisfies strategyproofness in expectation, robust
proportionality, and unanimity for piecewise constant valuations. For the case
of two agents, it is robust envy-free, robust proportional, strategyproof, and
polynomial-time. Many of our results extend to more general settings in cake
cutting that allow for variable claims and initial endowments. We also show a
few impossibility results to complement our algorithms.Comment: 39 page
An Algorithmic Framework for Strategic Fair Division
We study the paradigmatic fair division problem of allocating a divisible
good among agents with heterogeneous preferences, commonly known as cake
cutting. Classical cake cutting protocols are susceptible to manipulation. Do
their strategic outcomes still guarantee fairness?
To address this question we adopt a novel algorithmic approach, by designing
a concrete computational framework for fair division---the class of Generalized
Cut and Choose (GCC) protocols}---and reasoning about the game-theoretic
properties of algorithms that operate in this model. The class of GCC protocols
includes the most important discrete cake cutting protocols, and turns out to
be compatible with the study of fair division among strategic agents. In
particular, GCC protocols are guaranteed to have approximate subgame perfect
Nash equilibria, or even exact equilibria if the protocol's tie-breaking rule
is flexible. We further observe that the (approximate) equilibria of
proportional GCC protocols---which guarantee each of the agents a
-fraction of the cake---must be (approximately) proportional. Finally, we
design a protocol in this framework with the property that its Nash equilibrium
allocations coincide with the set of (contiguous) envy-free allocations
A Pipeline for Volume Electron Microscopy of the Caenorhabditis elegans Nervous System.
The "connectome," a comprehensive wiring diagram of synaptic connectivity, is achieved through volume electron microscopy (vEM) analysis of an entire nervous system and all associated non-neuronal tissues. White et al. (1986) pioneered the fully manual reconstruction of a connectome using Caenorhabditis elegans. Recent advances in vEM allow mapping new C. elegans connectomes with increased throughput, and reduced subjectivity. Current vEM studies aim to not only fill the remaining gaps in the original connectome, but also address fundamental questions including how the connectome changes during development, the nature of individuality, sexual dimorphism, and how genetic and environmental factors regulate connectivity. Here we describe our current vEM pipeline and projected improvements for the study of the C. elegans nervous system and beyond
Parameterized Algorithmics for Computational Social Choice: Nine Research Challenges
Computational Social Choice is an interdisciplinary research area involving
Economics, Political Science, and Social Science on the one side, and
Mathematics and Computer Science (including Artificial Intelligence and
Multiagent Systems) on the other side. Typical computational problems studied
in this field include the vulnerability of voting procedures against attacks,
or preference aggregation in multi-agent systems. Parameterized Algorithmics is
a subfield of Theoretical Computer Science seeking to exploit meaningful
problem-specific parameters in order to identify tractable special cases of in
general computationally hard problems. In this paper, we propose nine of our
favorite research challenges concerning the parameterized complexity of
problems appearing in this context
- …