1,121 research outputs found
A lazy approach to on-line bipartite matching
We present a new approach, called a lazy matching, to the problem of on-line
matching on bipartite graphs. Imagine that one side of a graph is given and the
vertices of the other side are arriving on-line. Originally, incoming vertex is
either irrevocably matched to an another element or stays forever unmatched. A
lazy algorithm is allowed to match a new vertex to a group of elements
(possibly empty) and afterwords, forced against next vertices, may give up
parts of the group. The restriction is that all the time each element is in at
most one group. We present an optimal lazy algorithm (deterministic) and prove
that its competitive ratio equals . The lazy approach allows us to break the barrier of , which is the
best competitive ratio that can be guaranteed by any deterministic algorithm in
the classical on-line matching
Prophet Inequalities with Limited Information
In the classical prophet inequality, a gambler observes a sequence of
stochastic rewards and must decide, for each reward ,
whether to keep it and stop the game or to forfeit the reward forever and
reveal the next value . The gambler's goal is to obtain a constant
fraction of the expected reward that the optimal offline algorithm would get.
Recently, prophet inequalities have been generalized to settings where the
gambler can choose items, and, more generally, where he can choose any
independent set in a matroid. However, all the existing algorithms require the
gambler to know the distribution from which the rewards are
drawn.
The assumption that the gambler knows the distribution from which
are drawn is very strong. Instead, we work with the much simpler
assumption that the gambler only knows a few samples from this distribution. We
construct the first single-sample prophet inequalities for many settings of
interest, whose guarantees all match the best possible asymptotically,
\emph{even with full knowledge of the distribution}. Specifically, we provide a
novel single-sample algorithm when the gambler can choose any elements
whose analysis is based on random walks with limited correlation. In addition,
we provide a black-box method for converting specific types of solutions to the
related \emph{secretary problem} to single-sample prophet inequalities, and
apply it to several existing algorithms. Finally, we provide a constant-sample
prophet inequality for constant-degree bipartite matchings.
We apply these results to design the first posted-price and multi-dimensional
auction mechanisms with limited information in settings with asymmetric
bidders
Geometry Helps to Compare Persistence Diagrams
Exploiting geometric structure to improve the asymptotic complexity of
discrete assignment problems is a well-studied subject. In contrast, the
practical advantages of using geometry for such problems have not been
explored. We implement geometric variants of the Hopcroft--Karp algorithm for
bottleneck matching (based on previous work by Efrat el al.) and of the auction
algorithm by Bertsekas for Wasserstein distance computation. Both
implementations use k-d trees to replace a linear scan with a geometric
proximity query. Our interest in this problem stems from the desire to compute
distances between persistence diagrams, a problem that comes up frequently in
topological data analysis. We show that our geometric matching algorithms lead
to a substantial performance gain, both in running time and in memory
consumption, over their purely combinatorial counterparts. Moreover, our
implementation significantly outperforms the only other implementation
available for comparing persistence diagrams.Comment: 20 pages, 10 figures; extended version of paper published in ALENEX
201
The mixing time of the switch Markov chains: a unified approach
Since 1997 a considerable effort has been spent to study the mixing time of
switch Markov chains on the realizations of graphic degree sequences of simple
graphs. Several results were proved on rapidly mixing Markov chains on
unconstrained, bipartite, and directed sequences, using different mechanisms.
The aim of this paper is to unify these approaches. We will illustrate the
strength of the unified method by showing that on any -stable family of
unconstrained/bipartite/directed degree sequences the switch Markov chain is
rapidly mixing. This is a common generalization of every known result that
shows the rapid mixing nature of the switch Markov chain on a region of degree
sequences. Two applications of this general result will be presented. One is an
almost uniform sampler for power-law degree sequences with exponent
. The other one shows that the switch Markov chain on the
degree sequence of an Erd\H{o}s-R\'enyi random graph is asymptotically
almost surely rapidly mixing if is bounded away from 0 and 1 by at least
.Comment: Clarification
Towards Resistance Sparsifiers
We study resistance sparsification of graphs, in which the goal is to find a
sparse subgraph (with reweighted edges) that approximately preserves the
effective resistances between every pair of nodes. We show that every dense
regular expander admits a -resistance sparsifier of size , and conjecture this bound holds for all graphs on nodes. In
comparison, spectral sparsification is a strictly stronger notion and requires
edges even on the complete graph.
Our approach leads to the following structural question on graphs: Does every
dense regular expander contain a sparse regular expander as a subgraph? Our
main technical contribution, which may of independent interest, is a positive
answer to this question in a certain setting of parameters. Combining this with
a recent result of von Luxburg, Radl, and Hein~(JMLR, 2014) leads to the
aforementioned resistance sparsifiers
Basis Token Consistency: A Practical Mechanism for Strong Web Cache Consistency
With web caching and cache-related services like CDNs and edge services playing an increasingly significant role in the modern internet, the problem of the weak consistency and coherence provisions in current web protocols is becoming increasingly significant and drawing the attention of the standards community [LCD01]. Toward this end, we present definitions of consistency and coherence for web-like environments, that is, distributed client-server information systems where the semantics of interactions with resource are more general than the read/write operations found in memory hierarchies and distributed file systems. We then present a brief review of proposed mechanisms which strengthen the consistency of caches in the web, focusing upon their conceptual contributions and their weaknesses in real-world practice. These insights motivate a new mechanism, which we call "Basis Token Consistency" or BTC; when implemented at the server, this mechanism allows any client (independent of the presence and conformity of any intermediaries) to maintain a self-consistent view of the server's state. This is accomplished by annotating responses with additional per-resource application information which allows client caches to recognize the obsolescence of currently cached entities and identify responses from other caches which are already stale in light of what has already been seen. The mechanism requires no deviation from the existing client-server communication model, and does not require servers to maintain any additional per-client state. We discuss how our mechanism could be integrated into a fragment-assembling Content Management System (CMS), and present a simulation-driven performance comparison between the BTC algorithm and the use of the Time-To-Live (TTL) heuristic.National Science Foundation (ANI-9986397, ANI-0095988
- …