2,661 research outputs found
Adding Isolated Vertices Makes some Online Algorithms Optimal
An unexpected difference between online and offline algorithms is observed.
The natural greedy algorithms are shown to be worst case online optimal for
Online Independent Set and Online Vertex Cover on graphs with 'enough' isolated
vertices, Freckle Graphs. For Online Dominating Set, the greedy algorithm is
shown to be worst case online optimal on graphs with at least one isolated
vertex. These algorithms are not online optimal in general. The online
optimality results for these greedy algorithms imply optimality according to
various worst case performance measures, such as the competitive ratio. It is
also shown that, despite this worst case optimality, there are Freckle graphs
where the greedy independent set algorithm is objectively less good than
another algorithm. It is shown that it is NP-hard to determine any of the
following for a given graph: the online independence number, the online vertex
cover number, and the online domination number.Comment: A footnote in the .tex file didn't show up in the last version. This
was fixe
On Conceptually Simple Algorithms for Variants of Online Bipartite Matching
We present a series of results regarding conceptually simple algorithms for
bipartite matching in various online and related models. We first consider a
deterministic adversarial model. The best approximation ratio possible for a
one-pass deterministic online algorithm is , which is achieved by any
greedy algorithm. D\"urr et al. recently presented a -pass algorithm called
Category-Advice that achieves approximation ratio . We extend their
algorithm to multiple passes. We prove the exact approximation ratio for the
-pass Category-Advice algorithm for all , and show that the
approximation ratio converges to the inverse of the golden ratio
as goes to infinity. The convergence is
extremely fast --- the -pass Category-Advice algorithm is already within
of the inverse of the golden ratio.
We then consider a natural greedy algorithm in the online stochastic IID
model---MinDegree. This algorithm is an online version of a well-known and
extensively studied offline algorithm MinGreedy. We show that MinDegree cannot
achieve an approximation ratio better than , which is guaranteed by any
consistent greedy algorithm in the known IID model.
Finally, following the work in Besser and Poloczek, we depart from an
adversarial or stochastic ordering and investigate a natural randomized
algorithm (MinRanking) in the priority model. Although the priority model
allows the algorithm to choose the input ordering in a general but well defined
way, this natural algorithm cannot obtain the approximation of the Ranking
algorithm in the ROM model
Temporal Ordered Clustering in Dynamic Networks: Unsupervised and Semi-supervised Learning Algorithms
In temporal ordered clustering, given a single snapshot of a dynamic network
in which nodes arrive at distinct time instants, we aim at partitioning its
nodes into ordered clusters such that for , nodes in cluster arrived
before nodes in cluster , with being a data-driven parameter
and not known upfront. Such a problem is of considerable significance in many
applications ranging from tracking the expansion of fake news to mapping the
spread of information. We first formulate our problem for a general dynamic
graph, and propose an integer programming framework that finds the optimal
clustering, represented as a strict partial order set, achieving the best
precision (i.e., fraction of successfully ordered node pairs) for a fixed
density (i.e., fraction of comparable node pairs). We then develop a sequential
importance procedure and design unsupervised and semi-supervised algorithms to
find temporal ordered clusters that efficiently approximate the optimal
solution. To illustrate the techniques, we apply our methods to the vertex
copying (duplication-divergence) model which exhibits some edge-case challenges
in inferring the clusters as compared to other network models. Finally, we
validate the performance of the proposed algorithms on synthetic and real-world
networks.Comment: 14 pages, 9 figures, and 3 tables. This version is submitted to a
journal. A shorter version of this work is published in the proceedings of
IEEE International Symposium on Information Theory (ISIT), 2020. The first
two authors contributed equall
Relaxing the Irrevocability Requirement for Online Graph Algorithms
Online graph problems are considered in models where the irrevocability
requirement is relaxed. Motivated by practical examples where, for example,
there is a cost associated with building a facility and no extra cost
associated with doing it later, we consider the Late Accept model, where a
request can be accepted at a later point, but any acceptance is irrevocable.
Similarly, we also consider a Late Reject model, where an accepted request can
later be rejected, but any rejection is irrevocable (this is sometimes called
preemption). Finally, we consider the Late Accept/Reject model, where late
accepts and rejects are both allowed, but any late reject is irrevocable. For
Independent Set, the Late Accept/Reject model is necessary to obtain a constant
competitive ratio, but for Vertex Cover the Late Accept model is sufficient and
for Minimum Spanning Forest the Late Reject model is sufficient. The Matching
problem has a competitive ratio of 2, but in the Late Accept/Reject model, its
competitive ratio is 3/2
Online Steiner Tree with Deletions
In the online Steiner tree problem, the input is a set of vertices that
appear one-by-one, and we have to maintain a Steiner tree on the current set of
vertices. The cost of the tree is the total length of edges in the tree, and we
want this cost to be close to the cost of the optimal Steiner tree at all
points in time. If we are allowed to only add edges, a tight bound of
on the competitiveness is known. Recently it was shown that if
we can add one new edge and make one edge swap upon every vertex arrival, we
can maintain a constant-competitive tree online.
But what if the set of vertices sees both additions and deletions? Again, we
would like to obtain a low-cost Steiner tree with as few edge changes as
possible. The original paper of Imase and Waxman had also considered this
model, and it gave a greedy algorithm that maintained a constant-competitive
tree online, and made at most edge changes for the first
requests. In this paper give the following two results.
Our first result is an online algorithm that maintains a Steiner tree only
under deletions: we start off with a set of vertices, and at each time one of
the vertices is removed from this set: our Steiner tree no longer has to span
this vertex. We give an algorithm that changes only a constant number of edges
upon each request, and maintains a constant-competitive tree at all times. Our
algorithm uses the primal-dual framework and a global charging argument to
carefully make these constant number of changes.
We then study the natural greedy algorithm proposed by Imase and Waxman that
maintains a constant-competitive Steiner tree in the fully-dynamic model (where
each request either adds or deletes a vertex). Our second result shows that
this algorithm makes only a constant number of changes per request in an
amortized sense.Comment: An extended abstract appears in the SODA 2014 conferenc
Seeding with Costly Network Information
We study the task of selecting nodes in a social network of size , to
seed a diffusion with maximum expected spread size, under the independent
cascade model with cascade probability . Most of the previous work on this
problem (known as influence maximization) focuses on efficient algorithms to
approximate the optimal seed set with provable guarantees, given the knowledge
of the entire network. However, in practice, obtaining full knowledge of the
network is very costly. To address this gap, we first study the achievable
guarantees using influence samples. We provide an approximation
algorithm with a tight (1-1/e){\mbox{OPT}}-\epsilon n guarantee, using
influence samples and show that this dependence on
is asymptotically optimal. We then propose a probing algorithm that queries
edges from the graph and use them to find a seed set with the
same almost tight approximation guarantee. We also provide a matching (up to
logarithmic factors) lower-bound on the required number of edges. To address
the dependence of our probing algorithm on the independent cascade probability
, we show that it is impossible to maintain the same approximation
guarantees by controlling the discrepancy between the probing and seeding
cascade probabilities. Instead, we propose to down-sample the probed edges to
match the seeding cascade probability, provided that it does not exceed that of
probing. Finally, we test our algorithms on real world data to quantify the
trade-off between the cost of obtaining more refined network information and
the benefit of the added information for guiding improved seeding strategies
- …