92 research outputs found
Lift-and-Round to Improve Weighted Completion Time on Unrelated Machines
We consider the problem of scheduling jobs on unrelated machines so as to
minimize the sum of weighted completion times. Our main result is a
-approximation algorithm for some fixed , improving upon the
long-standing bound of 3/2 (independently due to Skutella, Journal of the ACM,
2001, and Sethuraman & Squillante, SODA, 1999). To do this, we first introduce
a new lift-and-project based SDP relaxation for the problem. This is necessary
as the previous convex programming relaxations have an integrality gap of
. Second, we give a new general bipartite-rounding procedure that produces
an assignment with certain strong negative correlation properties.Comment: 21 pages, 4 figure
A Hierarchical Multi User Data Share Environment To Protect Data Privacy And Defend Unauthorized Access
A hierarchical access control technique M-HABE and an altered three-layer structure is proposed. Contrasting from the current standards, for example, the HABE algorithm and the first three-layer structure, the novel plan for the most part concentrates on the information handling, storing and accessing, which is intended to guarantee the application clients with lawful access experts to get relating detecting information and to limit illegal clients and unapproved legitimate clients access the information, the proposed promising worldview makes it to a great degree appropriate for the versatile distributed computing based worldview. What ought to be underlined is that the most critical feature of all in thin work can be depicted as that the adjusted three-layer structure is intended for settling the security issues
Improved approximation algorithms for inventory problems
We give new approximation algorithms for the submodular joint replenishment problem and the inventory routing problem, using an iterative rounding approach. In both problems, we are given a set of N items and a discrete time horizon of T days in which given demands for the items must be satisfied. Ordering a set of items incurs a cost according to a set function, with properties depending on the problem under consideration. Demand for an item at time t can be satisfied by an order on any day prior to t, but a holding cost is charged for storing the items during the intermediate period; the goal is to minimize the sum of the ordering and holding cost. Our approximation factor for both problems is [Formula Presented]; this improves exponentially on the previous best results
Stochastic Budget Optimization in Internet Advertising
Internet advertising is a sophisticated game in which the many advertisers
"play" to optimize their return on investment. There are many "targets" for the
advertisements, and each "target" has a collection of games with a potentially
different set of players involved. In this paper, we study the problem of how
advertisers allocate their budget across these "targets". In particular, we
focus on formulating their best response strategy as an optimization problem.
Advertisers have a set of keywords ("targets") and some stochastic information
about the future, namely a probability distribution over scenarios of cost vs
click combinations. This summarizes the potential states of the world assuming
that the strategies of other players are fixed. Then, the best response can be
abstracted as stochastic budget optimization problems to figure out how to
spread a given budget across these keywords to maximize the expected number of
clicks.
We present the first known non-trivial poly-logarithmic approximation for
these problems as well as the first known hardness results of getting better
than logarithmic approximation ratios in the various parameters involved. We
also identify several special cases of these problems of practical interest,
such as with fixed number of scenarios or with polynomial-sized parameters
related to cost, which are solvable either in polynomial time or with improved
approximation ratios. Stochastic budget optimization with scenarios has
sophisticated technical structure. Our approximation and hardness results come
from relating these problems to a special type of (0/1, bipartite) quadratic
programs inherent in them. Our research answers some open problems raised by
the authors in (Stochastic Models for Budget Optimization in Search-Based
Advertising, Algorithmica, 58 (4), 1022-1044, 2010).Comment: FINAL versio
Approximation Algorithms for the Directed k-Tour and k-Stroll Problems
We consider two natural generalizations of the Asymmetric Traveling Salesman problem: the k-Stroll and the k-Tour problems. The input to the k-Stroll problem is a directed n-vertex graph with nonnegative edge lengths, an integer k, and two special vertices s and t. The goal is to find a minimum-length s-t walk, containing at least k distinct vertices. The k-Tour problem can be viewed as a special case of k-Stroll, where s = t. That is, the walk is required to be a tour, containing some pre-specified vertex s. When k = n, the k-Stroll problem becomes equivalent to Asymmetric Traveling Salesman Path, and k-Tour to Asymmetric Traveling Salesman. Our main result is a polylogarithmic approximation algorithm for the k-Stroll problem. Prior to our work, only bicriteria (O(log 2 k), 3)-approximation algorithms have been known, producing walks whose length is bounded by 3OPT, while the number of vertices visited is Ω(k / log 2 k). We also show a simple O(log 2 n / log log n)-approximation algorithm for the k-Tour problem. The best previously known approximation algorithms achieved min(O(log 3 k), O(log 2 n · log k / log log n))-approximation in polynomial time, and O(log 2 k)-approximation in quasipolynomial time.
Incremental Medians via Online Bidding
In the k-median problem we are given sets of facilities and customers, and
distances between them. For a given set F of facilities, the cost of serving a
customer u is the minimum distance between u and a facility in F. The goal is
to find a set F of k facilities that minimizes the sum, over all customers, of
their service costs.
Following Mettu and Plaxton, we study the incremental medians problem, where
k is not known in advance, and the algorithm produces a nested sequence of
facility sets where the kth set has size k. The algorithm is c-cost-competitive
if the cost of each set is at most c times the cost of the optimum set of size
k. We give improved incremental algorithms for the metric version: an
8-cost-competitive deterministic algorithm, a 2e ~ 5.44-cost-competitive
randomized algorithm, a (24+epsilon)-cost-competitive, poly-time deterministic
algorithm, and a (6e+epsilon ~ .31)-cost-competitive, poly-time randomized
algorithm.
The algorithm is s-size-competitive if the cost of the kth set is at most the
minimum cost of any set of size k, and has size at most s k. The optimal
size-competitive ratios for this problem are 4 (deterministic) and e
(randomized). We present the first poly-time O(log m)-size-approximation
algorithm for the offline problem and first poly-time O(log m)-size-competitive
algorithm for the incremental problem.
Our proofs reduce incremental medians to the following online bidding
problem: faced with an unknown threshold T, an algorithm submits "bids" until
it submits a bid that is at least the threshold. It pays the sum of all its
bids. We prove that folklore algorithms for online bidding are optimally
competitive.Comment: conference version appeared in LATIN 2006 as "Oblivious Medians via
Online Bidding
Complexity of Discrete Energy Minimization Problems
Discrete energy minimization is widely-used in computer vision and machine
learning for problems such as MAP inference in graphical models. The problem,
in general, is notoriously intractable, and finding the global optimal solution
is known to be NP-hard. However, is it possible to approximate this problem
with a reasonable ratio bound on the solution quality in polynomial time? We
show in this paper that the answer is no. Specifically, we show that general
energy minimization, even in the 2-label pairwise case, and planar energy
minimization with three or more labels are exp-APX-complete. This finding rules
out the existence of any approximation algorithm with a sub-exponential
approximation ratio in the input size for these two problems, including
constant factor approximations. Moreover, we collect and review the
computational complexity of several subclass problems and arrange them on a
complexity scale consisting of three major complexity classes -- PO, APX, and
exp-APX, corresponding to problems that are solvable, approximable, and
inapproximable in polynomial time. Problems in the first two complexity classes
can serve as alternative tractable formulations to the inapproximable ones.
This paper can help vision researchers to select an appropriate model for an
application or guide them in designing new algorithms.Comment: ECCV'16 accepte
GRMA: Generalized Range Move Algorithms for the efficient optimization of MRFs
Markov Random Fields (MRF) have become an
important tool for many vision applications, and the optimization
of MRFs is a problem of fundamental importance.
Recently, Veksler and Kumar et al. proposed the range move
algorithms, which are some of the most successful optimizers.
Instead of considering only two labels as in previous
move-making algorithms, they explore a large search space
over a range of labels in each iteration, and significantly
outperform previous move-making algorithms. However, two
problems have greatly limited the applicability of range
move algorithms: 1) They are limited in the energy functions
they can handle (i.e., only truncated convex functions); 2)
They tend to be very slow compared to other move-making
algorithms (e.g., �-expansion and ��-swap). In this paper,
we propose two generalized range move algorithms (GRMA)
for the efficient optimization of MRFs. To address the
first problem, we extend the GRMAs to more general energy
functions by restricting the chosen labels in each move so
that the energy function is submodular on the chosen subset.
Furthermore, we provide a feasible sufficient condition for
choosing these subsets of labels. To address the second
problem, we dynamically obtain the iterative moves by solving
set cover problems. This greatly reduces the number of
moves during the optimization.We also propose a fast graph
construction method for the GRMAs. Experiments show
that the GRMAs offer a great speedup over previous range
move algorithms, while yielding competitive solutions
Maximum Flows on Disjoint Paths
Abstract. We consider the question: What is the maximum flow achievable in a network if the flow must be decomposable into a collection of edgedisjoint paths? Equivalently, we wish to find a maximum weighted packing of disjoint paths, where the weight of a path is the minimum capacity of an edge on the path. Our main result is an Ω(log n) lower bound on the approximability of the problem. We also show this bound is tight to within a constant factor. Surprisingly, the lower bound applies even for the simple case of undirected, planar graphs. Our results extend to the case in which the flow must decompose into at most k disjoint paths. There we obtain Θ(log k) upper and lower approximability bounds. 1
- …