82,372 research outputs found
An Algorithm for Koml\'os Conjecture Matching Banaszczyk's bound
We consider the problem of finding a low discrepancy coloring for sparse set
systems where each element lies in at most t sets. We give an efficient
algorithm that finds a coloring with discrepancy O((t log n)^{1/2}), matching
the best known non-constructive bound for the problem due to Banaszczyk. The
previous algorithms only achieved an O(t^{1/2} log n) bound. The result also
extends to the more general Koml\'{o}s setting and gives an algorithmic
O(log^{1/2} n) bound
Chance Constrained Mixed Integer Program: Bilinear and Linear Formulations, and Benders Decomposition
In this paper, we study chance constrained mixed integer program with
consideration of recourse decisions and their incurred cost, developed on a
finite discrete scenario set. Through studying a non-traditional bilinear mixed
integer formulation, we derive its linear counterparts and show that they could
be stronger than existing linear formulations. We also develop a variant of
Jensen's inequality that extends the one for stochastic program. To solve this
challenging problem, we present a variant of Benders decomposition method in
bilinear form, which actually provides an easy-to-use algorithm framework for
further improvements, along with a few enhancement strategies based on
structural properties or Jensen's inequality. Computational study shows that
the presented Benders decomposition method, jointly with appropriate
enhancement techniques, outperforms a commercial solver by an order of
magnitude on solving chance constrained program or detecting its infeasibility
Latin hypercube sampling with inequality constraints
In some studies requiring predictive and CPU-time consuming numerical models,
the sampling design of the model input variables has to be chosen with caution.
For this purpose, Latin hypercube sampling has a long history and has shown its
robustness capabilities. In this paper we propose and discuss a new algorithm
to build a Latin hypercube sample (LHS) taking into account inequality
constraints between the sampled variables. This technique, called constrained
Latin hypercube sampling (cLHS), consists in doing permutations on an initial
LHS to honor the desired monotonic constraints. The relevance of this approach
is shown on a real example concerning the numerical welding simulation, where
the inequality constraints are caused by the physical decreasing of some
material properties in function of the temperature
Randomized Composable Core-sets for Distributed Submodular Maximization
An effective technique for solving optimization problems over massive data
sets is to partition the data into smaller pieces, solve the problem on each
piece and compute a representative solution from it, and finally obtain a
solution inside the union of the representative solutions for all pieces. This
technique can be captured via the concept of {\em composable core-sets}, and
has been recently applied to solve diversity maximization problems as well as
several clustering problems. However, for coverage and submodular maximization
problems, impossibility bounds are known for this technique \cite{IMMM14}. In
this paper, we focus on efficient construction of a randomized variant of
composable core-sets where the above idea is applied on a {\em random
clustering} of the data. We employ this technique for the coverage, monotone
and non-monotone submodular maximization problems. Our results significantly
improve upon the hardness results for non-randomized core-sets, and imply
improved results for submodular maximization in a distributed and streaming
settings.
In summary, we show that a simple greedy algorithm results in a
-approximate randomized composable core-set for submodular maximization
under a cardinality constraint. This is in contrast to a known impossibility result for (non-randomized) composable core-set. Our
result also extends to non-monotone submodular functions, and leads to the
first 2-round MapReduce-based constant-factor approximation algorithm with
total communication complexity for either monotone or non-monotone
functions. Finally, using an improved analysis technique and a new algorithm
, we present an improved -approximation algorithm
for monotone submodular maximization, which is in turn the first
MapReduce-based algorithm beating factor in a constant number of rounds
Closing the Gap Between Short and Long XORs for Model Counting
Many recent algorithms for approximate model counting are based on a
reduction to combinatorial searches over random subsets of the space defined by
parity or XOR constraints. Long parity constraints (involving many variables)
provide strong theoretical guarantees but are computationally difficult. Short
parity constraints are easier to solve but have weaker statistical properties.
It is currently not known how long these parity constraints need to be. We
close the gap by providing matching necessary and sufficient conditions on the
required asymptotic length of the parity constraints. Further, we provide a new
family of lower bounds and the first non-trivial upper bounds on the model
count that are valid for arbitrarily short XORs. We empirically demonstrate the
effectiveness of these bounds on model counting benchmarks and in a
Satisfiability Modulo Theory (SMT) application motivated by the analysis of
contingency tables in statistics.Comment: The 30th Association for the Advancement of Artificial Intelligence
(AAAI-16) Conferenc
Sum of squares lower bounds for refuting any CSP
Let be a nontrivial -ary predicate. Consider a
random instance of the constraint satisfaction problem on
variables with constraints, each being applied to randomly
chosen literals. Provided the constraint density satisfies , such
an instance is unsatisfiable with high probability. The \emph{refutation}
problem is to efficiently find a proof of unsatisfiability.
We show that whenever the predicate supports a -\emph{wise uniform}
probability distribution on its satisfying assignments, the sum of squares
(SOS) algorithm of degree
(which runs in time ) \emph{cannot} refute a random instance of
. In particular, the polynomial-time SOS algorithm requires
constraints to refute random instances of
CSP when supports a -wise uniform distribution on its satisfying
assignments. Together with recent work of Lee et al. [LRS15], our result also
implies that \emph{any} polynomial-size semidefinite programming relaxation for
refutation requires at least constraints.
Our results (which also extend with no change to CSPs over larger alphabets)
subsume all previously known lower bounds for semialgebraic refutation of
random CSPs. For every constraint predicate~, they give a three-way hardness
tradeoff between the density of constraints, the SOS degree (hence running
time), and the strength of the refutation. By recent algorithmic results of
Allen et al. [AOW15] and Raghavendra et al. [RRS16], this full three-way
tradeoff is \emph{tight}, up to lower-order factors.Comment: 39 pages, 1 figur
Dependent randomized rounding for clustering and partition systems with knapsack constraints
Clustering problems are fundamental to unsupervised learning. There is an
increased emphasis on fairness in machine learning and AI; one representative
notion of fairness is that no single demographic group should be
over-represented among the cluster-centers. This, and much more general
clustering problems, can be formulated with "knapsack" and "partition"
constraints. We develop new randomized algorithms targeting such problems, and
study two in particular: multi-knapsack median and multi-knapsack center. Our
rounding algorithms give new approximation and pseudo-approximation algorithms
for these problems. One key technical tool, which may be of independent
interest, is a new tail bound analogous to Feige (2006) for sums of random
variables with unbounded variances. Such bounds are very useful in inferring
properties of large networks using few samples
- …