10,601 research outputs found
Submodular Maximization by Simulated Annealing
We consider the problem of maximizing a nonnegative (possibly non-monotone)
submodular set function with or without constraints. Feige et al. [FOCS'07]
showed a 2/5-approximation for the unconstrained problem and also proved that
no approximation better than 1/2 is possible in the value oracle model.
Constant-factor approximation was also given for submodular maximization
subject to a matroid independence constraint (a factor of 0.309 Vondrak
[FOCS'09]) and for submodular maximization subject to a matroid base
constraint, provided that the fractional base packing number is at least 2 (a
1/4-approximation, Vondrak [FOCS'09]).
In this paper, we propose a new algorithm for submodular maximization which
is based on the idea of {\em simulated annealing}. We prove that this algorithm
achieves improved approximation for two problems: a 0.41-approximation for
unconstrained submodular maximization, and a 0.325-approximation for submodular
maximization subject to a matroid independence constraint.
On the hardness side, we show that in the value oracle model it is impossible
to achieve a 0.478-approximation for submodular maximization subject to a
matroid independence constraint, or a 0.394-approximation subject to a matroid
base constraint in matroids with two disjoint bases. Even for the special case
of cardinality constraint, we prove it is impossible to achieve a
0.491-approximation. (Previously it was conceivable that a 1/2-approximation
exists for these problems.) It is still an open question whether a
1/2-approximation is possible for unconstrained submodular maximization
Constrained Bayesian Optimization for Automatic Chemical Design
Automatic Chemical Design is a framework for generating novel molecules with
optimized properties. The original scheme, featuring Bayesian optimization over
the latent space of a variational autoencoder, suffers from the pathology that
it tends to produce invalid molecular structures. First, we demonstrate
empirically that this pathology arises when the Bayesian optimization scheme
queries latent points far away from the data on which the variational
autoencoder has been trained. Secondly, by reformulating the search procedure
as a constrained Bayesian optimization problem, we show that the effects of
this pathology can be mitigated, yielding marked improvements in the validity
of the generated molecules. We posit that constrained Bayesian optimization is
a good approach for solving this class of training set mismatch in many
generative tasks involving Bayesian optimization over the latent space of a
variational autoencoder.Comment: Previous versions accepted to the NIPS 2017 Workshop on Bayesian
Optimization (BayesOpt 2017) and the NIPS 2017 Workshop on Machine Learning
for Molecules and Material
Bounded Degree Approximations of Stochastic Networks
We propose algorithms to approximate directed information graphs. Directed
information graphs are probabilistic graphical models that depict causal
dependencies between stochastic processes in a network. The proposed algorithms
identify optimal and near-optimal approximations in terms of Kullback-Leibler
divergence. The user-chosen sparsity trades off the quality of the
approximation against visual conciseness and computational tractability. One
class of approximations contains graphs with specified in-degrees. Another
class additionally requires that the graph is connected. For both classes, we
propose algorithms to identify the optimal approximations and also near-optimal
approximations, using a novel relaxation of submodularity. We also propose
algorithms to identify the r-best approximations among these classes, enabling
robust decision making
Hierarchical Clustering with Structural Constraints
Hierarchical clustering is a popular unsupervised data analysis method. For
many real-world applications, we would like to exploit prior information about
the data that imposes constraints on the clustering hierarchy, and is not
captured by the set of features available to the algorithm. This gives rise to
the problem of "hierarchical clustering with structural constraints".
Structural constraints pose major challenges for bottom-up approaches like
average/single linkage and even though they can be naturally incorporated into
top-down divisive algorithms, no formal guarantees exist on the quality of
their output. In this paper, we provide provable approximation guarantees for
two simple top-down algorithms, using a recently introduced optimization
viewpoint of hierarchical clustering with pairwise similarity information
[Dasgupta, 2016]. We show how to find good solutions even in the presence of
conflicting prior information, by formulating a constraint-based regularization
of the objective. We further explore a variation of this objective for
dissimilarity information [Cohen-Addad et al., 2018] and improve upon current
techniques. Finally, we demonstrate our approach on a real dataset for the
taxonomy application.Comment: In Proc. 35th International Conference on Machine Learning (ICML
2018
Submodular Maximization Beyond Non-negativity: Guarantees, Fast Algorithms, and Applications
It is generally believed that submodular functions -- and the more general
class of -weakly submodular functions -- may only be optimized under
the non-negativity assumption . In this paper, we show that once
the function is expressed as the difference , where is monotone,
non-negative, and -weakly submodular and is non-negative modular,
then strong approximation guarantees may be obtained. We present an algorithm
for maximizing under a -cardinality constraint which produces a
random feasible set such that , whose running time is , i.e., independent of . We
extend these results to the unconstrained setting by describing an algorithm
with the same approximation guarantees and faster runtime. The main techniques underlying our algorithms
are two-fold: the use of a surrogate objective which varies the relative
importance between and throughout the algorithm, and a geometric sweep
over possible values. Our algorithmic guarantees are complemented by a
hardness result showing that no polynomial-time algorithm which accesses
through a value oracle can do better. We empirically demonstrate the success of
our algorithms by applying them to experimental design on the Boston Housing
dataset and directed vertex cover on the Email EU dataset.Comment: submitted to ICML 201
Exploring the Scope of Unconstrained Via Minimization by Recursive Floorplan Bipartitioning
Random via failure is a major concern for post-fabrication reliability and
poor manufacturing yield. A demanding solution to this problem is redundant via
insertion during post-routing optimization. It becomes very critical when a
multi-layer routing solution already incurs a large number of vias. Very few
global routers addressed unconstrained via minimization (UVM) problem, while
using minimal pattern routing and layer assignment of nets. It also includes a
recent floorplan based early global routability assessment tool STAIRoute
\cite{karb2}.
This work addresses an early version of unconstrained via minimization
problem during early global routing by identifying a set of minimal bend
routing regions in any floorplan, by a new recursive bipartitioning framework.
These regions facilitate monotone pattern routing of a set of nets in the
floorplan by STAIRoute. The area/number balanced floorplan bipartitionining is
a multi-objective optimization problem and known to be NP-hard \cite{majum2}.
No existing approaches considered bend minimization as an objective and some of
them incurred higher runtime overhead. In this paper, we present a Greedy as
well as randomized neighbor search based staircase wave-front propagation
methods for obtaining optimal bipartitioning results for minimal bend routing
through multiple routing layers, for a balanced trade-off between routability,
wirelength and congestion.
Experiments were conducted on MCNC/GSRC floorplanning benchmarks for studying
the variation of early via count obtained by STAIRoute for different values of
the trade-off parameters () in this multi-objective optimization
problem, using metal layers. We studied the impact of ()
values on each of the objectives as well as their linear combination function
of these objectives.Comment: A draft aimed at ACM TODAES journal, 25 pages with 16 figures and 2
table
Achieving High Coverage for Floating-point Code via Unconstrained Programming (Extended Version)
Achieving high code coverage is essential in testing, which gives us
confidence in code quality. Testing floating-point code usually requires
painstaking efforts in handling floating-point constraints, e.g., in symbolic
execution. This paper turns the challenge of testing floating-point code into
the opportunity of applying unconstrained programming --- the mathematical
solution for calculating function minimum points over the entire search space.
Our core insight is to derive a representing function from the floating-point
program, any of whose minimum points is a test input guaranteed to exercise a
new branch of the tested program. This guarantee allows us to achieve high
coverage of the floating-point program by repeatedly minimizing the
representing function.
We have realized this approach in a tool called CoverMe and conducted an
extensive evaluation of it on Sun's C math library. Our evaluation results show
that CoverMe achieves, on average, 90.8% branch coverage in 6.9 seconds,
drastically outperforming our compared tools: (1) Random testing, (2) AFL, a
highly optimized, robust fuzzer released by Google, and (3) Austin, a
state-of-the-art coverage-based testing tool designed to support floating-point
code.Comment: Extended version of Fu and Su's PLDI'17 paper. arXiv admin note: text
overlap with arXiv:1610.0113
D-Optimal Input Design for Nonlinear FIR-type Systems:A Dispersion-based Approach
Optimal input design is an important step of the identification process in
order to reduce the model variance. In this work a D-optimal input design
method for finite-impulse-response-type nonlinear systems is presented. The
optimization of the determinant of the Fisher information matrix is expressed
as a convex optimization problem. This problem is then solved using a
dispersion-based optimization scheme, which is easy to implement and converges
monotonically to the optimal solution. Without constraints, the optimal design
cannot be realized as a time sequence. By imposing that the design should lie
in the subspace described by a symmetric and non-overlapping set, a realizable
design is found. A graph-based method is used in order to find a time sequence
that realizes this optimal constrained design. These methods are illustrated on
a numerical example of which the results are thoroughly discussed. Additionally
the computational speed of the algorithm is compared with the general convex
optimizer cvx.Comment: 15 pages, 11 figure
Online Submodular Maximization with Preemption
Submodular function maximization has been studied extensively in recent years
under various constraints and models. The problem plays a major role in various
disciplines. We study a natural online variant of this problem in which
elements arrive one-by-one and the algorithm has to maintain a solution obeying
certain constraints at all times. Upon arrival of an element, the algorithm has
to decide whether to accept the element into its solution and may preempt
previously chosen elements. The goal is to maximize a submodular function over
the set of elements in the solution.
We study two special cases of this general problem and derive upper and lower
bounds on the competitive ratio. Specifically, we design a -competitive
algorithm for the unconstrained case in which the algorithm may hold any subset
of the elements, and constant competitive ratio algorithms for the case where
the algorithm may hold at most elements in its solution.Comment: 32 pages, an extended abstract of this work appeared in SODA 201
A New Look at Survey Propagation and its Generalizations
This paper provides a new conceptual perspective on survey propagation, which
is an iterative algorithm recently introduced by the statistical physics
community that is very effective in solving random k-SAT problems even with
densities close to the satisfiability threshold. We first describe how any SAT
formula can be associated with a novel family of Markov random fields (MRFs),
parameterized by a real number \rho \in [0,1]. We then show that applying
belief propagation--a well-known ``message-passing'' technique for estimating
marginal probabilities--to this family of MRFs recovers a known family of
algorithms, ranging from pure survey propagation at one extreme (\rho = 1) to
standard belief propagation on the uniform distribution over SAT assignments at
the other extreme (\rho = 0). Configurations in these MRFs have a natural
interpretation as partial satisfiability assignments, on which a partial order
can be defined. We isolate cores as minimal elements in this partial ordering,
which are also fixed points of survey propagation and the only assignments with
positive probability in the MRF for \rho=1. Our experimental results for k=3
suggest that solutions of random formulas typically do not possess non-trivial
cores. This makes it necessary to study the structure of the space of partial
assignments for \rho<1 and investigate the role of assignments that are very
close to being cores. To that end, we investigate the associated lattice
structure, and prove a weight-preserving identity that shows how any MRF with
\rho>0 can be viewed as a ``smoothed'' version of the uniform distribution over
satisfying assignments (\rho=0). Finally, we isolate properties of Gibbs
sampling and message-passing algorithms that are typical for an ensemble of
k-SAT problems.Comment: v2:typoes and reference corrections; v3: expanded expositio
- …