864 research outputs found
Fine-grained Search Space Classification for Hard Enumeration Variants of Subset Problems
We propose a simple, powerful, and flexible machine learning framework for
(i) reducing the search space of computationally difficult enumeration variants
of subset problems and (ii) augmenting existing state-of-the-art solvers with
informative cues arising from the input distribution. We instantiate our
framework for the problem of listing all maximum cliques in a graph, a central
problem in network analysis, data mining, and computational biology. We
demonstrate the practicality of our approach on real-world networks with
millions of vertices and edges by not only retaining all optimal solutions, but
also aggressively pruning the input instance size resulting in several fold
speedups of state-of-the-art algorithms. Finally, we explore the limits of
scalability and robustness of our proposed framework, suggesting that
supervised learning is viable for tackling NP-hard problems in practice.Comment: AAAI 201
Learning to Prune Instances of Steiner Tree Problem in Graphs
We consider the Steiner tree problem on graphs where we are given a set of
nodes and the goal is to find a tree sub-graph of minimum weight that contains
all nodes in the given set, potentially including additional nodes. This is a
classical NP-hard combinatorial optimisation problem. In recent years, a
machine learning framework called learning-to-prune has been successfully used
for solving a diverse range of combinatorial optimisation problems. In this
paper, we use this learning framework on the Steiner tree problem and show that
even on this problem, the learning-to-prune framework results in computing
near-optimal solutions at a fraction of the time required by commercial ILP
solvers. Our results underscore the potential of the learning-to-prune
framework in solving various combinatorial optimisation problems
A reusable iterative optimization software library to solve combinatorial problems with approximate reasoning
Real world combinatorial optimization problems such as scheduling are
typically too complex to solve with exact methods. Additionally, the problems
often have to observe vaguely specified constraints of different importance,
the available data may be uncertain, and compromises between antagonistic
criteria may be necessary. We present a combination of approximate reasoning
based constraints and iterative optimization based heuristics that help to
model and solve such problems in a framework of C++ software libraries called
StarFLIP++. While initially developed to schedule continuous caster units in
steel plants, we present in this paper results from reusing the library
components in a shift scheduling system for the workforce of an industrial
production plant.Comment: 33 pages, 9 figures; for a project overview see
http://www.dbai.tuwien.ac.at/proj/StarFLIP
End-to-End Neural Network Compression via Regularized Latency Surrogates
Neural network (NN) compression via techniques such as pruning, quantization
requires setting compression hyperparameters (e.g., number of channels to be
pruned, bitwidths for quantization) for each layer either manually or via
neural architecture search (NAS) which can be computationally expensive. We
address this problem by providing an end-to-end technique that optimizes for
model's Floating Point Operations (FLOPs) or for on-device latency via a novel
latency surrogate. Our algorithm is versatile and can
be used with many popular compression methods including pruning, low-rank
factorization, and quantization. Crucially, it is fast and runs in almost the
same amount of time as single model training; which is a significant training
speed-up over standard NAS methods. For BERT compression on GLUE fine-tuning
tasks, we achieve reduction in FLOPs with only drop in
performance. For compressing MobileNetV3 on ImageNet-1K, we achieve
reduction in FLOPs, and reduction in on-device latency without drop in
accuracy, while still requiring less training compute than SOTA
compression techniques. Finally, for transfer learning on smaller datasets, our
technique identifies - cheaper architectures than
standard MobileNetV3, EfficientNet suite of architectures at almost the same
training cost and accuracy
Recommended from our members
Mayo: A Framework for Auto-generating Hardware Friendly Deep Neural Networks
Deep Neural Networks (DNNs) have proved to be a conve-
nient and powerful tool for a wide range of problems. How-
ever, the extensive computational and memory resource re-
quirements hinder the adoption of DNNs in resource-con-
strained scenarios. Existing compression methods have been
shown to significantly reduce the computation and mem-
ory requirements of many popular DNNs. These methods,
however, remain elusive to non-experts, as they demand ex-
tensive manual tuning of hyperparameters. The effects of
combining various compression techniques lack exploration
because of the large design space. To alleviate these chal-
lenges, this paper proposes an automated framework, Mayo,
which is built on top of TensorFlow and can compress DNNs
with minimal human intervention. First, we present over-
riders which are recursively-compositional and can be con-
figured to effectively compress individual components (e.g.
weights, biases, layer computations and gradients) in a DNN.
Second, we introduce novel heuristics and a global search al-
gorithm to efficiently optimize hyperparameters. We demon-
strate that without any manual tuning, Mayo generates a sparse
ResNet-18 that is 5.13Γ smaller than the baseline with no
loss in test accuracy. By composing multiple overriders,
our tool produces a sparse 6-bit CIFAR-10 classifier with
only 0.16% top-1 accuracy loss and a 34Γ compression rate.
Mayo and all compressed models are publicly available. To
our knowledge, Mayo is the first framework that supports
overlapping multiple compression techniques and automati-
cally optimizes hyperparameters in them.EPSR
- β¦