140 research outputs found
Speeding up convolutional networks pruning with coarse ranking
Channel-based pruning has achieved significant successes in accelerating deep
convolutional neural network, whose pipeline is an iterative three-step
procedure: ranking, pruning and fine-tuning. However, this iterative procedure
is computationally expensive. In this study, we present a novel computationally
efficient channel pruning approach based on the coarse ranking that utilizes
the intermediate results during fine-tuning to rank the importance of filters,
built upon state-of-the-art works with data-driven ranking criteria. The goal
of this work is not to propose a single improved approach built upon a specific
channel pruning method, but to introduce a new general framework that works for
a series of channel pruning methods. Various benchmark image datasets
(CIFAR-10, ImageNet, Birds-200, and Flowers-102) and network architectures
(AlexNet and VGG-16) are utilized to evaluate the proposed approach for object
classification purpose. Experimental results show that the proposed method can
achieve almost identical performance with the corresponding state-of-the-art
works (baseline) while our ranking time is negligibly short. In specific, with
the proposed method, 75% and 54% of the total computation time for the whole
pruning procedure can be reduced for AlexNet on CIFAR-10, and for VGG-16 on
ImageNet, respectively. Our approach would significantly facilitate pruning
practice, especially on resource-constrained platforms.Comment: Submitted to ICIP 201
Structured Pruning for Efficient ConvNets via Incremental Regularization
Parameter pruning is a promising approach for CNN compression and
acceleration by eliminating redundant model parameters with tolerable
performance loss. Despite its effectiveness, existing regularization-based
parameter pruning methods usually drive weights towards zero with large and
constant regularization factors, which neglects the fact that the
expressiveness of CNNs is fragile and needs a more gentle way of regularization
for the networks to adapt during pruning. To solve this problem, we propose a
new regularization-based pruning method (named IncReg) to incrementally assign
different regularization factors to different weight groups based on their
relative importance, whose effectiveness is proved on popular CNNs compared
with state-of-the-art methods.Comment: Accepted by NIPS 2018 workshop on "Compact Deep Neural Network
Representation with Industrial Applications
AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Model compression is a critical technique to efficiently deploy neural
network models on mobile devices which have limited computation resources and
tight power budgets. Conventional model compression techniques rely on
hand-crafted heuristics and rule-based policies that require domain experts to
explore the large design space trading off among model size, speed, and
accuracy, which is usually sub-optimal and time-consuming. In this paper, we
propose AutoML for Model Compression (AMC) which leverage reinforcement
learning to provide the model compression policy. This learning-based
compression policy outperforms conventional rule-based compression policy by
having higher compression ratio, better preserving the accuracy and freeing
human labor. Under 4x FLOPs reduction, we achieved 2.7% better accuracy than
the handcrafted model compression policy for VGG-16 on ImageNet. We applied
this automated, push-the-button compression pipeline to MobileNet and achieved
1.81x speedup of measured inference latency on an Android phone and 1.43x
speedup on the Titan XP GPU, with only 0.1% loss of ImageNet Top-1 accuracy
SimGNN: A Neural Network Approach to Fast Graph Similarity Computation
Graph similarity search is among the most important graph-based applications,
e.g. finding the chemical compounds that are most similar to a query compound.
Graph similarity computation, such as Graph Edit Distance (GED) and Maximum
Common Subgraph (MCS), is the core operation of graph similarity search and
many other applications, but very costly to compute in practice. Inspired by
the recent success of neural network approaches to several graph applications,
such as node or graph classification, we propose a novel neural network based
approach to address this classic yet challenging graph problem, aiming to
alleviate the computational burden while preserving a good performance.
The proposed approach, called SimGNN, combines two strategies. First, we
design a learnable embedding function that maps every graph into a vector,
which provides a global summary of a graph. A novel attention mechanism is
proposed to emphasize the important nodes with respect to a specific similarity
metric. Second, we design a pairwise node comparison method to supplement the
graph-level embeddings with fine-grained node-level information. Our model
achieves better generalization on unseen graphs, and in the worst case runs in
quadratic time with respect to the number of nodes in two graphs. Taking GED
computation as an example, experimental results on three real graph datasets
demonstrate the effectiveness and efficiency of our approach. Specifically, our
model achieves smaller error rate and great time reduction compared against a
series of baselines, including several approximation algorithms on GED
computation, and many existing graph neural network based models. To the best
of our knowledge, we are among the first to adopt neural networks to explicitly
model the similarity between two graphs, and provide a new direction for future
research on graph similarity computation and graph similarity search.Comment: WSDM 201
Conditional Random Fields as Recurrent Neural Networks
Pixel-level labelling tasks, such as semantic segmentation, play a central
role in image understanding. Recent approaches have attempted to harness the
capabilities of deep learning techniques for image recognition to tackle
pixel-level labelling tasks. One central issue in this methodology is the
limited capacity of deep learning techniques to delineate visual objects. To
solve this problem, we introduce a new form of convolutional neural network
that combines the strengths of Convolutional Neural Networks (CNNs) and
Conditional Random Fields (CRFs)-based probabilistic graphical modelling. To
this end, we formulate mean-field approximate inference for the Conditional
Random Fields with Gaussian pairwise potentials as Recurrent Neural Networks.
This network, called CRF-RNN, is then plugged in as a part of a CNN to obtain a
deep network that has desirable properties of both CNNs and CRFs. Importantly,
our system fully integrates CRF modelling with CNNs, making it possible to
train the whole deep network end-to-end with the usual back-propagation
algorithm, avoiding offline post-processing methods for object delineation. We
apply the proposed method to the problem of semantic image segmentation,
obtaining top results on the challenging Pascal VOC 2012 segmentation
benchmark.Comment: This paper is published in IEEE ICCV 201
A Review on Deep Learning Techniques Applied to Semantic Segmentation
Image semantic segmentation is more and more being of interest for computer
vision and machine learning researchers. Many applications on the rise need
accurate and efficient segmentation mechanisms: autonomous driving, indoor
navigation, and even virtual or augmented reality systems to name a few. This
demand coincides with the rise of deep learning approaches in almost every
field or application target related to computer vision, including semantic
segmentation or scene understanding. This paper provides a review on deep
learning methods for semantic segmentation applied to various application
areas. Firstly, we describe the terminology of this field as well as mandatory
background concepts. Next, the main datasets and challenges are exposed to help
researchers decide which are the ones that best suit their needs and their
targets. Then, existing methods are reviewed, highlighting their contributions
and their significance in the field. Finally, quantitative results are given
for the described methods and the datasets in which they were evaluated,
following up with a discussion of the results. At last, we point out a set of
promising future works and draw our own conclusions about the state of the art
of semantic segmentation using deep learning techniques.Comment: Submitted to TPAMI on Apr. 22, 201
BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization
Mixed-precision quantization can potentially achieve the optimal tradeoff
between performance and compression rate of deep neural networks, and thus,
have been widely investigated. However, it lacks a systematic method to
determine the exact quantization scheme. Previous methods either examine only a
small manually-designed search space or utilize a cumbersome neural
architecture search to explore the vast search space. These approaches cannot
lead to an optimal quantization scheme efficiently. This work proposes
bit-level sparsity quantization (BSQ) to tackle the mixed-precision
quantization from a new angle of inducing bit-level sparsity. We consider each
bit of quantized weights as an independent trainable variable and introduce a
differentiable bit-sparsity regularizer. BSQ can induce all-zero bits across a
group of weight elements and realize the dynamic precision reduction, leading
to a mixed-precision quantization scheme of the original model. Our method
enables the exploration of the full mixed-precision space with a single
gradient-based optimization process, with only one hyperparameter to tradeoff
the performance and compression. BSQ achieves both higher accuracy and higher
bit reduction on various model architectures on the CIFAR-10 and ImageNet
datasets comparing to previous methods.Comment: Published as a conference paper at ICLR 202
A Taxonomy of Deep Convolutional Neural Nets for Computer Vision
Traditional architectures for solving computer vision problems and the degree
of success they enjoyed have been heavily reliant on hand-crafted features.
However, of late, deep learning techniques have offered a compelling
alternative -- that of automatically learning problem-specific features. With
this new paradigm, every problem in computer vision is now being re-examined
from a deep learning perspective. Therefore, it has become important to
understand what kind of deep networks are suitable for a given problem.
Although general surveys of this fast-moving paradigm (i.e. deep-networks)
exist, a survey specific to computer vision is missing. We specifically
consider one form of deep networks widely used in computer vision -
convolutional neural networks (CNNs). We start with "AlexNet" as our base CNN
and then examine the broad variations proposed over time to suit different
applications. We hope that our recipe-style survey will serve as a guide,
particularly for novice practitioners intending to use deep-learning techniques
for computer vision.Comment: Published in Frontiers in Robotics and AI (http://goo.gl/6691Bm
Object Detection in 20 Years: A Survey
Object detection, as of one the most fundamental and challenging problems in
computer vision, has received great attention in recent years. Its development
in the past two decades can be regarded as an epitome of computer vision
history. If we think of today's object detection as a technical aesthetics
under the power of deep learning, then turning back the clock 20 years we would
witness the wisdom of cold weapon era. This paper extensively reviews 400+
papers of object detection in the light of its technical evolution, spanning
over a quarter-century's time (from the 1990s to 2019). A number of topics have
been covered in this paper, including the milestone detectors in history,
detection datasets, metrics, fundamental building blocks of the detection
system, speed up techniques, and the recent state of the art detection methods.
This paper also reviews some important detection applications, such as
pedestrian detection, face detection, text detection, etc, and makes an in-deep
analysis of their challenges as well as technical improvements in recent years.Comment: This work has been submitted to the IEEE TPAMI for possible
publicatio
AdaDeep: A Usage-Driven, Automated Deep Model Compression Framework for Enabling Ubiquitous Intelligent Mobiles
Recent breakthroughs in Deep Neural Networks (DNNs) have fueled a
tremendously growing demand for bringing DNN-powered intelligence into mobile
platforms. While the potential of deploying DNNs on resource-constrained
platforms has been demonstrated by DNN compression techniques, the current
practice suffers from two limitations: 1) merely stand-alone compression
schemes are investigated even though each compression technique only suit for
certain types of DNN layers; and 2) mostly compression techniques are optimized
for DNNs' inference accuracy, without explicitly considering other
application-driven system performance (e.g., latency and energy cost) and the
varying resource availability across platforms (e.g., storage and processing
capability). To this end, we propose AdaDeep, a usage-driven, automated DNN
compression framework for systematically exploring the desired trade-off
between performance and resource constraints, from a holistic system level.
Specifically, in a layer-wise manner, AdaDeep automatically selects the most
suitable combination of compression techniques and the corresponding
compression hyperparameters for a given DNN. Thorough evaluations on six
datasets and across twelve devices demonstrate that AdaDeep can achieve up to
latency reduction, energy-efficiency improvement, and
storage reduction in DNNs while incurring negligible accuracy
loss. Furthermore, AdaDeep also uncovers multiple novel combinations of
compression techniques
- …