194,399 research outputs found
Adaptive Quantization for Deep Neural Network
In recent years Deep Neural Networks (DNNs) have been rapidly developed in
various applications, together with increasingly complex architectures. The
performance gain of these DNNs generally comes with high computational costs
and large memory consumption, which may not be affordable for mobile platforms.
Deep model quantization can be used for reducing the computation and memory
costs of DNNs, and deploying complex DNNs on mobile equipment. In this work, we
propose an optimization framework for deep model quantization. First, we
propose a measurement to estimate the effect of parameter quantization errors
in individual layers on the overall model prediction accuracy. Then, we propose
an optimization process based on this measurement for finding optimal
quantization bit-width for each layer. This is the first work that
theoretically analyse the relationship between parameter quantization errors of
individual layers and model accuracy. Our new quantization algorithm
outperforms previous quantization optimization methods, and achieves 20-40%
higher compression rate compared to equal bit-width quantization at the same
model prediction accuracy.Comment: 9 pages main paper + 5 pages supplementary, 8 figures, conferenc
Solving the Insoluble: A New Rule for Quantization
The rules of canonical quantization normally offer good results, but
sometimes they fail, e.g., leading to quantum triviality ( free) for certain
examples that are classically nontrivial ( free). A new procedure, called
Enhanced Quantization, relates classical models with their quantum partners
differently and leads to satisfactory results for all systems. This paper
features enhanced quantization procedures and provides highlights of two
examples, a rotationally symmetric model and an ultralocal scalar model, for
which canonical quantization fails while enhanced quantization succeeds.Comment: 10 pages, several minor corrections, contribution to 2017 Coherent
States workshop as a CIRM conference proceeding
Semilogarithmic Nonuniform Vector Quantization of Two-Dimensional Laplacean Source for Small Variance Dynamics
In this paper high dynamic range nonuniform two-dimensional vector quantization model for Laplacean source was provided. Semilogarithmic A-law compression characteristic was used as radial scalar compression characteristic of two-dimensional vector quantization. Optimal number value of concentric quantization domains (amplitude levels) is expressed in the function of parameter A. Exact distortion analysis with obtained closed form expressions is provided. It has been shown that proposed model provides high SQNR values in wide range of variances, and overachieves quality obtained by scalar A-law quantization at same bit rate, so it can be used in various switching and adaptation implementations for realization of high quality signal compression
The Gervais-Neveu-Felder equation and the quantum Calogero-Moser systems
We quantize the spin Calogero-Moser model in the -matrix formalism. The
quantum -matrix of the model is dynamical. This -matrix has already
appeared in Gervais-Neveu's quantization of Toda field theory and in Felder's
quantization of the Knizhnik-Zamolodchikov-Bernard equation.Comment: Comments and References adde
HAQ: Hardware-Aware Automated Quantization with Mixed Precision
Model quantization is a widely used technique to compress and accelerate deep
neural network (DNN) inference. Emergent DNN hardware accelerators begin to
support mixed precision (1-8 bits) to further improve the computation
efficiency, which raises a great challenge to find the optimal bitwidth for
each layer: it requires domain experts to explore the vast design space trading
off among accuracy, latency, energy, and model size, which is both
time-consuming and sub-optimal. Conventional quantization algorithm ignores the
different hardware architectures and quantizes all the layers in a uniform way.
In this paper, we introduce the Hardware-Aware Automated Quantization (HAQ)
framework which leverages the reinforcement learning to automatically determine
the quantization policy, and we take the hardware accelerator's feedback in the
design loop. Rather than relying on proxy signals such as FLOPs and model size,
we employ a hardware simulator to generate direct feedback signals (latency and
energy) to the RL agent. Compared with conventional methods, our framework is
fully automated and can specialize the quantization policy for different neural
network architectures and hardware architectures. Our framework effectively
reduced the latency by 1.4-1.95x and the energy consumption by 1.9x with
negligible loss of accuracy compared with the fixed bitwidth (8 bits)
quantization. Our framework reveals that the optimal policies on different
hardware architectures (i.e., edge and cloud architectures) under different
resource constraints (i.e., latency, energy and model size) are drastically
different. We interpreted the implication of different quantization policies,
which offer insights for both neural network architecture design and hardware
architecture design.Comment: CVPR 2019. The first three authors contributed equally to this work.
Project page: https://hanlab.mit.edu/projects/haq
Deformation Quantization and Reduction
This note is an overview of the Poisson sigma model (PSM) and its
applications in deformation quantization. Reduction of coisotropic and
pre-Poisson submanifolds, their appearance as branes of the PSM, quantization
in terms of L-infinity and A-infinity algebras, and bimodule structures are
recalled. As an application, an "almost" functorial quantization of Poisson
maps is presented if no anomalies occur. This leads in principle to a novel
approach for the quantization of Poisson-Lie groups.Comment: 23 pages, 3 figures; added references, corrected typo
Absence of magnetically-induced fractional quantization in atomic contacts
Using the mechanically controlled break junction technique at low
temperatures and under cryogenic vacuum conditions we have studied atomic
contacts of several magnetic (Fe, Co and Ni) and non-magnetic (Pt) metals,
which recently were claimed to show fractional conductance quantization. In the
case of pure metals we see no quantization of the conductance nor
half-quantization, even when high magnetic fields are applied. On the other
hand, features in the conductance similar to (fractional) quantization are
observed when the contact is exposed to gas molecules. Furthermore, the absence
of fractional quantization when the contact is bridged by H_2 indicates the
current is never fully polarized for the metals studied here. Our results are
in agreement with recent model calculations.Comment: 4 pages, 3 figure
- …
