52 research outputs found
From Hashing to CNNs: Training BinaryWeight Networks via Hashing
Deep convolutional neural networks (CNNs) have shown appealing performance on
various computer vision tasks in recent years. This motivates people to deploy
CNNs to realworld applications. However, most of state-of-art CNNs require
large memory and computational resources, which hinders the deployment on
mobile devices. Recent studies show that low-bit weight representation can
reduce much storage and memory demand, and also can achieve efficient network
inference. To achieve this goal, we propose a novel approach named BWNH to
train Binary Weight Networks via Hashing. In this paper, we first reveal the
strong connection between inner-product preserving hashing and binary weight
networks, and show that training binary weight networks can be intrinsically
regarded as a hashing problem. Based on this perspective, we propose an
alternating optimization method to learn the hash codes instead of directly
learning binary weights. Extensive experiments on CIFAR10, CIFAR100 and
ImageNet demonstrate that our proposed BWNH outperforms current state-of-art by
a large margin
Supply, Sales and Inventory Management System Based Java and SQL Server
The development of the inventory management system mainly includes the establishment and maintenance of the database and the development of the Java. The purpose is to facilitate the store operators to input and modify the data of incoming and outgoing merchandise, customers and suppliers, easy and fast incoming and outgoing merchandise data search, complete incoming and outgoing merchandise receipt and payment fund management, and flexible incoming and outgoing merchandise inventory statistics. This paper presents the detailed design of this management system, with particular emphasis on the database part of the design, including requirement analysis, conceptual structure design, logical structure design, physical structure design, etc., and detailed testing of the system's functions
DATE: Dual Assignment for End-to-End Fully Convolutional Object Detection
Fully convolutional detectors discard the one-to-many assignment and adopt a
one-to-one assigning strategy to achieve end-to-end detection but suffer from
the slow convergence issue. In this paper, we revisit these two assignment
methods and find that bringing one-to-many assignment back to end-to-end fully
convolutional detectors helps with model convergence. Based on this
observation, we propose {\em \textbf{D}ual \textbf{A}ssignment} for end-to-end
fully convolutional de\textbf{TE}ction (DATE). Our method constructs two
branches with one-to-many and one-to-one assignment during training and speeds
up the convergence of the one-to-one assignment branch by providing more
supervision signals. DATE only uses the branch with the one-to-one matching
strategy for model inference, which doesn't bring inference overhead.
Experimental results show that Dual Assignment gives nontrivial improvements
and speeds up model convergence upon OneNet and DeFCN. Code:
https://github.com/YiqunChen1999/date
PalQuant: Accelerating High-precision Networks on Low-precision Accelerators
Recently low-precision deep learning accelerators (DLAs) have become popular
due to their advantages in chip area and energy consumption, yet the
low-precision quantized models on these DLAs bring in severe accuracy
degradation. One way to achieve both high accuracy and efficient inference is
to deploy high-precision neural networks on low-precision DLAs, which is rarely
studied. In this paper, we propose the PArallel Low-precision Quantization
(PalQuant) method that approximates high-precision computations via learning
parallel low-precision representations from scratch. In addition, we present a
novel cyclic shuffle module to boost the cross-group information communication
between parallel low-precision groups. Extensive experiments demonstrate that
PalQuant has superior performance to state-of-the-art quantization methods in
both accuracy and inference speed, e.g., for ResNet-18 network quantization,
PalQuant can obtain 0.52\% higher accuracy and 1.78 speedup
simultaneously over their 4-bit counter-part on a state-of-the-art 2-bit
accelerator. Code is available at \url{https://github.com/huqinghao/PalQuant}.Comment: accepted by ECCV202
Spiking NeRF: Making Bio-inspired Neural Networks See through the Real World
Spiking neuron networks (SNNs) have been thriving on numerous tasks to
leverage their promising energy efficiency and exploit their potentialities as
biologically plausible intelligence. Meanwhile, the Neural Radiance Fields
(NeRF) render high-quality 3D scenes with massive energy consumption, and few
works delve into the energy-saving solution with a bio-inspired approach. In
this paper, we propose spiking NeRF (SpikingNeRF), which aligns the radiance
ray with the temporal dimension of SNN, to naturally accommodate the SNN to the
reconstruction of Radiance Fields. Thus, the computation turns into a
spike-based, multiplication-free manner, reducing the energy consumption. In
SpikingNeRF, each sampled point on the ray is matched onto a particular time
step, and represented in a hybrid manner where the voxel grids are maintained
as well. Based on the voxel grids, sampled points are determined whether to be
masked for better training and inference. However, this operation also incurs
irregular temporal length. We propose the temporal condensing-and-padding (TCP)
strategy to tackle the masked samples to maintain regular temporal length,
i.e., regular tensors, for hardware-friendly computation. Extensive experiments
on a variety of datasets demonstrate that our method reduces the
energy consumption on average and obtains comparable synthesis quality with the
ANN baseline
- …