30 research outputs found
The Lottery Ticket Hypothesis for Vision Transformers
The conventional lottery ticket hypothesis (LTH) claims that there exists a
sparse subnetwork within a dense neural network and a proper random
initialization method, called the winning ticket, such that it can be trained
from scratch to almost as good as the dense counterpart. Meanwhile, the
research of LTH in vision transformers (ViTs) is scarcely evaluated. In this
paper, we first show that the conventional winning ticket is hard to find at
weight level of ViTs by existing methods. Then, we generalize the LTH for ViTs
to input images consisting of image patches inspired by the input dependence of
ViTs. That is, there exists a subset of input image patches such that a ViT can
be trained from scratch by using only this subset of patches and achieve
similar accuracy to the ViTs trained by using all image patches. We call this
subset of input patches the winning tickets, which represent a significant
amount of information in the input. Furthermore, we present a simple yet
effective method to find the winning tickets in input patches for various types
of ViT, including DeiT, LV-ViT, and Swin Transformers. More specifically, we
use a ticket selector to generate the winning tickets based on the
informativeness of patches. Meanwhile, we build another randomly selected
subset of patches for comparison, and the experiments show that there is clear
difference between the performance of models trained with winning tickets and
randomly selected subsets
Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Vision transformers (ViTs) have recently obtained success in many
applications, but their intensive computation and heavy memory usage at both
training and inference time limit their generalization. Previous compression
algorithms usually start from the pre-trained dense models and only focus on
efficient inference, while time-consuming training is still unavoidable. In
contrast, this paper points out that the million-scale training data is
redundant, which is the fundamental reason for the tedious training. To address
the issue, this paper aims to introduce sparsity into data and proposes an
end-to-end efficient training framework from three sparse perspectives, dubbed
Tri-Level E-ViT. Specifically, we leverage a hierarchical data redundancy
reduction scheme, by exploring the sparsity under three levels: number of
training examples in the dataset, number of patches (tokens) in each example,
and number of connections between tokens that lie in attention weights. With
extensive experiments, we demonstrate that our proposed technique can
noticeably accelerate training for various ViT architectures while maintaining
accuracy. Remarkably, under certain ratios, we are able to improve the ViT
accuracy rather than compromising it. For example, we can achieve 15.2% speedup
with 72.6% (+0.4) Top-1 accuracy on Deit-T, and 15.7% speedup with 79.9% (+0.1)
Top-1 accuracy on Deit-S. This proves the existence of data redundancy in ViT.Comment: AAAI 202
Quantum Neural Network Compression
Model compression, such as pruning and quantization, has been widely applied
to optimize neural networks on resource-limited classical devices. Recently,
there are growing interest in variational quantum circuits (VQC), that is, a
type of neural network on quantum computers (a.k.a., quantum neural networks).
It is well known that the near-term quantum devices have high noise and limited
resources (i.e., quantum bits, qubits); yet, how to compress quantum neural
networks has not been thoroughly studied. One might think it is straightforward
to apply the classical compression techniques to quantum scenarios. However,
this paper reveals that there exist differences between the compression of
quantum and classical neural networks. Based on our observations, we claim that
the compilation/traspilation has to be involved in the compression process. On
top of this, we propose the very first systematical framework, namely CompVQC,
to compress quantum neural networks (QNNs).In CompVQC, the key component is a
novel compression algorithm, which is based on the alternating direction method
of multipliers (ADMM) approach. Experiments demonstrate the advantage of the
CompVQC, reducing the circuit depth (almost over 2.5 %) with a negligible
accuracy drop (<1%), which outperforms other competitors. Another promising
truth is our CompVQC can indeed promote the robustness of the QNN on the
near-term noisy quantum devices
Multi-Scenario Simulations of Land Use and Habitat Quality Based on a PLUS-InVEST Model: A Case Study of Baoding, China
Habitat quality and ecosystem service value (ESV) are important foundations for sustainable development. Baoding, as the strategic hinterland of Beijing–Tianjin–Hebei, is of great significance to regional ecological conservation and sustainable urban development. Based on land-use data from 2000 to 2020, the land-use scenarios of natural development (ND), water protection (WP), forest rehabilitation (FR), and cultivated land protection (CP) in 2030 were predicted by the PLUS model and adopt the InVEST model and equivalent ESV table to assess ecological sustainability. The results show that: (1) From 2000 to 2020, the construction land in Baoding has increased by 812 km2, and the cultivated land and forest land decreased by 708 km2 and 154 km2. Habitat quality is obviously deteriorating in 4.66% of the city. (2) Under different scenarios, the order of habitat quality is CP > FR > WP > ND. The habitat quality under each scenario is dominated by medium habitat quality. (3) Under different scenarios, the order of ESV is FR > CP> WP > ND. The fluctuation of forest land and cultivated land scale is affecting the ESV. (4) CP and FR will form a land-use pattern that has “high ecological quality and value”, which better balances the economic development and ecological protection of Baoding. This research study will provide a reference for the effective allocation of land resources and will guide the formulation of urban land space planning policy in Baoding