38 research outputs found
Training Group Orthogonal Neural Networks with Privileged Information
Learning rich and diverse representations is critical for the performance of
deep convolutional neural networks (CNNs). In this paper, we consider how to
use privileged information to promote inherent diversity of a single CNN model
such that the model can learn better representations and offer stronger
generalization ability. To this end, we propose a novel group orthogonal
convolutional neural network (GoCNN) that learns untangled representations
within each layer by exploiting provided privileged information and enhances
representation diversity effectively. We take image classification as an
example where image segmentation annotations are used as privileged information
during the training process. Experiments on two benchmark datasets -- ImageNet
and PASCAL VOC -- clearly demonstrate the strong generalization ability of our
proposed GoCNN model. On the ImageNet dataset, GoCNN improves the performance
of state-of-the-art ResNet-152 model by absolute value of 1.2% while only uses
privileged information of 10% of the training images, confirming effectiveness
of GoCNN on utilizing available privileged knowledge to train better CNNs.Comment: Proceedings of the IJCAI-1
Efficient CNN with uncorrelated Bag of Features pooling
Despite the superior performance of CNN, deploying them on low computational
power devices is still limited as they are typically computationally expensive.
One key cause of the high complexity is the connection between the convolution
layers and the fully connected layers, which typically requires a high number
of parameters. To alleviate this issue, Bag of Features (BoF) pooling has been
recently proposed. BoF learns a dictionary, that is used to compile a histogram
representation of the input. In this paper, we propose an approach that builds
on top of BoF pooling to boost its efficiency by ensuring that the items of the
learned dictionary are non-redundant. We propose an additional loss term, based
on the pair-wise correlation of the items of the dictionary, which complements
the standard loss to explicitly regularize the model to learn a more diverse
and rich dictionary. The proposed strategy yields an efficient variant of BoF
and further boosts its performance, without any additional parameters.Comment: 6 pages, 2 Figure
Single-channel EEG classification of sleep stages based on REM microstructure
Rapid-eye movement (REM) sleep, or paradoxical sleep, accounts for 20โ25% of total night-time sleep in healthy adults and may be related, in pathological cases, to parasomnias. A large percentage of Parkinson's disease patients suffer from sleep disorders, including REM sleep behaviour disorder and hypokinesia; monitoring their sleep cycle and related activities would help to improve their quality of life. There is a need to accurately classify REM and the other stages of sleep in order to properly identify and monitor parasomnias. This study proposes a method for the identification of REM sleep from raw single-channel electroencephalogram data, employing novel features based on REM microstructures. Sleep stage classification was performed by means of random forest (RF) classifier, K-nearest neighbour (K-NN) classifier and random Under sampling boosted trees (RUSBoost); the classifiers were trained using a set of published and novel features. REM detection accuracy ranges from 89% to 92.7%, and the classifiers achieved a F-1 score (REM class) of about 0.83 (RF), 0.80 (K-NN), and 0.70 (RUSBoost). These methods provide encouraging outcomes in automatic sleep scoring and REM detection based on raw single-channel electroencephalogram, assessing the feasibility of a home sleep monitoring device with fewerย channels
ํฉ์ฑ๊ณฑ ์ปค๋ ์ ๊ทํ๋ฅผ ์ํ ๊ณ ๋ฅธ ๊ฐ๋๋ถ์ฐ๋ฐฉ๋ฒ
ํ์๋
ผ๋ฌธ(๋ฐ์ฌ) -- ์์ธ๋ํ๊ต๋ํ์ : ์์ฐ๊ณผํ๋ํ ์๋ฆฌ๊ณผํ๋ถ, 2022. 8. ๊ฐ๋ช
์ฃผ.In this thesis, we propose new convolutional kernel regularization methods. Along with the development of deep learning, there have been attempts to effectively regularize a convolutional layer, which is an important basic module of deep neural networks. Convolutional neural networks (CNN) are excellent at abstracting input data, but deepening causes gradient vanishing or explosion issues and produces redundant features. An approach to solve these issues is to directly regularize convolutional kernel weights of CNN. Its basic idea is to convert a convolutional kernel weight into a matrix and make the row or column vectors of the matrix orthogonal. However, this approach has some shortcomings. Firstly, it requires appropriate manipulation because overcomlete issue occurs when the number of vectors is larger than the dimension of vectors. As a method to deal with this issue, we define the concept of evenly dispersed state and propose PH0 and MST regularizations using this. Secondly, prior regularizations which enforce the Gram matrix of a matrix to be an identity matrix might not be an optimal approach for orthogonality of the matrix. We point out that these rather reduces the update of angles between some two vectors when two vectors are adjacent. Therefore, to complement for this issue, we propose EADK and EADC regularizations which update directly the angle. Through various experiments, we demonstrate that EADK and EADC regularizations outperform prior methods in some neural network architectures and, in particular, EADK has fast learning time.์ด ๋
ผ๋ฌธ์์๋ ํฉ์ฑ๊ณฑ์ปค๋์ ๋ํ ์๋ก์ด ์ ๊ทํ ๋ฐฉ๋ฒ๋ค์ ์ ์ํ๋ค. ๋ฅ๋ฌ๋์ ๋ฐ๋ฌ๊ณผ ๋๋ถ์ด ์ ๊ฒฝ๋ง์ ๊ฐ์ฅ ๊ธฐ๋ณธ์ ์ธ ๋ชจ๋์ธ ํฉ์ฑ๊ณฑ ๋ ์ด์ด๋ฅผ ํจ๊ณผ์ ์ผ๋ก ์ ๊ทํ ํ๋ ค๋ ์๋๋ค์ด ์์ด ์๋ค. ํฉ์ฑ๊ณฑ์ ๊ฒฝ๋ง๋ ์ธํ๋ฐ์ดํฐ๋ฅผ ์ถ์ํํ๋๋ฐ ํ์ํ์ง๋ง ๋คํธ์ํฌ์ ๊น์ด๊ฐ ๊น์ด์ง๋ฉด ๊ทธ๋ ๋์ธํธ ์๋ฉธ์ด๋ ํญ๋ฐ ๋ฌธ์ ๋ฅผ ์ผ์ผํค๊ณ ์ค๋ณต๋ ํผ์ณ๋ค์ ๋ง๋ ๋ค. ์ด๋ฌํ ๋ฌธ์ ๋ค์ ํด๊ฒฐํ๊ธฐ ์ํ ์ ๊ทผ๋ฒ ์ค ํ๋๋ ์ง์ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง์ ํฉ์ฑ๊ณฑ์ปค๋์ ์ง์ ์ ๊ทํ ํ๋ ๊ฒ์ด๋ค. ์ด ๋ฐฉ๋ฒ์ ํฉ์ฑ๊ณฑ์ปค๋์ ์ด๋ค ํ๋ ฌ๋ก ๋ณํํ๊ณ ํ๋ ฌ์ ํ ๋๋ ์ด๋ค์ ๋ฒกํฐ๋ค์ ์ง๊ต์ํค๋ ๊ฒ์ด๋ค. ๊ทธ๋ฌ๋ ์ด๋ฌํ ์ ๊ทผ๋ฒ์ ๋ช๊ฐ์ง ๋จ์ ์ด ์๋ค. ์ฒซ์งธ๋ก, ๋ฒกํฐ์ ์๊ฐ ๋ฒกํฐ์ ์ฐจ์๋ณด๋ค ๋ง์ ๋๋ ๋ชจ๋ ๋ฒกํฐ๋ฅผ ์ง๊ตํ ์ํฌ ์ ์๊ฒ ๋๋ฏ๋ก ์ ์ ํ ๊ธฐ๋ฒ๋ค์ ํ์๋ก ํ๋ค. ์ด ๋ฌธ์ ๋ฅผ ๋ค๋ฃจ๊ธฐ ์ํ ํ ๊ฐ์ง ๋ฐฉ๋ฒ์ผ๋ก ์ฐ๋ฆฌ๋ ๋ถ์ฐ ์ํ๋ผ๋ ๊ฐ๋
์ ์ ์ํ๊ณ ์ด ๊ฐ๋
์ ํ์ฉํ PH0์ MST ์ ๊ทํ๋ฒ์ ์ ์ํ๋ค. ๋์งธ๋ก, ๊ทธ๋ํ๋ ฌ์ ํญ๋ฑํ๋ ฌ๋ก ๊ทผ์ฌ์ํค๋ ๋ฐฉ๋ฒ์ ์ฌ์ฉํ๋ ๊ธฐ์กด ์ ๊ทํ๋ฒ์ด ๋ฒกํฐ๋ค์ ์ง๊ตํ์ํค๋ ์ต์ ์ ๋ฐฉ๋ฒ์ด ์๋ ์ ์๋ค๋ ์ ์ด๋ค. ์ฆ, ๊ธฐ์กด์ ์ ๊ทํ๋ฒ์ด ๋ ๋ฒกํฐ๊ฐ ๊ฐ๊น์ธ ๋๋ ์คํ๋ ค ๊ฐ๋์ ์
๋ฐ์ดํธ๋ฅผ ์ค์ด๊ฒ ๋๋ค.๋ฐ๋ผ์ ์ด๋ฅผ ๋ณด์ํ๊ธฐ ์ํ์ฌ ์ฐ๋ฆฌ๋ ๊ฐ๋๋ฅผ ์ง์ ์
๋ฐ์ดํธํ๋ EADK์ EADC ์ ๊ทํ๋ฒ์ ์ ์ํ๋ค. ๊ทธ๋ฆฌ๊ณ ๋ค์ํ ์คํ์ ํตํด EADK์ EADC ์ ๊ทํ๋ฒ์ด ๋ค์์ ์ ๊ฒฝ๋ง๊ตฌ์กฐ์์ ๊ธฐ์กด์ ๋ฐฉ๋ฒ๋ค๋ณด๋ค ์ฐ์ํ ์ฑ๋ฅ์ ๋ณด์ด๊ณ ํนํ EADK๋ ๋น ๋ฅธ ํ์ต์๊ฐ์ ๊ฐ์ง๋ค๋ ๊ฒ์ ํ์ธํ๋ค.Abstract i
1 Introduction 1
2 Preliminaries 4
2.1 Two Ways of Understanding CNN Layers as Matrix Operations 5
2.1.1 Kernel Matrix 6
2.1.2 Convolution Matrix 7
2.2 Soft Orthogonality 11
2.2.1 SO Regularization 11
2.2.2 DSO Regularization 12
2.3 Mutual Coherence 13
2.3.1 MC Regularization 13
2.4 Spectral Restricted Isometry Property 13
2.4.1 Restricted Isometry Property 13
2.4.2 SRIP Regularization 15
2.5 Orthogonal Convolutional Neural Networks 18
2.5.1 OCNN Regularizaiton 18
3 Topological Dispersing Regularizations 22
3.1 Evenly Dispersed State 23
3.1.1 Dispersing Vectors on Sphere 23
3.1.2 Evenly Dispersed State in the Real Projective Spaces 25
3.2 Persistent Homology Regularization 33
3.2.1 Cech and Vietoris-Rips Complexes 35
3.2.2 Persistent Homology 36
3.2.3 PH0 Regularization 38
3.3 Minimum Spanning Tree Regularization 39
3.3.1 Minimum Spanning Tree 39
3.3.2 MST Regularization 41
4 Evenly Angle Dispersing Regularizations 42
4.1 Analysis of Soft Orthogonality 43
4.1.1 Analysis of Soft Orthogonality 43
4.2 Evenly Angle Dispersing Regularizations 47
4.2.1 Evenly Angle Dispersing Regularization with Kernel Matrix 47
4.2.2 Evenly Angle Dispersing Regularization with Convolution Matrix 52
5 Algorithms & Experiments 54
5.1 Algorithms 55
5.1.1 PH0 and MST 55
5.1.2 EADK 57
5.1.3 EADC 58
5.2 Experiments 59
5.2.1 Analysis for Angle Dispersing 59
5.2.2 Experimental Setups 62
5.2.3 Classification Accuracy 68
5.2.4 Additional Experiments 76
6 Conclusion 78
The bibliography 80
Abstract (in Korean) 85๋ฐ
Pruning Ternary Quantization
Inference time, model size, and accuracy are three key factors in deep model
compression.
Most of the existing work addresses these three key factors separately as it
is difficult to optimize them all at the same time.
For example, low-bit quantization aims at obtaining a faster model; weight
sharing quantization aims at improving compression ratio and accuracy; and
mixed-precision quantization aims at balancing accuracy and inference time. To
simultaneously optimize bit-width, model size, and accuracy, we propose pruning
ternary quantization (PTQ): a simple, effective, symmetric ternary quantization
method. We integrate L2 normalization, pruning, and the weight decay term to
reduce the weight discrepancy in the gradient estimator during quantization,
thus producing highly compressed ternary weights. Our method brings the highest
test accuracy and the highest compression ratio. For example, it produces a
939kb (49) 2bit ternary ResNet-18 model with only 4\% accuracy drop on
the ImageNet dataset. It compresses 170MB Mask R-CNN to 5MB (34) with
only 2.8\% average precision drop. Our method is verified on image
classification, object detection/segmentation tasks with different network
structures such as ResNet-18, ResNet-50, and MobileNetV2