26 research outputs found

    ํšจ์œจ์ ์ธ ์‹ฌ์ธต ์‹ ๊ฒฝ๋ง์„ ์œ„ํ•œ ์–‘์žํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋ฐ ๋ฐฉ๋ฒ•๋ก 

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€,2020. 2. ์œ ์Šน์ฃผ.Deep neural networks (DNN) are becoming increasingly popular and widely adopted for various applications. Energy efficiency of neural networks is critically important for both edge devices and servers. It is imperative to optimize neural networks in terms of both speed and energy consumption while maintaining the accuracy of the network. Quantization is one of the most effective optimization techniques. By reducing the bit-width of activations and weights, both the speed and energy can be improved by executing more computations using the same amount of memory access and computational resources (e.g. silicon chip area and battery). It is expected that computations with 4-bit and lower precision will contribute to the energy efficient and real-time characteristics of future deep learning applications. One major drawback of quantization is the drop in accuracy, resulting from the reduction in the degree of freedom of data representation. Recently, there have been several studies that demonstrated that the inference of DNNs can be accurately done by using 8-bit precision. However, many studies show that the network quantized into 4-bit or less precision suffers from significant quality degradation. Especially, the state-of-the art networks cannot be quantized easily due to their optimized structure. In this dissertation, several methods are proposed that use different approaches to minimize the reduction in the accuracy of the quantized DNNs. Weighted- entropy-based quantization is designed to fully utilize the limited number of quantization levels by maximizing the weighted information of the quantized data. This work shows the potential of multi-bit quantization for both activation and weight. Value-aware quantization, or outlier-aware quantization is designed to support sub-4-bit quantization, while allowing a small amount (1 ~ 3 %) of large values in high precision. This helps the quantized data to maintain the statistics, e.g. mean and variance corresponding to the full-precision, thus minimizing the accuracy drop after quantization. The dedicated hardware accelerator, called OLAccel, is also proposed to maximize the performance of the network quantized by the outlier-aware quantization. The hardware takes advantage of the benefit of reduced precision, i.e. 4-bit, with minimal accuracy drop by the proposed quantization algorithm. Precision-highway is the structural concept that forms an end-to-end high-precision information flow while performing ultra-low-precision computations. This minimizes the accumulated quantization error, which helps to improve the accuracy of the network even with extremely low precision. BLast, the training methodology, and differentiable and unified quantization (DuQ), a novel quantization algorithm, are designed to support sub-4-bit quantization for the optimized mobile networks, i.e. MobileNet-v3. These methods allow the MobileNet-v3 network to be quantized into 4-bit for both activation and weight with negligible accuracy loss.๋”ฅ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ (DNN)๋Š” ํ™œ์šฉ ๋ฒ”์œ„๋ฅผ ์ ์ฐจ ๋„“ํ˜€๊ฐ€๋ฉฐ ๋‹ค์–‘ํ•œ ๋ถ„์•ผ์— ์ ์šฉ๋˜๊ณ  ์žˆ๋‹ค. ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋Š” ์„œ๋ฒ„ ๋ฟ๋งŒ ์•„๋‹ˆ๋ผ ์ž„๋ฒ ๋””๋“œ ๊ธฐ๊ธฐ์—์„œ๋„ ๋„๋ฆฌ ํ™œ์šฉ๋˜๊ณ  ์žˆ์œผ๋ฉฐ ์ด๋กœ์ธํ•ด ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ํšจ์œจ์„ฑ์„ ๋†’์ด๋Š” ๊ฒƒ์€ ์ ์  ๋” ์ค‘์š”ํ•ด์ง€๋Š” ์ค‘์ด๋‹ค. ์ด์ œ ์ •ํ™•๋„๋ฅผ ์œ ์ง€ํ•˜๋ฉด์„œ ์†๋„๋ฅผ ๋น ๋ฅด๊ฒŒ ํ•˜๊ณ  ์—๋„ˆ์ง€ ์†Œ๋ชจ๋ฅผ ์ค„์ด๋Š” ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ์ตœ์ ํ™”๋Š” ํ•„์ˆ˜์  ์š”์†Œ๋กœ ์ž๋ฆฌ์žก์•˜๋‹ค. ์–‘์žํ™”๋Š” ๊ฐ€์žฅ ํšจ๊ณผ์ ์ธ ์ตœ์ ํ™” ๊ธฐ๋ฒ• ์ค‘ ํ•˜๋‚˜์ด๋‹ค. ๋‰ด๋Ÿฐ์˜ ํ™œ์„ฑ๋„ (activation) ๋ฐ ํ•™์Šต ๊ฐ€์ค‘์น˜ (weight)๋ฅผ ์ €์žฅํ•˜๋Š”๋ฐ ํ•„์š”ํ•œ ๋น„ํŠธ ์ˆ˜๋ฅผ ์ค„์ž„์œผ๋กœ์จ ๋™์ผํ•œ ์–‘์˜ ๋ฐ์ดํ„ฐ ์ ‘๊ทผ๊ณผ ์—ฐ์‚ฐ ๋น„์šฉ (์นฉ ๋ฉด์  ๋ฐ ์—๋„ˆ์ง€ ์†Œ๋ชจ ๋“ฑ)์œผ๋กœ ๋” ๋งŽ์€ ์—ฐ์‚ฐ์ด ๊ฐ€๋Šฅํ•ด์ง€๋ฉฐ ์ด๋กœ์ธํ•ด ์†๋„์™€ ์—๋„ˆ์ง€ ์†Œ๋ชจ๋ฅผ ๋™์‹œ์— ์ตœ์ ํ™”ํ•  ์ˆ˜ ์žˆ๋‹ค. ์ถ”ํ›„ ๋”ฅ ๋Ÿฌ๋‹์„ ํ™œ์šฉํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ํ•„์š”ํ•  ๊ฒƒ์œผ๋กœ ์˜ˆ์ธก๋˜๋Š” ์—๋„ˆ์ง€ ํšจ์œจ ๋ฐ ์—ฐ์‚ฐ ์†๋„๋ฅผ ๋งŒ์กฑ์‹œํ‚ค๊ธฐ ์œ„ํ•ด์„œ 4 ๋น„ํŠธ ํ˜น์€ ๋” ์ ์€ ์ •๋ฐ€๋„ ๊ธฐ๋ฐ˜์˜ ์–‘์žํ™” ์—ฐ์‚ฐ์ด ์ง€๋Œ€ํ•œ ๊ณตํ—Œ์„ ํ•  ๊ฒƒ์œผ๋กœ ๊ธฐ๋Œ€๋œ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ์–‘์žํ™”์˜ ๊ฐ€์žฅ ์ค‘์š”ํ•œ ๋‹จ์  ์ค‘ ํ•˜๋‚˜๋Š” ๋ฐ์ดํ„ฐ์˜ ํ‘œํ˜„ํ˜•์„ ์ œํ•œํ•˜์—ฌ ์ž์œ ๋„๊ฐ€ ๋–จ์–ด์ง€๊ฒŒ ๋จ์œผ๋กœ์„œ ๋ฐœ์ƒํ•˜๋Š” ์ •ํ™•๋„์˜ ์†์‹ค์ด๋‹ค. ์ด๋Ÿฌํ•œ ๋‹จ์ ์„ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ๋‹ค์–‘ํ•œ ์—ฐ๊ตฌ๋“ค์ด ์ง„ํ–‰์ค‘์ด๋‹ค. ์ตœ๊ทผ ์ผ๋ถ€ ์—ฐ๊ตฌ๋“ค์€ 8 ๋น„ํŠธ์˜ ์ •๋ฐ€๋„์—์„œ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋ฅผ ํ™œ์šฉํ•ด ๊ฒฐ๊ณผ๋ฅผ ์ถ”๋ก  (inference)ํ•˜๋Š”๋ฐ ์ •ํ™•๋„ ์†์‹ค์ด ๊ฑฐ์˜ ์—†์Œ์„ ๋ณด๊ณ ํ•˜๊ณ  ์žˆ๋‹ค. ๋ฐ˜๋ฉด ๊ทธ ์™ธ์˜ ๋‹ค์–‘ํ•œ ์—ฐ๊ตฌ๋“ค์„ ํ†ตํ•ด 4 ๋น„ํŠธ ํ˜น์€ ๋” ๋‚ฎ์€ ์ •๋ฐ€๋„์—์„œ ์–‘์žํ™”๋ฅผ ์ ์šฉํ–ˆ์„ ๋•Œ ๋งŽ์€ ๋„คํŠธ์›Œํฌ๋“ค์˜ ์ •ํ™•๋„๊ฐ€ ํฌ๊ฒŒ ์†์ƒ๋˜๋Š” ํ˜„์ƒ๋„ ํ•จ๊ป˜ ๋ณด๊ณ ๋˜๊ณ  ์žˆ๋‹ค. ํŠนํžˆ ์ตœ๊ทผ ์ œ์•ˆ๋œ ๋„คํŠธ์›Œํฌ๋“ค์˜ ๊ฒฝ์šฐ ์„ฑ๋Šฅ ํ–ฅ์ƒ์„ ์œ„ํ•ด ๋„์ž…ํ•œ ์ตœ์ ํ™”๋œ ๊ตฌ์กฐ๊ฐ€ ์–‘์žํ™” ํ•˜๊ธฐ ์–ด๋ ค์šด ํŠน์„ฑ์„ ๊ฐ€์ ธ ์ด๋Ÿฌํ•œ ํ˜„์ƒ์ด ์‹ฌํ™”๋œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์–‘์žํ™”๋œ DNN์˜ ์ •ํ™•๋„ ์†์‹ค์„ ์ตœ์†Œํ™”ํ•˜๊ธฐ์œ„ํ•œ ๋‹ค์–‘ํ•œ ๋ฐฉ๋ฒ•๋“ค์„ ์ œ์•ˆํ•˜์˜€๋‹ค. ๊ฐ€์ค‘ ์—”ํŠธ๋กœํ”ผ ๊ธฐ๋ฐ˜ ์–‘์žํ™” (Weighted-entropy-based quantization)์€ ์ œํ•œ๋œ ๊ฐœ์ˆ˜์˜ ์–‘์žํ™” ๋ ˆ๋ฒจ์„ ์ตœ๋Œ€ํ•œ ํ™œ์šฉํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ์–‘์žํ™”๋œ ๋ฐ์ดํ„ฐ์˜ ์ •๋ณด๋Ÿ‰์„ ์ตœ๋Œ€ํ™”ํ•˜๋Š” ๋ฐฉํ–ฅ์œผ๋กœ ์–‘์žํ™”๋ฅผ ์ง„ํ–‰ํ•˜๋„๋ก ์„ค๊ณ„๋˜์—ˆ๋‹ค. ์ด ์—ฐ๊ตฌ๋ฅผ ํ†ตํ•ด ์•„์ฃผ ๊นŠ์€ ๋„คํŠธ์›Œํฌ์—์„œ๋„ ๋‰ด๋Ÿฐ์˜ ํ™œ์„ฑ๋„์™€ ํ•™์Šต ๊ฐ€์ค‘์น˜ ๋ชจ๋‘์˜ ์–‘์žํ™”๊ฐ€ ์ ์šฉ ๊ฐ€๋Šฅํ•จ์„ ๋ณด์˜€๋‹ค. ๊ฐ’-์˜์‹ ์–‘์žํ™” (value-aware quantization), ํ˜น์€ ์˜ˆ์™ธ-์˜์‹ ์–‘์žํ™” (outlier-aware quantization)๋Š” ๋นˆ๋„๋Š” ๋‚ฎ์ง€๋งŒ ํฐ ๊ฐ’์„ ๊ฐ€์ง€๋Š” ๋ฐ์ดํ„ฐ๋ฅผ ํฐ ์ •๋ฐ€๋„๋กœ ์ €์žฅํ•˜๋Š” ๋Œ€์‹  ๋‚˜๋จธ์ง€ ๋ฐ์ดํ„ฐ์— 4 ๋น„ํŠธ ์ดํ•˜์˜ ์–‘์žํ™”๋ฅผ ์ ์šฉํ•˜๋„๋ก ์„ค๊ณ„๋œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด๋‹ค. ์ด๋Š” ์›๋ณธ ๋ฐ์ดํ„ฐ์˜ ํ‰๊ท ๊ณผ ๋ถ„์‚ฐ ๊ฐ™์€ ํŠน์„ฑ์ด ์–‘์žํ™”๋œ ํ›„์—๋„ ์œ ์ง€ํ•˜๋„๋ก ๋„์™€์ฃผ์–ด ์–‘์žํ™”๋œ ๋„คํŠธ์›Œํฌ์˜ ์ •ํ™•๋„๋ฅผ ์œ ์ง€ํ•˜๋Š”๋ฐ ๊ธฐ์—ฌํ•œ๋‹ค. ์ด์— ๋”ํ•˜์—ฌ OLAccel์ด๋ผ ๋ช…๋ช…๋œ ํŠนํ™” ๊ฐ€์†๊ธฐ๋ฅผ ์ œ์•ˆํ•˜์˜€๋‹ค. ์ด ๊ฐ€์†๊ธฐ๋Š” ๊ฐ’-์˜์‹ ์–‘์žํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ํ†ตํ•ด ์–‘์žํ™”๋œ ๋„คํŠธ์›Œํฌ๋ฅผ ๊ฐ€์†ํ•จ์œผ๋กœ์จ ์ •ํ™•๋„ ๊ฐ์†Œ๋Š” ์ตœ์†Œํ™” ํ•˜๋ฉด์„œ ๋‚ฎ์€ ์ •๋ฐ€๋„์˜ ์„ฑ๋Šฅ ์ด๋“์„ ์ตœ๋Œ€ํ™”ํ•œ๋‹ค. ๊ณ ์ •๋ฐ€๋„-ํ†ต๋กœ ๊ตฌ์กฐ (precision-highway)๋Š” ๋„คํŠธ์›Œํฌ์˜ ๊ตฌ์กฐ๋ฅผ ๊ฐœ์„ ํ•˜์—ฌ ์ดˆ์ €์ •๋ฐ€๋„ ์—ฐ์‚ฐ์„ ์ˆ˜ํ–‰ํ•˜๋ฉด์„œ๋„ ๊ณ ์ •๋ฐ€๋„ ์ •๋ณด ํ†ต๋กœ๋ฅผ ์ƒ์„ฑํ•œ๋‹ค. ์ด๋Š” ์–‘์žํ™”๋กœ ์ธํ•˜์—ฌ ์—๋Ÿฌ๊ฐ€ ๋ˆ„์ ๋˜๋Š” ํ˜„์ƒ์„ ์™„ํ™”ํ•˜์—ฌ ๋งค์šฐ ๋‚ฎ์€ ์ •๋ฐ€๋„์—์„œ ์ •ํ™•๋„๋ฅผ ๊ฐœ์„ ํ•˜๋Š”๋ฐ ๊ธฐ์—ฌํ•œ๋‹ค. ํ•™์Šต ๊ธฐ๋ฒ•์ธ BLast์™€ ๋ฏธ๋ถ„ ๊ฐ€๋Šฅํ•˜๊ณ  ํ†ตํ•ฉ๋œ ์–‘์žํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜ (DuQ)๋Š” MobileNet-v3๊ณผ ๊ฐ™์€ ์ตœ์ ํ™”๋œ ๋ชจ๋ฐ”์ผํ–ฅ ๋„คํŠธ์›Œํฌ๋ฅผ ์ตœ์ ํ™”ํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ์ œ์•ˆ๋˜์—ˆ๋‹ค. ์ด ๋ฐฉ๋ฒ•๋“ค์„ ํ†ตํ•ด ๋ฏธ๋ฏธํ•œ ์ •ํ™•๋„ ์†์‹ค๋งŒ์œผ๋กœ MobileNet-v3์˜ ํ™œ์„ฑ๋„ ๋ฐ ํ•™์Šต ๊ฐ€์ค‘์น˜ ๋ชจ๋‘๋ฅผ 4 ๋น„ํŠธ ์ •๋ฐ€๋„๋กœ ์–‘์žํ™”ํ•˜๋Š”๋ฐ ์„ฑ๊ณตํ•˜์˜€๋‹ค.Chapter 1. Introduction 1 Chapter 2. Background and RelatedWork 4 Chapter 3. Weighted-entropy-based Quantization 15 3.1 Introduction 15 3.2 Motivation 17 3.3 Quantization based on Weighted Entropy 20 3.3.1 Weight Quantization 20 3.3.2 Activation Quantization 24 3.3.3 IntegratingWeight/Activation Quantization into the Training Algorithm 27 3.4 Experiment 28 3.4.1 Image Classification: AlexNet, GoogLeNet and ResNet-50/101 28 3.4.2 Object Detection: R-FCN with ResNet-50 35 3.4.3 Language Modeling: An LSTM 37 3.5 Conclusion 38 Chapter 4. Value-aware Quantization for Training and Inference of Neural Networks 40 4.1 Introduction 40 4.2 Motivation 41 4.3 Proposed Method 43 4.3.1 Quantized Back-Propagation 44 4.3.2 Back-Propagation of Full-Precision Loss 46 4.3.3 Potential of Further Reduction in Computation Cost 47 4.3.4 Local Sorting in Data Parallel Training 48 4.3.5 ReLU and Value-aware Quantization (RV-Quant) 49 4.3.6 Activation Annealing 50 4.3.7 Quantized Inference 50 4.4 Experiments 51 4.4.1 Training Results 52 4.4.2 Inference Results 59 4.4.3 LSTM Language Model 61 4.5 Conclusions 62 Chapter 5. Energy-efficient Neural Network Accelerator Based on Outlier-aware Low-precision Computation 63 5.1 Introduction 63 5.2 Proposed Architecture 65 5.2.1 Overall Structure 65 5.2.2 Dataflow 68 5.2.3 PE Cluster 72 5.2.4 Normal PE Group 72 5.2.5 Outlier PE Group and Cluster Output Tri-buffer 75 5.3 Evaluation Methodology 78 5.4 Experimental Results 80 5.5 Conclusion 90 Chapter 6. Precision Highway for Ultra Low-Precision Quantization 92 6.1 Introduction 92 6.2 Proposed Method 93 6.2.1 Precision Highway on Residual Network 94 6.2.2 Precision Highway on Recurrent Neural Network 96 6.2.3 Practical Issues with Precision Highway 98 6.3 Training 99 6.3.1 LinearWeight Quantization based on Laplace Distribution Model 99 6.3.2 Fine-tuning for Weight/Activation Quantization 100 6.4 Experiments 101 6.4.1 Experimental Setup 101 6.4.2 Analysis of Accumulated Quantization Error 101 6.4.3 Loss Surface Analysis of Quantized Model Training 103 6.4.4 Evaluating the Accuracy of Quantized Model 103 6.4.5 Hardware Cost Evaluation of Quantized Model 108 6.5 Conclusion 109 Chapter 7. Towards Sub-4-bit Quantization of Optimized Mobile Netowrks 114 7.1 Introduction 114 7.2 BLast Training 117 7.2.1 Notation 118 7.2.2 Observation 118 7.2.3 Activation Instability Metric 120 7.2.4 BLast Training 122 7.3 Differentiable and Unified Quantization 124 7.3.1 Rounding and Truncation Errors 124 7.3.2 Limitations of State-of-the-Art Methods 124 7.3.3 Proposed Method: DuQ 126 7.3.4 Handling Negative Values 128 7.4 Experiments 131 7.4.1 Accuracy on ImageNet Dataset 131 7.4.2 Discussion on Fused-BatchNorm 133 7.4.3 Ablation Study 134 7.5 Conclusion 137 Chapter 8 Conclusion 138 Bibliography 141 ๊ตญ๋ฌธ์ดˆ๋ก 154 Acknowledgements 157Docto

    INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold

    Full text link
    Binary Neural Networks (BNNs) have emerged as a promising solution for reducing the memory footprint and compute costs of deep neural networks. BNNs, on the other hand, suffer from information loss because binary activations are limited to only two values, resulting in reduced accuracy. To improve the accuracy, previous studies have attempted to control the distribution of binary activation by manually shifting the threshold of the activation function or making the shift amount trainable. During the process, they usually depended on statistical information computed from a batch. We argue that using statistical data from a batch fails to capture the crucial information for each input instance in BNN computations, and the differences between statistical information computed from each instance need to be considered when determining the binary activation threshold of each instance. Based on the concept, we propose the Binary Neural Network with INSTAnce-aware threshold (INSTA-BNN), which decides the activation threshold value considering the difference between statistical data computed from a batch and each instance. The proposed INSTA-BNN outperforms the baseline by 2.5% and 2.3% on the ImageNet classification task with comparable computing cost, achieving 68.0% and 71.7% top-1 accuracy on ResNet-18 and MobileNetV1 based models, respectively.Comment: 19 pages, 7 figures; excluded axessibility packag

    OWQ: Lessons learned from activation outliers for weight quantization in large language models

    Full text link
    Large language models (LLMs) with hundreds of billions of parameters show impressive results across various language tasks using simple prompt tuning and few-shot examples, without the need for task-specific fine-tuning. However, their enormous size requires multiple server-grade GPUs even for inference, creating a significant cost barrier. To address this limitation, we introduce a novel post-training quantization method for weights with minimal quality degradation. While activation outliers are known to be problematic in activation quantization, our theoretical analysis suggests that we can identify factors contributing to weight quantization errors by considering activation outliers. We propose an innovative PTQ scheme called outlier-aware weight quantization (OWQ), which identifies vulnerable weights and allocates high-precision to them. Our extensive experiments demonstrate that the 3.01-bit models produced by OWQ exhibit comparable quality to the 4-bit models generated by OPTQ

    Temporal Dynamic Quantization for Diffusion Models

    Full text link
    The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques struggle to maintain performance even in 8-bit precision due to the diffusion model's unique property of temporal variation in activation. We introduce a novel quantization method that dynamically adjusts the quantization interval based on time step information, significantly improving output quality. Unlike conventional dynamic quantization techniques, our approach has no computational overhead during inference and is compatible with both post-training quantization (PTQ) and quantization-aware training (QAT). Our extensive experiments demonstrate substantial improvements in output quality with the quantized diffusion model across various datasets

    Symmetry Regularization andย Saturating Nonlinearity forย Robust Quantization

    No full text
    Robust quantization improves the tolerance of networks for various implementations, allowing reliable output in different bit-widths or fragmented low-precision arithmetic. In this work, we perform extensive analyses to identify the sources of quantization error and present three insights to robustify a network against quantization: reduction of error propagation, range clamping for error minimization, and inherited robustness against quantization. Based on these insights, we propose two novel methods called symmetry regularization (SymReg) and saturating nonlinearity (SatNL). Applying the proposed methods during training can enhance the robustness of arbitrary neural networks against quantization on existing post-training quantization (PTQ) and quantization-aware training (QAT) algorithms and enables us to obtain a single weight flexible enough to maintain the output quality under various conditions. We conduct extensive studies on CIFAR and ImageNet datasets and validate the effectiveness of the proposed methods.1

    Tag2Pix: Line Art Colorization Using Text Tag With SECat and Changing Loss

    Full text link
    Line art colorization is expensive and challenging to automate. A GAN approach is proposed, called Tag2Pix, of line art colorization which takes as input a grayscale line art and color tag information and produces a quality colored image. First, we present the Tag2Pix line art colorization dataset. A generator network is proposed which consists of convolutional layers to transform the input line art, a pre-trained semantic extraction network, and an encoder for input color information. The discriminator is based on an auxiliary classifier GAN to classify the tag information as well as genuineness. In addition, we propose a novel network structure called SECat, which makes the generator properly colorize even small features such as eyes, and also suggest a novel two-step training method where the generator and discriminator first learn the notion of object and shape and then, based on the learned notion, learn colorization, such as where and how to place which color. We present both quantitative and qualitative evaluations which prove the effectiveness of the proposed method.Comment: Accepted to ICCV 201
    corecore