검색 상세

Quantized Gradual Aware Training for Image Classification

초록/요약

With the rapid development of articial intelligence, the size and computational complexity of deep learning models have increased sig- nicantly, necessitating various techniques to efciently manage these models. In particular, quantization techniques play a crucial role in re- ducing memory usage and improving computational speed by convert- ing model weights and activation values to low-bit precision. However, existing Quantization-Aware Training (QAT) methods face limited prac- ticality due to their complex training procedures and additional compu- tational costs. In this paper, we propose a simple yet effective Quan- tized Gradual Aware Training (QGAT) technique to address these limi- tations. QGAT gradually quantizes the channels of a CNN to minimize information loss during the quantization process. This approach aims to maximize the benets of quantization while mitigating the performance degradation of the model. In our experiments, we trained ResNet-20 and ResNet-32 models on the CIFAR-10 and CIFAR-100 datasets, demon- strating that QGAT can enhance the model’s efciency and lightweight characteristics while maintaining a comparable level of accuracy to tra- ditional models. These ndings suggest that the proposed technique can facilitate the deployment of deep learning models in resource-constrained environments. This study highlights the practicality and effectiveness of QGAT, and we plan to extend its application to various models and datasets in future work to verify its generalizability. Keywords: Quantization, Classication, Deep neural network, transfer learning, Knowledge Distillation

more

목차

1. Introduction 1
2. Related Work 6
2.1 Quantization-Aware Training 6
2.2 Fisher Information Matrix 7
2.3 Straight-Through Estimator (STE) 8
2.4 Quantized Knowledge Distillation 9
3. Proposed Method 12
3.1 Distillation from High-Bit to Low-Bit Models 13
3.2 Quantization-Gradual Aware Training (QGAT) 15
4. Experimental Results and Discussions 19
4.1 Dataset and Evaluation Protocol 19
4.2 Implemental Details 20
4.3 Performance Comparison Table 21
4.4 Ablation Experiments 23
5. Conclusion 25
Bibliography 26

more