篇名 |
Combined Knowledge Distillation Framework: Breaking Down Knowledge Barriers
|
---|---|
並列篇名 | Combined Knowledge Distillation Framework: Breaking Down Knowledge Barriers |
作者 | Shuiping Ni、Wendi Wang、Mingfu Zhu、Xinliang Ma、Yizhe Zhang |
英文摘要 | Knowledge distillation, one of the most prominent methods in model compression, has successfully balanced small model sizes and high performance. However, it has been observed that knowledge distillation predominantly focuses on acquiring knowledge concealed within the dataset and the external knowledge imparted by the teacher. In contrast, self-distillation concerns itself with the utilization of internal network knowledge. Neither approach fails to fully harness the potential of knowledge. Therefore, this paper introduces the combined knowledge c framework that combines knowledge distillation with self-distillation. Within this framework, we introduce multiple shallow classifiers, combined with an attention module, to exploit internal and external knowledge. To enhance the efficiency with which the network utilizes knowledge. Experimental results demonstrate that by comprehensively leveraging network knowledge, distillation effectiveness can be enhanced, resulting in further improvements in network accuracy. Additionally, we applied the framework to lightweight neural networks with group convolution, the framework continues to perform exceptionally well.
|
起訖頁 | 109-122 |
關鍵詞 | model compression、knowledge distillation、self-distillation、lightweight、neural network |
刊名 | 電腦學刊 |
期數 | 202408 (35:4期) |
DOI |
|
QR Code | |
該期刊 上一篇
| Attention Mechanism Based Spatial-Temporal Graph Convolution Network for Traffic Prediction |
該期刊 下一篇
| Based on Three-Stage Data Envelopment Analysis of The Efficiency Evaluation Model of Specialty Construction |