Volume- 12
Issue- 3
Year- 2024
DOI: 10.55524/ijircst.2024.12.3.17 | DOI URL: https://doi.org/10.55524/ijircst.2024.12.3.17 Crossref
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0) (http://creativecommons.org/licenses/by/4.0)
Article Tools: Print the Abstract | Indexing metadata | How to cite item | Email this article | Post a Comment
Elly Yijun Zhu , Chao Zhao, Haoyu Yang, Jing Li, Yue Wu, Rui Ding
Knowledge distillation is a model compression technique that enhances the performance and efficiency of a smaller model (student model) by transferring knowledge from a larger model (teacher model). This technique utilizes the outputs of the teacher model, such as soft labels, intermediate features, or attention weights, as additional supervisory signals to guide the learning process of the student model. By doing so, knowledge distillation reduces computational resources and storage space requirements while maintaining or surpassing the accuracy of the teacher model. Research on knowledge distillation has evolved significantly since its inception in the 1980s, especially with the introduction of soft labels by Hinton and colleagues in 2015. Various advancements have been made, including methods to extract richer knowledge, knowledge sharing among models, integration with other compression techniques, and application in diverse domains like natural language processing and reinforcement learning. This article provides a comprehensive review of knowledge distillation, covering its concepts, methods, applications, challenges, and future directions.
San Francisco Bay University, USA
No. of Downloads: 39 | No. of Views: 891
Xinghui Fei, Yulu Wang, Lu Dai, Mingxiu Sui .
May 2024 - Vol 12, Issue 3
Qishi Zhan, Yuhan Ma, Erdi Gao, Dan Sun, Haowei Yang.
May 2024 - Vol 12, Issue 3
Shweta Sinha.
March 2024 - Vol 12, Issue 2