Journal of Agricultural Big Data ›› 2025, Vol. 7 ›› Issue (2): 144-154.doi: 10.19788/j.issn.2096-6369.000106

Previous Articles     Next Articles

A Review of the Evolution and Applications of AI Knowledge Distillation Technology

MAO KeBiao1,2(), DAI Wang2, GUO ZhongHua2, SUN XueHong2, XIAO LiuRui2   

  1. 1. School of Physics and Electronic-Electrical Engineering, Ningxia University, Yinchuan 750021, China
    2. State Key Laboratory of Efficient Utilization of Arid and Semi-arid Arable Land in Northern China, Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100081, China
  • Received:2025-03-28 Accepted:2025-05-06 Online:2025-06-26 Published:2025-06-23

Abstract:

Knowledge Distillation (KD) in Artificial Intelligence (AI) achieves model lightweighting through a teacher-student framework, emerging as a key technology to address the performance-efficiency bottleneck in deep learning. This paper systematically analyzes KD’s theoretical framework from the perspective of algorithm evolution, categorizing knowledge transfer paths into four paradigms: response-based, feature-based, relation-based, and structure-based. It establishes a comparative evaluation system for dynamic and static KD methods. We deeply explore innovative mechanisms such as cross-modal feature alignment, adaptive distillation architectures, and multi-teacher collaborative validation, while analyzing fusion strategies like progressive knowledge transfer and adversarial distillation. Through empirical analysis in computer vision and natural language processing, we assess KD’s practicality in scenarios like image classification, semantic segmentation, and text generation. Notably, we highlight KD’s potential in agriculture and geosciences, enabling efficient deployment in resource-constrained settings for precision agriculture and geospatial analysis. Current models often face issues like ambiguous knowledge selection mechanisms and insufficient theoretical interpretability. Accordingly, we discuss the feasibility of automated distillation systems and multimodal knowledge fusion, offering new technical pathways for edge intelligence deployment and privacy computing, particularly suited for agricultural intelligence and geoscience research.

Key words: knowledge distillation, model compression, knowledge transfer, dynamic optimization, multimodal learning