Section 01
[Introduction] Slimming Models, Saving Watts: An Energy-Aware Knowledge Distillation Framework for Large Language Models
This research framework targets large language models such as Llama 3.1, systematically evaluating the accuracy, efficiency, and energy consumption performance of three knowledge distillation methods: responsive, feature-based, and relational. It is specifically designed for HPC clusters and Slurm environments. The framework fills the gap in traditional knowledge distillation research regarding the systematic evaluation of energy efficiency, deeply integrating energy consumption measurement with KD effect assessment, and providing a standardized tool for green AI research.