Section 01
[Overview] Knowledge Distillation Energy Efficiency Evaluation Framework: Slimming Large Models While Saving Power
This article introduces the open-source project Slimming-Models-Saving-Watts, a knowledge distillation research framework for HPC cluster environments. It supports three mainstream distillation paradigms and integrates GPU/CPU energy consumption telemetry, providing a quantitative evaluation tool for energy efficiency optimization of large language models. Its goal is to resolve the conflict between model scale and computational resource consumption.