Section 01
【Introduction】Model Compression and Reasoning Consistency: Do Distilled Models Truly "Reason Correctly"?
This study focuses on the reasoning consistency issue of compressed models after knowledge distillation, exploring whether compressed models truly understand the essence of problems when giving correct answers, rather than just mimicking surface patterns. Through multi-dimensional evaluation methods such as GradCAM (attention visualization), CKA (representational similarity analysis), and calibration analysis, it finds that test set accuracy cannot guarantee reasoning consistency, different distillation strategies have a significant impact on consistency, and puts forward practical suggestions for improving the distillation process and model selection, emphasizing the need to ensure model reasoning quality while improving efficiency.