Section 01
BitCal-TTS: Guide to Confidence Calibration Solutions for Quantized Reasoning Models
BitCal-TTS addresses the premature termination problem caused by inaccurate confidence calibration when quantized reasoning models run at 4-bit precision. Through mechanisms like bit-conditional recalibration and inference stability proxy, it achieves an accuracy improvement of 3.7% for the 7B model and 2.8% for the 14B model on GSM8K, while reducing the premature termination rate.