Section 01
ESSAM Method Guide: Zeroth-Order Fine-Tuning Fusing ES and SAM to Enhance Large Model Mathematical Reasoning
This article introduces ESSAM, a zeroth-order fine-tuning method combining Evolution Strategy (ES) and Sharpness-Aware Maximization (SAM), specifically designed to enhance the mathematical reasoning ability of large language models. Traditional backpropagation-based fine-tuning has high computational pressure; while zeroth-order optimization is memory-efficient, it underperforms on complex tasks. By fusing the two technologies, ESSAM maintains the memory advantages of zeroth-order methods while improving optimization effectiveness.