Section 01
LED: A New Latent Space Decoding Method to Restore Exploration Capability of Reasoning Models (Introduction)
This article introduces Latent Exploration Decoding (LED), an innovative method aimed at solving the problem of excessive conservatism in large reasoning models after post-training. By introducing exploratory noise into the model's latent representation space, LED restores the model's exploration capability while maintaining reasoning quality. The related research has been accepted by ICML 2026. Keywords: Reasoning models, latent space decoding, exploration capability, post-training optimization, Transformer, ICML 2026.