Section 01
CadLLM: An Innovative Method to Improve Inference Throughput of Diffusion Language Models Without Training
CadLLM is the open-source implementation of an ACL 2026 Findings paper, which proposes a plug-and-play controller that dynamically adjusts decoding strategies using the model's own lightweight confidence signals. This method achieves up to 2.28x throughput improvement on GSM8K, MATH, MBPP, and HumanEval benchmarks while maintaining competitive accuracy. It is training-free and compatible with existing diffusion language models (e.g., LLaDA, DREAM).