Section 01
[Introduction] LSE-MTP: Addressing MTP Structural Hallucination to Build Consistent World Models
The consistency of internal world models in Large Language Models (LLMs) is a core debate in the AI field. Traditional Multi-Token Prediction (MTP) can learn structured representations, but it has a structural hallucination problem (discrete token supervision leads to shortcuts in the latent space, violating environmental constraints). This study proposes the Latent Semantic Enhancement Multi-Token Prediction (LSE-MTP) method, which bridges the gap between discrete tokens and continuous state representations by anchoring to real hidden state trajectories, effectively resolving structural hallucinations and improving the consistency and robustness of world models.