Section 01
Accumulative Decoding: An Innovative Decoding Method to Reduce Hallucinations in Vision-Language Models Without Training (Introduction)
Accumulative Decoding is a training-free decoding technique for large vision-language models (LVLMs). Its core advantage is that it requires no additional training or data—only by improving the decoding process during inference and accumulating multiple sampling results can it reduce model hallucinations and improve output accuracy. This method addresses the problem of LVLMs generating non-existent content or misinterpreting images in image understanding tasks, and is applicable to scenarios such as image question answering and visual reasoning.