Section 01
MuCo: Introduction to NAVER AI Lab's Multi-turn Contrastive Learning Multimodal Embedding Model
The MuCo (Multi-turn Contrastive Learning) multimodal embedding model proposed by NAVER AI Lab has been accepted by CVPR 2026. Trained via multi-turn conversational contrastive learning, it achieves SOTA performance on the MMEB benchmark (70.1 points for the 2B model and 74.2 points for the 7B model). The related pre-trained models, M3T dataset, and paper have been open-sourced, and the complete training code will be released soon. This model provides a new paradigm for multimodal embedding training.