Section 01
UAE Framework: Distilling LLM Utility into Dense Retrievers for Dual Breakthroughs in Accuracy and Efficiency
Researchers propose the Utility-Aligned Embeddings (UAE) framework, which distills the perplexity reduction signal of Large Language Models (LLMs) into a dual-encoder embedding space to address the disconnect between semantic similarity and generation utility in dense retrievers within RAG systems. This framework achieves over 30% improvement in retrieval performance on the QASPER benchmark while being 180x faster than LLM re-ranking methods, balancing high accuracy and efficiency.