Section 01
Main Floor: The Consistency Dilemma of LLM Recommendation Explanations—Reliable Explainer or Unreliable Narrator?
Recent research systematically evaluates the explanation consistency and sensitivity of large language models (LLMs) in group recommendation tasks. It finds that different models show significant differences in generating recommendation reasons, with some models exhibiting the characteristics of an "unreliable narrator," sounding an alarm for the application of LLMs in high-risk recommendation scenarios (e.g., healthcare, finance). The study focuses on explanation consistency (whether explanations remain consistent under the same recommendation decision) and sensitivity (whether explanation adjustments are reasonable when inputs change slightly). Through multi-model comparison experiments in group recommendation scenarios, key conclusions are drawn, emphasizing the need to prioritize the stability and credibility of explanations.