Section 01
[Introduction] The Impact of Politeness on LLM Response Quality: Core Overview of Cross-Lingual Multi-Model Research
The title of this paper is The Impact of Politeness on LLMs: A Cross-Lingual, Multi-Model Study Using the PLUM Corpus, which focuses on whether polite language affects the response quality of large language models (LLMs). The study covers 3 languages (English, Hindi, Spanish), 5 models (Gemini-Pro, GPT-4o Mini, Claude 3.7 Sonnet, DeepSeek-Chat, Llama 3), and 22,500 prompt-response pairs. Key findings: Polite prompts improve response quality by an average of about 11%, but the effect varies across languages (e.g., Hindi prefers respectful and indirect expressions, while Spanish favors firm and confident ones) and models (e.g., Llama 3 is most sensitive to tone, GPT-4o Mini is more robust). Additionally, the tone of dialogue history affects the quality of current responses. This study aims to reveal the rules of interaction between politeness and LLMs, providing references for users' communication strategies and developers' model optimization.