Section 01
【Introduction】Core Introduction to the LLM Resilience Evaluation Framework llm-resilience-eval
llm-resilience-eval is an open-source framework focused on evaluating the response stability of large language models (LLMs) under semantics-preserving perturbations, supporting test scenarios such as paraphrasing, flattery, distractor, and confirmation challenge. This framework aims to address the problem of inconsistent responses caused by minor input changes in real-world LLM applications, enhancing model reliability and AI safety.