Section 01
LLM Red Team Evaluation Platform: Building a Security Testing System for Language Models (Introduction)
This article introduces the LLM Red Team Evaluation Platform, a modular red team evaluation framework for large language models. It systematically tests model performance in dimensions such as hallucination, instruction following, reasoning consistency, and adversarial robustness through automated evaluation and mutation attacks, aiming to build a security testing system for language models and ensure the safety of AI systems.