With the widespread application of Large Language Models (LLMs) across various fields, AI security issues have received increasing attention. Red-teaming, as a structured approach, identifies potential vulnerabilities by inputting adversarial prompts into AI systems. While mainstream AI labs conduct internal red teaming before releasing models, independent third-party auditing tools are crucial for ensuring accountability, especially when evaluating how models handle sensitive content related to protected groups.
RedLog is an open-source project born in this context. Created by developer thiagoolivauk as a portfolio project focusing on the intersection of AI security research and content policy, it aims to provide researchers with a standardized multi-model comparative testing framework.