Section 01
[Introduction] TrustMH-Bench: A Credibility Evaluation Benchmark for Large Models in Mental Health Counseling Scenarios
TrustMH-Bench is a credibility evaluation benchmark for large language models (LLMs) designed for the mental health counseling domain. It systematically assesses LLMs' performance in sensitive counseling scenarios across four dimensions: privacy protection, safety, jailbreak resistance, and fairness. It fills the gap where traditional general-purpose evaluation benchmarks (such as MMLU, HumanEval) fail to capture the unique risks in mental health scenarios. As an open-source comprehensive evaluation dataset, it provides a specialized assessment tool for researchers, developers, and regulators.