Section 01
[Main Floor] HEPTA: Guide to the Automated Benchmark Framework for Evaluating LLM Teaching Effectiveness in HCI Education
HEPTA (AI HCI Education Performance Test) is an automated benchmark framework specifically designed to evaluate the teaching effectiveness of large language models (LLMs) in the field of human-computer interaction (HCI) education. As LLMs are widely applied in education, how to objectively assess their teaching effectiveness in specialized fields like HCI has become a key issue. HEPTA fills the gap in traditional benchmarks (such as MMLU and HumanEval) that lack dedicated evaluation of teaching quality. Through a systematic framework, it tests AI performance in HCI education, with core components including evaluation dimension design, test dataset construction, and automated evaluation mechanisms, providing a scientific basis for relevant groups.