Social reasoning games such as Werewolf/Mafia and Avalon have long been important experimental scenarios for studying human social intelligence, deception detection, and strategic reasoning. These games require players to make decisions under incomplete information—they must hide their identities while inferring the truth through observing others' words and actions.
With the continuous improvement of large language model capabilities, researchers have begun to explore the performance of AI systems in tasks requiring complex social reasoning. Can LLMs understand the subtleties of deception? Can they infer others' true intentions through dialogue? Can they maintain strategic consistency under pressure? These questions not only concern the boundaries of AI capabilities but also provide a new perspective for us to understand intelligence itself.
The social-inference project was born in this context. It combines classic social reasoning game mechanisms with modern LLM technology to create a unique AI behavior research platform.