Zing Forum

Reading

TrustgameLLM: Do Large Language Models 'Treat People Differently Based on Their Background' When Playing the Trust Game?

An innovative study uses trust game experiments to reveal whether large language models (LLMs) adjust their cooperation strategies when facing virtual opponents of different genders and nationalities.

大语言模型信任博弈AI偏见社会身份公平性行为经济学GitHub项目
Published 2026-05-13 19:14Recent activity 2026-05-13 19:18Estimated read 5 min
TrustgameLLM: Do Large Language Models 'Treat People Differently Based on Their Background' When Playing the Trust Game?
1

Section 01

[Introduction] TrustgameLLM Study: Do Large Language Models 'Treat People Differently Based on Social Identity'?

This study uses the classic trust game experimental framework to systematically examine whether large language models (LLMs) exhibit differentiated strategies based on social identities such as gender and nationality in interactive decision-making. The results show that LLMs do adjust their cooperative behaviors according to the social identities of virtual opponents, revealing potential bias patterns in training data, which is of great significance for AI fairness research.

2

Section 02

Background: AI Bias Issues and Basic Concepts of the Trust Game

When we interact with LLMs like ChatGPT, do we get different responses based on user background differences? The TrustgameLLM project conducts research on this question. The trust game is a classic paradigm in behavioral economics: Player A transfers part of their funds to Player B (the amount is tripled), and Player B decides how much to return. The core is the win-win of trust and reciprocity or the loss from selfishness.

3

Section 03

Research Design: Game Experiments Between LLMs and Virtual Humans

The study has LLMs play one side of the game, and the other side is a virtual human with different gender (male/female) and nationality characteristics (presented only in text, no behavioral differences). Core hypothesis: If LLMs are influenced by identity cues, there will be systematic differences in the amount of investment (level of trust) when facing opponents of different identities.

4

Section 04

Experimental Findings: LLMs Have Biases Based on Social Identity

The results show that LLMs adjust their cooperation strategies: 1. Gender differences: Some models invest more in opponents of specific genders, suggesting gender stereotypes in training data; 2. Nationality bias: Different levels of trust towards opponents of different nationalities, reflecting uneven distribution of national images in training corpora. These biases come from statistical patterns in training data and are not intentionally implanted by developers.

5

Section 05

Technical Implementation: Reproducible Experimental Framework

The TrustgameLLM project provides complete code and datasets, with features including: standardized interfaces supporting calls to multiple mainstream LLMs; configurable identity characteristics of virtual players; complete recording of decision-making processes and results; control experiments to distinguish identity effects from random fluctuations, facilitating reproduction and expansion.

6

Section 06

Significance: Real-World Challenges for AI Fairness

The significance of the research goes far beyond academia: LLMs are widely used in scenarios such as customer service, recruitment, and credit. If there are identity biases in simple games, unfair consequences may be amplified in complex scenarios (e.g., biased decisions in loan approvals due to gender/nationality cues).

7

Section 07

Future Directions: Paths to Building More Fair AI

Future research can be carried out from four aspects: 1. Trace the sources of bias and identify key corpora in training data; 2. Develop debiasing technologies (fine-tuning/inference stages); 3. Establish bias detection benchmarks to improve transparency; 4. Expand cross-cultural research to reveal the cultural specificity of biases.