Playing repeated games with Large Language Models

LLMs are increasingly used in applications where they interact with humans and other agents. We propose to use behavioural game theory to study LLM's cooperation and coordination behaviour. We let different LLMs play finitely repeated games with each other, with human-like strategies, and actual human players. Our results show that LLMs perform particularly well at self-interested games like the iterated Prisoner's Dilemma family. However, they behave sub-optimally in games that require coordination, like the Battle of the Sexes. We verify that these behavioural signatures are stable across robustness checks. We additionally show how GPT-4's behaviour can be modulated by providing additional information about its opponent and by using a "social chain-of-thought" (SCoT) strategy. This also leads to better scores and more successful coordination when interacting with human players. These results enrich our understanding of LLM's social behaviour and pave the way for a behavioural game theory for machines.
View on arXiv@article{akata2025_2305.16867, title={ Playing repeated games with Large Language Models }, author={ Elif Akata and Lion Schulz and Julian Coda-Forno and Seong Joon Oh and Matthias Bethge and Eric Schulz }, journal={arXiv preprint arXiv:2305.16867}, year={ 2025 } }