ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2312.08484
24
2
v1v2 (latest)

Q-learners Can Provably Collude in the Iterated Prisoner's Dilemma

13 December 2023
Quentin Bertrand
Juan Agustin Duque
Emilio Calvano
Gauthier Gidel
ArXiv (abs)PDFHTML
Main:9 Pages
7 Figures
Bibliography:3 Pages
5 Tables
Appendix:12 Pages
Abstract

The deployment of machine learning systems in the market economy has triggered academic and institutional fears over potential tacit collusion between fully automated agents. Multiple recent economics studies have empirically shown the emergence of collusive strategies from agents guided by machine learning algorithms. In this work, we prove that multi-agent Q-learners playing the iterated prisoner's dilemma can learn to collude. The complexity of the cooperative multi-agent setting yields multiple fixed-point policies for QQQ-learning: the main technical contribution of this work is to characterize the convergence towards a specific cooperative policy. More precisely, in the iterated prisoner's dilemma, we show that with optimistic Q-values, any self-play Q-learner can provably learn a cooperative policy called Pavlov, also referred to as win-stay, lose-switch policy, which strongly differs from the vanilla Pareto-dominated always defect policy.

View on arXiv
Comments on this paper