ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1706.05296
16
982

Value-Decomposition Networks For Cooperative Multi-Agent Learning

16 June 2017
P. Sunehag
Guy Lever
A. Gruslys
Wojciech M. Czarnecki
V. Zambaldi
Max Jaderberg
Marc Lanctot
Nicolas Sonnerat
Joel Z Leibo
K. Tuyls
T. Graepel
ArXivPDFHTML
Abstract

We study the problem of cooperative multi-agent reinforcement learning with a single joint reward signal. This class of learning problems is difficult because of the often large combined action and observation spaces. In the fully centralized and decentralized approaches, we find the problem of spurious rewards and a phenomenon we call the "lazy agent" problem, which arises due to partial observability. We address these problems by training individual agents with a novel value decomposition network architecture, which learns to decompose the team value function into agent-wise value functions. We perform an experimental evaluation across a range of partially-observable multi-agent domains and show that learning such value-decompositions leads to superior results, in particular when combined with weight sharing, role information and information channels.

View on arXiv
Comments on this paper