Learning Sparse Polymatrix Games in Polynomial Time and Sample Complexity

We consider the problem of learning sparse polymatrix games from observations of strategic interactions. We show that a polynomial time method based on -group regularized logistic regression recovers a game, whose Nash equilibria are the -Nash equilibria of the game from which the data was generated (true game), in samples of strategy profiles --- where is the maximum number of pure strategies of a player, is the number of players, and is the maximum degree of the game graph. Under slightly more stringent separability conditions on the payoff matrices of the true game, we show that our method learns a game with the exact same Nash equilibria as the true game. We also show that samples are necessary for any method to consistently recover a game, with the same Nash-equilibria as the true game, from observations of strategic interactions. We verify our theoretical results through simulation experiments.
View on arXiv