کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
410611 | 679154 | 2009 | 14 صفحه PDF | دانلود رایگان |

Markov games is a framework which can be used to formalise nn-agent reinforcement learning (RL). Littman (Markov games as a framework for multi-agent reinforcement learning, in: Proceedings of the 11th International Conference on Machine Learning (ICML-94), 1994.) uses this framework to model two-agent zero-sum problems and, within this context, proposes the minimax-QQ algorithm. This paper reviews RL algorithms for two-player zero-sum Markov games and introduces a new, simple, fast, algorithm, called QL2QL2. QL2QL2 is compared to several standard algorithms (QQ-learning, Minimax and minimax-QQ) implemented with the QashQash library written in Python. The experiments show that QL2QL2 converges empirically to optimal mixed policies, as minimax-QQ, but uses a surprisingly simple and cheap updating rule.
Journal: Neurocomputing - Volume 72, Issues 7–9, March 2009, Pages 1494–1507