Revisiting log-linear learning: Asynchrony, completeness and payoff-based implementation
成果类型:
Article
署名作者:
Marden, Jason R.; Shamma, Jeff S.
署名单位:
University System of Georgia; Georgia Institute of Technology
刊物名称:
GAMES AND ECONOMIC BEHAVIOR
ISSN/ISSBN:
0899-8256
DOI:
10.1016/j.geb.2012.03.006
发表日期:
2012
页码:
788-808
关键词:
Potential games
equilibrium selection
Distributed control
摘要:
Log-linear learning is a learning algorithm that provides guarantees on the percentage of time that the action profile will be at a potential maximizer in potential games. The traditional analysis of log-linear learning focuses on explicitly computing the stationary distribution and hence requires a highly structured environment. Since the appeal of log-linear learning is not solely the explicit form of the stationary distribution, we seek to address to what degree one can relax the structural assumptions while maintaining that only potential function maximizers are stochastically stable. In this paper, we introduce slight variants of log-linear learning that provide the desired asymptotic guarantees while relaxing the structural assumptions to include synchronous updates, time-varying action sets, and limitations in information available to the players. The motivation for these relaxations stems from the applicability of log-linear learning to the control of multi-agent systems where these structural assumptions are unrealistic from an implementation perspective. (C) 2012 Elsevier Inc. All rights reserved.