共 50 条
Rationality of Learning Algorithms in Repeated Normal-Form Games
被引:0
|作者:
Bajaj, Shivam
[1
]
Das, Pranoy
[1
]
Vorobeychik, Yevgeniy
[2
]
Gupta, Vijay
[1
]
机构:
[1] Purdue Univ, Elmore Family Sch Elect & Comp Engn, W Lafayette, IN 47907 USA
[2] Washington Univ St Louis, Dept Comp Sci & Engn, St Louis, MO 63130 USA
来源:
关键词:
Game theory;
learning in games;
agents- based systems;
D O I:
10.1109/LCSYS.2024.3486631
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Many learning algorithms are known to converge to an equilibrium for specific classes of games if the same learning algorithm is adopted by all agents. However, when the agents are self-interested, a natural question is whether the agents have an incentive to unilaterally shift to an alternative learning algorithm. We capture such incentives as an algorithm's rationality ratio, which is the ratio of the highest payoff an agent can obtain by unilaterally deviating from a learning algorithm to its payoff from following it. We define a learning algorithm to be c-rational if its rationality ratio is at most c irrespective of the game. We show that popular learning algorithms such as fictitious play and regret-matching are not c-rational for any constant c >= 1. We also show that if an agent can only observe the actions of the other agents but not their payoffs, then there are games for which c-rational algorithms do not exist. We then propose a framework that can build upon any existing learning algorithm and establish, under mild assumptions, that our proposed algorithm is (i) c-rational for a given c >= 1 (ii) the strategies of the agents converge to an equilibrium, with high probability, if all agents follow it.
引用
收藏
页码:2409 / 2414
页数:6
相关论文