Bounding fixed points of set-based Bellman operator and Nash equilibria of stochastic games

被引:3
|
作者
Li, Sarah H. Q. [1 ]
Adje, Assale [2 ]
Garoche, Pierre-Loic [3 ]
Acikmese, Behcet [1 ]
机构
[1] Univ Washington, William E Boeing Dept Aeronaut & Astronaut, Seattle, WA 98195 USA
[2] Univ Perpignan, LAMPS, Via Domitia, Perpignan, France
[3] Univ Toulouse, ENAC, Toulouse, France
关键词
Markov decision process; Learning theory; Stochastic control; Multi-agent systems; Learning in games; Decision making and autonomy; MARKOV; STABILITY;
D O I
10.1016/j.automatica.2021.109685
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Motivated by uncertain parameters encountered in Markov decision processes (MDPs) and stochastic games, we study the effect of parameter-uncertainty on Bellman operator-based algorithms under a set-based framework. Specifically, we first consider a family of MDPs where the cost parameters are in a given compact set; we then define a Bellman operator acting on a set of value functions to produce a new set of value functions as the output under all possible variations in the cost parameter. We prove the existence of a fixed point of this set-based Bellman operator by showing that the operator is contractive on a complete metric space, and explore its relationship with the corresponding family of MDPs and stochastic games. Additionally, we show that given interval set-bounded cost parameters, we can form exact bounds on the set of optimal value functions. Finally, we utilize our results to bound the value function trajectory of a player in a stochastic game. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 50 条