Identify the Nash Equilibrium in Static Games with Random Payoffs

被引:0
|
作者
Zhou, Yichi [1 ]
Li, Jialian [1 ]
Zhu, Jun [1 ]
机构
[1] Tsinghua Univ, CBICR Ctr, Dept Comp Sci & Tech, TNList Lab,State Key Lab Intell Tech & Syst, Beijing, Peoples R China
关键词
MULTIARMED BANDIT; COMPLEXITY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem on how to learn the pure Nash Equilibrium of a two-player zero-sum static game with random payoffs under unknown distributions via efficient payoff queries. We introduce a multi-armed bandit model to this problem due to its ability to find the best arm efficiently among random arms and propose two algorithms for this problem-LUCB-G based on the confidence bounds and a racing algorithm based on successive action elimination We provide an analysis on the sample complexity lower bound when the Nash Equilibrium exists.
引用
收藏
页数:10
相关论文
共 50 条