共 21 条
No-regret dynamics in the Fenchel game: a unified framework for algorithmic convex optimization
被引:2
|作者:
Wang, Jun-Kun
[1
]
Abernethy, Jacob
[2
]
Levy, Kfir Y.
[3
]
机构:
[1] Yale Univ, Dept Comp Sci, New Haven, CT 06511 USA
[2] Georgia Inst Technol, Sch Comp Sci, Atlanta, GA USA
[3] Technion Israel Inst Technol, Dept Elect & Comp Engn, Haifa, Israel
基金:
以色列科学基金会;
美国国家科学基金会;
关键词:
Online learning;
No-regret learning;
Zero-sum game;
Convex optimization;
Frank-Wolfe method;
Nesterov's accelerated gradient methods;
Momentum methods;
VARIATIONAL-INEQUALITIES;
1ST-ORDER METHODS;
SPLITTING METHOD;
POINT ALGORITHM;
ONLINE;
CONVERGENCE;
GRADIENT;
RATES;
D O I:
10.1007/s10107-023-01976-y
中图分类号:
TP31 [计算机软件];
学科分类号:
081202 ;
0835 ;
摘要:
We develop an algorithmic framework for solving convex optimization problems using no-regret game dynamics. By converting the problem of minimizing a convex function into an auxiliary problem of solving a min-max game in a sequential fashion, we can consider a range of strategies for each of the two-players who must select their actions one after the other. A common choice for these strategies are so-called no-regret learning algorithms, and we describe a number of such and prove bounds on their regret. We then show that many classical first-order methods for convex optimization-including average-iterate gradient descent, the Frank-Wolfe algorithm, Nesterov's acceleration methods, the accelerated proximal method-can be interpreted as special cases of our framework as long as each player makes the correct choice of no-regret strategy. Proving convergence rates in this framework becomes very straightforward, as they follow from plugging in the appropriate known regret bounds. Our framework also gives rise to a number of new first-order methods for special cases of convex optimization that were not previously known.
引用
收藏
页码:203 / 268
页数:66
相关论文