A Simple Nearly Optimal Restart Scheme For Speeding Up First-Order Methods

被引:0
|
作者
James Renegar
Benjamin Grimmer
机构
[1] Cornell University,School of Operations Research and Information Engineering
关键词
First-order method; Restarting; Convex optimization; Parallelization; Convergence rates; 90C25; 90C52;
D O I
暂无
中图分类号
学科分类号
摘要
We present a simple scheme for restarting first-order methods for convex optimization problems. Restarts are made based only on achieving specified decreases in objective values, the specified amounts being the same for all optimization problems. Unlike existing restart schemes, the scheme makes no attempt to learn parameter values characterizing the structure of an optimization problem, nor does it require any special information that would not be available in practice (unless the first-order method chosen to be employed in the scheme itself requires special information). As immediate corollaries to the main theorems, we show that when some well-known first-order methods are employed in the scheme, the resulting complexity bounds are nearly optimal for particular—yet quite general—classes of problems.
引用
收藏
页码:211 / 256
页数:45
相关论文
共 50 条