Constant Regret, Generalized Mixability, and Mirror Descent

被引:0
|
作者
Mhammedi, Zakaria [1 ,2 ]
Williamson, Robert C. [1 ,2 ]
机构
[1] Australian Natl Univ, Res Sch Comp Sci, Canberra, ACT, Australia
[2] DATA61, Sydney, NSW, Australia
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018) | 2018年 / 31卷
基金
澳大利亚研究理事会;
关键词
PREDICTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the setting of prediction with expert advice; a learner makes predictions by aggregating those of a group of experts. Under this setting, and for the right choice of loss function and "mixing" algorithm, it is possible for the learner to achieve a constant regret regardless of the number of prediction rounds. For example, a constant regret can be achieved for mixable losses using the aggregating algorithm. The Generalized Aggregating Algorithm (GAA) is a name for a family of algorithms parameterized by convex functions on simplices (entropies), which reduce to the aggregating algorithm when using the Shannon entropy S. For a given entropy Phi, losses for which a constant regret is possible using the GAA are called Phi-mixable. Which losses are Phi-mixable was previously left as an open question. We fully characterize Phi-mixability and answer other open questions posed by [6]. We show that the Shannon entropy S is fundamental in nature when it comes to mixability; any Phi-mixable loss is necessarily S-mixable, and the lowest worst-case regret of the GAA is achieved using the Shannon entropy. Finally, by leveraging the connection between the mirror descent algorithm and the update step of the GAA, we suggest a new adaptive generalized aggregating algorithm and analyze its performance in terms of the regret bound.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Shifting Regret, Mirror Descent, and Matrices
    Gyorgy, Andras
    Szepesvari, Csaba
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [2] No-regret Caching via Online Mirror Descent
    Salem, Tareq Si
    Neglia, Giovanni
    Ioannidis, Stratis
    ACM TRANSACTIONS ON MODELING AND PERFORMANCE EVALUATION OF COMPUTING SYSTEMS, 2023, 8 (04)
  • [3] No-Regret Caching via Online Mirror Descent
    Salem, Tareq Si
    Neglia, Giovanni
    Ioannidis, Straus
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [4] Minimizing Weighted Counterfactual Regret with Optimistic Online Mirror Descent
    Xu, Hang
    Li, Kai
    Liu, Bingyun
    Fu, Haobo
    Fei, Qiang
    Xing, Junliang
    Cheng, Jian
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 5272 - 5280
  • [5] Equivalence Analysis between Counterfactual Regret Minimization and Online Mirror Descent
    Liu, Weiming
    Jiang, Huacong
    Li, Bin
    Li, Houqiang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [6] Dynamic Regret of Online Mirror Descent for Relatively Smooth Convex Cost Functions
    Eshraghi, Nima
    Liang, Ben
    IEEE CONTROL SYSTEMS LETTERS, 2022, 6 : 2395 - 2400
  • [7] A generalized online mirror descent with applications to classification and regression
    Orabona, Francesco
    Crammer, Koby
    Cesa-Bianchi, Nicolo
    MACHINE LEARNING, 2015, 99 (03) : 411 - 435
  • [8] A generalized online mirror descent with applications to classification and regression
    Francesco Orabona
    Koby Crammer
    Nicolò Cesa-Bianchi
    Machine Learning, 2015, 99 : 411 - 435
  • [9] Improving Dynamic Regret in Distributed Online Mirror Descent Using Primal and Dual Information
    Eshraghi, Nima
    Liang, Ben
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 168, 2022, 168
  • [10] Mirror Descent Maximizes Generalized Margin and Can Be Implemented Efficiently
    Sun, Haoyuan
    Ahn, Kwangjun
    Thrampoulidis, Christos
    Azizan, Navid
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,