On Private and Robust Bandits

被引:0
|
作者
Wu, Yulian [1 ]
Zhou, Xingyu [2 ]
Tao, Youming [3 ]
Wang, Di [1 ]
机构
[1] KAUST, Thuwal, Saudi Arabia
[2] Wayne State Univ, Wayne, NJ USA
[3] Shandong Univ, Jinan, Peoples R China
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
关键词
MULTIARMED BANDIT;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study private and robust multi-armed bandits (MABs), where the agent receives Huber's contaminated heavy-tailed rewards and meanwhile needs to ensure differential privacy. We consider both the finite k-th raw moment and the finite k-th central moment settings for heavy-tailed rewards distributions with k >= 2. We first present its minimax lower bound, characterizing the information-theoretic limit of regret with respect to privacy budget, contamination level, and heavy-tailedness. Then, we propose a meta-algorithm that builds on a private and robust mean estimation sub-routine PRM that essentially relies on reward truncation and the Laplace mechanism. For the above two different heavy-tailed settings, we give corresponding schemes of PRM, which enable us to achieve nearly-optimal regrets. Moreover, our two proposed truncation-based or histogram-based PRM schemes achieve the optimal trade-off between estimation accuracy, privacy and robustness. Finally, we support our theoretical results and show the effectiveness of our algorithms with experimental studies.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] (Locally) Differentially Private Combinatorial Semi-Bandits
    Chen, Xiaoyu
    Zheng, Kai
    Zhou, Zixin
    Yang, Yunchang
    Chen, Wei
    Wang, Liwei
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [22] Stochastic Linear Bandits Robust to Adversarial Attacks
    Bogunovic, Ilija
    Losalka, Arpan
    Krause, Andreas
    Scarlett, Jonathan
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [23] Dynamic Global Sensitivity for Differentially Private Contextual Bandits
    Wang, Huazheng
    Zhao, David
    Wang, Hongning
    PROCEEDINGS OF THE 16TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2022, 2022, : 179 - 187
  • [24] Robust Pure Exploration in Linear Bandits with Limited Budget
    Alieva, Ayya
    Cutkosky, Ashok
    Das, Abhimanyu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [25] Robust Heavy-Tailed Linear Bandits Algorithm
    Ma L.
    Zhao P.
    Zhou Z.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2023, 60 (06): : 1385 - 1395
  • [26] Mercenaries, Pirates, Bandits and Empires: Private Violence in Historical Context
    Riemann, Malte
    RUSI JOURNAL, 2011, 156 (02): : 112 - 113
  • [27] Differentially Private Multi-Armed Bandits in the Shuffle Model
    Tenenbaum, Jay
    Kaplan, Haim
    Mansour, Yishay
    Stemmer, Uri
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [28] Efficient and Robust High-Dimensional Linear Contextual Bandits
    Chen, Cheng
    Luo, Luo
    Zhang, Weinan
    Yu, Yong
    Lian, Yijiang
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 4259 - 4265
  • [29] A Differentially Private Approach for Budgeted Combinatorial Multi-Armed Bandits
    Wang, Hengzhi
    Cui, Laizhong
    Wang, En
    Liu, Jiangchuan
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2025, 22 (01) : 424 - 439
  • [30] Robust Stochastic Multi-Armed Bandits with Historical Data
    Yacobi, Sarah Boufelja
    Bounefouf, Djallel
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 959 - 965