Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent

被引:0
|
作者
Blanchard, Peva [1 ]
El Mhamdi, El Mandi [1 ]
Guerraoui, Rachid [1 ]
Stainer, Julien [1 ]
机构
[1] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
基金
瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the resilience to Byzantine failures of distributed implementations of Stochastic Gradient Descent (SGD). So far, distributed machine learning frameworks have largely ignored the possibility of failures, especially arbitrary (i.e., Byzantine) ones. Causes of failures include software bugs, network asynchrony, biases in local datasets, as well as attackers trying to compromise the entire system. Assuming a set of n workers, up to f being Byzantine, we ask how resilient can SGD be, without limiting the dimension, nor the size of the parameter space. We first show that no gradient aggregation rule based on a linear combination of the vectors proposed by the workers (i.e, current approaches) tolerates a single Byzantine failure. We then formulate a resilience property of the aggregation rule capturing the basic requirements to guarantee convergence despite f Byzantine workers. We propose Krum, an aggregation rule that satisfies our resilience property, which we argue is the first provably Byzantine-resilient algorithm for distributed SGD. We also report on experimental evaluations of Krum.
引用
收藏
页数:11
相关论文
共 50 条
  • [2] Distributed Byzantine Tolerant Stochastic Gradient Descent in the Era of Big Data
    Jin, Richeng
    He, Xiaofan
    Dai, Huaiyu
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [3] Byzantine Fault-Tolerant Parallelized Stochastic Gradient Descent for Linear Regression
    Gupta, Nirupam
    Vaidya, Nitin H.
    2019 57TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2019, : 415 - 420
  • [4] Byzantine Stochastic Gradient Descent
    Alistarh, Dan
    Allen-Zhu, Zeyuan
    Li, Jerry
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] Brief Announcement: Byzantine-Tolerant Machine Learning
    Blanchard, Peva
    El Mhamdi, El Mahdi
    Guerraoui, Rachid
    Stainer, Julien
    PROCEEDINGS OF THE ACM SYMPOSIUM ON PRINCIPLES OF DISTRIBUTED COMPUTING (PODC'17), 2017, : 455 - 457
  • [6] Byzantine Fault-Tolerant Distributed Machine Learning with Norm-Based Comparative Gradient Elimination
    Gupta, Nirupam
    Liu, Shuo
    Vaidya, Nitin
    51ST ANNUAL IEEE/IFIP INTERNATIONAL CONFERENCE ON DEPENDABLE SYSTEMS AND NETWORKS (DSN-W 2021), 2021, : 175 - 181
  • [7] Byzantine-Tolerant Distributed Coordinate Descent
    Data, Deepesh
    Diggavi, Suhas
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 2724 - 2728
  • [8] Byzantine Fault Tolerant Distributed Stochastic Gradient Descent Based on Over-the-Air Computation
    Park, Sangjun
    Choi, Wan
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2022, 70 (05) : 3204 - 3219
  • [9] Stochastic Gradient Descent and Its Variants in Machine Learning
    Netrapalli, Praneeth
    JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 201 - 213
  • [10] Stochastic Gradient Descent and Its Variants in Machine Learning
    Praneeth Netrapalli
    Journal of the Indian Institute of Science, 2019, 99 : 201 - 213