Distributed Byzantine Tolerant Stochastic Gradient Descent in the Era of Big Data

被引:0
|
作者
Jin, Richeng [1 ]
He, Xiaofan [2 ]
Dai, Huaiyu [1 ]
机构
[1] North Carolina State Univ, Dept ECE, Raleigh, NC 27695 USA
[2] Wuhan Univ, Elect Informat Sch, Wuhan, Hubei, Peoples R China
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The recent advances in sensor technologies and smart devices enable the collaborative collection of a sheer volume of data from multiple information sources. As a promising tool to efficiently extract useful information from such big data, machine learning has been pushed to the forefront and seen great success in a wide range of relevant areas such as computer vision, health care, and financial market analysis. To accommodate the large volume of data, there is a surge of interest in the design of distributed machine learning, among which stochastic gradient descent (SGD) is one of the mostly adopted methods. Nonetheless, distributed machine learning methods may be vulnerable to Byzantine attack, in which the adversary can deliberately share falsified information to disrupt the intended machine learning procedures. In this work, two asynchronous Byzantine tolerant SGD algorithms are proposed, in which the honest collaborative workers are assumed to store the model parameters derived from their own local data and use them as the ground truth. The proposed algorithms can deal with an arbitrary number of Byzantine attackers and are provably convergent. Simulation results based on a real-world dataset are presented to verify the theoretical results and demonstrate the effectiveness of the proposed algorithms.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] ON DISTRIBUTED STOCHASTIC GRADIENT DESCENT FOR NONCONVEX FUNCTIONS IN THE PRESENCE OF BYZANTINES
    Bulusu, Saikiran
    Khanduri, Prashant
    Sharma, Pranay
    Varshney, Pramod K.
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3137 - 3141
  • [32] Distributed Differentially Private Stochastic Gradient Descent: An Empirical Study
    Hegedus, Istvan
    Jelasity, Mark
    2016 24TH EUROMICRO INTERNATIONAL CONFERENCE ON PARALLEL, DISTRIBUTED, AND NETWORK-BASED PROCESSING (PDP), 2016, : 566 - 573
  • [33] A Novel Distributed Variant of Stochastic Gradient Descent and Its Optimization
    Wang, Yi-qi
    Zhao, Ya-wei
    Shi, Zhan
    Yin, Jian-ping
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COMPUTER SCIENCE (AICS 2016), 2016, : 486 - 492
  • [34] Convergence in High Probability of Distributed Stochastic Gradient Descent Algorithms
    Lu, Kaihong
    Wang, Hongxia
    Zhang, Huanshui
    Wang, Long
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (04) : 2189 - 2204
  • [35] Fast Convergence for Stochastic and Distributed Gradient Descent in the Interpolation Limit
    Mitra, Partha P.
    2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 1890 - 1894
  • [36] CONTROLLING STOCHASTIC GRADIENT DESCENT USING STOCHASTIC APPROXIMATION FOR ROBUST DISTRIBUTED OPTIMIZATION
    Jain, Adit
    Krishnamurthy, Vikram
    NUMERICAL ALGEBRA CONTROL AND OPTIMIZATION, 2025, 15 (01): : 173 - 195
  • [37] Byzantine-Resilient Stochastic Gradient Descent for Distributed Learning: A Lipschitz-Inspired Coordinate-wise Median Approach
    Yang, Haibo
    Zhang, Xin
    Fang, Minghong
    Liu, Jia
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 5832 - 5837
  • [38] Distributed Coordinate Descent Method for Learning with Big Data
    Richtarik, Peter
    Takac, Martin
    JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17
  • [39] Data-Dependent Stability of Stochastic Gradient Descent
    Kuzborskij, Ilja
    Lampert, Christoph H.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [40] Distributed coordinate descent method for learning with big data
    Richtárik, Peter
    Takáč, Martin
    Journal of Machine Learning Research, 2016, 17