vqSGD: Vector Quantized Stochastic Gradient Descent

被引:5
|
作者
Gandikota, Venkata [1 ]
Kane, Daniel [2 ,3 ]
Maity, Raj Kumar [4 ]
Mazumdar, Arya [5 ]
机构
[1] Syracuse Univ, Dept Elect Engn & Comp Sci, Syracuse, NY 13244 USA
[2] Univ Calif San Diego, Dept Comp Sci, La Jolla, CA 92093 USA
[3] Univ Calif San Diego, Dept Math, La Jolla, CA 92093 USA
[4] UMass Amherst, Coll Informat & Comp Sci, Amherst, MA 01003 USA
[5] Univ Calif San Diego, Halicioglu Data Sci Inst, La Jolla, CA 92093 USA
关键词
Vector quantization; communication efficiency; mean estimation; stochastic gradient descent (SGD);
D O I
10.1109/TIT.2022.3161620
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this work, we present a family of vector quantization schemes vqSGD (Vector-Quantized Stochastic Gradient Descent) that provide an asymptotic reduction in the communication cost with convergence guarantees in first-order distributed optimization. In the process we derive the following fundamental information theoretic fact: Theta(d/R-2) bits are necessary and sufficient (up to an additive O(log d) term) to describe an unbiased estimator (g) over cap (g) for any g in the d-dimensional unit sphere, under the constraint that parallel to(g) over cap (g)parallel to(2) <= R almost surely, R > 1. In particular, we consider a randomized scheme based on the convex hull of a point set, that returns an unbiased estimator of a d-dimensional gradient vector with almost surely bounded norm. We provide multiple efficient instances of our scheme, that are near optimal, and require o(d) bits of communication at the expense of tolerable increase in error. The instances of our quantization scheme are obtained using well-known families of binary error-correcting codes and provide a smooth tradeoff between the communication and the estimation error of quantization. Furthermore, we show that vqSGD also offers automatic privacy guarantees.
引用
收藏
页码:4573 / 4587
页数:15
相关论文
共 50 条
  • [1] vqSGD: Vector Quantized Stochastic Gradient Descent
    Gandikota, Venkata
    Kane, Daniel
    Maity, Raj Kumar
    Mazumdar, Arya
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [2] Stochastic Gradient Descent Support Vector Clustering
    Tung Pham
    Hang Dang
    Trung Le
    Hoang-Thai Le
    PROCEEDINGS OF 2015 2ND NATIONAL FOUNDATION FOR SCIENCE AND TECHNOLOGY DEVELOPMENT CONFERENCE ON INFORMATION AND COMPUTER SCIENCE NICS 2015, 2015, : 88 - 93
  • [3] Differentially Quantized Gradient Descent
    Lin, Chung-Yi
    Kostina, Victoria
    Hassibi, Babak
    2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 1200 - 1205
  • [4] An Exact Quantized Decentralized Gradient Descent Algorithm
    Reisizadeh, Amirhossein
    Mokhtari, Aryan
    Hassani, Hamed
    Pedarsani, Ramtin
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (19) : 4934 - 4947
  • [5] Distributed Optimization with Gradient Descent and Quantized Communication
    Rikos, Apostolos I.
    Jiang, Wei
    Charalambous, Themistoklis
    Johansson, Karl H.
    IFAC PAPERSONLINE, 2023, 56 (02): : 5900 - 5906
  • [6] Unforgeability in Stochastic Gradient Descent
    Baluta, Teodora
    Nikolic, Ivica
    Jain, Racchit
    Aggarwal, Divesh
    Saxena, Prateek
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 1138 - 1152
  • [7] Preconditioned Stochastic Gradient Descent
    Li, Xi-Lin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (05) : 1454 - 1466
  • [8] Stochastic Reweighted Gradient Descent
    El Hanchi, Ayoub
    Stephens, David A.
    Maddison, Chris J.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [9] Stochastic gradient descent tricks
    Bottou, Léon
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2012, 7700 LECTURE NO : 421 - 436
  • [10] Byzantine Stochastic Gradient Descent
    Alistarh, Dan
    Allen-Zhu, Zeyuan
    Li, Jerry
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31