Communication-Efficient Distributed Learning of Discrete Probability Distributions

被引:0
|
作者
Diakonikolas, Ilias [1 ]
Grigorescu, Elena [2 ]
Li, Jerry [3 ]
Natarajan, Abhiram [2 ]
Onak, Krzysztof [4 ]
Schmidt, Ludwig [3 ]
机构
[1] USC, CS, Los Angeles, CA 90007 USA
[2] Purdue Univ, CS, W Lafayette, IN 47907 USA
[3] MIT, EECS & CSAIL, Cambridge, MA 02139 USA
[4] IBM Res Corp, Albany, NY USA
关键词
DENSITY-ESTIMATION; MULTIVARIATE HISTOGRAMS; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We initiate a systematic investigation of distribution learning (density estimation) when the data is distributed across multiple servers. The servers must communicate with a referee and the goal is to estimate the underlying distribution with as few bits of communication as possible. We focus on non-parametric density estimation of discrete distributions with respect to the l(1) and ,l(2) norms. We provide the first non-trivial upper and lower bounds on the communication complexity of this basic estimation task in various settings of interest. Specifically, our results include the following: 1. When the unknown discrete distribution is unstructured and each server has only one sample, we show that any blackboard protocol (i.e., any protocol in which servers interact arbitrarily using public messages) that learns the distribution must essentially communicate the entire sample. 2. For the case of structured distributions, such as k-histograms and monotone distributions, we design distributed learning algorithms that achieve significantly better communication guarantees than the naive ones, and obtain tight upper and lower bounds in several regimes. Our distributed learning algorithms run in near-linear time and are robust to model misspecification. Our results provide insights on the interplay between structure and communication efficiency for a range of fundamental distribution estimation tasks.
引用
收藏
页数:11
相关论文
共 50 条
  • [11] Communication-Efficient Quantum Algorithm for Distributed Machine Learning
    Tang, Hao
    Li, Boning
    Wang, Guoqing
    Xu, Haowei
    Li, Changhao
    Barr, Ariel
    Cappellaro, Paola
    Li, Ju
    PHYSICAL REVIEW LETTERS, 2023, 130 (15)
  • [12] Communication-Efficient Distributed Deep Metric Learning with Hybrid Synchronization
    Su, Yuxin
    Lyu, Michael
    King, Irwin
    CIKM'18: PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2018, : 1463 - 1472
  • [13] Guest Editorial Communication-Efficient Distributed Learning Over Networks
    Cao, Xuanyu
    Basar, Tamer
    Diggavi, Suhas
    Eldar, Yonina C.
    Letaief, Khaled B.
    Poor, H. Vincent
    Zhang, Junshan
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2023, 41 (04) : 845 - 850
  • [14] LAG: Lazily Aggregated Gradient for Communication-Efficient Distributed Learning
    Chen, Tianyi
    Giannakis, Georgios B.
    Sun, Tao
    Yin, Wotao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [15] Communication-Efficient Coded Distributed Multi-Task Learning
    Tang, Hua
    Hu, Haoyang
    Yuan, Kai
    Wu, Youlong
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [16] FedBCD: A Communication-Efficient Collaborative Learning Framework for Distributed Features
    Liu, Yang
    Zhang, Xinwei
    Kang, Yan
    Li, Liping
    Chen, Tianjian
    Hong, Mingyi
    Yang, Qiang
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2022, 70 : 4277 - 4290
  • [17] Communication-Efficient Topology Orchestration for Distributed Learning in UAV Networks
    Liang, Zixuan
    Lyu, Xinchen
    Ren, Chenshan
    Li, Na
    Li, Kai
    20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024, 2024, : 662 - 667
  • [18] Intermittent Pulling With Local Compensation for Communication-Efficient Distributed Learning
    Wang, Haozhao
    Qu, Zhihao
    Guo, Song
    Gao, Xin
    Li, Ruixuan
    Ye, Baoliu
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2022, 10 (02) : 779 - 791
  • [19] Communication-Efficient Gradient Coding for Straggler Mitigation in Distributed Learning
    Kadhe, Swanand
    Koyluoglu, O. Ozan
    Ramchandran, Kannan
    2020 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2020, : 2634 - 2639
  • [20] CE-SGD: Communication-Efficient Distributed Machine Learning
    Tao, Zeyi
    Xia, Qi
    Li, Qun
    Cheng, Songqing
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,