Communication-Efficient Distributed Learning of Discrete Probability Distributions

被引:0
|
作者
Diakonikolas, Ilias [1 ]
Grigorescu, Elena [2 ]
Li, Jerry [3 ]
Natarajan, Abhiram [2 ]
Onak, Krzysztof [4 ]
Schmidt, Ludwig [3 ]
机构
[1] USC, CS, Los Angeles, CA 90007 USA
[2] Purdue Univ, CS, W Lafayette, IN 47907 USA
[3] MIT, EECS & CSAIL, Cambridge, MA 02139 USA
[4] IBM Res Corp, Albany, NY USA
关键词
DENSITY-ESTIMATION; MULTIVARIATE HISTOGRAMS; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We initiate a systematic investigation of distribution learning (density estimation) when the data is distributed across multiple servers. The servers must communicate with a referee and the goal is to estimate the underlying distribution with as few bits of communication as possible. We focus on non-parametric density estimation of discrete distributions with respect to the l(1) and ,l(2) norms. We provide the first non-trivial upper and lower bounds on the communication complexity of this basic estimation task in various settings of interest. Specifically, our results include the following: 1. When the unknown discrete distribution is unstructured and each server has only one sample, we show that any blackboard protocol (i.e., any protocol in which servers interact arbitrarily using public messages) that learns the distribution must essentially communicate the entire sample. 2. For the case of structured distributions, such as k-histograms and monotone distributions, we design distributed learning algorithms that achieve significantly better communication guarantees than the naive ones, and obtain tight upper and lower bounds in several regimes. Our distributed learning algorithms run in near-linear time and are robust to model misspecification. Our results provide insights on the interplay between structure and communication efficiency for a range of fundamental distribution estimation tasks.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Communication-Efficient Distributed Learning: An Overview
    Cao, Xuanyu
    Basar, Tamer
    Diggavi, Suhas
    Eldar, Yonina C.
    Letaief, Khaled B.
    Poor, H. Vincent
    Zhang, Junshan
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2023, 41 (04) : 851 - 873
  • [2] More communication-efficient distributed sparse learning
    Zhou, Xingcai
    Yang, Guang
    INFORMATION SCIENCES, 2024, 668
  • [3] More communication-efficient distributed sparse learning
    Zhou, Xingcai
    Yang, Guang
    Information Sciences, 2024, 668
  • [4] Communication-Efficient Distributed Cooperative Learning With Compressed Beliefs
    Toghani, Mohammad Taha
    Uribe, Cesar A.
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2022, 9 (03): : 1215 - 1226
  • [5] Communication-Efficient and Resilient Distributed Q-Learning
    Xie, Yijing
    Mou, Shaoshuai
    Sundaram, Shreyas
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3351 - 3364
  • [6] Local Stochastic ADMM for Communication-Efficient Distributed Learning
    ben Issaid, Chaouki
    Elgabli, Anis
    Bennis, Mehdi
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 1880 - 1885
  • [7] Communication-Efficient and Privacy-Aware Distributed Learning
    Gogineni, Vinay Chakravarthi
    Moradi, Ashkan
    Venkategowda, Naveen K. D.
    Werner, Stefan
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 705 - 720
  • [8] Ordered Gradient Approach for Communication-Efficient Distributed Learning
    Chen, Yicheng
    Sadler, Brian M.
    Blum, Rick S.
    PROCEEDINGS OF THE 21ST IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC2020), 2020,
  • [9] Communication-Efficient and Byzantine-Robust Distributed Learning
    Ghosh, Avishek
    Maity, Raj Kumar
    Kadhe, Swanand
    Mazumdar, Arya
    Ramchandran, Kannan
    2020 INFORMATION THEORY AND APPLICATIONS WORKSHOP (ITA), 2020,
  • [10] Communication-efficient Distributed Learning for Large Batch Optimization
    Liu, Rui
    Mozafari, Barzan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,