Communication-Efficient Distributed Learning of Discrete Probability Distributions

被引:0
|
作者
Diakonikolas, Ilias [1 ]
Grigorescu, Elena [2 ]
Li, Jerry [3 ]
Natarajan, Abhiram [2 ]
Onak, Krzysztof [4 ]
Schmidt, Ludwig [3 ]
机构
[1] USC, CS, Los Angeles, CA 90007 USA
[2] Purdue Univ, CS, W Lafayette, IN 47907 USA
[3] MIT, EECS & CSAIL, Cambridge, MA 02139 USA
[4] IBM Res Corp, Albany, NY USA
关键词
DENSITY-ESTIMATION; MULTIVARIATE HISTOGRAMS; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We initiate a systematic investigation of distribution learning (density estimation) when the data is distributed across multiple servers. The servers must communicate with a referee and the goal is to estimate the underlying distribution with as few bits of communication as possible. We focus on non-parametric density estimation of discrete distributions with respect to the l(1) and ,l(2) norms. We provide the first non-trivial upper and lower bounds on the communication complexity of this basic estimation task in various settings of interest. Specifically, our results include the following: 1. When the unknown discrete distribution is unstructured and each server has only one sample, we show that any blackboard protocol (i.e., any protocol in which servers interact arbitrarily using public messages) that learns the distribution must essentially communicate the entire sample. 2. For the case of structured distributions, such as k-histograms and monotone distributions, we design distributed learning algorithms that achieve significantly better communication guarantees than the naive ones, and obtain tight upper and lower bounds in several regimes. Our distributed learning algorithms run in near-linear time and are robust to model misspecification. Our results provide insights on the interplay between structure and communication efficiency for a range of fundamental distribution estimation tasks.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Communication-Efficient Distributed Mining of Association Rules
    Assaf Schuster
    Ran Wolff
    Data Mining and Knowledge Discovery, 2004, 8 : 171 - 196
  • [42] Communication-Efficient Distributed PCA by Riemannian Optimization
    Huang, Long-Kai
    Pan, Sinno Jialin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [43] Double Quantization for Communication-Efficient Distributed Optimization
    Huang, Longbo
    PROCEEDINGS OF THE 13TH EAI INTERNATIONAL CONFERENCE ON PERFORMANCE EVALUATION METHODOLOGIES AND TOOLS ( VALUETOOLS 2020), 2020, : 2 - 2
  • [44] Communication-Efficient Distributed Dual Coordinate Ascent
    Jaggi, Martin
    Smith, Virginia
    Takac, Martin
    Terhorst, Jonathan
    Krishnan, Sanjay
    Hofmann, Thomas
    Jordan, Michael, I
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [45] Communication-Efficient Distributed Optimization with Quantized Preconditioners
    Alimisis, Foivos
    Davies, Peter
    Alistarh, Dan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [46] Communication-Efficient Δ-Stepping for Distributed Computing Systems
    Zhang, Haomeng
    Xie, Junfei
    Zhang, Xinyu
    2023 19TH INTERNATIONAL CONFERENCE ON WIRELESS AND MOBILE COMPUTING, NETWORKING AND COMMUNICATIONS, WIMOB, 2023, : 369 - 374
  • [47] Communication-efficient distributed mining of association rules
    Schuster, A
    Wolff, R
    SIGMOD RECORD, 2001, 30 (02) : 473 - 484
  • [48] Communication-efficient Massively Distributed Connected Components
    Lamm, Sebastian
    Sanders, Peter
    2022 IEEE 36TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2022), 2022, : 302 - 312
  • [49] Gradient Sparsification for Communication-Efficient Distributed Optimization
    Wangni, Jianqiao
    Wang, Jialei
    Liu, Ji
    Zhang, Tong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [50] Communication-efficient Conformal Prediction for Distributed Datasets
    Riquelme-Granada, Nery
    Luo, Zhiyuan
    Khuong An Nguyen
    CONFORMAL AND PROBABILISTIC PREDICTION WITH APPLICATIONS, VOL 179, 2022, 179