Communication-Efficient Distributed Learning of Discrete Probability Distributions

被引:0
|
作者
Diakonikolas, Ilias [1 ]
Grigorescu, Elena [2 ]
Li, Jerry [3 ]
Natarajan, Abhiram [2 ]
Onak, Krzysztof [4 ]
Schmidt, Ludwig [3 ]
机构
[1] USC, CS, Los Angeles, CA 90007 USA
[2] Purdue Univ, CS, W Lafayette, IN 47907 USA
[3] MIT, EECS & CSAIL, Cambridge, MA 02139 USA
[4] IBM Res Corp, Albany, NY USA
关键词
DENSITY-ESTIMATION; MULTIVARIATE HISTOGRAMS; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We initiate a systematic investigation of distribution learning (density estimation) when the data is distributed across multiple servers. The servers must communicate with a referee and the goal is to estimate the underlying distribution with as few bits of communication as possible. We focus on non-parametric density estimation of discrete distributions with respect to the l(1) and ,l(2) norms. We provide the first non-trivial upper and lower bounds on the communication complexity of this basic estimation task in various settings of interest. Specifically, our results include the following: 1. When the unknown discrete distribution is unstructured and each server has only one sample, we show that any blackboard protocol (i.e., any protocol in which servers interact arbitrarily using public messages) that learns the distribution must essentially communicate the entire sample. 2. For the case of structured distributions, such as k-histograms and monotone distributions, we design distributed learning algorithms that achieve significantly better communication guarantees than the naive ones, and obtain tight upper and lower bounds in several regimes. Our distributed learning algorithms run in near-linear time and are robust to model misspecification. Our results provide insights on the interplay between structure and communication efficiency for a range of fundamental distribution estimation tasks.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Communication-Efficient Distributed Learning via Sparse and Adaptive Stochastic Gradient
    Deng, Xiaoge
    Li, Dongsheng
    Sun, Tao
    Lu, Xicheng
    IEEE TRANSACTIONS ON BIG DATA, 2025, 11 (01) : 234 - 246
  • [32] Adaptive Top-K in SGD for Communication-Efficient Distributed Learning
    Ruan, Mengzhe
    Yan, Guangfeng
    Xiao, Yuanzhang
    Song, Linqi
    Xu, Weitao
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 5280 - 5285
  • [33] EDEN: Communication-Efficient and Robust Distributed Mean Estimation for Federated Learning
    Vargaftik, Shay
    Ben Basat, Ran
    Portnoy, Amit
    Mendelson, Gal
    Ben-Itzhak, Yaniv
    Mitzenmacher, Michael
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [34] Communication-Efficient Coded Computing for Distributed Multi-Task Learning
    Hu, Haoyang
    Wu, Youlong
    Shi, Yuanming
    Li, Songze
    Jiang, Chunxiao
    Zhang, Wei
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2023, 71 (07) : 3861 - 3875
  • [35] Communication-Efficient and Byzantine-Robust Distributed Learning with Error Feedback
    Ghosh A.
    Maity R.K.
    Kadhe S.
    Mazumdar A.
    Ramchandran K.
    IEEE Journal on Selected Areas in Information Theory, 2021, 2 (03): : 942 - 953
  • [36] Communication-Efficient Distributed Deep Learning with Merged Gradient Sparsification on GPUs
    Shi, Shaohuai
    Wang, Qiang
    Chu, Xiaowen
    Li, Bo
    Qin, Yang
    Liu, Ruihao
    Zhao, Xinxiao
    IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2020, : 406 - 415
  • [37] Communication-Efficient and Distributed Learning Over Wireless Networks: Principles and Applications
    Park, Jihong
    Samarakoon, Sumudu
    Elgabli, Anis
    Kim, Joongheon
    Bennis, Mehdi
    Kim, Seong-Lyun
    Debbah, Merouane
    PROCEEDINGS OF THE IEEE, 2021, 109 (05) : 796 - 819
  • [38] Efficient-Adam: Communication-Efficient Distributed Adam
    Chen, Congliang
    Shen, Li
    Liu, Wei
    Luo, Zhi-Quan
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 : 3257 - 3266
  • [39] Double Quantization for Communication-Efficient Distributed Optimization
    Yu, Yue
    Wu, Jiaxiang
    Huang, Longbo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [40] Towards Communication-Efficient Distributed Background Subtraction
    Hung Ngoc Phan
    Synh Viet-Uyen Ha
    Phuong Hoai Ha
    RECENT CHALLENGES IN INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2022, 2022, 1716 : 490 - 502