Distributed Matrix Completion

被引:50
|
作者
Teflioudi, Christina [1 ]
Makari, Faraz [1 ]
Gemulla, Rainer [1 ]
机构
[1] Max Planck Inst Informat, D-66123 Saarbrucken, Germany
关键词
parallel and distributed matrix factorization; stochastic gradient descent; ALS; recommender systems;
D O I
10.1109/ICDM.2012.120
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We discuss parallel and distributed algorithms for large-scale matrix completion on problems with millions of rows, millions of columns, and billions of revealed entries. We focus on in-memory algorithms that run on a small cluster of commodity nodes; even very large problems can be handled effectively in such a setup. Our DALS, ASGD, and DSGD++ algorithms are novel variants of the popular alternating least squares and stochastic gradient descent algorithms; they exploit thread-level parallelism, in-memory processing, and asynchronous communication. We provide some guidance on the asymptotic performance of each algorithm and investigate the performance of both our algorithms and previously proposed MapReduce algorithms in large-scale experiments. We found that DSGD++ outperforms competing methods in terms of overall runtime, memory consumption, and scalability. Using DSGD++, we can factor a matrix with 10B entries on 16 compute nodes in around 40 minutes.
引用
收藏
页码:655 / 664
页数:10
相关论文
共 50 条
  • [1] Distributed Matrix Completion and Robust Factorization
    Mackey, Lester
    Talwalkar, Ameet
    Jordan, Michael I.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2015, 16 : 913 - 960
  • [2] Distributed matrix completion and robust factorization
    Stanford University, Department of Statistics, 390 Serra Mall, Stanford
    CA
    94305, United States
    不详
    CA
    90095, United States
    不详
    CA
    94720, United States
    J. Mach. Learn. Res., (913-960):
  • [3] Differentially Private Matrix Completion via Distributed Matrix Factorization
    Zhou, Haotian
    Liu, Xiao-Yang
    Fu, Cai
    Shang, Chen
    Chang, Xinyi
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (IEEE TRUSTCOM) / 12TH IEEE INTERNATIONAL CONFERENCE ON BIG DATA SCIENCE AND ENGINEERING (IEEE BIGDATASE), 2018, : 1628 - 1631
  • [4] DISTRIBUTED NUCLEAR NORM MINIMIZATION FOR MATRIX COMPLETION
    Mardani, Morteza
    Mateos, Gonzalo
    Giannakis, Georgios B.
    2012 IEEE 13TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC), 2012, : 354 - 358
  • [5] Blind Distributed Beamforming via Matrix Completion
    Vlachos, Evangelos
    Berberidis, Kostas
    2016 IEEE 17TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC), 2016,
  • [6] DISTRIBUTED BEARING ESTIMATION VIA MATRIX COMPLETION
    Waters, Andrew
    Cevher, Volkan
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 2590 - 2593
  • [7] An adaptive distributed completion algorithm for network latency matrix
    Wang, Cong
    Zhang, Feng-Li
    Wang, Rui-Jin
    Li, Min
    Yang, Xiao-Xiang
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2014, 36 (04): : 840 - 846
  • [8] Scaling Stratified Stochastic Gradient Descent for Distributed Matrix Completion
    Abubaker N.
    Karsavuran M.O.
    Aykanat C.
    IEEE Transactions on Knowledge and Data Engineering, 2023, 35 (10) : 10603 - 10615
  • [9] Distributed Spectral Clustering based on Euclidean Distance Matrix Completion
    Scardapane, Simone
    Altilio, Rosa
    Panella, Massimo
    Uncini, Aurelio
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 3093 - 3100
  • [10] BYZANTINE-RESILIENT DISTRIBUTED LARGE-SCALE MATRIX COMPLETION
    Lin, Feng
    Ling, Qing
    Xiong, Zhiwei
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 8167 - 8171