Distributed-Memory Parallel JointNMF

被引:1
|
作者
Eswar, Srinivas [1 ]
Cobb, Benjamin [2 ]
Hayashi, Koby [2 ]
Kannan, Ramakrishnan [3 ]
Ballard, Grey [4 ]
Vuduc, Richard [2 ]
Park, Haesun [2 ]
机构
[1] Argonne Natl Lab, Lemont, IL 60439 USA
[2] Georgia Inst Technol, Sch Computat Sci & Engn, Atlanta, GA 30332 USA
[3] Oak Ridge Natl Lab, Oak Ridge, TN USA
[4] Wake Forest Univ, Dept Comp Sci, Winston Salem, NC 27101 USA
基金
美国国家科学基金会; 美国能源部;
关键词
High Performance Computing; Multimodal Inputs; Nonnegative Matrix Factorization; NONNEGATIVE MATRIX; COMMUNICATION; MPI;
D O I
10.1145/3577193.3593733
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Joint Nonnegative Matrix Factorization (JointNMF) is a hybrid method for mining information from datasets that contain both feature and connection information. We propose distributed-memory parallelizations of three algorithms for solving the JointNMF problem based on Alternating Nonnegative Least Squares, Projected Gradient Descent, and Projected Gauss-Newton. We extend well-known communication-avoiding algorithms using a single processor grid case to our coupled case on two processor grids. We demonstrate the scalability of the algorithms on up to 960 cores (40 nodes) with 60% parallel efficiency. The more sophisticated Alternating Nonnegative Least Squares (ANLS) and Gauss-Newton variants outperform the first-order gradient descent method in reducing the objective on large-scale problems. We perform a topic modelling task on a large corpus of academic papers that consists of over 37 million paper abstracts and nearly a billion citation relationships, demonstrating the utility and scalability of the methods.
引用
收藏
页码:301 / 312
页数:12
相关论文
共 50 条