Asynchronous Distributed-Memory Parallel Algorithms for Influence Maximization

被引:0
|
作者
Singhal, Shubhendra Pal [1 ]
Hati, Souvadra [1 ]
Young, Jeffrey [1 ]
Sarkar, Vivek [1 ]
Hayashi, Akihiro [1 ]
Vuduc, Richard [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
关键词
Influence maximization; FA-BSP; PGAS; IMM; NETWORKS;
D O I
10.1109/SC41406.2024.00108
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Influence maximization (IM) is the problem of finding the k most influential nodes in a graph. We propose distributed-memory parallel algorithms for the two main kernels of a state-of-the-art implementation of one IM algorithm, influence maximization via martingales (IMM). The baseline relies on a bulk-synchronous parallel approach and uses replication to reduce communication and achieve approximate load balance, at the cost of synchronization and high memory requirements. By contrast, our method fully distributes the data, thereby improving memory scalability, and uses fine-grained asynchronous parallelism to improve network utilization and the cost of doing more communication. We show our design and implementation can achieve up to 29.6x speedup over the MPI-based state-of-the-art on synthetic and real-world network graphs. Moreover, ours is the first implementation that can run IMM to find influencers in the 'twitter' graph (41M nodes and 1.4B edges) in 200 seconds using 8K CPU cores of NERSC Perlmutter supercomputer.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] ASYNCHRONOUS PARALLEL ARC CONSISTENCY ALGORITHMS ON A DISTRIBUTED-MEMORY MACHINE
    CONRAD, JM
    AGRAWAL, DP
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 1995, 24 (01) : 27 - 40
  • [2] Lifting sequential graph algorithms for distributed-memory parallel computation
    Gregor, D
    Lumsdaine, A
    ACM SIGPLAN NOTICES, 2005, 40 (10) : 423 - 437
  • [3] PARALLEL MATRIX TRANSPOSE ALGORITHMS ON DISTRIBUTED-MEMORY CONCURRENT COMPUTERS
    CHOI, JY
    DONGARRA, JJ
    WALKER, DW
    PARALLEL COMPUTING, 1995, 21 (09) : 1387 - 1405
  • [4] Distributed-Memory Parallel JointNMF
    Eswar, Srinivas
    Cobb, Benjamin
    Hayashi, Koby
    Kannan, Ramakrishnan
    Ballard, Grey
    Vuduc, Richard
    Park, Haesun
    PROCEEDINGS OF THE 37TH INTERNATIONAL CONFERENCE ON SUPERCOMPUTING, ACM ICS 2023, 2023, : 301 - 312
  • [5] Parallel Asynchronous Distributed-Memory Maximal Independent Set Algorithm with Work Ordering
    Kanewala, Thejaka
    Zalewski, Marcin
    Lumsdaine, Andrew
    2017 IEEE 24TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING (HIPC), 2017, : 52 - 61
  • [6] A framework for generating distributed-memory parallel programs for block recursive algorithms
    Gupta, SKS
    Huang, CH
    Sadayappan, P
    Johnson, RW
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 1996, 34 (02) : 137 - 153
  • [7] Comparison of backfilling algorithms for job scheduling in distributed-memory parallel systems
    Department of Computer Science, Bowling Green State University, Bowling Green, OH 43403
    Comput. Educ. J., 2007, 4 (22-31):
  • [8] Parallel ILP for distributed-memory architectures
    Nuno A. Fonseca
    Ashwin Srinivasan
    Fernando Silva
    Rui Camacho
    Machine Learning, 2009, 74 : 257 - 279
  • [9] PARALLEL ANNEALING ON DISTRIBUTED-MEMORY SYSTEMS
    LEE, FH
    STILES, GS
    SWAMINATHAN, V
    PROGRAMMING AND COMPUTER SOFTWARE, 1995, 21 (01) : 1 - 8
  • [10] Parallel ILP for distributed-memory architectures
    Fonseca, Nuno A.
    Srinivasan, Ashwin
    Silva, Fernando
    Camacho, Rui
    MACHINE LEARNING, 2009, 74 (03) : 257 - 279