Adaptively scheduling parallel loops in distributed shared-memory systems

被引:33
|
作者
Yan, Y [1 ]
Jin, CM [1 ]
Zhang, XD [1 ]
机构
[1] INTERVOICE INC,DALLAS,TX
基金
美国国家科学基金会;
关键词
adaptive scheduling algorithms; dynamic information; load balancing; parallel loops; processor affinity; shared-memory systems;
D O I
10.1109/71.569656
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Using runtime information of load distributions and processor affinity, we propose an adaptive scheduling algorithm and its variations from different control mechanisms. The proposed algorithm applies different degrees of aggressiveness to adjust loop scheduling granularities, aiming at improving the execution performance of parallel loops by making scheduling decisions that match the real workload distributions at runtime. We experimentally compared the performance of our algorithm and its variations with several existing scheduling algorithms on two parallel machines: the KSR-1 and the Convex Exemplar. The kernel application programs we used for performance evaluation were carefully selected for different classes of parallel loops. Our results show that using runtime information to adaptively adjust scheduling granularity is an effective way to handle loops with a wide range of load distributions when no prior knowledge of the execution can be used. The overhead caused by collecting runtime information is insignificant in comparison with the performance improvement. Our experiments show that the adaptive algorithm and its five variations outperformed the existing scheduling algorithms.
引用
收藏
页码:70 / 81
页数:12
相关论文
共 50 条
  • [11] Performance evaluation of or-parallel logic programming systems on distributed shared-memory architectures
    Calegario, VM
    Dutra, ID
    EURO-PAR'99: PARALLEL PROCESSING, 1999, 1685 : 1484 - 1491
  • [12] A comparative evaluation of hybrid distributed shared-memory systems
    Moga, Adrian
    Dubois, Michel
    JOURNAL OF SYSTEMS ARCHITECTURE, 2009, 55 (01) : 43 - 52
  • [13] A Distributed-memory Parallelization of a Shared-memory Parallel Ensemble Kalman Filter
    Rostami, M. Ali
    Buecker, H. Martin
    Vogt, Christian
    Seidler, Ralf
    Neuhaeuser, David
    Rath, Volker
    16TH INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING (SYNASC 2014), 2014, : 455 - 462
  • [14] A fault tolerant self-scheduling scheme for parallel loops on shared memory systems
    Wang, Yizhuo
    Nicolau, Alexandru
    Cammarota, Rosario
    Veidenbaum, Alexander V.
    2012 19TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING (HIPC), 2012,
  • [15] THE IMPACT OF PARALLEL LOOP SCHEDULING STRATEGIES ON PREFETCHING IN A SHARED-MEMORY MULTIPROCESSOR
    LILJA, DJ
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 1994, 5 (06) : 573 - 584
  • [16] Scheduling user-level threads on distributed shared-memory multiprocessors
    Polychronopoulos, ED
    Papatheodorou, TS
    EURO-PAR'99: PARALLEL PROCESSING, 1999, 1685 : 358 - 368
  • [17] SAFE SELF-SCHEDULING - A PARALLEL LOOP SCHEDULING SCHEME FOR SHARED-MEMORY MULTIPROCESSORS
    LIU, J
    SALETORE, VA
    LEWIS, TG
    INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, 1994, 22 (06) : 589 - 616
  • [18] SHARED-MEMORY PERFORMANCE OF MULTIPLE COMPUTER TERMINALS IN PARALLEL DISTRIBUTED INFORMATION-PROCESSING SYSTEMS
    REDDI, AV
    COMPUTER PERFORMANCE, 1984, 5 (01): : 55 - 63
  • [19] Performance of hierarchical processor scheduling in shared-memory multiprocessor systems
    Dandamudi, SP
    Ayachi, S
    IEEE TRANSACTIONS ON COMPUTERS, 1999, 48 (11) : 1202 - 1213
  • [20] Parallel Mining of Correlated Heavy Hitters on Distributed and Shared-Memory Architectures
    Pulimeno, Marco
    Epicoco, Italo
    Cafaro, Massimo
    Melle, Catiuscia
    Aloisio, Giovanni
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 5111 - 5118