A methodology for detailed performance modeling of reduction computations on SMP machines

被引:3
|
作者
Jin, RM [1 ]
Agrawal, G [1 ]
机构
[1] Ohio State Univ, Dept Comp & Informat Sci, Columbus, OH 43210 USA
基金
美国国家科学基金会;
关键词
parallel processing; shared memory; memory hierarchy; data mining;
D O I
10.1016/j.peva.2004.10.017
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we revisit the problem of performance prediction on SMP machines, motivated by the need for selecting parallelization strategy for random write reductions. Such reductions frequently arise in data mining algorithms. In our previous work, we have developed a number of techniques for parallelizing this class of reductions. Our previous work has shown that each of the three techniques, full replication, optimized full locking, and cache-sensitive, can outperform others depending upon problem, dataset, and machine parameters. Therefore, an important question is, "Can we predict the performance of these techniques for a given problem, dataset, and machine?". This paper addresses this question by developing an analytical performance model that captures a two-level cache, coherence cache misses, TLB misses, locking overheads, and contention for memory. Analytical model is combined with results from micro-benchmarking to predict performance on real machines. We have validated our model on two different SMP machines. Our results show that our model effectively captures the impact of memory hierarchy (two-level cache and TLB) as well as the factors that limit parallelism (contention for locks, memory contention, and coherence cache misses). The difference between predicted and measured performance is within 20% in almost all cases. Moreover, the model is quite accurate in predicting the relative performance of the three parallelization techniques. (c) 2004 Elsevier B.V. All rights reserved.
引用
收藏
页码:73 / 105
页数:33
相关论文
共 50 条
  • [1] Performance evaluation of BLAST on SMP machines
    Kim, Hong-Soog
    Kim, Hae-Jin
    Han, Dong-Soo
    FRONTIERS OF HIGH PERFORMANCE COMPUTING AND NETWORKING - ISPA 2006 WORKSHOPS, PROCEEDINGS, 2006, 4331 : 668 - +
  • [2] Detailed analytical modeling for electromagnetic performance in actively shielded superconducting machines
    Shin, Kyung-Hun
    Bang, Tae-Kyoung
    Choi, Jang-Young
    Cho, Han-Wook
    Haran, Kiruba S.
    AIP ADVANCES, 2021, 11 (02)
  • [3] A MODELING METHODOLOGY FOR THE ANALYSIS OF CONCURRENT SYSTEMS AND COMPUTATIONS
    KAPELNIKOV, A
    MUNTZ, RR
    ERCEGOVAC, MD
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 1989, 6 (03) : 568 - 597
  • [4] Performance Modeling and Mapping of Sparse Computations
    Bliss, Nadya T.
    Mohindra, Sanjeev
    O'Reilly, Una-May
    PROCEEDINGS OF THE HPCMP USERS GROUP CONFERENCE 2008, 2008, : 448 - +
  • [5] Performance modeling of deterministic transport computations
    Kerbyson, DJ
    Hoisie, A
    Pautz, SD
    PERFORMANCE ANALYSIS AND GRID COMPUTING, 2004, : 21 - 39
  • [6] MODELING THE ERRORS OF MULTIAXIS MACHINES - A GENERAL METHODOLOGY
    SOONS, JA
    THEUWS, FC
    SCHELLEKENS, PH
    PRECISION ENGINEERING-JOURNAL OF THE AMERICAN SOCIETY FOR PRECISION ENGINEERING, 1992, 14 (01): : 5 - 19
  • [7] Washing Machines in Europe - Detailed Assessment of Consumption and Performance
    Stamminger, Rainer
    Schmitz, Angelika
    TENSIDE SURFACTANTS DETERGENTS, 2016, 53 (01) : 70 - 86
  • [8] A METHODOLOGY FOR PERFORMANCE ANALYSIS OF PARALLEL COMPUTATIONS WITH LOOPING CONSTRUCTS
    KAPELNIKOV, A
    MUNTZ, RR
    ERCEGOVAC, MD
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 1992, 14 (02) : 105 - 120
  • [9] Performance of different comunication mechanisms for FEM computations on PC-based cluster with SMP nodes
    Olas, T
    Lacinski, L
    Karczewski, K
    Tomas, A
    Wyrzykowski, R
    PAR ELEC 2002: INTERNATIONAL CONFERENCE ON PARALLEL COMPUTING IN ELECTRICAL ENGINEERING, 2002, : 305 - 311
  • [10] Modeling and simulative performance analysis of SMP and clustered computer architectures
    Burns, MW
    George, AD
    Wallace, BA
    SIMULATION, 2000, 74 (02) : 84 - 96