DyCache: Dynamic Multi-Grain Cache Management for Irregular Memory Accesses on GPU

被引:6
|
作者
Guo, Hui [1 ]
Huang, Libo [2 ]
Lu, Yashuai [4 ]
Ma, Sheng [2 ]
Wang, Zhiying [3 ]
机构
[1] Natl Univ Def Technol, Changsha 410073, Hunan, Peoples R China
[2] Natl Univ Def Technol, Sch Comp, Changsha 410073, Hunan, Peoples R China
[3] Natl Univ Def Technol, Comp Engn, Dept Comp, Changsha 410073, Hunan, Peoples R China
[4] Space Engn Univ, Beijing 101416, Peoples R China
来源
IEEE ACCESS | 2018年 / 6卷
关键词
Accelerator architectures; cache memory; fine-grain cache management; GPGPU computing; irregular memory access; memory divergence; memory management;
D O I
10.1109/ACCESS.2018.2818193
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
GPU utilizes the wide cache-line (128B) on-chip cache to provide high bandwidth and efficient memory accesses for applications with regularly-organized data structures. However, emerging applications exhibit a lot of irregular control flows and memory access patterns. Irregular memory accesses generate many fine-grain memory accesses to L1 data cache. This mismatching between fine-grain data accesses and the coarse-grain cache design makes the on-chip memory space more constrained and as a result, the frequency of cache line replacement increases and Ll data cache is utilized inefficiently. Fine-grain cache management is proposed to provide efficient cache management to improve the efficiency of data array utilization. Unlike other static fine-grain cache managements, we propose a dynamic multi-grain cache management, called DyCache, to resolve the inefficient use of L1 data cache. Through monitoring the memory access pattern of applications, DyCache can dynamically alter the cache management granularity in order to improve the performance of GPU for applications with irregular memory accesses while not impact the performance for regular applications. Our experiment demonstrates that DyCache can achieve a 40% geometric mean improvement on IPC for applications with irregular memory accesses against the baseline cache (128B), while for applications with regular memory accesses, DyCache does not degrade the performance.
引用
收藏
页码:38881 / 38891
页数:11
相关论文
共 37 条
  • [21] SuperNeurons: Dynamic GPU Memory Management for Training Deep Neural Networks
    Wang L.
    Ye J.
    Zhao Y.
    Wu W.
    Li A.
    Song S.L.
    Xu Z.
    Kraska T.
    2018, Association for Computing Machinery, 2 Penn Plaza, Suite 701, New York, NY 10121-0701, United States (53): : 41 - 53
  • [22] Efficient Multi-GPU Memory Management for Deep Learning Acceleration
    Kim, Youngrang
    Lee, Jaehwan
    Kim, Jik-Soo
    Jei, Hyunseung
    Roh, Hongchan
    2018 IEEE 3RD INTERNATIONAL WORKSHOPS ON FOUNDATIONS AND APPLICATIONS OF SELF* SYSTEMS (FAS*W), 2018, : 37 - 43
  • [23] POSTER: BACM: Barrier-Aware Cache Management for Irregular Memory-Intensive GPGPU Workloads
    Liu, Yuxi
    Zhao, Xia
    Yu, Zhibin
    Wang, Zhenlin
    Wang, Xiaolin
    Luo, Yingwei
    Eeckhout, Lieven
    2017 26TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT), 2017, : 140 - 141
  • [24] Dynamic Memory Management for GPU-based training of Deep Neural Networks
    Shriram, S. B.
    Garg, Anshuj
    Kulkarni, Purushottam
    2019 IEEE 33RD INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2019), 2019, : 200 - 209
  • [25] Dynamic scratch-pad memory management for irregular array access patterns
    Chen, G.
    Ozturk, O.
    Kandemir, M.
    Karakoy, M.
    2006 DESIGN AUTOMATION AND TEST IN EUROPE, VOLS 1-3, PROCEEDINGS, 2006, : 929 - +
  • [26] GPUdmm: A High-Performance and Memory-Oblivious GPU Architecture Using Dynamic Memory Management
    Kim, Youngsok
    Lee, Jaewon
    Jo, Jae-Eon
    Kim, Jangwoo
    2014 20TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE COMPUTER ARCHITECTURE (HPCA-20), 2014, : 546 - 557
  • [27] Supporting Secure Multi-GPU Computing with Dynamic and Batched Metadata Management
    Na, Seonjin
    Kim, Jungwoo
    Lee, Sunho
    Huh, Jaehyuk
    2024 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE, HPCA 2024, 2024, : 204 - 217
  • [28] GPS: A Global Publish-Subscribe Model for Multi-GPU Memory Management
    Muthukrishnan, Harini
    Lustig, Daniel
    Nellans, David
    Wenisch, Thomas
    PROCEEDINGS OF 54TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, MICRO 2021, 2021, : 46 - 58
  • [29] Dynamic Cache Management in Multi-Core Architectures through Run-time Adaptation
    Hameed, Fazal
    Bauer, Lars
    Henkel, Joerg
    DESIGN, AUTOMATION & TEST IN EUROPE (DATE 2012), 2012, : 485 - 490
  • [30] WADE: Writeback-Aware Dynamic Cache Management for NVM-Based Main Memory System
    Wang, Zhe
    Shan, Shuchang
    Cao, Ting
    Gu, Junli
    Xu, Yi
    Mu, Shuai
    Xie, Yuan
    Jimenez, Daniel A.
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2013, 10 (04)