Transparent Dual Memory Compression Architecture

被引:9
|
作者
Kim, Seikwon [1 ,2 ]
Lee, Seonyoung [1 ]
Kim, Taehoon [1 ]
Huh, Jaehyuk [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Comp, Daejeon, South Korea
[2] Samsung Elect Co Ltd, SW Ctr, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
memory compression; dual compression technique; OS transparency; locality awareness;
D O I
10.1109/PACT.2017.12
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The increasing memory requirements of big data applications have been driving the precipitous growth of memory capacity in server systems. To maximize the efficiency of external memory, HW-based memory compression techniques have been proposed to increase effective memory capacity. Although such memory compression techniques can improve the memory efficiency significantly, a critical trade-off exists in the HW-based compression techniques. As the memory blocks need to be decompressed as quickly as possible to serve cache misses, latency-optimized techniques apply compression at the cacheline granularity, achieving the decompression latency of less than a few cycles. However, such latency-optimized techniques can lose the potential high compression ratios of capacity-optimized techniques, which compress larger memory blocks with longer latency algorithms. Considering the fundamental trade-off in the memory compression, this paper proposes a transparent dual memory compression (DMC) architecture, which selectively uses two compression algorithms with distinct latency and compression characteristics. Exploiting the locality of memory accesses, the proposed architecture compresses less frequently accessed blocks with a capacity-optimized compression algorithm, while keeping recently accessed blocks compressed with a latency-optimized one. Furthermore, instead of relying on the support from the virtual memory system to locate compressed memory blocks, the study advocates a HW-based translation between the uncompressed address space and compressed physical space. This OS-transparent approach eliminates conflicts between compression efficiency and large page support adopted to reduce TLB misses. The proposed compression architecture is applied to the Hybrid Memory Cube (HMC) with a logic layer under the stacked DRAMs. The experimental results show that the proposed compression architecture provides 54% higher compression ratio than the state-of-the-art latency-optimized technique, with no performance degradation over the baseline system without compression.
引用
收藏
页码:206 / 218
页数:13
相关论文
共 50 条
  • [1] Enabling Transparent Memory-Compression for Commodity Memory Systems
    Young, Vinson
    Kariyappa, Sanjay
    Qureshi, Moinuddin K.
    2019 25TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE COMPUTER ARCHITECTURE (HPCA), 2019, : 570 - 581
  • [2] ASSOCIATIVE MEMORY ARCHITECTURE FOR VIDEO COMPRESSION
    IDRIS, F
    PANCHANATHAN, S
    IEE PROCEEDINGS-COMPUTERS AND DIGITAL TECHNIQUES, 1995, 142 (01): : 55 - 64
  • [3] Memory management for dual-addressing memory architecture
    Hung, Ting-Wei
    Chen, Yen-Hao
    Liu, Yi-Yu
    IEICE ELECTRONICS EXPRESS, 2013, 10 (15):
  • [4] Architecture for Transparent Binary Acceleration of Loops with Memory Accesses
    Paulino, Nuno
    Ferreira, Joao Canas
    Cardoso, Joao M. P.
    RECONFIGURABLE COMPUTING: ARCHITECTURES, TOOLS AND APPLICATIONS, 2013, 7806 : 122 - 133
  • [5] Parallelization of image compression on distributed memory architecture
    Daoudi, EM
    Jaâra, EL
    Cherif, N
    HIGH PERFORMANCE COMPUTING AND NETWORKING, PROCEEDINGS, 2000, 1823 : 580 - 583
  • [6] A processor-in-memory architecture for multimedia compression
    Jasionowski, Brandon J.
    Lay, Michelle K.
    Margala, Martin
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2007, 15 (04) : 478 - 483
  • [7] Flexible Memory: A Novel Main Memory Architecture with Block-level Memory Compression
    Cao, Yanan
    Chen, Long
    Zhang, Zhao
    PROCEEDINGS OF THE 2015 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, ARCHITECTURE AND STORAGE (NAS), 2015, : 285 - 294
  • [8] A Novel Compression Based Neuronal Architecture for Memory Encoding
    Kathpalia, Aditi
    Nagaraj, Nithin
    ICDCN '19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING AND NETWORKING, 2019, : 365 - 370
  • [9] Application-Transparent Near-Memory Processing Architecture with Memory Channel Network
    Alian, Mohammad
    Min, Seung Won
    Asgharimoghaddam, Hadi
    Dhar, Ashutosh
    Wang, Dong Kai
    Roewer, Thomas
    McPadden, Adam
    O'Halloran, Oliver
    Chen, Deming
    Xiong, Jinjun
    Kim, Daehoon
    Hwu, Wen-mei
    Kim, Nam Sung
    2018 51ST ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2018, : 802 - 814
  • [10] Dual Adaptation Supports a Parallel Architecture of Motor Memory
    Lee, Jeong-Yoon
    Schweighofer, Nicolas
    JOURNAL OF NEUROSCIENCE, 2009, 29 (33): : 10396 - 10404