GEM: Ultra-Efficient Near-Memory Reconfigurable Acceleration for Read Mapping by Dividing and Predictive Scattering

被引:0
|
作者
Chen, Longlong [1 ]
Zhu, Jianfeng [1 ]
Peng, Guiqiang [1 ]
Liu, Mingxu [2 ]
Wei, Shaojun [1 ]
Liu, Leibo [1 ]
机构
[1] Tsinghua Univ, Sch Integrated Circuits, Beijing 100084, Peoples R China
[2] Beijing Superstring Acad Memory Technol, Dept DRAM, Beijing 100176, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Data-centric computing; dividing and scattering; genome assembly; near memory computing; read mapping; reconfigurable computing; SEQUENCE; CLASSIFICATION; ARCHITECTURE; ALIGNMENT;
D O I
10.1109/TPDS.2023.3309462
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Read mapping, which maps billions of reads to a reference DNA, poses a significant performance bottleneck in genomic analysis. Current accelerators for read mapping are primarily bounded by the intensive and random memory access to huge datasets. Near-data processing (NDP) infrastructures are promising to provide extremely high bandwidth. However, existing frameworks failed to reach this potential due to poor locality and high redundancy. Our idea is to introduce prediction under the insight that candidate mapping positions become predictable when the reference is organized in coarse-grain slices. We present GEM (Genomic Memory), an ultra-efficient near-memory accelerator for read mapping. GEM adopts a novel data-centric framework, named dividing-and-predictive-scattering (DPS), which synthesizes information of seed existence to predict the target mapping locations to reduce memory access redundancy. During preparation, DPS divides the reference into coarse-grained slices and creates predictive filters to assess the likelihood of reads belonging to each slice. During mapping, DPS predicts and scatters reads to considerably fewer slices compared than without prediction. By employing small on-chip SRAM-based predictors with high accuracy, DPS minimizes unnecessary DRAM access and data movement from remote memory. In essence, DPS trades pre-seeding predictors for localized access patterns and low redundancy, hence achieving high throughput for data-intensive applications. We implement GEM by integrating coarse-grain reconfigurable architectures (CGRAs) in the logic layer of a 3D-stacked DRAM infrastructure, utilizing the massive banks as slices. GEM leverages CGRAs for their flexibility in supporting various algorithms tailored to different datasets. Bloom filters are leveraged for slice prediction, providing an error rate below 1%. Evaluation results demonstrate that GEM reduces memory requests by 95% and alignments by 87%, achieving a throughput improvement of 15.3 x and 11.0 x compared to compute-centric and broadcast-based baselines on the same NDP platform. Overall, GEM achieves a 3.5 x 3.5 x throughput improvement and 2.1 x 2.1 x energy efficiency compared to state-of-the-art ASIC accelerators.
引用
收藏
页码:3059 / 3072
页数:14
相关论文
共 4 条
  • [1] Ultra Efficient Acceleration for De Novo Genome Assembly via Near-Memory Computing
    Zhou, Minxuan
    Wu, Lingxi
    Li, Muzhou
    Moshiri, Niema
    Skadron, Kevin
    Rosing, Tajana
    30TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT 2021), 2021, : 199 - 212
  • [2] HeNCoG: A Heterogeneous Near-memory Computing Architecture for Energy Efficient GCN Acceleration
    Hwang, Seung-Eon
    Song, Duyeong
    Park, Jongsun
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [3] Dagger: Efficient and Fast RPCs in Cloud Microservices with Near-Memory Reconfigurable NICs
    Lazarev, Nikita
    Xiang, Shaojie
    Adit, Neil
    Zhang, Zhiru
    Delimitrou, Christina
    ASPLOS XXVI: TWENTY-SIXTH INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS, 2021, : 36 - 51
  • [4] Dagger: Towards Efficient RPCs in Cloud Microservices With Near-Memory Reconfigurable NICs
    Lazarev, Nikita
    Adit, Neil
    Xiang, Shaojie
    Zhang, Zhiru
    Delimitrou, Christina
    IEEE COMPUTER ARCHITECTURE LETTERS, 2020, 19 (02) : 134 - 138