ExaHDF5: Delivering Efficient Parallel I/O on Exascale Computing Systems

被引:31
|
作者
Byna, Suren [1 ]
Breitenfeld, M. Scot [2 ]
Dong, Bin [1 ]
Koziol, Quincey [1 ]
Pourmal, Elena [2 ]
Robinson, Dana [2 ]
Soumagne, Jerome [2 ]
Tang, Houjun [1 ]
Vishwanath, Venkatram [3 ]
Warren, Richard [2 ]
机构
[1] Lawrence Berkeley Natl Lab, Berkeley, CA 94597 USA
[2] HDF Grp, Champaign, IL 61820 USA
[3] Argonne Natl Lab, Lemont, IL 60439 USA
关键词
parallel I; O; Hierarchical Data Format version 5 (HDF5); I; O performance; virtual object layer; HDF5; optimizations;
D O I
10.1007/s11390-020-9822-9
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Scientific applications at exascale generate and analyze massive amounts of data. A critical requirement of these applications is the capability to access and manage this data efficiently on exascale systems. Parallel I/O, the key technology enables moving data between compute nodes and storage, faces monumental challenges from new applications, memory, and storage architectures considered in the designs of exascale systems. As the storage hierarchy is expanding to include node-local persistent memory, burst buffers, etc., as well as disk-based storage, data movement among these layers must be efficient. Parallel I/O libraries of the future should be capable of handling file sizes of many terabytes and beyond. In this paper, we describe new capabilities we have developed in Hierarchical Data Format version 5 (HDF5), the most popular parallel I/O library for scientific applications. HDF5 is one of the most used libraries at the leadership computing facilities for performing parallel I/O on existing HPC systems. The state-of-the-art features we describe include: Virtual Object Layer (VOL), Data Elevator, asynchronous I/O, full-featured single-writer and multiple-reader (Full SWMR), and parallel querying. In this paper, we introduce these features, their implementations, and the performance and feature benefits to applications and other libraries.
引用
收藏
页码:145 / 160
页数:16
相关论文
共 50 条
  • [21] An Efficient Algorithm for Resource Allocation in Parallel and Distributed Computing Systems
    El-Zoghdy, S. F.
    Nofal, M.
    Shohla, M. A.
    El-sawy, A.
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2013, 4 (02) : 251 - 259
  • [22] Efficient parallel optimization of volume meshes on heterogeneous computing systems
    Zuofu Cheng
    Eric Shaffer
    Raine Yeh
    George Zagaris
    Luke Olson
    Engineering with Computers, 2017, 33 : 717 - 726
  • [23] EFFICIENT MODELING OF CONSTRAINED MULTIBODY SYSTEMS FOR APPLICATION WITH PARALLEL COMPUTING
    ANDERSON, KS
    ZEITSCHRIFT FUR ANGEWANDTE MATHEMATIK UND MECHANIK, 1993, 73 (7-8): : T871 - T874
  • [24] I/O efficient: computing SCCs in massive graphs
    Zhang, Zhiwei
    Yu, Jeffrey Xu
    Qin, Lu
    Chang, Lijun
    Lin, Xuemin
    VLDB JOURNAL, 2015, 24 (02): : 245 - 270
  • [25] Contract & Expand: I/O Efficient SCCs Computing
    Zhang, Zhiwei
    Qin, Lu
    Yu, Jeffrey Xu
    2014 IEEE 30TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE), 2014, : 208 - 219
  • [26] I/O-Efficient Statistical Computing with RIOT
    Zhang, Yi
    Zhang, Weiping
    Yang, Jun
    26TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING ICDE 2010, 2010, : 1157 - 1160
  • [27] I/O efficient: computing SCCs in massive graphs
    Zhiwei Zhang
    Jeffrey Xu Yu
    Lu Qin
    Lijun Chang
    Xuemin Lin
    The VLDB Journal, 2015, 24 : 245 - 270
  • [28] A parallel I/O mechanism for distributed systems
    Baer, T
    Wyckoff, P
    2004 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING, 2004, : 63 - 69
  • [29] Priority scheduling in parallel I/O systems
    Venugopal, CR
    Rao, SSSP
    Patkar, S
    INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED PROCESSING TECHNIQUES AND APPLICATIONS, VOLS I-V, PROCEEDINGS, 1999, : 2554 - 2560
  • [30] (HOEIN)-O-2: A Hierarchical Hybrid Optical/Electrical Interconnection Network for Exascale Computing Systems
    Lu, Yunfeng
    Gu, Huaxi
    Chakrabarty, Krishnendu
    Yang, Yintang
    IEEE TRANSACTIONS ON MULTI-SCALE COMPUTING SYSTEMS, 2018, 4 (04): : 722 - 733