Simulating cortical networks on heterogeneous multi-GPU systems

被引:7
|
作者
Nere, Andrew [1 ]
Franey, Sean [1 ]
Hashmi, Atif [1 ]
Lipasti, Mikko [1 ]
机构
[1] Univ Wisconsin, Dept Elect & Comp Engn, Madison, WI 53706 USA
基金
美国国家科学基金会;
关键词
Cortical learning algorithms; CUDA; GPGPU; Profiling systems; RECEPTIVE-FIELDS; FUNCTIONAL ARCHITECTURE; MODEL;
D O I
10.1016/j.jpdc.2012.02.006
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Recent advances in neuroscientific understanding have highlighted the highly parallel computation power of the mammalian neocortex. In this paper we describe a GPGPU-accelerated implementation of an intelligent learning model inspired by the structural and functional properties of the neocortex. Furthermore, we consider two inefficiencies inherent to our initial implementation and propose software optimizations to mitigate such problems. Analysis of our application's behavior and performance provides important insights into the GPGPU architecture, including the number of cores, the memory system, atomic operations, and the global thread scheduler. Additionally, we create a runtime profiling tool for the cortical network that proportionally distributes work across the host CPU as well as multiple GPGPUs available to the system. Using the profiling tool with these optimizations on Nvidia's CUDA framework, we achieve up to 60 x speedup over a single-threaded CPU implementation of the model. (c) 2012 Elsevier Inc. All rights reserved.
引用
收藏
页码:953 / 971
页数:19
相关论文
共 50 条
  • [1] Dynamic load balancing on heterogeneous multi-GPU systems
    Acosta, Alejandro
    Blanco, Vicente
    Almeida, Francisco
    COMPUTERS & ELECTRICAL ENGINEERING, 2013, 39 (08) : 2591 - 2602
  • [2] Benchmarking multi-GPU applications on modern multi-GPU integrated systems
    Bernaschi, Massimo
    Agostini, Elena
    Rossetti, Davide
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (14):
  • [3] Modelling Multi-GPU Systems
    Spampinato, Daniele G.
    Elster, Anne C.
    Natvig, Thorvald
    PARALLEL COMPUTING: FROM MULTICORES AND GPU'S TO PETASCALE, 2010, 19 : 562 - 569
  • [4] GreenMD: Energy-efficient Matrix Decomposition on Heterogeneous Multi-GPU Systems
    Zamani, Hadi
    Bhuyan, Laxmi
    Chen, Jieyang
    Chen, Zizhong
    ACM TRANSACTIONS ON PARALLEL COMPUTING, 2023, 10 (02)
  • [5] Understanding Scalability of Multi-GPU Systems
    Feng, Yuan
    Jeon, Hyeran
    15TH WORKSHOP ON GENERAL PURPOSE PROCESSING USING GPU, GPGPU 2023, 2023, : 36 - 37
  • [6] Heterogeneous Computational Model for Landform Attributes Representation on Multicore and Multi-GPU Systems
    Boratto, Murilo
    Alonso, Pedro
    Ramiro, Carla
    Barreto, Marcos
    PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE, ICCS 2012, 2012, 9 : 47 - 56
  • [7] GPU-Centered Parallel Model on Heterogeneous Multi-GPU Clusters
    Wang, Feng
    PROCEEDINGS OF 2012 2ND INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND NETWORK TECHNOLOGY (ICCSNT 2012), 2012, : 1865 - 1868
  • [8] Multi-GPU System Design with Memory Networks
    Kim, Gwangsun
    Lee, Minseok
    Jeong, Jiyun
    Kim, John
    2014 47TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2014, : 484 - 495
  • [9] Data Parallel Skeletons for GPU Clusters and Multi-GPU Systems
    Ernsting, Steffen
    Kuchen, Herbert
    APPLICATIONS, TOOLS AND TECHNIQUES ON THE ROAD TO EXASCALE COMPUTING, 2012, 22 : 509 - 518
  • [10] Suffix Array Construction on Multi-GPU Systems
    Bueren, Florian
    Juenger, Daniel
    Kobus, Robin
    Hundt, Christian
    Schmidt, Bertil
    HPDC'19: PROCEEDINGS OF THE 28TH INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, 2019, : 183 - 194