Accurate, Low-latency, Efficient SAR Automatic Target Recognition on FPGA

被引:7
|
作者
Zhang, Bingyi [1 ]
Kannan, Rajgopal [2 ]
Prasanna, Viktor [1 ]
Busart, Carl [2 ]
机构
[1] Univ Southern Calif, Los Angeles, CA 90007 USA
[2] DEVCOM US Army Res Lab, Washington, DC USA
基金
美国国家科学基金会;
关键词
SAR ATR; graph neural network (GNN); hardware architecture; NETWORK;
D O I
10.1109/FPL57034.2022.00013
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Synthetic aperture radar (SAR) automatic target recognition (ATR) is the key technique for remote-sensing image recognition. The state-of-the-art convolutional neural networks (CNNs) for SAR ATR suffer from high computation cost and large memory footprint, making them unsuitable to be deployed on resource-limited platforms, such as small/micro satellites. In this paper, we propose a comprehensive GNN-based model-architecture co-design on FPGA to address the above issues. Model design: we design a novel graph neural network (GNN) for SAR ATR. The proposed GNN model incorporates GraphSAGE layer operators and attention mechanism, achieving comparable accuracy as the state-of-the-art work with near 1/100 computation cost. Then, we propose a pruning approach including weight pruning and input pruning. While weight pruning through lasso regression reduces most parameters without accuracy drop, input pruning eliminates most input pixels with negligible accuracy drop. Architecture design: to fully unleash the computation parallelism within the proposed model, we develop a novel unified hardware architecture that can execute various computation kernels (feature aggregation, feature transformation, graph pooling). The proposed hardware design adopts the Scatter-Gather paradigm to efficiently handle the irregular computation patterns of various computation kernels. We deploy the proposed design on an embedded FPGA (AMD Xilinx ZCU104) and evaluate the performance using MSTAR dataset. Compared with the state-of-the-art CNNs, the proposed GNN achieves comparable accuracy with 1/3258 computation cost and 1/83 model size. Compared with the state-of-the-art CPU/GPU, our FPGA accelerator achieves 14.8x/2.5x speedup (latency) and is 62x/39x more energy efficient.
引用
收藏
页码:1 / 8
页数:8
相关论文
共 50 条
  • [1] Low-latency transformer model for streaming automatic speech recognition
    Miao, Haoran
    Cheng, Gaofeng
    Zhang, Pengyuan
    ELECTRONICS LETTERS, 2022, 58 (01) : 44 - 46
  • [2] ADERLNet: Adaptive Denoising Enhancement Representation Learning for Low-Latency and High-Accurate Target Detection on SAR Sensors
    Yang, Yahu
    Ma, Chen
    Li, Guoqi
    Wang, Yu
    Zhang, Li
    Song, Shenmin
    IEEE SENSORS JOURNAL, 2024, 24 (05) : 6430 - 6450
  • [3] An Efficient and Accurate Link Latency Monitoring Method for Low-Latency Software-Defined Networks
    Liao, Lingxia
    Leung, Victor C. M.
    Chen, Min
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2019, 68 (02) : 377 - 391
  • [4] Low-Latency Distance Protective Relay on FPGA
    Wang, Yifan
    Dinavahi, Venkata
    IEEE TRANSACTIONS ON SMART GRID, 2014, 5 (02) : 896 - 905
  • [5] Low-Latency Distance Protective Relay on FPGA
    Wang, Yifan
    Dinavahi, Venkata
    2014 IEEE PES GENERAL MEETING - CONFERENCE & EXPOSITION, 2014,
  • [6] Biologically inspired heterogeneous learning for accurate, efficient and low-latency neural network
    Wang, Bo
    Zhang, Yuxuan
    Li, Hongjue
    Dou, Hongkun
    Guo, Yuchen
    Deng, Yue
    NATIONAL SCIENCE REVIEW, 2025, 12 (01)
  • [7] FPGA Implementation of Low-Latency Recursive Median Filter
    Peng, Bo
    Zhou, Yuzhu
    Li, Qiang
    Lin, Maosong
    Weng, Jiankui
    Zeng, Qiang
    2022 21ST INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2022), 2022, : 312 - 318
  • [8] Low-Latency CoAP Processing in FPGA for the Internet of Things
    Brasilino, Lucas R. B.
    Swany, Martin
    2019 INTERNATIONAL CONFERENCE ON INTERNET OF THINGS (ITHINGS) AND IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) AND IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) AND IEEE SMART DATA (SMARTDATA), 2019, : 1057 - 1064
  • [9] LOW-LATENCY CONNECTED COMPONENT LABELING USING AN FPGA
    Ito, Yasuaki
    Nakano, Koji
    INTERNATIONAL JOURNAL OF FOUNDATIONS OF COMPUTER SCIENCE, 2010, 21 (03) : 405 - 425
  • [10] Low-Latency Modular Packet Header Parser for FPGA
    Pus, Viktor
    Kekely, Lukas
    Korenek, Jan
    PROCEEDINGS OF THE EIGHTH ACM/IEEE SYMPOSIUM ON ARCHITECTURES FOR NETWORKING AND COMMUNICATIONS SYSTEMS (ANCS'12), 2012, : 77 - 78