Pruning for Power: Optimizing Energy Efficiency in IoT with Neural Network Pruning

被引:0
|
作者
Widmann, Thomas [1 ]
Merkle, Florian [1 ]
Nocker, Martin [1 ]
Schoettle, Pascal [1 ]
机构
[1] MCI Management Ctr Innsbruck, Innsbruck, Austria
基金
奥地利科学基金会;
关键词
D O I
10.1007/978-3-031-34204-2_22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Internet of Things (IoT) has rapidly emerged as a crucial driver of the digital economy, generating massive amounts of data. Machine learning (ML) is an important technology to extract insights from the data generated by IoT devices. Deploying ML on low-power devices such as microcontroller units (MCUs) improves data protection, reduces bandwidth, and enables on-device data processing. However, the requirements of ML algorithms exceed the processing power, memory, and energy consumption capabilities of these devices. One solution to adapt ML networks to the limited capacities of MCUs is network pruning, the process of removing unnecessary connections or neurons from a neural network. In this work, we investigate the effect of unstructured and structured pruning methods on energy consumption. A series of experiments is conducted using a Raspberry Pi Pico to classify the FashionMNIST dataset with a LeNet-5-like convolutional neural network while applying unstructured magnitude and structured APoZ pruning approaches with various model compression rates from two to 64. We find that unstructured pruning out of the box has no effect on energy consumption, while structured pruning reduces energy consumption with increasing model compression. When structured pruning is applied to remove 75% of the model parameters, inference consumes 59.06% less energy, while the accuracy declines by 3.01 %. We further develop an adaption of the TensorFlow Lite framework that realizes the theoretical improvements for unstructured pruning, reducing the energy consumption by 37.59% with a decrease of only 1.58% in accuracy when 75% of the parameters are removed. Our results show that both approaches are feasible to significantly reduce the energy consumption of MCUs, leading to various possible sweet spots within the trade-off between accuracy and energy consumption.
引用
收藏
页码:251 / 263
页数:13
相关论文
共 50 条
  • [41] Fuse Devices for Pruning in Memristive Neural Network
    Kim, Tae-Hyeon
    Hong, Kyungho
    Kim, Sungjoon
    Park, Jinwoo
    Youn, Sangwook
    Lee, Jong-Ho
    Park, Byung-Gook
    Kim, Hyungjin
    Choi, Woo Young
    IEEE ELECTRON DEVICE LETTERS, 2023, 44 (03) : 520 - 523
  • [42] Neural Network Compression and Acceleration by Federated Pruning
    Pei, Songwen
    Wu, Yusheng
    Qiu, Meikang
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 173 - 183
  • [43] An Improved Pruning Algorithm for Fuzzy Neural Network
    Ai Fangju
    INFORMATION TECHNOLOGY APPLICATIONS IN INDUSTRY II, PTS 1-4, 2013, 411-414 : 2031 - 2036
  • [44] Neural Network Pruning for Biomedical Image Segmentation
    Jeong, Taehee
    Bollavararn, Manasa
    Delaye, Elliott
    Sirasao, Ashish
    MEDICAL IMAGING 2021: IMAGE-GUIDED PROCEDURES, ROBOTIC INTERVENTIONS, AND MODELING, 2021, 11598
  • [45] Thinning of convolutional neural network with mixed pruning
    Yang, Wenzhu
    Jin, Lilei
    Wang, Sile
    Cu, Zhenchao
    Chen, Xiangyang
    Chen, Liping
    IET IMAGE PROCESSING, 2019, 13 (05) : 779 - 784
  • [46] DRP:Discrete Rank Pruning for Neural Network
    Pei, Songwen
    Luo, Jie
    Liang, Sheng
    NETWORK AND PARALLEL COMPUTING, NPC 2022, 2022, 13615 : 168 - 179
  • [47] Crossbar-Aware Neural Network Pruning
    Liang, Ling
    Deng, Lei
    Zeng, Yueling
    Hu, Xing
    Ji, Yu
    Ma, Xin
    Li, Guoqi
    Xie, Yuan
    IEEE ACCESS, 2018, 6 : 58324 - 58337
  • [48] Optimizing Turning Parameters based on Correlation Pruning Neural Networks
    Wang Wu
    Zhang Yuan-min
    2009 IITA INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS ENGINEERING, PROCEEDINGS, 2009, : 319 - 322
  • [49] Zero-Keep Filter Pruning for Energy Efficient Deep Neural Network
    Woo, Yunhee
    Kim, Dongyoung
    Jeong, Jaemin
    Ko, Young-Woong
    Lee, Jeong-Gun
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1288 - 1292
  • [50] Revolutionizing neural network efficiency: introducing FPAC for filter pruning via attention consistency
    Mana, Suja Cherukullapurath
    Rajesh, Sudha
    Governor, Kalaiarasi
    Chandrasekaran, Hemalatha
    Murugesan, Kanipriya
    NEURAL COMPUTING & APPLICATIONS, 2023, 36 (2): : 639 - 652