Efficient Low-rank Backpropagation for Vision Transformer Adaptation

被引:0
|
作者
Yang, Yuedong [1 ]
Chiang, Hung-Yueh [1 ]
Li, Guihong [1 ]
Marculescu, Diana [1 ]
Marculescu, Radu [1 ]
机构
[1] Univ Texas Austin, Chandra Family Dept Elect & Comp Engn, Austin, TX 78712 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The increasing scale of vision transformers (ViT) has made the efficient fine-tuning of these large models for specific needs a significant challenge in various applications. This issue originates from the computationally demanding matrix multiplications required during the backpropagation process through linear layers in ViT. In this paper, we tackle this problem by proposing a new Low-rank Back-Propagation viaWalsh-Hadamard Transformation (LBP-WHT) method. Intuitively, LBP-WHT projects the gradient into a low-rank space and carries out backpropagation. This approach substantially reduces the computation needed for adapting ViT, as matrix multiplication in the low-rank space is far less resource-intensive. We conduct extensive experiments with different models (ViT, hybrid convolution-ViT model) on multiple datasets to demonstrate the effectiveness of our method. For instance, when adapting an EfficientFormer-L1 model on CIFAR100, our LBP-WHT achieves 10.4% higher accuracy than the state-of-the-art baseline, while requiring 9 MFLOPs less computation. As the first work to accelerate ViT adaptation with low-rank backpropagation, our LBP-WHT method is complementary to many prior efforts and can be combined with them for better performance. Code: https://github.com/SLDGroup/LBP- WHT
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Provably Efficient Algorithm for Nonstationary Low-Rank MDPs
    Cheng, Yuan
    Yang, Jing
    Liang, Yingbin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [32] Efficient low-rank solution of generalized Lyapunov equations
    Shank, Stephen D.
    Simoncini, Valeria
    Szyld, Daniel B.
    NUMERISCHE MATHEMATIK, 2016, 134 (02) : 327 - 342
  • [33] COMPACTER: Efficient Low-Rank Hypercomplex Adapter Layers
    Mahabadi, Rabeeh Karimi
    Henderson, James
    Ruder, Sebastian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [34] Efficient Optimization for Low-Rank Integrated Bilinear Classifiers
    Kobayashi, Takumi
    Otsu, Nobuyuki
    COMPUTER VISION - ECCV 2012, PT II, 2012, 7573 : 474 - 487
  • [35] A Fast and Efficient Algorithm for Low-rank Approximation of a Matrix
    Nguyen, Nam H.
    Do, Thong T.
    Tran, Trac D.
    STOC'09: PROCEEDINGS OF THE 2009 ACM SYMPOSIUM ON THEORY OF COMPUTING, 2009, : 215 - 224
  • [36] An efficient Kalman filter for the identification of low-rank systems
    Dogariu, Laura-Maria
    Paleologu, Constantin
    Benesty, Jacob
    Ciochina, Silviu
    SIGNAL PROCESSING, 2020, 166
  • [37] Efficient methods for grouping vectors into low-rank clusters
    Rangan, Aaditya V.
    JOURNAL OF COMPUTATIONAL PHYSICS, 2011, 230 (14) : 5684 - 5703
  • [38] Low-Rank Adaptation of Pre-Trained Large Vision Models for Improved Lung Nodule Malignancy Classification
    Veasey, Benjamin P.
    Amini, Amir A.
    IEEE OPEN JOURNAL OF ENGINEERING IN MEDICINE AND BIOLOGY, 2025, 6 : 296 - 304
  • [39] Efficient Low-Rank Spectrotemporal Decomposition using ADMM
    Schamberg, Gabriel
    Ba, Demba
    Wagner, Mark
    Coleman, Todd
    2016 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP (SSP), 2016,
  • [40] StruNet: Perceptual and low-rank regularized transformer for medical image denoising
    Ma, Yuhui
    Yan, Qifeng
    Liu, Yonghuai
    Liu, Jiang
    Zhang, Jiong
    Zhao, Yitian
    MEDICAL PHYSICS, 2023, 50 (12) : 7654 - 7669