Efficient Low-rank Backpropagation for Vision Transformer Adaptation

被引:0
|
作者
Yang, Yuedong [1 ]
Chiang, Hung-Yueh [1 ]
Li, Guihong [1 ]
Marculescu, Diana [1 ]
Marculescu, Radu [1 ]
机构
[1] Univ Texas Austin, Chandra Family Dept Elect & Comp Engn, Austin, TX 78712 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The increasing scale of vision transformers (ViT) has made the efficient fine-tuning of these large models for specific needs a significant challenge in various applications. This issue originates from the computationally demanding matrix multiplications required during the backpropagation process through linear layers in ViT. In this paper, we tackle this problem by proposing a new Low-rank Back-Propagation viaWalsh-Hadamard Transformation (LBP-WHT) method. Intuitively, LBP-WHT projects the gradient into a low-rank space and carries out backpropagation. This approach substantially reduces the computation needed for adapting ViT, as matrix multiplication in the low-rank space is far less resource-intensive. We conduct extensive experiments with different models (ViT, hybrid convolution-ViT model) on multiple datasets to demonstrate the effectiveness of our method. For instance, when adapting an EfficientFormer-L1 model on CIFAR100, our LBP-WHT achieves 10.4% higher accuracy than the state-of-the-art baseline, while requiring 9 MFLOPs less computation. As the first work to accelerate ViT adaptation with low-rank backpropagation, our LBP-WHT method is complementary to many prior efforts and can be combined with them for better performance. Code: https://github.com/SLDGroup/LBP- WHT
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Hydra: Multi-head low-rank adaptation for parameter efficient fine-tuning
    Kim, Sanghyeon
    Yang, Hyunmo
    Kim, Yunghyun
    Hong, Youngjoon
    Park, Eunbyung
    NEURAL NETWORKS, 2024, 178
  • [22] Dropout Mixture Low-Rank Adaptation for Visual Parameters-Efficient Fine-Tuning
    Fang, Zhengyi
    Wang, Yue
    Yi, Ran
    Ma, Lizhuang
    COMPUTER VISION-ECCV 2024, PT VII, 2025, 15065 : 369 - 386
  • [23] Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning
    Hu, Yahao
    Xie, Yifei
    Wang, Tianfeng
    Chen, Man
    Pan, Zhisong
    MATHEMATICS, 2023, 11 (20)
  • [24] Low-rank lottery tickets: finding efficient low-rank neural networks via matrix differential equations
    Schotthoefer, Steffen
    Zangrando, Emanuele
    Kusch, Jonas
    Ceruti, Gianluca
    Tudisco, Francesco
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [25] Uniform Low-Rank Representation for Unsupervised Visual Domain Adaptation
    Liu, Pengcheng
    Yang, Peipei
    Huang, Kaiqi
    Tan, Tieniu
    PROCEEDINGS 3RD IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION ACPR 2015, 2015, : 216 - 220
  • [26] LOW-RANK PLUS DIAGONAL ADAPTATION FOR DEEP NEURAL NETWORKS
    Zhao, Yong
    Li, Jinyu
    Gong, Yifan
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 5005 - 5009
  • [27] Low-Rank Adaptation Applied to Multiclass Diabetic Retinopathy Classification
    Ferreira-Caballero, Sebastian
    Pinto-Roa, Diego P.
    Noguera, Jose Luis Vazquez
    Ayala, Jordan
    Gardel-Sotomayor, Pedro E.
    Perez-Estigarribia, Pastor
    2024 L LATIN AMERICAN COMPUTER CONFERENCE, CLEI 2024, 2024,
  • [28] LoRA-NIR: Low-Rank Adaptation of Vision Transformers for Remote Sensing With Near-Infrared Imagery
    Ulku, Irem
    Tanriover, O. Ozgur
    Akagunduz, Erdem
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21
  • [29] Efficient video hashing based on low-rank frames
    Chen, Zhenhai
    Tang, Zhenjun
    Zhang, Xinpeng
    Sun, Ronghai
    Zhang, Xianquan
    IET IMAGE PROCESSING, 2022, 16 (02) : 344 - 355
  • [30] Efficient low-rank solution of generalized Lyapunov equations
    Stephen D. Shank
    Valeria Simoncini
    Daniel B. Szyld
    Numerische Mathematik, 2016, 134 : 327 - 342