Focus Your Attention when Few-Shot Classification

被引:0
|
作者
Wang, Haoqing [1 ]
Jie, Shibo [1 ]
Deng, Zhi-Hong [1 ]
机构
[1] Peking Univ, Sch Intelligence Sci & Technol, Beijing, Peoples R China
关键词
NEURAL-NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since many pre-trained vision transformers emerge and provide strong representation for various downstream tasks, we aim to adapt them to few-shot image classification tasks in this work. The input images typically contain multiple entities. The model may not focus on the class-related entities for the current few-shot task, even with fine-tuning on support samples, and the noise information from the class-independent entities harms performance. To this end, we first propose a method that uses the attention and gradient information to automatically locate the positions of key entities in the support images, denoted as position prompts. Then we employ the cross-entropy loss between their many-hot presentation and the attention logits to optimize the model to focus its attention on the key entities during fine-tuning. This ability then can generalize to the query samples. Our method is applicable to different vision transformers (e.g., columnar or pyramidal ones), and also to different pre-training ways (e.g., single-modal or vision-language pre-training). Extensive experiments show that our method can improve the performance of full or parameter-efficient fine-tuning methods on few-shot tasks. Code is available at https://github.com/Haoqing-Wang/FORT.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Cross Attention Network for Few-shot Classification
    Hou, Ruibing
    Chang, Hong
    Ma, Bingpeng
    Shan, Shiguang
    Chen, Xilin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Multiscale attention for few-shot image classification
    Zhou, Tong
    Dong, Changyin
    Song, Junshu
    Zhang, Zhiqiang
    Wang, Zhen
    Chang, Bo
    Chen, Dechun
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (02)
  • [3] Few-shot classification with Fork Attention Adapter
    Sun, Jieqi
    Li, Jian
    PATTERN RECOGNITION, 2024, 156
  • [4] Few-Shot Few-Shot Learning and the role of Spatial Attention
    Lifchitz, Yann
    Avrithis, Yannis
    Picard, Sylvaine
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2693 - 2700
  • [5] Selectively Augmented Attention Network for Few-Shot Image Classification
    Li, Xiaoxu
    Wang, Xiangyang
    Zhu, Rui
    Ma, Zhanyu
    Cao, Jie
    Xue, Jing-Hao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1180 - 1192
  • [6] Transductive Graph-Attention Network for Few-shot Classification
    Pan, Lili
    Liu, Weifeng
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 190 - 195
  • [7] Few-Shot Scene Classification with Attention Mechanism in Remote Sensing
    Zhang, Duona
    Zhao, Hongjia
    Lu, Yuanyao
    Cui, Jian
    Zhang, Baochang
    Computer Engineering and Applications, 2024, 60 (04) : 173 - 182
  • [8] Attribute- and attention-guided few-shot classification
    Ziquan Wang
    Hui Li
    Zikai Zhang
    Feng Chen
    Jia Zhai
    Multimedia Systems, 2024, 30
  • [9] Total Relation Network with Attention for Few-Shot Image Classification
    Li X.-X.
    Liu Z.-Y.
    Wu J.-J.
    Cao J.
    Ma Z.-Y.
    Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (02): : 371 - 384
  • [10] Intra-inter channel attention for few-shot classification
    Yang L.
    Zhang T.
    Wang Y.
    Gu X.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2023, 31 (21): : 3145 - 3155