Pro-Tuning: Unified Prompt Tuning for Vision Tasks

被引:7
|
作者
Nie, Xing [1 ,2 ]
Ni, Bolin [1 ,2 ]
Chang, Jianlong [3 ]
Meng, Gaofeng [1 ,2 ,4 ]
Huo, Chunlei [5 ,6 ]
Xiang, Shiming [1 ,2 ]
Tian, Qi [3 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence S, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Huawei Cloud & AI, Beijing 100095, Peoples R China
[4] HK Inst Sci & Innovat, CAS Ctr Artificial Intelligence & Robot, Hong Kong, Peoples R China
[5] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
[6] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
关键词
Task analysis; Adaptation models; Tuning; Computational modeling; Transformers; Visualization; Training; Prompt-based learning; representation learning; task-specific knowledge; transfer learning;
D O I
10.1109/TCSVT.2023.3327605
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In computer vision, fine-tuning is the de-facto approach to leverage pre-trained vision models to perform downstream tasks. However, deploying it in practice is quite challenging, due to adopting parameter inefficient global update and heavily relying on high-quality downstream data. Recently, prompt-based learning, which adds the task-relevant prompt to adapt the pre-trained models to downstream tasks, has drastically boosted the performance of many natural language downstream tasks. In this work, we extend this notable transfer ability benefited from prompt into vision models as an alternative to fine-tuning. To this end, we propose parameter-efficient Prompt tuning (Pro-tuning) to adapt diverse frozen pre-trained models to a wide variety of downstream vision tasks. The key to Pro-tuning is prompt-based tuning, i.e., learning task-specific vision prompts for downstream input images with the pre-trained model frozen. By only training a small number of additional parameters, Pro-tuning can generate compact and robust downstream models both for CNN-based and transformer-based network architectures. Comprehensive experiments evidence that the proposed Pro-tuning outperforms fine-tuning on a broad range of vision tasks and scenarios, including image classification (under generic objects, class imbalance, image corruption, natural adversarial examples, and out-of-distribution generalization), and dense prediction tasks such as object detection and semantic segmentation.
引用
收藏
页码:4653 / 4667
页数:15
相关论文
共 50 条
  • [41] Universal Prompt Tuning for Graph Neural Networks
    Fang, Taoran
    Zhang, Yunchao
    Yang, Yang
    Wang, Chunping
    Chen, Lei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [42] Consistent Prompt Tuning for Generalized Category Discovery
    Yang, Muli
    Yin, Jie
    Gu, Yanan
    Deng, Cheng
    Zhang, Hanwang
    Zhu, Hongyuan
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [43] Knowledge Prompt-tuning for Sequential Recommendation
    Zhai, Jianyang
    Zheng, Xiawu
    Wang, Chang-Dong
    Li, Hui
    Tian, Yonghong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 6451 - 6461
  • [44] Point Prompt Tuning for Temporally Language Grounding
    Zeng, Yawen
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2003 - 2007
  • [45] ASR MODEL ADAPTATION WITH DOMAIN PROMPT TUNING
    Zou, Pengpeng
    Ye, Jianhao
    Zhou, Hongbin
    2024 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING, IALP 2024, 2024, : 406 - 410
  • [46] Continual Prompt Tuning for Dialog State Tracking
    Zhu, Qi
    Li, Bing
    Mi, Fei
    Zhu, Xiaoyan
    Huang, Minlie
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1124 - 1137
  • [47] Prompt Tuning with Contradictory Intentions for Sarcasm Recognition
    Liu, Yiyi
    Zhang, Ruqing
    Fan, Yixing
    Guo, Jiafeng
    Cheng, Xueqi
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 328 - 339
  • [48] On Transferability of Prompt Tuning for Natural Language Processing
    Su, Yusheng
    Wang, Xiaozhi
    Qin, Yujia
    Chan, Chi-Min
    Lin, Yankai
    Wang, Huadong
    Wen, Kaiyue
    Liu, Zhiyuan
    Li, Peng
    Li, Juanzi
    Hou, Lei
    Sun, Maosong
    Zhou, Jie
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3949 - 3969
  • [49] Hard Sample Aware Prompt-Tuning
    Xu, Yuanjian
    An, Qi
    Zhang, Jiahuan
    Li, Peng
    Nie, Zaiqing
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 12356 - 12369
  • [50] PTAU: Prompt Tuning for Attributing Unanswerable Questions
    Liao, Jinzhi
    Zhao, Xiang
    Zheng, Jianming
    Li, Xinyi
    Cai, Fei
    Tang, Jiuyang
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1219 - 1229