Instance-aware Dynamic Prompt Tuning for Pre-trained Point Cloud Models

被引:6
|
作者
Zha, Yaohua [1 ]
Wang, Jinpeng [1 ]
Dai, Tao [2 ]
Bin Chen [3 ]
Wang, Zhi [1 ]
Xia, Shu-Tao [4 ]
机构
[1] Tsinghua Univ, Tsinghua Shenzhen Int Grad Sch, Beijing, Peoples R China
[2] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen, Peoples R China
[3] Harbin Inst Technol, Harbin, Peoples R China
[4] Shenzhen Res Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01302
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained point cloud models have found extensive applications in 3D understanding tasks like object classification and part segmentation. However, the prevailing strategy of full fine-tuning in downstream tasks leads to large per-task storage overhead for model parameters, which limits the efficiency when applying large-scale pre-trained models. Inspired by the recent success of visual prompt tuning (VPT), this paper attempts to explore prompt tuning on pre-trained point cloud models, to pursue an elegant balance between performance and parameter efficiency. We find while instance-agnostic static prompting, e.g. VPT, shows some efficacy in downstream transfer, it is vulnerable to the distribution diversity caused by various types of noises in real-world point cloud data. To conquer this limitation, we propose a novel Instance-aware Dynamic Prompt Tuning (IDPT) strategy for pre-trained point cloud models. The essence of IDPT is to develop a dynamic prompt generation module to perceive semantic prior features of each point cloud instance and generate adaptive prompt tokens to enhance the model's robustness. Notably, extensive experiments demonstrate that IDPT outperforms full fine-tuning in most tasks with a mere 7% of the trainable parameters, providing a promising solution to parameter-efficient learning for pre-trained point cloud models. Code is available at https://github.com/zyh16143998882/ICCV23-IDPT.
引用
收藏
页码:14115 / 14124
页数:10
相关论文
共 50 条
  • [21] Personalised soft prompt tuning in pre-trained language models: Bridging multitask transfer learning and crowdsourcing learning
    Tian, Zeshu
    Zhang, Hongli
    Wang, Yan
    KNOWLEDGE-BASED SYSTEMS, 2024, 305
  • [22] SDPT: Synchronous Dual Prompt Tuning for Fusion-Based Visual-Language Pre-trained Models
    Zhou, Yang
    Wu, Yongjian
    Saiyin, Jiya
    Wei, Bingzheng
    Lai, Maode
    Chang, Eric
    Xu, Yan
    COMPUTER VISION - ECCV 2024, PT XLIX, 2025, 15107 : 340 - 356
  • [23] Span Fine-tuning for Pre-trained Language Models
    Bao, Rongzhou
    Zhang, Zhuosheng
    Zhao, Hai
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 1970 - 1979
  • [24] Dynamic Knowledge Distillation for Pre-trained Language Models
    Li, Lei
    Lin, Yankai
    Ren, Shuhuai
    Li, Peng
    Zhou, Jie
    Sun, Xu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 379 - 389
  • [25] DictPrompt: Comprehensive dictionary-integrated prompt tuning for pre-trained language model
    Cao, Rui
    Wang, Yihao
    Gao, Ling
    Yang, Meng
    KNOWLEDGE-BASED SYSTEMS, 2023, 273
  • [26] Matching tasks to objectives: Fine-tuning and prompt-tuning strategies for encoder-decoder pre-trained language models
    Pouramini, Ahmad
    Faili, Hesham
    APPLIED INTELLIGENCE, 2024, 54 (20) : 9783 - 9810
  • [27] Generative Visual Prompt: Unifying Distributional Control of Pre-Trained Generative Models
    Wu, Chen Henry
    Motamed, Saman
    Srivastava, Shaunak
    De la Torre, Fernando
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [28] Co2PT: Mitigating Bias in Pre-trained Language Models through Counterfactual Contrastive Prompt Tuning
    Dong, Xiangjue
    Zhu, Ziwei
    Wang, Zhuoer
    Teleki, Maria
    Caverlee, James
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5859 - 5871
  • [29] Bi-tuning: Efficient Transfer from Pre-trained Models
    Zhong, Jincheng
    Ma, Haoyu
    Wang, Ximei
    Kou, Zhi
    Long, Mingsheng
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT V, 2023, 14173 : 357 - 373
  • [30] Fine-Tuning Pre-Trained Language Models with Gaze Supervision
    Deng, Shuwen
    Prasse, Paul
    Reich, David R.
    Scheffer, Tobias
    Jager, Lena A.
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 217 - 224