Instance-aware Dynamic Prompt Tuning for Pre-trained Point Cloud Models

被引:6
|
作者
Zha, Yaohua [1 ]
Wang, Jinpeng [1 ]
Dai, Tao [2 ]
Bin Chen [3 ]
Wang, Zhi [1 ]
Xia, Shu-Tao [4 ]
机构
[1] Tsinghua Univ, Tsinghua Shenzhen Int Grad Sch, Beijing, Peoples R China
[2] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen, Peoples R China
[3] Harbin Inst Technol, Harbin, Peoples R China
[4] Shenzhen Res Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01302
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained point cloud models have found extensive applications in 3D understanding tasks like object classification and part segmentation. However, the prevailing strategy of full fine-tuning in downstream tasks leads to large per-task storage overhead for model parameters, which limits the efficiency when applying large-scale pre-trained models. Inspired by the recent success of visual prompt tuning (VPT), this paper attempts to explore prompt tuning on pre-trained point cloud models, to pursue an elegant balance between performance and parameter efficiency. We find while instance-agnostic static prompting, e.g. VPT, shows some efficacy in downstream transfer, it is vulnerable to the distribution diversity caused by various types of noises in real-world point cloud data. To conquer this limitation, we propose a novel Instance-aware Dynamic Prompt Tuning (IDPT) strategy for pre-trained point cloud models. The essence of IDPT is to develop a dynamic prompt generation module to perceive semantic prior features of each point cloud instance and generate adaptive prompt tokens to enhance the model's robustness. Notably, extensive experiments demonstrate that IDPT outperforms full fine-tuning in most tasks with a mere 7% of the trainable parameters, providing a promising solution to parameter-efficient learning for pre-trained point cloud models. Code is available at https://github.com/zyh16143998882/ICCV23-IDPT.
引用
收藏
页码:14115 / 14124
页数:10
相关论文
共 50 条
  • [41] Dialog summarization for software collaborative platform via tuning pre-trained models
    Fan, Guodong
    Chen, Shizhan
    Wu, Hongyue
    Gao, Cuiyun
    Xiao, Jianmao
    Xue, Xiao
    Feng, Zhiyong
    JOURNAL OF SYSTEMS AND SOFTWARE, 2023, 204
  • [42] Ranking and Tuning Pre-trained Models: A New Paradigm for Exploiting Model Hubs
    You, Kaichao
    Liu, Yong
    Zhang, Ziyang
    Wang, Jianmin
    Jordan, Michael I.
    Long, Mingsheng
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [43] Semantic Importance-Aware Communications Using Pre-Trained Language Models
    Guo, Shuaishuai
    Wang, Yanhu
    Li, Shujing
    Saeed, Nasir
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (09) : 2328 - 2332
  • [44] Latency-Aware Generative Semantic Communications With Pre-Trained Diffusion Models
    Qiao, Li
    Mashhadi, Mahdi Boloursaz
    Gao, Zhen
    Foh, Chuan Heng
    Xiao, Pei
    Bennis, Mehdi
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2024, 13 (10) : 2652 - 2656
  • [45] Exploiting Syntactic Information to Boost the Fine-tuning of Pre-trained Models
    Liu, Chaoming
    Zhu, Wenhao
    Zhang, Xiaoyu
    Zhai, Qiuhong
    2022 IEEE 46TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2022), 2022, : 575 - 582
  • [46] InstaSAM: Instance-Aware Segment Any Nuclei Model with Point Annotations
    Nam, Siwoo
    Namgung, Hyun
    Jeong, Jaehoon
    Luna, Miguel
    Kim, Soopil
    Chikontwe, Philip
    Park, Sang Hyun
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT IV, 2024, 15004 : 232 - 242
  • [47] Learning instance-aware object detection using determinantal point processes
    Kim, Nuri
    Lee, Donghoon
    Oh, Songhwai
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2020, 201
  • [48] Self-supervised Bidirectional Prompt Tuning for Entity-enhanced Pre-trained Language Model
    Zou, Jiaxin
    Xu, Xianghong
    Hou, Jiawei
    Yang, Qiang
    Zheng, Hai-Tao
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [49] Refining Pre-Trained Motion Models
    Sun, Xinglong
    Harley, Adam W.
    Guibas, Leonidas J.
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 4932 - 4938
  • [50] Efficiently Robustify Pre-Trained Models
    Jain, Nishant
    Behl, Harkirat
    Rawat, Yogesh Singh
    Vineet, Vibhav
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5482 - 5492