Fine-tuning 3D foundation models for geometric object retrieval

被引:0
|
作者
Van den Herrewegen, Jarne [1 ,2 ]
Tourwe, Tom [1 ]
Ovsjanikov, Maks [3 ]
Wyffels, Francis [2 ]
机构
[1] Oqton AI, Edegem, Belgium
[2] Ghent Univ Imec, AI & Robot Lab, IDLab AIRO, Zwijnaarde, Belgium
[3] Ecole Polytech, LIX, Palaiseau, France
来源
COMPUTERS & GRAPHICS-UK | 2024年 / 122卷
关键词
Object retrieval; Deep learning; 3D; Transfer learning; Foundation models; Self-supervised learning; NEURAL-NETWORK;
D O I
10.1016/j.cag.2024.103993
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Foundation models, such as ULIP-2 (Xue et al., 2023) recently projected forward the field of 3D deep learning. These models are trained with significantly more data and show superior representation learning capacity in many downstream tasks like 3D shape classification and few-shot part segmentation. A particular characteristic of the recent 3D foundation models is that they are typically multi-modal, , and involve image (2D) as well as caption (text) branches. This leads to an intricate interplay that benefits all modalities. At the same time, the nature of the 3D encoders alone, involved in these foundation models is not well-understood. Specifically, there is little analysis on the utility of both pre-trained 3D features provided by these models, or their capacity to adapt to new downstream 3D data. Furthermore, existing studies typically focus on label-oriented downstream tasks, such as shape classification, and ignore other critical applications, such as 3D content-based object retrieval. In this paper, we fill this gap and show, for the first time, how 3D foundation models can be leveraged for strong 3D-to-3D retrieval performance on seven different datasets, on par with state-of-the-art view-based architectures. We evaluate both the pre-trained foundation models, as well as their fine-tuned versions using downstream data. We compare supervised fine-tuning using classification labels against two self-supervised label-free fine-tuning methods. Importantly, we introduce and describe a methodology for fine-tuning, as we found this to be crucial to make transfer learning from 3D foundation models work in a stable manner.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Fine-Tuning Regression Forests Votes for Object Alignment in the Wild
    Yang, Heng
    Patras, Ioannis
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (02) : 619 - 631
  • [42] FINE-TUNING RHEOLOGICAL PROPERTIES OF 3D BIOPRINTABLE HYBRID HYDROGEL BY CONTROLLING COMPONENT SOLID CONTENT
    Habib, Ahasan
    Khoda, Bashir
    Tuladhar, Slesha
    Sarah, Rokeya
    PROCEEDINGS OF ASME 2024 19TH INTERNATIONAL MANUFACTURING SCIENCE AND ENGINEERING CONFERENCE, MSEC2024, VOL 1, 2024,
  • [43] An initial prediction and fine-tuning model based on improving GCN for 3D human motion prediction
    He, Zhiquan
    Zhang, Lujun
    Wang, Hengyou
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2023, 17
  • [44] A Fast and Accurate 3D Fine-Tuning Convolutional Neural Network for Alzheimer's Disease Diagnosis
    Tang, Hao
    Yao, Erlin
    Tan, Guangming
    Guo, Xiuhua
    ARTIFICIAL INTELLIGENCE (ICAI 2018), 2018, 888 : 115 - 126
  • [45] Threshold Fine-Tuning and 3D Characterisation of Porous Media Using X-ray Nanotomography
    Ostadi, Hossein
    Rama, Pratap
    Liu, Yu
    Chen, Rui
    Zhang, Xiaoxian
    Jiang, Kyle
    CURRENT NANOSCIENCE, 2010, 6 (02) : 226 - 231
  • [46] Fine-tuning Dynamic Cross-linking for Enhanced 3D Bioprinting of Hyaluronic Acid Hydrogels
    Tavakoli, Shima
    Krishnan, Nithiyanandan
    Mokhtari, Hamidreza
    Oommen, Oommen P.
    Varghese, Oommen P.
    ADVANCED FUNCTIONAL MATERIALS, 2024, 34 (04)
  • [47] Equi-Tuning: Group Equivariant Fine-Tuning of Pretrained Models
    Basu, Sourya
    Sattigeri, Prasanna
    Ramamurthy, Karthikeyan Natesan
    Chenthamarakshan, Vijil
    Varshney, Kush R.
    Varshney, Lav R.
    Das, Payel
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 6788 - 6796
  • [48] 3D models retrieval and indexing
    Mahmoudi, S
    Daoudi, M
    PROCEEDINGS OF THE 6TH JOINT CONFERENCE ON INFORMATION SCIENCES, 2002, : 972 - 975
  • [49] Phased Instruction Fine-Tuning for Large Language Models
    Pang, Wei
    Zhou, Chuan
    Zhou, Xiao-Hua
    Wang, Xiaojie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5735 - 5748
  • [50] How fine can fine-tuning be? Learning efficient language models
    Radiya-Dixit, Evani
    Wang, Xin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2435 - 2442