Fine-tuning 3D foundation models for geometric object retrieval

被引:0
|
作者
Van den Herrewegen, Jarne [1 ,2 ]
Tourwe, Tom [1 ]
Ovsjanikov, Maks [3 ]
Wyffels, Francis [2 ]
机构
[1] Oqton AI, Edegem, Belgium
[2] Ghent Univ Imec, AI & Robot Lab, IDLab AIRO, Zwijnaarde, Belgium
[3] Ecole Polytech, LIX, Palaiseau, France
来源
COMPUTERS & GRAPHICS-UK | 2024年 / 122卷
关键词
Object retrieval; Deep learning; 3D; Transfer learning; Foundation models; Self-supervised learning; NEURAL-NETWORK;
D O I
10.1016/j.cag.2024.103993
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Foundation models, such as ULIP-2 (Xue et al., 2023) recently projected forward the field of 3D deep learning. These models are trained with significantly more data and show superior representation learning capacity in many downstream tasks like 3D shape classification and few-shot part segmentation. A particular characteristic of the recent 3D foundation models is that they are typically multi-modal, , and involve image (2D) as well as caption (text) branches. This leads to an intricate interplay that benefits all modalities. At the same time, the nature of the 3D encoders alone, involved in these foundation models is not well-understood. Specifically, there is little analysis on the utility of both pre-trained 3D features provided by these models, or their capacity to adapt to new downstream 3D data. Furthermore, existing studies typically focus on label-oriented downstream tasks, such as shape classification, and ignore other critical applications, such as 3D content-based object retrieval. In this paper, we fill this gap and show, for the first time, how 3D foundation models can be leveraged for strong 3D-to-3D retrieval performance on seven different datasets, on par with state-of-the-art view-based architectures. We evaluate both the pre-trained foundation models, as well as their fine-tuned versions using downstream data. We compare supervised fine-tuning using classification labels against two self-supervised label-free fine-tuning methods. Importantly, we introduce and describe a methodology for fine-tuning, as we found this to be crucial to make transfer learning from 3D foundation models work in a stable manner.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] 3D OBJECT RETRIEVAL BY 3D CURVE MATCHING
    Feinen, Christian
    Czajkowska, Joanna
    Grzegorzek, Marcin
    Latecki, Longin Jan
    2014 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2014, : 2749 - 2753
  • [22] Exemplar Fine-Tuning for 3D Human Model Fitting Towards In-the-Wild 3D Human Pose Estimation
    Joo, Hanbyul
    Neverova, Natalia
    Vedaldi, Andrea
    2021 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2021), 2021, : 42 - 52
  • [23] A framework for 3D object segmentation and retrieval using local geometric surface features
    Dimou, Dimitrios
    Moustakas, Konstantinos
    2018 INTERNATIONAL CONFERENCE ON CYBERWORLDS (CW), 2018, : 102 - 107
  • [24] FINE-TUNING OF ECONOMIC-FORECASTING MODELS
    HUJER, R
    CREMER, R
    KNEPEL, H
    JAHRBUCHER FUR NATIONALOKONOMIE UND STATISTIK, 1979, 194 (01): : 41 - 70
  • [25] Improving fine-tuning in composite Higgs models
    Banerjee, Avik
    Bhattacharyya, Gautam
    Ray, Tirtha Sankar
    PHYSICAL REVIEW D, 2017, 96 (03)
  • [26] Quantum fine-tuning in stringy quintessence models
    Hertzberg, Mark R.
    Sandora, McCullen
    Trodden, Mark
    PHYSICS LETTERS B, 2019, 797
  • [27] The fine-tuning cost of the likelihood in SUSY models
    Ghilencea, D. M.
    Ross, G. G.
    NUCLEAR PHYSICS B, 2013, 868 (01) : 65 - 74
  • [28] Learning from models beyond fine-tuning
    Zheng, Hongling
    Shen, Li
    Tang, Anke
    Luo, Yong
    Hu, Han
    Du, Bo
    Wen, Yonggang
    Tao, Dacheng
    NATURE MACHINE INTELLIGENCE, 2025, 7 (01) : 6 - 17
  • [29] Efficient Fine-Tuning of BERT Models on the Edge
    Vucetic, Danilo
    Tayaranian, Mohammadreza
    Ziaeefard, Maryam
    Clark, James J.
    Meyer, Brett H.
    Gross, Warren J.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1838 - 1842
  • [30] Device-Edge Cooperative Fine-Tuning of Foundation Models as a 6G Service
    Wu, Hai
    Chen, Xu
    Huang, Kaibin
    IEEE WIRELESS COMMUNICATIONS, 2024, 31 (03) : 60 - 67