More Photos are All You Need: Semi-Supervised Learning for Fine-Grained Sketch Based Image Retrieval

被引:44
|
作者
Bhunia, Ayan Kumar [1 ]
Chowdhury, Pinaki Nath [1 ,2 ]
Sain, Aneeshan [1 ,2 ]
Yang, Yongxin [1 ,2 ]
Xiang, Tao [1 ,2 ]
Song, Yi-Zhe [1 ,2 ]
机构
[1] Univ Surrey, CVSSP, SketchX, Guildford, Surrey, England
[2] iFlyTek Surrey Joint Res Ctr Artificial Intellige, Guildford, Surrey, England
关键词
D O I
10.1109/CVPR46437.2021.00423
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A fundamental challenge faced by existing Fine-Grained Sketch-Based Image Retrieval (FG-SBIR) models is the data scarcity - model performances are largely bottlenecked by the lack of sketch-photo pairs. Whilst the number of photos can be easily scaled, each corresponding sketch still needs to be individually produced. In this paper, we aim to mitigate such an upper-bound on sketch data, and study whether unlabelled photos alone (of which they are many) can be cultivated for performance gain. In particular, we introduce a novel semi-supervised framework for cross-modal retrieval that can additionally leverage large-scale unlabelled photos to account for data scarcity. At the center of our semi-supervision design is a sequential photo-to-sketch generation model that aims to generate paired sketches for unlabelled photos. Importantly, we further introduce a discriminator-guided mechanism to guide against unfaithful generation, together with a distillation loss-based regularizer to provide tolerance against noisy training samples. Last but not least, we treat generation and retrieval as two conjugate problems, where a joint learning procedure is devised for each module to mutually benefit from each other. Extensive experiments show that our semi-supervised model yields a significant performance boost over the state-of-the-art supervised alternatives, as well as existing methods that can exploit unlabelled photos for FG-SBIR.
引用
收藏
页码:4245 / 4254
页数:10
相关论文
共 50 条
  • [21] You'll Never Walk Alone: A Sketch and Text Duet for Fine-Grained Image Retrieval
    Koley, Subhadeep
    Bhunia, Ayan Kumar
    Sahli, Aneeshan
    Chowdhury, Pinaki Nath
    Xiang, Tao
    Song, Yi-Zhe
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 16509 - 16519
  • [22] Fine-Grained Semi-Supervised Labeling of Large Shape Collections
    Huang, Qi-Xing
    Su, Hao
    Guibas, Leonidas
    ACM TRANSACTIONS ON GRAPHICS, 2013, 32 (06):
  • [23] Co-Training Semi-Supervised Learning for Fine-Grained Air Quality Analysis
    Zhao, Yaning
    Wang, Li
    Zhang, Nannan
    Huang, Xiangwei
    Yang, Lunke
    Yang, Wenbiao
    ATMOSPHERE, 2023, 14 (01)
  • [24] Semi-supervised node classification via fine-grained graph auxiliary augmentation learning
    Lv, Jia
    Song, Kaikai
    Ye, Qiang
    Tian, Guangjian
    PATTERN RECOGNITION, 2023, 137
  • [25] Fine-grained Multi-label Sexism Classification Using Semi-supervised Learning
    Abburi, Harika
    Parikh, Pulkit
    Chhaya, Niyati
    Varma, Vasudeva
    WEB INFORMATION SYSTEMS ENGINEERING, WISE 2020, PT II, 2020, 12343 : 531 - 547
  • [26] An algorithm for semi-supervised learning in image retrieval
    Lu, K
    Zhao, JD
    Cai, D
    PATTERN RECOGNITION, 2006, 39 (04) : 717 - 720
  • [27] Image Retrieval Using Semi-Supervised Learning
    Zhu Songhao
    Liang Zhiwei
    PROCEEDINGS OF THE 29TH CHINESE CONTROL CONFERENCE, 2010, : 2924 - 2929
  • [28] Fine-grained interactive attention learning for semi-supervised white blood cell classification
    Ha, Yan
    Du, Zeyu
    Tian, Junfeng
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2022, 75
  • [29] Multi-feature fusion for fine-grained sketch-based image retrieval
    Zhu, Ming
    Zhao, Chen
    Wang, Nian
    Tang, Jun
    Yan, Pu
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 82 (24) : 38067 - 38076
  • [30] Deep Multimodal Embedding Model for Fine-grained Sketch-based Image Retrieval
    Huang, Fei
    Cheng, Yong
    Jin, Cheng
    Zhang, Yuejie
    Zhang, Tao
    SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 929 - 932