CIGAR: Cross-Modality Graph Reasoning for Domain Adaptive Object Detection

被引:17
|
作者
Liu, Yabo [1 ,2 ]
Wang, Jinghua [1 ]
Huang, Chao [3 ]
Wang, Yaowei [2 ]
Xu, Yong [1 ,2 ]
机构
[1] Harbin Inst Technol, Shenzhen, Peoples R China
[2] Peng Cheng Lab, Shenzhen, Peoples R China
[3] Sun Yat Sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen, Peoples R China
关键词
D O I
10.1109/CVPR52729.2023.02277
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised domain adaptive object detection (UDA-OD) aims to learn a detector by generalizing knowledge from a labeled source domain to an unlabeled target domain. Though the existing graph-based methods for UDA-OD perform well in some cases, they cannot learn a proper node set for the graph. In addition, these methods build the graph solely based on the visual features and do not consider the linguistic knowledge carried by the semantic prototypes, e.g., dataset labels. To overcome these problems, we propose a cross-modality graph reasoning adaptation (CIGAR) method to take advantage of both visual and linguistic knowledge. Specifically, our method performs cross-modality graph reasoning between the linguistic modality graph and visual modality graphs to enhance their representations. We also propose a discriminative feature selector to find the most discriminative features and take them as the nodes of the visual graph for both efficiency and effectiveness. In addition, we employ the linguistic graph matching loss to regulate the update of linguistic graphs and maintain their semantic representation during the training process. Comprehensive experiments validate the effectiveness of our proposed CIGAR.
引用
收藏
页码:23776 / 23786
页数:11
相关论文
共 50 条
  • [21] CRKD: Enhanced Camera-Radar Object Detection with Cross-modality Knowledge Distillation
    Zhao, Lingjun
    Song, Jingyu
    Skinner, Katherine A.
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 15470 - 15480
  • [22] DCMNet: Discriminant and cross-modality network for RGB-D salient object detection
    Wang, Fasheng
    Wang, Ruimin
    Sun, Fuming
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 214
  • [23] Cross-modality interaction for few-shot multispectral object detection with semantic knowledge
    Huang, Lian
    Peng, Zongju
    Chen, Fen
    Dai, Shaosheng
    He, Ziqiang
    Liu, Kesheng
    NEURAL NETWORKS, 2024, 173
  • [24] Cross-Modality Knowledge Distillation Network for Monocular 3D Object Detection
    Hong, Yu
    Dai, Hang
    Ding, Yong
    COMPUTER VISION, ECCV 2022, PT X, 2022, 13670 : 87 - 104
  • [25] Cross-modality Discrepant Interaction Network for RGB-D Salient Object Detection
    Zhang, Chen
    Cong, Runmin
    Lin, Qinwei
    Ma, Lin
    Li, Feng
    Zhao, Yao
    Kwong, Sam
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2094 - 2102
  • [26] Cross-modality interaction for few-shot multispectral object detection with semantic knowledge
    Huang, Lian
    Peng, Zongju
    Chen, Fen
    Dai, Shaosheng
    He, Ziqiang
    Liu, Kesheng
    Neural Networks, 2024, 173
  • [27] Asymmetric cross-modality interaction network for RGB-D salient object detection
    Su, Yiming
    Gao, Haoran
    Wang, Mengyin
    Wang, Fasheng
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 275
  • [28] Unsupervised Domain Adaptation for Cross-Modality Cerebrovascular Segmentation
    Wang, Yinuo
    Meng, Cai
    Tang, Zhouping
    Bai, Xiangzhuo
    Ji, Ping
    Bai, Xiangzhi
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (04) : 2871 - 2884
  • [29] SiamSMN: Siamese Cross-Modality Fusion Network for Object Tracking
    Han, Shuo
    Gao, Lisha
    Wu, Yue
    Wei, Tian
    Wang, Manyu
    Cheng, Xu
    INFORMATION, 2024, 15 (07)
  • [30] Relation Matters: Foreground-Aware Graph-Based Relational Reasoning for Domain Adaptive Object Detection
    Chen, Chaoqi
    Li, Jiongcheng
    Zhou, Hong-Yu
    Han, Xiaoguang
    Huang, Yue
    Ding, Xinghao
    Yu, Yizhou
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3677 - 3694