Compositional Feature Augmentation for Unbiased Scene Graph Generation

被引:11
|
作者
Li, Lin [1 ,2 ]
Chen, Guikun [1 ]
Xiao, Jun [1 ]
Yang, Yi [1 ]
Wang, Chunping [3 ]
Chen, Long [2 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[3] FinVolut, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01982
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scene Graph Generation (SGG) aims to detect all the visual relation triplets <sub, pred, obj> in a given image. With the emergence of various advanced techniques for better utilizing both the intrinsic and extrinsic information in each relation triplet, SGG has achieved great progress over the recent years. However, due to the ubiquitous long-tailed predicate distributions, today's SGG models are still easily biased to the head predicates. Currently, the most prevalent debiasing solutions for SGG are re-balancing methods, e.g., changing the distributions of original training samples. In this paper, we argue that all existing re-balancing strategies fail to increase the diversity of the relation triplet features of each predicate, which is critical for robust SGG. To this end, we propose a novel Compositional Feature Augmentation (CFA) strategy, which is the first unbiased SGG work to mitigate the bias issue from the perspective of increasing the diversity of triplet features. Specifically, we first decompose each relation triplet feature into two components: intrinsic feature and extrinsic feature, which correspond to the intrinsic characteristics and extrinsic contexts of a relation triplet, respectively. Then, we design two different feature augmentation modules to enrich the feature diversity of original relation triplets by replacing or mixing up either their intrinsic or extrinsic features from other samples. Due to its model-agnostic nature, CFA can be seamlessly incorporated into various SGG frameworks. Extensive ablations have shown that CFA achieves a new state-of-the-art performance on the trade-off between different metrics.
引用
收藏
页码:21628 / 21638
页数:11
相关论文
共 50 条
  • [21] Unbiased scene graph generation using the self-distillation method
    Bo Sun
    Zhuo Hao
    Lejun Yu
    Jun He
    The Visual Computer, 2024, 40 : 2381 - 2390
  • [22] Unbiased Scene Graph Generation via Two-Stage Causal Modeling
    Sun, Shuzhou
    Zhi, Shuaifeng
    Liao, Qing
    Heikkila, Janne
    Liu, Li
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12562 - 12580
  • [23] Resistance Training Using Prior Bias: Toward Unbiased Scene Graph Generation
    Chen, Chao
    Zhan, Yibing
    Yu, Baosheng
    Liu, Liu
    Luo, Yong
    Du, Bo
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 212 - 220
  • [24] Dual-Branch Hybrid Learning Network for Unbiased Scene Graph Generation
    Zheng, Chaofan
    Gao, Lianli
    Lyu, Xinyu
    Zeng, Pengpeng
    El Saddik, Abdulmotaleb
    Shen, Heng Tao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (03) : 1743 - 1756
  • [25] Attention redirection transformer with semantic oriented learning for unbiased scene graph generation
    Zhang, Ruonan
    An, Gaoyun
    Cen, Yigang
    Ruan, Qiuqi
    PATTERN RECOGNITION, 2025, 158
  • [26] PPDL: Predicate Probability Distribution based Loss for Unbiased Scene Graph Generation
    Li, Wei
    Zhang, Haiwei
    Bai, Qijie
    Zhao, Guoqing
    Jiang, Ning
    Yuan, Xiaojie
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 19425 - 19434
  • [27] Bridging Visual and Textual Semantics: Towards Consistency for Unbiased Scene Graph Generation
    Zhang, Ruonan
    An, Gaoyun
    Hao, Yiqing
    Wu, Dapeng Oliver
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (11) : 7102 - 7119
  • [28] PCPL: Predicate-Correlation Perception Learning for Unbiased Scene Graph Generation
    Yan, Shaotian
    Shen, Chen
    Jin, Zhongming
    Huang, Jianqiang
    Jiang, Rongxin
    Chen, Yaowu
    Hua, Xian-Sheng
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 265 - 273
  • [29] Triple Correlations-Guided Label Supplementation for Unbiased Video Scene Graph Generation
    Wang, Wenqing
    Gao, Kaifeng
    Luo, Yawei
    Jiang, Tao
    Gao, Fei
    Shao, Jian
    Sun, Jianwen
    Xiao, Jun
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5153 - 5163
  • [30] A New Training Data Organization Form and Training Mode for Unbiased Scene Graph Generation
    Xu, Hongbo
    Wang, Lichun
    Xu, Kai
    Fu, Fangyu
    Yin, Baocai
    Huang, Qingming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 5295 - 5305