Cross-Modal Feature Fusion-Based Knowledge Transfer for Text-Based Person Search

被引:1
|
作者
You, Kaiyang [1 ,2 ]
Chen, Wenjing [3 ]
Wang, Chengji [1 ,2 ]
Sun, Hao [1 ,2 ]
Xie, Wei [1 ,2 ]
机构
[1] Cent China Normal Univ, Sch Comp Sci, Hubei Prov Key Lab Artificial Intelligence & Smart, Wuhan 430079, Peoples R China
[2] Cent China Normal Univ, Natl Language Resources Monitoring & Res Ctr Netwo, Wuhan 430079, Peoples R China
[3] Hubei Univ Technol, Sch Comp Sci, Wuhan 430068, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Knowledge transfer; Visualization; Transformers; Data mining; Task analysis; Sun; Text-based person search; knowledge imbalance; knowledge transfer; cross-modal fusion; TRANSFORMER;
D O I
10.1109/LSP.2024.3449222
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Text-based person search aims to retrieve corresponding images of person from a large gallery based on text descriptions. Existing methods strive to bridge the modality gap between images and texts and have made promising progress. However, these approaches disregard the knowledge imbalance between images and texts caused by the reporting bias. To resolve this issue, we present a cross-modal feature fusion-based knowledge transfer network to balance identity information between images and texts. First, we design an identity information emphasis module to enhance person-relevant information and suppress person-irrelevant information. Second, we design an intermediate modal-guided knowledge transfer module to balance the knowledge between images and texts. Experimental results on CUHK-PEDES, ICFG-PEDE, and RSTPReid datasets demonstrate that our method achieves state-of-the-art performance.
引用
收藏
页码:2230 / 2234
页数:5
相关论文
共 50 条
  • [31] Text-based Person Search via Virtual Attribute Learning
    Wang C.-J.
    Su J.-W.
    Luo Z.-M.
    Cao D.-L.
    Lin Y.-J.
    Li S.-Z.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (05): : 2035 - 2050
  • [32] Hierarchical Gumbel Attention Network for Text-based Person Search
    Zheng, Kecheng
    Liu, Wu
    Liu, Jiawei
    Zha, Zheng-Jun
    Mei, Tao
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3441 - 3449
  • [33] An Adaptive Correlation Filtering Method for Text-Based Person Search
    Sun, Mengyang
    Suo, Wei
    Wang, Peng
    Niu, Kai
    Liu, Le
    Lin, Guosheng
    Zhang, Yanning
    Wu, Qi
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (10) : 4440 - 4455
  • [34] Noise correspondence with evidence learning for text-based person search
    Xie, Yihan
    Zhang, Baohua
    Li, Yang
    Shan, Chongrui
    Wang, Shun
    Zhang, Jiale
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (05):
  • [35] PaSeMix: A Multi-modal Partitional Semantic Data Augmentation Method for Text-Based Person Search
    Yuan, Xinpan
    Li, Jiabao
    Gan, Wenguang
    Xia, Wei
    Weng, Yanbin
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14864 : 468 - 479
  • [36] Feature fusion-based collaborative learning for knowledge distillation
    Li, Yiting
    Sun, Liyuan
    Gou, Jianping
    Du, Lan
    Ou, Weihua
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2021, 17 (11)
  • [37] Diverse Person: Customize Your Own Dataset for Text-Based Person Search
    Song, Zifan
    Hu, Guosheng
    Zhao, Cairong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4943 - 4951
  • [38] Drug-Target Affinity Prediction Based on Cross-Modal Fusion of Text and Graph
    Yang, Jucheng
    Ren, Fushun
    APPLIED SCIENCES-BASEL, 2025, 15 (06):
  • [39] Text-based Person Search without Parallel Image-Text Data
    Bai, Yang
    Wang, Jingyao
    Cao, Min
    Chen, Chen
    Cao, Ziqiang
    Nie, Liqiang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 757 - 767
  • [40] A cross-modal fusion network based on graph feature learning for multimodal emotion recognition
    Cao Xiaopeng
    Zhang Linying
    Chen Qiuxian
    Ning Hailong
    Dong Yizhuo
    The Journal of China Universities of Posts and Telecommunications, 2024, 31 (06) : 16 - 25