A Weighted Cross-Modal Feature Aggregation Network for Rumor Detection

被引:0
|
作者
Li, Jia [1 ]
Hu, Zihan [1 ]
Yang, Zhenguo [1 ]
Lee, Lap-Kei [2 ]
Wang, Fu Lee [2 ]
机构
[1] Guangdong Univ Technol, Guangzhou, Peoples R China
[2] Hong Kong Metropolitan Univ, Hong Kong, Peoples R China
关键词
Rumor detection; Adversarial training; Cross-modal alignment;
D O I
10.1007/978-981-97-2266-2_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a Weighted Cross-modal Aggregation network (WCAN) for rumor detection in order to combine highly correlated features in different modalities and obtain a unified representation in the same space. WCAN exploits an adversarial training method to add perturbations to text features to enhance model robustness. Specifically, we devise a weighted cross-modal aggregation (WCA) module that measures the distance between text, image and social graph modality distributions using KL divergence, which leverages correlations between modalities. By using MSE loss, the fusion features are progressively closer to the original features of the image and social graph while taking into account all of the information from each modality. In addition, WCAN includes a feature fusion module that uses dual-modal co-attention blocks to dynamically adjust features from three modalities. Experiments are conducted on two datasets, WEIBO and PHEME, and the experimental results demonstrate the superior performance of the proposed method.
引用
收藏
页码:42 / 53
页数:12
相关论文
共 50 条
  • [31] CAFE: Robust Detection of Malicious Macro based on Cross-modal Feature Extraction
    Bao, Huaifeng
    Wang, Xingyu
    Li, Wenhao
    Xu, Jinpeng
    Yin, Peng
    Wang, Wen
    Liu, Feng
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 2534 - 2540
  • [32] Cross-modal and multi-level feature refinement network for RGB-D salient object detection
    Gao, Yue
    Dai, Meng
    Zhang, Qing
    VISUAL COMPUTER, 2023, 39 (09): : 3979 - 3994
  • [33] GCANet: A Cross-Modal Pedestrian Detection Method Based on Gaussian Cross Attention Network
    Peng, Peiran
    Mu, Feng
    Yan, Peilin
    Song, Liqiang
    Li, Hui
    Chen, Yu
    Li, Jianan
    Xu, Tingfa
    INTELLIGENT COMPUTING, VOL 2, 2022, 507 : 520 - 530
  • [34] Cross-Modal and Cross-Level Attention Interaction Network for Salient Object Detection
    Wang F.
    Su Y.
    Wang R.
    Sun J.
    Sun F.
    Li H.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (06): : 2907 - 2920
  • [35] A cross-modal fusion network based on graph feature learning for multimodal emotion recognition
    Cao Xiaopeng
    Zhang Linying
    Chen Qiuxian
    Ning Hailong
    Dong Yizhuo
    The Journal of China Universities of Posts and Telecommunications, 2024, 31 (06) : 16 - 25
  • [36] Hybrid Network Based on Cross-Modal Feature Fusion for Diagnosis of Alzheimer's Disease
    Qiu, Zifeng
    Yang, Peng
    Wang, Tianfu
    Lei, Baiying
    ETHICAL AND PHILOSOPHICAL ISSUES IN MEDICAL IMAGING, MULTIMODAL LEARNING AND FUSION ACROSS SCALES FOR CLINICAL DECISION SUPPORT, AND TOPOLOGICAL DATA ANALYSIS FOR BIOMEDICAL IMAGING, EPIMI 2022, ML-CDS 2022, TDA4BIOMEDICALIMAGING, 2022, 13755 : 87 - 99
  • [37] Grouping by feature of cross-modal flankers in temporal ventriloquism
    Klimova, Michaela
    Nishida, Shin'ya
    Roseboom, Warrick
    SCIENTIFIC REPORTS, 2017, 7
  • [38] Grouping by feature of cross-modal flankers in temporal ventriloquism
    Michaela Klimova
    Shin’ya Nishida
    Warrick Roseboom
    Scientific Reports, 7
  • [39] Learning Coupled Feature Spaces for Cross-modal Matching
    Wang, Kaiye
    He, Ran
    Wang, Wei
    Wang, Liang
    Tan, Tieniu
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 2088 - 2095
  • [40] Joint feature fusion hashing for cross-modal retrieval
    Cao, Yuxia
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 6149 - 6162