Multi-Level Multi-Modal Cross-Attention Network for Fake News Detection

被引:31
|
作者
Ying, Long [1 ]
Yu, Hui [1 ]
Wang, Jinguang [2 ]
Ji, Yongze [3 ]
Qian, Shengsheng [4 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp & Software, Nanjing 210044, Peoples R China
[2] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230601, Peoples R China
[3] China Univ Petr, Sch Informat Sci & Engn, Beijing 102249, Peoples R China
[4] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
来源
IEEE ACCESS | 2021年 / 9卷
基金
中国国家自然科学基金;
关键词
Feature extraction; Semantics; Visualization; Task analysis; Bit error rate; Convolutional neural networks; Social networking (online); Multi-level neural networks; fake news detection; multi-modal fusion;
D O I
10.1109/ACCESS.2021.3114093
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the development of the Mobile Internet, more and more users publish multi-modal posts on social media platforms. Fake news detection has become an increasingly challenging task. Although there are many works using deep schemes to extract and combine textual and visual representation in the post, most existing methods do not sufficiently utilize the complementary multi-modal information containing semantic concepts and entities to complement and enhance each modality. Moreover, these methods do not model and incorporate the rich multi-level semantics of text information to improve fake news detection tasks. In this paper, we propose a novel end-to-end Multi-level Multi-modal Cross-attention Network (MMCN) which exploits the multi-level semantics of textual content and jointly integrates the relationships of duplicate and different modalities (textual and visual modality) of social multimedia posts in a unified framework. Pre-trained BERT and ResNet models are employed to generate high-quality representations for text words and image regions respectively. A multi-modal cross-attention network is then designed to fuse the feature embeddings of the text words and image regions by simultaneously considering data relationships in duplicate and different modalities. Specially, due to different layers of the transformer architecture have different feature representations, we employ a multi-level encoding network to capture the rich multi-level semantics to enhance the presentations of posts. Extensive experiments on the two public datasets (WEIBO and PHEME) demonstrate that compared with the state-of-the-art models, the proposed MMCN has an advantageous performance.
引用
收藏
页码:132363 / 132373
页数:11
相关论文
共 50 条
  • [31] Fake News Detection Based on Multi-Modal Classifier Ensemble
    Shao, Yi
    Sun, Jiande
    Zhang, Tianlin
    Jiang, Ye
    Ma, Jianhua
    Li, Jing
    1ST ACM INTERNATIONAL WORKSHOP ON MULTIMEDIA AI AGAINST DISINFORMATION, MAD 2022, 2022, : 78 - 86
  • [32] Leveraging Supplementary Information for Multi-Modal Fake News Detection
    Ho, Chia-Chun
    Dai, Bi-Ru
    2023 INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGIES FOR DISASTER MANAGEMENT, ICT-DM, 2023, : 50 - 54
  • [33] IS CROSS-ATTENTION PREFERABLE TO SELF-ATTENTION FOR MULTI-MODAL EMOTION RECOGNITION?
    Rajan, Vandana
    Brutti, Alessio
    Cavallaro, Andrea
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4693 - 4697
  • [34] Embracing Domain Differences in Fake News: Cross-domain Fake News Detection using Multi-modal Data
    Silva, Amila
    Luo, Ling
    Karunasekera, Shanika
    Leckie, Christopher
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 557 - 565
  • [35] MICN: Multi-level Induced Cross-Attention Network for Breast Lesion Segmentation
    Ye, Xianjun
    Qu, Xiaofeng
    Xu, Zhenyi
    Kang, Yu
    2024 INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS, ICARM 2024, 2024, : 795 - 800
  • [36] Entity-Oriented Multi-Modal Alignment and Fusion Network for Fake News Detection
    Li, Peiguang
    Sun, Xian
    Yu, Hongfeng
    Tian, Yu
    Yao, Fanglong
    Xu, Guangluan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 24 : 3455 - 3468
  • [37] Positive Unlabeled Fake News Detection via Multi-Modal Masked Transformer Network
    Wang, Jinguang
    Qian, Shengsheng
    Hu, Jun
    Hong, Richang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 234 - 244
  • [38] Cross-Attention Model for Multi-modal Bio-Signal Processing
    Heesoo, Son
    Sangseok, Lee
    Sael, Lee
    2022 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (IEEE BIGCOMP 2022), 2022, : 43 - 46
  • [39] MAFE: Multi-modal Alignment via Mutual Information Maximum Perspective in Multi-modal Fake News Detection
    Qin, Haimei
    Jing, Yaqi
    Duan, Yunqiang
    Jiang, Lei
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1515 - 1521
  • [40] A joint hierarchical cross-attention graph convolutional network for multi-modal facial expression recognition
    Xu, Chujie
    Du, Yong
    Wang, Jingzi
    Zheng, Wenjie
    Li, Tiejun
    Yuan, Zhansheng
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (01)