Modeling inter-modal incongruous sentiment expressions for multi-modal sarcasm detection

被引:2
|
作者
Ou, Lisong [1 ,2 ,3 ]
Li, Zhixin [1 ,2 ]
机构
[1] Guangxi Normal Univ, Key Lab Educ Blockchain & Intelligent Technol, Minist Educ, Guilin 541004, Peoples R China
[2] Guangxi Normal Univ, Guangxi Key Lab Multisource Informat Min & Secur, Guilin 541004, Peoples R China
[3] Guilin Univ Technol, Sch Math & Stat, Guilin 541004, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-modal sarcasm detection; Graph convolutional network; Cross-modal mapping; External knowledge; Cross-correlation graphs;
D O I
10.1016/j.neucom.2024.128874
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modal sarcasm detection (MSD) presents a formidable and intricate endeavor. Despite strides made by extant models, two principal hurdles persist: Firstly, prevailing methodologies merely address superficial disparities between textual inputs and associated images, neglecting nuanced inter-modal combinations. Secondly, satirical instances frequently involve intricate emotional expressions, highlighting the imperative of leveraging emotional cues across modalities to discern satirical nuances. Accordingly, this research proposes the utilization of a deep graph convolutional network that integrates cross-modal mapping information to effectively identify significant incongruent sentiment expressions across various modalities for the purpose of multi-modal sarcasm detection. Specifically, we first design a cross-modal mapping network, which obtains the interaction information between these two modalities by mapping text feature vectors and image feature vectors two by two to compensate for the lack of multi-modal data in the fusion process. Additionally, we employ external knowledge of ANPS as abridge to construct cross-correlation graphs through highly correlated satirical cues and their connection weights between image and text modalities. Afterward, the GCN architecture with retrieval-based attentional mechanisms will effectively capture satirical cues. The experiments conducted on two publicly available datasets demonstrate a significant enhancement in the performance of our method when compared to numerous contemporary models.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Multi-Modal Pedestrian Detection with Large Misalignment Based on Modal-Wise Regression and Multi-Modal IoU
    Wanchaitanawong, Napat
    Tanaka, Masayuki
    Shibata, Takashi
    Okutomi, Masatoshi
    PROCEEDINGS OF 17TH INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS (MVA 2021), 2021,
  • [42] INTER-MODAL EXPLORATION AND KNOWLEDGE IN INFANCY
    SPELKE, ES
    OWSLEY, CJ
    INFANT BEHAVIOR & DEVELOPMENT, 1979, 2 (01): : 13 - 27
  • [43] What Does Your Smile Mean? Jointly Detecting Multi-Modal Sarcasm and Sentiment Using Quantum Probability
    Liu, Yaochen
    Zhang, Yazhou
    Li, Qiuchi
    Wang, Benyou
    Song, Dawei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 871 - 880
  • [44] INTER-MODAL TRANSPORTATION GARAGE FACILITY
    PACELLI, AJ
    TRANSPORTATION ENGINEERING JOURNAL OF ASCE, 1980, 106 (04): : 401 - 413
  • [45] Is Multi-Modal Necessarily Better? Robustness Evaluation of Multi-Modal Fake News Detection
    Chen, Jinyin
    Jia, Chengyu
    Zheng, Haibin
    Chen, Ruoxi
    Fu, Chenbo
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (06): : 3144 - 3158
  • [46] KINESTHETIC AND INTER-MODAL TILT AFTEREFFECTS
    MORANT, RB
    MISTOVICH, M
    ACTA PSYCHOLOGICA, 1963, 21 (01) : 24 - 34
  • [48] INTER-MODAL MATCHING BY HUMAN NEONATES
    MELTZOFF, AN
    BORTON, RW
    NATURE, 1979, 282 (5737) : 403 - 404
  • [49] OPTIMAL PRICING WITH INTER-MODAL COMPETITION
    BRAEUTIGAM, RR
    AMERICAN ECONOMIC REVIEW, 1979, 69 (01): : 38 - 49
  • [50] Toward's Arabic Multi-modal Sentiment Analysis
    Alqarafi, Abdulrahman S.
    Adeel, Ahsan
    Gogate, Mandar
    Dashitpour, Kia
    Hussain, Amir
    Durrani, Tariq
    COMMUNICATIONS, SIGNAL PROCESSING, AND SYSTEMS, 2019, 463 : 2378 - 2386