Multimodal dual perception fusion framework for multimodal affective analysis

被引:0
|
作者
Lu, Qiang [1 ]
Sun, Xia [1 ]
Long, Yunfei [2 ]
Zhao, Xiaodi [1 ]
Zou, Wang [1 ]
Feng, Jun [1 ]
Wang, Xuxin [1 ]
机构
[1] Northwest Univ, Sch Informat Sci & Technol, Xian 710127, Peoples R China
[2] Univ Essex, Sch Comp Sci & Elect Engn, Colchester CO4 3SQ, England
关键词
Multimodal sentiment analysis; Sarcasm detection; Fake news detection; Multimodal affective analysis; Multimodal dual perception fusion;
D O I
10.1016/j.inffus.2024.102747
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The misuse of social platforms and the difficulty in regulating post contents have culminated in a surge of negative sentiments, sarcasms, and the rampant spread of fake news. In response, Multimodal sentiment analysis, sarcasm detection and fake news detection based on image and text have attracted considerable attention recently. Due to that these areas share semantic and sentiment features and confront related fusion challenges in deciphering complex human expressions across different modalities, integrating these multimodal classification tasks that share commonalities across different scenarios into a unified framework is expected to simplify research in sentiment analysis, and enhance the effectiveness of classification tasks involving both semantic and sentiment modeling. Therefore, we consider integral components of a broader spectrum of research known as multimodal affective analysis towards semantics and sentiment, and propose a novel multimodal dual perception fusion framework (MDPF). Specifically, MDPF contains three core procedures: (1) Generating bootstrapping language-image Knowledge to enrich origin modality space, and utilizing cross- modal contrastive learning for aligning text and image modalities to understand underlying semantics and interactions. (2) Designing dynamic connective mechanism to adaptively match image-text pairs and jointly employing gaussian-weighted distribution to intensify semantic sequences. (3) Constructing a cross-modal graph to preserve the structured information of both image and text data and share information between modalities, while introducing sentiment knowledge to refine the edge weights of the graph to capture cross- modal sentiment interaction. We evaluate MDPF on three publicly available datasets across three tasks, and the empirical results demonstrate the superiority of our proposed model.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] A multimodal generative and fusion framework for recognizing faculty homepages
    Yu, Guanyuan
    Li, Qing
    Wang, Jun
    Zhang, Di
    Liu, Yuehao
    INFORMATION SCIENCES, 2020, 525 : 205 - 220
  • [32] Multimodal Data Fusion Framework For Fake News Detection
    Athira, A. B.
    Tiwari, Abhishek
    Kumar, S. D. Madhu
    Chacko, Anu Mary
    2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [33] A multimodal fusion framework for children's storytelling systems
    Wang, Danli
    Zhang, Jie
    Dai, Guozhong
    TECHNOLOGIES FOR E-LEARNING AND DIGITAL ENTERTAINMENT, PROCEEDINGS, 2006, 3942 : 585 - 588
  • [34] Effective Diagnosis of Alzheimer's Disease via Multimodal Fusion Analysis Framework
    Bi, Xia-an
    Cai, Ruipeng
    Wang, Yang
    Liu, Yingchao
    FRONTIERS IN GENETICS, 2019, 10
  • [35] Software Framework for Multimodal Fusion in Ubiquitous Computing Applications
    Vildjiounaite, Elena
    Kyllonen, Vesa
    2009 INTERNATIONAL CONFERENCE ON NETWORK AND SERVICE SECURITY, 2009, : 34 - 39
  • [36] Sound Emblems for Affective Multimodal Output of a Robotic Tutor: A Perception Study
    Hastie, Helen
    Dente, Pasquale
    Kuester, Dennis
    Kappas, Arvid
    ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, : 256 - 260
  • [37] Multimodal dynamic fusion framework: Multilevel feature fusion guided by prompts
    Pan, Lei
    Wu, Huan-Qing
    EXPERT SYSTEMS, 2024, 41 (11)
  • [38] MUON: multimodal omics analysis framework
    Danila Bredikhin
    Ilia Kats
    Oliver Stegle
    Genome Biology, 23
  • [39] MUON: multimodal omics analysis framework
    Bredikhin, Danila
    Kats, Ilia
    Stegle, Oliver
    GENOME BIOLOGY, 2022, 23 (01)
  • [40] MAG: a smart gloves system based on multimodal fusion perception
    Hong Cui
    Zhiquan Feng
    Jinglan Tian
    Dehui Kong
    Zishuo Xia
    Weina Li
    CCF Transactions on Pervasive Computing and Interaction, 2023, 5 : 411 - 429