Hate-UDF: Explainable Hateful Meme Detection With Uncertainty-Aware Dynamic Fusion

被引:0
|
作者
Lei, Xia [1 ]
Wang, Siqi [2 ]
Fan, Yongkai [1 ]
Shang, Wenqian [1 ]
机构
[1] Commun Univ China, State Key Lab Media Convergence & Commun, Beijing, Peoples R China
[2] Hebei Normal Univ, Coll Comp & Cyber Secur, Shijiazhuang, Peoples R China
关键词
detection; dynamic fusion; hateful meme; interpretable; multi-modal;
D O I
10.1002/spe.3403
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
BackgroundWith the increasing integration of Artificial Intelligence (AI) and Internet of Things (IoT), the dissemination of multimodal data is undergoing revolutionary changes. To mitigate the societal risks posed by the rapid spread of malicious multimodal data, such as hateful memes, it is crucial to develop effective detection methods for such data. Existing detection models often struggle with data quality issues and lack interpretability, limiting their effectiveness in content moderation tasks.AimsThis paper aims to propose an explainable hateful meme detection model by uncertainty-aware dynamic fusion. The goal is to enhance both generalization performance and interpretability, addressing the limitations of conventional static fusion methods and existing algorithms for hateful meme detection.Materials & MethodsTo mitigate the societal risks posed by the rapid spread of malicious multimodal data, such as hateful memes, it is crucial to develop effective detection methods for such data. However, existing algorithms for hateful meme detection frequently overlook the data quality and the interpretability of model. To adress these challenges, this paper proposes Hate-UDF, an explainable hateful meme detection model with uncertainty-aware dynamic fusion, providing both high generalization ability and interpretability. This method dynamically evaluates the uncertainty of different modalities, obtains dynamic weights, and utilizes them to weight the feature values for fusion, thereby obtaining a uncertainty-aware dynamic fusion method with provable upper bounds on generalization error. Furthermore, an analysis of the dynamic weights can explain the modality on which the model primarily relies for detection, thereby providing a method that is both explainable and reliable.ResultsWe compare the performance of Hate-UDF with three general models and three State of the Art (SOTA) models in the field of hateful meme detection on the Facebook Hateful Memes (FHM) and the Multimedia Automatic Misogyny Identification (MAMI) datasets. Hate-UDF achieved state-of-the-art performance, surpassing existing models on both datasets. Specifically, it improved accuracy and AUC by 7.56% and 2.8% on FHM and by 3.34% and 0.17% on MAMI compared with the current SOTA model, respectively. Additionally, we demonstrate that the visual modality is more important than the textual modality in the hateful meme detection model, and we explain the primary reason behind this by visualization.DiscussionThe model dynamically adapts to modality quality, enhancing reliability and reducing the risk of misclassification. Its interpretability, achieved through visualizations of modality and feature attributions, provides valuable insights for content moderation systems and highlights the importance of image modality in detecting hateful meme. While Hate-UDF provides an explainable and reliable method for detecting hateful memes, it may still learn biases from the training data, potentially leading to the over-detection of content from certain groups or communities. Future research must focus on improving the fairness and ethical responsibilities of the model's decisions.ConclusionThis paper introduces the model of Hate-UDF, a dynamic fusion method based on uncertainty, designed to improve multimodal fusion issues in existing hateful meme detection models. The model determines the reliability of different modal information by assessing their uncertainty and generates dynamic weights accordingly. By comparing these weights, the model can identify which modality is most influential in detecting malicious content. Therefore, the Hate-UDF model not only has interpretability but also its generalization performance has been validated.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Uncertainty-Aware and Explainable Human Error Detection in the Operation of Nuclear Power Plants
    Reddy, Bhavya
    Gursel, Ezgi
    Daniels, Katy
    Khojandi, Anahita
    Baalis Coble, Jamie
    Agarwal, Vivek
    Boring, Ronald
    Yadav, Vaibhav
    Madadi, Mahboubeh
    NUCLEAR TECHNOLOGY, 2024, 210 (12) : 2312 - 2330
  • [2] Uncertainty-aware explainable AI as a foundational paradigm for digital twins
    Cohen, Joseph
    Huan, Xun
    FRONTIERS IN MECHANICAL ENGINEERING-SWITZERLAND, 2024, 9
  • [3] An Uncertainty-Aware Encoder for Aspect Detection
    Nguyen, Thi-Nhung
    Nguyen, Kiem-Hieu
    Song, Young-In
    Cao, Tuan-Dung
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 797 - 806
  • [4] Uncertainty-Aware Time Series Anomaly Detection
    Wiessner, Paul
    Bezirganyan, Grigor
    Sellami, Sana
    Chbeir, Richard
    Bungartz, Hans-Joachim
    FUTURE INTERNET, 2024, 16 (11)
  • [5] Uncertainty-Aware LSTM Based Dynamic Flight Fault Detection for UAV Actuator
    Guo, Kai
    Wang, Na
    Liu, Datong
    Peng, Xiyuan
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [6] Uncertainty-Aware LSTM Based Dynamic Flight Fault Detection for UAV Actuator
    Guo, Kai
    Wang, Na
    Liu, Datong
    Peng, Xiyuan
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [7] Dynamic predictions of postoperative complications from explainable, uncertainty-aware, and multi-task deep neural networks
    Shickel, Benjamin
    Loftus, Tyler J.
    Ruppert, Matthew
    Upchurch, Gilbert R.
    Ozrazgat-Baslanti, Tezcan
    Rashidi, Parisa
    Bihorac, Azra
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [8] Dynamic predictions of postoperative complications from explainable, uncertainty-aware, and multi-task deep neural networks
    Benjamin Shickel
    Tyler J. Loftus
    Matthew Ruppert
    Gilbert R. Upchurch
    Tezcan Ozrazgat-Baslanti
    Parisa Rashidi
    Azra Bihorac
    Scientific Reports, 13
  • [9] Uncertainty-aware and explainable machine learning for early prediction of battery degradation trajectory
    Rieger, Laura Hannemose
    Flores, Eibar
    Nielsen, Kristian Frellesen
    Norby, Poul
    Ayerbe, Elixabete
    Winther, Ole
    Vegge, Tejs
    Bhowmik, Arghya
    DIGITAL DISCOVERY, 2023, 2 (01): : 112 - 122
  • [10] Uncertainty-Aware Deep Ensembles for Reliable and Explainable Predictions of Clinical Time Series
    Wickstrom, Kristoffer
    Mikalsen, Karl Oyvind
    Kampffmeyer, Michael
    Revhaug, Arthur
    Jenssen, Robert
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (07) : 2435 - 2444