HIVE: Evaluating the Human Interpretability of Visual Explanations

被引:38
|
作者
Kim, Sunnie S. Y. [1 ]
Meister, Nicole [1 ]
Ramaswamy, Vikram V. [1 ]
Fong, Ruth [1 ]
Russakovsky, Olga [1 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
来源
基金
美国国家科学基金会;
关键词
Interpretability; Explainable AI (XAI); Human studies; Evaluation framework; Human-centered AI;
D O I
10.1007/978-3-031-19775-8_17
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As AI technology is increasingly applied to high-impact, high-risk domains, there have been a number of new methods aimed at making AI models more human interpretable. Despite the recent growth of interpretability work, there is a lack of systematic evaluation of proposed techniques. In this work, we introduce HIVE (Human Interpretability of Visual Explanations), a novel human evaluation framework that assesses the utility of explanations to human users in AI-assisted decision making scenarios, and enables falsifiable hypothesis testing, cross-method comparison, and human-centered evaluation of visual interpretability methods. To the best of our knowledge, this is the first work of its kind. Using HIVE, we conduct IRB-approved human studies with nearly 1000 participants and evaluate four methods that represent the diversity of computer vision interpretability works: GradCAM, BagNet, ProtoPNet, and ProtoTree. Our results suggest that explanations engender human trust, even for incorrect predictions, yet are not distinct enough for users to distinguish between correct and incorrect predictions. We open-source HIVE to enable future studies and encourage more human-centered approaches to interpretability research. HIVE can be found at https://princetonvisualai.github.io/HIVE.
引用
收藏
页码:280 / 298
页数:19
相关论文
共 50 条
  • [31] Evaluating Human-like Explanations for Robot Actions in Reinforcement Learning Scenarios
    Cruz, Francisco
    Young, Charlotte
    Dazeley, Richard
    Vamplew, Peter
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 894 - 901
  • [32] Adversarial Counterfactual Visual Explanations
    Jeanneret, Guillaume
    Simon, Loic
    Jurie, Frederic
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 16425 - 16435
  • [33] Exploring Coherence in Visual Explanations
    Alikhani, Malihe
    Stone, Matthew
    IEEE 1ST CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2018), 2018, : 272 - 277
  • [34] Visual Explanations of Probabilistic Reasoning
    Erwig, Martin
    Walkingshaw, Eric
    2009 IEEE SYMPOSIUM ON VISUAL LANGUAGES AND HUMAN-CENTRIC COMPUTING, PROCEEDINGS, 2009, : 23 - 27
  • [35] Diffusion Visual Counterfactual Explanations
    Augustin, Maximilian
    Boreiko, Valentyn
    Croce, Francesco
    Hein, Matthias
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [36] Enhancing Interpretability in Molecular Property Prediction with Contextual Explanations of Molecular Graphical Depictions
    Bertolini, Marco
    Zhao, Linlin
    Montanari, Floriane
    Clevert, Djork-Arne
    AI IN DRUG DISCOVERY, AIDD 2024, 2025, 14894 : 1 - 12
  • [37] Visual interpretability of bioimaging deep learning models
    Rotem, Oded
    Zaritsky, Assaf
    NATURE METHODS, 2024, 21 (08) : 1394 - 1397
  • [38] VOICE: Variance of Induced Contrastive Explanations to Quantify Uncertainty in Neural Network Interpretability
    Prabhushankar, Mohit
    Alregib, Ghassan
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2025, 19 (01) : 19 - 31
  • [39] A holistic approach to interpretability in financial lending: Models, visualizations, and summary-explanations
    Chen, Chaofan
    Lin, Kangcheng
    Rudin, Cynthia
    Shaposhnik, Yaron
    Wang, Sijia
    Wang, Tong
    DECISION SUPPORT SYSTEMS, 2022, 152
  • [40] EVALUATING COGNITIVE EXPLANATIONS OF COMMUNICATIVE PHENOMENA
    GREENE, JO
    QUARTERLY JOURNAL OF SPEECH, 1984, 70 (03) : 241 - 254