HEX: Human-in-the-loop explainability via deep reinforcement learning

被引:0
|
作者
Lash, Michael T. [1 ]
机构
[1] Univ Kansas, Sch Business, Analyt Informat & Operat Area, 1654 Naismith Dr, Lawrence, KS 66045 USA
关键词
Explainability; Interpretability; Human-in-the-loop; Deep reinforcement learning; Machine learning; Behavioral machine learning; Decision support; EXPLANATIONS; ALGORITHMS; MODELS;
D O I
10.1016/j.dss.2024.114304
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The use of machine learning (ML) models in decision-making contexts, particularly those used in high-stakes decision-making, are fraught with issue and peril since a person - not a machine - must ultimately be held accountable for the consequences of decisions made using such systems. Machine learning explainability (MLX) promises to provide decision-makers with prediction-specific rationale, assuring them that the model-elicited predictions are made for the right reasons and are thus reliable. Few works explicitly consider this key human-in the-loop (HITL) component, however. In this work we propose HEX, a human-in-the-loop deep reinforcement learning approach to MLX. HEX incorporates 0-distrust projection to synthesize decider-specific explainers that produce explanations strictly in terms of a decider's preferred explanatory features using any classification model. Our formulation explicitly considers the decision boundary of the ML model in question using proposed explanatory point mode of explanation, thus ensuring explanations are specific to the ML model in question. We empirically evaluate HEX against other competing methods, finding that HEX is competitive with the state-of-the-art and outperforms other methods in human-in-the-loop scenarios. We conduct a randomized, controlled laboratory experiment utilizing actual explanations elicited from both HEX and competing methods. We causally establish that our method increases decider's trust and tendency to rely on trusted features.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Human-in-the-Loop Reinforcement Learning: A Survey and Position on Requirements, Challenges, and Opportunities
    Retzlaff, Carl Orge
    Das, Srijita
    Wayllace, Christabel
    Mousavi, Payam
    Afshari, Mohammad
    Yang, Tianpei
    Saranti, Anna
    Angerschmid, Alessa
    Taylor, Matthew E.
    Holzinger, Andreas
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2024, 79 : 359 - 415
  • [22] Toward Human-in-the-Loop PID Control Based on CACLA Reinforcement Learning
    Zhong, Junpei
    Li, Yanan
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2019, PT III, 2019, 11742 : 605 - 613
  • [23] PARTNER: Human-in-the-Loop Entity Name Understanding with Deep Learning
    Qian, Kun
    Raman, Poornima Chozhiyath
    Li, Yunyao
    Popa, Lucian
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13634 - 13635
  • [24] Human-in-the-Loop Reinforcement Learning: A Survey and Position on Requirements, Challenges, and Opportunities
    Retzlaff C.O.
    Das S.
    Wayllace C.
    Mousavi P.
    Afshari M.
    Yang T.
    Saranti A.
    Angerschmid A.
    Taylor M.E.
    Holzinger A.
    Journal of Artificial Intelligence Research, 2024, 79 : 359 - 415
  • [25] Human-in-the-loop Extraction of Interpretable Concepts in Deep Learning Models
    Zhao, Zhenge
    Xu, Panpan
    Scheidegger, Carlos
    Ren, Liu
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2022, 28 (01) : 780 - 790
  • [26] Iterative Deep Learning Based Unbiased Stereology With Human-in-the-Loop
    Alahmari, Saeed S.
    Goldgof, Dmitry
    Hall, Lawrence O.
    Dave, Palak
    Phoulady, Hady Ahmady
    Mouton, Peter R.
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 665 - 670
  • [27] Explainability of Deep Reinforcement Learning Method with Drones
    Cetin, Ender
    Barrado, Cristina
    Pastor, Enric
    2023 IEEE/AIAA 42ND DIGITAL AVIONICS SYSTEMS CONFERENCE, DASC, 2023,
  • [28] A Hybrid Human-in-the-Loop Deep Reinforcement Learning Method for UAV Motion Planning for Long Trajectories with Unpredictable Obstacles
    Zhang, Sitong
    Li, Yibing
    Ye, Fang
    Geng, Xiaoyu
    Zhou, Zitao
    Shi, Tuo
    DRONES, 2023, 7 (05)
  • [29] Human-in-the-loop active learning via brain computer interface
    Netzer, Eitan
    Geva, Amir B.
    ANNALS OF MATHEMATICS AND ARTIFICIAL INTELLIGENCE, 2020, 88 (11-12) : 1191 - 1205
  • [30] Human-in-the-loop active learning via brain computer interface
    Eitan Netzer
    Amir B. Geva
    Annals of Mathematics and Artificial Intelligence, 2020, 88 : 1191 - 1205