Stable and actionable explanations of black-box models through factual and counterfactual rules

被引:15
|
作者
Guidotti, Riccardo [1 ]
Monreale, Anna [1 ]
Ruggieri, Salvatore [1 ]
Naretto, Francesca [2 ]
Turini, Franco [1 ]
Pedreschi, Dino [1 ]
Giannotti, Fosca [2 ]
机构
[1] Univ Pisa, Dept Comp Sci, Largo B Pontecorvo 3, I-56127 Pisa, PI, Italy
[2] Scuola Normale Super Pisa, Piazza Cavalieri 7, I-56126 Pisa, PI, Italy
基金
英国工程与自然科学研究理事会; 欧盟地平线“2020”; 欧洲研究理事会;
关键词
Explainable AI; Local explanations; Model-agnostic explanations; Rule-based explanations; Counterfactuals; INSTANCE SELECTION; ALGORITHMS;
D O I
10.1007/s10618-022-00878-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years have witnessed the rise of accurate but obscure classification models that hide the logic of their internal decision processes. Explaining the decision taken by a black-box classifier on a specific input instance is therefore of striking interest. We propose a local rule-based model-agnostic explanation method providing stable and actionable explanations. An explanation consists of a factual logic rule, stating the reasons for the black-box decision, and a set of actionable counterfactual logic rules, proactively suggesting the changes in the instance that lead to a different outcome. Explanations are computed from a decision tree that mimics the behavior of the black-box locally to the instance to explain. The decision tree is obtained through a bagging-like approach that favors stability and fidelity: first, an ensemble of decision trees is learned from neighborhoods of the instance under investigation; then, the ensemble is merged into a single decision tree. Neighbor instances are synthetically generated through a genetic algorithm whose fitness function is driven by the black-box behavior. Experiments show that the proposed method advances the state-of-the-art towards a comprehensive approach that successfully covers stability and actionability of factual and counterfactual explanations.
引用
收藏
页码:2825 / 2862
页数:38
相关论文
共 50 条
  • [1] Factual and Counterfactual Explanations for Black Box Decision Making
    Guidotti, Riccardo
    Monreale, Anna
    Giannotti, Fosca
    Pedreschi, Dino
    Ruggieri, Salvatore
    Turini, Franco
    IEEE INTELLIGENT SYSTEMS, 2019, 34 (06) : 14 - 22
  • [2] Learning Groupwise Explanations for Black-Box Models
    Gao, Jingyue
    Wang, Xiting
    Wang, Yasha
    Yan, Yulan
    Xie, Xing
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2396 - 2402
  • [3] Feature Importance Explanations for Temporal Black-Box Models
    Sood, Akshay
    Craven, Mark
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8351 - 8360
  • [4] Considerations when learning additive explanations for black-box models
    Tan, Sarah
    Hooker, Giles
    Koch, Paul
    Gordo, Albert
    Caruana, Rich
    MACHINE LEARNING, 2023, 112 (09) : 3333 - 3359
  • [5] Considerations when learning additive explanations for black-box models
    Sarah Tan
    Giles Hooker
    Paul Koch
    Albert Gordo
    Rich Caruana
    Machine Learning, 2023, 112 : 3333 - 3359
  • [6] A Generic Framework for Black-box Explanations
    Henin, Clement
    Le Metayer, Daniel
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 3667 - 3676
  • [7] Generative causal explanations of black-box classifiers
    O'Shaughnessy, Matthew
    Canal, Gregory
    Connor, Marissa
    Davenport, Mark
    Rozell, Christopher
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [8] Comparing Explanations from Glass-Box and Black-Box Machine-Learning Models
    Kuk, Michal
    Bobek, Szymon
    Nalepa, Grzegorz J.
    COMPUTATIONAL SCIENCE - ICCS 2022, PT III, 2022, 13352 : 668 - 675
  • [9] Explaining the black-box smoothly-A counterfactual approach
    Singla, Sumedha
    Eslami, Motahhare
    Pollack, Brian
    Wallace, Stephen
    Batmanghelich, Kayhan
    MEDICAL IMAGE ANALYSIS, 2023, 84
  • [10] DiConStruct: Causal Concept-based Explanations through Black-Box Distillation
    Moreira, Ricardo
    Bono, Jacopo
    Cardoso, Mario
    Saleiro, Pedro
    Figueiredo, Mario
    Bizarro, Pedro
    CAUSAL LEARNING AND REASONING, VOL 236, 2024, 236 : 740 - 768