Principles and Practice of Explainable Machine Learning

被引:375
作者
Belle, Vaishak [1 ,2 ]
Papantonis, Ioannis [1 ]
机构
[1] Univ Edinburgh, Sch Informat, Edinburgh, Midlothian, Scotland
[2] Alan Turing Inst, London, England
来源
FRONTIERS IN BIG DATA | 2021年 / 4卷
基金
英国工程与自然科学研究理事会;
关键词
survey; explainable AI; black-box models; transparent models; machine learning; NEURAL-NETWORKS; RULE EXTRACTION; EXPLANATION; DECISIONS;
D O I
10.3389/fdata.2021.688969
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Artificial intelligence (AI) provides many opportunities to improve private and public life. Discovering patterns and structures in large troves of data in an automated manner is a core component of data science, and currently drives applications in diverse areas such as computational biology, law and finance. However, such a highly positive impact is coupled with a significant challenge: how do we understand the decisions suggested by these systems in order that we can trust them? In this report, we focus specifically on data-driven methods-machine learning (ML) and pattern recognition models in particular-so as to survey and distill the results and observations from the literature. The purpose of this report can be especially appreciated by noting that ML models are increasingly deployed in a wide range of businesses. However, with the increasing prevalence and complexity of methods, business stakeholders in the very least have a growing number of concerns about the drawbacks of models, data-specific biases, and so on. Analogously, data science practitioners are often not aware about approaches emerging from the academic literature or may struggle to appreciate the differences between different methods, so end up using industry standards such as SHAP. Here, we have undertaken a survey to help industry practitioners (but also data scientists more broadly) understand the field of explainable machine learning better and apply the right tools. Our latter sections build a narrative around a putative data scientist, and discuss how she might go about explaining her models by asking the right questions. From an organization viewpoint, after motivating the area broadly, we discuss the main developments, including the principles that allow us to study transparent models vs. opaque models, as well as model-specific or model-agnostic post-hoc explainability approaches. We also briefly reflect on deep learning models, and conclude with a discussion about future research directions.
引用
收藏
页数:25
相关论文
共 111 条
[1]  
Adebayo J., 2016, FATML WORKSH 2016 NE
[2]  
Agrahari R., 2018, SCI REP-UK
[3]  
[Anonymous], 2016, ARXIV160509196
[4]  
[Anonymous], 2016, 2016 IEEE 26 INT WOR
[5]  
[Anonymous], 2014, ARXIV14085456
[6]   Reverse Engineering the Neural Networks for Rule Extraction in Classification Problems [J].
Augasta, M. Gethsiyal ;
Kathirvalavakumar, T. .
NEURAL PROCESSING LETTERS, 2012, 35 (02) :131-150
[7]   Interpretation of nonlinear relationships between process variables by use of random forests [J].
Auret, Lidia ;
Aldrich, Chris .
MINERALS ENGINEERING, 2012, 35 :27-42
[8]  
Bastani O., 2017, ABS170609773 ARXIV
[9]   STATISTICAL INFERENCE FOR PROBABILISTIC FUNCTIONS OF FINITE STATE MARKOV CHAINS [J].
BAUM, LE ;
PETRIE, T .
ANNALS OF MATHEMATICAL STATISTICS, 1966, 37 (06) :1554-&
[10]  
Belle V., 2019, 9 INT WORKSH STAT RE