Interpretation of linear classifiers by means of feature relevance bounds

被引:7
|
作者
Goepfert, Christina [1 ]
Pfannschmidt, Lukas [1 ]
Goepfert, Jan Philip [1 ]
Hammer, Barbara [1 ]
机构
[1] Cognit Interact Technol, Inspirat 1, D-33619 Bielefeld, Germany
关键词
Feature relevance; Feature selection; Interpretability; All-relevant; Linear classification; FEATURE-SELECTION;
D O I
10.1016/j.neucom.2017.11.074
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Research on feature relevance and feature selection problems goes back several decades, but the importance of these areas continues to grow as more and more data becomes available, and machine learning methods are used to gain insight and interpret, rather than solely to solve classification or regression problems. Despite the fact that feature relevance is often discussed, it is frequently poorly defined, and the feature selection problems studied are subtly different. Furthermore, the problem of finding all features relevant for a classification problem has only recently started to gain traction, despite its importance for interpretability and integrating expert knowledge. In this paper, we attempt to unify commonly used concepts and to give an overview of the main questions and results. We formalize two interpretations of the all-relevant problem and propose a polynomial method to approximate one of them for the important hypothesis class of linear classifiers, which also enables a distinction between strongly and weakly relevant features. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:69 / 79
页数:11
相关论文
共 50 条
  • [1] Valid Interpretation of Feature Relevance for Linear Data Mappings
    Frenay, Benoit
    Hofmann, Daniela
    Schulz, Alexander
    Biehl, Michael
    Hammer, Barbara
    2014 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DATA MINING (CIDM), 2014, : 149 - 156
  • [2] New bounds and approximations for the error of linear classifiers
    Rueda, L
    PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS AND APPLICATIONS, 2004, 3287 : 342 - 349
  • [3] Feature Shaping for Linear SVM Classifiers
    Forman, George
    Scholz, Martin
    Rajaram, Shyamsundar
    KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2009, : 299 - 307
  • [4] DIVERGENCE AND LINEAR CLASSIFIERS FOR FEATURE SELECTION
    CARDILLO, GP
    FU, KS
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1967, AC12 (06) : 780 - &
  • [5] Upper bounds for error rates of linear combinations of classifiers
    Murua, A
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (05) : 591 - 602
  • [6] SQ Lower Bounds for Learning Mixtures of Linear Classifiers
    Diakonikolas, Ilias
    Kane, Daniel M.
    Sun, Yuxin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Generalization Error Bounds for Multiclass Sparse Linear Classifiers
    Levy, Tomer
    Abramovich, Felix
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [8] The Impact of Feature Importance Methods on the Interpretation of Defect Classifiers
    Rajbahadur, Gopi Krishnan
    Wang, Shaowei
    Oliva, Gustavo A.
    Kamei, Yasutaka
    Hassan, Ahmed E.
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2021, 48 (07) : 2245 - 2261
  • [9] The Impact of Feature Importance Methods on the Interpretation of Defect Classifiers
    Rajbahadur, Gopi Krishnan
    Wang, Shaowei
    Oliva, Gustavo A.
    Kamei, Yasutaka
    Hassan, Ahmed E.
    IEEE Transactions on Software Engineering, 2022, 48 (07) : 2245 - 2261
  • [10] Transfer bounds for linear feature learning
    Maurer, Andreas
    MACHINE LEARNING, 2009, 75 (03) : 327 - 350