Bayesian approach to feature selection and parameter tuning for support vector machine classifiers

被引:50
|
作者
Gold, C [1 ]
Holub, A
Sollich, P
机构
[1] CALTECH, Pasadena, CA 91125 USA
[2] Kings Coll London, Dept Math, London WC2R 2LS, England
关键词
D O I
10.1016/j.neunet.2005.06.044
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A Bayesian point of view of SVM classifiers allows the definition of a quantity analogous to the evidence in probabilistic models. By maximizing this one can systematically tune hyperparameters and, via automatic relevance determination (ARD), select relevant input features. Evidence gradients are expressed as averages over the associated posterior and can be approximated using Hybrid Monte Carlo (HMC) sampling. We describe how a Nystrom approximation of the Gram matrix can be used to speed up sampling times significantly while maintaining almost unchanged classification accuracy. In experiments on classification problems with a significant number of irrelevant features this approach to ARD can give a significant improvement in classification performance over more traditional, non-ARD, SVM systems. The final tuned hyperparameter values provide a useful criterion for pruning irrelevant features, and we define a measure of relevance with which to determine systematically how many features should be removed. This use of ARD for hard feature selection can improve classification accuracy in non-ARD SVMs. In the majority of cases, however, we find that in data sets constructed by human domain experts the performance of non-ARD SVMs is largely insensitive to the presence of some less relevant features. Eliminating such features via ARD then does not improve classification accuracy, but leads to impressive reductions in the number of features required, by up to 75%.(1) (c) 2005 Elsevier Ltd. All rights reserved.
引用
收藏
页码:693 / 701
页数:9
相关论文
共 50 条
  • [41] Electromagnetism-like algorithm for support vector machine parameter tuning
    Aleksandar Kartelj
    Nenad Mitić
    Vladimir Filipović
    Dušan Tošić
    Soft Computing, 2014, 18 : 1985 - 1998
  • [42] Parameter Tuning via Kernel Matrix Approximation for Support Vector Machine
    Yang, Chenhao
    Ding, Lizhong
    Liao, Shizhong
    JOURNAL OF COMPUTERS, 2012, 7 (08) : 2047 - 2054
  • [43] Electromagnetism-like algorithm for support vector machine parameter tuning
    Kartelj, Aleksandar
    Mitic, Nenad
    Filipovic, Vladimir
    Tosic, Dusan
    SOFT COMPUTING, 2014, 18 (10) : 1985 - 1998
  • [44] A feature selection Newton method for support vector machine classification
    Fung, GM
    Mangasarian, OL
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2004, 28 (02) : 185 - 202
  • [45] Automated parameter selection for support vector machine decision tree
    Choi, Gyunghyun
    Bae, Suk Joo
    NEURAL INFORMATION PROCESSING, PT 2, PROCEEDINGS, 2006, 4233 : 746 - 753
  • [46] A novel feature selection method for twin support vector machine
    Bai, Lan
    Wang, Zhen
    Shao, Yuan-Hai
    Deng, Nai-Yang
    KNOWLEDGE-BASED SYSTEMS, 2014, 59 : 1 - 8
  • [47] Sparse Support Vector Machine with Lp Penalty for Feature Selection
    Lan Yao
    Feng Zeng
    Dong-Hui Li
    Zhi-Gang Chen
    Journal of Computer Science and Technology, 2017, 32 : 68 - 77
  • [48] Novel linear search for support vector machine parameter selection
    Hong-xia Pang
    Wen-de Dong
    Zhi-hai Xu
    Hua-jun Feng
    Qi Li
    Yue-ting Chen
    Journal of Zhejiang University SCIENCE C, 2011, 12 : 885 - 896
  • [49] Novel linear search for support vector machine parameter selection
    Hongxia PANG Wende DONG Zhihai XU Huajun FENG Qi LI Yueting CHEN State Key Laboratory of Optical Instrumentation Zhejiang University Hangzhou China
    Journal of Zhejiang University-Science C(Computers & Electronics), 2011, 12 (11) : 885 - 896