Calibrated lazy associative classification

被引:21
|
作者
Veloso, Adriano [1 ]
Meira, Wagner, Jr. [1 ]
Goncalves, Marcos [1 ]
Almeida, Humberto M. [1 ]
Zaki, Mohammed [2 ]
机构
[1] Univ Fed Minas Gerais, Dept Comp Sci, Belo Horizonte, MG, Brazil
[2] Rensselaer Polytech Inst, Dept Comp Sci, Troy, NY 12181 USA
关键词
Classification; MDL; Calibration;
D O I
10.1016/j.ins.2010.03.007
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Classification is a popular machine learning task. Given an example x and a class c, a classifier usually works by estimating the probability of x being member of c (i.e., membership probability). Well calibrated classifiers are those able to provide accurate estimates of class membership probabilities, that is, the estimated probability (p) over cap (c vertical bar x) is close to p(c vertical bar(p) over cap (c vertical bar x)), which is the true, (unknown) empirical probability of x being member of c given that the probability estimated by the classifier is (p) over cap (c vertical bar x). Calibration is not a necessary property for producing accurate classifiers, and, thus, most of the research has focused on direct accuracy maximization strategies rather than on calibration. However, non-calibrated classifiers are problematic in applications where the reliability associated with a prediction must be taken into account. In these applications, a sensible use of the classifier must be based on the reliability of its predictions, and, thus, the classifier must be well calibrated. In this paper we show that lazy associative classifiers (LAC) are well calibrated using an MM.:based entropy minimization method. We investigate important applications where such characteristics (i.e., accuracy and calibration) are relevant, and we demonstrate empirically that LAC outperforms other classifiers, such as SVMs, Naive Bayes, and Decision Trees (even after these classifiers are calibrated). Additional highlights of LAC include the ability to incorporate reliable predictions for improving training, and the ability to refrain from doubtful predictions. (C) 2010 Elsevier Inc. All rights reserved.
引用
收藏
页码:2656 / 2670
页数:15
相关论文
共 50 条
  • [11] Lazy approach for ordinal classification
    Wang, JF
    Wang, XZ
    PROCEEDINGS OF THE 11TH JOINT INTERNATIONAL COMPUTER CONFERENCE, 2005, : 395 - 398
  • [12] A lazy bagging approach to classification
    Zhu, Xingquan
    Yang, Ying
    PATTERN RECOGNITION, 2008, 41 (10) : 2980 - 2992
  • [13] A lazy approach to pruning classification rules
    Baralis, E
    Garza, P
    2002 IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2002, : 35 - 42
  • [14] Evolutionary Lazy Learning for Naive Bayes Classification
    Bai, Yu
    Wang, Haishuai
    Wu, Jia
    Zhang, Yun
    Jiang, Jing
    Long, Guodong
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 3124 - 3129
  • [15] A lazy data mining approach for protein classification
    Merschmann, Luiz
    Plastino, Alexandre
    IEEE TRANSACTIONS ON NANOBIOSCIENCE, 2007, 6 (01) : 36 - 42
  • [16] Shapelet Discovery by Lazy Time Series Classification
    Zhang, Wei
    Wang, Zhihai
    Yuan, Jidong
    Hao, Shilei
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2020, 2020
  • [17] Lazy Learning for Classification Based on Query Projections
    Han, Yiqiu
    Lam, Wai
    PROCEEDINGS OF THE FIFTH SIAM INTERNATIONAL CONFERENCE ON DATA MINING, 2005, : 227 - 238
  • [18] A Lazy Learning Approach for Building Classification Models
    Galvan, Ines M.
    Valls, Jose M.
    Garcia, Miguel
    Isasi, Pedro
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2011, 26 (08) : 773 - 786
  • [19] Classification of symbolic objects: A lazy learning approach
    Appice, Annalisa
    D'Amato, Claudia
    Esposito, Floriana
    Malerba, Donato
    INTELLIGENT DATA ANALYSIS, 2006, 10 (04) : 301 - 324
  • [20] An Empirical Study of Lazy Multilabel Classification Algorithms
    Spyromitros, E.
    Tsoumakas, G.
    Vlahavas, I.
    ARTIFICIAL INTELLIGENCE: THEORIES, MODELS AND APPLICATIONS, SETN 2008, 2008, 5138 : 401 - 406