IGTree: Using trees for compression and classification in lazy learning algorithms

被引:65
|
作者
Daelemans, W [1 ]
VandenBosch, A [1 ]
Weijters, T [1 ]
机构
[1] MAASTRICHT UNIV,MATRIKS,MAASTRICHT,NETHERLANDS
关键词
lazy learning; eager learning; decision trees; information gain; data compression; instance base indexing;
D O I
10.1023/A:1006506017891
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We describe the IGTree learning algorithm, which compresses an instance base into a tree structure. The concept of information gain is used as a heuristic function for performing this compression. IGTree produces trees that, compared to other lazy learning approaches, reduce storage requirements and the time required to compute classifications. Furthermore, we obtained similar or better generalization accuracy with IGTree when trained on two complex linguistic tasks, viz. letter-phoneme transliteration and part-of-speech-tagging, when compared to alternative lazy learning and decision tree approaches (viz., IB1, information-gain-weighted IB1, and C4.5). A third experiment, with the task of word hyphenation, demonstrates that when the mutual differences in information gain of features is too small, IGTree as well as information-gain-weighted IB1 perform worse than IB1. These results indicate that IGTree is a useful algorithm for problems characterized by the availability of a large number of training instances described by symbolic features with sufficiently differing information gain values.
引用
收藏
页码:407 / 423
页数:17
相关论文
共 50 条
  • [1] IGTree: Using Trees for Compression and Classification in Lazy Learning Algorithms
    Walter Daelemans
    Antal Van Den Bosch
    Ton Weijters
    Artificial Intelligence Review, 1997, 11 : 407 - 423
  • [2] Discretisation in Lazy Learning Algorithms
    Kai Ming Ting
    Artificial Intelligence Review, 1997, 11 : 157 - 174
  • [3] Discretisation in lazy learning algorithms
    Ting, KM
    ARTIFICIAL INTELLIGENCE REVIEW, 1997, 11 (1-5) : 157 - 174
  • [4] A Lazy Approach for Machine Learning Algorithms
    Galván, Inés M.
    Valls, José M.
    Lecomte, Nicolas
    Isasi, Pedro
    IFIP Advances in Information and Communication Technology, 2009, 296 : 517 - 522
  • [5] A Lazy Approach for Machine Learning Algorithms
    Galvan, Ines M.
    Valls, Jose M.
    Lecomte, Nicolas
    Isasi, Pedro
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS III, 2009, : 517 - 522
  • [6] An Empirical Study of Lazy Multilabel Classification Algorithms
    Spyromitros, E.
    Tsoumakas, G.
    Vlahavas, I.
    ARTIFICIAL INTELLIGENCE: THEORIES, MODELS AND APPLICATIONS, SETN 2008, 2008, 5138 : 401 - 406
  • [7] Lazy learning for improving ranking of decision trees
    Liang, Han
    Yan, Yuhong
    AI 2006: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2006, 4304 : 1173 - +
  • [8] LLAC: Lazy Learning in Associative Classification
    Ibrahim, S. P. Syed
    Chandran, K. R.
    Nataraj, R. V.
    ADVANCES IN COMPUTING AND COMMUNICATIONS, PT I, 2011, 190 : 631 - 638
  • [9] On the consistency of information filters for lazy learning algorithms
    Brighton, H
    Mellish, C
    PRINCIPLES OF DATA MINING AND KNOWLEDGE DISCOVERY, 1999, 1704 : 283 - 288
  • [10] Lazy Learning for Multi-class Classification Using Genetic Programming
    Jabeen, Hajira
    Baig, Abdul Rauf
    ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS: WITH ASPECTS OF ARTIFICIAL INTELLIGENCE, 2012, 6839 : 177 - +