IGTree: Using trees for compression and classification in lazy learning algorithms

被引:65
|
作者
Daelemans, W [1 ]
VandenBosch, A [1 ]
Weijters, T [1 ]
机构
[1] MAASTRICHT UNIV,MATRIKS,MAASTRICHT,NETHERLANDS
关键词
lazy learning; eager learning; decision trees; information gain; data compression; instance base indexing;
D O I
10.1023/A:1006506017891
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We describe the IGTree learning algorithm, which compresses an instance base into a tree structure. The concept of information gain is used as a heuristic function for performing this compression. IGTree produces trees that, compared to other lazy learning approaches, reduce storage requirements and the time required to compute classifications. Furthermore, we obtained similar or better generalization accuracy with IGTree when trained on two complex linguistic tasks, viz. letter-phoneme transliteration and part-of-speech-tagging, when compared to alternative lazy learning and decision tree approaches (viz., IB1, information-gain-weighted IB1, and C4.5). A third experiment, with the task of word hyphenation, demonstrates that when the mutual differences in information gain of features is too small, IGTree as well as information-gain-weighted IB1 perform worse than IB1. These results indicate that IGTree is a useful algorithm for problems characterized by the availability of a large number of training instances described by symbolic features with sufficiently differing information gain values.
引用
收藏
页码:407 / 423
页数:17
相关论文
共 50 条
  • [31] Optimizing weighted lazy learning and Naive Bayes classification using differential evolution algorithm
    Bai, Yu
    Bain, Michael
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 13 (6) : 3005 - 3024
  • [32] Intelligent selection of instances for prediction functions in lazy learning algorithms
    Utah State Univ, Logan, United States
    Artif Intell Rev, 1-5 (175-191):
  • [33] Intelligent Selection of Instances for Prediction Functions in Lazy Learning Algorithms
    Jianping Zhang
    Yee-Sat Yim
    Jumming Yang
    Artificial Intelligence Review, 1997, 11 : 175 - 191
  • [34] Optimizing weighted lazy learning and Naive Bayes classification using differential evolution algorithm
    Yu Bai
    Michael Bain
    Journal of Ambient Intelligence and Humanized Computing, 2022, 13 : 3005 - 3024
  • [35] Lazy fine-tuning algorithms for naive Bayesian text classification
    El Hindi, Khalil M.
    Aljulaidan, Reem R.
    AlSalman, Hussien
    APPLIED SOFT COMPUTING, 2020, 96
  • [36] Comparison of lazy classification algorithms based on deterministic and inhibitory decision rules
    Delimata, Pawel
    Moshkov, Mikhail
    Skowron, Andrzej
    Suraj, Zbigniew
    ROUGH SETS AND KNOWLEDGE TECHNOLOGY, 2008, 5009 : 55 - +
  • [37] Intelligent selection of instances for prediction functions in lazy learning algorithms
    Zhang, JP
    Yim, YS
    Yang, JM
    ARTIFICIAL INTELLIGENCE REVIEW, 1997, 11 (1-5) : 175 - 191
  • [38] Classification and function estimation of protein by using data compression and genetic algorithms
    Chiba, S
    Sugawara, K
    Watanabe, T
    PROCEEDINGS OF THE 2001 CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1 AND 2, 2001, : 839 - 844
  • [39] Zonda wind classification using machine learning algorithms
    Otero, Federico
    Araneo, Diego
    INTERNATIONAL JOURNAL OF CLIMATOLOGY, 2021, 41 (S1) : E342 - E353
  • [40] Classification of Cardiac Arrhythmias Using Machine Learning Algorithms
    Garcia-Aquino, Christian
    Mujica-Vargas, Dante
    Matuz-Cruz, Manuel
    TELEMATICS AND COMPUTING, WITCOM 2021, 2021, 1430 : 174 - 185