Complement-Class Harmonized Naive Bayes Classifier

被引:2
|
作者
Alenazi, Fahad S. [1 ]
El Hindi, Khalil [1 ]
AsSadhan, Basil [2 ]
机构
[1] King Saud Univ, Dept Comp Sci, Riyadh 11543, Saudi Arabia
[2] King Saud Univ, Dept Elect Engn, Riyadh 11421, Saudi Arabia
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 08期
关键词
scarce data; harmonic average; attribute weighting; Naive Bayes;
D O I
10.3390/app13084852
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Naive Bayes (NB) classification performance degrades if the conditional independence assumption is not satisfied or if the conditional probability estimate is not realistic due to the attributes of correlation and scarce data, respectively. Many works address these two problems, but few works tackle them simultaneously. Existing methods heuristically employ information theory or applied gradient optimization to enhance NB classification performance, however, to the best of our knowledge, the enhanced model generalization capability deteriorated especially on scant data. In this work, we propose a fine-grained boosting of the NB classifier to identify hidden and potential discriminative attribute values that lead the NB model to underfit or overfit on the training data and to enhance their predictive power. We employ the complement harmonic average of the conditional probability terms to measure their distribution divergence and impact on the classification performance for each attribute value. The proposed method is subtle yet significant enough in capturing the attribute values' inter-correlation (between classes) and intra-correlation (within the class) and elegantly and effectively measuring their impact on the model's performance. We compare our proposed complement-class harmonized Naive Bayes classifier (CHNB) with the state-of-the-art Naive Bayes and imbalanced ensemble boosting methods on general and imbalanced machine-learning benchmark datasets, respectively. The empirical results demonstrate that CHNB significantly outperforms the compared methods.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] Mining housekeeping genes with a Naive Bayes classifier
    De Ferrari, Luna
    Aitken, Stuart
    BMC GENOMICS, 2006, 7 (1)
  • [42] Mining housekeeping genes with a Naive Bayes classifier
    Luna De Ferrari
    Stuart Aitken
    BMC Genomics, 7
  • [43] An Aggregated Fuzzy Naive Bayes Data Classifier
    Tutuncu, G. Yazgi
    Kayaalp, Necla
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2015, 286 : 17 - 27
  • [44] DECOMPOSABLE NAIVE BAYES CLASSIFIER FOR PARTITIONED DATA
    Khedr, Ahmed M.
    COMPUTING AND INFORMATICS, 2012, 31 (06) : 1511 - 1531
  • [45] Mixture of latent multinomial naive Bayes classifier
    Harzevili, Nima Shiri
    Alizadeh, Sasan H.
    APPLIED SOFT COMPUTING, 2018, 69 : 516 - 527
  • [46] Extended Naive Bayes classifier for mixed data
    Hsu, Chung-Chian
    Huang, Yan-Ping
    Chang, Keng-Wei
    EXPERT SYSTEMS WITH APPLICATIONS, 2008, 35 (03) : 1080 - 1083
  • [47] Alpha Skew Gaussian Naive Bayes Classifier
    Ara, Anderson
    Louzada, Francisco
    INTERNATIONAL JOURNAL OF INFORMATION TECHNOLOGY & DECISION MAKING, 2022, 21 (01) : 441 - 462
  • [48] Prediction of conotoxin superfamilies by the Naive Bayes classifier
    Huo, Haiyan
    Yang, Lei
    Proceedings - 2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics, CISP-BMEI 2017, 2017, 2018-January : 1 - 5
  • [49] Prediction of Conotoxin Superfamilies by the Naive Bayes Classifier
    Huo, Haiyan
    Yang, Lei
    2017 10TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI), 2017,
  • [50] The Application of Naive Bayes Classifier in Name Disambiguation
    Li, Na
    Han, Jin
    CLOUD COMPUTING AND SECURITY, PT II, 2017, 10603 : 611 - 618