Linear Approximation of F-Measure for the Performance Evaluation of Classification Algorithms on Imbalanced Data Sets

被引:7
|
作者
Wong, Tzu-Tsung [1 ]
机构
[1] Natl Cheng Kung Univ, Inst Informat Management, 1 Ta Sheuh Rd, Tainan 701, Taiwan
关键词
Classification; cross validation; F-measure; imbalanced data set; sampling distribution; STATISTICAL COMPARISONS; K-FOLD; CLASSIFIERS; RECALL;
D O I
10.1109/TKDE.2020.2986749
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Accuracy is a popular measure for evaluating the performance of classification algorithms tested on ordinary data sets. When a data set is imbalanced, F-measure will be a better choice than accuracy for this purpose. Since F-measure is calculated as the harmonic mean of recall and precision, it is difficult to find the sampling distribution of F-measure for evaluating classification algorithms. Since the values of recall and precision are dependent, their joint distribution is assumed to follow a bivariate normal distribution in this study. When the evaluation method is k-fold cross validation, a linear approximation approach is proposed to derive the sampling distribution of F-measure. This approach is used to design methods for comparing the performance of two classification algorithms tested on single or multiple imbalanced data sets. The methods are tested on ten imbalanced data sets to demonstrate their effectiveness. The weight of recall provided by this linear approximation approach can help us to analyze the characteristics of classification algorithms.
引用
收藏
页码:753 / 763
页数:11
相关论文
共 50 条
  • [21] F-measure curves: A tool to visualize classifier performance under imbalance
    Soleymani, Roghayeh
    Granger, Eric
    Fumera, Giorgio
    PATTERN RECOGNITION, 2020, 100
  • [22] From Cost-Sensitive Classification to Tight F-measure Bounds
    Bascol, Kevin
    Emonet, Remi
    Fromont, Elisa
    Habrard, Amaury
    Metzler, Guillaume
    Sebban, Marc
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [23] GP classification under imbalanced data sets: Active sub-sampling and AUC approximation
    Doucette, John
    Heywood, Malcolm I.
    GENETIC PROGRAMMING, PROCEEDINGS, 2008, 4971 : 266 - +
  • [24] The improved AdaBoost algorithms for imbalanced data classification
    Wang, Wenyang
    Sun, Dongchu
    INFORMATION SCIENCES, 2021, 563 : 358 - 374
  • [25] Monotonic classification: An overview on algorithms, performance measures and data sets
    Cano, Jose-Ramon
    Antonio Gutierrez, Pedro
    Krawczyk, Bartosz
    Wozniak, Michal
    Garcia, Salvador
    NEUROCOMPUTING, 2019, 341 : 168 - 182
  • [26] A Performance Evaluation of Classification Algorithms for Big Data
    Hai, Mo
    Zhang, You
    Zhang, Youjin
    5TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND QUANTITATIVE MANAGEMENT, ITQM 2017, 2017, 122 : 1100 - 1107
  • [27] Comparing ϕ and the F-measure as performance metrics for software-related classifications
    Luigi Lavazza
    Sandro Morasca
    Empirical Software Engineering, 2022, 27
  • [28] Deep F-Measure Maximization in Multi-label Classification: A Comparative Study
    Decubber, Stijn
    Mortier, Thomas
    Dembczynski, Krzysztof
    Waegeman, Willem
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT I, 2019, 11051 : 290 - 305
  • [29] Comparing φ and the F-measure as performance metrics for software-related classifications
    Lavazza, Luigi
    Morasca, Sandro
    EMPIRICAL SOFTWARE ENGINEERING, 2022, 27 (07)
  • [30] Optimizing F-Measure with Non-Convex Loss and Sparse Linear Classifiers
    Chinta, Punya Murthy
    Balamurugan, P.
    Shevade, Shirish
    Murty, M. Narasimha
    2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,