DSPOTE: Density-induced Selection Probability-based Oversampling TEchnique for Imbalanced Learning

被引:0
|
作者
Wei, Zhen [1 ]
Zhang, Li [1 ]
Zhao, Lei [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215006, Peoples R China
关键词
SAMPLING METHOD; SMOTE; CLASSIFICATION;
D O I
10.1109/ICPR56361.2022.9956583
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In imbalanced learning, oversampling is incredibly prevalent. However, it is disappointing that existing oversampling methods have their own limitations, such as new synthetic samples may be uninformative or noisy. To better address imbalanced learning tasks, this paper proposes a novel oversampling method, named Density-induced Selection Probability-based Over-sampling TEchnique (DSPOTE). To increase the number of samples in the minority class, DSPOTE designs a novel scheme for filtering noisy samples based on the Chebychev distance and a new way of calculating selection probability based on relative density. DSPOTE first filters noisy samples and then gets borderline ones from the minority class. Next, DSPOTE calculates the selection probabilities for all borderline samples and applies these probabilities to pick up borderline samples. Finally, DSPOTE generates synthetic samples for the minority class based on the selected borderline ones. Experimental results indicate that our method has good performance in terms of metrics, recall and AUC (Area Under the Curve), when compared with other eight methods.
引用
收藏
页码:2165 / 2171
页数:7
相关论文
共 50 条
  • [21] Perturbation-based oversampling technique for imbalanced classification problems
    Jianjun Zhang
    Ting Wang
    Wing W. Y. Ng
    Witold Pedrycz
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 773 - 787
  • [22] CCO: A Cluster Core-Based Oversampling Technique for Improved Class-Imbalanced Learning
    Mondal, Priyobrata
    Ansari, Faizanuddin
    Das, Swagatam
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, : 1 - 13
  • [23] Improving interpolation-based oversampling for imbalanced data learning
    Zhu, Tuanfei
    Lin, Yaping
    Liu, Yonghe
    KNOWLEDGE-BASED SYSTEMS, 2020, 187
  • [24] A theoretical distribution analysis of synthetic minority oversampling technique (SMOTE) for imbalanced learning
    Elreedy, Dina
    Atiya, Amir F.
    Kamalov, Firuz
    MACHINE LEARNING, 2024, 113 (07) : 4903 - 4923
  • [25] A Novel Oversampling Method for Imbalanced Datasets Based on Density Peaks Clustering
    Cao, Jie
    Shi, Yong
    TEHNICKI VJESNIK-TECHNICAL GAZETTE, 2021, 28 (06): : 1813 - 1819
  • [26] Oversampling technique based on fuzzy representativeness difference for classifying imbalanced data
    Ren, Ruonan
    Yang, Youlong
    Sun, Liqin
    APPLIED INTELLIGENCE, 2020, 50 (08) : 2465 - 2487
  • [27] A No Parameter Synthetic Minority Oversampling Technique Based on Finch for Imbalanced Data
    Xu, Shoukun
    Li, Zhibang
    Yuan, Baohua
    Yang, Gaochao
    Wang, Xueyuan
    Li, Ning
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT IV, 2023, 14089 : 367 - 378
  • [28] Oversampling technique based on fuzzy representativeness difference for classifying imbalanced data
    Ruonan Ren
    Youlong Yang
    Liqin Sun
    Applied Intelligence, 2020, 50 : 2465 - 2487
  • [29] Boosting the oversampling methods based on differential evolution strategies for imbalanced learning
    Korkmaz, Sedat
    Sahman, Mehmet Akif
    Cinar, Ahmet Cevahir
    Kaya, Ersin
    APPLIED SOFT COMPUTING, 2021, 112
  • [30] EDOS: Entropy Difference-based Oversampling Approach for Imbalanced Learning
    Li, Lusi
    He, Haibo
    Li, Jie
    Li, Weijun
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,