Multi-label borderline oversampling technique

被引:10
|
作者
Teng, Zeyu [1 ]
Cao, Peng [2 ,3 ]
Huang, Min [1 ]
Gao, Zheming [1 ]
Wang, Xingwei [2 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang 110819, Liaoning, Peoples R China
[2] Northeastern Univ, Coll Comp Sci & Engn, Shenyang 110169, Liaoning, Peoples R China
[3] Northeastern Univ, Key Lab Intelligent Comp Med Image, Minist Educ, Shenyang 110169, Liaoning, Peoples R China
关键词
Multi-label learning; Class imbalance; Borderline sample; Oversampling; CLASSIFICATION; IMBALANCE; RANKING; MACHINE; SMOTE;
D O I
10.1016/j.patcog.2023.109953
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Class imbalance problem commonly exists in multi-label classification (MLC) tasks. It has non-negligible im-pacts on the classifier performance and has drawn extensive attention in recent years. Borderline oversampling has been widely used in single-label learning as a competitive technique in dealing with class imbalance. Nevertheless, the borderline samples in multi-label data sets (MLDs) have not been studied. Hence, this paper deeply discussed the borderline samples in MLDs and found they have different neighboring relationships with class borders, which makes their roles different in the classifier training. For that, they are divided into two types named the self-borderline samples and the cross-borderline samples. Further, a novel MLDs resampling approach called Multi-Label Borderline Oversampling Technique (MLBOTE) is proposed for multi -label imbalanced learning. MLBOTE identifies three types of seed samples, including interior, self-borderline, and cross-borderline samples, and different oversampling mechanisms are designed for them, respectively. Meanwhile, it regards not only the minority classes but also the classes suffering from one-vs-rest imbalance as those in need of oversampling. Experiments on eight data sets with nine MLC algorithms and three base classifiers are carried out to compare MLBOTE with some state-of-art MLDs resampling techniques. The results show MLBOTE outperforms other methods in various scenarios.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Multi-Label Annotation of Music
    Ahsan, Hiba
    Kumar, Vijay
    Jawahar, C. V.
    2015 EIGHTH INTERNATIONAL CONFERENCE ON ADVANCES IN PATTERN RECOGNITION (ICAPR), 2015, : 150 - 154
  • [42] Multi-label Ensemble Learning
    Shi, Chuan
    Kong, Xiangnan
    Yu, Philip S.
    Wang, Bai
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT III, 2011, 6913 : 223 - 239
  • [43] Multi-Label Adversarial Perturbations
    Song, Qingquan
    Jin, Haifeng
    Huang, Xiao
    Hu, Xia
    2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, : 1242 - 1247
  • [44] Interactive Multi-label Segmentation
    Santner, Jakob
    Pock, Thomas
    Bischof, Horst
    COMPUTER VISION-ACCV 2010, PT I, 2011, 6492 : 397 - 410
  • [45] Privileged Multi-label Learning
    You, Shan
    Xu, Chang
    Wang, Yunhe
    Xu, Chao
    Tao, Dacheng
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3336 - 3342
  • [46] Copula Multi-label Learning
    Liu, Weiwei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [47] The advances in multi-label classification
    Chen, Shijun
    Gao, Lin
    2014 INTERNATIONAL CONFERENCE ON MANAGEMENT OF E-COMMERCE AND E-GOVERNMENT (ICMECG), 2014, : 240 - 245
  • [48] Multi-label contrastive hashing
    Wei Z.
    Jin K.
    Zhang Z.
    Zhou X.
    Pattern Recognition, 2024, 149
  • [49] On the consistency of multi-label learning
    Gao, Wei
    Zhou, Zhi-Hua
    ARTIFICIAL INTELLIGENCE, 2013, 199 : 22 - 44
  • [50] Multi-label Deepfake Classification
    Singh, Inder Pal
    Mejri, Nesryne
    Nguyen, Van Dat
    Ghorbel, Enjie
    Aouada, Djamila
    2023 IEEE 25TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, MMSP, 2023,