Progressive Training Technique with Weak-Label Boosting for Fine-Grained Classification on Unbalanced Training Data

被引:0
|
作者
Jin, Yuhui [1 ]
Wang, Zuyun [2 ]
Liao, Huimin [2 ]
Zhu, Sainan [3 ]
Tong, Bin [3 ]
Yin, Yu [4 ]
Huang, Jian [1 ]
机构
[1] Beihang Univ, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[2] Beijing Transportat Comprehens Law Enforcement Co, Beijing 100044, Peoples R China
[3] China Inst Geoenvironm Monitoring, Beijing 100081, Peoples R China
[4] Peking Univ, Affiliated High Sch, Beijing 102218, Peoples R China
关键词
unbalanced training data; progressive training; weak-label boosting; instance-aware hard ID mining strategy; feature-mapping loss;
D O I
10.3390/electronics11111684
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In practical classification tasks, the sample distribution of the dataset is often unbalanced; for example, this is the case in a dataset that contains a massive quantity of samples with weak labels and for which concrete identification is unavailable. Even in samples with exact labels, the number of samples corresponding to many labels is small, resulting in difficulties in learning the concepts through a small number of labeled samples. In addition, there is always a small interclass variance and a large intraclass variance among categories. Weak labels, few-shot problems, and fine-grained analysis are the key challenges affecting the performance of the classification model. In this paper, we develop a progressive training technique to address the few-shot challenge, along with a weak-label boosting method, by considering all of the weak IDs as negative samples of every predefined ID in order to take full advantage of the more numerous weak-label data. We introduce an instance-aware hard ID mining strategy in the classification loss and then further develop the global and local feature-mapping loss to expand the decision margin. We entered the proposed method into the Kaggle competition, which aims to build an algorithm to identify individual humpback whales in images. With a few other common training tricks, the proposed approach won first place in the competition. All three problems (weak labels, few-shot problems, and fine-grained analysis) exist in the dataset used in the competition. Additionally, we applied our method to CUB-2011 and Cars-196, which are the most widely-used datasets for fine-grained visual categorization tasks, and achieved respective accuracies of 90.1% and 94.9%. This experiment shows that the proposed method achieves the optimal effect compared with other common baselines, and verifies the effectiveness of our method. Our solution has been made available as an open source project.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] FPAN: Fine-grained and progressive attention localization network for data retrieval
    Chen, Sijia
    Song, Bin
    Guo, Jie
    Zhang, Yanling
    Du, Xiaojiang
    Guizani, Mohsen
    COMPUTER NETWORKS, 2018, 143 : 98 - 111
  • [32] Data reweighting net for web fine-grained image classification
    Liu, Yifeng
    Wu, Zhenxin
    Lo, Sio-long
    Chen, Zhenqiang
    Ke, Gang
    Yue, Chuan
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (33) : 79985 - 80005
  • [33] A Data Augmentation Based ViT for Fine-Grained Visual Classification
    Yuan, Shuozhi
    Guo, Wenming
    Han, Fang
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT II, 2023, 14255 : 1 - 12
  • [34] A Fine-Grained Access Control Model with Secure Label on Data Resource
    Gao, Lijie
    Liu, Lianzhong
    Jin, Ze
    Han, Chunyan
    2013 3RD INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND NETWORK TECHNOLOGY (ICCSNT), 2013, : 14 - 18
  • [35] FGRCAT: A fine-grained reasoning framework through causality and adversarial training
    Guo, Hanghui
    Di, Shimin
    Chen, Zhangze
    Pan, Changfan
    Meng, Chaojun
    Zhu, Jia
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 272
  • [36] Fine-Grained Entity Type Classification by Jointly Learning Representations and Label Embeddings
    Abhishek
    Anand, Ashish
    Awekar, Amit
    15TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2017), VOL 1: LONG PAPERS, 2017, : 797 - 807
  • [37] Coarse2Fine: Fine-grained Text Classification on Coarsely-grained Annotated Data
    Mekala, Dheeraj
    Gangal, Varun
    Shang, Jingbo
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 583 - 594
  • [38] Snuba: Automating Weak Supervision to Label Training Data
    Varma, Paroma
    Re, Christopher
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2018, 12 (03): : 223 - 236
  • [39] Combining simulated data, foundation models, and few real samples for training fine-grained object detectors
    Heslinga, Friso G.
    Eker, Thijs A.
    Fokkinga, Ella P.
    van Woerden, Jan Erik
    Ruis, Frank A.
    den Hollander, Richard J. M.
    Schutte, Klamer
    SYNTHETIC DATA FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING: TOOLS, TECHNIQUES, AND APPLICATIONS II, 2024, 13035
  • [40] Fine-Grained Entity Typing via Label Noise Reduction and Data Augmentation
    Li, Haoyang
    Lin, Xueling
    Chen, Lei
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT I, 2021, 12681 : 356 - 374