A sparse focus framework for visual fine-grained classification

被引:0
|
作者
YongXiong Wang
Guangjun Li
Li Ma
机构
[1] University of Shanghai for Science and Technology,School of Optical
来源
关键词
Fine-grained visual recognition; Sparse focus framework; Network sparsity; Network pruning;
D O I
暂无
中图分类号
学科分类号
摘要
The location of discriminative features and reduction of model complexity are the two main research directions in fine-grained image classification. The manual annotation of object is very labor-intensive, and the commonly used model compression methods usually reduce the classification accuracy while compressing the model. In this paper, we propose a Sparse Focus Framework(SFF) based on Bilinear Convolutional Neural Network(BCNN), which includes self-focus module and sparse scaling factors. The focus module like the focus function of human beings automatically locates the object from background without manual labeling, and only a small amount of computing resources are occupied in the guaranteed accuracy. The sparse scaling factor for each channel can evaluate the importance of feature channels, which is adopted in pruning of channels. The large number of parameters and calculations in the fine-grained classification model can been effectively reduced by pruning method adopted in our network, which can obtain a sparse structure to prevent overfitting while maintaining classification performance. Our experimental results show that our model obtains accuracy of 90.2%, 84.5% and 92.0% on FGVC-aircraft, Stanford dogs and Stanford cars, respectively. Compared with the highest classification accuracy obtained by the same classification network B-CNN[D,D], the accuracy gains with 6.1%, 4.1% and 1.4% respectively. Moreover, the channel-level sparsity effectively reduces 30% of the network parameters and nearly 13% of the computation.
引用
收藏
页码:25271 / 25289
页数:18
相关论文
共 50 条
  • [41] Fine-Grained Visual Classification via Internal Ensemble Learning Transformer
    Xu, Qin
    Wang, Jiahui
    Jiang, Bo
    Luo, Bin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 9015 - 9028
  • [42] Dual-Dependency Attention Transformer for Fine-Grained Visual Classification
    Cui, Shiyan
    Hui, Bin
    SENSORS, 2024, 24 (07)
  • [43] Progressive Erasing Network with consistency loss for fine-grained visual classification
    Peng, Jin
    Wang, Yongxiong
    Zhou, Zeping
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 87
  • [44] Progressive Co-Attention Network for Fine-Grained Visual Classification
    Zhang, Tian
    Chang, Dongliang
    Ma, Zhanyu
    Guo, Jun
    2021 INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2021,
  • [45] Classification-Specific Parts for Improving Fine-Grained Visual Categorization
    Korsch, Dimitri
    Bodesheim, Paul
    Denzler, Joachim
    PATTERN RECOGNITION, DAGM GCPR 2019, 2019, 11824 : 62 - 75
  • [46] Image local structure information learning for fine-grained visual classification
    Jin Lu
    Weichuan Zhang
    Yali Zhao
    Changming Sun
    Scientific Reports, 12
  • [47] Multi-directional guidance network for fine-grained visual classification
    Yang, Shengying
    Jin, Yao
    Lei, Jingsheng
    Zhang, Shuping
    VISUAL COMPUTER, 2024, 40 (11): : 8113 - 8124
  • [48] Integrating Scene Text and Visual Appearance for Fine-Grained Image Classification
    Bai, Xiang
    Yang, Mingkun
    Lyu, Pengyuan
    Xu, Yongchao
    Luo, Jiebo
    IEEE ACCESS, 2018, 6 : 66322 - 66335
  • [49] Hierarchical Fine-Grained Visual Classification Leveraging Consistent Hierarchical Knowledge
    Liu, Yuting
    Yang, Liu
    Wang, Yu
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT I, ECML PKDD 2024, 2024, 14941 : 279 - 295
  • [50] Image local structure information learning for fine-grained visual classification
    Lu, Jin
    Zhang, Weichuan
    Zhao, Yali
    Sun, Changming
    SCIENTIFIC REPORTS, 2022, 12 (01)