Fine-Grained Car Recognition Model Based on Semantic DCNN Features Fusion

被引:0
|
作者
Yang J. [1 ]
Cao H. [1 ]
Wang R. [1 ]
Xue L. [1 ]
机构
[1] School of Computer and Information, Hefei University of Technology, Hefei
来源
Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics | 2019年 / 31卷 / 01期
关键词
Car recognition; Convolutional neural networks; Deep learning; Fine-grained car recognition; Fine-grained recognition; Image classification;
D O I
10.3724/SP.J.1089.2019.17130
中图分类号
学科分类号
摘要
As the deep convolution neural networks (DCNN) lack the ability of representation of semantic information, while visual differences between classes are small and concentrated on key semantic parts during the fine-grained categorization, this paper proposes a model based on fusing semantic information of DCNN features, which is composed of the detection sub-network and the classification sub-network. Firstly, by use of the detection sub-network we capture the definite position of car object and each semantic parts through Faster RCNN. Secondly, the classification sub-network extracts the overall car object features and semantic parts features of the object via DCNN, then processes the joint and fusion of features by using small kernel convolution. Finally, we obtain final recognition result through deep neural network. The recognition accuracy of our model is 78.74% in Stanford BMW-10 dataset, which is 13.39% higher than the VGG network method and 85.94% in the Stanford cars-197 dataset. And the recognition accuracy of the transfer learning models in BMVC car-types dataset is 98.27%, which is 3.77% higher than the best recognition result of the dataset. Experimental results show that our model avoids the dependence of the fine-grained car recognition on the positions of car object and semantic parts, with high recognition accuracy and versatility. © 2019, Beijing China Science Journal Publishing Co. Ltd. All right reserved.
引用
收藏
页码:141 / 157
页数:16
相关论文
共 42 条
  • [21] Belhumeur P.N., Chen D.Z., Feiner S., Et al., Searching the world's herbaria: a system for visual identification of plant species, Proceedings of the 10th European Conference on Computer Vision, pp. 116-129, (2008)
  • [22] Krause J., Stark M., Deng J., Et al., 3D object representations for fine-grained categorization, Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 554-561, (2013)
  • [23] Krause J., Gebru T., Deng J., Et al., Learning features and parts for fine-grained recognition, Proceedings of the 22nd IEEE International Conference on Pattern Recognition, pp. 26-33, (2014)
  • [24] Gavves E., Fernando B., Snoek C.G.M., Et al., Local alignments for fine-grained categorization, International Journal of Computer Vision, 111, 2, pp. 191-212, (2015)
  • [25] Zhang N., Farrell R., Iandola F.N., Et al., Deformable part descriptors for fine-grained recognition and attribute prediction, Proceedings of the IEEE International Conference on Computer Vision, pp. 729-736, (2013)
  • [26] Pearce G., Pears N., Automatic make and model recognition from frontal images of cars, Proceedings of the 8th IEEE International Conference on Advanced Video and Signal Based Surveillance, pp. 373-378, (2011)
  • [27] Lazebnik S., Schmid C., Ponce J., Et al., Beyond bags of features: spatial pyramid matching for recognizing natural scene categories, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2169-2178, (2006)
  • [28] Wang J.J., Yang J.C., Yu K., Et al., Locality-constrained linear coding for image classification, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3360-3367, (2010)
  • [29] Deng J., Krause J., Li F.F., Et al., Fine-grained crowdsourcing for fine-grained recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580-587, (2013)
  • [30] Xie S.N., Yang T.B., Wang X.Y., Et al., Hyper-class augmented and regularized deep learning for fine-grained image classification, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2645-2654, (2015)