Deep Hashing Retrieval Algorithm Combing Attention Model and Bimodal Gaussian Distribution

被引:0
|
作者
Li Z. [1 ]
Zhang P. [1 ]
Liu Y. [1 ]
Li H. [2 ]
机构
[1] College of Computer & Communication Engineering, China University of Petroleum (East China), Qingdao
[2] Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences, Beijing
关键词
Attention model; Bimodal Gaussian distribution; Deep hashing; Image retrieval;
D O I
10.3724/SP.J.1089.2020.17889
中图分类号
学科分类号
摘要
Hash retrieval has attracted wide attention because of its small storage space and fast retrieval speed. There are two problems in deep hashing methods: Deep hash codes are essentially binary features, and the coding length is short, so their feature representation abilities are limited; In addition, existing deep hashing algorithms cannot directly learn discrete hash codes by backpropagation, and usually relax discrete values to continuous values in their optimization procedure, so this leads to quantization errors. Aiming at above problems, we propose a deep hashing retrieval algorithm combining attention model and bimodal Gaussian distribution. The network structure with spatial and channel attention model, focusing on important features and suppressing unnecessary features, enhances the feature representation abilities of hash codes; To solve the quantization error problem, the bimodal Gaussian distribution with the mean of either +1 or -1 is used as the prior distribution. We refer to the idea of variational auto-encoder, and propose to constrain the hash codes distribution to obey the prior distribution with KL divergence. The mean average precision of our method on three benchmark databases CIFAR-10, ImageNet-100 and NUS-WIDE is better than other methods of comparison, which verifies the effectiveness of the algorithm in this paper. © 2020, Beijing China Science Journal Publishing Co. Ltd. All right reserved.
引用
收藏
页码:759 / 768
页数:9
相关论文
共 24 条
  • [1] Indyk P, Motwani R., Approximate nearest neighbors: towards removing the curse of dimensionality, Proceedings of the 30th Annual ACM Symposium on Theory of the Computing, pp. 604-613, (1998)
  • [2] Heo J, Lin Z, Yoon S, Et al., Distance encoded product quantization, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2139-2146, (2014)
  • [3] Gionis A, Indyk P, Motwani R, Et al., Similarity search in high dimensions via hashing, Proceedings of the 25th International Conference on Very Large Data Bases, pp. 518-529, (1999)
  • [4] Wang J D, Zhang T, Song J K, Et al., A survey on learning to hash, IEEE Transactions on Pattern Analysis and Machine Intelligence, 40, 4, pp. 769-790, (2018)
  • [5] Krizhevsky A, Sutskever I, Hinton G E, Et al., ImageNet classification with deep convolutional neural networks, Communications of the ACM, 60, 6, pp. 84-90, (2017)
  • [6] Simonyan K, Zisserman A., Very deep convolutional networks for large-scale image recognition
  • [7] He K M, Zhang X Y, Ren S Q, Et al., Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [8] Xia R K, Pan Y, Lai H J, Et al., Supervised hashing for image retrieval via image representation learning, Proceedings of the National Conference on Artificial Intelligence, pp. 2156-2162, (2014)
  • [9] Lai H J, Pan Y, Liu Y, Et al., Simultaneous feature learning and hash coding with deep neural networks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3270-3278, (2015)
  • [10] Zhu H, Long M S, Wang J M, Et al., Deep Hashing network for efficient similarity retrieval, Proceedings of the National Conference on Artificial Intelligence, pp. 2415-2421, (2016)