Deep Hashing Retrieval Algorithm Combing Attention Model and Bimodal Gaussian Distribution

被引:0
|
作者
Li Z. [1 ]
Zhang P. [1 ]
Liu Y. [1 ]
Li H. [2 ]
机构
[1] College of Computer & Communication Engineering, China University of Petroleum (East China), Qingdao
[2] Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences, Beijing
关键词
Attention model; Bimodal Gaussian distribution; Deep hashing; Image retrieval;
D O I
10.3724/SP.J.1089.2020.17889
中图分类号
学科分类号
摘要
Hash retrieval has attracted wide attention because of its small storage space and fast retrieval speed. There are two problems in deep hashing methods: Deep hash codes are essentially binary features, and the coding length is short, so their feature representation abilities are limited; In addition, existing deep hashing algorithms cannot directly learn discrete hash codes by backpropagation, and usually relax discrete values to continuous values in their optimization procedure, so this leads to quantization errors. Aiming at above problems, we propose a deep hashing retrieval algorithm combining attention model and bimodal Gaussian distribution. The network structure with spatial and channel attention model, focusing on important features and suppressing unnecessary features, enhances the feature representation abilities of hash codes; To solve the quantization error problem, the bimodal Gaussian distribution with the mean of either +1 or -1 is used as the prior distribution. We refer to the idea of variational auto-encoder, and propose to constrain the hash codes distribution to obey the prior distribution with KL divergence. The mean average precision of our method on three benchmark databases CIFAR-10, ImageNet-100 and NUS-WIDE is better than other methods of comparison, which verifies the effectiveness of the algorithm in this paper. © 2020, Beijing China Science Journal Publishing Co. Ltd. All right reserved.
引用
收藏
页码:759 / 768
页数:9
相关论文
共 24 条
  • [11] Cao Z J, Long M S, Wang J M, Et al., HashNet: deep learning to hash by continuation, Proceedings of the IEEE International Conference on Computer Vision, pp. 5609-5618, (2017)
  • [12] Li W J, Wang S, Kang W C, Et al., Feature learning based deep supervised hashing with pairwise labels, Proceedings of International Joint Conference on Artificial Intelligence, pp. 1711-1717, (2016)
  • [13] Kingma D P, Welling M., Auto-encoding variational Bayes
  • [14] Gong Y, Lazebnik S., Iterative quantization: a procrustean approach to learning binary codes, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 817-824, (2011)
  • [15] Weiss Y, Torralba A, Fergus R, Et al., Spectral hashing, Proceedings of the International Conference on Neural Information Processing Systems, pp. 1753-1760, (2008)
  • [16] Kulis B, Darrell T., Learning to hash with binary reconstructive embeddings, Proceedings of the International Conference on Neural Information Processing Systems, pp. 1042-1050, (2009)
  • [17] Liu W, Wang J, Ji R R, Et al., Supervised hashing with kernels, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2074-2081, (2012)
  • [18] Jin S, Yao H X, Sun X S, Et al., Deep saliency hashing
  • [19] Wang F, Jiang M Q, Qian C, Et al., Residual attention network for image classification, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6450-6458, (2017)
  • [20] Hu J, Shen L, Sun G., Squeeze-and-excitation networks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132-7141, (2018)