DENSE CHAINED ATTENTION NETWORK FOR SCENE TEXT RECOGNITION

被引:0
|
作者
Gao, Yunze [1 ]
Chen, Yingying
Wang, Jinqiao
Tang, Ming
Lu, Hanqing
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Inst Automat, Beijing 100190, Peoples R China
关键词
text recognition; attention; convolution-deconvolution;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Reading text in the wild is a challenging task in computer vision. Scene text suffers from various background noise, including shadow, irrelevant symbols and background texture. In order to reduce the disturbance of background noise, we propose a dense chained attention network with stacked attention modules for scene text recognition. Each attention module learns the attention map that is adapted to corresponding features to enhance the foreground text and suppress the background noise. Besides, the attention branch is designed with the convolution-deconvolution structure which rapidly captures global information to guide the discriminative feature selection. We stack multiple attention modules to gradually refine the attention maps and capture both the low-level appearance feature and the high-level semantic information. Extensive experiments on the standard benchmarks, the Street View Text, IIIT5K, and ICDAR datasets validate the superiority of the proposed method. The dense chained attention network achieves state-of-the-art or highly competitive recognition performance.
引用
收藏
页码:679 / 683
页数:5
相关论文
共 50 条
  • [31] Weakly Supervised Attention Rectification for Scene Text Recognition
    Gu, Chengyu
    Wang, Shilin
    Zhu, Yiwei
    Huang, Zheng
    Chen, Kai
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 779 - 786
  • [32] FACLSTM:ConvLSTM with focused attention for scene text recognition
    Qingqing WANG
    Ye HUANG
    Wenjing JIA
    Xiangjian HE
    Michael BLUMENSTEIN
    Shujing LYU
    Yue LU
    Science China(Information Sciences), 2020, 63 (02) : 35 - 48
  • [33] Sequential alignment attention model for scene text recognition
    Wu, Yan
    Fan, Jiaxin
    Tao, Renshuai
    Wang, Jiakai
    Qin, Haotong
    Liu, Aishan
    Liu, Xianglong
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 80
  • [34] Dual Relation Network for Scene Text Recognition
    Li, Ming
    Fu, Bin
    Chen, Han
    He, Junjun
    Qiao, Yu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4094 - 4107
  • [35] Decoupled Attention Network for Text Recognition
    Wang, Tianwei
    Zhu, Yuanzhi
    Jin, Lianwen
    Luo, Canjie
    Chen, Xiaoxue
    Wu, Yaqiang
    Wang, Qianying
    Cai, Mingxiang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12216 - 12224
  • [36] CarveNet: a channel-wise attention-based network for irregular scene text recognition
    Guibin Wu
    Zheng Zhang
    Yongping Xiong
    International Journal on Document Analysis and Recognition (IJDAR), 2022, 25 : 177 - 186
  • [37] Look Back Again: Dual Parallel Attention Network for Accurate and Robust Scene Text Recognition
    Fu, Zilong
    Xie, Hongtao
    Jin, Guoqing
    Guo, Junbo
    PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR '21), 2021, : 638 - 644
  • [38] CarveNet: a channel-wise attention-based network for irregular scene text recognition
    Wu, Guibin
    Zhang, Zheng
    Xiong, Yongping
    INTERNATIONAL JOURNAL ON DOCUMENT ANALYSIS AND RECOGNITION, 2022, 25 (3) : 177 - 186
  • [39] Look back again: Dual parallel attention network for accurate and robust scene text recognition
    Fu, Zilong
    Xie, Hongtao
    Jin, Guoqing
    Guo, Junbo
    ICMR 2021 - Proceedings of the 2021 International Conference on Multimedia Retrieval, 2021, : 638 - 644
  • [40] Flexible scene text recognition based on dual attention mechanism
    Tian, Zhiqiang
    Wang, Chunhui
    Xiao, Youzi
    Lin, Yuping
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (22):