Simultaneous Script Identification and Handwriting Recognition via Multi-Task Learning of Recurrent Neural Networks

被引:23
|
作者
Chen, Zhuo [1 ,2 ]
Wu, Yichao [1 ,2 ]
Yin, Pei [1 ]
Liu, Cheng-Lin [1 ,2 ]
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Inst Automat, 95 Zhongguan East Rd, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
multi-task learning; SepMDLSTM; script identification; language identification; handwritten text recognition;
D O I
10.1109/ICDAR.2017.92
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a method for simultaneous script identification and handwritten text line recognition in multi-task learning framework. Firstly, we use Separable Multi-Dimensional Long Short-Term Memory (SepMDLSTM) to encode the input text line images based on convolutional feature extraction. Then, the extracted features are fed into two classification modules for script identification and multi-script text recognition, respectively. All the network parameters are trained end-to-end by multi-task learning where the script identification task and the text recognition task are aimed to minimize the Negative Log Likelihood (NLL) loss and Connectionist Temporal Classification (CTC) loss, respectively. We evaluated the performance of the proposed method on handwritten text line datasets of three languages, namely, IAM (English), Rimes (French) and IFN/ENIT (Arabic). Experimental results demonstrate the multi-task learning framework performs superiorly for both script identification and text recognition. Particularly, the accuracy of script identification is higher than 99.9% and the character error rate (CER) of text recognition is even lower than that of some single-script text recognition systems.
引用
收藏
页码:525 / 530
页数:6
相关论文
共 50 条
  • [31] Multi-Task Learning for Face Recognition via Mutual Information Minimization
    Qian, Weizhu
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 1015 - 1021
  • [32] Implicit Discourse Relation Classification via Multi-Task Neural Networks
    Liu, Yang
    Li, Sujian
    Zhang, Xiaodong
    Sui, Zhifang
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2750 - 2756
  • [33] Multi-Task Learning with Capsule Networks
    Lei, Kai
    Fu, Qiuai
    Liang, Yuzhi
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [34] Cell tracking using deep neural networks with multi-task learning
    He, Tao
    Mao, Hua
    Guo, Jixiang
    Yi, Zhang
    IMAGE AND VISION COMPUTING, 2017, 60 : 142 - 153
  • [35] Evolutionary Multi-task Learning for Modular Training of Feedforward Neural Networks
    Chandra, Rohitash
    Gupta, Abhishek
    Ong, Yew-Soon
    Goh, Chi-Keong
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT II, 2016, 9948 : 37 - 46
  • [36] Simple, Efficient and Convenient Decentralized Multi-task Learning for Neural Networks
    Pilet, Amaury Bouchra
    Frey, Davide
    Taiani, Francois
    ADVANCES IN INTELLIGENT DATA ANALYSIS XIX, IDA 2021, 2021, 12695 : 37 - 49
  • [37] Improving generalization ability of neural networks ensemble with multi-task learning
    State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210093, China
    不详
    J. Comput. Inf. Syst., 2006, 4 (1235-1240):
  • [38] Randomized Neural Networks Based Decentralized Multi-Task Learning via Hybrid Multi-Block ADMM
    Ye, Yu
    Xiao, Ming
    Skoglund, Mikael
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 2844 - 2857
  • [39] Direct Learning With Multi-Task Neural Networks for Treatment Effect Estimation
    Zhu, Fujin
    Lu, Jie
    Lin, Adi
    Xuan, Junyu
    Zhang, Guangquan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (03) : 2457 - 2470
  • [40] Rapid Adaptation for Deep Neural Networks through Multi-Task Learning
    Huang, Zhen
    Li, Jinyu
    Siniscalchi, Sabato Marco
    Chen, I-Fan
    Wu, Ji
    Lee, Chin-Hui
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 3625 - 3629