A Self-attention Based Model for Offline Handwritten Text Recognition

被引:2
|
作者
Nam Tuan Ly [1 ]
Trung Tan Ngo [1 ]
Nakagawa, Masaki [1 ]
机构
[1] Tokyo Univ Agr & Technol, Tokyo, Japan
来源
关键词
Self-attention; Multi-head; Handwritten text recognition; CNN; BLSTM; CTC; SEQUENCE;
D O I
10.1007/978-3-031-02444-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Offline handwritten text recognition is an important part of document analysis and it has been receiving a lot of attention from numerous researchers for decades. In this paper, we present a self-attention-based model for offline handwritten textline recognition. The proposed model consists of three main components: a feature extractor by CNN; an encoder by a BLSTM network and a self-attention module; and a decoder by CTC. The self-attention module is complementary to RNN in the encoder and helps the encoder to capture long-range and multi-level dependencies across an input sequence. According to the extensive experiments on the two datasets of IAM Handwriting and Kuzushiji, the proposed model achieves better accuracy than the state-of-the-art models. The self-attention map visualization shows that the self-attention mechanism helps the encoder capture long-range and multi-level dependencies across an input sequence.
引用
下载
收藏
页码:356 / 369
页数:14
相关论文
共 50 条
  • [31] A Self-attention Based LSTM Network for Text Classification
    Jing, Ran
    2019 3RD INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2019), 2019, 1207
  • [32] Self-attention Based Text Matching Model with Generative Pre-training
    Zhang, Xiaolin
    Lei, Fengpei
    Yu, Shengji
    2021 IEEE INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, INTL CONF ON CLOUD AND BIG DATA COMPUTING, INTL CONF ON CYBER SCIENCE AND TECHNOLOGY CONGRESS DASC/PICOM/CBDCOM/CYBERSCITECH 2021, 2021, : 84 - 91
  • [33] Occluded offline handwritten Chinese character inpainting via generative adversarial network and self-attention mechanism
    Song, Ge
    Li, Jianwu
    Wang, Zheng
    NEUROCOMPUTING, 2020, 415 : 146 - 156
  • [34] Deep Convolutional Neural Network Based Hidden Markov Model for Offline Handwritten Chinese Text Recognition
    Wang, Zi-Rui
    Du, Jun
    Hu, Jin-Shui
    Hu, Yu-Long
    PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2017, : 816 - 821
  • [35] Lexicon-based Offline Recognition of Amharic Words in Unconstrained Handwritten Text
    Assabie, Yaregal
    Bigun, Josef
    19TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOLS 1-6, 2008, : 3043 - 3046
  • [36] Topic Language Model Adaption for Recognition of Homologous Offline Handwritten Chinese Text Image
    Wang, Yanwei
    Ding, Xiaoqing
    Liu, Changsong
    IEEE SIGNAL PROCESSING LETTERS, 2014, 21 (05) : 550 - 553
  • [37] Deformable Self-Attention for Text Classification
    Ma, Qianli
    Yan, Jiangyue
    Lin, Zhenxi
    Yu, Liuhong
    Chen, Zipeng
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 1570 - 1581
  • [38] Offline recognition of syntax-constrained cursive handwritten text
    González, J
    Salvador, I
    Toselli, AH
    Juan, A
    Vidal, E
    Casacuberta, F
    ADVANCES IN PATTERN RECOGNITION, 2000, 1876 : 143 - 153
  • [39] Multiple classifier methods for offline handwritten text line recognition
    Bertolami, Roman
    Bunke, Horst
    MULTIPLE CLASSIFIER SYSTEMS, PROCEEDINGS, 2007, 4472 : 72 - +
  • [40] Offline Handwritten Text Recognition Using Support Vector Machines
    Rajnoha, Martin
    Burget, Radim
    Dutta, Malay Kishore
    2017 4TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND INTEGRATED NETWORKS (SPIN), 2017, : 132 - 136