Multiple attention-based encoder-decoder networks for gas meter character recognition

被引:3
|
作者
Li, Weidong [1 ,2 ]
Wang, Shuai [1 ,2 ]
Ullah, Inam [1 ,2 ]
Zhang, Xuehai [1 ,2 ]
Duan, Jinlong [1 ,2 ]
机构
[1] Henan Univ Technol, Minist Educ, Key Lab Grain Informat Proc & Control, Zhengzhou 450001, Peoples R China
[2] Henan Univ Technol, Coll Informat Sci & Engn, Zhengzhou 450001, Peoples R China
关键词
SCENE; REPRESENTATION; SYSTEM;
D O I
10.1038/s41598-022-14434-0
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Factories swiftly and precisely grasp the real-time data of the production instrumentation, which is the foundation for the development and progress of industrial intelligence in industrial production. Weather, light, angle, and other unknown circumstances, on the other hand, impair the image quality of meter dials in natural environments, resulting in poor dial image quality. The remote meter reading system has trouble recognizing dial pictures in extreme settings, challenging it to meet industrial production demands. This paper provides multiple attention and encoder-decoder-based gas meter recognition networks (MAEDR) for this problem. First, from the acquired dial photos, the dial images with extreme conditions such as overexposure, artifacts, blurring, incomplete display of characters, and occlusion are chosen to generate the gas meter dataset. Then, a new character recognition network is proposed utilizing multiple attention and an encoder-decoder structure. Convolutional neural networks (CNN) extract visual features from dial images, encode visual features employing multi-head self-attention and position information, and facilitate feature alignment using the connectionist temporal classification (CTC) method. A novel two-step attention decoder is presented to improve the accuracy of recognition results. convolutional block attention module (CBAM) reweights the visual features from the CNN and the semantic features computed by the encoder to improve model performance; long short-term memory attention (LSTM attention) focuses on the relationship between feature sequences. According to experimental data, our system can effectively and efficiently identify industrial gas meter picture digits with 91.1% identification accuracy, faster inference speed, and higher accuracy than standard algorithms. The accuracy and practicality of the recognition can fulfill the needs of instrument data detection and recognition in industrial production, and it has a wide range of applications.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Attention-Based Personalized Encoder-Decoder Model for Local Citation Recommendation
    Yang, Libin
    Zhang, Zeqing
    Cai, Xiaoyan
    Dai, Tao
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2019, 2019
  • [22] Modeling User Session and Intent with an Attention-based Encoder-Decoder Architecture
    Loyola, Pablo
    Liu, Chen
    Hirate, Yu
    [J]. PROCEEDINGS OF THE ELEVENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'17), 2017, : 147 - 151
  • [23] Attention-based encoder-decoder model for answer selection in question answering
    Nie, Yuan-ping
    Han, Yi
    Huang, Jiu-ming
    Jiao, Bo
    Li, Ai-ping
    [J]. FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2017, 18 (04) : 535 - 544
  • [24] An attention-based row-column encoder-decoder model for text recognition in Japanese historical documents
    Ly, Nam Tuan
    Nguyen, Cuong Tuan
    Nakagawa, Masaki
    [J]. PATTERN RECOGNITION LETTERS, 2020, 136 : 134 - 141
  • [25] Multivariate time series forecasting via attention-based encoder-decoder framework
    Du, Shengdong
    Li, Tianrui
    Yang, Yan
    Horng, Shi-Jinn
    [J]. NEUROCOMPUTING, 2020, 388 (388) : 269 - 279
  • [26] Mining Implicit Intention Using Attention-Based RNN Encoder-Decoder Model
    Li, ChenXing
    Du, YaJun
    Wang, SiDa
    [J]. INTELLIGENT COMPUTING METHODOLOGIES, ICIC 2017, PT III, 2017, 10363 : 413 - 424
  • [27] Accurate water quality prediction with attention-based bidirectional LSTM and encoder-decoder
    Bi, Jing
    Chen, Zexian
    Yuan, Haitao
    Zhang, Jia
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [28] Attention-Based Encoder-Decoder End-to-End Neural Diarization With Embedding Enhancer
    Chen, Zhengyang
    Han, Bing
    Wang, Shuai
    Qian, Yanmin
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 1636 - 1649
  • [29] IMPROVED MULTI-STAGE TRAINING OF ONLINE ATTENTION-BASED ENCODER-DECODER MODELS
    Garg, Abhinav
    Gowda, Dhananjaya
    Kumar, Ankur
    Kim, Kwangyoun
    Kumar, Mehul
    Kim, Chanwoo
    [J]. 2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 70 - 77
  • [30] Hybrid Attention-Based Encoder-Decoder Fully Convolutional Network for PolSAR Image Classification
    Fang, Zheng
    Zhang, Gong
    Dai, Qijun
    Xue, Biao
    Wang, Peng
    [J]. REMOTE SENSING, 2023, 15 (02)