Attention-Based Bimodal Neural Network Speech Recognition System on FPGA

被引:0
|
作者
Chen, Aiwu [1 ]
机构
[1] College of Intelligent Manufacturing (CIM), Hunan University of Science and Engineering (HUSE), Yongzhou,425199, China
来源
Informatica (Slovenia) | 2025年 / 49卷 / 13期
关键词
Audiovisual - Gates (transistor) - Neural networks - Speech enhancement - Speech recognition;
D O I
10.31449/inf.v49i13.7154
中图分类号
学科分类号
摘要
To further improve the accuracy of speech recognition technology, a neural network speech recognition system based on a field programmable gate array is designed. Firstly, a neural network audiovisual bimodal speech recognition algorithm based on an attention mechanism is designed. Then, a speech recognition platform based on on-site programmable gate arrays is built. The results showed that the word error rate and the character error rate of this research algorithm were 3.17% and 1.56%, respectively, which were significantly lower than the traditional Lip-Reading Network algorithm's 26.24% and 12.56%. The algorithm converged quickly when the training rounds were less than 10 and tended to stabilize when it was 20. The proposed speech recognition platform used many DSP units in its design, with a utilization rate of 83.2%, the lowest power consumption of 2.21W, the highest energy efficiency ratio of 26.15, and the shortest processing time and faster running speed. In summary, the research algorithm can reasonably allocate learning weights, improve training speed, and has certain feasibility and effectiveness because of introducing attention mechanism. It has good application effects in speech recognition, which helps to improve the accuracy of language recognition algorithms and promote communication between humans and machines. © 2025 Slovene Society Informatika. All rights reserved.
引用
收藏
页码:1 / 12
相关论文
共 50 条
  • [1] Speech Emotion Recognition via Generation using an Attention-based Variational Recurrent Neural Network
    Baruah, Murchana
    Banerjee, Bonny
    INTERSPEECH 2022, 2022, : 4710 - 4714
  • [2] Attention-Based Models for Speech Recognition
    Chorowski, Jan
    Bahdanau, Dzmitry
    Serdyuk, Dmitriy
    Cho, Kyunghyun
    Bengio, Yoshua
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [3] Attention-based deep neural network for driver behavior recognition
    Xiao, Weichu
    Liu, Hongli
    Ma, Ziji
    Chen, Weihong
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 132 : 152 - 161
  • [4] Attention-based convolutional neural network for deep face recognition
    Hefei Ling
    Jiyang Wu
    Junrui Huang
    Jiazhong Chen
    Ping Li
    Multimedia Tools and Applications, 2020, 79 : 5595 - 5616
  • [5] Attention-based convolutional neural network for deep face recognition
    Ling, Hefei
    Wu, Jiyang
    Huang, Junrui
    Chen, Jiazhong
    Li, Ping
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (9-10) : 5595 - 5616
  • [6] AN ATTENTION-BASED NEURAL NETWORK APPROACH FOR SINGLE CHANNEL SPEECH ENHANCEMENT
    Hao, Xiang
    Shan, Changhao
    Xu, Yong
    Sun, Sining
    Xie, Lei
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6895 - 6899
  • [7] An Online Attention-Based Model for Speech Recognition
    Fan, Ruchao
    Zhou, Pan
    Chen, Wei
    Jia, Jia
    Liu, Gang
    INTERSPEECH 2019, 2019, : 4390 - 4394
  • [8] Thank you for attention: A survey on attention-based artificial neural networks for automatic speech recognition
    Karmakar, Priyabrata
    Teng, Shyh Wei
    Lu, Guojun
    INTELLIGENT SYSTEMS WITH APPLICATIONS, 2024, 23
  • [9] AHRNN: Attention-Based Hybrid Robust Neural Network for emotion recognition
    Xu, Ke
    Liu, Bin
    Tao, Jianhua
    Lv, Zhao
    Fan, Cunhang
    Song, Leichao
    COGNITIVE COMPUTATION AND SYSTEMS, 2022, 4 (01) : 85 - 95
  • [10] Multistage Spatial Attention-Based Neural Network for Hand Gesture Recognition
    Miah, Abu Saleh Musa
    Hasan, Md. Al Mehedi
    Shin, Jungpil
    Okuyama, Yuichi
    Tomioka, Yoichi
    COMPUTERS, 2023, 12 (01)