Cognitively Inspired Feature Extraction and Speech Recognition for Automated Hearing Loss Testing

被引:0
|
作者
Shibli Nisar
Muhammad Tariq
Ahsan Adeel
Mandar Gogate
Amir Hussain
机构
[1] National University of Computer and Emerging Sciences,School of Mathematics and Computer Science
[2] Princeton University,Edinburgh Napier University
[3] University of Stirling,Taibah Valley
[4] deepCI,undefined
[5] University of Wolverhampton,undefined
[6] School of Computing,undefined
[7] Taibah University,undefined
来源
Cognitive Computation | 2019年 / 11卷
关键词
Hearing loss; Speech recognition; Machine learning; Automation; Cognitive radio;
D O I
暂无
中图分类号
学科分类号
摘要
Hearing loss, a partial or total inability to hear, is one of the most commonly reported disabilities. A hearing test can be carried out by an audiologist to assess a patient’s auditory system. However, the procedure requires an appointment, which can result in delays and practitioner fees. In addition, there are often challenges associated with the unavailability of equipment and qualified practitioners, particularly in remote areas. This paper presents a novel idea that automatically identifies any hearing impairment based on a cognitively inspired feature extraction and speech recognition approach. The proposed system uses an adaptive filter bank with weighted Mel-frequency cepstral coefficients for feature extraction. The adaptive filter bank implementation is inspired by the principle of spectrum sensing in cognitive radio that is aware of its environment and adapts to statistical variations in the input stimuli by learning from the environment. Comparative performance evaluation demonstrates the potential of our automated hearing test method to achieve comparable results to the clinical ground truth, established by the expert audiologist’s tests. The overall absolute error of the proposed model when compared with the expert audiologist test is less than 4.9 dB and 4.4 dB for the pure tone and speech audiometry tests, respectively. The overall accuracy achieved is 96.67% with a hidden Markov model (HMM). The proposed method potentially offers a second opinion to audiologists, and serves as a cost-effective pre-screening test to predict hearing loss at an early stage. In future work, authors intend to explore the application of advanced deep learning and optimization approaches to further enhance the performance of the automated testing prototype considering imperfect datasets with real-world background noise.
引用
收藏
页码:489 / 502
页数:13
相关论文
共 50 条
  • [41] Feature Extraction Analysis on Indonesian Speech Recognition System
    Wisesty, Untari N.
    Adiwijaya
    Astuti, Widi
    2015 3rd International Conference on Information and Communication Technology (ICoICT), 2015, : 54 - 58
  • [42] Applying sparse KPCA for feature extraction in speech recognition
    Lima, A
    Zen, H
    Nankaku, Y
    Tokuda, K
    Kitamura, T
    Resende, FG
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2005, E88D (03): : 401 - 409
  • [43] Discriminative temporal feature extraction for robust speech recognition
    Shen, JL
    ELECTRONICS LETTERS, 1997, 33 (19) : 1598 - 1600
  • [44] Soft Margin Feature Extraction for Automatic Speech Recognition
    Li, Jinyu
    Lee, Chin-Hui
    INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 293 - 296
  • [45] A Salient Feature Extraction Algorithm for Speech Emotion Recognition
    Liang, Ruiyu
    Tao, Huawei
    Tang, Guichen
    Wang, Qingyun
    Zhao, Li
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2015, E98D (09): : 1715 - 1718
  • [46] APPLYING FEATURE EXTRACTION OF SPEECH RECOGNITION ON VOIP AUDITING
    Wang, Xuan
    Lin, Jiancheng
    Sun, Yong
    Gan, Haibo
    Yao, Lin
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2009, 5 (07): : 1851 - 1856
  • [47] On the use of kernel PCA for feature extraction in speech recognition
    Lima, A
    Zen, H
    Nankaku, Y
    Miyajima, C
    Tokuda, K
    Kitamura, T
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2004, E87D (12) : 2802 - 2811
  • [48] Survey on Acoustic Modeling and Feature Extraction for Speech Recognition
    Garg, Anjali
    Sharma, Poonam
    PROCEEDINGS OF THE 10TH INDIACOM - 2016 3RD INTERNATIONAL CONFERENCE ON COMPUTING FOR SUSTAINABLE GLOBAL DEVELOPMENT, 2016, : 2291 - 2295
  • [49] Distinctive phonetic feature extraction for robust speech recognition
    Fukuda, T
    Yamamoto, W
    Nitta, T
    2003 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL II, PROCEEDINGS: SPEECH II; INDUSTRY TECHNOLOGY TRACKS; DESIGN & IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS; NEURAL NETWORKS FOR SIGNAL PROCESSING, 2003, : 25 - 28
  • [50] Human vision inspired feature extraction for facial expression recognition
    Hamid Sadeghi
    Abolghasem-A. Raie
    Multimedia Tools and Applications, 2019, 78 : 30335 - 30353