N-BEST ENTROPY BASED DATA SELECTION FOR ACOUSTIC MODELING

被引:0
|
作者
Itoh, Nobuyasu [1 ]
Sainath, Tara N. [2 ]
Liang, Dan Ning [3 ]
Zhou, Lie [3 ]
Ramabhadran, Bhuvana [2 ]
机构
[1] IBM Japan Ltd, IBM Res Tokyo, Yamato 2428502, Japan
[2] IBM Corp, Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
[3] IBM Res Corp, Beijing 100193, Peoples R China
关键词
N-best entropy; Acoustic modeling; Active learning; Data selection; Speech recognition;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper presents a strategy for efficiently selecting informative data from large corpora of untranscribed speech. Confidence-based selection methods (i.e., selecting utterances we are least confident about) have been a popular approach, though they only look at the top hypothesis when selecting utterances and tend to select outliers, therefore, not always improving overall recognition accuracy. Alternatively, we propose a method for selecting data looking at competing hypothesis by computing entropy of N-best hypothesis decoded by the baseline acoustic model. In addition we address the issue of outliers by calculating how representative a specific utterance is to all other unselected utterances via a tf-idf score. Experiments show that N-best entropy based selection (%relative 5.8 in 400-hour corpus) outperformed other conventional selection strategies; confidence based and lattice entropy based, and that tf-idfbased representativeness improved the model further (%relative 6.2). A comparison with random selection is also presented. Finally model size impact is discussed.
引用
收藏
页码:4133 / 4136
页数:4
相关论文
共 50 条
  • [41] Results of the N-Best 2008 Dutch Speech Recognition Evaluation
    van Leeuwen, David A.
    Kessens, Judith
    Sanders, Eric
    van den Heuvel, Henk
    INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 2531 - +
  • [42] Using N-Best Lists and Confusion Networks for Meeting Summarization
    Xie, Shasha
    Liu, Yang
    IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2011, 19 (05): : 1160 - 1169
  • [43] N-BEST HYPOTHESES RERANKING FOR TEXT-TO-SQL SYSTEMS
    Zeng, Lu
    Parthasarathi, Sree Hari Krishnan
    Hakkani-Tur, Dilek
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 663 - 670
  • [44] A Comparison of Two N-Best Extraction Methods for Weighted Tree Automata
    Bjorklund, Johanna
    Drewes, Frank
    Jonsson, Anna
    IMPLEMENTATION AND APPLICATION OF AUTOMATA, CIAA 2018, 2018, 10977 : 97 - 108
  • [45] N-best based supervised and unsupervised adaptation for native and non-native speakers in cars
    Nguyen, P
    Gelin, P
    Junqua, JC
    Chien, JT
    ICASSP '99: 1999 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS VOLS I-VI, 1999, : 173 - 176
  • [46] A new localization method for wireless sensor network nodes based on N-best rank sequence
    Pei Z.-M.
    Deng Z.-D.
    Xu S.
    Xu X.
    Zidonghua Xuebao/ Acta Automatica Sinica, 2010, 36 (02): : 199 - 207
  • [47] Character confidence based on N-best list for keyword spotting in online Chinese handwritten documents
    Zhang, Heng
    Wang, Da-Han
    Liu, Cheng-Lin
    PATTERN RECOGNITION, 2014, 47 (05) : 1880 - 1890
  • [48] THE N-BEST ALGORITHM - AN EFFICIENT PROCEDURE FOR FINDING TOP-N SENTENCE HYPOTHESES
    CHOW, YL
    SCHWARTZ, R
    SPEECH AND NATURAL LANGUAGE, 1989, : 199 - 202
  • [49] Finite-to-Infinite N-Best POMDP for Spoken Dialogue Management
    Wu, Guohua
    Yuan, Caixia
    Leng, Bing
    Wang, Xiaojie
    CHINESE COMPUTATIONAL LINGUISTICS AND NATURAL LANGUAGE PROCESSING BASED ON NATURALLY ANNOTATED BIG DATA (CCL 2015), 2015, 9427 : 369 - 380
  • [50] A stochastic sparse representation: n-best approximation to random signals and computation
    Qu, Wei
    Qian, Tao
    Deng, Guan-Tie
    APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2021, 55 : 185 - 198