Maximizing information content in feature extraction

被引:15
|
作者
Padmanabhan, M [1 ]
Dharanipragada, S
机构
[1] Renaissance Technol, E Setauket, NY 11733 USA
[2] Citadel Investment Grp, Chicago, IL 60603 USA
来源
关键词
classifiers; optimal feature projections; optimum feature extraction; penalized mutual information; speech recognition;
D O I
10.1109/TSA.2005.848876
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we consider the problem of quantifying the amount of information contained in a set of features, to discriminate between various classes. We explore these ideas in the context of a speech recognition system, where an important classification sub-problem is to predict the phonetic class, given an observed acoustic feature vector. The connection between information content and speech recognition system performance is first explored in the context of various feature extraction schemes used in speech recognition applications. Subsequently, the idea of optimizing the information content to improve recognition accuracy is generalized to a linear projection of the underlying features. We show that several prior methods to compute linear transformations (such as linear/heteroscedastic discriminant analysis) can be interpreted in this general framework of maximizing the information content. Subsequently, we extend this reasoning and propose a new objective function to maximize a penalized mutual information (pMI) measure. This objective function is seen to be very well correlated with the word error rate of the final system. Finally experimental results are provided that show that the proposed pMI projection consistently outperforms other methods for a variety of cases, leading to relative improvements in the word error rate of 5%-16% over earlier methods.
引用
收藏
页码:512 / 519
页数:8
相关论文
共 50 条
  • [1] Feature extraction by maximizing the average neighborhood margin
    Wang, Fei
    Zhang, Changshui
    2007 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-8, 2007, : 1173 - +
  • [2] An evaluation of feature extraction for query-by-content audio information retrieval
    Yu, Yi
    Downie, J. Stephen
    Joe, Kazuki
    ISM WORKSHOPS 2007: NINTH IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA - WORKSHOPS, PROCEEDINGS, 2007, : 297 - +
  • [3] Feature Selection by Maximizing Part Mutual Information
    Gao, Wanfu
    Hu, Liang
    Zhang, Ping
    2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND MACHINE LEARNING (SPML 2018), 2018, : 120 - 127
  • [4] Feature Selection by Maximizing Independent Classification Information
    Wang, Jun
    Wei, Jin-Mao
    Yang, Zhenglu
    Wang, Shu-Qin
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2017, 29 (04) : 828 - 841
  • [5] Maximizing the spread of information through content optimization
    Lin, Lei
    Du, Yihua
    Zhao, Shibo
    Jiang, Wenkang
    Tang, Qirui
    Xu, Li
    INTELLIGENT SYSTEMS WITH APPLICATIONS, 2024, 24
  • [6] Maximizing the Information Content of Experiments in Systems Biology
    Liepe, Juliane
    Filippi, Sarah
    Komorowski, Michal
    Stumpf, Michael P. H.
    PLOS COMPUTATIONAL BIOLOGY, 2013, 9 (01)
  • [7] Maximizing the information content in combinatorial libraries.
    Gallion, SL
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 1997, 214 : 31 - COMP
  • [8] Orthogonal locality minimizing globality maximizing projections for feature extraction
    Nie, Feiping
    Xiang, Shiming
    Song, Yangqiu
    Zhang, Changshui
    OPTICAL ENGINEERING, 2009, 48 (01)
  • [9] Information theoretic feature extraction for ATR
    Massachusetts Inst of Technology, Cambridge, United States
    Conf Rec Asilomar Conf Signals Syst Comput, (1245-1249):
  • [10] Feature matrix models for information extraction
    Hua, D
    Liu, F
    Chen, DC
    DMIN '05: Proceedings of the 2005 International Conference on Data Mining, 2005, : 229 - 235