A Parallel Training Algorithm for Hierarchical Pitman-Yor Process Language Models

被引:0
|
作者
Huang, Songfang [1 ]
Renals, Steve [1 ]
机构
[1] Univ Edinburgh, Ctr Speech Technol Res, Edinburgh EH8 9AB, Midlothian, Scotland
关键词
language model; Pitman-Yor processes; hierarchical Bayesian models; parallel training; meetings;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Hierarchical Pitman Yor Process Language Model (HPYLM) is a Bayesian language model based on a non-parametric prior, the Pitman-Yor Process. It has been demonstrated, both theoretically and practically, that the HPYLM can provide better smoothing for language modeling, compared with state-of-the-art approaches such as interpolated Kneser-Ney and modified Kneser-Ney smoothing. However, estimation of Bayesian language models is expensive in terms of both computation time and memory; the inference is approximate and requires a number of iterations to converge. In this paper, we present a parallel training algorithm for the HPYLM, which enables the approach to be applied in the context of automatic speech recognition, using large training corpora with large vocabularies. We demonstrate the effectiveness of the proposed algorithm by estimating language models from corpora for meeting transcription containing over 200 million words, and observe significant reductions in perplexity and word error rate.
引用
收藏
页码:2663 / 2666
页数:4
相关论文
共 50 条
  • [41] A Pitman-Yor Process Self-Aggregated Topic Model for Short Texts of Social Media
    Niu, Yue
    Zhang, Hongjie
    Li, Jing
    [J]. IEEE ACCESS, 2021, 9 : 129011 - 129021
  • [42] Inducing Word and Part-of-Speech with Pitman-Yor Hidden Semi-Markov Models
    Uchiumi, Kei
    Tsukahara, Hiroshi
    Mochihashi, Daichi
    [J]. PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, 2015, : 1774 - 1782
  • [43] Pitman-Yor process mixture model for community structure exploration considering latent interaction patterns*
    Wang, Jing
    Li, Kan
    [J]. CHINESE PHYSICS B, 2021, 30 (12)
  • [44] A Markov random field-regulated Pitman-Yor process prior for spatially constrained data clustering
    Chatzis, Sotirios P.
    [J]. PATTERN RECOGNITION, 2013, 46 (06) : 1595 - 1603
  • [45] Beyond the Chinese Restaurant and Pitman-Yor processes: Statistical Models with double power-law behavior
    Ayed, Fadhel
    Lee, Juho
    Caron, Francois
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [46] Joint modeling of causal phrases-sentiments-aspects using Hierarchical Pitman Yor Process
    Yadavilli, V. R. P. S. Sastry
    Seshadri, Karthick
    Bhattu, S. Nagesh
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (04)
  • [47] A Novel 3D Model Recognition Approach using Pitman-Yor Process Mixtures of Beta-Liouville Distributions
    Fan, Wentao
    Al-Osaimi, Faisal R.
    Bouguila, Nizar
    [J]. 2016 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2016, : 1986 - 1989
  • [48] Simultaneous clustering and feature selection via nonparametric Pitman–Yor process mixture models
    Wentao Fan
    Nizar Bouguila
    [J]. International Journal of Machine Learning and Cybernetics, 2019, 10 : 2753 - 2766
  • [49] Adaptive Bayesian Density Estimation in Lp-metrics with Pitman-Yor or Normalized Inverse-Gaussian Process Kernel Mixtures
    Scricciolo, Catia
    [J]. BAYESIAN ANALYSIS, 2014, 9 (02): : 475 - 520
  • [50] Sparse Parallel Training of Hierarchical Dirichlet Process Topic Models
    Terenin, Alexander
    Magnusson, Mans
    Jonsson, Leif
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2925 - 2934