• DocumentCode
    49744
  • Title

    Hierarchical Pitman–Yor–Dirichlet Language Model

  • Author

    Jen-Tzung Chien

  • Author_Institution
    Dept. of Electr. & Comput. Eng., Nat. Chiao Tung Univ., Hsinchu, Taiwan
  • Volume
    23
  • Issue
    8
  • fYear
    2015
  • fDate
    Aug. 2015
  • Firstpage
    1259
  • Lastpage
    1272
  • Abstract
    Probabilistic models are often viewed as insufficiently expressive because of strong limitation and assumption on the probabilistic distribution and the fixed model complexity. Bayesian nonparametric learning pursues an expressive probabilistic representation based on the nonparametric prior and posterior distributions with less assumption-laden approach to inference. This paper presents a hierarchical Pitman-Yor-Dirichlet (HPYD) process as the nonparametric priors to infer the predictive probabilities of the smoothed n-grams with the integrated topic information. A metaphor of hierarchical Chinese restaurant process is proposed to infer the HPYD language model (HPYD-LM) via Gibbs sampling. This process is equivalent to implement the hierarchical Dirichlet process-latent Dirichlet allocation (HDP-LDA) with the twisted hierarchical Pitman-Yor LM (HPY-LM) as base measures. Accordingly, we produce the power-law distributions and extract the semantic topics to reflect the properties of natural language in the estimated HPYD-LM. The superiority of HPYD-LM to HPY-LM and other language models is demonstrated by the experiments on model perplexity and speech recognition.
  • Keywords
    Bayes methods; learning (artificial intelligence); nonparametric statistics; speech recognition; statistical distributions; Bayesian nonparametric learning; Gibbs sampling; HDP-LDA; HPYD language model; HPYD-LM; base measures; expressive probabilistic representation; fixed model complexity; hierarchical Chinese restaurant process; hierarchical Dirichlet process-latent Dirichlet allocation; hierarchical Pitman-Yor-Dirichlet language model; integrated topic information; model perplexity; natural language; nonparametric prior distribution; posterior distributions; power-law distributions; predictive probabilities; probabilistic distribution; probabilistic models; semantic topics; smoothed n-grams; speech recognition; twisted hierarchical Pitman-Yor LM; Adaptation models; Bayes methods; Context; Data models; Semantics; Speech; Speech processing; Bayesian nonparametrics; language model; speech recognition; topic model; unsupervised learning;
  • fLanguage
    English
  • Journal_Title
    Audio, Speech, and Language Processing, IEEE/ACM Transactions on
  • Publisher
    ieee
  • ISSN
    2329-9290
  • Type

    jour

  • DOI
    10.1109/TASLP.2015.2428632
  • Filename
    7098357