• DocumentCode
    37431
  • Title

    Learning Topic Models by Belief Propagation

  • Author

    Jia Zeng ; Cheung, William K. ; Jiming Liu

  • Author_Institution
    Sch. of Comput. Sci. & Technol., Soochow Univ., Suzhou, China
  • Volume
    35
  • Issue
    5
  • fYear
    2013
  • fDate
    May-13
  • Firstpage
    1121
  • Lastpage
    1134
  • Abstract
    Latent Dirichlet allocation (LDA) is an important hierarchical Bayesian model for probabilistic topic modeling, which attracts worldwide interest and touches on many important applications in text mining, computer vision and computational biology. This paper represents the collapsed LDA as a factor graph, which enables the classic loopy belief propagation (BP) algorithm for approximate inference and parameter estimation. Although two commonly used approximate inference methods, such as variational Bayes (VB) and collapsed Gibbs sampling (GS), have gained great success in learning LDA, the proposed BP is competitive in both speed and accuracy, as validated by encouraging experimental results on four large-scale document datasets. Furthermore, the BP algorithm has the potential to become a generic scheme for learning variants of LDA-based topic models in the collapsed space. To this end, we show how to learn two typical variants of LDA-based topic models, such as author-topic models (ATM) and relational topic models (RTM), using BP based on the factor graph representations.
  • Keywords
    Bayes methods; belief networks; graph theory; inference mechanisms; learning (artificial intelligence); sampling methods; text analysis; LDA learning; LDA-based topic models; VB method; approximate inference; collapsed GS method; collapsed Gibbs sampling method; collapsed LDA; computational biology; computer vision; factor graph representations; hierarchical Bayesian model; large-scale document datasets; latent Dirichlet allocation; loopy BP algorithm; loopy belief propagation algorithm; parameter estimation; probabilistic topic modeling; text mining; variational Bayes method; Approximation algorithms; Approximation methods; Computational modeling; Hidden Markov models; Indexes; Inference algorithms; Joints; Bayesian networks; Gibbs sampling; Latent Dirichlet allocation; Markov random fields; belief propagation; factor graph; hierarchical Bayesian models; message passing; topic models; variational Bayes;
  • fLanguage
    English
  • Journal_Title
    Pattern Analysis and Machine Intelligence, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0162-8828
  • Type

    jour

  • DOI
    10.1109/TPAMI.2012.185
  • Filename
    6291721