• DocumentCode
    4030
  • Title

    Stationary and Transition Probabilities in Slow Mixing, Long Memory Markov Processes

  • Author

    Asadi, Meysam ; Torghabeh, Ramezan Paravi ; Santhanam, Narayana P.

  • Author_Institution
    Dept. of Electr. Eng., Univ. of Hawaii at Manoa, Honolulu, HI, USA
  • Volume
    60
  • Issue
    9
  • fYear
    2014
  • fDate
    Sept. 2014
  • Firstpage
    5682
  • Lastpage
    5701
  • Abstract
    We observe a length-n sample generated by an unknown, stationary ergodic Markov process (model) over a finite alphabet A. Given any string w of symbols from A we want estimates of the conditional probability distribution of symbols following w, as well as the stationary probability of w. Two distinct problems that complicate estimation in this setting are: 1) long memory and 2) slow mixing, which could happen even with only one bit of memory. Any consistent estimator in this setting can only converge pointwise over the class of all ergodic Markov models. Namely, given any estimator and any sample size n, the underlying model could be such that the estimator performs poorly on a sample of size n with high probability. But can we look at a length-n sample and identify if an estimate is likely to be accurate? Since the memory is unknown a-priori, a natural approach is to estimate a potentially coarser model with memory kn = O(log n). As n grows, pointwise consistent estimates that hold eventually almost surely (e.a.s.) are known so long as the scaling of kn is not superlogarithmic in n. Here, rather than e.a.s. convergence results, we want the best answers possible with a length-n sample. Combining results in universal compression with Aldous´ coupling arguments, we obtain sufficient conditions on the lengthn sample (even for slow mixing models) to identify when naive: 1) estimates of the conditional probabilities and 2) estimates related to the stationary probabilities are accurate, and also bound the deviations of the naive estimates from true values.
  • Keywords
    Markov processes; probability; Markov process model; coarser model; conditional probability distribution; ergodic Markov models; finite alphabet; long memory Markov processes; slow mixing; stationary probabilities; transition probabilities; Context; Convergence; Estimation; History; Internet; Markov processes; Probability distribution; Context-tree weighting; Markov processes; coupling; pointwise consistency; universal compression;
  • fLanguage
    English
  • Journal_Title
    Information Theory, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9448
  • Type

    jour

  • DOI
    10.1109/TIT.2014.2334594
  • Filename
    6868219