• DocumentCode
    740494
  • Title

    Hilberg Exponents: New Measures of Long Memory in the Process

  • Author

    Debowski, Lukasz

  • Author_Institution
    Polish Academy of Sciences, Institute of Computer Science, Warszaw, Poland
  • Volume
    61
  • Issue
    10
  • fYear
    2015
  • Firstpage
    5716
  • Lastpage
    5726
  • Abstract
    This paper concerns the rates of power law growth of mutual information computed for a stationary measure or for a universal code. The rates are called Hilberg exponents, and four such quantities are defined for each measure and each code: two random exponents and two expected exponents. A particularly interesting case arises for the conditional algorithmic mutual information. In this case, the random Hilberg exponents are almost surely constant on ergodic sources and are bounded by the expected Hilberg exponents. This property is the second-order analog of the Shannon–McMillan–Breiman theorem, proved without invoking the ergodic theorem. It carries over to Hilberg exponents for the underlying probability measure via Shannon–Fano coding and Barron inequality. Moreover, the expected Hilberg exponents can be linked for different universal codes. Namely, if one code dominates another, the expected Hilberg exponents are greater for the former than for the latter. This paper is concluded by an evaluation of Hilberg exponents for certain sources, such as the mixture Bernoulli process and the Santa Fe processes.
  • Keywords
    Approximation algorithms; Complexity theory; Entropy; IP networks; Mutual information; Natural languages; Q measurement; Kolmogorov complexity; Mutual information; ergodic processes; mutual information; universal coding;
  • fLanguage
    English
  • Journal_Title
    Information Theory, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9448
  • Type

    jour

  • DOI
    10.1109/TIT.2015.2470675
  • Filename
    7214292