• DocumentCode
    873080
  • Title

    On universal simulation of information sources using training data

  • Author

    Merhav, Neri ; Weinberger, Marcelo J.

  • Author_Institution
    Dept. of Electr. Eng., Technion-Israel Inst. of Technol., Haifa, Israel
  • Volume
    50
  • Issue
    1
  • fYear
    2004
  • Firstpage
    5
  • Lastpage
    20
  • Abstract
    We consider the problem of universal simulation of an unknown random process, or information source, of a certain parametric family, given a training sequence from that source and given a limited budget of purely random bits. The goal is to generate another random sequence (of the same length or shorter), whose probability law is identical to that of the given training sequence, but with minimum statistical dependency (minimum mutual information) between the input training sequence and the output sequence. We derive lower bounds on the mutual information that are shown to he achievable by conceptually simple algorithms proposed here. We show that the behavior of the minimum achievable mutual information depends critically on the relative amount of random bits and on the lengths of the input sequence and the output sequence. While in the ordinary (nonuniversal) simulation problem, the number of random bits per symbol must exceed the entropy rate H of the source in order to simulate it faithfully, in the universal simulation problem considered here, faithful preservation of the probability law is not a problem, yet the same minimum rate of H random bits per symbol is still needed to essentially eliminate the statistical dependency between the input sequence and the output sequence. The results are extended to more general information measures.
  • Keywords
    binary sequences; minimum entropy methods; probability; random number generation; random sequences; simulation; entropy rate; general information measures; information sources; lower bounds; minimum mutual information; minimum statistical dependency; parametric family; probability law; random bits; random number generators; random sequence; sequence lengths; training sequence; universal simulation; unknown random process; Entropy; Image generation; Information theory; Laboratories; Mutual information; Probability; Random number generation; Random processes; Random sequences; Training data;
  • fLanguage
    English
  • Journal_Title
    Information Theory, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9448
  • Type

    jour

  • DOI
    10.1109/TIT.2003.821993
  • Filename
    1262613