• DocumentCode
    1007685
  • Title

    Approximation theory of output statistics

  • Author

    Han, Te Sun ; Verdu, Sergio

  • Author_Institution
    Graduate Sch. of Inf. Syst., Univ. of Electro-Commun., Tokyo, Japan
  • Volume
    39
  • Issue
    3
  • fYear
    1993
  • fDate
    5/1/1993 12:00:00 AM
  • Firstpage
    752
  • Lastpage
    772
  • Abstract
    Given a channel and an input process with output statistics that approximate the original output statistics with arbitrary accuracy, the randomness of the input processes is studied. The notion of resolvability of a channel, defined as the number of random bits required per channel use in order to generate an input that achieves arbitrarily accurate approximation of the output statistics for any given input process, is introduced. A general formula for resolvability that holds regardless of the channel memory structure is obtained. It is shown that for most channels, resolvability is equal to the Shannon capacity. By-products of the analysis are a general formula for the minimum achievable source coding rate of any finite-alphabet source and a strong converse of the identification coding theorem, which holds for any channel that satisfies the strong converse of the channel coding theorem
  • Keywords
    approximation theory; channel capacity; encoding; information theory; statistical analysis; Shannon capacity; approximation theory; channel coding theorem; channel memory structure; finite-alphabet source; identification coding theorem; input process; minimum achievable source coding rate; output statistics; resolvability; Approximation methods; Noise generators; Random number generation; Random processes; Random variables; Source coding; Statistical distributions; Statistics; Sun; Tellurium;
  • fLanguage
    English
  • Journal_Title
    Information Theory, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9448
  • Type

    jour

  • DOI
    10.1109/18.256486
  • Filename
    256486