• DocumentCode
    2020893
  • Title

    A Simple Proof of the Entropy-Power Inequality via Properties of Mutual Information

  • Author

    Rioul, O.

  • Author_Institution
    Dept. ComElec, Paris Tech Inst. & CNRS LTCI, Paris
  • fYear
    2007
  • fDate
    24-29 June 2007
  • Firstpage
    46
  • Lastpage
    50
  • Abstract
    While most useful information theoretic inequalities can be deduced from the basic properties of entropy or mutual information, Shannon´s entropy power inequality (EPI) seems to be an exception: available information theoretic proofs of the EPI hinge on integral representations of differential entropy using either Fisher´s information (FI) or minimum mean-square error (MMSE). In this paper, we first present a unified view of proofs via FI and MMSE, showing that they are essentially dual versions of the same proof, and then fill the gap by providing a new, simple proof of the EPI, which is solely based on the properties of mutual information and sidesteps both FI or MMSE representations.
  • Keywords
    entropy; least mean squares methods; Fisher information; Shannon entropy power inequality; information theoretic inequality; minimum mean-square error; Additive noise; Broadcasting; Entropy; Fasteners; MIMO; Mutual information; Probability density function; Random variables; Source coding; Telecommunications;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Theory, 2007. ISIT 2007. IEEE International Symposium on
  • Conference_Location
    Nice
  • Print_ISBN
    978-1-4244-1397-3
  • Type

    conf

  • DOI
    10.1109/ISIT.2007.4557202
  • Filename
    4557202