DocumentCode
1456519
Title
Extensions of Fisher Information and Stam´s Inequality
Author
Lutwak, Erwin ; Lv, Songjun ; Yang, Deane ; Zhang, Gaoyong
Author_Institution
Dept. of Math., Polytech. Inst. of New York Univ., Brooklyn, NY, USA
Volume
58
Issue
3
fYear
2012
fDate
3/1/2012 12:00:00 AM
Firstpage
1319
Lastpage
1327
Abstract
We explain how the classical notions of Fisher information of a random variable and Fisher information matrix of a random vector can be extended to a much broader setting. We also show that Stam´s inequality for Fisher information and Shannon entropy, as well as the more generalized versions proved earlier by the authors, are all special cases of more general sharp inequalities satisfied by random vectors. The extremal random vectors, which we call generalized Gaussians, contain Gaussians as a limiting case but are noteworthy because they are heavy-tailed.
Keywords
Gaussian processes; entropy; matrix algebra; random processes; Fisher information matrix; Gaussian process; Shannon entropy; Stam inequality; random variable; random vector; Covariance matrix; Entropy; Linear matrix inequalities; Random variables; Symmetric matrices; Transforms; Vectors; Entropy; Fisher information; Rényi entropy; Shannon entropy; Shannon theory; Stam inequality; information measure; information theory;
fLanguage
English
Journal_Title
Information Theory, IEEE Transactions on
Publisher
ieee
ISSN
0018-9448
Type
jour
DOI
10.1109/TIT.2011.2177563
Filename
6157091
Link To Document