A simple proof of the entropy-power inequality via properties of mutual information

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

While most useful information theoretic inequalities can be deduced from the basic properties of entropy or mutual information, Shannon's entropy power inequality (EPI) seems to be an exception: available information theoretic proofs of the EPI hinge on integral representations of differential entropy using either Fisher's information (FI) or minimum mean-square error (MMSE). In this paper, we first present a unified view of proofs via FI and MMSE, showing that they are essentially dual versions of the same proof, and then fill the gap by providing a new, simple proof of the EPI, which is solely based on the properties of mutual information and sidesteps both FI or MMSE representations.

Original languageEnglish
Title of host publicationProceedings - 2007 IEEE International Symposium on Information Theory, ISIT 2007
Pages46-50
Number of pages5
DOIs
Publication statusPublished - 1 Dec 2007
Externally publishedYes
Event2007 IEEE International Symposium on Information Theory, ISIT 2007 - Nice, France
Duration: 24 Jun 200729 Jun 2007

Publication series

NameIEEE International Symposium on Information Theory - Proceedings
ISSN (Print)2157-8101

Conference

Conference2007 IEEE International Symposium on Information Theory, ISIT 2007
Country/TerritoryFrance
CityNice
Period24/06/0729/06/07

Fingerprint

Dive into the research topics of 'A simple proof of the entropy-power inequality via properties of mutual information'. Together they form a unique fingerprint.

Cite this