Information Without Rolling Dice

The deterministic notions of capacity and entropy are studied in the context of communication and storage of information using square-integrable and bandlimited signals subject to perturbation. The (E, δ)-capacity that extends the Kolmogorov E-capacity to packing sets of overlap at most δ is introdu...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on information theory Vol. 63; no. 3; pp. 1349 - 1363
Main Authors Lim, Taehyung J., Franceschetti, Massimo
Format Journal Article
LanguageEnglish
Published New York IEEE 01.03.2017
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text
ISSN0018-9448
1557-9654
DOI10.1109/TIT.2016.2647602

Cover

More Information
Summary:The deterministic notions of capacity and entropy are studied in the context of communication and storage of information using square-integrable and bandlimited signals subject to perturbation. The (E, δ)-capacity that extends the Kolmogorov E-capacity to packing sets of overlap at most δ is introduced and compared with the Shannon capacity. The functional form of the results indicates that in both Kolmogorov and Shannon's settings, capacity and entropy grow linearly with the number of degrees of freedom, but only logarithmically with the signal to noise ratio. This basic insight transcends the details of the stochastic or deterministic description of the information theoretic model. For δ = 0, the analysis leads to a tight asymptotic expression of the Kolmogorov E-entropy of bandlimited signals. A deterministic notion of error exponent is introduced. Applications of the theory are briefly discussed.
Bibliography:SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 14
ObjectType-Article-1
ObjectType-Feature-2
ISSN:0018-9448
1557-9654
DOI:10.1109/TIT.2016.2647602