Herzog–Schönheim conjecture

From formulasearchengine
Revision as of 16:21, 20 April 2013 by en>Yobot (WP:CHECKWIKI error fixes - Replaced endash with hyphen in sortkey per WP:MCSTJR using AWB (9100))
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

In information theory, Fano's inequality (also known as the Fano converse and the Fano lemma) relates the average information lost in a noisy channel to the probability of the categorization error. It was derived by Robert Fano in the early 1950s while teaching a Ph.D. seminar in information theory at MIT, and later recorded in his 1961 textbook.

It is used to find a lower bound on the error probability of any decoder as well as the lower bounds for minimax risks in density estimation.

Fano's inequality

Let the random variables X and Y represent input and output messages with a joint probability P(x,y). Let e represent an occurrence of error; i.e., that XY. Fano's inequality is

H(X|Y)H(P(e))+P(e)log(|𝒳|1),

where 𝒳 denotes the support of X,

H(X|Y)=i,jP(xi,yj)logP(xi|yj)

is the conditional entropy,

P(e)=P(XY)

is the probability of the communication error, and

H(e)=P(e)logP(e)(1P(e))log(1P(e))

is the corresponding binary entropy.

Alternative formulation

Let X be a random variable with density equal to one of r+1 possible densities f1,,fr+1. Furthermore, the Kullback-Leibler divergence between any pair of densities cannot be too large,

DKL(fifj)β for all i=j.

Let ψ(X){1,,r+1} be an estimate of the index. Then

supiPi(ψ(X)=i)1β+log2logr

where Pi is the probability induced by fi

Generalization

The following generalization is due to Ibragimov and Khasminskii (1979), Assouad and Birge (1983).

Let F be a class of densities with a subclass of r + 1 densities ƒθ such that for any θ ≠ θ

fθfθL1α,
DKL(fθfθ)β.

Then in the worst case the expected value of error of estimation is bound from below,

supfFEfnfL1α2(1nβ+log2logr)

where ƒn is any density estimator based on a sample of size n.

References

  • P. Assouad, "Deux remarques sur l'estimation," Comptes Rendus de L'Academie des Sciences de Paris, Vol. 296, pp. 1021–1024, 1983.
  • L. Birge, "Estimating a density under order restrictions: nonasymptotic minimax risk," Technical report, UER de Sciences Economiques, Universite Paris X, Nanterre, France, 1983.
  • T. Cover, J. Thomas, Elements of Information Theory. pp. 43.
  • L. Devroye, A Course in Density Estimation. Progress in probability and statistics, Vol 14. Boston, Birkhauser, 1987. ISBN 0-8176-3365-0, ISBN 3-7643-3365-0.
  • R. Fano, Transmission of information; a statistical theory of communications. Cambridge, Mass., M.I.T. Press, 1961. ISBN 0-262-06001-9
  • R. Fano, Fano inequality Scholarpedia, 2008.
  • I. A. Ibragimov, R. Z. Has′minskii, Statistical estimation, asymptotic theory. Applications of Mathematics, vol. 16, Springer-Verlag, New York, 1981. ISBN 0-387-90523-5