Berezin integral: Difference between revisions
en>Rjwilmsi m →References: Format plain DOIs using AWB (8087) |
en>Initapp No edit summary |
||
Line 1: | Line 1: | ||
In [[information theory]], '''Pinsker's inequality''', named after its inventor [[Mark Semenovich Pinsker]], is an [[inequality (mathematics)|inequality]] that bounds the [[total variation distance]] (or statistical distance) in terms of the [[Kullback-Leibler divergence]]. | |||
The inequality is tight up to constant factors.<ref>{{cite book|title=Information Theory: Coding Theorems for Discrete Memoryless Systems|first1=Imre|last1=Csiszár|first2=János|last2=Körner|publisher=Cambridge University Press|year=2011|isbn=9781139499989|page=44|url=http://books.google.com/books?id=2gsLkQlb8JAC&pg=PA44}}</ref> | |||
Pinsker's inequality states that, if ''P'' and ''Q'' are two [[probability distribution]]s, then | |||
:<math>\delta(P,Q) \le \sqrt{\frac{1}{2} D_{\mathrm{KL}}(P\|Q)}</math> | |||
where | |||
:<math>\delta(P,Q)=\sup \{ |P(A) - Q(A)| : A\text{ is an event to which probabilities are assigned.} \}</math> | |||
is the [[Total variation distance of probability measures|total variation distance]] (or statistical distance) between ''P'' and ''Q'' and | |||
:<math>D_{\mathrm{KL}}(P\|Q) = \sum_i \ln\left(\frac{P(i)}{Q(i)}\right) P(i)\!</math> | |||
is the [[Kullback-Leibler divergence]] in [[Nat (information)|nats]]. | |||
Pinsker first proved the inequality with a worse constant. The inequality in the above form was proved independently by [[Solomon Kullback|Kullback]], [[Imre Csiszár|Csiszár]], and [[Johannes Kemperman|Kemperman]].<ref>{{cite book|last=Tsybakov|first=Alexandre|title=Introduction to Nonparametric Estimation|year=2009|publisher=Springer|isbn=9780387790527|page=132}}</ref> | |||
An inverse of the inequality cannot hold: for every <math>\epsilon > 0</math>, there are distributions with <math>\delta(P,Q)\le\epsilon</math> but <math>D_{\mathrm{KL}}(P\|Q) = \infty</math>.<ref>The divergence becomes infinite whenever one of the two distributions assigns probability zero to an event while the other assigns it a nonzero probability (no matter how small); see e.g. {{cite book|title=Data Complexity in Pattern Recognition|first1=Mitra|last1=Basu|first2=Tin Kam|last2=Ho|publisher=Springer|year=2006|isbn=9781846281723|page=161|url=http://books.google.com/books?id=GflBKbzym9oC&pg=PA161}}.</ref> | |||
==References== | |||
{{Reflist}} | |||
==Additional reading== | |||
* Thomas M. Cover and Joy A. Thomas: ''Elements of Information Theory'', 2nd edition, Willey-Interscience, 2006 | |||
* Nicolo Cesa-Bianchi and Gábor Lugosi: ''Prediction, Learning, and Games'', Cambridge University Press, 2006 | |||
[[Category:Information theory]] | |||
[[Category:Probabilistic inequalities]] |
Latest revision as of 13:14, 29 January 2014
In information theory, Pinsker's inequality, named after its inventor Mark Semenovich Pinsker, is an inequality that bounds the total variation distance (or statistical distance) in terms of the Kullback-Leibler divergence. The inequality is tight up to constant factors.[1]
Pinsker's inequality states that, if P and Q are two probability distributions, then
where
is the total variation distance (or statistical distance) between P and Q and
is the Kullback-Leibler divergence in nats.
Pinsker first proved the inequality with a worse constant. The inequality in the above form was proved independently by Kullback, Csiszár, and Kemperman.[2]
An inverse of the inequality cannot hold: for every , there are distributions with but .[3]
References
43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.
Additional reading
- Thomas M. Cover and Joy A. Thomas: Elements of Information Theory, 2nd edition, Willey-Interscience, 2006
- Nicolo Cesa-Bianchi and Gábor Lugosi: Prediction, Learning, and Games, Cambridge University Press, 2006
- ↑ 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.
My blog: http://www.primaboinca.com/view_profile.php?userid=5889534 - ↑ 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.
My blog: http://www.primaboinca.com/view_profile.php?userid=5889534 - ↑ The divergence becomes infinite whenever one of the two distributions assigns probability zero to an event while the other assigns it a nonzero probability (no matter how small); see e.g. 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.
My blog: http://www.primaboinca.com/view_profile.php?userid=5889534.