Block code: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>Mogism
m Sphere packings and lattices: Cleanup/Typo fixing, typo(s) fixed: so called → so-called using AWB
en>BD2412
Line 1: Line 1:
In [[probability theory]], '''Hoeffding's inequality''' provides an [[upper bound]] on the [[probability]] that the sum of [[random variables]] deviates from its [[expected value]].
Marvella is what you can call her but it's not the most feminine name out there. Body developing is 1 of the issues I love most. Her husband and her live in Puerto Rico but she will have to move one working day or an additional. My working day job is a meter reader.<br><br>My site; diet meal delivery ([http://r.fbtest.net/dietmealdelivery74619 source website])
Hoeffding's inequality was proved by [[Wassily Hoeffding]] in 1963.<ref>{{harvtxt|Hoeffding|1963}}</ref>
 
Hoeffding's inequality is a special case of the [[Azuma–Hoeffding inequality]], and it is more general than the [[Bernstein inequalities in probability theory|Bernstein inequality]], proved by [[Sergei Bernstein]] in 1923. They are also special cases of [[McDiarmid's inequality]].
 
== Special case of Bernoulli random variables ==
 
Hoeffding's inequality can be applied to the important special case of identically distributed [[Bernoulli trial|Bernoulli random variables]], and this is how the inequality is often used in [[combinatorics]] and [[computer science]].
We consider a coin that shows heads with probability <math>p</math> and tails with probability <math>1-p</math>.
We toss the coin <math>n</math> times.
The [[expected value|expected]] number of times the coin comes up heads is <math>p\cdot n</math>.
Furthermore, the probability that the coin comes up heads at most <math>k</math> times can be exactly quantified by the following expression:
 
:<math>\Pr\Big(n \text{ coin tosses yield heads at most } k \text{ times}\Big)= \sum_{i=0}^{k} \binom{n}{i} p^i (1-p)^{n-i}\,.</math>
 
In the case that <math>k=(p-\epsilon) n</math> for some <math>\epsilon > 0</math>,
Hoeffding's inequality bounds this probability by a term that is exponentially small in <math>\epsilon^2 \cdot n</math>:
:<math>\Pr\Big(n \text{ coin tosses yield heads at most } (p-\epsilon) n \text{ times}\Big)\leq\exp\big(-2\epsilon^2 n\big)\,.</math>
 
Similarly, in the case that <math>k=(p+\epsilon) n</math> for some <math>\epsilon > 0</math>,
Hoeffding's inequality bounds the probability that we see at least <math>\epsilon n</math> more tosses that show heads than we would expect:
:<math>\Pr\Big(n \text{ coin tosses yield heads at least } (p+\epsilon) n \text{ times}\Big)\leq\exp\big(-2\epsilon^2 n\big)\,.</math>
 
Hence Hoeffding's inequality implies that the number of heads that we see is concentrated around its mean, with exponentially small tail.
:<math>\Pr\Big(n \text{ coin tosses yield heads between } (p-\epsilon)n \text{ and } (p+\epsilon)n \text{ times}\Big)\geq 1-2\exp\big(-2\epsilon^2 n\big)\,.</math>
 
== General case ==
 
Let
 
:<math>X_1, \dots, X_n \!</math>
 
be [[independent random variables]].
Assume that the <math>X_i</math> are [[almost sure]]ly bounded; that is, assume for <math>1 \leq i \leq n</math> that
 
:<math>\Pr(X_i \in [a_i, b_i]) = 1. \!</math>
 
We define the empirical mean of these variables
 
:<math>\overline X = \frac{1}{n}(X_1 + \cdots + X_n).</math>
 
Theorem 2 of {{harvtxt|Hoeffding|1963}} proves the inequalities
 
:<math>\Pr(\overline X - \mathrm{E}[\overline X] \geq t) \leq \exp \left( - \frac{2n^2t^2}{\sum_{i=1}^n (b_i - a_i)^2} \right),\!</math>
:<math>\Pr(|\overline X - \mathrm{E}[\overline X]| \geq t) \leq 2\exp \left( - \frac{2n^2t^2}{\sum_{i=1}^n (b_i - a_i)^2} \right),\!</math>
 
which are valid for positive values of ''t''. Here <math>\mathrm{E}[\overline X]</math> is the [[expected value]] of <math>\overline X</math>.
The inequalities can be also stated in terms of the sum
 
:<math>S = X_1 + \cdots + X_n</math>
 
of the random variables:
 
:<math>\Pr(S - \mathrm{E}[S] \geq t) \leq \exp \left( - \frac{2t^2}{\sum_{i=1}^n (b_i - a_i)^2} \right),\!</math>
:<math>\Pr(|S - \mathrm{E}[S]| \geq t) \leq 2\exp \left( - \frac{2t^2}{\sum_{i=1}^n (b_i - a_i)^2} \right).\!</math>
 
Note that the inequalities also hold when the <math>X_i</math> have been obtained using sampling without replacement; in this case the random variables are not independent anymore. A proof of this statement can be found in Hoeffding's paper. For slightly better bounds in the case of sampling without replacement, see for instance the paper by {{harvtxt|Serfling|1974}}.
 
==See also==
*[[Bennett's inequality]]
*[[Chebyshev's inequality]]
*[[Markov's inequality]]
*[[Chernoff bounds]]
*[[Hoeffding's lemma]]
 
==Notes==
{{reflist}}
 
==References==
{{refbegin}}
* {{cite journal
| first1=Robert J. | last1=Serfling
| title=Probability Inequalities for the Sum in Sampling without Replacement
| journal=The Annals of Statistics
| pages=39–48
| year=1974
| ref=harv
| volume=2
| number=1
| doi=10.1214/aos/1176342611}}
* {{cite journal
| first1=Wassily | last1=Hoeffding
| title=Probability inequalities for sums of bounded random variables
| journal=Journal of the American Statistical Association
| pages=13–30
|date=March 1963
| ref=harv
| volume=58
| number=301
| jstor=2282952}}
 
  {{refend}}
 
[[Category:Probabilistic inequalities]]

Revision as of 02:45, 20 February 2014

Marvella is what you can call her but it's not the most feminine name out there. Body developing is 1 of the issues I love most. Her husband and her live in Puerto Rico but she will have to move one working day or an additional. My working day job is a meter reader.

My site; diet meal delivery (source website)