Decimal degrees: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
Precision: using better unit prefixes
 
(One intermediate revision by one other user not shown)
Line 1: Line 1:
The '''Hartley function''' is a measure of uncertainty, introduced by [[Ralph Hartley]] in 1928. If we pick a sample from a finite set ''A'' uniformly at random, the information revealed after we know the outcome is given by the Hartley function
They contact me Emilia. I am a meter reader but I plan on changing it. For a whilst I've been in South Dakota and my mothers and fathers reside nearby. What I adore doing is taking part in baseball but I haven't produced a dime with it.<br><br>Here is my web page :: [http://www.cam4teens.com/blog/84472 home std test kit]
:<math> H_0(A) := \mathrm{log}_b \vert A \vert .</math>
If the [[base (exponentiation)|base]] of the [[logarithm]] is 2, then the uncertainty is measured in bits. If it is the [[natural logarithm]], then the unit is [[Nat (information)|nats]]. (Hartley himself used a base-ten logarithm, and this unit of information is sometimes called the '''[[Ban (information)|hartley]]''' in his honor.)  It is also known as the Hartley entropy.
 
== Hartley function, Shannon's entropy, and Rényi entropy ==
The Hartley function coincides with the [[Shannon entropy]] (as well as with the Rényi entropies of all orders) in the case of a uniform probability distribution. It is actually a special case of the [[Rényi entropy]] since:
:<math>H_0(X) = \frac 1 {1-0} \log \sum_{i=1}^{|X|} p_i^0 = \log |X|.</math>
 
But it can also be viewed as a primitive construction, since, as emphasized by Kolmogorov and Rényi (see George, J. Klirr's "Uncertainty and information", p.423), the Hartley function can be defined without introducing any notions of probability.
 
==Characterization of the Hartley function==
The Hartley function only depends on the number of elements in a set, and hence can be viewed as a function on natural numbers. Rényi showed that the Hartley function in base 2 is the only function mapping natural numbers to real numbers that satisfies
 
# <math>H(mn) = H(m)+H(n)</math> (additivity)
# <math>H(m) \leq H(m+1)</math> (monotonicity)
# <math>H(2)=1</math> (normalization)
 
Condition 1 says that the uncertainty of the Cartesian product of two finite sets ''A'' and ''B'' is the sum of uncertainties of ''A'' and ''B''. Condition 2 says that larger set has larger uncertainty.
 
==Derivation of the Hartley function==
We want to show that the Hartley function, log<sub>2</sub>(''n''), is the only function mapping natural numbers to real numbers that satisfies
 
# <math>H(mn) = H(m)+H(n)\,</math> (additivity)
# <math>H(m) \leq H(m+1)\,</math> (monotonicity)
# <math>H(2)=1\,</math> (normalization)
 
Let ''&fnof;'' be a function on positive integers that satisfies the above three properties. From the additive property, we can show that for any integer ''n'' and ''k'',
 
:<math>f(n^k) = kf(n).\,</math>
 
Let ''a'', ''b'', and ''t'' be any positive integers. There is a unique integer ''s'' determined by
 
:<math>a^s \leq b^t \leq a^{s+1}. \qquad(1)</math>
 
Therefore,
 
:<math>s \log_2 a\leq t \log_2 b \leq (s+1) \log_2 a \, </math>
 
and
 
:<math>\frac{s}{t} \leq \frac{\log_2 b}{\log_2 a} \leq \frac{s+1}{t}.</math>
 
On the other hand, by monotonicity,
 
:<math>f(a^s) \leq f(b^t) \leq f(a^{s+1}). \, </math>
 
Using Equation (1), we get
 
:<math>s f(a) \leq t f(b) \leq (s+1) f(a),\,</math>
 
and
 
:<math>\frac{s}{t} \leq \frac{f(a)}{f(b)} \leq \frac{s+1}{t}.</math>
 
Hence,
 
:<math>\Big\vert \frac{f(a)}{f(b)} - \frac{\log_2(a)}{\log_2(b)} \Big\vert \leq \frac{1}{t}.</math>
 
Since ''t'' can be arbitrarily large, the difference on the left hand side of the above inequality must be zero,
 
:<math>\frac{f(a)}{f(b)} = \frac{\log_2(a)}{\log_2(b)}.</math>
 
So,
 
:<math>f(a) = \mu \log_2(a)\,</math>
 
for some constant ''μ'', which must be equal to 1 by the normalization property.
 
==See also==
* [[Rényi entropy]]
 
{{PlanetMath attribution|id=6070|title=Hartley function}}
{{PlanetMath attribution|id=6082|title=Derivation of Hartley function}}
 
[[Category:Information theory]]

Latest revision as of 21:57, 18 December 2014

They contact me Emilia. I am a meter reader but I plan on changing it. For a whilst I've been in South Dakota and my mothers and fathers reside nearby. What I adore doing is taking part in baseball but I haven't produced a dime with it.

Here is my web page :: home std test kit