Fermion doubling: Difference between revisions
en>Gaius Cornelius m →The Nielsen–Ninomiya theorem: Fix indefinite article, replaced: an nonlocal → a nonlocal using AWB |
en>Addbot m Bot: Migrating 2 interwiki links, now provided by Wikidata on d:q5444417 |
||
Line 1: | Line 1: | ||
{{Probability distribution| | |||
name =Scaled inverse chi-squared| | |||
type =density| | |||
pdf_image =[[File:Scaled inverse chi squared.svg|250px]]| | |||
cdf_image =[[File:Scaled inverse chi squared cdf.svg|250px]]| | |||
parameters =<math>\nu > 0\,</math><BR /><math>\tau^2 > 0\,</math> | | |||
support =<math>x \in (0, \infty)</math>| | |||
pdf =<math>\frac{(\tau^2\nu/2)^{\nu/2}}{\Gamma(\nu/2)}~ | |||
\frac{\exp\left[ \frac{-\nu \tau^2}{2 x}\right]}{x^{1+\nu/2}} </math>| | |||
cdf =<math>\Gamma\left(\frac{\nu}{2},\frac{\tau^2\nu}{2x}\right) | |||
\left/\Gamma\left(\frac{\nu}{2}\right)\right.</math>| | |||
mean =<math>\frac{\nu \tau^2}{\nu-2}</math> for <math>\nu >2\,</math>| | |||
median =| | |||
mode =<math>\frac{\nu \tau^2}{\nu+2}</math>| | |||
variance =<math>\frac{2 \nu^2 \tau^4}{(\nu-2)^2 (\nu-4)}</math>for <math>\nu >4\,</math>| | |||
skewness =<math>\frac{4}{\nu-6}\sqrt{2(\nu-4)}</math>for <math>\nu >6\,</math>| | |||
kurtosis =<math>\frac{12(5\nu-22)}{(\nu-6)(\nu-8)}</math>for <math>\nu >8\,</math>| | |||
entropy =<math>\frac{\nu}{2} | |||
\!+\!\ln\left(\frac{\tau^2\nu}{2}\Gamma\left(\frac{\nu}{2}\right)\right)</math> | |||
<math>\!-\!\left(1\!+\!\frac{\nu}{2}\right)\psi\left(\frac{\nu}{2}\right)</math>| | |||
mgf =<math>\frac{2}{\Gamma(\frac{\nu}{2})}\left(\frac{-\tau^2\nu t}{2}\right)^{\!\!\frac{\nu}{4}}\!\!K_{\frac{\nu}{2}}\left(\sqrt{-2\tau^2\nu t}\right)</math>| | |||
char =<math>\frac{2}{\Gamma(\frac{\nu}{2})}\left(\frac{-i\tau^2\nu t}{2}\right)^{\!\!\frac{\nu}{4}}\!\!K_{\frac{\nu}{2}}\left(\sqrt{-2i\tau^2\nu t}\right)</math>| | |||
}} | |||
The '''scaled inverse chi-squared distribution''' is the distribution for ''x'' = 1/''s''<sup>2</sup>, where ''s''<sup>2</sup> is a sample mean of the squares of ν independent [[normal distribution|normal]] random variables that have mean 0 and inverse variance 1/σ<sup>2</sup> = τ<sup>2</sup>. The distribution is therefore parametrised by the two quantities ν and τ<sup>2</sup>, referred to as the ''number of chi-squared degrees of freedom'' and the ''scaling parameter'', respectively. | |||
This family of scaled inverse chi-squared distributions is closely related to two other distribution families, those of the [[inverse-chi-squared distribution]] and the [[inverse gamma distribution]]. Compared to the inverse-chi-squared distribution, the scaled distribution has an extra parameter ''τ''<sup>2</sup>, which scales the distribution horizontally and vertically, representing the inverse-variance of the original underlying process. Also, the scale inverse chi-squared distribution is presented as the distribution for the inverse of the ''mean'' of ν squared deviates, rather than the inverse of their ''sum''. The two distributions thus have the relation that if | |||
:<math>X \sim \mbox{Scale-inv-}\chi^2(\nu, \tau^2)</math> then <math> \frac{X}{\tau^2 \nu} \sim \mbox{inv-}\chi^2(\nu)</math> | |||
Compared to the inverse gamma distribution, the scaled inverse chi-squared distribution describes the same data distribution, but using a different [[parametrization]], which may be more convenient in some circumstances. Specifically, if | |||
:<math>X \sim \mbox{Scale-inv-}\chi^2(\nu, \tau^2)</math> then <math>X \sim \textrm{Inv-Gamma}\left(\frac{\nu}{2}, \frac{\nu\tau^2}{2}\right)</math> | |||
Either form may be used to represent the [[maximum entropy]] distribution for a fixed first inverse [[Moment (mathematics)|moment]] <math>(E(1/X))</math> and first logarithmic moment <math>(E(\ln(X))</math>. | |||
The scaled inverse chi-squared distribution also has a particular use in [[Bayesian statistics]], somewhat unrelated to its use as a predictive distribution for ''x'' = 1/''s''<sup>2</sup>. Specifically, the scaled inverse chi-squared distribution can be used as a [[conjugate prior]] for the [[variance]] parameter of a [[normal distribution]]. In this context the scaling parameter is denoted by σ<sub>0</sub><sup>2</sup> rather than by τ<sup>2</sup>, and has a different interpretation. The application has been more usually presented using the [[inverse gamma distribution]] formulation instead; however, some authors, following in particular Gelman ''et al.'' (1995/2004) argue that the inverse chi-squared parametrisation is more intuitive. | |||
==Characterization== | |||
The [[probability density function]] of the scaled inverse chi-squared distribution extends over the domain <math>x>0</math> and is | |||
:<math> | |||
f(x; \nu, \tau^2)= | |||
\frac{(\tau^2\nu/2)^{\nu/2}}{\Gamma(\nu/2)}~ | |||
\frac{\exp\left[ \frac{-\nu \tau^2}{2 x}\right]}{x^{1+\nu/2}} | |||
</math> | |||
where <math>\nu</math> is the [[degrees of freedom (statistics)|degrees of freedom]] parameter and <math>\tau^2</math> is the [[scale parameter]]. The cumulative distribution function is | |||
:<math>F(x; \nu, \tau^2)= | |||
\Gamma\left(\frac{\nu}{2},\frac{\tau^2\nu}{2x}\right) | |||
\left/\Gamma\left(\frac{\nu}{2}\right)\right.</math> | |||
:<math>=Q\left(\frac{\nu}{2},\frac{\tau^2\nu}{2x}\right)</math> | |||
where <math>\Gamma(a,x)</math> is the [[incomplete Gamma function]], <math>\Gamma(x)</math> is the [[Gamma function]] and <math>Q(a,x)</math> is a [[incomplete Gamma function|regularized Gamma function]]. The [[Characteristic function (probability theory)|characteristic function]] is | |||
:<math>\varphi(t;\nu,\tau^2)=</math> | |||
:<math>\frac{2}{\Gamma(\frac{\nu}{2})}\left(\frac{-i\tau^2\nu t}{2}\right)^{\!\!\frac{\nu}{4}}\!\!K_{\frac{\nu}{2}}\left(\sqrt{-2i\tau^2\nu t}\right) ,</math> | |||
where <math>K_{\frac{\nu}{2}}(z)</math> is the modified [[Bessel function]] of the second kind. | |||
==Parameter estimation== | |||
The [[maximum likelihood]] estimate of <math>\tau^2</math> is | |||
:<math>\tau^2 = n/\sum_{i=1}^n \frac{1}{x_i}.</math> | |||
The [[maximum likelihood]] estimate of <math>\frac{\nu}{2}</math> can be found using [[Newton's method]] on: | |||
:<math>\ln(\frac{\nu}{2}) + \psi(\frac{\nu}{2}) = \sum_{i=1}^n \ln(x_i) - n \ln(\tau^2) ,</math> | |||
where <math>\psi(x)</math> is the [[digamma function]]. An initial estimate can be found by taking the formula for mean and solving it for <math>\nu.</math> Let <math>\bar{x} = \frac{1}{n}\sum_{i=1}^n x_i</math> be the sample mean. Then an initial estimate for <math>\nu</math> is given by: | |||
:<math>\frac{\nu}{2} = \frac{\bar{x}}{\bar{x} - \tau^2}.</math> | |||
==Bayesian estimation of the variance of a Normal distribution== | |||
The scaled inverse chi-squared distribution has a second important application, in the Bayesian estimation of the variance of a Normal distribution. | |||
According to [[Bayes theorem]], the [[posterior probability distribution]] for quantities of interest is proportional to the product of a [[prior distribution]] for the quantities and a [[likelihood function]]: | |||
:<math>p(\sigma^2|D,I) \propto p(\sigma^2|I) \; p(D|\sigma^2)</math> | |||
where ''D'' represents the data and ''I'' represents any initial information about σ<sup>2</sup> that we may already have. | |||
The simplest scenario arises if the mean μ is already known; or, alternatively, if it is the [[conditional probability|conditional distribution]] of σ<sup>2</sup> that is sought, for a particular assumed value of μ. | |||
Then the likelihood term ''L''(σ<sup>2</sup>|''D'') = ''p''(''D''|σ<sup>2</sup>) has the familiar form | |||
:<math>\mathcal{L}(\sigma^2|D,\mu) = \frac{1}{\left(\sqrt{2\pi}\sigma\right)^n} \; \exp \left[ -\frac{\sum_i^n(x_i-\mu)^2}{2\sigma^2} \right]</math> | |||
Combining this with the rescaling-invariant prior p(σ<sup>2</sup>|''I'') = 1/σ<sup>2</sup>, which can be argued (e.g. [[Jeffreys_prior#Gaussian_distribution_with_standard_deviation_parameter|following Jeffreys]]) to be the least informative possible prior for σ<sup>2</sup> in this problem, gives a combined posterior probability | |||
:<math>p(\sigma^2|D, I, \mu) \propto \frac{1}{\sigma^{n+2}} \; \exp \left[ -\frac{\sum_i^n(x_i-\mu)^2}{2\sigma^2} \right]</math> | |||
This form can be recognised as that of a scaled inverse chi-squared distribution, with parameters ν = ''n'' and τ<sup>2</sup> = ''s''<sup>2</sup> = (1/''n'') Σ (x<sub>i</sub>-μ)<sup>2</sup> | |||
Gelman ''et al'' remark that the re-appearance of this distribution, previously seen in a sampling context, may seem remarkable; but given the choice of prior the "result is not surprising".<ref>Gelman ''et al'' (1995), ''Bayesian Data Analysis'' (1st ed), p.68</ref> | |||
In particular, the choice of a rescaling-invariant prior for σ<sup>2</sup> has the result that the probability for the ratio of σ<sup>2</sup> / ''s''<sup>2</sup> has the same form (independent of the conditioning variable) when conditioned on ''s''<sup>2</sup> as when conditioned on σ<sup>2</sup>: | |||
:<math>p(\tfrac{\sigma^2}{s^2}|s^2) = p(\tfrac{\sigma^2}{s^2}|\sigma^2)</math> | |||
In the sampling-theory case, conditioned on σ<sup>2</sup>, the probability distribution for (1/s<sup>2</sup>) is a scaled inverse chi-squared distribution; and so the probability distribution for σ<sup>2</sup> conditioned on ''s''<sup>2</sup>, given a scale-agnostic prior, is also a scaled inverse chi-squared distribution. | |||
=== Use as an informative prior === | |||
If more is known about the possible values of σ<sup>2</sup>, a distribution from the scaled inverse chi-squared family, such as Scale-inv-χ<sup>2</sup>(''n''<sub>0</sub>, ''s''<sub>0</sub><sup>2</sup>) can be a convenient form to represent a less uninformative prior for σ<sup>2</sup>, as if from the result of ''n''<sub>0</sub> previous observations (though ''n''<sub>0</sub> need not necessarily be a whole number): | |||
:<math>p(\sigma^2|I^\prime, \mu) \propto \frac{1}{\sigma^{n_0+2}} \; \exp \left[ -\frac{n_0 s_0^2}{2\sigma^2} \right]</math> | |||
Such a prior would lead to the posterior distribution | |||
:<math>p(\sigma^2|D, I^\prime, \mu) \propto \frac{1}{\sigma^{n+n_0+2}} \; \exp \left[ -\frac{\sum{ns^2 + n_0 s_0^2}}{2\sigma^2} \right]</math> | |||
which is itself a scaled inverse chi-squared distribution. The scaled inverse chi-squared distributions are thus a convenient [[conjugate prior]] family for σ<sup>2</sup> estimation. | |||
=== Estimation of variance when mean is unknown === | |||
If the mean is not known, the most uninformative prior that can be taken for it is arguably the translation-invariant prior ''p''(μ|''I'') ∝ const., which gives the following joint posterior distribution for μ and σ<sup>2</sup>, | |||
:<math>\begin{align} | |||
p(\mu, \sigma^2|D, I) \; \propto \; & \frac{1}{\sigma^{n+2}} \; \exp \left[ -\frac{\sum_i^n(x_i-\mu)^2}{2\sigma^2} \right] \\ | |||
= \; & \frac{1}{\sigma^{n+2}} \; \exp \left[ -\frac{\sum_i^n(x_i-\bar{x})^2}{2\sigma^2} \right] \; \exp \left[ -\frac{\sum_i^n(\mu -\bar{x})^2}{2\sigma^2} \right] | |||
\end{align}</math> | |||
The marginal posterior distribution for σ<sup>2</sup> is obtained from the joint posterior distribution by integrating out over μ, | |||
:<math>\begin{align} | |||
p(\sigma^2|D, I) \; \propto \; & \frac{1}{\sigma^{n+2}} \; \exp \left[ -\frac{\sum_i^n(x_i-\bar{x})^2}{2\sigma^2} \right] \; \int_{-\infty}^{\infty} \exp \left[ -\frac{\sum_i^n(\mu -\bar{x})^2}{2\sigma^2} \right] d\mu\\ | |||
= \; & \frac{1}{\sigma^{n+2}} \; \exp \left[ -\frac{\sum_i^n(x_i-\bar{x})^2}{2\sigma^2} \right] \; \sqrt{2 \pi \sigma^2 / n} \\ | |||
\propto \; & (\sigma^2)^{-(n+1)/2} \; \exp \left[ -\frac{(n-1)s^2}{2\sigma^2} \right] | |||
\end{align}</math> | |||
This is again a scaled inverse chi-squared distribution, with parameters <math>\scriptstyle{n-1}\;</math> and <math>\scriptstyle{s^2 = \sum (x_i - \bar{x})^2/(n-1)}</math>. | |||
==Related distributions== | |||
* If <math>X \sim \mbox{Scale-inv-}\chi^2(\nu, \tau^2)</math> then <math> k X \sim \mbox{Scale-inv-}\chi^2(\nu, k \tau^2)\, </math> | |||
* If <math>X \sim \mbox{inv-}\chi^2(\nu) \, </math> ([[Inverse-chi-squared distribution]]) then <math>X \sim \mbox{Scale-inv-}\chi^2(\nu, 1/\nu) \,</math> | |||
* If <math>X \sim \mbox{Scale-inv-}\chi^2(\nu, \tau^2)</math> then <math> \frac{X}{\tau^2 \nu} \sim \mbox{inv-}\chi^2(\nu) \, </math> ([[Inverse-chi-squared distribution]]) | |||
* If <math>X \sim \mbox{Scale-inv-}\chi^2(\nu, \tau^2)</math> then <math>X \sim \textrm{Inv-Gamma}\left(\frac{\nu}{2}, \frac{\nu\tau^2}{2}\right)</math> ([[Inverse-gamma distribution]]) | |||
* Scaled inverse chi square distribution is a special case of type 5 [[Pearson distribution]] | |||
== References == | |||
* Gelman A. ''et al'' (1995), ''Bayesian Data Analysis'', pp 474–475; also pp 47, 480 | |||
{{reflist}} | |||
{{ProbDistributions|continuous-semi-infinite}} | |||
{{DEFAULTSORT:Scaled-Inverse-Chi-Squared Distribution}} | |||
[[Category:Continuous distributions]] | |||
[[Category:Exponential family distributions]] | |||
[[Category:Probability distributions]] |
Revision as of 04:18, 14 March 2013
Template:Probability distribution The scaled inverse chi-squared distribution is the distribution for x = 1/s2, where s2 is a sample mean of the squares of ν independent normal random variables that have mean 0 and inverse variance 1/σ2 = τ2. The distribution is therefore parametrised by the two quantities ν and τ2, referred to as the number of chi-squared degrees of freedom and the scaling parameter, respectively.
This family of scaled inverse chi-squared distributions is closely related to two other distribution families, those of the inverse-chi-squared distribution and the inverse gamma distribution. Compared to the inverse-chi-squared distribution, the scaled distribution has an extra parameter τ2, which scales the distribution horizontally and vertically, representing the inverse-variance of the original underlying process. Also, the scale inverse chi-squared distribution is presented as the distribution for the inverse of the mean of ν squared deviates, rather than the inverse of their sum. The two distributions thus have the relation that if
Compared to the inverse gamma distribution, the scaled inverse chi-squared distribution describes the same data distribution, but using a different parametrization, which may be more convenient in some circumstances. Specifically, if
Either form may be used to represent the maximum entropy distribution for a fixed first inverse moment and first logarithmic moment .
The scaled inverse chi-squared distribution also has a particular use in Bayesian statistics, somewhat unrelated to its use as a predictive distribution for x = 1/s2. Specifically, the scaled inverse chi-squared distribution can be used as a conjugate prior for the variance parameter of a normal distribution. In this context the scaling parameter is denoted by σ02 rather than by τ2, and has a different interpretation. The application has been more usually presented using the inverse gamma distribution formulation instead; however, some authors, following in particular Gelman et al. (1995/2004) argue that the inverse chi-squared parametrisation is more intuitive.
Characterization
The probability density function of the scaled inverse chi-squared distribution extends over the domain and is
where is the degrees of freedom parameter and is the scale parameter. The cumulative distribution function is
where is the incomplete Gamma function, is the Gamma function and is a regularized Gamma function. The characteristic function is
where is the modified Bessel function of the second kind.
Parameter estimation
The maximum likelihood estimate of is
The maximum likelihood estimate of can be found using Newton's method on:
where is the digamma function. An initial estimate can be found by taking the formula for mean and solving it for Let be the sample mean. Then an initial estimate for is given by:
Bayesian estimation of the variance of a Normal distribution
The scaled inverse chi-squared distribution has a second important application, in the Bayesian estimation of the variance of a Normal distribution.
According to Bayes theorem, the posterior probability distribution for quantities of interest is proportional to the product of a prior distribution for the quantities and a likelihood function:
where D represents the data and I represents any initial information about σ2 that we may already have.
The simplest scenario arises if the mean μ is already known; or, alternatively, if it is the conditional distribution of σ2 that is sought, for a particular assumed value of μ.
Then the likelihood term L(σ2|D) = p(D|σ2) has the familiar form
Combining this with the rescaling-invariant prior p(σ2|I) = 1/σ2, which can be argued (e.g. following Jeffreys) to be the least informative possible prior for σ2 in this problem, gives a combined posterior probability
This form can be recognised as that of a scaled inverse chi-squared distribution, with parameters ν = n and τ2 = s2 = (1/n) Σ (xi-μ)2
Gelman et al remark that the re-appearance of this distribution, previously seen in a sampling context, may seem remarkable; but given the choice of prior the "result is not surprising".[1]
In particular, the choice of a rescaling-invariant prior for σ2 has the result that the probability for the ratio of σ2 / s2 has the same form (independent of the conditioning variable) when conditioned on s2 as when conditioned on σ2:
In the sampling-theory case, conditioned on σ2, the probability distribution for (1/s2) is a scaled inverse chi-squared distribution; and so the probability distribution for σ2 conditioned on s2, given a scale-agnostic prior, is also a scaled inverse chi-squared distribution.
Use as an informative prior
If more is known about the possible values of σ2, a distribution from the scaled inverse chi-squared family, such as Scale-inv-χ2(n0, s02) can be a convenient form to represent a less uninformative prior for σ2, as if from the result of n0 previous observations (though n0 need not necessarily be a whole number):
Such a prior would lead to the posterior distribution
which is itself a scaled inverse chi-squared distribution. The scaled inverse chi-squared distributions are thus a convenient conjugate prior family for σ2 estimation.
Estimation of variance when mean is unknown
If the mean is not known, the most uninformative prior that can be taken for it is arguably the translation-invariant prior p(μ|I) ∝ const., which gives the following joint posterior distribution for μ and σ2,
The marginal posterior distribution for σ2 is obtained from the joint posterior distribution by integrating out over μ,
This is again a scaled inverse chi-squared distribution, with parameters and .
Related distributions
- If then
- If (Inverse-chi-squared distribution) then
- If then (Inverse-chi-squared distribution)
- If then (Inverse-gamma distribution)
- Scaled inverse chi square distribution is a special case of type 5 Pearson distribution
References
- Gelman A. et al (1995), Bayesian Data Analysis, pp 474–475; also pp 47, 480
43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.
55 yrs old Metal Polisher Records from Gypsumville, has interests which include owning an antique car, summoners war hack and spelunkering. Gets immense motivation from life by going to places such as Villa Adriana (Tivoli).
my web site - summoners war hack no survey ios
- ↑ Gelman et al (1995), Bayesian Data Analysis (1st ed), p.68