Skew-Hamiltonian matrix: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>Addbot
m Bot: Migrating 1 interwiki links, now provided by Wikidata on d:q3851920
en>Rich Farmbrough
No edit summary
 
Line 1: Line 1:
In [[probability theory]] and [[statistics]], a '''conditional variance''' is the [[variance]] of a [[conditional probability distribution]]. Particularly in [[econometrics]], the conditional variance is also known as the ''[[scedastic function]]'' or ''skedastic function''. Conditional variances are important parts of [[autoregressive conditional heteroskedasticity]] (ARCH) models.  
Friends call him Royal Seyler. I am a production and distribution officer. Playing croquet is something I will by no means give up. Years ago we moved to Kansas.<br><br>Here is my weblog ... extended auto warranty [[http://maltarenting.com/ActivityFeed/MyProfile/tabid/60/userId/175295/Default.aspx more information]]
 
==Definition==
The conditional variance of a [[random variable]] ''Y'' given that the value of a random variable ''X'' takes the value ''x'' is
 
:<math>\operatorname{Var}(Y|X=x) = \operatorname{E}((Y - \operatorname{E}(Y\mid X=x))^{2}\mid X=x),</math>
 
where E is the [[expectation operator]] with respect to the [[conditional distribution]] of ''Y'' given that the ''X'' takes the value ''x''. An alternative notation for this is :<math>\operatorname{Var}_{Y\mid X}(Y|x).</math>
 
The above may be stated in the alternative form that, based on the [[conditional distribution]] of ''Y'' given that the ''X'' takes the value ''x'', the conditional variance is the [[variance]] of this [[probability distribution]].
 
==Components of variance==
The [[law of total variance]] says
 
:<math>\operatorname{Var}(Y) = \operatorname{E}(\operatorname{Var}(Y\mid X))+\operatorname{Var}(\operatorname{E}(Y\mid X)),</math>
 
where, for example, <math>\operatorname{Var}(Y|X)</math> is understood to mean that the value ''x'' at which the conditional variance would be evaluated is allowed to be a [[random variable]], ''X''. In this "law", the inner expectation or variance is taken with respect to ''Y'' conditional on ''X'', while the outer expectation or variance is taken with respect to ''X''. This expression represents the overall variance of ''Y'' as the sum of two components, involving a prediction of ''Y'' based on ''X''. Specifically, let the predictor be the least-mean-squares prediction based on ''X'', which is the [[conditional expectation]] of ''Y'' given ''X''. Then the two components are:
:*the average of the variance of ''Y'' about the prediction based on ''X'', as ''X'' varies;
:*the variance of the prediction based on ''X'', as ''X'' varies.
 
[[Category:Statistical deviation and dispersion]]
[[Category:Statistical terminology]]
[[Category:Theory of probability distributions]]
 
{{statistics-stub}}
{{probability-stub}}

Latest revision as of 02:25, 22 June 2014

Friends call him Royal Seyler. I am a production and distribution officer. Playing croquet is something I will by no means give up. Years ago we moved to Kansas.

Here is my weblog ... extended auto warranty [more information]