Kolmogorov's inequality: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>DrMicro
 
en>Egsan Bacon
Proof: Repairing links to disambiguation pages - You can help!
 
Line 1: Line 1:
Nice to satisfy you, my title is Numbers Held although I don't really like becoming called like that. North Dakota is exactly where me and my husband live. Since she was eighteen she's been operating as a meter reader but she's always wanted her own company. To perform baseball is the hobby he will by no means stop doing.<br><br>my weblog; [http://www.articlecube.com/profile/Nydia-Lyons/532548 articlecube.com]
'''Trust region''' is a term used in [[optimization (mathematics)|mathematical optimization]] to denote the subset of the region of the [[objective function]] to be optimized that is approximated using a model function (often a [[quadratic function|quadratic]]). If an adequate model of the objective function is found within the trust region then the region is expanded; conversely, if the approximation is poor then the region is contracted. Trust region methods are also known as '''restricted step methods'''.
 
The fit is evaluated by comparing the ratio of expected improvement from the model approximation with the actual improvement observed in the objective function. Simple thresholding of the ratio is used as the criteria for expansion and contraction&mdash;a model function is "trusted" only in the region where it provides a reasonable approximation.
 
Trust region methods are in some sense dual to [[line search]] methods: trust region methods first choose a step size (the size of the trust region) and then a step direction while line search methods first choose a step direction and then a step size.
 
The earliest use of the term seems to be by Sorensen (1982).
 
==Example==
 
Conceptually, in the [[Levenberg–Marquardt algorithm]], the objective function is iteratively approximated by a [[quadratic surface]], then using a linear solve, the estimate is updated. This alone may not converge nicely if the initial guess is too far from the optimum. For this reason, the algorithm instead restricts each step, preventing it from stepping "too far". It operationalizes "too far" as follows. Rather than solving <math>A \Delta x = b</math> for <math>\Delta x</math>, it solves <math>(A + \lambda \operatorname{diag}(A) ) \Delta x = b</math> where <math>\operatorname{diag}(A)</math> is the diagonal matrix with the same diagonal as ''A'' and λ is a parameter that controls the trust-region size. Geometrically, this adds a parabaloid centered at <math>\Delta x=0</math> to the [[quadratic form]], resulting in a smaller step.  
 
The trick is to change the trust-region size (λ). At each iteration, the damped quadratic fit predicts a certain reduction in the cost function, <math>\Delta f_{\mathrm{pred}}</math>, which we would expect to be a smaller reduction than the true reduction. Given <math>\Delta x</math> we can evaluate
: <math>\Delta f_{\mathrm{actual}} = f(x+\Delta x) - f(x)</math>
By looking at the ratio <math>\Delta f_{\mathrm{pred}}/\Delta f_{\mathrm{actual}}</math> we can adjust the trust-region size. In general, we expect <math>\Delta f_{\mathrm{pred}}</math> to be a bit less than <math>\Delta f_{\mathrm{actual}}</math> and so the ratio would be between, say, 0.25 and 0.5. If the ratio is more than 0.5, then we aren't damping the step much, so expand the trust region (decrease λ), and iterate. If the ratio is smaller than 0.25, then the true function is diverging "too much" from the trust-region approximation, so shrink the trust region (increase λ) and try again.
 
==References==
 
* Andrew R. Conn, Nicholas I. M. Gould, Philippe L. Toint "[http://books.google.com/books?id=5kNC4fqssYQC Trust-Region Methods (MPS-SIAM Series on Optimization)]".
* Byrd, R. H, R. B. Schnabel, and G. A. Schultz. "[http://epubs.siam.org/doi/abs/10.1137/0724076 A trust region algorithm for nonlinearly constrained optimization]", SIAM J. Numer. Anal., 24 (1987), pp.&nbsp;1152–1170.
* Sorensen, D. C.: "[http://epubs.siam.org/doi/abs/10.1137/0719026 Newton’s Method with a Model Trust Region Modification]", SIAM J. Numer. Anal., 19(2), 409–426 (1982).
* Yuan, Y. "[http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.45.9964 A review of trust region algorithms for optimization]" in ICIAM 99: Proceedings of the Fourth International Congress on Industrial & Applied Mathematics, Edinburgh, 2000  Oxford University Press, USA.
 
== External links ==
* [http://www.applied-mathematics.net/optimization/optimizationIntro.html Kranf site: Trust Region Algorithms]
 
{{Optimization algorithms}}
[[Category:Mathematical optimization]]
[[Category:Optimization algorithms and methods]]

Latest revision as of 13:53, 10 October 2013

Trust region is a term used in mathematical optimization to denote the subset of the region of the objective function to be optimized that is approximated using a model function (often a quadratic). If an adequate model of the objective function is found within the trust region then the region is expanded; conversely, if the approximation is poor then the region is contracted. Trust region methods are also known as restricted step methods.

The fit is evaluated by comparing the ratio of expected improvement from the model approximation with the actual improvement observed in the objective function. Simple thresholding of the ratio is used as the criteria for expansion and contraction—a model function is "trusted" only in the region where it provides a reasonable approximation.

Trust region methods are in some sense dual to line search methods: trust region methods first choose a step size (the size of the trust region) and then a step direction while line search methods first choose a step direction and then a step size.

The earliest use of the term seems to be by Sorensen (1982).

Example

Conceptually, in the Levenberg–Marquardt algorithm, the objective function is iteratively approximated by a quadratic surface, then using a linear solve, the estimate is updated. This alone may not converge nicely if the initial guess is too far from the optimum. For this reason, the algorithm instead restricts each step, preventing it from stepping "too far". It operationalizes "too far" as follows. Rather than solving AΔx=b for Δx, it solves (A+λdiag(A))Δx=b where diag(A) is the diagonal matrix with the same diagonal as A and λ is a parameter that controls the trust-region size. Geometrically, this adds a parabaloid centered at Δx=0 to the quadratic form, resulting in a smaller step.

The trick is to change the trust-region size (λ). At each iteration, the damped quadratic fit predicts a certain reduction in the cost function, Δfpred, which we would expect to be a smaller reduction than the true reduction. Given Δx we can evaluate

Δfactual=f(x+Δx)f(x)

By looking at the ratio Δfpred/Δfactual we can adjust the trust-region size. In general, we expect Δfpred to be a bit less than Δfactual and so the ratio would be between, say, 0.25 and 0.5. If the ratio is more than 0.5, then we aren't damping the step much, so expand the trust region (decrease λ), and iterate. If the ratio is smaller than 0.25, then the true function is diverging "too much" from the trust-region approximation, so shrink the trust region (increase λ) and try again.

References

External links

Template:Optimization algorithms