Shannon Criteria

From formulasearchengine
Revision as of 01:00, 30 January 2014 by en>Monkbot (Fix CS1 deprecated date parameter errors)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

In numerical mathematics, the Uzawa iteration is an algorithm for solving saddle point problems. It is named after Hirofumi Uzawa and was originally introduced in the context of concave programming.[1]

Basic idea

We consider a saddle point problem of the form

(ABB*)(x1x2)=(b1b2),

where A is a symmetric positive-definite matrix. Multiplying the first row by B*A1 and subtracting from the second row yields the upper-triangular system

(ABS)(x1x2)=(b1b2B*A1b1),

where S:=B*A1B denotes the Schur complement. Since S is symmetric positive-definite, we can apply standard iterative methods like the gradient descent method or the conjugate gradient method to

Sx2=B*A1b1b2

in order to compute x2. The vector x1 can be reconstructed by solving

Ax1=b1Bx2.

It is possible to update x1 alongside x2 during the iteration for the Schur complement system and thus obtain an efficient algorithm.

Implementation

We start the conjugate gradient iteration by computing the residual

r2:=b2B*A1b1Sx2=b2B*A1(b1Bx2)=b2B*x1,

of the Schur complement system, where

x1:=A1(b1Bx2)

denotes the upper half of the solution vector matching the initial guess x2 for its lower half. We complete the initialization by choosing the first search direction

p2:=r2.

In each step, we compute

a2:=Sp2=B*A1Bp2=B*p1

and keep the intermediate result

p1:=A1Bp2

for later. The scaling factor is given by

α:=p2*r2/p2*a2

and leads to the updates

x2:=x2+αp2,r2:=r2αa2.

Using the intermediate result p1 saved earlier, we can also update the upper part of the solution vector

x1:=x1αp1.

Now we only have to construct the new search direction by the Gram–Schmidt process, i.e.,

β:=r2*a2/p2*a2,p2:=r2βp2.

The iteration terminates if the residual r2 has become sufficiently small or if the norm of p2 is significantly smaller than r2 indicating that the Krylov subspace has been almost exhausted.

Modifications and extensions

If solving the linear system Ax=b exactly is not feasible, inexact solvers can be applied. [2] [3] [4]

If the Schur complement system is ill-conditioned, preconditioners can be employed to improve the speed of convergence of the underlying gradient method.[2] [5]

Inequality constraints can be incorporated, e.g., in order to handle obstacle problems.[5]

Literature

  1. H. Uzawa, Iterative methods for concave programming, in K. J. Arrow, L. Hurwicz, H. Uzawa, Studies in linear and nonlinear programming, Stanford University Press 1958
  2. 2.0 2.1 H. C. Elman and G. H. Golub, Inexact and preconditioned Uzawa algorithms for saddle point problems, SIAM J. Num. Anal. 31(6):1645–1661 (1994)
  3. J. H. Bramble, J. E. Pasciak and A. T. Vassilev, Analysis of the inexact Uzawa algorithm for saddle point problems, SIAM J. Num. Anal. 34:1072–1982 (1997)
  4. W. Zulehner, Analysis of iterative methods for saddle point problems. A unified approach, Math. Comp. 71:479–505 (1998)
  5. 5.0 5.1 C. Gräser and R. Kornhuber, On preconditioned Uzawa-type iterations for a saddle point problem with inequality constraints, Domain Decomposition Methods in Science and Engineering XVI, Lec. Not. Comp. Sci. Eng. 55:91–102 (2007)