Mikhail Yakovlevich Suslin

From formulasearchengine
Revision as of 22:02, 26 October 2013 by en>Chimpionspeak
Jump to navigation Jump to search

The theory of optimal control is concerned with operating a dynamic system at minimum cost. The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called the LQ problem. One of the main results in the theory is that the solution is provided by the linear-quadratic regulator (LQR), a feedback controller whose equations are given below. The LQR is an important part of the solution to the LQG problem. Like the LQR problem itself, the LQG problem is one of the most fundamental problems in control theory.

General description

This means that the settings of a (regulating) controller governing either a machine or process (like an airplane or chemical reactor) are found by using a mathematical algorithm that minimizes a cost function with weighting factors supplied by a human (engineer). The "cost" (function) is often defined as a sum of the deviations of key measurements from their desired values. In effect this algorithm finds those controller settings that minimize the undesired deviations, like deviations from desired altitude or process temperature. Often the magnitude of the control action itself is included in this sum so as to keep the energy expended by the control action itself limited.

In effect, the LQR algorithm takes care of the tedious work done by the control systems engineer in optimizing the controller. However, the engineer still needs to specify the weighting factors and compare the results with the specified design goals. Often this means that controller synthesis will still be an iterative process where the engineer judges the produced "optimal" controllers through simulation and then adjusts the weighting factors to get a controller more in line with the specified design goals.

The LQR algorithm is, at its core, just an automated way of finding an appropriate state-feedback controller. As such it is not uncommon to find that control engineers prefer alternative methods like full state feedback (also known as pole placement) to find a controller over the use of the LQR algorithm. With these the engineer has a much clearer linkage between adjusted parameters and the resulting changes in controller behavior. Difficulty in finding the right weighting factors limits the application of the LQR based controller synthesis.


Finite-horizon, continuous-time LQR

For a continuous-time linear system, defined on t[t0,t1], described by

x˙=Ax+Bu

with a quadratic cost function defined as

J=12xT(t1)F(t1)x(t1)+t0t1(xTQx+uTRu)dt

the feedback control law that minimizes the value of the cost is

u=Kx

where K is given by

K=R1BTP(t)

and P is found by solving the continuous time Riccati differential equation.

ATP(t)+P(t)AP(t)BR1BTP(t)+Q=P˙(t)

with the boundary condition

P(t1)=F(t1).

The first order conditions for Jmin are

(i) State equation

x˙=Ax+Bu

(ii) Co-state equation

λ˙=Qx+ATλ

(iii) Stationary equation

0=Ru+BTλ

(iv) Boundary conditions

x(t0)=x0

and λ(t1)=F(t1)x(t1)

Infinite-horizon, continuous-time LQR

For a continuous-time linear system described by

x˙=Ax+Bu

with a cost functional defined as

J=0(xTQx+uTRu)dt

the feedback control law that minimizes the value of the cost is

u=Kx

where K is given by

K=R1BTP

and P is found by solving the continuous time algebraic Riccati equation

ATP+PAPBR1BTP+Q=0

Finite-horizon, discrete-time LQR

For a discrete-time linear system described by [1]

xk=Axk1+Buk

with a performance index defined as

J=k=0N(xkTQxk+ukTRuk)

the optimal control sequence minimizing the performance index is given by

uk=Fkxk1

where

Fk=(R+BTPkB)1BTPkA

and Pk is found iteratively backwards in time by the dynamic Riccati equation

Pk1=Q+AT(PkPkB(R+BTPkB)1BTPk)A

from initial condition PN=Q.

Infinite-horizon, discrete-time LQR

For a discrete-time linear system described by

xk+1=Axk+Buk

with a performance index defined as

J=k=0(xkTQxk+ukTRuk)

the optimal control sequence minimizing the performance index is given by

uk=Fxk

where

F=(R+BTPB)1BTPA

and P is the unique positive definite solution to the discrete time algebraic Riccati equation (DARE)

P=Q+AT(PPB(R+BTPB)1BTP)A.

Note that one way to solve this equation is by iterating the dynamic Riccati equation of the finite-horizon case until it converges.

References

  1. 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.

    My blog: http://www.primaboinca.com/view_profile.php?userid=5889534
  • 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.

    My blog: http://www.primaboinca.com/view_profile.php?userid=5889534
  • 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.

    My blog: http://www.primaboinca.com/view_profile.php?userid=5889534

External links