Sie sind auf Seite 1von 23

# Optimal Control

Motivation
Bellmans Principle of Optimality
Discrete-Time Systems
Continuous-Time Systems
Control
Illustrative Examples
Motivation
Control design based on pole-placement
often has non unique solutions
Best locations for eigenvalues are
difficult to determine
Optimal control minimizes a performance
index based on time response
Control gains result from solving the
optimal control problem
f ( x ) qx 2 bx c

f (x) xT Qx bT x c
Where Q is a symmetric (QT=Q) nxn matrix
and b is an nx1 vector

## It can be shown that the Jacobian of f is

f f f f
2 xT
Q b T

x x1 x 2 x n
f (x ) x 1 2 x 2 2 x 1x 2 2 x 1
2 2

Matrix representation:
1 1 x 1 x1
f (x) x 1 x 2 2 0
1 2 x 2 x2
The value of x that minimizes f(x) (denoted by x*)
sets f
2xT Q bT 0 2Qx b 0
x
1
or equivalently x Q 1b
2

## Provided that the Hessian of f,

2f 2f 2f

21 x
2
x 1x 2 x 1x n
f
T f 2f 2f

H x 1x 2 x 2
2
x 2 x n 2Q
x x
2
f 2f

2f
x x x 2 x n x n
2
1 n

is positive definite
Positive Definite Matrixes
Definition: A symmetric matrix H is said to
be positive definite (denoted by H>0) if
xTHx>0 for any non zero vector x (semi
positive definite if it only satisfies xTHx0
for any x (denoted by H 0) ).
Positive definiteness (Sylvester) test: H
is positive definite iff all the principal
minors of H are positive:
h11 h12 h1n
h11 h12 h21 h22 h2 n
h11 0, 0,, 0
h21 h22
hn1 hn 2 hnn
1 1 x 1 x1
f (x) x 1 x 2 2 0
1 2 x2 x2

Q

Optimal solution:
1
1 1 1 1 1 2 2
x Q b
2 2 1 2 0 1
5 20

2 2
20
10
H 2Q 0 4 18

2 4
20
3 10 16
5 1
2 14

0
1
0
1 20 -1
12
10 1
Thus x* minimizes f(x)

-1
10

x2
0

5
5
0 1
10 8
-1
5 20
20 6
-2 10 10
4
-3 20
20 2
-4
0
-5
-5 0 5
x1
(LQ) Optimal Control
Given discrete-time state equation
x(k 1) Gx(k) Hu(k)

## Find control sequence u(k) to minimize

1 T

1 N 1 T
J x (N)Sx(N) x (k)Qx (k) uT (k)Ru(k)
2 2 k 0

S, Q 0, R 0 and symmetric
Performance Index (PI)
Control objective is to make x small by
penalizing large inputs and states
PI makes a compromise between
performance and control action

uTRu

xTQx

t
Principle of Optimality
4
2 7
5
9
1
3 8
6
Bellmans Principle of Optimality: At any
intermediate state xi in an optimal path
from x0 to xf, the policy from xi to goal xf
must itself constitute optimal policy
Discrete-Time LQ Formulation
Optimization of the last input (k=N-1):
1 T 1 T
JN x (N)Sx(N) u (N 1)Ru(N 1)
2 2

## Where x(N)=Gx(n-1)+Hu(n-1). The optimal

input at the last step is obtained by setting
JN x(N)
x (N)S
T
uT (N 1)R 0
u(N 1) u(N 1)

## u(N 1) Kx(N 1), K R H SH SG

T 1
Optimal Value of JN-1
Substituting the optimal value of u(N-1) in JN gives
JN*
1
2

x(N 1)T (G HK)T PN (G HK) K TRK x(N 1)

## The optimal value of u(N-2) may be obtained by

minimizing
1 1
JN1 J x(N 1) Qx (N 1) u(N 2)T Rx(N 2)
*
N
T

2 2
1 1
JN1 x(N 1)T PN1x(N 1) u(N 2)T Rx(N 2)
2 2
where PN1 (G HK)T PN (G HK) KTRK Q

## But JN-1 is of the same form as JN with the

indexes decremented by one.
Summary of Discrete-Time LQ
Solution
Control law:
u(k) Kk x(k), Kk R HTPk 1HBTPk 1G

Ricatti Equation
Pk G HKk Pk 1G HKk Q KkTRKk , k N 1,,1,0
T

Optimal Cost:

J x
1
x(0)T P0 x(0)
2
LQ Solution

## Control law is a time varying state

feedback law
Matrix Pk can be computed recursively
from the Ricatti equation by decrementing
the index k from N to 0.
In most cases P and K have steady-state
solutions as N approaches infinity
Matlab Example
y
Find the discrete-time u
(T=0.1) optimal controller M=1
that minimizes
99
J 5 y (100) y 2 (k ) u 2 (k )
2

k 0

## Solution: State Equation

x 1 position, x 2 velocity, u force
x 1 x 2 d x 1 0 1 x 1 0
u
x2 u dt x 2 0 0 x 2 1
Discretized Equation and PI
Weighting Matrices

Discretized Equation:

x1(k 1) 1 T x1(k ) T 2 2
x (k 1) 0 1 x (k ) u(k )
2 2 T

## Performance Index Weigthing Matrices:

1 T 10 0 1 90 T 2 0
J x (100) x (100) x ( k ) x ( k ) 2u 2
(k )
2 0 0 2 k 0 0 0

R
PN Q
System Definition in Matlab

## %System: dx1/dt=x2, dx2/dt=u

%System Matrices
Ac=[0 1;0 0]; Bc=[0;1];
[G,H]=c2d(Ac,Bc,0.1);
%Performance Index Matrices
N=100;
PN=[10 0;0 0]; Q=[1 0;0 0]; R=2;
Ricatti Equation Computation

## %Initialize gain K and S matrices

P=zeros(2,2,N+1); K=zeros(N,2);
P(:,:,N+1)=PN;

## %Computation of gain K and S matrices

for k=N:-1:1
Pkp1=P(:,:,k+1);
Kk=(R+H'*Pkp1*H)\(H'*Pkp1*G);
Gcl=G-H*Kk;
Pk=Gcl'*Pkp1*Gcl+Q+Kk'*R*Kk;
K(k,:)=Kk; P(:,:,k)=Pk;
end
LQ Controller Simulation

%Simulation
x=zeros(N+1,2); x0=[1;0];
x(1,:)=x0';

for k=1:N
xk=x(k,:)';
uk=-K(k,:)*xk;
xkp1=G*xk+H*uk;
x(k+1,:)=xkp1';
end

%plot results
Plot of P and K Matrices
25
P11
Elements of P 20 P12
P22
15

10

0
0 2 4 6 8 10

1.5
K1
K2
Elements of K

0.5

0
0 2 4 6 8 10
Time (sec)
Plot of State Variables
1
x1
x2
States 0.5

-0.5
0 2 4 6 8 10

0.2

-0.2
Input

-0.4

-0.6

-0.8
0 2 4 6 8 10
Time (sec)

## Given discrete-time state equation

x(k 1) Gx(k ) Hu(k )

## Find control sequence u(k) to minimize

1 T
J x (k )Qx (k ) uT (k )Ru(k )
2 k 0

S, Q 0, R 0 and symmetric

## Solution is obtained as the limiting

case of Ricatti Eq.
Summary of LQR Solution

Control law:
u(k) Kx(k),
K R HTPH HTPG
Ricatti Equation
P G HK PG HK Q K TRK
T

Optimal Cost:
1
J x(0)T Px(0)
2