Professional Documents
Culture Documents
OPTIMAL CONTROL
AIM
To provide an understanding of
the principles of optimisation
techniques in the static and
dynamic contexts.
1
LEARNING OBJECTIVES
On completion of the module the student
should be able to demonstrate:
- an understanding of the basic
principles of optimisation and the ability to
apply them to linear and non-linear
unconstrained and constrained static
problems,
- an understanding of the fundamentals
of optimal control, system identification
and parameter estimation. 2
CONTENT
(20 lectures common to BEng and MSc)
Principles of optimisation, essential features and
examples of application.
(1 lecture)
Basic principles of static optimisation theory,
unconstrained and constrained optimisation, Lagrange
multipliers, necessary and sufficient conditions for
optimality, limitations of analytical methods.
(3 lectures)
Numerical solution of static optimisation problems:
unidimensional search methods; unconstrained
multivariable optimisation, direct and indirect methods,
gradient and Newton type approaches; non-linear
programming with constraints.
(4 lectures)
3
Introduction to genetic optimisation
algorithms.
(1 lecture)
Basic concepts of linear programming.
(1 lecture)
On-line optimisation, integrated system
optimisation and parameter estimation.
(1 lecture)
4
The optimal control problem for continuous
dynamic systems: calculus of variations, the
maximum principle, and two-point boundary
value problems. The linear quadratic regulator
problem and the matrix Riccati equation.
(4 lectures)
Introduction to system identification, parameter
estimation and self-adaptive control
(3 lectures)
Introduction to the Kalman filter.
(2 lectures)
5
(10 Lectures MSc Students only)
6
LABORATORY WORK
The module will be illustrated by laboratory
exercises and demonstrations on the use of
MATLAB and the associated Optimization and
Control Tool Boxes for solving unconstrained and
constrained static optimisation problems and for
solving linear quadratic regulator problems.
ASSESSMENT
Via written examination.
MSc only - also laboratory session and report
7
READING LIST
P.E Gill, W. Murray and M.H. Wright: Practical
Optimization" (Academic Press, 1981)
T.E. Edgar. D.M. Himmelblau and L.S. Lasdon:
"Optimization of Chemical Processes 2nd Edition (McGraw-
Hill, 2001)
M.S. Bazaraa, H.D. Sherali and C.M. Shetty: Nonlinear
Programming - Theory and Algorithms 2nd Edition (Wiley
Interscience, 1993)
J. Nocedal and S.J. Wright: Numerical Optimization,
(Springer, 1999)
J.E. Dennis, Jr. and R.B. Schnabel: Numerical methods
for Unconstrained Optimization and Nonlinear Equations
(SIAM Classics in Applied Mathematics, 1996) (Prentice
Hall, 1983)
K.J. Astrom and B Wittenmark: "Adaptive Control", 2nd
Edition (Prentice Hall, 1993)
F.L. Lewis and V S Syrmos: "Optimal Control" 82nd
Edition (Wiley, 1995)
PRINCIPLES OF OPTIMISATION
Typical engineering problem: You have a
process that can be represented by a
mathematical model. You also have a
performance criterion such as minimum cost.
The goal of optimisation is to find the values of
the variables in the process that yield the best
value of the performance criterion.
Two ingredients of an optimisation problem:
(i) process or model
(ii) performance criterion 9
Some typical performance
criteria:
maximum profit
minimum cost
minimum effort
minimum error
minimum waste
maximum throughput
best product quality
Note the need to express the performance
criterion in mathematical form.
10
Static optimisation: variables
have numerical values, fixed
with respect to time.
Dynamic optimisation:
variables are functions of
time.
11
Essential Features
Every optimisation problem contains
three essential categories:
1. At least one objective function
to be optimised
2. Equality constraints
3. Inequality constraints
12
By a feasible solution we mean a set of variables which
satisfy categories 2 and 3. The region of feasible solutions
is called the feasible region.
x2 nonlinear linear
inequality equality
constraints constraint
linear
inequality
constraint
nonlinear
inequality
constraint
x1
linear inequality constraint
13
An optimal solution is a set of values
of the variables that are contained in
the feasible region and also provide
the best value of the objective
function in category 1.
14
Mathematical Description
Minimize : f (x) objective function
h(x) 0 equality constraints
Subject to:
g(x) 0 inequality constraints
where x n , is a vector of n variables (x1, x2 , , xn )
h(x) is a vector of equalities of dimension m1
g(x) is a vector of inequalities of dimension m2
15
Steps Used To Solve Optimisation Problems
1. Analyse the process in order to make a list of all the
variables.
2. Determine the optimisation criterion and specify
the objective function.
3. Develop the mathematical model of the process to
define the equality and inequality constraints.
Identify the independent and dependent variables
to obtain the number of degrees of freedom.
4. If the problem formulation is too large or complex
simplify it if possible.
5. Apply a suitable optimisation technique.
6. Check the result and examine its sensitivity to
changes in model parameters and assumptions. 16
Classification of
Optimisation Problems
Properties of f(x)
single variable or multivariable
linear or nonlinear
sum of squares
quadratic
smooth or non-smooth
sparsity
17
Properties of h(x) and g(x)
simple bounds
smooth or non-smooth
sparsity
linear or nonlinear
no constraints
18
Properties of variables x
time variant or invariant
continuous or discrete
take only integer values
mixed
19
Obstacles and Difficulties
Objective function and/or the constraint
functions may have finite discontinuities in
the continuous parameter values.
Objective function and/or the constraint
functions may be non-linear functions of the
variables.
Objective function and/or the constraint
functions may be defined in terms of
complicated interactions of the variables. This
may prevent calculation of unique values of
the variables at the optimum. 20
Objective function and/or the constraint
functions may exhibit nearly flat behaviour
for some ranges of variables or exponential
behaviour for other ranges. This causes the
problem to be insensitive, or too sensitive.
The problem may exhibit many local optima
whereas the global optimum is sought. A
solution may be obtained that is less
satisfactory than another solution elsewhere.
Absence of a feasible region.
Model-reality differences. 21
Typical Examples of Application
static optimisation
Plant design (sizing and layout).
Operation (best steady-state operating
condition).
Parameter estimation (model fitting).
Allocation of resources.
Choice of controller parameters (e.g.
gains, time constants) to minimise a given
performance index (e.g. overshoot, settling
time, integral of error squared). 22
dynamic optimisation
(c) lim f ( x ) f ( xo )
x xo 24
If f(x) is continuous at every point in a
region R, then f(x) is said to be continuous
throughout R.
f(x) is discontinuous.
f(x)
27
Multivariate Functions -
Surface and Contour Plots
We shall be concerned with basic properties of
a scalar function f(x) of n variables (x1,...,xn).
29
contour map of z e x1 (4 x12 2 x22 4 x1 x2 2 x2 1)
2
3 4 5
6
1.5 1.7 1.8 2 z = 20
1
1.8
0.5 1.7
2
x2 0
saddle point 1.0
-0.5 0.7
0.4
3 0.2
-1
-1.5 4
-2 5
6
-2.5
local minimum
-3
-3 -2 -1
x1 0 1 2
30
31
32
Example: Surface and
Contour Plots of Peaks
Function
z 3(1 x1 ) exp x ( x2 1)
2 2
1
2
10(0.2 x1 x x )exp( x x )
3
1
5
2
2
1
2
2
1 3exp ( x1 1) x
2 2
2
33
10
5 multimodal!
0
z
-5
-10
20
15 20
10 15
10
x2 5 5
0 0 x1
100
90
global max
80
70
local min saddle
x2 60
50
40
30
local max
20
At a stationary point:
f (x) 0 (a null vector)
37
Example
f ( x) x1 x22 x2 cos x1
L
f ( x ) O
M x P Lx x sin x O
2
f ( x ) M P 1
M 2
P
2 1
M f ( x ) P N
2 x x cos x Q
M
Nx P 2 Q
1 2 1
2 x1 x2 cos x1 0
38
Note: If f ( x) is a constant vector,
f(x) is then linear.
e.g.
f (x) c x f( x ) = c
T
39
Hessian Matrix (Curvature Matrix
The second derivative of a n - variable
function is defined by the n2 partial
derivatives:
F I
f (x)
G J
H K
xi x j
, i 1,, n; j 1, , n
written as:
f ( x)
2
f ( x)
2
, i j, , i j.
xi x j xi2
40
These n2 second partial derivatives are
usually represented by a square, symmetric
matrix, termed the Hessian matrix, denoted
by:
L
f ( x)
M
2
2 f ( x) O
P
M x 2
1 x1xn P
H( x) f ( x) M
2
P
M
f ( x)
M
2
f ( x) P
2
x P
Nx x
1 n
2
n Q
41
Example: For the previous example:
LM f
2
f
2
O
P
f ( x) M
2 x 2
1 x1x2P
L
M
x2 cos x1 2 x2 sin x1 O
P
M
M
f
2
f
2
P
P N2 x2 sin x1 2 x1 Q
Nx x
1 2 x22 Q
Note: If the Hessian matrix of f(x) is a
constant matrix, f(x) is then quadratic,
expressed as:
f (x) 2 x Hx c x
1 T T
f (x) Hx c, 2 f (x) H
42
Convex and Concave Functions
A function is called concave over a given
region R if:
f ( x ) 0
x
xa xb
f ( x ) 0
xb x44
xa
2 f
If f ( x ) 0 then f ( x ) is concave.
x 2
2 f
If f ( x ) 0 then f ( x ) is convex.
x 2
47
Example f ( x) 2 x 3x1 x2 2 x
2
1
2
2
f ( x) 2 f ( x) 2 f ( x)
4 x1 3 x2 4 3
x1 x12
x1 x2
f ( x) 2 f ( x)
3x1 4 x2 4
x2 x22
4 3 4 3
H ( x) , 1 4, 2 7
3 4 3 4
4 3
eigenvalues: | I 2 H | 2 8 7 0
3 4
1 1, 2 7. Hence, f (x) is strictly convex.
48
Convex Region
xa
xb
convex
xa region non convex
region xb
N 4 4 x 22 x 21x Q 2
1
yielding three stationary points located by setting
f (x) 0 and solving numerically:
x*=(x1,x2) f(x*) eigenvalues classification
2 f (x)
of
A.(-1.05,1.03) -0.51 10.5 3.5 global min
B.(1.94,3.85) 0.98 37.0 0.97 local min
C.(0.61,1.49) 2.83 7.0 -2.56 saddle
where:
L
2 12 x 4 x
f ( x) M
2
2
1 2 2 4 x1 O
P
N 2 4 x 1 4 Q 52
contour map
6
3
x2
2
0
-4 -3 -2 -1 0 1 2 3 4
x1
53
Interpretation of the Objective Function
in Terms of its Quadratic Approximation
If a function of two variables can be approximated
within a region of a stationary point by a quadratic
function:
f ( x1 , x2 ) 2 x1 x2
1
L O
L
P
M O
h11 h12 x1
M P c1 c2
x1
L
M O
P
N Q
NQ
h12 h22 x2 x2 NQ
h x h x h12 x1 x2 c1 x1 c2 x2
1 2
2 11 1
1 2
2 22 2
f (x) x Hx c x
1
2
T T
55
Case 1: Equal Eigenvalues - circular contours
which are interpreted as a circular hill (max)(-ve
eigenvalues) or circular valley (min) (+ve
eigenvalues)
2
1.5
0.5
x2
0
-0.5
-1
-1.5
-2
-2 -1.5 -1 -0.5 0 x1 0.5 1 1.5 2 56
57
Case 2: Unequal Eigenvalues (same sign)
elliptical contours which are interpreted as
an elliptical hill (max)(-ve eigenvalues) or
elliptical valley (min)(+ve eigenvalues)
2
1.5
0.5
0
X2
-0.5
-1
-1.5
-2
-2 -1.5 -1 -0.5 0 0.5 1 1.5 2 58
x1
59
Case 3: Eigenvalues of opposite sign but
equal in magnitude - symmetrical saddle.
2
1.5
0.5
0
x2
-0.5
-1
-1.5
-2
-2 -1.5 -1 -0.5 0 0.5 1 1.5 2 60
x1
61
Case 4: Eigenvalues of opposite sign but
unequal in magnitude - asymmetrical saddle.
2
1.5
0.5
0
x2
-0.5
-1
-1.5
-2
-2 -1.5 -1 -0.5 0 0.5 1 1.5 62 2
x1
63
Optimisation with Equality
Constraints
min f (x); x n
x
subject to: h(x) 0; m constraints (m n)
Elimination of variables:
f ( x2 )
0 6(6 3x2 ) 10 x2 0
x2
28 x2 36 x 1286
*
2.
6 3x *
Then using (c): x
*
1 1071
2
.
2
s. t. h( x1 , x2 ) 0
h h
x1 x2 f h f h
with
h h x1 x1 x2 x2
x1 x2
where is known as a Lagrange multiplier.
67
If an augmented objective function, called the
Lagrangian is defined as:
L( x1 , x2 , ) f ( x1 , x2 ) h( x1 , x2 )
we can solve the constrained optimisation problem by
solving:
L f
h
0
U
|
x1 x1 x1
L f h V
provides equations (a) and (b)
||
0
x2 x2 x2 W
L
h( x1 , x2 ) 0 re-statement of equality constraint
68
Generalizing : To solve the problem:
min f (x); x n
x
subject to: h(x) 0; m constraints (m n)
x L(x, ) x f (x) 0
x
L(x, ) h(x) 0
69
Example Consider the previous example
again. The Lagrangian is:-
L 4 x 5x (2 x1 3x2 6)
2
1
2
2
L
8 x1 2 0 (a)
x1
L
10 x2 3 0 (b)
x2
L
2 x1 3x2 6 0 (c)
Substituting (a) and (b) into (c) gives: 70
3 9 30
x1 , x2 6 0 4.281
4 10 2 10 7
15 90
Hence, x1 1071
. , x2 1286
.
14 70
which agrees with the previous result.
4
0
x2
-1
-2
-3
-4 71
-4 -3 -2 -1 0 1 2 3 4
x1
Necessary Conditions for a Local Extremum of
an Optimisation Problem Subject to Equality
and Inequality Constraints
(Kuhn-Tucker Conditions)
74
Limitations of Analytical Methods
The computations needed to evaluate the above
conditions can be extensive and intractable.
Furthermore, the resulting simultaneous
equations required for solving x*, * and * are
often nonlinear and cannot be solved without
resorting to numerical methods.
g2 (x) ( x1 3) 2 ( x2 1) 2 0
g3 (x) x1 2
g 4 ( x ) x2 0 76
contours and constraints
8
7 -60
-50 -40 -30
6
(1.00,4.90) -25 -20
5 )
-10
4 h1(x)=0
-5
0
3 g1(x)=0
x2 5
2 10
1
x2=0
0
-1
-2
-2 0 2 4 6 8
x1 77
We test each Kuhn-Tucker condition in turn:
(a) All functions are seen by inspection to be twice
differentiable.
(b) We assume the Lagrange multipliers exist
(c) Are the constraints satisfied?
h1: 25 (100
. ) (4.90) 0.01 0 yes
2 2
. ) (100
g1: 10(100 . ) 10(4.90) (4.90) 34 0.01 0 yes binding
2 2
2
g2 : (1.00-3) (4.90 1) 19.21 0 yes not active
2
78
To test the rest of the conditions we need to
determine the Lagrange multipliers using the
stationarity conditions. First we note that from
condition (e) we require:
g j (x ) 0, j 1,2,3,4
*
j
*
79
Now consider the stationarity condition:
x L(x, , ) 0 where, since 2* 3* 0
L=4 x1 x22 12 1 (25 x12 x22 ) 1 (10 x1 x12 10 x2 x22 34)
Hence:
x1 L 4 2*1 x1* 1* (10 2 x1* ) 4 2*1 8 1*
x2 L 2 x2* 2*1 x2* 1* (10 2 x2* ) 9.8 9.8*1 0.2 1*
L
M
2 8O L
M OL4 O
P
*
P
N9.8 0.2Q M P 1015
N QN9.8Q
1
*
1
. , *
1
*
1 = 0.754
80
Now we can check the remaining conditions:
(d) Are *j 0, j 1,2,3,4 ?
1* 0.754, *2 *3 *4 0
Hence, the answer is yes
(e) Are *j g j (x* ) 0, j 1,2,3,4 ?
Yes, because we have already used this above.
81
Hence, all the Kuhn-Tucker conditions are
satisfied and we can have confidence in the
solution :
x* = (1.00,4.90)
82