1 optimisation and optimal control aim to provide an understanding of the principles of optimisation...

82
1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts.

Upload: myles-griffin

Post on 17-Dec-2015

212 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

1

OPTIMISATION AND OPTIMAL CONTROL

AIM

To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts.

Page 2: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

2

LEARNING OBJECTIVES

On completion of the module the student should be able to demonstrate:

 - an understanding of the basic principles of optimisation and the ability to apply them to linear and non-linear unconstrained and constrained static problems,

 - an understanding of the fundamentals of optimal control, system identification and parameter estimation.

Page 3: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

3

CONTENT(20 lectures common to BEng and MSc)

Principles of optimisation, essential features and examples of application.

(1 lecture)

Basic principles of static optimisation theory, unconstrained and constrained optimisation, Lagrange multipliers, necessary and sufficient conditions for optimality, limitations of analytical methods.

(3 lectures)

Numerical solution of static optimisation problems: unidimensional search methods; unconstrained multivariable optimisation, direct and indirect methods, gradient and Newton type approaches; non-linear programming with constraints.

(4 lectures)

Page 4: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

4

Introduction to genetic optimisation algorithms.

(1 lecture)

 Basic concepts of linear programming.

(1 lecture)

 On‑line optimisation, integrated system optimisation and parameter estimation.

(1 lecture)

Page 5: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

5

The optimal control problem for continuous dynamic systems: calculus of variations, the maximum principle, and two-point boundary value problems. The linear quadratic regulator problem and the matrix Riccati equation.

(4 lectures)

Introduction to system identification, parameter estimation and self-adaptive control

(3 lectures)

Introduction to the Kalman filter.

(2 lectures)

Page 6: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

6

(10 Lectures MSc Students only)

Page 7: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

7

LABORATORY WORK

The module will be illustrated by laboratory exercises and demonstrations on the use of MATLAB and the associated Optimization and Control Tool Boxes for solving unconstrained and constrained static optimisation problems and for solving linear quadratic regulator problems.

 

ASSESSMENT Via written examination.

 MSc only - also laboratory session and report

Page 8: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

8

READING LIST P.E Gill, W. Murray and M.H. Wright: Practical Optimization" (Academic Press, 1981) T.E. Edgar. D.M. Himmelblau and L.S. Lasdon: "Optimization of Chemical Processes“ 2nd Edition (McGraw-Hill, 2001) M.S. Bazaraa, H.D. Sherali and C.M. Shetty: “Nonlinear Programming - Theory and Algorithms” 2nd Edition (Wiley Interscience, 1993) J. Nocedal and S.J. Wright: “Numerical Optimization”, (Springer, 1999) J.E. Dennis, Jr. and R.B. Schnabel: “Numerical methods for Unconstrained Optimization and Nonlinear Equations” (SIAM Classics in Applied Mathematics, 1996) (Prentice Hall, 1983) K.J. Astrom and B Wittenmark: "Adaptive Control", 2nd Edition (Prentice Hall, 1993) F.L. Lewis and V S Syrmos: "Optimal Control" 2nd Edition (Wiley, 1995)

Page 9: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

9

PRINCIPLES OF OPTIMISATION Typical engineering problem: You have a process that can be represented by a mathematical model. You also have a performance criterion such as minimum cost. The goal of optimisation is to find the values of the variables in the process that yield the best value of the performance criterion.

  Two ingredients of an optimisation problem:

(i) process or model

(ii) performance criterion

Page 10: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

10

Some typical performance criteria:        maximum profit

        minimum cost

        minimum effort

        minimum error

        minimum waste

        maximum throughput

        best product quality

 Note the need to express the performance criterion in mathematical form.

Page 11: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

11

Static optimisation: variables have numerical values, fixed with respect to time.

Dynamic optimisation: variables are functions of time.

Page 12: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

12

Essential Features

Every optimisation problem contains three essential categories:

1.        At least one objective function to be optimised

2.        Equality constraints

3.        Inequality constraints

Page 13: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

13

By a feasible solution we mean a set of variables which satisfy categories 2 and 3. The region of feasible solutions is called the feasible region.

feas

ible

regi

on

linearinequalityconstraint

linearequalityconstraint

nonlinearinequalityconstraint

nonlinearinequalityconstraints

x2

x1linear inequality constraint

Page 14: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

14

An optimal solution is a set of values of the variables that are contained in the feasible region and also provide the best value of the objective function in category 1.

For a meaningful optimisation problem the model needs to be underdetermined.

Page 15: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

15

Mathematical Description

1 2

Minimize : ( ) objective function

( ) equality constraintsSubject to: ( ) inequality constraints

where , is a vector of n variables ( , , , )

( ) is a ve

nn

f

x x x

x

h x 0g x 0

x

h x

1

2

ctor of equalities of dimension

( ) is a vector of inequalities of dimension

m

mg x

Page 16: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

16

Steps Used To Solve Optimisation Problems

1. Analyse the process in order to make a list of all the variables.

2. Determine the optimisation criterion and specify the objective function.

3. Develop the mathematical model of the process to define the equality and inequality constraints. Identify the independent and dependent variables to obtain the number of degrees of freedom.

4. If the problem formulation is too large or complex simplify it if possible.

5. Apply a suitable optimisation technique.

6. Check the result and examine it’s sensitivity to changes in model parameters and assumptions.

Page 17: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

17

Classification of Optimisation Problems

Properties of f(x)

    single variable or multivariable

    linear or nonlinear

    sum of squares

    quadratic

    smooth or non-smooth

    sparsity

Page 18: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

18

Properties of h(x) and g(x)

        simple bounds

        smooth or non-smooth

        sparsity

        linear or nonlinear

        no constraints

Page 19: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

19

Properties of variables x

•time variant or invariant

•continuous or discrete

•take only integer values

•mixed

Page 20: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

20

Obstacles and Difficulties

• Objective function and/or the constraint functions may have finite discontinuities in the continuous parameter values.

• Objective function and/or the constraint functions may be non-linear functions of the variables.

• Objective function and/or the constraint functions may be defined in terms of complicated interactions of the variables. This may prevent calculation of unique values of the variables at the optimum.

Page 21: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

21

• Objective function and/or the constraint functions may exhibit nearly “flat” behaviour for some ranges of variables or exponential behaviour for other ranges. This causes the problem to be insensitive, or too sensitive.

• The problem may exhibit many local optima whereas the global optimum is sought. A solution may be obtained that is less satisfactory than another solution elsewhere.

• Absence of a feasible region.

• Model-reality differences.

Page 22: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

22

Typical Examples of Application static optimisation • Plant design (sizing and layout).

• Operation (best steady-state operating condition).

• Parameter estimation (model fitting).

• Allocation of resources.

• Choice of controller parameters (e.g. gains, time constants) to minimise a given performance index (e.g. overshoot, settling time, integral of error squared).

Page 23: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

23

dynamic optimisation

• Determination of a control signal u(t) to transfer a dynamic system from an initial state to a desired final state to satisfy a given performance index.

• Optimal plant start-up and/or shut down.

• Minimum time problems

Page 24: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

24

BASIC PRINCIPLES OF STATIC OPTIMISATION THEORY

Continuity of Functions

Functions containing discontinuities can cause difficulty in solving optimisation problems.

Definition: A function of a single variable x is continuous at a point xo if:

(a) exists

(b) exists

(c)

f x

f x

f x f x

o

x x

x xo

o

o

( )

lim ( )

lim ( ) ( )

Page 25: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

25

If f(x) is continuous at every point in a region R, then f(x) is said to be continuous throughout R.

x

f(x)

x

f(x)

f(x) is discontinuous.

f(x) is continuous, but( ) ( )dff x x

dx is not.

Page 26: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

26

Unimodal and Multimodal Functions

A unimodal function f(x) (in the range specified for x) has a single extremum (minimum or maximum).A multimodal function f(x) has two or more extrema.

There is a distinction between the global extremum (the biggest or smallest between a set of extrema) and local extrema (any extremum). Note: many numericalprocedures terminate at a local extremum.

If ( ) 0f x at the extremum, the point iscalled a stationary point.

Page 27: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

27

f(x)

x

local max (stationary)

local min (stationary)

global max (not stationary)

stationary point(saddle point)

global min (stationary)

A multimodal function

Page 28: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

28

Multivariate Functions - Surface and Contour Plots

We shall be concerned with basic properties of a scalar function f(x) of n variables (x1,...,xn).If n = 1, f(x) is a univariate functionIf n > 1, f(x) is a multivariate function.

n 1

For any multivariate function, the equationz = f(x) defines a surface in n+1 dimensional space .

Page 29: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

29

In the case n = 2, the points z = f(x1,x2) represent a three dimensional surface. Let c be a particular value of f(x1,x2). Then f(x1,x2) = c defines a curve in x1 and x2 on the plane z = c.

If we consider a selection of different values of c, we obtain a family of curves which provide a contour map of the function z = f(x1,x2).

Page 30: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

30

1 2 21 2 1 2 2(4 2 4 2 1)xz e x x x x x contour map of

-3 -2 -1 0 1 2-3

-2.5

-2

-1.5

-1

-0.5

0

0.5

1

1.5

2

x2

x1

0.20.4

0.71.0

1.7

1.8 2

3 4 56

z = 20

1.8

2

1.7

3

4

56

local minimum

saddle point

Page 31: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

31

Page 32: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

32

Page 33: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

33

Example: Surface and Contour Plots of “Peaks”

Function

2 2 21 1 2

3 5 2 21 1 2 1 2

2 21 2

3(1 ) exp ( 1)

10(0.2 )exp( )

1 3exp ( 1)

z x x x

x x x x x

x x

Page 34: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

34

0 5

10 15

20

0 5

10 15

20 -10

-5

0

5

10

x2

z

x1

multimodal!

10 20 30 40 50 60 70 80 90

10

20

30

40

50

60

70

80

90

100

x1

x2

global max

global min

local min

local max

local max

saddle

Page 35: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

35

Page 36: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

36

Gradient VectorThe slope of f(x) at a pointx x in the directionof the ith co-ordinate axis is

The n-vector of these partial derivatives is termed the gradient vector of f, denoted by:

L

N

MMMMM

O

Q

PPPPPf

f

x

f

xn

( )

( )

( )x

x

x

1

(a column vector)

f

xi

( )x

x x

Page 37: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

37

The gradient vector at a point x xis normal to the the contour through that point in the direction of increasing f.

x

f ( )xincreasing f

At a stationary point:

f ( )x 0 (a null vector)

Page 38: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

38

Example

f x x x x

f

f

xf

x

x x x

x x x

( ) cos

( )

( )

( )sin

cos

x

x

x

x

L

N

MMMM

O

Q

PPPP

LNM

OQP

1 22

2 1

1

2

22

2 1

1 2 12

and the stationary point (points) are given by the simultaneous solution(s) of:-

x x x

x x x22

2 1

1 2 1

0

2 0

sin

cos

Page 39: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

39

e.g.

f T( )x c x x c f( )=

Note: If f ( )x is a constant vector,

f(x) is then linear.

Page 40: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

40

Hessian Matrix (Curvature Matrix

The second derivative of a n - variable function is defined by the n2 partial derivatives:

x

f

xi n j n

i j

( ), , , ; , ,

xFHG

IKJ 1 1

written as:

2 2

2

f

x xi j

f

xi j

i j i

( ), ,

( ), .

x x

Page 41: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

41

These n2 second partial derivatives are usually represented by a square, symmetric matrix, termed the Hessian matrix, denoted by:

H x x

x x

x x

2

2 2

2 2( ) ( )

( ) ( )

( ) ( )

L

N

MMMMM

O

Q

PPPPPf

f

x

f

x x

f

x x

f

x

n

n n

12

1

12

Page 42: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

42

Example: For the previous example:

L

N

MMMM

O

Q

PPPP

LNM

OQP

2

2

12

2

1 22

1 2

2

22

2 1 2 1

2 1 1

2

2 2f

f

x

f

x xf

x x

f

x

x x x x

x x x( )

cos sin

sinx

Note: If the Hessian matrix of f(x) is a constant matrix, f(x) is then quadratic, expressed as:

f

f f

T T( )

( ) , ( )

x x Hx c x

x Hx c x H

12

2

Page 43: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

43

Convex and Concave Functions

A function is called concave over a given region R if:

f f f

Ra b a b

a b

( ( ) ) ( ) ( ) ( )

, , .

x x x x

x x

1 1

0 1where: and

The function is strictly concave if is replaced by >.

A function is called convex (strictly convex) if is replaced by (<).

Page 44: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

44

concave function

convex function

xxa xb

f(x)

f x( ) 0

xxa xb

f(x)

f x( ) 0

Page 45: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

45

I f then is concave.

I f then is convex.

f xf

xf x

f xf

xf x

( ) ( )

( ) ( )

2

2

2

2

0

0

For a multivariate function f(x) the conditions are:-

Strictly convex +ve defconvex +ve semi defconcave -ve semi defstrictly concave -ve def

f(x) H(x) Hessian matrix

Page 46: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

46

Tests for Convexity and Concavity

H is +ve def (+ve semi def) iff

x Hx x 0T 0 0 ( ), .

H is -ve def (-ve semi def) iff

x Hx x 0T 0 0 ( ), .

Convenient tests: H(x) is strictly convex (+ve def) (convex) (+ve semi def)) if:

1. all eigenvalues of H(x) are or 2. all principal determinants of H(x) are

0 0 ( )

0 0 ( )

Page 47: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

47

H(x) is strictly concave (-ve def)(concave (- ve semi def)) if:

1. all eigenvalues of H(x) are or 2. the principal determinants of H(x) are alternating in sign:

0 0 ( )

1 2 3

1 2 3

0 0 0

0 0 0

, , ,

( , , , )

Page 48: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

48

Example f x x x x x( ) 2 3 212

1 2 22

2 2

1 2 21 1 21

2

1 2 22 2

( ) ( ) ( )4 3 4 3

( ) ( )3 4 4

f f fx x

x x xx

f fx x

x x

x x x

x x

1 2

4 3 4 3( ) , 4, 7

3 4 3 4

H x

22

1 2

4 3eigenvalues: | | 8 7 0

3 4

1 Hence, ( ) is strictly conv, 7. e .xf

I H

x

Page 49: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

49

Convex Regionxa

xb

non convexregion

A convex set of points exist if for any two points, xa

and xb, in a region, all points:x x x a b( ) ,1 0 1

on the straight line joining xa and xb are in the set.If a region is completely bounded by concave functions then the functions form a convex region.

xa

xb

convexregion

Page 50: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

50

Necessary and Sufficient Conditions for an Extremum of

an Unconstrained FunctionA condition N is necessary for a result R if R can be true only if N is true.

R N

A condition S is sufficient for a result R if R istrue if S is true.

S R

A condition T is necessary and sufficient for a result R iff T is true.

T R

Page 51: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

51

There are two necessary and a single sufficientconditions to guarantee that x* is an extremum of afunction f(x) at x = x*:1. f(x) is twice continuously differentiable at x*.

2. , i.e. a stationary point exists at x*.

3. is +ve def for a minimum to exist at x*, or -ve def for a maximum to exist at x*

f ( )*x 0

2 f ( ) ( )* *x H x

1 and 2 are necessary conditions; 3 is a sufficient condition.Note: an extremum may exist at x* even though it is not possible to demonstrate the fact using the three conditions.

Page 52: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

52

Example: Consider:

The gradient vector is:

LNM

OQPf

x x x x x

x x x( )

.x

4 5 2 2 4 4

4 4 2 21 2 1

31 2

2 1 12

yielding three stationary points located by setting and solving numerically: f ( )x 0

x*=(x1,x2) f(x*) eigenvaluesof

2 f ( )x

classification

A.(-1.05,1.03) -0.51 10.5 3.5 global min

B.(1.94,3.85) 0.98 37.0 0.97 local min

C.(0.61,1.49) 2.83 7.0 -2.56 saddle

where:

LNM

OQP

2 12

2 1

1

2 12 4 2 4

2 4 4f

x x x

x( )x

f x x x x x x x x x( ) .x 4 4 5 4 2 2 21 2 12

22

1 2 14

12

2

Page 53: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

53

contour map

-4 -3 -2 -1 0 1 2 3 40

1

2

3

4

5

6

x2

x1

Page 54: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

54

Interpretation of the Objective Function in Terms of its Quadratic

ApproximationIf a function of two variables can be approximated within a region of a stationary point by a quadratic function:

f x x x xh h

h h

x

xc c

x

x

h x h x h x x c x c x

( , )1 212 1 2

11 12

12 22

1

21 2

1

2

12 11 1

2 12 22 2

212 1 2 1 1 2 2

LNM

OQPLNM

OQP

LNM

OQP

then the eigenvalues and eigenvectors of:

H( , ) ( , )* * * *x x f x xh h

h h1 22

1 211 12

12 22

LNM

OQP

can be used to interpret the nature of f(x1,x2) at:

x x x x1 1 2 2 * *,

Page 55: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

55

They provide information on the shape of

f(x1,x2) at

x x x x1 1 2 2 * *, If H( , )* *x x1 2 is +ve def, theeigenvectors are at right angles (orthogonal) andcorrespond to the principal axes of elliptical contours of f(x1,x2). A valley or ridge lies in the direction of the eigenvector associated with a relative small eigenvalue.

These interpretations can be generalized to the multivariate quadratic approximation:

f T T( )x x Hx c x 12

Page 56: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

56

Case 1: Equal Eigenvalues - circular contours which are interpreted as a circular hill (max)(-ve eigenvalues) or circular valley (min) (+ve eigenvalues)

-2 -1.5 -1 -0.5 0 0.5 1 1.5 2 -2

-1.5

-1

-0.5

0

0.5

1

1.5

2

x2

x1

Page 57: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

57

Page 58: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

58

Case 2: Unequal Eigenvalues (same sign) – elliptical contours which are interpreted as an elliptical hill (max)(-ve eigenvalues) or elliptical valley (min)(+ve eigenvalues)

-2 -1.5 -1 -0.5 0 0.5 1 1.5 2 -2

-1.5

-1

-0.5

0

0.5

1

1.5

2

x1

X2

Page 59: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

59

Page 60: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

60

Case 3: Eigenvalues of opposite sign but equal in magnitude - symmetrical saddle.

-2 -1.5 -1 -0.5 0 0.5 1 1.5 2 -2

-1.5

-1

-0.5

0

0.5

1

1.5

2

x1

x2

Page 61: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

61

Page 62: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

62

Case 4: Eigenvalues of opposite sign but unequal in magnitude - asymmetrical saddle.

-2 -1.5 -1 -0.5 0 0.5 1 1.5 2 -2

-1.5

-1

-0.5

0

0.5

1

1.5

2

x1

x2

Page 63: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

63

Page 64: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

64

Optimisation with Equality Constraints

min ( );

( ) ;x

x x

h x 0

nf

subject to: m constraints (m n)

Elimination of variables:

example: min ( ),x xf x x

x x1 2

4 5

2 3 6

12

22

1 2

x

(a)

s.t. (b)

and substituting into (a) :- f x x x( ) ( )2 22

226 3 5

Using (b) to eliminate x1 gives:xx

126 3

2

(c)

Page 65: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

65

At a stationary point

f x

xx x

x x

( )( )

.*

2

22 2

2 2

0 6 6 3 10 0

28 36 1286

Then using (c): xx

126 3

21071*

*

.

Hence, the stationary point (min) is: (1.071, 1.286)

Page 66: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

66

The Lagrange Multiplier MethodConsider a two variable problem with a single equality constraint:

min ( , )

( , )

,x xf x x

h x x1 2

1 2

1 2 0s.t.

At a stationary point we may write:

(a)

(b)

dff

xdx

f

xdx

dhh

xdx

h

xdx

f

x

f

xh

x

h

x

dx

dx

L

N

MMMM

O

Q

PPPPLNM

OQP

LNMOQP

11

22

11

22

1 2

1 2

1

2

0

0

0

0

Page 67: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

67

If:

f

x

f

xh

x

h

x

1 2

1 2

0 nontrivial nonunique solutionsfor dx1 and dx2 will exist.

This is achieved by setting

h

x

h

xh

x

h

x

fx

hx

fx

hx

1 2

1 2

1 1 2 2

with

where is known as a Lagrange multiplier.

Page 68: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

68

If an augmented objective function, called theLagrangian is defined as:

L x x f x x h x x( , , ) ( , ) ( , )1 2 1 2 1 2

we can solve the constrained optimisation problem bysolving:

L

x

f

x

h

xL

x

f

x

h

x

Lh x x

1 1 1

2 2 2

1 2

0

0

0

UV||W||

provides equations (a) and (b)

re-statement of equality constraint( , )

Page 69: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

69

Generalizing : To solve the problem:

min ( );

subject to: ( ) ; constraints ( )

f

m m n

n

xx x

h x 0

define the Lagrangian:

( , ) ( ) ( ), T mL f x x h x

and the stationary point (points) is obtained from:-

( )( , ) ( )

( , ) ( )

T

L f

L

x x

h xx x 0

x

x h x 0

Page 70: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

70

Example Consider the previous example

again. The Lagrangian is:-

L x x x x

L

xx

L

xx

Lx x

4 5 2 3 6

8 2 0

10 3 0

2 3 6 0

12

22

1 2

11

22

1 2

( )

(a)

(b)

(c)

Substituting (a) and (b) into (c) gives:

Page 71: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

71

which agrees with the previous result.

x x

x x

1 2

1 2

4

3

10 2

9

106 0

30

74 281

15

141071

90

701286

, .

. , .

Hence,

-4 -3 -2 -1 0 1 2 3 4 -4

-3

-2

-1

0

1

2

3

4

x1

x2

Page 72: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

72

Necessary Conditions for a Local Extremum of

an Optimisation Problem Subject to Equality and Inequality Constraints(Kuhn-Tucker Conditions)

Consider the problem:min ( );

s.t.: ( ) ; equalities ( )

( ) ; inequalities

f

m m n

p

n

xx x

h x 0

g x 0

Here, we define the Lagrangian:

( , , ) ( ) ; , T T m pL f x x h(x) g(x)

The necessary conditions for x* to be a local extremum of f(x) are:-

Page 73: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

73

(a) f(x), hj(x), gj(x) are all twice differentiable at x*

(b) The Lagrange multipliers exist

(c) All constraints are satisfied at x*h x 0 x g x 0 x( ) ( ) ; ( *) ( )* * * h gj j0 0

(d) The Lagrange multipliers (at x*) for the inequality constraints are not negative, i.e.

j*

j* 0

(e) The binding (active) inequality constraints are zero, the inactive inequality constraints are > 0, and the associated j’s are 0 at x*, i.e.

j jg* *( )x 0

(f) The Lagrangian function is at a stationary point( , , ) 0L x x

Page 74: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

74

Notes:

1. Further analysis or investigation is required to determine if the

extremum is a minimum (or maximum)

2. If f(x) is convex, h(x) are linear and g(x)

are concave: x* will be a global extremum.

Page 75: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

75

Limitations of Analytical Methods

The computations needed to evaluate the above conditions can be extensive and intractable. Furthermore, the resulting simultaneous equations required for solving x*, * and * are often nonlinear and cannot be solved without resorting to numerical methods.

The results may be inconclusive.

For these reasons, we often have to resort to numerical methods for solving optimisation problems, using computer codes (e.g.. MATLAB)

Page 76: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

76

ExampleDetermine if the potential minimumx* = (1.00,4.90) satisfies the Kuhn Tucker conditions for the problem:

min ( )

( )

( )

( ) ( ) ( )

( )

( )

xx

x

x

x

x

x

f x x

h x x

g x x x x

g x x

g x

g x

4 12

25 0

10 10 34 0

3 1 0

2

0

1 22

1 12

22

1 1 12

2 22

2 12

22

3 1

4 2

s.t.

Page 77: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

77

-2 0 2 4 6 8-2

-1

0

1

2

3

4

5

6

7

8

x1

x2

(1.00,4.90))

x2=0

h1(x)=0

g1(x)=0

105

0-5

-10

-20-25

-30-40-50-60

contours and constraints

Page 78: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

78

We test each Kuhn-Tucker condition in turn:(a) All functions are seen by inspection to be twice differentiable.

(b) We assume the Lagrange multipliers exist

(c) Are the constraints satisfied?

h yes

g yes binding

g yes not active

g yes not active

g yes not active

12 2

12 2

22

3

4

25 100 4 90 0 01 0

10 100 100 10 4 90 4 90 34 0 01 0

4 90 1 19 21 0

1

4 90 0

: ( . ) ( . ) .

: ( . ) ( . ) ( . ) ( . ) .

: ( ( . ) .

:

: .

1.00-3)

.00 -2

2

Page 79: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

79

To test the rest of the conditions we need to determine the Lagrange multipliers using the stationarity conditions. First we note that from condition (e) we require:

j jg j* *( ) , , , ,x 0 1 2 3 4

1 1

2 2

3 3

4 4

0

0

0

0

* *

* *

* *

* *

( )

( )

( )

( )

can have any value because

must be zero because

must be zero because

must be zero because

g

g

g

g

x

x

x

x

Page 80: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

80

Hence:

LNM

OQPLNM

OQP

LNM

OQP

x

x

L x x

L x x x

1

2

4 2 10 2 4 2 8

2 2 10 2 9 8 9 8 0 2

2 8

9 8 0 2

4

9 81015

1 1 1 1 1 1

2 1 2 1 2 1 1

1

1

1 1

* * * * * *

* * * * * * *

*

*

* *

( )

( ) . . .

. . .. , =0.754

Now consider the stationarity condition:* *2 3

2 2 2 2 21 2 1 1 2 1 1 1 2 2

( , , ) 0 where, since 0

=4 12 (25 ) (10 10 34)

L

L x x x x x x x x

x x

Page 81: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

81

Now we can check the remaining conditions:

(d) Are j j* , , , , ? 0 1 2 3 4

1 2 3 40 754 0* * * *. ,

Hence, the answer is yes

(e) Are j jg j* *( ) , , , , ?x 0 1 2 3 4

Yes, because we have already used this above.

(f) Is the Lagrangian function at a stationary point?

Yes, because we have already used this above.

Page 82: 1 OPTIMISATION AND OPTIMAL CONTROL AIM To provide an understanding of the principles of optimisation techniques in the static and dynamic contexts

82

Hence, all the Kuhn-Tucker conditions aresatisfied and we can have confidence in thesolution :

x* = (1.00,4.90)