stochastic control theory and some of its industrial...

12
Stochastic Control Theory and Some of Its Industrial Applications Åström, Karl Johan 1975 Document Version: Publisher's PDF, also known as Version of record Link to publication Citation for published version (APA): Åström, K. J. (1975). Stochastic Control Theory and Some of Its Industrial Applications. (Technical Reports TFRT-7074). Department of Automatic Control, Lund Institute of Technology (LTH). General rights Unless other specific re-use rights are stated the following general rights apply: Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal Read more about Creative commons licenses: https://creativecommons.org/licenses/ Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim.

Upload: others

Post on 05-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

LUND UNIVERSITY

PO Box 117221 00 Lund+46 46-222 00 00

Stochastic Control Theory and Some of Its Industrial Applications

Åström, Karl Johan

1975

Document Version:Publisher's PDF, also known as Version of record

Link to publication

Citation for published version (APA):Åström, K. J. (1975). Stochastic Control Theory and Some of Its Industrial Applications. (Technical ReportsTFRT-7074). Department of Automatic Control, Lund Institute of Technology (LTH).

General rightsUnless other specific re-use rights are stated the following general rights apply:Copyright and moral rights for the publications made accessible in the public portal are retained by the authorsand/or other copyright owners and it is a condition of accessing publications that users recognise and abide by thelegal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private studyor research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal

Read more about Creative commons licenses: https://creativecommons.org/licenses/Take down policyIf you believe that this document breaches copyright please contact us providing details, and we will removeaccess to the work immediately and investigate your claim.

Page 2: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

STOCHASTIC CONTROL THEORY AND SOME OFIT5 INDUSTRIAL APPLICATIONS

oK. J. ASTROM

Report 7501 (C) Jonuori 1975Lund lnstitúte õf TêiFinologyDeportment of Automotic Control

Page 3: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

STOCHASTTC CONTROL THEORY AND SOME OF TTS TNDUSÎRTAL APPLT-CATIONS

by K. J. .Â,ström¡I

Contents

Invited survey paper at the IFAC symposium on StochasticControl, Budapest, September 25-27, 1974. Thís work hasbeen partially supported by the Swedj-sh Board for Tech-nical Development under Contract 733546.

Page

1. Tntroduction

Linear Theory

Minimum Variance ControlReview of Assumpti_ons

3. Nonl j-near Theory

4. Linear systems with constant but unknohrn parameters 4

One-Step and Certainty Equivalence ControlDual Control

5. Se1f-Tuning Regulators

6. Conclusions

7. References

1

2 2

2

3

5

6

4

6

7

I

t

Page 4: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

I

STOCHASTIC CONTROL THEORY AND SOME OF ITS TNDUSTRIAL APPLTCATIONS

K. J. Âström

Department of Automatic Control, Lund Institute of Technology

5-220 07 Lund, Sweden

ABSTRACT

The purpose of this paper is to give a surveyof some results from stochastic control theory andto discuss their application t.o industrial controlproblems. The intention ís to state the results,discuss them from the point of view of applica-tions and to give some experiences from practicaluse. The review is not comprehensive.

1. INTRODUCTION

The paper is limi.ted to discrete lime syslems. Themotivation is that digital computers are used Eo

implement the conËro1 algorithms. Mathematicallythe theory becomes much simpler. Some formulaswill, however, formally be more complicated thanin the continuous time case. Several importantproblems, like the selection of the sampling interval, are also neglected by this assumption.

The paper is organized as follows. Control oflinear sysEems with a quadraLic criterion is dis-cussed in Section 2. The standard problem is for-rnulated and the solution staEed. An importantspecial case of the linear theory, which leads tothe so-cal1ed minimum variance strategy, is alsodiscussed. The linear theory resulLs in a regula-tor ¡vhich is a linear, dynamical system. This isa convenient regulator in several practical prob-lems. The major difficulty in applying thel.inear theory is to obtain suitable models for theprocess dynamics and the disturbances, Such modelscan be obtained from plant experiments using sys-tem identífication techniques. Some applícationsof the linear theory are also given in Section 3.

A review of nonlinear theory is given in Section 4.By assuming that the process and its environmentcan be characterized as Markov processes, the opti-mal strategies are given by the Bellman equationobtained by Dynamic Progranuning. The equaËion isunfortunately unpleasant both analytically andnumerically. The maín obstacle is dimensionality.To carry out the analysis it is necessary to intro-duce a hyperstate, which is a probability distri-bution on the original state space of the problem.For this reason ít has not yet been possible tosolve any realistic problens usíng Lhese methods.A few problems have, however, been solved numeri-

discussed in Section 4. In thís case the nonlineartheory gives dual control strategies ín the senseof Feldbaum. The ínsight obtained from the nonlin-ear theory can be exploited to construct suboptimalstrategies r¡hich can be implemented. One class ofsuboptimal regulators, the so-called self-tuningregulators, are discussed in Section 5. These regu-larors can be used to control a linear system withconstant but unknown parameters and will thus makeit possible to compensate for lack of proeess know-ledge by rnore sophisticated control algorithms. Asindicated by the nonlinear theory, the regulatorswill have some limitations. One drawback is thatthey are not dual controls. This is discussed inconnection with some applications of the regulatorsin Section 5. Some concluding remarks on the appli-cability of stochastic control theory are finallygiven in Section 6.

2. LINEAR THEORY

The well-known linear stochastic control theorydeals with control of a system which can be de-scribed by the stochastic difference equation

x(t+l) =Ax(t) +Bu(t) +v(t)(2.L)

y(t)=Cx(t)+e(t)

where t€T = {...-1, 0, 1,...}, x(t) is a nxl statevecEor, u(t) a pxl vector of control variables andy(t) a rxl vector of measured outputs. The stochas-Eic processes {v(t), t€Ti and {e(t), t€T}, calledprocess disturbances and measurement errors respec-tively, are sequences of uncorrelated, random vari-ables with zero mean values and the covariances

cov[v(t), v(t)] = Rf

cov[v(t), e(t)] = RfZ

covIe(t), e(t)] = RZ

The initial state of (2.1) is assumed to be a ran-dom variable with mean m and covariance Ro.

The performance of the control system is describedby the scalar loss function

N-1T

t=ll = fo*rtnloox(N) + ¡*r1t¡qr*1r¡ + ,rrlt¡qrr,1t¡ J

(2.2)cal ly .

One particular problem, namely control of a linearsystem r^¡ith consLant but unknown parameters, is

where the matrices Q^, Q. and Q. are symmetric andnon-negative. Since"l iå a ranáom variable, thecriËerion of the cootrol is taken as to minimize tle

Page 5: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

2.¿,i'(

exPected loss.

The admissible strategies are such that u(t) ismeasurable with respect to the sígma-algebra /¡tgenerated by {y(s), 0 ( s < t-1}, for each t€T'

The problem can be solved under two sets of assump-

tions. In one case the controls are linear func-tions of the measurements. It is then sufficientto assume second order properties of the random

Drocesses. If the random processes are assumed toie jointly gaussían, it is no longer necessary to."rir*" thât the controls are linear' The solutionis formally the same in both câses ' It is givenbv the seDaration theorem, which says thal the op-

timal control is

u(t) = -r, *ltlt-r¡rh"t" i(tlt-f) is the best estimate of lhe state*(a) gi""" y(t), y(t-l),... . The esEimate isgiven bY the Kalman filter

*{t*rlt) = e x(tlt-r) + b u(t) + K[v(t) - x(tlt-r)]

x(0) = m

The matrices K and L are obtained from solutionsof Riccati equations. The formulas first derivedby Kalman are standard texËboolt rnaterial' See e'g'Kíakernaak and sivan (L972) and Âström (1970)

'where references to original work as well as inter-pretations are given. A critical review of theiirr""t problem is found in trrritsenhausen (1971) ' A

special issue of IEEE transactions, Athans (f971)'is also devoted to the linear, quadratic problem'This issue contains many valuable papers coveringboth theory and aPPlications.

There are several trivial but useful extensionsthât can be made. The process disturbance and themeasurement errors rnay have unknown mean values 'This is taken care of by augmenting the state vec-tor with Ehe vector z = E v(t) with the associatedstate equation

z(t+1) = z(t)

Correlated disturbances with rational sPecËral den-

sities can be handled in a similar way by stateaugmentation.

The solution of the linear, stochastic controlproblem thus gives a regulator in terms of a lineæ,äynamical system. The dynamics of the regulatorarises from the Kalman filter.

Linear control theory is attractive for controlengineers because it gives a regulalor which ís a

liiear, dynamical system; a structure which has

been known to be useful for a long tin\e' The com-

plexity of the regulator directly reflects thecomplexity of the model and the criterion' The

matiices i< and L will in general be time varying'Hor^rever, if the system has constant parameters and

if N + æ, the matrices K and L will become con-stant matrices under reasonable conditions' This

The special case of linear control theory, obtainêdrfr"" 'aft" output is a scalar and there is no penaltyon the

"o.rttãl, is particularly simpl-e' By choos-

ing an innovat.ionts representation of the distur-bances and changing the basis in the state space

the model (2.1) can be transformed into

Minimum Variance Control

x ( t+t) u(t)+ e(r)

0

01

1

0

b-Ibz

kt

2kz

k.,-1

kn

x(t)+

1

nn

0 0 ...10 0 ... 0

n-1b

b

y(r)=x(t)+e(t)

Elimination of the state variables gives the follovring inpuL output relation

v('L) + a'v('c-')+ ' .".ili .'";.,:-l;;.1.;"".,:-1,

"'

By introducing the PolYnomials

n n-lA(z) = z" + ãlz - + ... * t'--1B(z)= bl""' *...*bn

c(z) = "n *

"!"n-L + ... + cn

and the forward shift operator q the model can alsobe written as

A(q)y(t) = b(c)u(t) + c(q)e(t) Q,2)

The conlrol law which minimizes the variance of theoutput i.e. the criterion

)e.r=Ey¿G) Q'3)

can of course easily be found from the general the-ory. The strategy can, however, also be found di-reätty by a very-simple argument as shovm in Åströrn(1970). rf the polynornial B(z) has no zeros outsidethe unit circle, the optirnal strategy is

u(r) =ffiQrt.> Q.4)

Notice that the stlategy is obtained without solvinga Riccati equation or without spectral factorization'Llhen the poiynomial B(z) has zeros outside the unitcircle, the Riccati equation will have several so-lutions, One solution corresponds to (Z '4) '

The control strategy which minimi.zes 9', will alsominimize

N^

(2.1)

solution, referred Eo as steacly scaEe corlEror,frequently used in aPPlícations.

12 = Eti xt=1

IS ,'1t) J \2.5)

Page 6: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

3.

1Nu+=uñl

Not.ice that the strategy whích minimizes the cri-terion

n')03=Ely'(t)+qru'(t)J

is also very easy Eo calculate' This strateBY is'howevur, ,roi th"- same as the strategy which mini-mizes

tv2(.) + qr,r2{t)J

It is easy to find cases where the strategy which

mínimizes f,, Bives an unstable, closed loop sys-

fluctuations ín the process the manufacEurer selectsthe set points of the regulators for the qualityvariables inside the tesi limits to ensure that hisp;;ã;;¿ is accepted. Bv reducing the variations in'the quality variables it is possible to move the

set points closer to the limits without changing .

tie prot.tility for acceptance' By doing this there

is a gain which can be exploited in many rrtays as

i""t"ãt"¿ production, reduction of raw-material or

;;;;;-;.;umption. rhe reduction of f luctuationsi"-qiäritv .rtii"bl.t is also of value by itself' rtis,'howevlr, often very difficult to express thisin monetary terns.

In some cases the more general quadratic criterioncan ¡e motivated from physical argurnents' One good

uxample is in the design of autopilots for shipswherl the average breaking force due to rudder ac-lions and angle of attack can be given approxima-ii""rv as a f,uadratic form' See Koyama (1967) ' rn*ã"a ä"tut ii is, however, difficult to assign

fenalties to different combinations of state and

"orrttol variables. The criterion (2'2) i-s therefcre

not always realistic'

The assumption that models for the process and the

disturbances are available is frequently the major

difficulty in applications. Even if a reasonablemodel for the process dynamics is a"'ailable, thecharacteristics of the disturbances are seldom kmwn'

To apply the linear theory it is Eherefore a neces-

sity'io-have methods to determine the desired mo-

dels from process experiments using.system identifi-cation tecirníques. ttt:-t presents little difficultyfor single output systems. In this case â canoni-cal form for the model can easily be given, and theparameters of this model can then be estimated' See

itströo, and Eykhoff (1971) and Evkhoff (I974) ' The

problem is more difficult because of the lack of,rrriqrru canonical forms. If the observability indi-ces or the controllability indices are known, it isstraightfotward to obLain a canonical form' Lack-

ing iiformation about the indices Èhere are however

maiy alternatives to explore. See e'g' Kalman

(1972) and MaYne (1974).

In many cases the difficulties mentioned above can

be oveicome; and the linear stochastic controlai"oty, in áombination with suitable syètem identi-fication techniques, is a useful tool for the con-trol engineer. This is clearly wítnessed by the in-cre"sinj number of apþlications. See Athans (1971) '

There are several applications of minimum variancecontrol which have covered all steps from processexperiments to on-line control. Applications topaier rnaehine control are given in Åström (1967) and

i'¡äfa "na Grimnes (1974). Control of interior cli-mãte is discussed in Jensen (1974). An applicationto design of autopilots for super tankers is done

by Källãtröm (1973). control of a pilot distillationpiant is described in Binder and Calvillo (L974)'The experience from all these applications havebeen that closed loop performânce can be predictedvery accurately from the models obtained' from planta-narimpnrs nn¡1 swstem identification. The plant

tem.

The following simple example v¡ill be used'

Exarnple 3.1

The minímun variance strategy for the system

y(t+l) + a y(t) = b u(t) + e(t+l) + c e(t) (3'6)

is given by

u(t)="o"t{t) (2'7)

Review of As

The fundamental assumptions made in the lineartheory are

o the process dYnamics is linearo the disturbances are stochasLic processes,

which are gaussian or of second ordero the. críterion is to minimize a quadratic func-

tiono the urodels for the Process and the disturban-

ces are known.

The assumptíon of linearity is quite reasonable\^7hen steady state control of industrial processesare considLred. The purpose of the control is tomaintain the process variables close to steadystate values. It is difficult to verify theassumption on the disturbances. lühen attempts tomodel disturbances are made, it is frequentlyfound that a stochastic process rnodel is adequateprovided that the model allows for drifting dis-iurbances. Such processes can be modeled by ínte-grators driven by processes with rational spectralãensities, for example the ARIMA process intro-duced by Box and Jenkins (1970). In practice itis also necessary to take into âccount that theremay occasionally be big disturbances'.

The quadratic criterion is often controvetsial' Inthe special case when Qr = 0 and Q, = 1 the crite-ríor '(2,2) is simply thÉ minirnizatton of the vari-ance of the outPut. This can be justified as a

criterion for steady state control using the fol-'lowínB ârsument. Throueh general trade rules, li-mits are imposed on important quality variables 'The custoner accePts a batch if his quality con-trol procedures show that the linits are exceeded

by a given fraction of the product. Due to the

experiments and the system identifícation are quitetime consuming, and they will also require skilledpersonnel.

Page 7: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

r=*ih(x(t),u(r))

3. NONLINEAR THEORY

The obvious nonlinear extension of the linear prob-lem discussed in the previous section is obtainedby replacing the linear difference equation (2.1)by the nonlinear equation

x(t+l) = f(x(t), u(t), v(t)) (r.tl

i(tl = c(x(t), u(t), e(t))

and by replacing the quadratic loss (2.2) by thegeneral loss function

(3.2)

Tf it is assumed that the disturbances are suchthat the stochastic process {x(t), y(t); t€T} is a

Markov process, the control problem can be formu-lated precisely and existence theorems can be givenin some cases. The theory is often formulated bygiving the transition probabilities of Ehe processdirectly instead of giving the nonlinear equations.The theory is presented in the Dvnamic Programmingframework. To do this the hyperstate

p.(A) = P{x(t)€Al Øt} (3.3)

igma-algebra generaled by {y(s) 'ntroducing the function

4.

constraints on computation capacity and controlforce, little more can be achieved by subtle chang-es in control logic."

The nonlinear theory has thus not had any Practicalapplications. In spite of this it will be shown.inthe next section that the nonlinear theory willgive some insight which indirectly may be of some

practical signif icance.

4. LTNEAR SYSTEMS WITH CONSTANT BUT UNKNOI,IN PARA-METERS

L7hen the linear theory was discussed in Section 2,it was remarked that it hTas a severe restrictionfrom the practical point of view to assume that theparameters of the models for the process dynamicsand the disturbances were known. The lacking know-ledge had to be supplied by a combination of expe-rimentation and statistics. An alternative to thiswould be to formulate the problem in such a waythat the lacking knorvledge of the process pârametersbecomes part of the problem statement. The linearquadratic problem discussed in Section 2 will thusbe considered, but the parameters of the models areassumed to be unknol^rn.

This problem can be regarded as a special case ofthe nonlinear problem by introducing the unknownparameters as nevT state variables. If the parame-ters are constants, the associated state equationsare simply given by

(3.4) 0(t+l) = 0(t)

It lurns out that the analysis to follol^¡ can equal-ly well be performed when the parameters are givenby

where /1<s<

theBy'. i:

siNtv. (n) h(x(t), u(t) | 4.)

the following funclional equaEion for V,- is obtairpdby standard Dynamic Programming argumenEs

vr(p) = min{lh(x,u)r.(dx) + Elvr+l (n.*r)l cEJ} ß.5)

This equation, which is called the Bellman equationafter Bellman (1961), is the starting point formany investigations. To have the right hand sideof (3.5) well defined it is necessary Lo have anequation for updating the conditional probabilitydistribution (3.3). Such equaLions were given byStratonovich (1960), Kushner (L967) and trrlonham(L964). Updating the conditional distribution isequivalent to a nonlinear filtering problem, whichhas been studied extensively. Numerical aspectsof this problem are recently given by Levieux(L974). The functional equation (3.5) has been in-vestigated in Florentin (1962), Åström (1964),Mortensen (f966), and Stratonovich (1963) .

The numerical solution of the functional equation(3.5) is very difficult because of the dimensíona-lity of state. For example, when "E¡3,

the hyp.t-state p is a probability distribution'over RJ. The

conceplually simple problem of storing the functionV thus becomes a major difficulty. This difficultyis the maín reason why the intensive interest innonlinear stochåstic control in the sixties endedin dispair. The situation is expressively staledin l^lonham (1968):

where 0 is a vector of unknown parameters, D is aknown matrix and tw(t), t€T] a white noise processirlhen D equals the identity matrix, the parametersare simply üIiener processes. If in addition r^¡ = 0the parameters are constants. The model (4.1) istherefore somer¡hat inconsistently referred to asthe case of "drifting parameters".

The assumption (4.1) is not particularly appealingfrom the point of view of applications because itmeans that the model has randomly varying parameters,but that Lhe statistical properties of the parame-ters are known. This is of course not â very real-istic situation.

From the theoretical point of view it is, however,interesting to observe that the case of constantparameteÍs is as difficult as the case of "driftingparameters".

It will now be discussed how far Ëhe nonlinear the-ory can be exploited in this particular case. Touse Dynamic Programming and the functional equation(3.5) it is necessary to introduce a hyperstatewhich.is a probability distribution over a space

= min E{k=t

0(t+t)=¡ s(r)+w(t) (4.1)

"To summarize: the only cure for dynamic distur-bances is tight feedback (high rates of informationprocessing) and large control forces. I^Iith fixed

and the number of unknown parameters. It is clearthat even a numerical solution of the functionalequation (3.5) is out of the question for any

Page 8: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

5

Jproblem of reasonable size. It is therefore ofinterest to consider further simplifications whichwill make it possible to reduce the dimension ofthe hyperstate. Two examples of such simplifiea-tions are given below.

Exanple 4.1

Consider the model (2.1) with C = I and e = 0. Allstate variables are thus measurable without error.Furthermore assume that the unknown Parameters areparameters of the matrices A and B and that thestochastic process {w(t)' t€T} is gaussian. The

conditíona1 distribution of the parameters giventhe sigma-algebra generated by {y(s), 1 < s ( t}is then gaussian and can be characterized by theconditional mean and two conditional covariances.See Farison (1967).

tr

Example 4.2

Consider the model (3.f). Assume that the para-meters ci are aII zero, that the unknown parame-ters are ã1t â2t ...¡ ê¡¡ b1, b2r..., br and that{e(t), t€T} are gaussian. Then the conditíonaldistribution of the parameters is gaussian. See

Mayne (1963).tr

The second example, which corresponds to-minimumvariance control, will now be explored in some

detail. The analysis follovrs Ä,strörn and lrritten-mark (1971).

Introduce the notations

o(r)=co1[ar(t)ar(t)...an(t) br(t)b2(t)...bn(t)]

e(t-1) =r-y(t-1) -y ( t-2) . . .-y (t-n)u,.:iìl Ii_iì ;

.

ao(t-1)=[-y(t-l) -y(t-2)...-y(t-n) ..:llil];;:

o(r) = E{o(r)lør}

p(r) = at to<t>-ô(r) I te(tl-ô<t) lrl atj

where rJ- is the sigma-algebra generated by y(t),y(t-l),i..

One-Step and Certainty Equivalence Control

I{aving obtained the hyperstate it is not straight-forward to do the minimization. Since

e y2(t+1) = r n{y2(t+I)IUJ = E E{[¡ u(t) +

,* eo(r)e(t) + e(t)l'lqJ (4.3)

is a quadratic function of u(t), it is straightfor-ward to show that the optimal control is

beoo + eopobu\E,/ =

b_ + pbb(4.4)

r^rhef e

P66 = Ele-oltb-bl

The control law (4.4) is called the one-steP con-trol because it minimizes the loss function (4.3),ffii-ch is the expected loss over one step only. Itis also called the cautious control.

Notice that if pbb = 0, then the control reduces to

u(r)=-S,r,b

oeu(r)=_ft.l

(4.s)

In the case of constant parameters the mÍnimumvariance control ís given by (2.4) i.e.

The process model

y(t) + ar(t-l)r(t-l)+...+arr(t-n)y(t-n) =

= br(t-l)u(t-1)+...+brr(t-l)u(t-n) + e(t)

can then be written as

y(t+t) = a(r)0(r) + e(r) = b(t)u(t) +

+ 9o(t)0(t) + e(t)

Notice that C(z) = I in the model (4.2). The con-trol (4.5) can be interpreted as the control ob-tained by computing the optimal control under thecondition that the parameters are known and thensirnply substituting the true parameters by theirestimates. The control (4.5) is therefor.e ca1ledthe certainty equivalence control.

Example 4.3

For the system (3.6) of Example 3.1 the one-stepcontrol (4.3)

ab + p-.u(t)=, o' (4.6)

b2 * Pbb

and the^certainty equivalence control (4.4) becomes

u(t) =*v(t) (4.7)

(4.2)

(4.2)

where the parameters are assumed given by (4.1).Recall that the case of unknov¡n but constant para-meters corresponds to D = I and v = 0. To applythe nonlinear theory the hyperstate (3.3) shouldfirst be determined. This appears difficult be-cause the stochastic process {y(t), t€T} is notgaussian. It \^ras, however, observed by Mayne(1963) that the hyperstate (3.3) is gaussian. See

reduction of the dimensionality of the problembecause the hyperstaEe can be represented by theconditíonal mean and the conditional covariance The control lar¡s (4.4) and (4.5) have empirically

been found to be useful, at least when they are

Page 9: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

âpp1íed to non-minimum phase "Y9!"t.t with constant

pãiã*"t"tt. The controllers wil1, however' have

åífficulties when they are applied to systems

where the parameter b changes its sign'^rThe con-

trol la\ts are very sirnilar when p66 << b' but verydifferent when p65 " Ê2. In par"ticular, the gain

of the certainty équívalence controller becomes

very large while the gaín of the one-stÎ'p controlb..ämes iery small for small values of b' The

ããi..i"ay eiuivalence controller will^thus givevery large control signals for small b' The one-

,i"i "o"itoller will, on the other hand, exhibit

the so-called turn-off effectrwhich means that the

control signal will be zero for long periods oftime. rntiitively this can be explained as fo1-lows. iolhen the estimate of b is poor i'e'o.. tr 62, th"tt the gain of the regulator (4'4)riÏf u" small. The control signal applied to theplant will then also be small. The estimate of b

ríf f "a

the same time be even r¡7orse' the gainiorãt, "a"'

This phenomena was observed in Åström-

wittenmark (1970). It is also discussed in Hughes

and Jàcobs (1974).

Dual Control

üIhen the parameters of the system (-3 '2) ate known'

the control stratègy (3.4) wiff minimize both thecriterion (3.3) and the criterion (3'5) ' -This isunfortunately not true in the case of unknown pa-

rameters. While it is fairly simple to find theone-step cont.rol for the system (4'2), it is a

much moie difficult task to find the control whichminimizes che criterion (3.5). In principle thesolution is given by the functional equation (3'5)'Introducing

N

v<ô<t>,P(t),Qo(t),t) = Min n{ tk=t

y2 1r+t¡lur\ (4.8)

The Bellman equation (3.5) then becomes

v(e (r) ,P(t)Qo(t) ,t) =

gipr{ tott)ô (t) l2 * a(t)P(È)ar(t)

2,^-s l¿

modet (4.2) changes its sign. The control given¡v (4.d) is a duã1 control in Feldbaum's sense'Simulations have shov¡n that the dual controllerãvoids the turn-off effeet of the one-steP control-ler and also the very large control signals of the

"årt.i"ty equivalencä control ' For any practical .

problem åf ieasonable size the control given by(¿.g)

"..t not be computed. It is therefore a good

research problem to find approximations which can

be computãd. In Tse et.al' (1973) and Tse and Bar-Shalom (1973) approximate solutions to the func-tional equation (3.5) are given for some examples

includini interception and soft-landing' The símu-

lations shor^t the superiority of the dual controllersas compared with the certainty equivalence control-lers. The computing time for the approximatedual controller is, however, 45 s on a UNIVAC 1108'Other approximaËions of the dual controller aregiven in-the paper by Bar-Shalo* 9!:11: (I974) atIhir "ytpotium

and by l^iittenmark (L974) '

6

SELF-TUNING REGULATORS

Example 5.1

Assume that the certainty equivalence controller(4.7) is applied to the system described by theequation (3.6). Let rv,i(t) denote the sample cova--t^-^^ ^r +l,a i--,,¡ ¡nàuthe ôrrtDut sisnats i.e.

a)

5

J vle {t+r),P (r+1),too (t+1), t+l) e

where

ô(t*r) = r ô(r) + r(t) s2 + p(t)P(t)çr(t) s

K(r) = o p(.)qr(.)[o2 + ç(t)p(t)çr(t)]-1

Qol (r+t) = - ç(t)ô(t) * o2 + ç(t)p(t)ar(t)

The properties of the one-step controller and the

cert;inly equivalence controller will now be ex-plored further in the case when the controllers are

ãpplied to a system with constan! but unknown para-rnài"tt. It will thus be assumed that D = T and

v = 0 in the equation (4.1) which describes the pa-

rameter variation. Since the regulators wete de-rived under the assumption that the parameters c1tc^- .... c in the model (3 '2) ate aLl zero, iE can

uÉ'"*p.át"å thaË the regulators will behave wellwhen ãpplied to systems with this property' rn thiscase the parameter estimates will be unbiased' Iflhe closed loop system is stable, the parameterestimates will then converge Ëo the model parame-

ters as time tends to infinity. The control laws

obtained in the limit will be the same as the con-Erol laws which could be computed if the model pa-

rameters were known a priori. Regulators r¡ith thisproperty will be cal1ed self-tuning controllers'äã""åteä""" conditions arefliven--ñ' the paper by

Ljung ãnd l^littenmark (1974) at this symposium'f¡átiãe that if the one-step controller and the cer-tainty equivalence controller both converge, theywill converge to the same regulaÈor because Pbb.-- 0as t + ó. The transient behaviour of the algorlthmsmay however differ sígníficantly'

In Âström and i,littenmark (1973) it was sho\'¡n thatboth the one-step controller and the certaintyequívalence controller are self-tuning when appliedto a systern described by the model (3'2) ' This isso*.whät surprising because the least squares esti-mate is biased when the polynomial C(z) in (3'2) isdifferent from unity. The basic idea behind thisresult can be illustrated by the following simpleexample.

2+o

1

/ñ+ dsÌ

(4.e)

s

eoi(t+l) =9o(i-1)(t) i -- 2, ..., 2t (4'10)

See Åström and i^Iittenmark (197f). The funcLionalequation (4.9), which will give the optimal stra-tegy, can be solved numerically íf the number ofparameters is sufficiently small, say one or atlh" *ott tno parameters. Solutions for specialcases ale g].ven rn IlorenLlrr \LJo¿,, r Dvrr¡rr¡ \Lzvr/,A.ström and l,Iittenmark (1971). The controllersobtained differ significantly from the one-stepcontrol and the certainty equivalence control inthose cases where the parameter b in the process

r1t

= + r y(r)u(r-r)t k=lyu

(t)

Page 10: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

7

/

1

r

The conditional nean values of the pârameters are

then given bY

.yr(1) + å{t)rrr(0, = t{t)rr,, (0)

ty,r(l) + a(t)rru(o) = b(t)ruu(o)

Since the control signals r^7ere generated by the

control law (4.7), we get

fn the particular case the norrnal equations whichgi"" ahä estimates may be ill conditioned' and dif-iiculties may arise if care is not taken hthen solv-ing them. SLe Nahorski (I974) ' The difficulty can

Ue"avoi¿e¿ by using pseudo inverses or by givingthe parameter 6 a fixed value'

Extensive simulations of the certainty equivalencecontroller and the one-step controller when appliedto a system hTith constant or slol4rly varying parame-

t"r" .r" given in l,Jieslander-ÍJittenmark (1971)ti^Iittenmarl (1973) and i^Iouters (197 4) in this sympo-

sj.um. Extensions of the algorithm to include penal-Ly on th" control are given.in Âström and llitten-märk (1974) and to multivariable systems in Peterkaand .qstrijm (1973) .

A self-tuning regulator will be an attractive solu-tion for a tineai stochastic control problern' Itsubstitutes the need for process identification by

a more complex control algorithm' The computationalreqr-rirum.nis both for the one-step control ând the

"eitainty equivalence control are very modest'

No practical control problem can be solved by theo-ty àlott". There will always be some elements ofsubjective judgement involved. For the certaintyequlvalence regulator the sampling interval, therod"l "omplexiiy

and the exPonential forgetting must

be selectåd. Experience has whown that these fac-tors are in many cases easily determined by engineer-ing judgement. The certainty equivalence regulatorhai therefore successfully been used in several ap-plications. Applications in the pâper industry areäescribed in Cãgrell and Hedqvist (1973, 1974),Borisson and l^Iittenmark (L974). Control of an orecrusher was discussed at this syrnposium, Borissonantl Syding (1974).

It has also been used to control an enthalphy ex-change6 Jensen (I974\ and as an autopilot for a

srrpui t"rrk"r,Källström (1974). In all these appli-

"riiu.r, the regulator has been running on-line on a

real industrial proeess. The simulation study by

lJouters (1g74) at this symposium also indicates thefeasibility as a regulator for a stírred tank reac-tãi. Simiiarly the paper by Vanacek (1974) at this,yrnporiu* indicates ihe applicabílity to airlfue1

(1) - åu!r t;,',-û <tlâ <tl ¡¡tr'l lv2 tr'l(s.1)

-t= - ! r Ia(r)-b(r)a(k)/b(k)]v(k)u(k)tk=l

vv

(r)yu

If the input and the output signals are bounded

and if the controller gain

a(t) = a(t)/b(t)

conveiges as t + -, it can be shown that the righthand sides of (5.1) converge Eo zero as t + @'

Hence if the closed loop system obtained with a

certainÈy equivalence controller is stable and ifthe controller gain converges âs t + ø, then theclosed loop system is such that

lim Í!æ1im rt+æ

vv

yu

(i)

(1)

(5.2)

(s .3)

The certainty equivalence controller thus atEempts

to dríve certain covariances of the inpu! and theoutput to zero. Assuming that the gain of the con-troller converges as time increases i'e'

a(t) /b (t) + cr

The closed loop system obtained in the limit thenbecomes

y(t+1) + (a-ob)y(t¡ = s(¡+1) + c e(t)

The condition (5.2) then gives controllers.

6. CONCLUSIONS* *l b

d = oz=

ora-llc

b

The second solution is impossible because it givesan unstable, closed looP system. l'Ie thus f indthat if the certainty equivalence controller con-veïges it r^till converge to a regulator' which isideãtical to the mínímum variance regulator (3'7)for the system (3.6).

Notice that the argument can also be extended tothe one-steP controller.

tr

Many process engineers claim that most of the índus-triäl'co.,trol pioblems (607" - 707") axe solved suf-ficiently well using ordinary PID regulators' Even

if this ,t.t"*.ttt is true, T belíeve that there are

control loops where more complex control algorithms.r. ptotit.ifu. Regulation of important qualityvariàbles where there is a significant pay-off inreduction of fluctuations is a typical example' One

possibility to obtain such control laws is to use

iin""t, stochastic control theory where the poten-tial of modeling dísturbances as random processes

ar" e*ptoited. The major difficulty is to obtain.pptopii"t" models for the process dynamics and the

dì"t.rib"nce". Such models can be obtainêd by sys-¡om irienlification. Linear stochascic control the-

In the example it is not^necessarl to assume thatthe parameter estimates a(t) and b(t) converge; itís only required that their ratior converges.

ory in combination with system identification has

beån successfylly applied to solve real industrialproblems.

Page 11: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

a

a\ó

7. REFERENCES

The nonlínear stochastic control theory does notyield strategies that are conveniently computed'íhe th.ory gi.tut, however, insight into nonlinearproblems. Ihis insight cân then be used to con-struct control strategies heuristically' It can

also be used to undersEand the behaviour of con-trol laws designed by other arguments ' A typicalcase is that of adaptive control where the non-iirr"", theory gives insight into the properties.of

"ãit.i" suboptimal regulãtors. In this way it is

possible to åbt.it self-tuning regulators and dualcontrollers. Even if much lheoretical work re-mains before these problems are fully understood'some useful applicaiions of available results have

already been made'

In surunary, while stochastic control theory is a

highly teltrnical area, which often is considered

"""" pl"ygtound for academics, it is my belief

that there are some results Ehat can be profitablyused to solve certain industrial control problems'

Digítal Computer", Preprints 4th IFAC/IFID Interna-tional Conference on oigital CompuLer Applicationsto Process Control, Zürich, Springer-Verlag, Berlin'

Bohlin, r. (1969):"optimal Dual Control of a Simple Process withUninown Gain", RePort TP l-B'196, IBl"l Systems Devel-

opment Division, Nordic Laboratory' Sweden'

Bohlin, T. (1970):"Information Pattern for l,inear Discrete-Time Mo-

dels with Stocbastic Coefficients", IEEE Trans

S:Þ' 104-106.

Borisson, U. and SYding, R. (1974): ..i'self-Trr.ri.rg control of an ore crusher", IFAC sym-

posium on SLochaslic Control Theory, Budapest'

Borisson, U. and Wittenmark, B' (1974):

"An Industrial Application of a Self-Tuning Regu-

lator", 4th IFAC Conference on Digital Computer

Applications to Process Control, Zürich'

Box, G. E. P. and Jenkins, G' M' (f970):iiiå" s"ti"s Analysis, Forecasting and Conlro1"'Halden DaY, San Fransisco'

Cegrell, T. and Hedqvist, f' (f973): ..;ii"""rár"r Adaptivã conirol of Paper Machines"'

3rd IFAC Symposium on ldentification and Systen

Parameter Estimation, Hague'

Cegrell, T. and Hedqvist, T' (1974): ^

"A-Net Approach to iontinuous Digester Control"'¿tn rracTirrD rnternational Conference on Digital¿;;o;;;;' Àpplications to Process control' Zürich'

Eykhoff, P. (1974) :

;å;;;.;'tdeniification - Parameter and sEate Esti-mation", Wi1eY, New York'

Farison, ¡' S. (1967):i'pãtã*"iut ldentification for a class of LinearDiscrele Systems", IEEE Trans 4:1-1 '

tO9 '

Fjeld, M. and Grimnes, K' (1974):;öÃpp'- A Norwegian Co-operation Project on ìlu1ti-variable control of a Kràft Paper ì{achine"'Preprints 4th IFAC/IFID International Gonference on

nigitat computer Applications.to Process Control'Zü;ich, Springer-Verlag, Berlin'

Florentín, J. J. (1962):;'õptimal lrobing, Adaptive control of-a Simple

fayesian Systemi, Int. J' on Control 13' 165-17i '

Hughes, D. J. and Jacobs, O' L' R' (1974):

"Turn-off, Escape and Probing in Non-lineal Stochas-

tic control" ' This s1'rnPosium'

Jensen, L. and Hänse1, R' (1974):"Compuier Control of an Entltalphy Exchanger"' Report

7412, Oivision of Automat ic Contrtrl, Lttntì Inst ' ofTechnologY.

Ã,ström, r' J. (1967):;'corprri.t Control of a Paper Machine - An Applica-tion of Linear Stochastic Control Theory", IBM J'Res. Dev. 11' 389-405.

Âström, K. J. (f970):"Introduction to Stochastic Control Theory", Acade-

mic Press, New York.

Âström, K. J. and EYkhoff, P' (197-l):"system rdentification - A survey", Automatica 7'L23-L62.

Âströrn, K. J. and i^IitEenmark, B' (1971):.

"Probláms of rdentification and Control", JMAÀ 34'90-113.

Åström, K. J. and witEerìmark' B' (f973):"on seif-Tuning Regulators", Automatica 9, 185-199'

Â,ström, K. J. and Wittenmark,'B' (1974):"Analyåis of a Self-Tuning Regulator for Nonminimum

Phase Systems", Submitted to IFAC Syrnposium on

Stochastic Control Theory, Budapest'

Athans, M (1971) (nditor):"Speciál Issue on Linear - Quadratic - Gaussian

Problem", IEEE Trans AC-16.

Athans, M (1971):"The Role and the Use of the Stochastic Linear -Quadratic - Gaussian problem in Control System

Design", rEEE Trans AC-16

Bar-Shalom, Y., Tse. E. and Larson, R' E' (1974):"Some Recent Advances in the Development of Closed

Loop StochasEic Control and Resource AllocationAlgorithms", This SYmPosium.

Bellman, R. (1961):"Adaptive Control Processes: A Guided 'lìour",Princeton Universicy Press, Princeton, N' J' Kalrnan, R. IÌ. (f 972)

Lr ll

or<1 . ì)iff .

Junc 14-23 'Binder, Z. and Calvi1lo, L (1974):"Dynamic Control of a Pilot Distillation Column by

Iiqns., NRL }l¿rtltt'rn¡tics Reseârcìì (lellter'igzr i" I-. I^reiss (e'd), "t)r'linar¡' Dif f e-

Page 12: Stochastic Control Theory and Some of Its Industrial ...lup.lub.lu.se/search/ws/files/48589629/7074.pdf · Stochastic Control Theory and Some of Its Industrial Applications Åström,

9

rential Equationstt, Academic Pressr New York'

Koyama, r. (f967):"on tnå Optimum Automatic Steering systems of Ships

ât Sea", J.S.N.A. Japan 122, L42-156.

Kushner, H. J. (1967):;;1y".*iá"f Equations for Optimal Nonlinear Filtering"J. of Math. Anal.and Appl., 3' I79-L90'

Kwakernaak, H. and Sivan, R. (1972):t'Linear Optimal Control", Wiley-Interscience, New

York.

Källström, C. (f973):"The Sea Scout Experiment", Report 7407(c), Div' ofAutomatic Control, Lund Institute of Teehnology'

Levieux, F. (1974):"Recursive Nonlinear Filtering: A Solution of theMathematical Problem and Appications", This sympo-

sium.

Ljung, L. and trrlittenmark, B. (1974):"Ã,syiitoti" Properties of Self-Tuning Regulators",Report 7404, Div.of Automatic Control, Lund Inst'of Technologv, also in this symposium.

Mayne, D. Q. (1963):"Optimal Non-stationary Estimation of the Parame-teis of a Linear System r¿ith Gaussian Inpurs",J. Electronics and Control 14' 101-112.

Mayne, D. Q. (1974):"Cànonical Forms for ldentification", ResearchReport 74/49, DepartmenË of Computing and Control,Imperial College of Science and Technology'

Mortensen' R. E. (1966):"Optimal Control of Continous-Time StochasEic Sys-tems", PhD Thesis, university of California, Ber-keley.

Peterka' V' (1970):"Adaptive ligital Regulation of a Noisy System",2nd IFAC Symposium on Identification and ProcessParameter Estimation, Prague' paper No. 6'2'

with Random Parameters via the Dual ControlApproach", IEEE Trans AC-18' 109-117'

Vanecek, A. (1974):"Coupled Air/Fuel Dynamic Optimízation and Tempe-

rature Regulation: A Case Study in Stochastic Con-

tro1", This symposium.

Wieslander, J. and tr^littenmark, B. (197f ):"An Approach to Adaptive Control Using Rcal Time

Identification", Automatíca 7, 2IL-2I7 '

InIitLenmark, B. (1973) :

"A Self-Tuníng Regulator", Report 7311, Div' ofAutomatic Control, Lund Institute of Technology,April 1973.

I^littenmark, B. (L97 4) :

"An AcEive Suboptimal Duâ1 Controller for Systems

with Stochastic Parametels", to appear in ACTA

(Canada) .

Irritsenhausen, H' S. (1971):"separation of Estímation and Control for DiscreLeTimà Systems", Proc. IEEE 59, L557-L566'

Wonham, W. M. (1964):"Some Applications of Stochastic Differentialfquations to Optimal Nonlinear Filtering", SIAM J'ConLrol 2, 347-369.

I^¡outers, I,J. R. E . (I974):"Parameter Adaptive Regulatory Control for Stochas-tic SISO Systems. Theory and Applications", Thissympo s ium .

Wonham, lI. M. (1968):"Optimal Stochastic Control", Symposium on Stochas-tic Problems in Control, June 26, L968, AmericanAutomatic Control Council.

Peterka, V. and Âström, K. J. (1973):"Control of Multivariable Systems with Unknownbut ConstanË Parameters", 3rd IFAC Symposium on

Identificatíon and System ParameEer Estimation,Hague.

Stratonovich, R. L. (1960)r"Conditional Markov Processes", Theory of Probabi-lity and its Application", 5' 156-178.

Stratonovích' R. L. (1963):"Most Recent Development of Dynamic ProgrammingTechníques and their Application to 0pcimal Sys-tem Design", 2nd Int. Congress of IFAC, Basle'

Tse, E., Bar-Shalom, Y. and Meier, III, L. (1973):"I^Iide-Sense Adaptive Dual Control for Nonlinear

c s

Tse, E. and Bar-Shalom, Y. (1973):"An Actively Adaptive Control for Linear Systems