sensitivity enhancing transformations for monitoring the process correlation structure

11
Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enhancing transformations for monitoring the process correlation structure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jprocont.2014.04.006 ARTICLE IN PRESS G Model JJPC-1761; No. of Pages 11 Journal of Process Control xxx (2014) xxx–xxx Contents lists available at ScienceDirect Journal of Process Control j ourna l ho me pa ge: www.elsevier.com/locate/jprocont Sensitivity enhancing transformations for monitoring the process correlation structure Tiago J. Rato, Marco S. Reis CIEPQPF, Department of Chemical Engineering, University of Coimbra, Rua Sílvio Lima 3030-790, Coimbra, Portugal a r t i c l e i n f o Article history: Received 31 October 2013 Received in revised form 3 April 2014 Accepted 4 April 2014 Available online xxx Keywords: Process monitoring Multivariate dynamical processes Variables transformation Partial correlations Marginal correlations a b s t r a c t In this article we address the general problem of monitoring the process cross- and auto-correlation structure through the incorporation of information about its internal structure in a pre-processing stage, where sensitivity enhancing transformations are applied to collected data. We have found out that the sensitivity of the monitoring statistics based on partial or marginal correlations in detecting structural changes is directly related to the nominal levels of the correlation coefficients during normal operation conditions (NOC). The highest sensitivities are obtained when the process variables involved are uncor- related, a situation that is hardly met in practice. However, not all transformations perform equally well in producing uncorrelated transformed variables with enhanced detection sensitivities. The most suc- cessful ones are based on the incorporation of the natural relationships connecting the process variables. In this context, a set of sensitivity enhancing transformations are proposed, which are based on a net- work reconstruction algorithm. These new transformations make use of fine structural information of the variables connectivity and therefore are able to improve the detection capability to local changes in correlation, leading to better performances when compared to current marginal-based methods, namely those based on latent variables models, such as PCA or PLS. Moreover, a novel monitoring statistic for the transformed variables variance proved to be very useful in the detection of structural changes result- ing from model mismatch. This statistic allows for the detection of multiple structural changes within the same monitoring scheme and with higher detection performances when compared to the current methods. © 2014 Published by Elsevier Ltd. 1. Introduction Statistical Process Control (SPC) is a pervasive industrial task with the aim to monitor the evolution of the overall process oper- ation state. It is implemented mostly by means of control charts designed to assess whether the process is only subject to normal causes of variation, inherent to its operation, or if a special cause of variation has occurred, that urges to be detected, diagnosed and fixed or accommodated. Special causes can result from equipment failure, abnormal changes in raw materials properties or any inter- nal or external events that affect process evolution and increase its variability, putting in risk not only product quality and process efficiency, but also the safety of people and the environment. There- fore, the rapid detection of faults and associated diagnosis of their origins are of outmost importance, and have attracted the inter- est of many researches and practitioners. However, even without Corresponding author. Tel.: +351 239 798 700; fax: +351 239 798 703. E-mail address: [email protected] (M.S. Reis). entering in an exhaustive enumeration of the multivariate method- ologies proposed through the decades, one can easily acknowledge that the proposed control charts were essentially designed to mon- itor the process mean [1–5], having some limited ability to address process variability, in particular concerning changes in the process correlation structure. In fact, only a comparatively low number of methodologies were proposed for explicitly addressing the prob- lem of detecting changes in correlation. The monitoring of the process dispersion and correlation is also a relevant issue since a process failure may not manifest itself so notoriously as a deviation from the nominal mean values, espe- cially due to the action of control systems fighting to maintain key process variables close to their target values. In multivariate pro- cesses the monitoring of process dispersion is usually achieved by monitoring statistics based on the generalized variance [6–8] or through the implementation of supervision schemes based on suc- cessive likelihood ratio tests [4,9–11]. Yet, these procedures tend to consider only the marginal covariance and ignore the variables inner associations. Even the SPC methodologies based on state- space models (such as CVA and subspace system identification), http://dx.doi.org/10.1016/j.jprocont.2014.04.006 0959-1524/© 2014 Published by Elsevier Ltd.

Upload: marco-s

Post on 30-Dec-2016

228 views

Category:

Documents


5 download

TRANSCRIPT

Page 1: Sensitivity enhancing transformations for monitoring the process correlation structure

J

Sc

TC

a

ARRAA

KPMVPM

1

wadcofifniefoe

h0

ARTICLE IN PRESSG ModelJPC-1761; No. of Pages 11

Journal of Process Control xxx (2014) xxx–xxx

Contents lists available at ScienceDirect

Journal of Process Control

j ourna l ho me pa ge: www.elsev ier .com/ locate / jprocont

ensitivity enhancing transformations for monitoring the processorrelation structure

iago J. Rato, Marco S. Reis ∗

IEPQPF, Department of Chemical Engineering, University of Coimbra, Rua Sílvio Lima 3030-790, Coimbra, Portugal

r t i c l e i n f o

rticle history:eceived 31 October 2013eceived in revised form 3 April 2014ccepted 4 April 2014vailable online xxx

eywords:rocess monitoringultivariate dynamical processes

ariables transformationartial correlationsarginal correlations

a b s t r a c t

In this article we address the general problem of monitoring the process cross- and auto-correlationstructure through the incorporation of information about its internal structure in a pre-processing stage,where sensitivity enhancing transformations are applied to collected data. We have found out that thesensitivity of the monitoring statistics based on partial or marginal correlations in detecting structuralchanges is directly related to the nominal levels of the correlation coefficients during normal operationconditions (NOC). The highest sensitivities are obtained when the process variables involved are uncor-related, a situation that is hardly met in practice. However, not all transformations perform equally wellin producing uncorrelated transformed variables with enhanced detection sensitivities. The most suc-cessful ones are based on the incorporation of the natural relationships connecting the process variables.In this context, a set of sensitivity enhancing transformations are proposed, which are based on a net-work reconstruction algorithm. These new transformations make use of fine structural information ofthe variables connectivity and therefore are able to improve the detection capability to local changes incorrelation, leading to better performances when compared to current marginal-based methods, namely

those based on latent variables models, such as PCA or PLS. Moreover, a novel monitoring statistic forthe transformed variables variance proved to be very useful in the detection of structural changes result-ing from model mismatch. This statistic allows for the detection of multiple structural changes withinthe same monitoring scheme and with higher detection performances when compared to the currentmethods.

© 2014 Published by Elsevier Ltd.

. Introduction

Statistical Process Control (SPC) is a pervasive industrial taskith the aim to monitor the evolution of the overall process oper-

tion state. It is implemented mostly by means of control chartsesigned to assess whether the process is only subject to normalauses of variation, inherent to its operation, or if a special causef variation has occurred, that urges to be detected, diagnosed andxed or accommodated. Special causes can result from equipment

ailure, abnormal changes in raw materials properties or any inter-al or external events that affect process evolution and increase

ts variability, putting in risk not only product quality and processfficiency, but also the safety of people and the environment. There-

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

ore, the rapid detection of faults and associated diagnosis of theirrigins are of outmost importance, and have attracted the inter-st of many researches and practitioners. However, even without

∗ Corresponding author. Tel.: +351 239 798 700; fax: +351 239 798 703.E-mail address: [email protected] (M.S. Reis).

ttp://dx.doi.org/10.1016/j.jprocont.2014.04.006959-1524/© 2014 Published by Elsevier Ltd.

entering in an exhaustive enumeration of the multivariate method-ologies proposed through the decades, one can easily acknowledgethat the proposed control charts were essentially designed to mon-itor the process mean [1–5], having some limited ability to addressprocess variability, in particular concerning changes in the processcorrelation structure. In fact, only a comparatively low number ofmethodologies were proposed for explicitly addressing the prob-lem of detecting changes in correlation.

The monitoring of the process dispersion and correlation is alsoa relevant issue since a process failure may not manifest itself sonotoriously as a deviation from the nominal mean values, espe-cially due to the action of control systems fighting to maintain keyprocess variables close to their target values. In multivariate pro-cesses the monitoring of process dispersion is usually achieved bymonitoring statistics based on the generalized variance [6–8] orthrough the implementation of supervision schemes based on suc-

hancing transformations for monitoring the process correlationnt.2014.04.006

cessive likelihood ratio tests [4,9–11]. Yet, these procedures tendto consider only the marginal covariance and ignore the variablesinner associations. Even the SPC methodologies based on state-space models (such as CVA and subspace system identification),

Page 2: Sensitivity enhancing transformations for monitoring the process correlation structure

ING ModelJ

2 Proces

a[twusrsobcRtdtaf[ispdcrvaatswivamvss

fpaeisa[

casstarwptc

2m

st

ARTICLEJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of

re based on the marginal cross-covariance between the variables12–14]. On the other hand, partial correlations have the poten-ial to describe such local associations, even though in a non-causalay, being already applied to retrieve the structural informationnderlying collected data [15–17]. Moreover, Rato and Reis [18]howed recently that monitoring statistics based on partial cor-elations have indeed the potential to enhance the detection oftructural changes, when compared to current state of the art meth-ds. In this preliminary work [18], the authors established thease approach to monitor structural changes through the appli-ation of partial correlations and concluded that the proposedMAX statistic (maximum norm of normalized partial correla-ions) was the one presenting better characteristics for detectingrifts in the process structure. Furthermore, it was also provedhat structural changes are easier to detect for uncorrelated vari-bles. These were obtained through a Cholesky decomposition,ollowing a similar approach to Hawkins and Maboudou-Tchao19]. Even though this approach can lead to a significant increasen detection performance in some circumstances, it does not con-ider the inner relationship between the variables. Therefore, theotential of effectively using the process correlated network toecorrelate the variables and for monitoring the process auto- andross-correlation structure was not fully exploited yet. In the cur-ent work we explore this possibility and propose a new class ofariables transformations based on the network reflecting the innerssociations between variables which are obtained through thenalysis of partial correlations, and that can finally lead to consis-ent superior detection performances. Additionally, a monitoringtatistic specifically designed to detect changes in the varianceill also be proposed here, in order to complement the monitor-

ng of partial correlations which are insensitive to changes on theariables dispersion. Departures from the system’s normal oper-tion conditions (NOC) model can also be detected by this newonitoring statistic. Therefore, used in combination with the pre-

iously proposed RMAX statistic, it becomes possible to detect a fullpectrum of faults specifically related with changes in the processtructure.

Some related approaches have already been proposed mostlyor fault diagnostics, based on the analysis of fault propagationathways over an existent causal network [20,21]. However, thesepplications are dedicated to the diagnostic stage, assuming thexistence of a previous detection phase and also the availabil-ty of a priori knowledge regarding variables connectivity. Theame applies to model based approaches to fault detection, where

process dynamic model is always assumed to be available22–24].

This article is organized as follows. In the next section, theurrent monitoring statistics based on the marginal covariancere briefly reviewed. Then, we describe the proposed monitoringtatistics based on partial correlations for monitoring the processtructure. The set of proposed sensitivity enhancing transforma-ions studied are introduced in the next section. We then conduct

performance comparison study encompassing both the cur-ent monitoring statistics and the ones proposed in this articlehen applied to systems with different degrees of complexity, andresent and discuss the results obtained. Finally, we summarizehe main contributions proposed in this paper and present ouronclusions.

. Statistical monitoring of process structure based onarginal covariance

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

The procedures for monitoring the multivariate process disper-ion are either based on the implementation of a likelihood ratioest or on the generalized variance and can be traced back to works

PRESSs Control xxx (2014) xxx–xxx

of Alt [6] and Alt and Smith [9]. In these articles, the authors proposeto monitor the statistic,

W = −p(n − 1) − (n − 1) ln( |S|

|�0|)

+ (n − 1)tr(�−10 S) (1)

where, p stands for the number of variables, n is the number ofobservations, �0 is the in-control covariance matrix and S is thesample covariance matrix. This monitoring statistic responds wellto perturbations on low dimensional linear stationary systems andhas the advantage of being invariant to any non-singular transfor-mation of the data [19]. However, as it requires the inversion of thein-control covariance matrix, it may be unsuitable for cases where�0 is ill-conditioned, as happens in moderate/high dimensionalprocesses.

Another procedure, based on the likelihood ratio test, was pro-posed by Levinson et al. [10] and explores the idea that undernormal operation conditions the covariance matrix estimated withthe complete data set should be equal to the one estimated by themean square of successive differences. In order to obtain the mon-itoring statistics one has to calculate first the in-control samplecovariance matrix, S0, from a reference data set with n0 obser-vations. After that, the ith test sample covariance matrix, S1,i,determined from a subgroup with n1 observations is combined withS0 in the pooled estimator of the covariance [10],

Sw,i = (n0 − 1)S0 + (n1 − 1)S1,i

n0 + n1 − 2(2)

The monitoring statistic is then computed as Gi = mMi where,

Mi = (n0 + n1 − 2) ln |Sw,i| − (n0 − 1) ln |S0| − (n1 − 1) ln∣∣S1,i

∣∣

m = 1 −(

1n0 − 1

+ 1n1 − 1

− 1n0 + n1 − 2

)(2p2 + 3p − 1

6(p + 1)

) (3)

To specifically monitor the sample variance, a more efficient moni-toring scheme proposed by Costa and Machado [25] can be applied.Their procedure is based on the sample covariance matrix, by tak-ing the maximum value of the sample variance of normalized data,i.e.,

VMAX = max{s21, s2

2, . . ., s2p} (4)

where s2i

= zTizi/n − 1 and zi = (xi − �i)/�i. The normalization of

the data is an important step on this procedure since it guaranteesthat all the sample variances have the same probability to exceed acertain UCL. Even though this procedure is simple and efficient, bydesign it is only capable to detect increases in variance.

3. Statistical monitoring of process structure based onpartial correlations

Hitherto, the current monitoring statistics are typically basedon measures of multivariate dispersion based on marginal informa-tion. It is well-known that such information leads to correlationsbetween variables that are not directly associated, as long as theyshare some common inducing variation sources or are part of thesame causality chain. Therefore, these methodologies are in prin-ciple unable to detect subtle and localized changes on the processstructure. One way to improve the detection capability to faultsspecifically related with the local process structure relies on the useof partial correlations. Partial correlations have the ability to iden-tify variables that are directly associated or connected, a featurethat was already extensively explored to reconstruct interaction

hancing transformations for monitoring the process correlationnt.2014.04.006

networks [15–17,26] as well as in classification problems [27–29].In the SPC context, it is expected that monitoring partial correla-tions can bring the benefits of increasing sensitivity to more subtlechanges in the process structure (due to the finer description of

Page 3: Sensitivity enhancing transformations for monitoring the process correlation structure

ING ModelJ

Proces

tllamtbdist

3

otiiasttbclavraclim(tcc

r

1

r

2

r

FmlfistmramT

R

R

ARTICLEJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of

he NOC structure) and to speed up the diagnosis stages (due to theocalized nature of the information they provide), which potentiallyeads to a significant reduction in the total time to detect, diagnosend accommodate or fix, a given process upset. With these aims inind, in the following subsections we present statistics for moni-

oring the fine process structure based on the information providedy partial correlations, along with their complementary statisticsedicated to monitor changes in variance. As will be demonstrated

n the results section, the combined use of these complementarytatistics can significantly increase the methods detection abilityo a large class of faults.

.1. R0MAX and R1MAX

The main idea behind the use of partial correlations is to bringut differences between directed and undirected variable’s rela-ionships by removing the common variables effects [27–29]. Thiss usually done by conditioning upon (i.e., controlling for, or hold-ng constant) one or several other variables (z) before checking anssociation between the two designated variables (x and y). Foruch, one can adopt a linear regression approach where x and y arehe responses and z (conditioning variables) the regressors, andhen compute the correlation between the residuals obtained inoth models for x and y. This correlation is known as the partialorrelation (rxy·z) of the original variables and is a measure of theevel of direct association between the components of x and y thatre uncorrelated with z. By conditioning over all the remainingariables, it is possible to infer the existence or not of a directelationship between the pair under analysis, from which a finernd more detailed picture of the true variables interaction networkan be obtained. Even though no causal directionalities are estab-ished at this point, the direct variables associations can be properlydentified using this type of procedure. Instead of fitting multiple

odels for each pair of variables, the following recursive formulasEqs. (5)–(7)) can also be applied for computing the partial correla-ion coefficients of lower orders (the order of a partial correlationoefficient is given by the number of variables conditioned in itsomputation). Similar equations exist for higher orders.

0th order partial correlation:

xy = cov(x, y)√var(x)var(y)

(5)

st order partial correlation:

xy.z = rxy − rxzryz√(1 − r2

xz)(1 − r2yz)

(6)

nd order partial correlation:

xy.zq = rxy.z − rxq.zryq.z√(1 − r2

xq.z)(1 − r2yq.z)

(7)

ollowing a reasoning analogous to network reconstructionethodologies, Rato and Reis [18] proposed to test partial corre-

ations against their NOC distributions. The procedure consists ofrst pre-processing and normalizing the partial correlations to atandard normal distribution with zero mean and unit variance, andhen, the maximum norm, ||x||∞ = max{|x1|, |x2|, . . ., |xn|}, (i.e., the

aximum of the absolute values) of these normalized partial cor-elations is considered as the monitoring statistic. This procedure ispplied to both 0th order partial correlation (corresponding to thearginal correlation coefficients) and 1st order partial correlations.

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

he monitoring statistics are defined as,

0MAX = ||w(r0)||∞ = max{|w(r0)|} (8)

1MAX = ||w(r1)||∞ = max{|w(r1)|} (9)

PRESSs Control xxx (2014) xxx–xxx 3

where w(·) is a normalization function based on the partial cor-relations distribution, r0 is the (p(p − 1)/2) × 1 column vectorcontaining all distinct correlation coefficients (0th order partial cor-relations) and r1 is the (p(p − 1)(p − 2)/2) × 1 column vector of 1storder partial correlation coefficients. The normalization function iseasy to construct when the number of samples (n) is large, sincethe qth order partial correlation coefficients, tend to be normallydistributed with zero mean and unit variance, when transformedaccording to [30],

w1 =√

n − q − 1(r − �)

1 − �2(10)

where r is the sample partial correlation and � is the populationpartial correlation.

Another useful normalization function is based on the Fisher’s ztransformation [30], which tends more rapidly to normality,

w2 =√

n − q − 1

2

[ln

(1 + r

1 − r

)− ln

(1 + �

1 − �

)](11)

In cases where it cannot be assumed that the partial correlationcoefficient follows a normal distribution, an estimation of the den-sity function can be applied, for instance based on kernel densityestimation, which can then be used to normalize the partial corre-lations by matching the empirical distribution quantiles with thoseof the standard normal distribution.

3.2. VnMAX

As stated above in this article, monitoring solely based onthe information provided by partial correlation coefficients is notenough to detect all faults, as some events may pass undetectedsuch as those affecting process variance. This is particularly rele-vant when transformed variables are used, since a model mismatchcan then be translated into changes on the transformed variables’variance. In this context, the VMAX statistic proposed by Costa andMachado [25] (see Section 2) is the one offering more potentialfor dealing with this problem. However, as it is defined, it is onlyable to detect a significant increase in the variance. Of course, insome situations this may be all that is necessary, but in the presentcontext we are interested in detecting any departure from nor-mal operation conditions. Therefore, we propose a modification ofthis statistic based on the inference properties of the variance in anormal distribution.

The test statistic for the hypothesis test that a sample varianceobtained from n observations of a random normally distributedvariable is equal to �2

0 , is given by Montgomery,

�20 = (n − 1)s2

�20

(12)

where �20 follows a chi-squared distribution with (n − 1) degrees

of freedom [31]. The one sided-test is already performed with theVMAX statistic, for detecting an increase in the variance (Section2). In order to set the conditions for maintaining a methodologyconsistent with RMAX and use a single control limit, the probabil-ity of exceeding symmetrical UCL and LCL must be the same, whichcan be easily achieved by taken the absolute value of a symmetricrandom variable. In this particular case, the Wilson–Hilferty trans-formation can be applied to the chi-squared distribution to obtain a

hancing transformations for monitoring the process correlationnt.2014.04.006

normally distributed random variable [32]. The use of this transfor-mation also facilitates future constructions of a combined statisticto monitor both variance and correlation coefficients, since corre-lation coefficients are already approximately normally distributed.

Page 4: Sensitivity enhancing transformations for monitoring the process correlation structure

ING ModelJ

4 Proces

d

wd

t

w

wm

V

wa

vcst(bc

4

cntdOctwbsdwYeiahesto[tane

lrot

˙

ARTICLEJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of

The Wilson–Hilferty transformation states that a chi-squaredistribution with � degrees of freedom is well approximated by:

2˛,� ≈ �

[z˛

(2

9�

)0.5+

(1 − 2

9�

)]3

(13)

here z˛ is the 100 × ˛% upper quantile from the standard normalistribution.

Given this relation, we propose to use the following transforma-ion:

s = (s2/�20 )

1/3 − (1 − (2/(9(n − 1))))√2/(9(n − 1))

(14)

hich is approximately distributed as N(0, 1). Consequently, theodified (normalized) VnMAX statistic becomes:

nMAX = ||ws(v)||∞ = max{|ws(v)|} (15)

here v is a (p × 1) column vector containing the variables’ vari-nce.

As a consequence of the above transformation, this modifiedersion is able to detect both increases and decreases in the pro-ess variance. It is also capable to detect changes in the processtructure, when any of the sensitivity enhancing transformationso be described in Section 4, are applied. As an alternative to Eq.14), a non-parametric density estimation methodology can alsoe employed as referred in Section 3.1 for the case of correlationoefficients.

. Sensitivity enhancing transformations (SET)

The detection of changes on (marginal or partial) correlationoefficients is highly dependent of their nominal values underormal operation conditions. For instance, when the intrinsic rela-ionship between two highly correlated variables suffers a smalleviation, their correlation coefficient remains almost unchanged.n the contrary, whenever two initially unrelated variables becomeorrelated to some extent, their correlation changes more abruptlyhan in other situations. A demonstration of this phenomenon, asell as an example of its impact on the correlation coefficients can

e found in Ref. [18]. This feature suggests that, in order to detectmall changes on the structure, it is advantageous to pre-processata in order to obtain uncorrelated variables. A similar principleas applied by Hawkins [33] with the regressed-adjusted variables.et, in his work, the variables transformation is intended just toliminate the contribution of preceding variables in order to bettersolate drifts in the mean value and not to enhance the detectionbility to changes in correlation. The use of uncorrelated variablesas also been applied to monitor the marginal covariance. Nev-rtheless, these procedures only adopt such transformations forimplification purposes (e.g. [19,34]). Moreover, they tend to usehe inverse of the covariance matrix, which may be ill-conditioned,r triangularization methods (like the Cholesky decomposition35]), which may not decorrelate data in the best way. Still, the lat-er approach showed to be quite useful, provided that the variablesre properly ordered [18], and therefore it will be briefly introducedext in order to better evaluate the performance of the sensitivitynhancing transformations proposed in this paper.

The Cholesky decomposition consists of a series of successiveinear regressions where the ith variable is regressed onto theemaining (i − 1) preceding variables, based on the factorization

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

f the covariance matrix, �, into a lower triangular matrix L, suchhat,

= L LT (16)

PRESSs Control xxx (2014) xxx–xxx

This matrix, can then be used to obtain uncorrelated variables withunit variance through the following transformation [35]:

u = L−1 (x − �) (17)

To account for the dynamic and non-linear features of the systems,Rato and Reis [18] suggested the inclusion of time-shifted vari-ables or polynomial terms (depending of the situation) in the datamatrix. Exemplifying for the case of including time-shifted vari-ables, these additional variables should be placed at the beginningof the extended data matrix, as follows:

X =[

X(l) · · · X(1) X(0)]

(18)

where X(j) is an n × p matrix of variables shifted j times into thepast (i.e., with j lags), so past variables become the regressors ofpresent variables.

After this step, the regular Cholesky decomposition is per-formed, resulting in a new set of uncorrelated variables accordingto Eqs. (19) and (20).

˜ = L L

T(19)

u = L−1

(x − �) (20)

Note that only the regression variables related to the presentstate are of interest (i.e., the last p variables in u), since they arethe ones that correspond to the residuals of the linear regressionof the variables in the current time, onto those from the past. Sincethis last transformation can be seen as an extended version of thetypical Cholesky decomposition, it will be referred as TChExt in therest of this paper.

As stated earlier, the Cholesky decomposition, is essentially atriangularization method that regresses the ith variable onto theother (i − 1) variables that preceded it. Therefore, the sequence bywhich variables are included in the model is defined by their par-ticular order of appearance in the data matrix. In this case, it isunlikely that such an arbitrary ordering scheme will provide thebest description of the system structure. In fact, the more mean-ingful variable ordering results from placing the more importantvariables (i.e., variables that affect most of the others or that areplaced on root nodes of the true variable interaction network) atthe beginning, by considering a priori knowledge of the process.As this information may not be always available, an alternativetransformation is proposed. The goal of the new transformationis to break the relevant variables relationships upon applicationof linear regression only on the variables that are indeed related,as exemplified on Fig. 1 (“breaking” means here the removal of anassociation by application of a proper model). For such, the relevantedges between variables must be first identified through a networkreconstruction technique, as described in the next paragraph (seealso Fig. 1(a) and (b)).

In order to identify the edges linking directly associated vari-ables, we suggest the use of partial correlation up to the 2ndorder, according to a procedure similar to that applied elsewhere[16,28,29]. The main difference relies on the inclusion of time-sifted variables or polynomial terms in the regression models,for addressing dynamic and non-linear process features, respec-tively. By conditioning on these additional variables, this algorithmbecomes capable to detect both dynamic and non-linear relation-ships, resulting in more accurate networks. After that, the causaldirections of these undirected edges are determined in a secondstage, by applying another algorithm based on the variables’ cross-correlation. Alternatively, one can also use Granger causality [21]or transfer entropy [20] to determine the variables causal direction.

hancing transformations for monitoring the process correlationnt.2014.04.006

The pseudo-code for these algorithms, are available in the journalwebsite as supplementary material.

The second step of the transformation involves the regressionof each variable onto its parents (see Fig. 1(c)), resulting in a final

Page 5: Sensitivity enhancing transformations for monitoring the process correlation structure

ARTICLE IN PRESSG ModelJJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of Process Control xxx (2014) xxx–xxx 5

Fig. 1. Implementation of the proposed transformation procedure: (a) original network, (b) identification of the relevant edges, (c) implementation of successive linearregressions involving the directly connected variables and (d) final model.

Table 1Nomenclature associated with the sensitivity enhancing transformations coveredin this study.

Transformation Description

TX Original variablesTChExt Variables transformed using the Cholesky

decomposition after incorporation oftime-shifted variables and/or polynomialterms, Eq. (20)

TNet Variables transformed using a set of linearregression models constructed only forrelated variables (see Fig. 1)

TNetCh Variables transformed using a set of linearregression models constructed only forrelated variables with an additional

rcsrmpudNoC

iimrn

cmt

5

thimpy

Table 2Definition of the monitoring statistics proposed in this work.

Sensitivity enhancing transformation Statistic

VnMAX RMAX

v r0 r1

TX VnMAXX R0MAXX R1MAXX

Cholesky decomposition of the resultingresiduals

egression model where only the directly connected variables areonsidered to obtain the new set of residual variables. On dynamicystems, time-shifted variables should also be included on theegression model and for non-linear systems, polynomial termsust also be added. As a possible additional step, a Cholesky decom-

osition can be applied to the residuals, in order to ensure thatncorrelated variables are indeed obtained and also to accommo-ate any missed relationship. On the rest of the article, these novelETwork-oriented transformations will be referred as TNet (with-ut a Cholesky decomposition of the residual) and TNetCh (with aholesky decomposition of the residual).

These new proposed transformations are able to integratenformation about the local structure of correlations, leading tomprovements in the monitoring statistics performance. Further-

ore, as the variables connectivity is estimated through networkeconstruction methods (using the algorithms proposed here), theature of the transformations remain fully data driven.

The complete set of sensitivity enhancing transformationsovered in this study, and the associated nomenclature, are sum-arized in Table 1. The proposed monitoring statistics for each

ransformation are presented in Table 2.

. Results

The effects of the proposed sensitivity enhancing transforma-ions in the performance of the developed monitoring statistics isere compared against their current counterparts already proposed

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

n the literature, namely the ones described in Section 2. Otheronitoring statistics were also considered, but excluded from this

resentation due to their lower relative performance. This anal-sis was performed for three simulated systems with different

TChExt VnMAXChExt R0MAXChExt R1MAXChExt

TNet VnMAXNet R0MAXNet R1MAXNet

TNetCh VnMAXNetCh R0MAXNetCh R1MAXNetCh

degrees of complexity in order to demonstrate the versatility of theproposed methodology, its robustness to misspecifications of thecausal structure and the consistently higher detection capabilityregarding other alternative frameworks.

For each case study, faults of different magnitudes were gen-erated and the associated curve of detection rate versus fault’smagnitude was determined. For the sake of making it possible toestablish a sound comparison between the different methods, eachcurve is summarized by a performance index (N) based on the areaunder the curve. Note that, as the fault’s magnitude increases, thedetection rate also increases, until a maximum value of 1 (i.e. allpoints are out-of-control). Therefore, a monitoring statistic with agreater area under the detection rate curve tends to 1 more rapidlyand will present higher detections rates. This index was computedfor each fault and normalized so that its values fall in the range [0,1], where 1 represents the best performance observed in each fault,by a given method. In this way the relative performance of differ-ent methods in each fault can be established and easily analyzed.Additionally, we have also assessed the statistical significance ofthe detection rates of the studied monitoring statistics through theapplication of a permutation test [36], from where the associatedp-values were determined. These results are available in the journalwebsite as supplementary material.

5.1. Case study 1: stationary non-linear system

To test the performance of the proposed monitoring methods(SET and statistics), the causal network proposed by Tamada et al.[37] was adopted. This network is composed by 16 nodes (or vari-ables) causally related according to the representation providedin Fig. 2. The variables relationships were approximated by poly-nomials according to Eq. (21), where εi is a white noise sequencewith a signal-to-noise ratio of 10 dB. By application of the identi-

hancing transformations for monitoring the process correlationnt.2014.04.006

fication and causal directionality inference algorithms referred inSection 4 to a data set with NOC data, the original causal networkwas obtained (full reconstruction). The variable transformation TNetwas then carried out taken this inferred structure into account, and

Page 6: Sensitivity enhancing transformations for monitoring the process correlation structure

ARTICLE ING ModelJJPC-1761; No. of Pages 11

6 T.J. Rato, M.S. Reis / Journal of Proces

11

10

9

2

45

6

7

8

12

13 14

15

3

1

16

Fig. 2. Graphical representation for the causal structure for the artificial networkused in case 1.

Table 3Definition of the faults and respective variables involved for the network systemwith a non-linear model structure. ı is a multiplicative factor that changes the modelparameters (under NOC, ı = 1).

Fault Variables relation changed

A g8 → g1 (g1 = 1.20ıg8 + 0.80g9 + ε1)B g1 → g3 (g3 = (ıg1 − 4)(g1 + 4) + ε3)C g8 → g10 (g10 = 0.02g2

8 + 0.44ıg8 + 0.82 + ε10)D g3 → g14 (g14 = 0.020ıg2

3 + 0.44g3 − 0.82 + ε14)

aut9eta

wttoddtlsrwST

nrmtTce

dt (1 + T3/KT3 )(1 + KT4 /T4)

E g8 → g11 (g11 = −0.053ıg38 − 0.00068g2

8 + 0.52g8 + 0.50 + ε11)F g3 → g15 (g15 = 1.40ıg3 + ε15)

3rd order polynomial model was finally fitted for each variablesing the identified variables’ parents as regressors (for instance, inhe case of variable 1 the regressors identified were variables 8 and). Transformation TChExt was also obtained from the same refer-nce data set by adding polynomial terms to the data matrix. Theseransformed variables, along with the original ones, were then useds inputs of the monitoring statistics of the process structure.

The system was perturbed with the faults presented in Table 3,here the magnitude of change, ı, was set to cause changes on

he model’s parameter on the range of ±10%. In each perturba-ion, 1000 sample covariance matrices were determined with 3000bservations each. Based on the sample covariance matrices, weetermined the monitoring statistics and their corresponding faultetection rates. The same procedure was repeated 10 times in ordero estimate the confidence levels of the detection rates. The controlimits for all the monitoring statistics were preliminarily adjustedo that all monitoring statistics present the same false detectionate of 1%. Subsequently, for each fault the performance index Nas obtained, resulting in the global distribution, as shown in Fig. 3.

imilarly, the results of the permutations tests are summarized inable S5 of the supplementary material.

Analyzing the results obtained, it is possible to observe thatone of the current monitoring statistics presented detectionsates higher than 5%. This result is a clear evidence that theseonitoring statistics are not able to deal with the systems’ fea-

ures and stress the need for more efficient monitoring schemes.

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

he proposed statistics, RMAX and VnMAX, were the only onesapable to effectively detect these faults. Furthermore, it is alsovident that not all of the sensitivity enhancing transformations

PRESSs Control xxx (2014) xxx–xxx

are appropriate. For instance, TChExt does not model correctly thesystems non-linearity, leading to partial correlations that do notfollow their assumed NOC probability distributions. Therefore, theRMAX statistic cannot be directly applied with this transforma-tion, since RMAX requires that all partial correlation coefficientshave the same probability of exceeding the control limits. The samesituation arises with the original untransformed variables. Never-theless, R0MAXChExt and R1MAXChExt still present some capabilitiesto detect faults and when the partial correlations are properlypre-processed with resource to an adequate estimation of their dis-tribution, the RMAX performance is significantly increased. Yet, theuse of the proposed transformation TNet is preferable because theactual variables’ relationships are explicitly considered in its con-struction. Transformation TNet also leads to uncorrelated variablesand partial correlations following normal distributions. Conse-quently, R0MAXNet and R1MAXNet can detect efficiently most ofthe simulated faults, with the exception of fault D which is onlydetected by VnMAXNet. Hence, transformation TNet is adequatefor modelling and monitoring this non-linear system, since eitherRMAXNet or VnMAXNet are able to rapidly detect any structuralchange, while current monitoring statistics in general fail in thistask.

g8 = ε8, g9 = ε9, g16 = ε16

g10 = 0.020g28 + 0.44g8 + 0.82 + ε10

g11 = −0.053g38 − 0.00068g2

8 + 0.52g8 + 0.50 + ε11

g1 = 1.20g8 + 0.80g9 + ε1

g2 = 0.60g1 + ε2

g3 = (ıg1 − 4)(g1 + 4) + ε3

g4 = 0.020g21 + 0.44g1 + 0.82 + ε4

g5 = −0.057g31 − 0.077g2

1 + 0.52g1 + 0.22 + ε5

g6 = 0.60g1 + ε6

g7 = 0.70g1 + ε7

g12 = 0.80g16 + 0.60g3 + ε12

g13 = 1.30g3 + ε13

g14 = 0.020g23 + 0.44g3 − 0.82 + ε14

g15 = 1.40g3 + ε15

(21)

5.2. Case study 2: gene network model

The test system considered in this case study was proposed byFuente et al. [16]. It regards the dynamic modelling of transcriptionlevels (Ti) in a gene network, and consists of the following set ofdifferential equations:

dT1

dt= V1

(1 + KT1 /T4)− k1T1 + �1T1

dT2

dt= V2

(1 + T1/KT1 )− k2T2 + �2T2

dT3

dt= V3

(1 + KT2 /T2)− k3T3 + �3T3

dT4

dt= V4

(1 + T2/KT2 )− k4T4 + �4T4

dT5 = V5 − k5T5 + �5T5

(22)

hancing transformations for monitoring the process correlationnt.2014.04.006

where Vi are maximal transcription rates, ki degradation rateconstants, Ki inhibition or activation constants and �i error termsdesigned to simulate biological variability. Parameters Vi, ki and

Page 7: Sensitivity enhancing transformations for monitoring the process correlation structure

ARTICLE IN PRESSG ModelJJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of Process Control xxx (2014) xxx–xxx 7

0

0.2

0.4

0.6

0.8

1

R1M

AXNet

R0M

AXNet

VnM

AXNet

R1M

AXX

R0M

AXX

R1M

AXChExt

R0M

AXChExt

VMAX

VnM

AXChExt

VnM

AXX

W

N

F withf ed me

Ka0ebdcc

asowr

ig. 3. Comparison of the statistics performance for case study 1 (network systemault detection curve obtained for all perturbations. Bars correspond to the associat

i were set to unity and the error terms (�i) were sampled from normal distribution with zero mean and standard deviation of.01. The underlying network is represented in Fig. 4. The recov-ry of the undirected network from data was already addressedy Fuente et al. [16] through the use of partial correlations. As theirections of the edges can also be obtained by simple analysis of theross-correlation between the variables, the complete network wasonsidered in the construction of transformations TNet and TNetCh.

The system was subject to perturbations on V1, V5, k1 and k5 in magnitude range of ±20%. The control limits for the monitoring

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

tatistics were adjusted to a false detection rate of 1% under normalperation conditions. Afterwards, 500 sample covariance matricesere computed from 3000 observations from which the detection

ates for each fault were determined. This procedure was repeated

T1

T5

T4

T2

T3

Fig. 4. Directed network representation of the gene network model.

a non-linear model structure): performance index N based on the area under thedian bounded by the 25th and 75th percentiles.

5 times in order to obtain approximate confidence intervals for thefault detention rates. From these results, the performance indexN presented in Fig. 5 was obtained. The results for the permuta-tion tests are available in Table S6 of the supplementary material.Analyzing these results, it is possible to verify that, for this par-ticular system, the current monitoring statistics presented higherdetection rates than RMAX based statistics, regardless of the trans-formation used. This can be explained by analysis of the originaluntransformed data, where only small changes in the correlationmatrix are observed. In fact, the most significant changes occurredin the process mean, since the faults drive the process to a newsteady state, and in a lesser extent in the process’ variance. There-fore, as none of the partial correlation based statistics account forchanges in variance, all of them show poor detection performancesunder these simulated scenarios.

For detecting a change on the variance, the VMAX statisticwas extended to detect both increases and decreases in processvariance. Its direct extension without transformation, VnMAXX,showed to be slightly worse than VMAX in detecting increasesin variance, due to the increase in the control limit. However,when the variance decreases, VnMAXX was capable to signal suchchange while VMAX is unable to detect it at all. ConsequentlyVnMAXX is globally better than VMAX when the detection of bothincreases and decreases are of interest, as can be seen in Fig. 5.The performance of VnMAX can be further improved by use of anappropriated transformation, namely transformations TChExt, TNetand TNetCh, which incorporate dynamic dependencies. For this par-ticular case, the relative performance of VnMAXChExt, VnMAXNetand VnMAXNetCh is very similar (see Fig. 5 and Table S6). This hap-pens because the order of the variables used to implement thetransformation TChExt is consistent with their causal relationships(see Fig. 4) and therefore the model obtained is adequate. How-ever, in the cases where any other variable ordering is used, theperformance of VnMAXChExt changes, while that for VnMAXNet andVnMAXNetCh remains the same since they are not dependent on theparticular order of appearance of the variables.

We would like to point out that perturbations detected byVnMAX are not only related to changes in the variance, but alsoto changes in the process’ structure when transformed variablesare used. In this case, if a change in the process’ structure occurs,

hancing transformations for monitoring the process correlationnt.2014.04.006

the transformation no longer describes the variables’ relationshipscorrectly. Therefore, non-accounted contributions will lead to devi-ations on the transformed variables’ variance, which will ultimatelybe captured by VnMAX. This behaviour shows the importance of

Page 8: Sensitivity enhancing transformations for monitoring the process correlation structure

ARTICLE IN PRESSG ModelJJPC-1761; No. of Pages 11

8 T.J. Rato, M.S. Reis / Journal of Process Control xxx (2014) xxx–xxx

0

0.2

0.4

0.6

0.8

1

VnM

AXChExt

VnM

AXNetCh

VnM

AXNet

W VnM

AXX

VMAX

R0M

AXChExt

R1M

AXChExt

R1M

AXNetCh

R0M

AXNetCh

R1M

AXNet

R0M

AXNet

R1M

AXX

R0M

AXX

N

Fig. 5. Comparison of the statistics performance for case study 2 (gene network model): pall perturbations. Bars correspond to the associated median bounded by the 25th and 75t

Table 4Jöbses’ model parameters.

Parameters Values Unit

KE 0.00383 m6 kg−2 h−1

c1 59.2085 kg m−3

c2 70.5565 kg m−3

KS 0.500 kg m−3

mS 2.160 kg kg−1 h−1

mP 1.10 kg kg−1 h−1

YSX 0.02445 kg kg−1

mw

5

dOi

w(CoCcitpov

In order to compare the monitoring statistics performance, thesystem was subject to changes on c1, KS and YSX. The range of theperturbations was of ±25% for c1 and YSX and ±50% for KS. The con-trol limits of the studied monitoring statistics were set to a false

CXCS

CP

CE

CXCS

CP

(a) (b)

YPX 0.05263 kg kg−1

�max 1.0 h−1

onitoring the variance in order to also detect structural changes,hen using SET.

.3. Case study 3: biologic production of ethanol

This non-linear dynamical system is based in the ethanol pro-uction from glucose fermentation by Zymomonas mobilis bacteria.ne of the models proposed to describe the dynamics of this process

s the Jöbses’ model [38], given by the following equations:

dCS

dt= −

(�maxCSCE

YSX (KS + CS)

)− mSCX + D(CS0 − CS)

dCX

dt=

(�maxCSCE

KS + CS

)+ D(CX0 − CX )

dCE

dt= KE(CP − c1)(CP − c2)

(CSCE

KS + CS

)+ D(CE0 − CE)

dCP

dt= −

(�maxCSCE

YPX (KS + CS)

)− mPCX + D(CP0 − CP)

(23)

here CS is the substract (glucose) concentration, CX is the biomassZymomonas mobilis), CP is the product (ethanol) concentration, andE is an auxiliary variable used to account for the lagged effectf ethanol concentration in the kinetic model. The variables CS0,X0, CE0 and CP0 complete the mass balance representing the inputoncentrations in the reactor, where normally only CS0 (substractnlet) is non-zero. The dilution rate (D) is the inverse of the resident

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

ime. The reactor volume was kept constant and all the remindingarameters proposed by Jöbses et al. [38] are listed in Table 4. Forur case study, only CS, CX and CP were considered to be measuredariables.

erformance index N based on the area under the fault detection curve obtained forh percentiles.

By analysis of the differential equations, the system model rep-resented by Eq. (23), can be translated into the causal networkpresented in Fig. 6(a), by considering that if a variable is presentin the right side of the differential equation, then it has an effect onthe main variable (that appears on the left side). This representationclearly shows a complex relationship between variables, as they areall directly or indirectly linked by closed-loops. The reconstructionof the correct dependency between the variables connective struc-ture has an impact in the monitoring statistics, since the derivationof the sensibility enhancing transformations depends on this map.

All the variable transformations described in Section 4 weredetermined based on a data set collected under normal oper-ation conditions. In what concerns to transformation TChExt, noparticular order was considered for the variables. As for transfor-mations TNet and TNetCh, the algorithm developed for identifyingthe variables edges and dependencies led to the graph shown inFig. 6(b). The differences between the reconstructed network andthe real one (presented in Fig. 6(a)), lie essentially in the fact thatonly one direction is considered together with the absence of theunmeasured variable, CE. However, even with these structural mis-matches, the transformation is able to break the major cross- andauto-correlations present in the original data, and thus it providesa reasonable description of the system for monitoring purposes.

hancing transformations for monitoring the process correlationnt.2014.04.006

Fig. 6. Jöbses’ model network: (a) the original causal network, (b) the estimatedcausal network. Circles represent measured variables and rectangles unmeasuredvariables.

Page 9: Sensitivity enhancing transformations for monitoring the process correlation structure

ARTICLE IN PRESSG ModelJJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of Process Control xxx (2014) xxx–xxx 9

0

0.2

0.4

0.6

0.8

1

R0M

AXNetCh

R1M

AXNetCh

R0M

AXNet

R1M

AXNet

R1M

AXChExt

R0M

AXChEx t

VnM

AXNe t

VnM

AXChExt

VnM

AXNetCh

W VnM

AXX

R0M

AXX

VMAX

R1M

AXX

N

F perforp 75th p

d5Trim

tnoottTaafCicbr

irartiAsTiFou

6

pvod

ig. 7. Comparison of the statistics performance for case study 3 (Jöbses’ model):

erturbations. Bars correspond to the associated median bounded by the 25th and

etection rate of 1% and the detection rates were determined from00 sample covariance matrices based on 3000 observations each.hese simulations were repeated 5 times in order to assess theesults consistency. As in the previous cases study, the performancendex N (Fig. 7) and permutations tests (Table S7 in supplementary

aterial) were computed and saved for analysis.Looking to the current monitoring statistics, the W and G statis-

ics were the ones that presented the best performance. Howeverone of the current monitoring statistics successfully detects faultsn KS and YSX, leading to fault detections rates below 5%. On thether hand, most of the proposed monitoring statistics are ableo signal all the simulated faults. Only the statistics based onhe original variables (TX) presented poor detection capabilities.his result is expected as the TX variables are highly correlated,nd therefore changes are difficult to detect. The effect of cross-nd auto-correlation is eliminated through the use of the trans-ormation TChExt, which includes time-shifted variables on theholesky decomposition. This is the simplest among the three stud-

ed sensitivity enhancing transformations for dynamic systems andonsequently both TNet and TNetCh transformations present evenetter performances, since they make use of more informationegarding the variables’ relationships.

The main difference between transformations TNet and TNetChs the additional Cholesky decomposition performed over the finalesiduals in the case of TNetCh. This extra step ensures that the vari-bles obtained are indeed uncorrelated and it accommodates anyelationship that is not being considered by TNet. This characteris-ic gives some robustness to the proposed transformations, sincet allows some flexibility in the estimation of the causal network.s stated before, the retrieved causal network for this case study isubstantially different from the real one. Yet, with transformationNetCh, the effects of the missed interactions are mitigated and anncrease in the monitoring statistics performance is observed (seeig. 7). This shows that the proposed monitoring statistics basedn this type of SET presents an interesting degree of robustness, aseful feature in real world practical applications.

. Discussion

The statistical process monitoring of process multivariate dis-

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

ersion is mostly done by analysis of either the generalizedariance, likelihood test ratio or other dispersion measures thatnly consider the marginal distributions of process data. Theseescriptions have a rather small resolution of the process inner

mance index N based on the area under the fault detection curve obtained for allercentiles.

structure and therefore such procedures are unable to effectivelydetect localized changes. This feature becomes more relevant whenthe system cannot be assumed to be approximately linear aroundits operation state, as in the case study considered in Section 5.1.

In order to incorporate the systems’ non-linearity and dynamicsin a more rigorous and consistent way, we proposed new sensitivityenhancing transformations that use partial correlations to retrievethe underlying process structure. The application of these transfor-mations lead to a new set of uncorrelated variables that accountfor the process structure and, due to their uncorrelated nature,makes fault detection easier. Moreover, as the system is comparedagainst a fix model, any drift from the reference model can also beperceived as a structural change. These types of faults are essen-tially manifested in changes in variance and can be detected by anextended version of the current monitoring statistics VMAX. Theproposed extension, VnMAX, is capable to detect both increasesand decreases in variance, complementing the RMAX statistics insuch a way that any structural change is detected by at least one ofthem. Therefore, the combined use of these statistics will enable tomonitor the full spectrum of structural changes conveyed by bothVnMAX and RMAX. Yet, on this study, they were treated separatelyto clarify and understand their complementariness.

Another important feature of the sensitivity enhancing trans-formations is their ability to increase the sensitivity to changesin correlation. This increase in sensitivity occurs because aroundthe state of zero correlation, any small change in the variablesrelationship is translated into a high deviation on the correlationcoefficient values. Furthermore, as time-shifted variables and/orpolynomial terms are added to the model, the transformationsalso break the systems dynamics and non-linearity, facilitating thedetection of faults related to these terms. Note that the conven-tional approach only analyses the linear relationships between thevariables and therefore cannot detect more complex relationships.Additionally, the proposed network-oriented transformations arefar superior than the method based on the Cholesky decomposi-tion proposed earlier, where the monitoring statistics performanceis highly dependent on the variables order. However, the new trans-formations require the estimation of the causal network, whichcan be a quite complex task. Still, our study shows that when theregression step is complemented by an additional Cholesky decom-

hancing transformations for monitoring the process correlationnt.2014.04.006

position over the regression residuals, the final model has somerobustness to a miss specification of the true network, namely theidentification of wrong edges and the incorrect attribution of acausal direction.

Page 10: Sensitivity enhancing transformations for monitoring the process correlation structure

ING ModelJ

1 Proces

fwpOfbdL(aamood

ciaMhad

fotbvnStoucbmeara

7

itTsdie

wsVacitaTstt

[

[

[

[

[

[

[

[

[

[

[

[

ARTICLEJPC-1761; No. of Pages 11

0 T.J. Rato, M.S. Reis / Journal of

For multivariate processes, the need of fitting a regression modelor each variable can be seen as a disadvantage. However, it isorth noticing that other similar approaches have already beenroposed, such as Hawking’s regressed-adjusted variables [33] orttestad’s regression components [39]. Other examples can also be

ound in Refs. [40,41]. However, none of these methods are driveny the inner process structure, with the aim of maximizing theetection of correlation drifts. In this study, we applied Ordinaryest Squares regression (OLS), but Principal Component RegressionPCR) or Partial Least Squares Regression (PLSR) could have beenpplied as well to solve potential collinearity problems that mayrise in the regression stage. Still, the resource to such regressionethodologies was not necessary, since the proposed methodol-

gy pre-selects the set of variables to be used as regressors basedn the inferred causal network of the process and these variablesid not present collinearity issues.

The procedures mentioned above do not consider the inner pro-ess structure and even Principal Component Analysis (PCA), whichs extensively used to explain the process variability, ends up with

linear combination of variables that may not be directly related.oreover, the typical multivariate control schemes based on PCA,

ave some limitations in the detection of changes in the correlationnd are easily outperformed by monitoring statistics specificallyevoted for such task, as for instance the W statistic [18].

After application of an appropriate sensitivity enhancing trans-ormation, the monitoring of the marginal correlation by R0MAXr of the 1st order partial correlations by R1MAX, seems to be sta-istically similar. This situation arises from the close relationshipetween these two quantities and also because the transformedariables are inherently uncorrelated, which means that there iso extra effect in controlling any pair of variables by a third one.till, in both cases, the inner process structure is explored, since theransformed variables result from the application of a model basedn partial correlation, which are well-known by their capability toncover the variables relationships. Furthermore, the whole con-ept of partial correlations is based on regression residuals obtainedy controlling two variables by a set of others. Consequently, thearginal correlation of the transformed variables can, to some

xtent, be considered as partial correlations of the original vari-bles, since the transformed variables are in fact the regressionesiduals of the original variables controlled by their parent vari-bles.

. Conclusions

In this work we have addressed the implementation of sensitiv-ty enhancing transformations in statistical process monitoring ofhe process structure by means of partial and marginal correlations.he use of partial correlations allows for a better description of theystems dynamics and non-linearity and consequently, a higheretection performance. Most of this improvement is due to the

ntroduction of process structure information into the sensitivitynhancing transformation.

In order to expand the detection capability of faults relatedith the process structure, we have also proposed the VnMAX

tatistic. This monitoring statistic is a generalization of the currentMAX statistic, and it is able to detect both changes in vari-nce and in the process structure, since model deviations lead tohanges in the transformed variables variance. This result is anotherndication that the sensitivity enhancing transformations increasehe sensitivity to detect changes in correlation, through RMAX,nd simultaneously assess the model validity, through VnMAX.

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

hese two monitoring statistics can also be easily combined into aingle monitoring scheme, able to detect both types of devia-ions and therefore consolidate the detection of structural changeshrough an unified monitoring scheme.

[

PRESSs Control xxx (2014) xxx–xxx

The proposed monitoring statistics also proved to be statisti-cally superior when compared to the current monitoring statisticstested, which makes this approach a feasible solution worthwhileconsidering in practical applications. Future work will encompassthe development of similar monitoring schemes for single obser-vations, instead of the use of rational groups of observations.

Acknowledgments

Tiago J. Rato acknowledges the Portuguese Founda-tion for Science and Technology for his PhD grant (grantSFRH/BD/65794/2009). Marco S. Reis also acknowledges financialsupport through project PTDC/EQU-ESI/108374/2008 co-financedby the Portuguese FCT and European Union’s FEDER through “EixoI do Programa Operacional Factores de Competitividade (POFC)” ofQREN (with ref. FCOMP-01-0124-FEDER-010397).

Appendix A. Supplementary data

Supplementary data associated with this article can be found,in the online version, at http://dx.doi.org/10.1016/j.jprocont.2014.04.006.

References

[1] V.B. Ghute, D.T. Shirke, A multivariate synthetic control chart for process dis-persion, Qual. Technol. Quant. Manage. 5 (3) (2008) 271–288.

[2] A.B. Yeh, D.K.J. Lin, R.N. McGrath, Multivariate control charts for monitor-ing covariance matrix: a review, Qual. Technol. Quant. Manage. 3 (4) (2006)415–436.

[3] B. Abbasi, et al., A transformation-based multivariate chart to monitor processdispersion, Int. J. Adv. Manuf. Technol. 44 (7) (2009) 748–756.

[4] C.-L. Yen, J.-J.H. Shiau, A.B. Yeh, Effective control charts for monitoring multi-variate process dispersion, Qual. Reliab. Eng. Int. 28 (4) (2012) 409–426.

[5] S. Bersimis, S. Psarakis, J. Panaretos, Multivariate statistical process controlcharts: an overview, Qual. Reliab. Eng. Int. 23 (5) (2007) 517–543.

[6] F.B. Alt, Multivariate quality control, in: S. Kotz (Ed.), Encyclopedia of StatisticalSciences, John Wiley & Sons, Inc., New York, 2005, pp. 5312–5323.

[7] M.A. Djauhari, Improved monitoring of multivariate process variability, J. Qual.Technol. 37 (1) (2005) 32–39.

[8] F. Aparisi, J. Jabaloyes, A. Carrión, Generalized variance chart design with adap-tive sample sizes. The bivariate case, Commun. Stat. Simul. Comput. 30 (4)(2001) 931–948.

[9] F.B. Alt, N.D. Smith, Multivariate process control, in: P.R. Krishnaiah, C.R. Rao(Eds.), Handbook of Statistics, Elsevier, 1988, pp. 333–351.

10] W.A. Levinson, D.S. Holmes, A.E. Mergen, Variation charts for multivariate pro-cesses, Qual. Eng. 14 (4) (2002) 539–545.

11] C.-L. Yen, J.-J.H. Shiau, A multivariate control chart for detecting increases inprocess dispersion, Stat. Sinica 20 (2010) 1683–1707.

12] E.L. Russell, L.H. Chiang, R.D. Braatz, Fault detection in industrial processesusing canonical variate analysis and dynamic principal component analysis,Chemometr. Intell. Lab. Syst. 51 (1) (2000) 81–93.

13] A. Negiz, A. C inar, Statistical monitoring of multivariable dynamic processeswith state-space models, AIChE J. 43 (8) (1997) 2002–2020.

14] R.J. Treasure, U. Kruger, J.E. Cooper, Dynamic multivariate statistical processcontrol using subspace identification, J. Process Control 14 (3) (2004) 279–292.

15] J.-P. Pellet, A. Elisseeff, A partial correlation-based algorithm for causal structurediscovery with continuous variables, in: Proceedings of the 7th InternationalConference on Intelligent Data Analysis, Springer-Verlag, Ljubljana, Slovenia,2007, pp. 229–239.

16] A.D.L. Fuente, et al., Discovery of meaningful associations in genomic data usingpartial correlation coefficients, Bioinformatics 20 (18) (2004) 3565–3574.

17] M. Kalisch, P. Bühlmann, Robustification of the PC-algorithm for directed acyclicgraphs, J. Comput. Graph. Stat. 17 (4) (2008) 773–789.

18] T.J. Rato, M.S. Reis, Non-causal data-driven monitoring of the process correla-tion structure: a comparison study with new methods (in submission).

19] D.M. Hawkins, E.M. Maboudou-Tchao, Multivariate exponentially weightedmoving covariance matrix, Technometrics 50 (2) (2008) 155–166.

20] M. Bauer, et al., Finding the direction of disturbance propagation in a chemicalprocess using transfer entropy, IEEE Trans. Control Syst. Technol. 15 (1) (2007)12–21.

21] T. Yuan, S.J. Qin, Root cause diagnosis of plant-wide oscillations using granger

hancing transformations for monitoring the process correlationnt.2014.04.006

causality, in: 8th IFAC International Symposium on Advanced Control of Chem-ical Processes, Singapore, 2012.

22] P.M. Frank, Fault diagnosis in dynamic systems using analytical and knowledge-based redundancy – a survey and some new results, Automatica 26 (3) (1990)459–474.

Page 11: Sensitivity enhancing transformations for monitoring the process correlation structure

ING ModelJ

Proces

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

ARTICLEJPC-1761; No. of Pages 11

T.J. Rato, M.S. Reis / Journal of

23] R. Isermann, Model-based fault-detection and diagnosis – status and applica-tions, Ann. Rev. Control 29 (1) (2005) 71–85.

24] R. Isermann, Process fault detection based on modeling and estimation meth-ods – a survey, Automatica 20 (4) (1984) 387–404.

25] A. Costa, M. Machado, A new chart based on sample variances for monitoringthe covariance matrix of multivariate processes, Int. J. Adv. Manufact. Technol.41 (7) (2009) 770–779.

26] A. Reverter, E.K.F. Chan, Combining partial correlation and an information the-ory approach to the reversed engineering of gene co-expression networks,Bioinformatics 24 (21) (2008) 2491–2497.

27] A.S. Melissa, K.R. Raghuraj, S. Lakshminarayanan, Partial correlation metricbased classifier for food product characterization, J. Food Eng. 90 (2) (2009)146–152.

28] K.R. Rao, S. Lakshminarayanan, Partial correlation based variable selectionapproach for multivariate data classification methods, Chemometr. Intell. Lab.Syst. 86 (1) (2007) 68–81.

29] R. Rao, S. Lakshminarayanan, Variable interaction network based variableselection for multivariate calibration, Anal. Chim. Acta 599 (1) (2007)

Please cite this article in press as: T.J. Rato, M.S. Reis, Sensitivity enstructure, J. Process Control (2014), http://dx.doi.org/10.1016/j.jproco

24–35.30] T.W. Anderson, An Introduction to Multivariate Statistical Analysis, 3rd ed.,

Wiley, New Jersey, 2003.31] D.C. Montgomery, Introduction to Statistical Quality Control, 6th ed., John

Wiley & Sons, New York, 2008.

[

[[

PRESSs Control xxx (2014) xxx–xxx 11

32] E.B. Wilson, M.M. Hilferty, The distribution of chi-square, in: Proceedings ofthe National Academy of Sciences of the United States of America, 1931, pp.684–688.

33] D.M. Hawkins, Regression adjustment for variables in multivariate quality con-trol, J. Qual. Technol. 25 (3) (1993) 170–182.

34] L. Huwang, A.B. Yeh, C.-W. Wu, Monitoring multivariate process vari-ability for individual observations, J. Qual. Technol. 39 (3) (2007)258–278.

35] W.H. Press, et al., Numerical Recipes: The Art of Scientific Computing, 3rd ed.,Cambridge University Press, Cambridge, 2007.

36] F. Pesarin, L. Salmaso, Permutation Tests for Complex Data: Theory, Applica-tions and Software, John Wiley & Sons Ltd., New York, 2010.

37] Y. Tamada, et al., Estimating gene networks from gene expression data bycombining Bayesian network model with promoter element detection, Bioin-formatics 19 (1) (2003) ii227–ii236.

38] I.M.L. Jöbses, H.C.H. Hiemstra, J.A. Roels, Fermentation kinetics ofZymomonas mobilis near zero growth rate, Biotechnol. Bioeng. 29 (4) (1987)502–512.

hancing transformations for monitoring the process correlationnt.2014.04.006

39] P. Ottestad, Component analysis an alternative system, Int. Stat. Rev. 43 (1)(1975) 83–107.

40] I.T. Jolliffe, Principal Component Analysis, 2nd ed., Springer, New York, 2002.41] J.E. Jackson, A User’s Guide to Principal Components, John Wiley & Sons, Inc.,

New York, 1991.