nnsi

12
Substructural identification using neural networks Chung-Bang Yun*, Eun Young Bahng Department of Civil Engineering, Korea Advanced Institute of Science and Technology, Taejon, 305-701, South Korea Received 18 February 1997; accepted 27 July 1999 Abstract In relation to the problems of damage detection and safety assessment of existing structures, the estimation of the element-level stiness parameters becomes an important issue. This study presents a method for estimating the stiness parameters of a complex structural system by using a backpropagation neural network. Several techniques are employed to overcome the issues associated with many unknown parameters in a large structural system. They are the substructural identification and the submatrix scaling factor. The natural frequencies and mode shapes are used as input patterns to the neural network for eective element-level identification particularly for the case with incomplete measurements of the mode shapes. The Latin hypercube sampling and the component mode synthesis methods are adapted for ecient generation of the patterns for training the neural network. Noise injection technique is also employed during the learning process to reduce the deterioration of the estimation accuracy due to measurement errors. Two numerical example analyses on a truss and a frame structures are presented to demonstrate the eectiveness of the present method. 7 2000 Elsevier Science Ltd. All rights reserved. Keywords: Substructuring identification; Neural networks; Modal data; Noise injection learning; Latin hypercube sampling; Com- ponent mode synthesis 1. Introduction Structural identification has become increasingly an important research topic in connection with damage assessment and safety evaluation of existing structures [1–4]. The stiness parameters such as axial and flex- ural rigidities can be identified based on the measured responses and excitations using various system identifi- cation techniques. Such inspection approaches provide nondestructive means of safety assessment. Recent development in this area is made possible by rapid advances in computer technology for data acquisition, signal processing and analysis. In general, the system identification methods inherently involve comprehen- sive search processes such as extended Kalman filter, recursive least squares or modal perturbation. Most of them are computationally expensive and tend to be nu- merically unstable for the case with a large-degrees-of- freedom system. In recent years, instead of the compre- hensive search processes, pattern-matching techniques using neural networks have drawn considerable atten- tion in the field of damage assessment and structural identification [5–12]. Neural networks have unique capability to be trained to recognize given patterns and to classify other untrained patterns. Hence, the neural networks can be used for estimating the structural par- ameters with proper training. Several researchers have dealt with neural network approaches for damage estimation of small degrees-of- Computers and Structures 77 (2000) 41–52 0045-7949/00/$ - see front matter 7 2000 Elsevier Science Ltd. All rights reserved. PII: S0045-7949(99)00199-6 www.elsevier.com/locate/compstruc * Corresponding author. Tel.: +82-42-869-3612; fax: +82- 42-869-3610. E-mail address: [email protected] (C.B. Yun).

Upload: ramprasad-srinivasan

Post on 11-Nov-2015

220 views

Category:

Documents


0 download

DESCRIPTION

substructural identification

TRANSCRIPT

  • Substructural identification using neural networks

    Chung-Bang Yun*, Eun Young Bahng

    Department of Civil Engineering, Korea Advanced Institute of Science and Technology, Taejon, 305-701, South Korea

    Received 18 February 1997; accepted 27 July 1999

    Abstract

    In relation to the problems of damage detection and safety assessment of existing structures, the estimation of theelement-level stiness parameters becomes an important issue. This study presents a method for estimating thestiness parameters of a complex structural system by using a backpropagation neural network. Several techniques

    are employed to overcome the issues associated with many unknown parameters in a large structural system. Theyare the substructural identification and the submatrix scaling factor. The natural frequencies and mode shapes areused as input patterns to the neural network for eective element-level identification particularly for the case withincomplete measurements of the mode shapes. The Latin hypercube sampling and the component mode synthesis

    methods are adapted for ecient generation of the patterns for training the neural network. Noise injectiontechnique is also employed during the learning process to reduce the deterioration of the estimation accuracy due tomeasurement errors. Two numerical example analyses on a truss and a frame structures are presented to

    demonstrate the eectiveness of the present method. 7 2000 Elsevier Science Ltd. All rights reserved.

    Keywords: Substructuring identification; Neural networks; Modal data; Noise injection learning; Latin hypercube sampling; Com-

    ponent mode synthesis

    1. Introduction

    Structural identification has become increasingly an

    important research topic in connection with damage

    assessment and safety evaluation of existing structures[14]. The stiness parameters such as axial and flex-

    ural rigidities can be identified based on the measured

    responses and excitations using various system identifi-

    cation techniques. Such inspection approaches providenondestructive means of safety assessment. Recent

    development in this area is made possible by rapid

    advances in computer technology for data acquisition,

    signal processing and analysis. In general, the systemidentification methods inherently involve comprehen-

    sive search processes such as extended Kalman filter,recursive least squares or modal perturbation. Most ofthem are computationally expensive and tend to be nu-

    merically unstable for the case with a large-degrees-of-freedom system. In recent years, instead of the compre-hensive search processes, pattern-matching techniquesusing neural networks have drawn considerable atten-

    tion in the field of damage assessment and structuralidentification [512]. Neural networks have uniquecapability to be trained to recognize given patterns and

    to classify other untrained patterns. Hence, the neuralnetworks can be used for estimating the structural par-ameters with proper training.

    Several researchers have dealt with neural networkapproaches for damage estimation of small degrees-of-

    Computers and Structures 77 (2000) 4152

    0045-7949/00/$ - see front matter 7 2000 Elsevier Science Ltd. All rights reserved.PII: S0045-7949(99 )00199-6

    www.elsevier.com/locate/compstruc

    * Corresponding author. Tel.: +82-42-869-3612; fax: +82-

    42-869-3610.

    E-mail address: [email protected] (C.B. Yun).

  • freedom models. Wu et al. applied the neural networksto recognize the locations and the extent of individual

    member damage of a simple three story frame [5].Szewczyk and Hajela proposed a modified counterpro-pagation neural network to estimate the stiness of in-

    dividual structural elements based on observable staticdisplacements under prescribed loads [6]. Their pilotstudies showed that the neural-network-based

    approach for damage assessment is promising. Tsouand Shen identified damage of a spring-mass systemwith eight degrees-of-freedom (DOF) that has closely

    spaced eigenvalues [8]. They constructed three multi-layer subnets which performed three tasks of inputpattern generation, damage location identification, anddamage severity determination. Their study showed

    good performance for on-line damage identification.Yagawa et al. applied neural network approach to esti-mate stable crack growth in welded specimens [9].

    Yoshimura et al. proposed the Generalized-Space-Lat-tice transformation of training patterns to relax the ill-posedness of the problem. As the eect of transform-

    ation, a training process converged fast, and estimationaccuracy was improved significantly [10].In this study, the neural network-based approach is

    extended to the estimation of stiness parameters of acomplex structural system with many unknowns. Sev-eral techniques, such as substructural identification[1315] and submatrix scaling factor [16] are employed

    to overcome the issues associated with manyunknowns. The concept of modal strain energy [17] isutilized for selection of modes to be used as input to

    the neural networks, and Latin hypercube samplingtechnique [18] and the component mode synthesismethod [19] are employed for ecient generation of

    training patterns. Two numerical example analyses ona truss and a frame structures are presented to demon-strate the eectiveness of the proposed method.

    2. Neural network-based structural identification

    Studies on neural networks have been motivated toimitate the way that the brain operates. It has recentlydrawn considerable attention in various fields of

    science and technology, such as character recognition,electro-communication, image processing, and indus-trial control problems [20]. Many researchers have

    developed various neural network models for dierentpurposes [21]. In this study, a model called backpropa-gation neural network (BPNN) or multilayer percep-

    tron is used for structural identification. The BPNNconsists of an input layer, hidden layers, and an outputlayer as in Fig. 1. The input and output relationship ofa neural network can be nonlinear as well as linear,

    and its characteristics are determined by the weightsassigned to the connections between nodes in two adja-cent layers. Changing these weights will change the

    input/output behavior of the network. Systematic waysof determining the weights of the network to achieve adesired input/output relationship are referred to as

    training or learning algorithm. In this study, the stan-dard backpropagation algorithm is used. The basicstrategy for developing a neural network-based

    approach to identify a structural system is to train theBPNN to recognize the element-level structural par-ameters from the measurement data on the structuralbehavior, such as natural frequencies and mode

    shapes.

    2.1. Substructural identification

    For the identification of a structure with manyunknowns, it is not practical to identify all of the par-

    ameters in the structure at a same time, because mostof the identification techniques require expensive com-putation that would be even prohibitive, as the numberof unknown parameters increases. Several researches

    were reported on identification of a part of a structureso as to reduce the size of the system under consider-ation [1315]. Those works were based on the reason-

    ing that the expected damages of a structure occur atseveral critical locations, hence it is more reasonable toconcentrate the identification at critical locations of

    the structure.The local-identification method is generally based on

    substructuring, in which the structure is subdivided

    Fig. 2. Substructuring for localized identification.Fig. 1. Architecture of backpropagation neural network.

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 415242

  • into several substructures and the identification is car-ried out on a substructure at a time. In the present

    study, the substructure to be identified is called as theinternal substructure, while the others as the externalsubstructures as shown in Fig. 2. Since the parameters

    to be estimated are limited to an internal substructure,it is expected that the numerical problems such asdivergence or falling into local minima may be

    avoided. Another advantage is that this approachrequires measurement data only on the substructure ofinterest, instead of on the whole structure.

    The two-stage identification is another idea toreduce the number of variables. The two-stage identifi-cation was also studied by the present authors [22]. Inthe study, Topole and Stubbss damage index method

    [23] was employed to select probable damaged mem-bers in the first stage identification. Then the damageseverities of the selected members were estimated using

    the neural network technique in the second stageidentification. However, there may be many caseswhere we can estimate which region shall be investi-

    gated through visual inspection or knowledge ofexperts without carrying out the first stage identifi-cation. Then the substructural identification can be

    performed on the interested part of the structure, whilethe parameters of the other parts are still unknown. Inthe worst case any information may not be obtained todetermine the substructure to be identified, then the

    substructuring technique can be applied to each sub-structure in turn.

    2.2. Submatrix scaling factors

    The number of the parameters to be estimated shall

    be kept reasonably small for successful identificationof a structure. Reduction of parameters will improvethe results of the identification and minimize the

    required measurement data. For the purpose of e-cient parameterization of the structure, the submatrixscaling technique is adopted in this study [16]. Usingthe technique, the stiness matrix of the system can be

    described by introducing the submatrix scaling factors(SSF) corresponding to the element level stinessmatrices as

    K Xqj1

    Kj Xqj1

    sjK0j 1

    where K is the stiness matrix of the structure in the

    present condition, Kj and K0j are the present and the

    reference values of the stiness submatrix for the jthelement transformed into the global coordinates. sj is

    the j-th SSF, and q is the total number of submatrices.Then, the identification of the stiness matrix can beperformed by estimating the submatrix scaling factors

    instead of all of the stiness matrix coecients. Directidentification of the system stiness matrix coecients

    may face with problems which violate the basic proper-ties of the stiness matrix related to the member con-nectivity, boundary condition, and those as

    symmetricity and positive definiteness. However, theuse of the SSFs described in Eq. (1) can avoid suchproblems.

    2.3. Input to neural networks

    A way of choosing the patterns representing the

    characteristics of the structure, which are to be used asthe input to neural networks, is one of the most im-portant subjects in this approach. Several researchershave used the various input patterns (input vectors)

    suitable for their purpose. For example, Wu et al. usedthe frequency spectrum for each DOF of the examplestructure for damage estimation [5]. This input has an

    advantage that does not need the modal parameteridentification from the measurements. Nonetheless, tocharacterize those spectral properties, many sampling

    data are required. Accordingly, a large number ofinput neurons in the neural networks are needed,which may reduce the eciency and accuracy of the

    training process. Tsou and Shen used the dynamic re-sidual vector that can be obtained from the modaldata [8]. It provides a simple and eective way todetect the damage and the length of the input pattern

    can be reduced significantly compared with the spec-trum data. However, it still has the restriction that themodes should be measured at every finite element

    DOF. In reality, the modes obtained from a typicalmodal survey are generally incomplete in the sensethat the mode vector coecients are available only at

    the measured DOFs. Therefore, it is desirable to usethe input patterns which are more suitable for the casewith partial measurement data.

    The natural frequencies and modes of a structureare used as the input patterns in this study. The choicehas been made based on the following advantages: (1)the length of the input pattern is limited to several

    times of the number of the measured DOFs; (2) thenatural frequencies represent global behaviors, whilethe mode vectors do local characteristics; and (3) they

    can be obtained from the measurements of the struc-tural behavior. The relationship between the modesand the stiness matrix can be written as

    KF MFL 2where M is the mass matrix, F is the modal matrix,and L is the diagonal matrix containing the eigen-values. The mass matrix is assumed to be knownexactly in this study.Using the substructuring technique and referring to

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 4152 43

  • Fig. 2, the modal matrix F can be partitioned as

    F 24FL1 FH1FL2 FH2FL3 F

    H3

    35 3where the subscripts 1, 2, and 3 denote substructures,

    while Substructure 2 is the internal substructure, andthe superscripts L and H represent lower and highermodes, where only the lower modes are generally used

    for the structural identification. Rewriting the ithmode in the partitioned modal matrix FL2 as ~ji for thesimplicity, the input pattern can be defined as

    Input pattern vector f fi, ~j1i, . . . , ~jni , i 1, . . . ,

    mg4

    where fi is the ith natural frequency, ~jji denotes the jthcomponent of ~ji which is normalized as ~j

    Ti ~ji 1, n is

    the number of DOFs measured for Substructure 2, andm is the number of modes to be included in the identi-fication.

    A useful relationship between the submatrix for theelement stiness and the modes may be obtained byexamining the strain energy of each submatrix withrespect to the modes [17]. The modal strain energy dis-

    tribution among the stiness submatrices for the ithmode can be evaluated as

    bji jTi K

    0j ji

    jTi Mji

    1

    o2i5

    where bji is the modal strain energy (MSE) coecientof submatrix K0j in the ith mode, ji is the ith mode inthe modal matrix F, and oi is the ith circular naturalfrequency. The submatrices containing large values ofthe MSEs for a mode indicate that the corresponding

    structural elements are major load carrying elementsfor the particular mode. Stiness changes in those el-ements will cause significant changes in the frequency

    and mode shape. Hence, detection of the stinesschanges may be more eective by examining the modaldata for those elements. On the other hand, the struc-tural elements having negligible MSEs reflect that it

    would be dicult to detect the stiness changes inthose elements using the corresponding modal data.Topole and Stubbs utilized the changes in the MSEs

    from the undamaged to the damaged structures to pre-dict structural damage [23]. In the present study, how-ever, the MSEs are used as indicators in selecting

    important vibration modes for the input patterns. TheMSEs distribution gives a good information for select-ing the modal data.

    2.4. Generation of training and testing patterns

    Since the neural network-based structural identifi-

    cation is highly dependent on the training patterns, itis very important to prepare well-examined data sets.The number of training patterns must be large enough

    to represent the structural system properly. However,for the computational eciency, it must be limited tobe reasonably small, because most of the compu-

    tational time for this approach is required for prepar-ing the training patterns and training the networks.When the number of the unknown parameters is P

    and each parameter has L sampling points, the size ofthe whole population is LP. This size increases expo-

    nentially, as the number of the unknown parametersdoes. To reduce the number of training patterns eec-tively, a sparse sampling algorithm such as Latin

    hypercube technique [18] is employed in this study,which makes the required number of samples reducedfrom LP to L.

    The training patterns for the proposed neural net-work-based method consist of the modal data as inputand the corresponding SSFs as output. To generate

    training patterns, a series of eigenanalyses are to beperformed. However, repeated computations of eigen-value problems for a complex structure can be very ex-

    pensive. Thus, the fixed-interface-component-mode-synthesis (CMS) method [19] incorporating the sub-structuring technique is employed. In the component

    mode synthesis method, each substructure is analyzedindependently for local mode shapes. Then, the com-

    ponent mode shapes are assembled to construct Ritzvectors, which are subsequently used to compute themode shapes for the whole structure. When the num-

    ber of DOFs is large, the computational eciency ofthe CMS method becomes more significant.The neural-network approach is an approximation

    technique that handles non-unique cases by eitherreturning one of the possible solutions or an averagetaken over all possible solutions [6]. Although the

    input information to the neural network is limited tothe components of the mode vectors for an internalsubstructure by applying the substructuring technique

    to the structural identification, most of the informationregarding the stiness parameters of the internal sub-structure is included in the components of the mode

    vectors of the internal substructure, rather than thoseof the external substructures. Thus, those stiness par-

    ameters may be estimated with a reasonable accuracyfrom the components of mode vectors of the internalsubstructure as the input to the neural network. On

    the other hand, as the range of the submatrix scalingfactors in the external substructure reduces, the esti-mation accuracy increases. If possible, the uncertainties

    of the unknowns in the external substructure shall bereduced by using all the available information.

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 415244

  • In the process of computing the mode vectors foreach pattern, the mode shapes may be obtained in

    dierent order compared with the baseline case, par-ticularly for the higher modes. In this study, the modeshapes of each pattern are rearranged to match the

    baseline shapes using the modal assurance criteria(MAC) [24], in which the MAC is defined as

    MACij j ~jTi ~jTj j2j ~jTi ~ji k ~jTj ~jj j

    6

    For instance, to select the mode matching to the ithbaseline shape ~ji , the mode ~jj corresponding to thelargest MACij shall be chosen. Additionally, the direc-

    tion (sign) of the mode shape must be kept same asthe corresponding baseline shape. The condition canbe satisfied, if the inner product of the mode and its

    baseline is always kept positive.

    2.5. Noise injection learning

    A crucial problem on the BPNN is its generalizationcapability. Usually, training patterns used for learningare only a limited number of samples selected from a

    population of input and output patterns. Hence, a net-work successfully trained to a given set of samplesmay not provide the desired input and output associ-

    ations for untrained patterns, particularly in the caseswith measurement noises and errors in the mode esti-mation. Concerning this problem, several researchers

    reported that adding some noise to input patternsduring backpropagation learning process can remark-ably enhance the generalization capability of the resul-tant networks, if the mapping from the input space to

    the output space is smooth [25,26]. Matsuoka foundthat the noise injected into the input reduces the sensi-tivity of the network to the variation of the input [25]:

    that is, it makes the mapping from the input space tothe output space be smooth.In this study, noises are artificially added to natural

    frequencies and mode shapes in generating the trainingand the testing patterns as

    f ni fi1 vi 7

    ~jnji ~jji1 vji 8

    where f ni and ~jnji are the ith natural frequency with

    noise and the j-th component of the corresponding

    modes with noise, and vi and vji are Gaussian randomnoises. The sequence of noise is generated by using asubroutine which returns a series of random numbers

    normally distributed with zero mean and a specifiedstandard deviation. Based on Eqs. (7) and (8), thenoise level can be defined as the value of the standard

    deviation. For instance, if the standard deviation is0.03, the noise level in the modal data can be referred

    as 3% in the root-mean-square (RMS) level.

    3. Illustrative numerical examples

    Two dierent types of structures were selected for

    example analyses to illustrate the applicability of theproposed approach. They are a two-span planar trussand a multi-storey frame structure.

    3.1. Truss structure

    A truss example is the two-span planar truss whichcontains 55 truss elements, 24 nodes and 44 nodal

    DOFs as shown in Fig. 3. It was assumed that thebaseline parameters were known. Values for the ma-terial and geometric properties are as follows: the elas-

    tic modulus=200 GPa; the cross-sectionalarea=0.03 m2; and the mass density=8000 kg/m3. Asthe first step of the identification procedure, the sti-

    ness matrix of the structure is represented as

    K X55j1

    sjK0j 9

    Then, the problem is to identify 55 unknown subma-trix scaling factors, sj, j = 1, . . . , 55, based on the

    measured modal data. Instead of identifying the wholeSSFs simultaneously, the structure was subdivided intothree substructures; an internal substructure and two

    external substructures as shown in Fig. 3. It wasassumed that the unknown SSFs for the elements inthe internal substructure were between 0.3 and 1.7,while those for the external substructures were between

    0.5 and 1.5. The mode shapes were assumed to bemeasured only at 10 test DOFs, which include the dis-placements in the x- and y-directions at five nodes in

    the internal-substructure.

    3.1.1. Modal data for input to BPNN

    For ecient training process, appropriate modaldata shall be selected as the input vector. It wasassumed that the first five modes of the internal-sub-structure were available at the test DOFs. The natural

    frequencies and mode shapes of the reference structure,in which all the SSFs are unity, are shown in Fig. 4.The first, second and fifth modes are found to be verti-

    cal, i.e., flexural modes, the third is a mixed mode withthe vertical and horizontal motions, and the fourth is ahorizontal mode. The MSE distribution in the first five

    modes of the reference structure is shown in Fig. 5. Itshows that the Elements 1, 2, 4, 5, 7, and 10 are carry-ing large fractions of the strain energy in most of

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 4152 45

  • modes. Especially in the third and fourth modes whichcontain the horizontal motions, the horizontal elementssuch as the Elements 2, 5, 7, and 10 have relativelylarge values of MSE. On the other hand, the MSE

    coecients associated with the Elements 3, 6, 8, 9, and11 are relatively small for all of the five modes, whichindicates that the accuracy of the SSF estimation for

    the elements using the first five modal data may bepoor. Based on the above observation, the first fourmodal data were selected as an input pattern for this

    example. However, other cases with dierent combi-nations of modes were also tested for the purpose ofcomparison.

    3.1.2. Training and testing patternsFor the generation of training patterns using the

    Latin hypercube sampling, 15 sampling points between

    0.3 and 1.7 were selected with an equal interval of 0.1for each SSF in the internal substructures, assumingthat each SSF is uniformly distributed. On the other

    hand, 11 sampling points between 0.5 and 1.5 weretaken with an interval of 0.1 for the SSFs in the exter-

    nal substructures. Consequently, the number of thepossible sample cases is 165 (=15 11), which wasnot judged to be enough to represent the system ade-

    quately. Hence, 10 iterations of the Latin hypercubesampling process were taken, and the 1650 training

    patterns were sampled. For the cases of testing pat-

    terns, the sampling points for the internal substructurewere taken as the same with those for the training

    data, however the number of sampling points for theexternal substructures was taken to be 10 in the range

    of 0.551.45 with an interval of 0.1. The dierences in

    sampling points were imposed, because the real testingpatterns are not necessarily the same as the training

    patterns. Thus, the number of possible sample cases

    for testing was 150 (15 10).Using the CMS method, natural frequencies and

    mode shapes were computed for sampled values of the

    SSFs. In the process of computing the mode vectorsfor the sampled training and testing patterns and re-

    arranging those in order of the baseline, the mode

    shapes of some patterns may be very dierent withthose of the baseline. Accordingly, those patterns were

    eliminated from the training and testing patterns.

    Actually 1200 of the sampled training patterns wereused for training process and 100 of the sampled test-

    ing patterns were used for testing process. The elapsetime for generating whole training data set was about

    53 sec by the Alpastation 2004/233 workstation. The

    variations of the SSFs in the external substructureswere considered in preparation for the training data

    set, though they were not included in the output vec-

    tor. Therefore, the output vector of the BPNN consistsof the SSFs in the internal substructure, while the

    input vector consists of the first four natural frequen-

    cies and the mode shape components for the internalstructure. The training and testing data set were pre-

    pared for the cases with 0, 3, 5, and 7% noise in RMSlevel. An example of training patterns was shown in

    Table 1.

    Fig. 3. Substructuring for localized identification.

    Fig. 4. First five mode shapes of a truss structure.

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 415246

  • 3.1.3. Training and testing processes

    It is important to choose the proper network size. If

    the network is too small, it may not be able to rep-

    resent the system adequately. On the other hand, if the

    network is too big, it becomes over-trained and may

    provide erroneous results for untrained patterns. In

    general, it is not straight forward to determine the best

    size of the networks for a given system. It may be

    found through trial and error process using knowledge

    about the system. A four-layer neural network as

    shown in Fig. 1 was selected for the present example.

    The number of the neurons in the input layers varies

    depending on the number of modes included: i.e. 11

    neurons for each mode. However, the numbers of the

    neurons in the first hidden, the second hidden and the

    output layers were taken to be fixed to 17, 13, and 11,

    respectively as shown in Fig. 1. Numerical investi-

    gations show that small changes of the numbers of the

    neurons in the hidden layers have little eect on the

    estimation results.

    Fig. 5. Modal strain energy distribution for internal substructure of a truss structure.

    Table 1

    One example of training patterns

    (a) Input pattern without noise

    5.352 0.362 0.340 0.050 0.268 0.113 0.002 0.299 0.290 0.495 0.4977.594 0.045 0.080 0.076 0.064 0.071 0.187 0.386 0.374 0.572 0.57110.415 0.347 0.345 0.321 0.418 0.468 0.121 0.175 0.196 0.283 0.31712.913 0.032 0.242 0.358 0.637 0.591 0.208 0.086 0.022 0.036 0.071

    (b) Input pattern with 3% noise in RMS

    5.727 0.358 0.333 0.050 0.268 0.105 0.003 0.299 0.279 0.496 0.5117.637 0.049 0.080 0.074 0.064 0.069 0.207 0.357 0.371 0.576 0.58110.967 0.327 0.328 0.318 0.430 0.485 0.122 0.159 0.209 0.305 0.29512.349 0.028 0.259 0.334 0.615 0.622 0.212 0.075 0.021 0.034 0.067

    (c) Output pattern (SSFs, sIj )

    0.5 1.5 1.7 0.3 0.9 1.1 0.6 1.6 1.1 1.1 1.6

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 4152 47

  • The architecture of the neural network and therequired number of training patterns are inter-related.

    However, there is no established method to determinethe relationship. According to the VapnikChervonen-kis dimension theory [21], the required number of

    training patterns may be roughly taken as two times ofthe number of weights. In this study, however, thenumber of the training patterns was taken as 1200

    through trial and error, which is approximately thenumber of weights including threshold (1153) as shownin Fig. 1. It is found that the values of the weightsafter the learning process are well-distributed without

    severe saturation. It means that the present architec-ture is suitable and the number of the training patternsis sucient.

    The training process took about 1000 epochs tolearn the pattern representation using the standard BP

    algorithm, in which the elapse time was about 220 s byan Alpastation computer system. Generally the train-ing process is a time consuming work. However, once

    the training is completed, the neural networks can beused very eciently and repeatedly for identifying thesystem parameters. Hence, it can be a very useful tool

    for structural health monitoring.Each of the training patterns was used once for

    training at an epoch, then the order of the trainingpatterns is randomly altered at the next epoch. An

    example of the tested outputs was listed in the Table 2.Fig. 6 shows the averages of the relative errors in theestimated SSFs for various cases with dierent number

    Table 2

    One example of test outputs

    NN Input 5.28 0.35 0.33 0.03 0.22 0.12 0.00 0.30 0.29 0.50 0.50( fj, ~jji) 8.01 0.03 0.14 0.07 0.11 0.07 0.08 0.30 0.27 0.62 0.61

    10.63 0.24 0.28 0.27 0.56 0.59 0.00 0.05 0.07 0.20 0.2311.92 0.15 0.12 0.20 0.76 0.52 0.04 0.06 0.20 0.02 0.09

    NN output sIj ) 1.47 0.89 1.44 0.90 1.13 0.37 0.35 1.27 0.45 0.62 1.52Target output 1.60 0.90 1.40 0.90 1.10 0.40 0.30 1.20 0.50 0.60 1.70

    Relative error (%)a 7.65 0.26 2.98 0.22 3.06 5.32 19.67 5.89 8.10 4.79 10.44

    a Relativeerror 100 jNN outputTarget outputTarget output

    j.

    Fig. 6. Average estimation errors for dierent modes used in input patterns (for testing data set without noise).

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 415248

  • of modes included in the input patterns. It can be

    found that the accuracy of the estimation improveswith increasing number of the modes included until thefourth mode. But adding the fifth mode did not

    improve the estimates. Hence, the first four modeswere included in the input to the BPNN.The eect of the measurement errors in the testing

    data set and the eectiveness of the noise injection

    learning were also investigated. Fig. 7 shows the esti-mation errors for the SSFs for dierent cases with var-ious noise levels in the training and testing data sets.

    The noises in the training patterns mean the noisesinjected to the input patterns during the training pro-cess, and those in the testing patterns indicate the

    measurement errors. Fig. 7 shows that the average esti-mation error for the noise-free testing data wasobtained as 8.9%, when the neural network has beentrained using the noise-free training data set. But, the

    average error for the testing data with 5% noise usingthe same neural network was found to be 18.3%,which implies that the measurement error causes de-

    terioration in the estimation. On the other hand, whenthe network has been trained using the training data

    set with 5% noise, the error for the testing data with5% noise was obtained as 13.5%, which is much smal-ler than the error using the noise-free training data.

    This indicates that it is desirable to use the 5% noisedata for training, if the measurement error is expectedto be 5%. As expected from the MSE distribution, the

    estimation errors for the Elements 3, 6, 8, 9, and 11are larger than those for the other elements. However,the results generally lie in a satisfactory range. Table 3shows the similar results for the cases with 3 and 7%

    measurement noises. The results also indicate that theestimation accuracy can be improved significantly, ifnoise with an intensity similar to the measurement

    Fig. 7. Estimation error for dierent noise injection learning.

    Table 3

    Average estimation errors (%) for dierent noise injection

    levels

    Noise levels in RMS In testing data

    0% 3% 5% 7%

    In training data

    0% 8.9 12.8 18.3 22.4

    3% 9.7 11.1 14.0 15.8

    5% 10.4 11.4 13.5 14.8

    7% 12.0 12.6 13.7 14.7

    Fig. 8. Learning curves without and with noise injection in

    training patterns (for testing data set with 7% noise).

    Fig. 9. Two-bay and ten-storey frame structure.

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 4152 49

  • noise level is imposed in the training process. Fig. 8 is

    the learning curves to depict how the noise injectionlearning works. The learning rate is relatively high atthe epochs from 1 to 100, and after that, it becomes

    lower. It can be observed from the curves that thenoise injection improves the results of estimationduring the learning process. Generally, the back-propa-

    gation algorithm requires some reasonable criteria tostop the weight adjustment. The number of learningepochs was used as the termination criterion in this

    paper, and it was taken as 1000 epochs, because thelearning curves showed that the decreasing rate of thetesting error became very flat after the epoch as inFig. 8. The level of the testing error may be used as a

    practical termination criterion, too.

    3.2. Frame structure

    The second example is a frame structure as shown inFig. 9. The planar frame structure contains 50 el-

    ements, 33 nodes and 90 nodal DOFs. The baselineparameters were listed in Table 4. The SSFs of the in-ternal substructure were estimated for twelve elements,

    marked in Fig. 9. The natural frequencies and modeshapes of the frame structure are shown in Fig. 10.The four measurement cases in Fig. 11 were investi-

    gated on the eect of rotational and translational com-ponents of mode shapes. The first four modes wereused as input patterns to the neural networks; 1200training and 100 testing patterns were actually used

    during the learning process, similarly with the trussexample.Fig. 12 shows the average estimation errors for four

    cases with dierent measured components of the mode

    vectors. When all the DOFs at 9 nodes were used asthe input (Case 1), the estimation accuracy was good.However, when only the x- and y-DOFs were included

    in the input pattern (Case 2), the estimation accuracybecame poor. On the other hand, when the transla-tional DOF at the mid point of each element in the in-

    ternal substructure was added into the input pattern(Case 3), the estimation accuracy was found to beimproved significantly, which indicates that the transla-

    tional data at the additional nodes may make up forthe information on the unmeasured rotational DOFs,which are usually dicult to be measured in the fieldexperiment. Additionally when the translational DOFs

    at two quarter points of each element were added intothe input pattern (Case 4), the estimation accuracy wasfound to be better than the case using the rotational

    DOFs (Case 1). The average relative errors for thetraining and testing data set with 5% noise were 8.8and 9.8%, respectively. It means that the generaliz-

    ation performance of the present neural networks isvery good.

    4. Conclusions

    A neural network-based substructural identificationwas presented for the estimation of the stiness par-ameters of a complex structural system, particularly

    for the case with noisy and incomplete measurement ofthe modal data. The proposed approach does notrequire any complicated formulation for model re-

    duction of the system, which is indispensable to theconventional approaches for substructural identifi-cation. On the other hand, this approach requires gen-eration of enough training patterns which can

    adequately represent the relationship between the sti-ness parameters and the modal information. The Latinhypercube sampling and the component-mode-syn-

    thesis methods were used for the ecient generation ofsuch training patterns.The numerical example analyses were carried out

    on a two-span truss and a multi-storey frame, andthe results were summarized as follows: (1) The sub-structuring technique and the concept of the subma-Fig. 10. First four mode shapes of a frame structure.

    Table 4

    Baseline parameters for frame examplea

    Shape Area (m2) I (m4) Density (kg/m3)

    Beams W24 15 1.04 102 5.62 104 7850Columns W14 145 2.78 102 7.12 104 7850

    a I is moment of inertia of the cross-sectional area; E is modulus of elasticity at 210 GPa.

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 415250

  • trix scaling factor were found to be very ecient to

    reduce the number of unknown stiness parameters

    to be estimated. (2) The modal strain energy coe-

    cients were eective indicators in selecting vibration

    modes for the input to the neural networks. (3) In

    the truss structure, the horizontal elements were

    major load carrying members for lower modes,

    hence the accuracy of the parameter estimation for

    the elements was better than the others. (4) For the

    case of the frame structure, the estimates became

    deteriorated, when the rotational DOFs at the nodes

    were not included in the input to the neural net-

    works. However, the estimation accuracy could be

    improved significantly, when additional measurement

    information on the translational DOFs within the el-

    ements was utilized. (5) The noise injection learning

    using noise at similar level to the measurement error

    was found to be very eective to improve the esti-

    mation accuracy. (6) The average relative estimation

    errors for testing data set with various noise intensi-

    ties were found to be in the range of 915%, which

    shows the applicability of the present method for the

    identification of large structural systems.

    Fig. 11. Cases of measured DOFs (sensor locations).

    Fig. 12. Average estimation error for dierent measured

    DOFs with 5% measurement noise (w: Case 1, +: Case 2, Q:Case 3, R: Case 4).

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 4152 51

  • Acknowledgements

    The authors would like to thank to the KoreaEarthquake Engineering Research Center and theKorea Science and Engineering Foundation for their

    financial support for this study.

    References

    [1] Natke HG, Yao JT-P. System identification approach in

    structural damage evaluation. In: Proceedings of ASCE

    Structural Congress 86, Preprint 17-1, 1986.

    [2] Hong K-S, Yun C-B. Improved method for frequency

    domain identification of structures. Engng Struct

    1993;15(3):17988.

    [3] Ghanem R, Shinozuka M. Structural-system identifi-

    cation I: theory. J Engng Mech ASCE 1995;121(2):255

    64.

    [4] Yun C-B, Lee H-J, Lee C-G. On sequential prediction

    error method for structural identification. J Engng Mech

    ASCE 1997;123(2):115122.

    [5] Wu X, Ghaboussi J, Garrett Jr JH. Use of neural net-

    works in detection of structural damage. Comput Struct

    1992;42(4):64959.

    [6] Szewczyk Z, Hajela P. Neural network based damage

    detection in structures. ASCE Journal of Computing and

    Civil Engineering 1994;8(2):16378.

    [7] Manning RA. Structural damage detection using active

    members and neural networks. AIAA Journal

    1994;32(6):13313.

    [8] Tsou P, Shen M-HH. Structural damage detection and

    identification using neural networks. AIAA Journal

    1994;32(1):17683.

    [9] Yagawa G, Matsuda A, Kawate H, Yoshimura S.

    Neural network appoach to estimate stable crack growth

    in welded specimens. International Journal of Pressure

    Vessels and Piping 1995;63:30318.

    [10] Yoshimura S, Matsuda A, Yagawa A. New regulariz-

    ation by transformation for neural network based inverse

    analyses and its application to structure identification.

    International Journal of Numerical Methods in

    Engineering 1996;39:395368.

    [11] Okuda H, Miyazaki H, Yagawa G. A neural network

    approach for modeling of viscoplastic material behaviors.

    Advanced Computer Applications, ASME/PVP

    1994;274:1415.

    [12] Pandey PC, Barai SV. Multilayer perceptron in damage

    detection of bridge structures. Comput Struct

    1995;54(4):597608.

    [13] Koh CG, See LM, Balendra T. Estimation of structural

    parameters in the time domain: a substructure approach.

    Earthquake Engng Struct Dyn 1991;20:787801.

    [14] Oreta AWC, Tanabe T-A. Element identification of

    member properties of framed structures. J Struct Engng

    ASCE 1994;120(7):196176.

    [15] Yun C-B, Lee H-J. Damage estimation of structures

    incorporating substructural identification. In:

    Proceedings of Asian-Pacific Symposium on Structural

    Reliability and its Applications (APSSRA), Tokyo

    Japan, 1995. pp. 220228.

    [16] Lim TW. Submatrix approach to stiness matrix correc-

    tion using modal test data. AIAA Journal

    1990;28(6):112330.

    [17] Lim TW. Structural damage detection using modal test

    data. AIAA Journal 1991;227174.

    [18] Press W, Teukolsky SA, Vetterling WT, Flannery BP. In:

    Numerical recipes in C The art of scientific comput-

    ing. New York: Cambridge University Press, 1992. pp.

    315.

    [19] Craig Jr RR, Bampton MCC. Coupling of substructures

    for dynamic analysis. AIAA Journal 1968;6(7):13139.

    [20] Haykin S. Neural networks A comprehensive foun-

    dation. New York: Macmillan, 1994.

    [21] Hush DR, Horne BG. Progress in supervised neural net-

    works. IEEE Signal Processing Magazine, 1993. p. 839.

    [22] Yun C-B, Bahng EY. Neural network approach to

    damage assessment of civil structure. In: Proceedings of

    the Structural Engineers World Congress (SEWC 98),

    San Francisco, California, USA, 1923 July 1998, Paper

    reference: T1312, 1998.

    [23] Topole KG, Stubbs N. Non-destructive damage evalu-

    ation of a structure from limited modal parameters.

    Earthquake Engng Struct Dyn 1995;24:142736.

    [24] Ewins DJ. In: Modal testing: theory and practice.

    England: Research Studies Press, 1986. pp. 2245.

    [25] Matsuoka K. Noise injection into inputs in back-propa-

    gation learning. IEEE Trans Systems, Man and

    Cybernetics 1992;22(3):43640.

    [26] Holmstrom L, Koistinen P. Using additive noise in back-

    propagation training. IEEE Trans Neural Networks

    1992;3(1):2438.

    C.-B. Yun, E.Y. Bahng / Computers and Structures 77 (2000) 415252