utilizing similarity information in industrial...

82
UNIVERSITATIS OULUENSIS ACTA C TECHNICA OULU 2009 C 317 Heli Koskimäki UTILIZING SIMILARITY INFORMATION IN INDUSTRIAL APPLICATIONS FACULTY OF TECHNOLOGY, DEPARTMENT OF ELECTRICAL AND INFORMATION ENGINEERING, UNIVERSITY OF OULU C 317 ACTA Heli Koskimäki

Upload: others

Post on 20-Jan-2021

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

ABCDEFG

UNIVERS ITY OF OULU P.O.B . 7500 F I -90014 UNIVERS ITY OF OULU F INLAND

A C T A U N I V E R S I T A T I S O U L U E N S I S

S E R I E S E D I T O R S

SCIENTIAE RERUM NATURALIUM

HUMANIORA

TECHNICA

MEDICA

SCIENTIAE RERUM SOCIALIUM

SCRIPTA ACADEMICA

OECONOMICA

EDITOR IN CHIEF

PUBLICATIONS EDITOR

Professor Mikko Siponen

University Lecturer Elise Kärkkäinen

Professor Hannu Heusala

Professor Olli Vuolteenaho

Senior Researcher Eila Estola

Information officer Tiina Pistokoski

University Lecturer Seppo Eriksson

Professor Olli Vuolteenaho

Publications Editor Kirsti Nurkkala

ISBN 978-951-42-9038-1 (Paperback)ISBN 978-951-42-9039-8 (PDF)ISSN 0355-3213 (Print)ISSN 1796-2226 (Online)

U N I V E R S I TAT I S O U L U E N S I SACTAC

TECHNICA

U N I V E R S I TAT I S O U L U E N S I SACTAC

TECHNICA

OULU 2009

C 317

Heli Koskimäki

UTILIZING SIMILARITY INFORMATION IN INDUSTRIAL APPLICATIONS

FACULTY OF TECHNOLOGY,DEPARTMENT OF ELECTRICAL AND INFORMATION ENGINEERING,UNIVERSITY OF OULU

C 317

ACTA

Heli K

oskimäki

C317etukansi.fm Page 1 Wednesday, February 11, 2009 1:43 PM

Page 2: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki
Page 3: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

A C T A U N I V E R S I T A T I S O U L U E N S I SC Te c h n i c a 3 1 7

HELI KOSKIMÄKI

UTILIZING SIMILARITY INFORMATION ININDUSTRIAL APPLICATIONS

Academic dissertation to be presented, with the assent ofthe Faculty of Technology of the University of Oulu, forpublic defence in Auditorium TS101, Linnanmaa, onMarch 13th, 2009, at 12 noon

OULUN YLIOPISTO, OULU 2009

Page 4: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Copyright © 2009Acta Univ. Oul. C 317, 2009

Supervised byProfessor Juha Röning

Reviewed byProfessor Xiaohui LiuProfessor Heikki Mannila

ISBN 978-951-42-9038-1 (Paperback)ISBN 978-951-42-9039-8 (PDF)http://herkules.oulu.fi/isbn9789514290398/ISSN 0355-3213 (Printed)ISSN 1796-2226 (Online)http://herkules.oulu.fi/issn03553213/

Cover designRaimo Ahonen

OULU UNIVERSITY PRESSOULU 2009

Page 5: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Koskimäki, Heli, Utilizing similarity information in industrial applicationsFaculty of Technology, Department of Electrical and Information Engineering, University ofOulu, P.O.Box 4500, FI-90014 University of Oulu, Finland Acta Univ. Oul. C 317, 2009Oulu, Finland

AbstractThe amount of digital data surrounding us has exploded within the past years. In industry, data aregathered from different production phases with the intent to use the data to improve the overallmanufacturing process. However, management and utilization of these huge data sets is notstraightforward. Thus, a computer-driven approach called data mining has become an attractiveresearch area. Using data mining methods, new and useful information can be extracted fromenormous data sets.

In this thesis, diverse industrial problems are approached using data mining methods based onsimilarity. Similarity information is shown to give an additional advantage in different phases ofmanufacturing. Similarity information is utilized with smaller-scale problems, but also in abroader perspective when aiming to improve the whole manufacturing process. Different ways ofutilizing similarity are also introduced. Methods are chosen to emphasize the similarity aspect;some of the methods rely entirely on similarity information, while other methods just preservesimilarity information as a result.

The actual problems covered in this thesis are from quality control, process monitoring,improvement of manufacturing efficiency and model maintenance. They are real-world problemsfrom two different application areas: spot welding and steel manufacturing. Thus, this thesisclearly shows how the industry can benefit from the presented data mining methods.

Keywords: data mining, manufacturing, process data, real-world application, similarity

Page 6: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki
Page 7: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Acknowledgements

This study was carried out in the Intelligent Systems Group (ISG) at the Department of

Electrical and Information Engineering of the University of Oulu, Finland, between the

years 2004 and 2009.

I am very grateful to Professor Juha Röning and Doctor Perttu Laurinen, who sup-

ported and encouraged me throughout this work. I also thank Professor Heikki Mannila

from Helsinki University of Technology and Professor Xiaohui Liu from Brunel Uni-

versity for reviewing this manuscript.

My colleagues from the Intelligent Systems Group also deserve my profound thanks.

Special thanks go to Eija Haapalainen, who was one of my best friends from the day

we began to work in the same room and helped keep me sane during these years. Also

Pekka Siirtola, Henna Mörsäri and many other colleagues have been friends and not

just people sharing the same workplace. I could not have dreamed of a better place to

work.

This work was financially supported by Infotech Graduate School, The National

Technology Agency of Finland, the European Union, the Kaupallisten ja teknillisten

tieteiden tukisäätiö and the Tauno Tönning Foundation. I gratefully acknowledge their

support. In addition, I would like to thank all the partners I have cooperated with during

these years.

I could not have been able to finish this thesis without my good friends. Although

they have not contributed to the research part of the thesis, they have always been there

for me. Especially I would like to thank Tuire, who helped me clear my head many

times. My mother and father have always let me search for my own path and have

encouraged me when necessary. Finally, I would like to thank my husband, Timo, who

has made my life complete.

Oulu, March 2009 Heli Koskimäki

5

Page 8: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

6

Page 9: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Abbreviations

APEN Average prediction error of the neighborhood

ASSRN Average of the squared standardized residuals in the neighborhood

C Current

CBR Case-based reasoning

F Force

HWH Harms + Wende GmbH & Co.KG

IEEE Institute of Electrical and Electronic Engineers

ISG Intelligent Systems Group

kA Kiloamper

KDD Knowledge discovery in databases

knn k nearest neighbor

MAE Mean absolute prediction error

mm Millimeter

PCA Principal component analysis

robL robustified Negative log-likelihood

PCB Printed circuit board

SBT Stanzbiegetechnik

SOM Self-organizing map

V Voltage

wknn Weighted k nearest neighbor

7

Page 10: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

8

Page 11: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

List of original publications

I Laurinen P, Junno H*, Tuovinen L & Röning J (2004) Studying the Quality of ResistanceSpot Welding Joints Using Bayesian Networks. Artificial Intelligence and Applications:705–711.

II Junno H*, Laurinen P, Tuovinen L & Röning J (2004) Studying the Quality of ResistanceSpot Welding Joints Using Self-Organising Maps. Fourth International ICSC Symposiumon Engineering of Intelligent Systems.

III Zettel D, Sampaio D, Link N, Braun A, Peschl M & Junno H* (2004) A Self OrganisingMap (SOM) Sensor for the Detection, Visualisation and Analysis of Process Drifts. PosterProceedings of the 27th Annual German Conference on Artificial Intelligence: 175–188.

IV Junno H*, Laurinen P, Haapalainen E, Tuovinen L, Röning J , Zettel D, Sampaio D, LinkN & Peschl M (2004) Resistance Spot Welding Process Identification and InitializationBased on Self-Organising Maps. 1st International Conference on Informatics in Control,Automation and Robotics: 296–299.

V Koskimäki H, Laurinen P, Haapalainen E, Tuovinen L & Röning J (2007) Application of theExtended knn Method to Resistance Spot Welding Process Identification and the Benefits ofProcess Information. IEEE Transactions on Industrial Electronics 54(5): 2823–2830.

VI Koskimäki H, Juutilainen I, Laurinen P& Röning J (2008) Detection of Correct Moment toModel Update. Lecture Notes in Electrical Engineering - Informatics in Control, Automa-tion and Robotics: 87–94.

VII Koskimäki H, Juutilainen I, Laurinen P & Röning J (2008) Two-level Clustering Approachto Training Data Instance Selection: a Case Study For the Steel Industry. International JointConference on Neural Networks: 3044–3049.

*Koskimäki née Junno

9

Page 12: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

10

Page 13: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Contents

Abstract

Acknowledgements 5

Abbreviations 7

List of original publications 9

Contents 11

1 Introduction 13

1.1 Similarity utilization within this work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

1.2 Contributions of the thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

2 Background 19

2.1 Data mining in industrial applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.1.1 Data set selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

2.1.2 Case-based reasoning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

2.1.3 Similarity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

2.1.4 Model adaption . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

2.2 Application areas of this thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

2.2.1 Resistance spot welding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

2.2.2 Steel manufacturing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

3 Methods 33

3.1 k nearest neighbors classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

3.2 k-means clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

3.3 Self-organizing maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

3.4 Bayesian belief networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

3.5 Similarity aspect of the methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

4 Utilizing similarity 39

4.1 Quality control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

4.1.1 Data set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

4.1.2 Quality control using Bayesian networks . . . . . . . . . . . . . . . . . . . . . . . . . 40

4.1.3 Quality control using Self-organizing maps . . . . . . . . . . . . . . . . . . . . . . 42

4.2 Process monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .45

4.3 Improvement of manufacturing efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

4.3.1 Data set and features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

11

Page 14: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

4.3.2 Process identification using self-organizing maps . . . . . . . . . . . . . . . . . 49

4.3.3 Extension for process identification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

4.3.4 Ways to improve efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

4.4 Model maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

4.4.1 Data set and regression models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .57

4.4.2 Finding a correct moment for model update . . . . . . . . . . . . . . . . . . . . . . 59

4.4.3 Selecting a suitable subset for training data . . . . . . . . . . . . . . . . . . . . . . .62

5 Summary and conclusions 67

References 68

Original publications 75

12

Page 15: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

1 Introduction

In recent years the amount of digital information around us has exploded. In everyday

life digital pictures, documents and music files are stored in computers. On the other

hand, in industry, huge databases have been developed to store information gathered

from manufacturing processes. The intent is to use this data, for example, to improve the

production process or to assess product quality. However, the main question is how to

handle all this data and find desired correlations from the data; for example, the human

eye cannot find correlations between multidimensional data and measured quality. Thus,

a computer-driven approach called data mining (also called knowledge discovery) has

become an attractive research area. The term can be defined as the nontrivial extraction

of implicit, previously unknown, and potentially useful information from data (Frawley

et al. 1992). Data mining research also addresses management of huge amounts of data.

In industry, different kinds of information need to be mined from gathered data. For

example, the intent can be to determine the quality of separate tasks, to monitor pro-

cesses or to use the information to improve the whole production process. In practice,

this means research in the area is concentrated on finding solutions to specific problems.

Several articles have been published concerning solutions to industrial problems using

computer learning methods. They have been used in manufacturing fault detection,

quality control, decision support and many other problems (Harding et al. 2006). The

approaches are scattered, although due to the diversity of these tasks, a universal solu-

tion is not even possible. However, some generic principles can be adapted to several

different research problems. In this thesis, an idea based on similarity of data points is

utilized in diverse industrial problems. The problems not only come from two different

application areas, they also vary inside the application areas. Similarity is utilized in

smaller-scale problems like quality control of individual spot welding joints, but also in

a broader perspective when aiming to improve the whole manufacturing process. On the

other hand, the way similarity information is utilized and corresponding methods are

used varies between these tasks, but the benefits of the similarity approach are clearly

shown in every case.

13

Page 16: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

1.1 Similarity utilization within this work

This thesis introduces how similarity information can be utilized intelligently in indus-

trial applications. Similarity between data points and sets is shown to provide valuable

information in different kinds of data mining problems. The applications introduced

here are distinct, but together they clearly show the benefits of this approach. The

problems covered are from quality control, process monitoring, improvement of manu-

facturing efficiency and model maintenance.

In many practical applications the quality of the final product is of greatest interest

(Harding et al. 2006). In addition, in most cases actual quality cannot be measured

directly, but has to be assessed using process-related data. This is an area for data

mining solutions. The approach introduced in this thesis constitutes finding similar

cases, studying their quality and using this information to assess the quality of unknown

data. In addition, using our approach, parameter settings that constitute low-quality

products can be seen directly.

In process monitoring, similarity is utilized when developing models for control-

ling ongoing processes. In the model training phase, data from different process states

are divided into areas comprised of similar observations whose quality is known. An

optimal area is then identified and in the control phase the similarity of every new ob-

servation is compared with this optimal area. If the new observation is not classified

as similar, a drift is noticed in the process. The similarity information is also used to

enable continuation of high-quality production. In practice, this means adjustment of

machining parameters based on this similarity information.

Improvement of manufacturing efficiency is also approached by utilizing similarity.

In this study data sets obtained from different processes are compared. New data are

compared with data already collected from different processes, the most similar process

among them is identified (process identification) and detailed information on the most

similar process is retrieved from an archive. The two main benefits achieved from

knowing the most similar process are: the quality of new products can be predicted and

improved using the stored information of a similar process and the time needed to set

up a new process can be substantially reduced by restoring process parameters that lead

to high-quality results. These benefits make the overall manufacturing process more

efficient, with time, money and effort being saved from the beginning of production.

In model maintenance, information within similar data of new observations is uti-

lized to decide if the models in use should be retrained. This is because collection of

14

Page 17: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

new kinds of data and running time decrease the reliability of the models, which can

in turn cause economic losses to manufacturers. Thus, having a method for deciding

when to update the models becomes an important step in improving modeling in the

long run. The optimal moment is sought based on estimation accuracies of the models

of the most similar data points and the actual number of similar data points. If there

are plenty of similar data points the allowed prediction errors are narrowed down, while

if there are only a few similar data points the prediction error could be higher without

affecting the need for model retraining. On the other hand, similarity information is

also utilized to reduce the amount of data needed to train the models without losing in-

teresting pieces of information. This is an important part of model maintenance, since

if the amount of gathered data increases it can become infeasible to use all the data, for

example, because of the memory and calculation capacities of the data mining methods.

Training data selection is based on a cluster structure where similar data points consti-

tute their own clusters. Using the cluster structure, the number of similar data points

can be reduced while rarer observations are all selected into the training data.

1.2 Contributions of the thesis

This section introduces the contributions of this thesis. The previous section described

how the similarity approach can be utilized in four different applications: quality con-

trol, process monitoring, improvement of manufacturing efficiency and model mainte-

nance. The main contribution of this thesis is to show that similarity information can be

utilized in diverse phases of manufacturing processes. The phases considered are from

steel manufacturing to joining different steels together, and from single quality control

tasks to improvement of whole manufacturing processes.

The other contributions are in novel research ideas and in the development of solu-

tions to be applied to real manufacturing problems. These contributions are introduced

next, one publication at a time. The contributions of the author of this thesis are dis-

cussed in the last paragraph.

This thesis consists of seven publications.

Publications I and II describe quality assessment of a spot welding process. In

them Bayesian networks and self-organizing maps are applied for the first time to spot

welding quality control. The methods were chosen because they do not make strong

assumptions about the distribution of the data and thus adjust easily to real data (the

conditional independence assumption often made with Bayesian networks is not con-

15

Page 18: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

sidered to be of high importance from the application point of view). In addition, the

results of the methods are simple to understand even without any profound knowledge

of the methods themselves, which is desirable when aiming to develop systems for use

in industrial production lines.

Publication III introduces a novel idea of using self-organizing maps to discover

process drift in spot welding. When a drift in a manufacturing process is noticed online,

certain counteractions can be performed immediately, enabling better quality during

the whole process. In addition, the publication also introduces a way to perform the

counteractions.

Publications IV and V deal with process identification. In process identification the

idea is to gather data from different manufacturing processes, and when collecting new

data, to retrieve a similar data set from a database and utilize the information of the

retrieved process. After a similar process is identified, two different steps can be taken:

(1) restore the process parameters leading to high-quality products and substantially

reduce the time needed to set up a new process or (2) seek the quality prediction and

improvement methods already developed for the similar process. In addition, a specific

similarity measure was developed to decide if the retrieved process is actually similar

enough that the information can be used. In the spot welding industry, no comparable

approach aiming to improve the whole manufacturing process could be found. Thus,

the process identification task introduced in this thesis is considered an important novel

step for the spot welding industry.

Publications VI and VII concentrate on maintenance activities for models. For ex-

ample, collection of new data and the passing of time decrease the reliability of models.

This can be solved by retraining models at optimally selected intervals. Publication VI

introduces a novel approach to finding the optimal time for model updating. Despite

an extensive search by the author, no previous research reported in the area of model

updating in a real application, where updating was developed without a periodic ap-

proach, was found. Thus, the developed approach can be assumed to be a significant

contribution to the area of model efficiency continuum. Publication VII describes a

new procedure for selecting a suitable subset of instances for model training so that the

reduction does not negatively affect the accuracy of the model. In addition, a novel

clustering method called coarse clustering is introduced in publication VII. Both publi-

cations demonstrate the effectiveness of the proposed method by using them with pre-

dictive regression models developed to optimize steel manufacturing process settings

beforehand.

16

Page 19: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

In publications where the author of this thesis is presented as the first author, the

study and discussion are mainly direct contributions of the author, while the studies

were planned and in the writing phase the contents of the publications were revised by

the other authors. In two of the publications the author of this thesis was not the first

author. However, in publication I the study was done by the author, although the results

were mainly reported by P. Laurinen. In publication III the basic idea was presented by

the author, although the study and discussion were done by the other authors.

17

Page 20: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

18

Page 21: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

2 Background

This chapter introduces the background for this thesis. It contains a short overview of

the main activity in the area of data mining and of applications in the manufacturing

industry in the 21th century; areas related the most to this thesis. In addition, the sub-

tasks of data mining related to the thesis are discussed. Data set selection, which is

very important in every practical application, is covered in subsection 2.1.1. Research

done by utilizing previous cases (case-based reasoning) and the background of similar-

ity are introduced in subsections 2.1.2 and 2.1.3, respectively. In addition, subsection

2.1.4 includes a short discussion about the adaptation capability of models. Section 2.2

introduces work done in two application areas: spot welding and steel manufacturing.

2.1 Data mining in industrial applications

As computers became more common and the amount of data gathered increased, an

idea for finding suitable patterns from data using a computer-driven approach was intro-

duced. In business, data mining was thought to give a competitive advantage (Frawley

et al. 1992). In a workshop of the U.S. National Science Foundation held in 1990, data

mining was positioned as one of the most promising research areas of the 1990s (Sil-

berschatz et al. 1990). It has not failed these expectations; data mining conferences are

held worldwide, hundreds of research papers are published in many journals and books

have been written, not only on actual data mining methods, but also on the background

support of these methods.

The term data mining was at some time thought of as a pattern recognition step

of the KDD process (knowledge discovery in databases) (Fayyad et al. 1996), which

is depicted in Figure 1. However, nowadays the term ’data mining’ is understood to

cover the whole process and can be applied to any data source, not just databases. The

synonym ’knowledge discovery’ is also used to describe the whole process (Kurgan &

Musilek 2006). The methods used in the data mining process consist of summarization,

classification, dependency modeling, clustering, link analysis and sequence analysis

(Wang 1999).

19

Page 22: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Fig 1. The steps of data mining process (Fayyad et al. 1996).

In manufacturing, data mining solutions are used to model, classify and make predic-

tions in numerous applications (Harding et al. 2006). The application areas covered

with data mining approach are (Wang 2007):

1. optimization of manufacturing yield,

2. quality control,

3. monitoring and diagnosis,

4. preventive machine maintenance,

5. manufacturing process control,

6. process planning and scheduling,

7. material requirement planning,

8. manufacturing system modeling,

9. safety evaluation,

10.assembly selection,

11.learning robotics.

The first six application areas are most related to this thesis. The main aim in every man-

ufacturing application is to make a profit, meaning that optimization of manufacturing

yield is the driving force behind every application. Of the individual application areas,

quality control is approached in publications I and II, and monitoring and diagnosis as

well as preventive machine maintenance and manufacturing process control are covered

in publication III. Item 6 is related to publications IV and V. Publications VI and VII

20

Page 23: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

are not directly related to any of the applications, although in them also, optimization

of manufacturing yield is the final aim. Thus, the study carried out in publications VI

and VII can be considered separated from the application outline.

Computer-aided quality control in industrial applications aims to check the qual-

ity of products in cases where quality measurement is expensive, unreasonable for hu-

mans or even impossible without breaking the final product. The machine learning

approaches most commonly used consist of neural networks, fuzzy logic, evolution-

ary algorithms and different classification methods. In addition, the application areas

tackled are spread out to widely cover different areas of manufacturing. Studies have

been conducted to estimate the quality of complex electronic products, for example,

sound speakers (Melin & Castillo 2002) and PCB boards (Yang & Tsai 2002). Also

quality in different areas of manufacture in basic industry are covered; many scientific

publications discuss the areas of pulp and paper (Achiche et al. 2006), steel (Wiltschi

et al. 2000) and wood (Silven et al. 2003) manufacturing. On the other hand, quality

assessment methods have been developed to control certain stages of production. For

example, the quality of individual spot welding joints (Xinmin et al. 2007) and the qual-

ity of the painting process (Filev 2002) are evaluated individually in car assembly. In

addition, many visual inspection methods have been developed to monitor the quality

of products on an assembly line, where human concentration can be distracted from the

task from time to time (Piuri et al. 2005).

In monitoring and diagnosis the main aim is to keep the manufacturing process ef-

ficient in the long run by diagnosing system failures and avoiding process disorders.

It is also related to preventive machine maintenance and process control. The solu-

tion approaches are divided into three different categories: rule-based, case-based and

model-based diagnostic systems (Hur et al. 2006). In the rule-based solution faults are

diagnosed using certain rules that tell the correlation between monitored equipment con-

dition and each possible fault. In case-based systems historical maintenance informa-

tion is utilized. In model-based systems different mathematical, statistical and machine

learning methods are used to diagnose faults based on the structure and properties of

the systems used. In data mining the solution is usually a combination of available in-

formation and machine learning methods. For example, in the semiconductor industry

faults can be detected using the k nearest neighbors method. The training data consist

of normal situations, and faulty situations are detected if the distance from a new case

to the normal cases is too big (He & Jin 2007). There are also studies where the sys-

tem is based on sensors and programmable logic controllers, and the data set for fault

21

Page 24: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

identification is gathered from them (Hou et al. 2003). In addition, plenty of effort has

been used for very specific aspects of process monitoring, for example, Rehorn et al.

(2005) review more than 100 scientific articles on solely tool condition monitoring sys-

tems in conventional cutting operations, including drilling, turning, end milling and

face milling.

Manufacturing process planning and scheduling are widely approached using case-

based reasoning methods. There are studies where artificial intelligence is utilized to

plan and schedule whole assembly lines, for example, in car manufacturing (Gusikhin

et al. 2007). However, this review emphasizes how the processes are initialized to pro-

duce high-quality products from the beginning. Thus, the research done to find optimal

setup parameters for production machinery is introduced in more detail. Sadoyan et al.

(2006) sought suitable parameters for the spray tooling process in tool and die making

using if/then rules. Tsao (2008) studied the correct setup value for the thrust force

of a step drill based on the step angle, the stage ratio, the feed rate and the spindle

speed. In addition, genetic algorithms have been used to select the optimal parameters

of a mold oscillation system in the continuous casting process of steel manufacturing

(Bhattacharya et al. 2007).

One feature is common to all the introduced manufacturing applications. The data

mining solutions are developed for specific problems and it is unlikely that the solutions

could be adapted easily to different applications. Thus, from the industrial point of view,

there is still plenty of research to be done. In addition, there are new application areas

where data mining is not yet utilized. Even in the traditional application areas there

is still a need for deeper technical solutions. The model’s maintenance activities and

manufacturing efficiency improvement introduced in this thesis are examples of such

deeper solutions.

2.1.1 Data set selection

As data mining methods are becoming more common, more interest is focused on the

background of these methods. Data preparation is considered to be essential for success-

ful mining applications (Pyle 1999). For example, the learning capability of machine

learning models depends heavily on the features given to the models.

The term ’feature identification’ is used to describe the search for suitable features.

It can be divided into two sometimes overlapping categories: feature extraction and

feature selection. In feature extraction the features affecting the problem are searched

22

Page 25: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

for. Features can be, for example, certain individual measurement values and/or, as

in many cases where the data consist of measurement signals, certain statistical or ge-

ometrical features that represent the signals (Zhang et al. 2006). These features can

be reworked, for example, using PCA to uncorrelate them or using feature selection

methods to choose a subset of features that are the most suitable for the problem. For

example, Haapalainen et al. (2006) introduced different methods for selecting the most

suitable feature set. The aim can be to minimize the size of the feature set (makes

the algorithms used work faster and/or the memory requirements smaller) or improv-

ing prediction accuracy. In many cases the smallest possible feature set that does not

significantly decrease prediction accuracy is chosen (Dash & Liu 1997).

When the available data set is considered as a matrix where the features are the

columns of the matrix and the observations constitute the rows, a distinct approach

comparable to feature selection is easy to rationalize. In feature selection the problem

of an overwhelming amount of data is tackled by removing meaningless columns from

the data set. In some cases the data set reduction has to be performed by deleting

insignificant rows from the data set (Liu & Motoda 2002). When the suitable features

are found but the number of available data points increases, it is difficult to use all the

data in data mining applications. The procedure in which the number of data points is

reduced is called instance selection, though instance selection is a less-treated research

problem than feature identification.

Although the question of instance selection is tackled in several articles from the

classification point of view (Brighton & Mellish 1999, 2002, Jankowski & Grochowski

2004) very few articles consider instance selection with other kinds of problems. In

addition, most of the methods introduced in classification-oriented articles are not ap-

plicable when class labels are not available. The few instance selection studies found

that are adaptable to a case where class labels are unavailable have solved the prob-

lem using genetic algorithms (Kim 2006) or with an approach based on tuning random

sampling. For example, a data set is selected using random sampling and tuned by

replacing random observations of the selected data with observations left out if the re-

placement improves the model’s accuracy (Wilson & Martinez 2000). However, in most

cases this approach would be impractical and highly time-consuming. Nevertheless, ap-

proaches based on k-means clustering have given promising results in the classification

area (Rokach et al. 2003, Altinçay & Ergün 2004) and they can also be adapted to a

case where class labels are not available.

23

Page 26: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

2.1.2 Case-based reasoning

Utilization of similar data sets is recognized as an important aspect in many different

research areas. In many of these studies the problem is referred to as case-based reason-

ing. Case-based reasoning (CBR) is an approach where new problems are solved using

prior experience (Lopez de Mantaras et al. 2006). It can also be considered a method

of data mining.

The first steps in case-based reasoning were taken by Roger Schank in 1982 and

Janet Kolodner, a researcher in Schank’s group. Kolodner developed the first actual

case-based reasoning system, CYRUS, in 1983. Elsewhere, Bruce Porter brought dif-

ferent aspects to the development of case-based reasoning in 1986 and the system de-

veloped using his research was called PROTOS (Aamodt & Plaza 1994).

Aamodt (Aamodt & Plaza 1994) has described CBR as a cyclical process of four

Rs:

1. RETRIEVE the most similar case or cases

2. REUSE the information and knowledge in that case to solve the problem

3. REVISE the proposed solution

4. RETAIN the parts of this experience likely to be useful for future problem solving.

Figure 2 shows the task-method decomposition of case-based reasoning. Although in

the decomposition the tasks are considered subtasks of case-based reasoning, many of

these tasks themselves are actually independent research areas, and their introduction

would not be sensible from the point of view of this thesis. Thus, the discussion con-

centrates on actual use cases in industrial applications.

24

Page 27: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Fig 2. A task-method composition of CBR (Aamodt & Plaza 1994).

The case-based reasoning approach has been used in many different industrial appli-

cations. Case-based reasoning is recognized to be a functional aid in many industrial

design tasks. For example, to reduce design costs, Boyle et al. (2004) used CBR for

fixture design. Woon et al. (2005) used CBR in the design of conveyors, for example,

to decide on suitable input parameters. Case-based reasoning systems have also been

found suitable in oil and gas well design (Kravis & Irrgang 2005). In addition, the case-

based reasoning approach is used for process planning of automotive panels (Chen et al.

2005).

A second area in industry where case-based reasoning methods are found effective

is quality improvement. In the metal industry case-based reasoning is used in quality

designing, where the most similar case is retrieved using the k nearest neighbors method

combined with fuzzy logic (Kim et al. 2004), and in defining a suitable temperature

value for blowing control using only the k nearest neighbors method (Kim et al. 2005).

The k nearest neighbors method is also used as a retrieval method for fault diagnosis

of industrial robots (Olsson et al. 2004) and to control water quality and billing in

the water supply industry (Saward 1999). In addition, a case-based reasoning system

is used to control an electric furnace for reduction of copper from slag (Moczulski &

Szulim 2004).

25

Page 28: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

2.1.3 Similarity

In case-based reasoning the solutions are based on similar previous events. Similarity

information can offer huge benefits in other data mining methods, also. To specify

similarity information, the points to be concentrated on are the features used to define

similarity, the similarity measures and the domain knowledge available. It should be

noted that their selection depends on the method used and the form of data set available.

When considering similarity from the point of view of suitable features, surface

similarity is used in many cases. In surface similarity the features are presented as

vectors or attribute-value pairs. Other commonly used methods are k nearest neighbors

(the number of cases to be retrieved is predefined, but the actual similarity of the cases

to the target problem varies) or different window methods (the number of cases to be

retrieved varies, but the similarity to the target problem is inside a predefined threshold)

(Lopez de Mantaras et al. 2006).

When using the k nearest neighbors method for retrieval of similar cases, the fea-

tures are assigned different weights. It has been shown that classification accuracy can

be improved by limiting the number of prototypes and by weighting features (Skalak

1994). Different methods have been developed to automatically decide the proper

weights for features, for example by Cardie (1993), Moore & Lee (1994), Skalak (1994)

and Aha & Bankert (1994). Many of the automated weight assignment methods prop-

erly give low weights to irrelevant features. However, by using performance feedback

the results improve compared with methods that do not use feedback (Wettschereck &

Aha 1995).

Feature weighting is also with other methods. Munoz-Avila & Hüllen (1996) intro-

duced a method where two cases are considered similar if a certain predefined number

of features match regardless of what the features are. However, because some of these

features are more important than others, they have to be weighted differently.

In addition, the similarity measures that are used should be considered. In the k

nearest neighbors method, for example, an Euclidean or Mahalanobis distance measure

is used to define the nearest neighbors (Mitchell 1997). These measures give a numeric

value to similarity. Also ordinal and Boolean values for similarity can be used (Osborne

& Bridge 1996). The similarity measure is normally chosen according to the problem

(Zhang et al. 2006).

Another very important aspect in similarity is deciding when the most similar is

not similar enough. In other words, when the new data point or data set is so different

26

Page 29: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

compared with the data already gathered that it could be categorized as a novel event.

This task is usually known as novelty detection.

Selection of a method for novelty detection depends on the statistical properties of

the data set. Parametric methods assume that the data distribution is Gaussian and the

problem can be modeled statistically based on data means and covariance (Markou &

Singh 2003). However, their effectiveness depends more on the fact that the data can

only support simple decision boundaries such as linear or quadratic (Hastie et al. 2001).

Also non-parametric approaches such as nearest neighbor, Parzen density and string

matching methods can be used for novelty detection (Markou & Singh 2003).

2.1.4 Model adaption

In some studies, for example (Gabrys et al. 2005), model adaptation has been consid-

ered as the model’s ability to learn behavior in areas from which information has not

been acquired. However, in this thesis, adaptation of the model is considered to be the

ability to react to time-dependent changes in the modeled causality. In practice, model

adaptation means retraining the model at optimally selected intervals. However, be-

cause the system has to adapt quickly to a new situation in order to avoid losses to the

industry, periodic retraining, used in many methods ((Haykin 1999), (Yang et al. 2004)),

is not considered to be the best approach. Moreover, there are also disadvantages if re-

training is done unnecessarily. For example, extra work is needed to take a new model

into use in the actual application environment. In the worst case, this can result in cod-

ing errors that affect the actual accuracy of the model. In spite of extensive literature

searches, studies that would be comparable with the approach were not found. Thus, it

can be assumed that the approach is new, at least in an actual industrial application.

2.2 Application areas of this thesis

This section introduces the background of the application areas where the approach of

this thesis is utilized. The next subsections explain what has been done in the areas

using soft computing and why data mining is so important. Although the introduced

studies from the application areas have not separately tackled the question of similarity,

it generally forms the basis of every modeling task. In model training it is assumed

that the available data represent the whole data space and thus the capability of the

models to estimate new data points is related to the existence of similar data points in

27

Page 30: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

the training data.

2.2.1 Resistance spot welding

Resistance welding is the most common way to join metal objects together. The most

common type of resistance welding, developed by Professor Elihu Thomson, is called

resistance spot welding. It is widely used, for example, in electrical and automotive

industries, because it is effective and economical. In addition, the spot welding process

is easy to automate (Rautaruukki 2002).

Figure 3 shows an example of spot welding. Two metal sheets are welded together.

Electrode force is applied to hold the sheets tightly together, and electrical current flows

through the electrodes and the material. The resistance of the material being welded is

much higher than the resistance of the electrodes. Thus, enough heat is generated to

melt the metal. The pressure on the electrodes forces the molten spots in the two pieces

of metal to unite, forming the final spot (nugget).

Fig 3. Principle of spot welding (Publication V).

The most important thing in this application area is to assure that the spots are produced

with good quality. Because the actual quality can only be measured using destructive

testing, meaning breaking the actual product, it can be done only randomly, and there is

a need for data mining solutions to ensure quality. Also important to manufacturers is

how to produce good-quality spots in cases where the product is made for the first time

or there is no clear information on how to produce it.

28

Page 31: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Research on computational quality assessment techniques in the field has concen-

trated on estimating the quality of welding by using neural networks, regression analysis

and mathematical methods. The studies have utilized different features extracted from

data. The variation of resistance over time (dynamic resistance pattern) has been an

important explanatory variable in many of the studies. Artificial neural network and

regression models have been generated based on the dynamic resistance pattern by, for

example, Aravinthan et al. (2001) and Cho & Rhee (2002, 2004). Cho & Rhee (2002)

compared regression analysis with neural networks, which demonstrated the superior

accuracy of the neural network estimator. Unfortunately, the sample consisted of only

60 measurements, and the leave-one-out method was used to measure the estimator’s

performance, which limits the significance of the conclusions. Also the new approach

of Cho & Rhee (2004), where a Hopfield network was used to classify a new exper-

iment into five different classes (two of them consisted of unsuccessful welds), lacks

confidence, because the test set was comprised of only ten experiments. In addition,

dynamic resistance has been used to monitor welding processes online (Garza & Das

2001). However, their data set was narrow, consisting of only 40 experiments. Stud-

ies using other input variables include approaches involving neural networks with tip

force, the number of weld cycles, the weld current and the upslope current (Ivezic et al.

1999). Also pre-weld information such as angular misalignment and fit-up faults of the

electrodes are used (Brown & Schwaber 2000). Common to all these studies is that the

quality control methods are developed merely to estimate quality, while using similarity

information also the parameter values affecting quality could be seen.

Although in many studies the quality control of spot welding is approached using

data mining solutions, no reported approaches aiming to improve the whole manufac-

turing process could be found. Thus, the process identification task, introduced in this

thesis, where the characteristics of a sample measured from a new process are com-

pared with information gathered from previous processes to find a similar process, is

considered an important step for the spot welding industry. This is because the retrieved

information can be used to return the process parameters leading to high-quality joints

(the time needed to set up a new process can be substantially reduced) or to predict the

quality of new welding spots using the stored information of a similar process.

29

Page 32: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

2.2.2 Steel manufacturing

Steel is the most important structural material used in industry; there is plenty of iron

(Fe) in the ground, steel is easy to manufacture, cheap, easily moldable, has high

strength and it is possibility to affect the mechanical properties of products by chang-

ing material compositions, process variables and mechanisms (Nevalainen 1970). Steel

manufacturing is a multi-step process where liquid steel is cast into steel slabs that are

then rolled into steel. Figure 4 shows an example of how steel slabs are processed after

coming out of the furnace.

Fig 4. Steel manufacturing (hot strip mill of Ruukki Steel) (Elsilä & Röning 2002).

Data mining inspired widespread research in the area of steel manufacturing. Different

properties affecting steel quality are studied separately. For example, tensile strength

and yield strength are modeled by joining information on the predicted mean and disper-

sion, aiming to solve beforehand the rejection probabilities in qualification tests (Juu-

tilainen & Röning 2004) or the actual limits inside which the strength value will be

(Juutilainen & Röning 2006). Research by Tamminen et al. (2008) concentrated on im-

pact toughness estimation. The surface defects of steel have also been considered. For

example, Haapamäki et al. (2005) studied the scale defects of steel and Jia et al. (2004)

examined seam defects.

On the other hand, models have been developed to control the intermediate phases

of manufacturing, for example, to predict the post roughing mill temperature of steel

slabs (Laurinen et al. 2001) and to monitor scheduled working times in relation to

observed times using multiagent technologies, making it possible to notice faulty situ-

ations rapidly (Fischer et al. 2004). In addition, optimization of coolant bow settings

in the spray cooling zone has been studied for possible casting process improvements

(Filipic & Robic 2004).

30

Page 33: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

As it was shown, data mining has been utilized in many different sectors of steel

manufacturing. However, the studies have concentrated on basic modeling of certain

tasks, while in this thesis the emphasis is on model maintenance. In the basic modeling

task the similarity aspect is not a necessary approach, but in model maintenance sim-

ilarity information proved its usability. While the model maintenance task introduced

in this thesis is a novel approach, it is not surprising that similarity information has not

been used this way previously. No comparable studies in which the ultimate goal is to

maintain steel quality prediction models automatically or even to improve model accu-

racy in the long run were found. Thus, also from the steel manufacturing point of view,

the steps taken in this thesis are valuable and welcome.

31

Page 34: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

32

Page 35: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

3 Methods

When considering data mining research, different machine learning methods are widely

used. Some of the methods are so-called black box methods, meaning that after the

training the method itself does not carry information on which are the most important

features, how the different features correlate with each other or whether there are similar

observations among the data. The last of these items is the most interesting point within

the scope of this thesis. Although every model can be thought to benefit from similarity,

for example, a model’s prediction capability can be considered to correlate with the

existence of similar data points in the training data, there are a smaller number of models

that do not lose similarity information. Thus, in this thesis, the methods used were

chosen not only based on their capability to solve the problem in question, but also to

introduce how similarity information can be brought forward.

From the basic applicability point of view, the methods used in this thesis, Bayesian

belief networks, a self-organizing map, a k nearest neighbor classifier and k-means clus-

tering, were chosen because they do not make strong assumptions about data distribu-

tion and therefore adjust to real-world problems where data distributions are not known.

An exception to this is the conditional independence assumption of Bayesian networks,

but the assumption is not considered to have a great effect on the adjustability of the

method with real-world data. Another advantage of the methods is that the results they

give are easy to interpret without a deep understanding of the underlying principles of

the methods. Thus, it is no coincidence that k-means clustering, the k nearest neighbor

classifier and a simplified version of the Bayesian networks (naive Bayes) were ranked

by the IEEE International Conference on Data Mining in 2006 among the top 10 most

influential data mining algorithms in the research community (Wu et al. 2008).

As the discussion on the similarity point of view of the methods assumes an under-

standing of the basic principles of the methods, the methods are introduced in the next

subsections and the similarity discussion is postponed until the last subsection. The

basic discussion about the methods is mainly focused on the principles of the methods

and the applications the methods are used for.

33

Page 36: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

3.1 k nearest neighbors classifier

The k nearest neighbors classifier (knn) is a method for which it has been proven that,

when k=1, the classification error is maximally twice the optimal Bayesian probability

of error (Cover & Hart 1967). However, the idea of k nearest neighbors classification

is quite simple: a data point is classified into the class where most of its k nearest

neighbors belong (Fix & Hodges Jr. 1951). The nearest neighbors are defined using, for

example, the Euclidean distance measure (Mitchell 1997).

Different modifications have been made to the basic algorithm. In the weighted

k nearest neighbors method (wknn) the class is defined using weighted sum voting,

meaning that the class is decided using not only the number of neighbors from certain

classes, but also their distance to the data point to be classified. The closer the neighbors

are, the more weight is given to the class they represent (Hechenbichler & Schliep 2004).

Also, scaling the features differently is a widely used method for increasing the weight

of the most significant features in the classification.

The k nearest neighbors classifier and its variations are widely used in real-world

applications. They have been used, for example, to categorizing texts (Wang & Wang

2007), to recognize human-performed daily activities based on triaxial accelerometer

and heart rate data (Pirttikangas et al. 2006) and to reduce the number of false alarms

in intrusion detection systems (Law & Kwok 2005). In industry, the k nearest neigh-

bors method has been applied to detect broken bars in asynchronous motors (Ondel

et al. 2006) and to detect surface defects on production wafers (Hunt et al. 2000). The

effectiveness of the k nearest neighbor method has also been noted when monitoring

semiconductor process tools, especially plasma etching (Chamness 2006).

3.2 k-means clustering

The k-means clustering method was developed to divide a given data set into a certain

prefixed number of clusters, k. It aims to minimize the distance inside the clusters and

maximize the distance between clusters by minimizing the squared error function

V =k

∑i=1

∑x j∈Si

|x j−µi|2, (1)

where Si, i = 1, . . . ,k are the k clusters and µi are the cluster centroids and also the

means of points x j ∈ Si. This is done using an iterative training sequence. The k cluster

34

Page 37: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

centroids are chosen randomly, after which every data point is attached to a cluster with

the closest centroid point. The new centers of the clusters are then calculated and the

centroid value is updated. The process is carried on iteratively until a point is reached

where no changes to cluster centroids happen or a certain number of iterations have

been done (Bishop 2006).

In the introduction of this chapter it was stated that the k-means method does not

make any assumptions about the distributions of the data even though the basic ver-

sion assumes that the different clusters can be separated by treating them as normally

distributed. However, the statement was based on the fact that by using different weight-

ing schemes and distance measures, the k-means method can be made to forget these

assumptions (Wu et al. 2008). For example, by using categorical weighting for some

variables or the mahalanobis distance as a distance metric, the form of the clusters can

be changed.

The actual application areas of the k-means clustering algorithm are also widespread.

It has been applied to face recognition (Hadid & Pietikäinen 2004) and to recommend

new music tunes to people based on their previous music downloading (Kim et al. 2007).

In addition, because k-means clustering is a data partition method, the solutions are

often combined with other data mining methods. For example, it has been applied

with neural networks to discover faults in satellite attitude determination systems (Cai

et al. 2007) and with genetic algorithms in anomaly intrusion detection (Marimuthu &

Shanmugam 2008). In industry, k-means clustering has been applied, for example, to

reduce state space complexity in supply networks of automotive production systems,

after which control rules can be learned more straightforwardly (Doring et al. 2007).

In semiconductor wafer fabrication process control, k-means have been used to prepro-

cess the data before applying genetic algorithms and support vector machines (Huang

& Chen 2006).

3.3 Self-organizing maps

A self-organizing map (SOM) is an unsupervised neural network method that visualizes

high-dimensional data in a low-dimensional (typically two-dimensional space). It was

developed by Professor Teuvo Kohonen in 1981. The SOM presents the statistical

dependencies of high-dimensional data in the form of geometric figures. This is done

by keeping the topologic and metric relations of the two-dimensional space as close as

possible to the relations of the initial high-dimensional space.

35

Page 38: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Self-organizing maps are widely used and more than 5000 scientific articles have

been based on them already before 2001 (Kaski et al. 1998, Oja et al. 2003). They

have been used, for example, in health monitoring (Tamminen et al. 2000), in financial

analysis (Kiviluoto & Bergius 1998) and in speech recognition (Kohonen et al. 1996).

Self-organizing maps have been combined with machine vision systems, for example, to

develop wood surface inspection methods (Niskanen et al. 2001) and to detect human

faces from pictures (Ikeda et al. 2002). Several areas of industry have noticed the

effectiveness of the method. For example, self-organizing maps are used to model a

steel production process and to study a continuous pulp digester (Alhoniemi et al. 1999).

Many process monitoring approaches also rely on self-organizing maps (Kohonen et al.

1996).

A SOM is usually formed of neurons on a regular low-dimensional grid with a

hexagonal or rectangular lattice. The neurons are model vectors mi = [mi1,mi2, . . . ,min],

where n is the dimension of the input space. Training is done by choosing a data sample

x and finding the closest model vector mc (the best-matching unit). When the best-

matching unit is found, it and its topologically closest neighbors are updated with the

equation

mi(t +1) = mi(t)+α(t)hci(t)[x(t)−mi(t)], (2)

where α(t) is the learning rate factor (a decreasing function of time) and hci(t) is the

neighborhood kernel centered on the winner unit c. Training continues by choosing a

new data sample and iterating the updating equation (Kohonen 2001).

3.4 Bayesian belief networks

Bayesian belief networks were developed to model dependencies between variables.

They have been applied to diverse problems, for example, failure analysis of inject

printers (Lee 2001) and sensor planning for mobile robot localization (Zhou & Sakane

2002). The quality of C++ programs (Masoud et al. 2004) and text-to-speech systems

(Goubanova 2001) are also studied using Bayesian belief networks.

A Bayesian belief network is a directed acyclic graph that consists of nodes and

links between them. The nodes represent variables and the links describe conditional

probabilities between them. It is applied mostly to discrete variables. Figure 5 shows

an example of a Bayesian network. Node A represent a discrete variable, which can

have states a1, . . . ,an. Every state has a probability P(ai) and the probabilities of the all

36

Page 39: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

states of variable A add up to one.

Fig 5. Bayesian belief networks.

In addition, nodes A and B in Figure 5 are considered parents of X and nodes C and D

represent children of X . The probability distribution of X is conditional on that of its

parents A and B and its children C and D. In other words, the probability of occurrence

of a certain state of X is dependent on the states of its parents and children (assuming

there is information available about their states). However, the states of the children

and parents affect a different amount. Using Bayesian networks, the biggest posterior

probabilities can be calculated for every variable (Duda et al. 2001).

A naive Bayes classifier is a simplified method of Bayes’ theorem based on the con-

ditional independence assumption that all attributes are independent, given the value

of the class variable. Although the assumptions simplify the problem, the results of

the method are in many cases even surprisingly good (Hastie et al. 2001, Hand & Yu

2001). It has actually been shown that in some cases the naive Bayes classifier outper-

forms more sophisticated methods, even when there are dependencies between variables

(Domingos & Pazzani 1997). In addition, it is very easy to construct. It does not acquire

any iterative parameter estimation and thus can be easily applied to huge data sets (Wu

et al. 2008). In practical applications the naive Bayes approach is used, for example,

in automatic video surveillance systems (Hongeng et al. 2000), to filter spam messages

(Androutsopoulos et al. 2000) and for text classification (Eyheramendy et al. 2003).

37

Page 40: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

3.5 Similarity aspect of the methods

From the similarity point of view, the methods were chosen because they do not lose

similarity information, but instead similarity is one of the driving forces of the methods.

The methods were introduced in order of their simplicity compared with the similarity

aspect, while the most straightforward method was introduced first.

The k nearest neighbor method was already mentioned in section 2.1.3 to be one

of the most common methods used in similarity research. The similarity aspect of the

method is quite obvious. Basic similarity is usually measured as distance between data

points, and what the k nearest neighbor classifier decides is based on the closest obser-

vations, which are naturally the most similar ones. Similarity and k-means clustering

are also highly related in that the observations are clustered based on their distances to

each other, indicating that the most similar observations are in the same or neighboring

clusters.

Self-organizing maps are trained to preserve topologic and metric relations as well

as possible, meaning that even in the two-dimensional plane the most similar map cells

are close to each other. The SOM was built as a visualization tool and is a more sophis-

ticated method compared to knn and k-means, but it also relies on distance information

in the training phase. In addition, when using the trained map to classify new observa-

tions, the best matching unit (the most similar map cell) is searched for and the decision

is made based on this unit.

Utilization of similarity is not as straightforward in Bayesian belief networks. Dis-

tance information is not used in the training phase, but instead the similarity aspect is

based on the structure of the method. Dependencies are estimated for discrete variables,

while the variables constitute different classes. These classes contain observations that

are similar to each other in relation to the variable. For every variable there are distinct

classes, but in the modeling phase the dependencies are estimated separately for every

class combination. Therefore, these class combinations contain only observations that

are similar to each other, and similarity information is preserved also with Bayesian

networks.

38

Page 41: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

4 Utilizing similarity

This chapter introduces the main contributions of this thesis. The advantages of utilizing

similarity information are presented clearly in four different applications in the manu-

facturing industry. Section 4.1 introduces the work done for quality control of welding

spots (publications I and II). Section 4.2 considers process monitoring (publication III),

while section 4.3 discusses how to improve manufacturing efficiency from the point

of view of spot welding (publications IV-V). The last section, 4.4, describes methods

developed for model maintenance; a case study was done for the steel manufacturing

industry (publications VI and VII).

4.1 Quality control

The main task here was to investigate how similarity information can be used to predict

the quality of welding spots in a welding process. Quality control was approached with

two different well-known methods: Bayesian networks and self-organizing maps. They

were chosen because they adapt easily to real industrial data: self-organizing maps do

not make any assumptions about the distribution of the data and although Bayesian

networks often make assumptions to simplify data, this does not significantly affect

the adjustability of the method. In addition, the results given by the methods are easy

to interpret, which was considered an important feature when aiming to implement

methods for actual welding production lines. Another very important aspect of the

methods was that, beside their given estimate, they preserve similarity information, in

which case feature sets constituting unsuccessful events can be discovered and avoided

in future production.

In both methods observations are grouped with similar observations, in Bayesian

networks using classes and in self-organizing maps by arranging them into the same

or close map elements. When the quality of a new observation is predicted, the most

similar observations are searched (the corresponding classes or map elements). The

quality of the most similar cases is studied and this information is used to estimate

the quality of unknown data points. The similarity information can also be used when

searching for the feature sets of unsuccessful events.

39

Page 42: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

4.1.1 Data set

The data used in the study were comprised of measurements from welding tests done at

Voest-Alpine Transport Montage Systeme GmbH, Austria. The data set contained ob-

servations from 192 experiments in which two metal sheets were welded together with

a resistance spot welding machine. Each of the observation sets contained measure-

ments of current, compression force and voltage signals recorded during the welding

and the diameter of the weld measured after welding (the diameter must exceed 4 mm

to represent a good joint). The signals were measured at intervals of 0.04 milliseconds

and each of them was composed of about 7000 values.

Naturally, it was not reasonable to use all these data points to study the methods,

thus suitable features were extracted from the signal curves. The extracted features are

introduced in detail in the next subsections.

4.1.2 Quality control using Bayesian networks

The first approach was to define the effects of different signal curves on the quality of

welding joints by using Bayesian networks. The study was done together with P. Lau-

rinen and also reported as an example in his thesis on a top-down approach to creating

and developing data mining solutions (Laurinen 2006). The idea of Bayesian networks

is to model conditional probabilities between variables and represent them as easily

understandable rules.

In the study, the features extracted were the quartile and median values of the signals.

Thus, every measured signal could be presented as three values constituting altogether

nine features when all three signals were treated. These continuous values were further

processed by dividing them into classes. The distributions of the continuous values

were used when creating the class structure. An example of the approach can be seen

in Figure 6 a), which shows a plot of the lower quartiles of the voltage for all the

observations. In addition, the diameter values were also divided into classes, where

class 1 contained all the unsuccessful welds (Figure 6 b)). This new data set was then

used to create Bayesian networks.

40

Page 43: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

a) b)

Fig 6. Division into classes for a) voltage and b) welding spot diameter (Publica-

tion I).

Different correlations between the diameters and some of the extracted features could

be clearly seen from the created Bayesian networks. For example, using only three

determining features, the median of the voltage (V_med), the upper quartile of the

current (C_uq) and the upper quartile of force (F_uq), the results showed which class

divisions lead to unsuccessful welds. Some of the most interesting class probabilities

of the welding spot diameter are shown in Table 1. The first columns in the table show

the classes of the features, the probabilities of the welding spot belonging to a certain

class are give in the columns marked with D and the last column shows how many of

the observations belonged in the configurations.

Table 1. Probabilities from the Bayesian network. V_med = voltage median, C_uq

= current upper quartile, F_uq = force upper quartile, D = welding spot diameter,

#obs = number of observations in the configuration.

Variables and their classes Welding spot diameter #obs

V_med C_uq F_uq D = 1 D = 2 D = 3

1 1 3 0.986 0.007 0.007 1

1 2 3 0.993 0.003 0.003 1

2 1 1 0.002 0.002 0.995 3

2 1 2 0.739 0.087 0.174 23

2 1 3 0.498 0.498 0.003 2

2 2 1 0.001 0.001 0.999 10

2 2 2 0.588 0.294 0.118 34

2 3 1 0.001 0.001 0.998 6

2 3 2 0.250 0.375 0.375 16

2 4 2 0.001 0.499 0.499 6

3 1 2 0.001 0.997 0.001 5

3 2 2 0.000 0.904 0.095 41

41

Page 44: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

It can be seen from the first rows, (1, 1, 3) and (1, 2, 3), of Table 1 that with the val-

ues situated in these classes, it is highly probable that the diameter of the welding spot

will be unsuccessful. Although in these cases only single observations were available, it

is highly probable that in future cases the same assumptions will hold. In rows (2, 1, 2),

(2, 1, 3), (2, 2, 2) and (2, 3, 2) the probabilities are more equally distributed amongst all

the classes, but more information can be obtained by using more determining features.

The rest of the configurations listed in the table show high probabilities of the welding

spots belonging to good-quality classes, which is also very important information.

Although the results in the study are not adjusted exactly to give the best results,

for example, features extracted in further studies could give more straightforward re-

sults, the basic idea and its usability in this kind of problem is obvious. Clear rules on

when unsuccessful welding spots occur or may occur make it possible for manufactur-

ers to avoid settings constituting them and in this way improve the overall quality of the

welding process.

4.1.3 Quality control using Self-organizing maps

The second approach to quality control was based on self-organizing maps. SOMs

perform clustering of feature patterns, where similar patterns form a cluster character-

ized by a representative pattern. The clusters are presented as a two-dimensional map

(U-matrix), with the map location distances reflecting the similarity of the clusters. In

addition, representative patterns can be drawn for every feature used to train the map

individually so that the nodes at same places correspond with each other. The map ar-

rangement allows direct interpretation and understanding of the interrelations between

the features and their relationship to the quality.

The study was mainly based on the means of ten equally long parts of the contin-

uously measured signals that were chosen as features. The legends used in the study

are: c = current, f = force and v = voltage, and the numbers 1,. . .,10 mark the means of

the respective tenths of the signal divided into ten parts of equal length. The study was

divided into three steps:

– Reducing the feature set (SOM trained with all the features and the quality measure),

– Studying the correlation between quality and the signal features (SOM trained with

a reduced feature set and the quality measure),

– Using the map to predict unsuccessful welds (SOM trained without the quality mea-

sure, using only 80 percent of the data).

42

Page 45: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

In the first step the correlations between the features were studied. If the map for-

mation was similar for two or more features, only one of them was left to represent the

effect of the features in the second step. In practice, this meant a reduction from 31

features to 10 features.

The map in Figure 7 shows the effects of this reduced feature set on the diameter

of welding spots. A comparison of the last feature map of diameters with the other

maps shows, for example, that small compression force value has a positive effect on

diameter. This can be seen through the large values of diameter and small values of

force occurring in the bottom right corner. Also the largest values of voltage v_3 occur

in that corner. When comparing these with the relations found using Bayesian networks

in subsection 4.1.2, it can be seen that the correlations hold. For example, in Table 1

the smallest values of force correlated positively with the size of the welding spot.

Fig 7. The trained map when the uncorrelated features of signals and diameter are

used as the training parameters (Publication II).

On the other hand, the small values of diameter appear at the top of the feature map.

The features that affect the diameters at the top are the smallest values of voltage, v_4

and v_5. The smallest values of current in the upper right corner and the largest val-

ues of force in the upper left corner also contribute to the small values of diameter.

Furthermore, these results are consistent with the results in the previous subsection.

43

Page 46: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

To estimate the quality prediction capability of the SOMs, the data were divided into

training data (80 percent of the data) and testing data (20 percent). The map was trained

with the training data; diameter information was not used as a training parameter. How-

ever, the diameter labels were assigned afterwards to the map elements representative

of the curves that belong to the corresponding cluster (Figure 8 a)). An area that con-

tains all the unsuccessful welds of the training data can be found in the figure (marked

with grey). When looking for the best-matching unit on the map for testing data points

and adding the diameter labels, it can be seen (Figure 8 b)) that all the unsuccessful

welds of the testing data are placed in the grey area derived from Figure 8 a). Only

three successful welding spots were misinterpreted, but they lie on the same nodes as

the successful welds in Figure 8 a). Depending on the area of application, the division

into successful and unsuccessful welds can be changed. If it is important that all the

unsuccessful welding spots are found, the division shown in the figure can be used, but

if some of the unsuccessful spots can be misinterpreted, the grey area can be smaller.

a) b)

Fig 8. The trained map when the uncorrelated features of signals and diameter are

used as the training parameters (Publication II).

44

Page 47: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

4.2 Process monitoring

In process control the basic idea was to monitor the manufacturing processes online. A

self-organizing map (SOM) was taught with observations from different process states

and an area of optimal process states was discovered. When a new process was launched

the new observations were mapped into the SOM. If a drift from the optimal area was

identified while the process was ongoing, necessary changes were made instantly to

enable continuation of high-quality production. The functionality of the system was

verified with spot welding data.

In this study similarity was utilized in both phases: process monitoring and process

adjustment. The self-organizing map was used to divide the training data into nodes

comprised of similar observations. For each node, the quality (poor, average, good)

was estimated and an optimal area was decided from the map. In the control phase the

best matching node (the most similar) was searched for every new observation. If the

most similar node was not from the optimal area, the process was adjusted by changing

the machining parameters based on the distance and angle between these two areas.

The principle of the system in a spot welding case can be seen in Figure 9. The

current and voltage signals gathered during every separate welding event were merged

into a resistance signal. The resistance signal was preprocessed and suitable features

were extracted from it using principal component analysis. The self-organizing map and

the developed drift detector were used to find the distance and angle to the optimal area

and the converter was used to adjust the current and pressure based on this information.

45

Page 48: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Fig 9. The principle of drift detector (Publication III).

Figure 10 shows an example of a trained self-organizing map. The qualities assigned to

each node are color-coded, meaning that red and green shades mark better quality. The

optimal area found is marked as a rectangle and the center of it is marked as a white

circle. When a drift is detected, like in this case when the best matching node (marked

as yellow circle) is pretty far from the optimal area, the distance and the angle to the

optimum center can be calculated.

46

Page 49: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Fig 10. Self-organizing map with process optimum area marked as rectangle. The

white circle marks the optimum center, while the yellow circle marks the winner

node for a new welding event (Publication III).

The results showed that this approach is applicable in a real-world problem. The overall

production duration of high-quality welding joints was lengthened and the system was

easily set to the appropriate manufacturing environment.

4.3 Improvement of manufacturing efficiency

In this task similarity information was used to identify similar welding processes. The

ultimate aim of the system was to gather vast amounts of information from different

manufacturers into a database from which the most similar process will be sought. After

the similar process is identified, its detailed information is retrieved from the archive.

The information can be retrieved for two distinct purposes: (1) to restore the process

parameters leading to high-quality joints and substantially reduce the time needed to

set up a new process or (2) to seek the quality prediction and improvement methods

already developed for the similar process. Thus, the developed system makes the overall

manufacturing process more efficient while time, money and effort are saved from the

beginning of production.

47

Page 50: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

The principle of this study can be seen in Figure 11. First the data from a new

process are gathered and pre-processed. The new data are compared with the data

already stored in the database and the most similar process is identified among them

(process identification). After that the similarity measure between the new data set and

the most similar process is calculated. If the processes are classified as similar based

on the similarity measure, the database restores the process parameters leading to high-

quality joints or the methods optimal for quality control and improvement. The study

on the database build-up is reported in (Tuovinen et al. 2007).

Fig 11. The principle of the study (Publication V).

4.3.1 Data set and features

In the study, the data set was provided by two welding machine manufacturers: Harms

+ Wende GmbH & Co. KG (HWH) (Harms+Wende www page) and Stanzbiegetech-

nik (SBT) (Stanzbiegetechnik www page). The HWH tests were done on materials

commonly used in the automotive industry, while SBT concentrates on welding thin

materials used in electronic components.

The data set contained altogether 3879 experiments from 20 processes (11 from

HWH and 9 from SBT, and they are marked as HWH1-HWH11 and SBT1-SBT9).

The experiments were welded using varying welding parameters, and each experiment

consisted of measurements of current and voltage signals recorded during the welding

process. The processes were divided into several configurations. The term ’configu-

48

Page 51: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

ration’ refers to a set of experiments (usually 15) welded with unchanged parameters.

For example, the experiments measured with current values of 8.0 kA, 7.2 kA and 6.4

kA while the other parameters remained unchanged constituted three different configu-

rations. The parameters that varied between the configurations were current, electrode

force, electrode wear and fitting.

The measured signal curves contained plenty of oscillatory motion and a pre-heating

section, and they were hence preprocessed before further manipulation. The pre-heating

parts of the curves were cut off, leaving only the signal curves recorded during the actual

welding phase. In addition, the curves were smoothened using the Reinsch algorithm

(Reinsch 1967, 1971). After pre-processing, suitable features were extracted from the

signal curves. As in section 4.1.3, the averages of ten parts of equal length were used

as features.

4.3.2 Process identification using self-organizing maps

The first steps to process identification were taken using self-organizing maps. The

visualization capability of the method was considered a benefit to actual application

usage. In addition, SOMs were used to find the correct initialization parameters, which

could be taken into use after identifying the closest process.

At this point the data set consisted of five SBT processes, and ten means were

calculated for the resistances of the signal curves. The classification ability of the self-

organizing maps was verified by dividing the data into training data and test data. The

map was trained with 80 percent of the available data, and in Figure 12 a) the obser-

vations of five test series are labeled according to corresponding map nodes with the

numbers 1-5. However, the division in the lower part of the map only points out the

differences inside processes 1 and 2, which are thought to be similar processes, and

it is therefore not considered an important division. In addition, the processes of an

independent test set were classified correctly using the map.

The results showed that self-organizing maps can be used for process identification.

49

Page 52: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

a) b)

Fig 12. The class labels for a) training data and b) test data (Publication IV).

After the study, more data were archived and the accuracy of the self-organizing maps

decreased. Thus, the visualization capability had to be forgotten and a more accurate

method of classification had to be sought.

4.3.3 Extension for process identification

Haapalainen et al. (2005) carried out a comparative study of different parametric and

non-parametric classification methods for spot welding process classification. In their

study, the k nearest neighbor method was found to be the most suitable for the pro-

cess identification task. Nevertheless, the approach was based on identifying the most

similar process and a more accurate examination of the similar process was not exe-

cuted. This means that, although a similar process is retrieved from a database using

the method, how can we be sure it is similar enough to be used to give suitable initializa-

tion parameters and quality control methods. Thus, the emphasis was shifted not only

to process identification but also on actual similarity between processes.

50

Page 53: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

In this study the knn method was used to find the most similar process. In addition,

an extension of the method was developed to find out if the most similar process could

be classified as similar, because it was recognized that some of the new processes could

be so dissimilar from the ones already stored in the database that leaving out the distance

information would cause serious misclassifications. The approach was also reported in

Junno et al. (2005).

The idea of similarity was approached by developing two boundaries, taking into

consideration that different processes were divided into several configurations (a set

of experiments welded with unchanged parameters). In practice, the first boundary

calculated the average distance to k (k = 5) nearest neighbors inside the configurations,

while the second one expressed the average distance to k nearest neighbors between two

neighboring configurations. The configurations by which the actual second boundary

value was calculated differed only with regard to the initialized current, because current

was the only parameter that varied between the configurations in every process and

can be assumed to vary in future processes. The final threshold, called the similarity

measure, was a combination of these boundaries.

A simplified example of the boundaries is presented in Figure 13. Actually, the

data set was 20-dimensional, but the idea is presented here using a two-dimensional

figure. The white circles refer to measurements from two neighboring configurations

of a process. The black circles refer to test points 1-3. The test points demonstrate the

new measurements to be classified, and the grey dashed circles around them mark the

boundaries.

51

Page 54: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Fig 13. A simplified example of the boundaries (Publication V).

The final similarity measure combines the information of the boundaries and provides

a more straightforward result to be used in industry. The final form of the similarity

measure was obtained by comparing these two boundaries with the average distance

from the data points of a new test set to the k nearest neighbors of the closest matching

process. If the new distance is marked as d and the boundaries are marked as b1 and b2,

then the similarity measure sim is

sim =d−b1b2−b1

. (3)

Now the thresholds for the similarity measure can be formulated more concretely. If

the similarity value is smaller than zero, the new test set is considered to be precisely

the same process as the closest matching process. If the similarity measure is between

zero and one, the closest process is considered similar, and knowledge of that process

can be used as tentative information of the new test set. If the similarity value is greater

than one, the closest matching process is classified as dissimilar.

In this study, each configuration of different processes was successively considered

as a new data set (information of the process to which the configuration belongs is not

taken into account) and compared with the information stored in the database, making

the setting resemble the actual operation of the system. To avoid repetition, only three

of the processes are introduced in detail. Table 2 shows the results obtained from the

52

Page 55: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

first five configurations of the processes. The abbreviations C1-C5 represent different

configurations. In addition, the first number of cells indicates the percentage of data

points of a configuration classified as each process, while the second number is a value

of the similarity measure.

Table 2. Classification of the first five configurations of a) HWH2, b) SBT3 and c)

SBT4.

Config a) HWH2 b) SBT3 c) SBT4

HWH3 HWH4 SBT9 SBT9

% SM* % SM* % SM* % SM*

C1 100 0.49 100 0.89 100 2.61

C2 93.3 0.37 6.7 1.02 100 0.86 100 2.97

C3 13.3 0.32 86.7 0.79 100 0.89 100 3.08

C4 100 0.88 100 0.86 100 3.08

C5 100 0.21 100 0.89 100 3.11

*Sm = Similarity measure

The HWH2 case in Table 2 indicates that most of the configurations are classified with

100 % accuracy as HWH3. The configurations C2 and C3 are exceptions to this. For

example, 93.3 percent of the experiments of configuration C2 are classified as process

HWH3 and 6.7 percent as process HWH4. In this case, the similarity measure is also

better for HWH3. However, most of the experiments in configuration C3 are classified

as HWH4, while the similarity measure is significantly smaller for HWH3. Hence, the

user needs to decide which should be considered more important, a small value of the

similarity measure or a large classification percentage. However, the system developed

in this study shows both similarity and percentage, which enables the user to make

the decision by applying information on both processes and choosing the most suitable

alternative.

In the SBT data set, the similarity measure showed that some of the most similar

processes were farther than others. An example of two processes known to be approx-

imately the same is shown in Table 2 b). Processes SBT3 and SBT9 are similar to

each other, but the SBT9 process was welded one year later. Table 2 b) shows the

classification results for SBT3. It can be seen that the similarity measures for all the

53

Page 56: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

configurations are smaller than one, and the processes are therefore classified as similar.

An opposite example can be seen in Table 2 c) for process SBT4, for which there are no

similar processes stored in the database. The similarity measures give the same infor-

mation. The configurations of process SBT4 are classified as SBT9, but the similarity

measures are much larger than one. These two processes are classified as dissimilar

using the similarity measure; this was also the desired result.

4.3.4 Ways to improve efficiency

When the most similar process is retrieved and it has been determined that it is simi-

lar enough, the benefits of the system introduced in Figure 11 can be achieved. The

simplest benefit is that the correct initialization parameters can be restored from the

database, which reduces the time needed for the machine setup. In addition, different

data mining methods developed for quality control or process drift estimation of the

most similar process can be employed as such or with small modifications, enabling

certain delivery of high-quality end products to the customers. These benefits are intro-

duced in more detail in the next subsections.

Restore of correct initialization parameters

When a similar process is found, the database is searched and the parameters leading

to high-quality joints are restored. This approach can substantially reduce the setup

time of a new process. In practice, even experienced welding machine users must find

suitable parameters through trial and error. Naturally, this means spending more time

and money, while the system developed here can restore the parameters faster and more

automatically. The ultimate aim of the system is to gather vast amounts of information

from different manufacturers into a database from which suitable initialization parame-

ters can be searched independent of the manufacturer. Thus, when a manufacturer be-

gins to weld new parts (expertise needed to select the suitable parameters is not found

inside the factory) the initialization parameters already used in a different factory can

be found from the database.

54

Page 57: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Restore of quality control and process monitoring methods

Quality control and improvement methods that already exist can be used for a new

process if the process that is found is classified as similar. A simple example of quality

prediction can be seen in Figure 14. The average quality of the k closest data points

of a new experiment is assigned as a quality value of the experiment while the results

are introduced configuration-wise (= as a mean of approximately 15 observations). In

Figure 14 a) the most similar process is classified as similar, while in Figure 14 b) the

most similar process is classified as dissimilar. It can be seen that the measured and

predicted values in similar case correspond well to each other. Naturally, when the

similarity measures are poor the results are remarkably worse. In that case, the quality

information of the closest matching process cannot be used, but because the similarity

measure is known, this mistake would not be made.

(a) (b)

Fig 14. Average measured and predicted quality values of experiments by different

configurations of a) similar processes and b) dissimilar processes (Publication V).

Figure 15 shows an example how detection and visualization of process drifts can be

utilized for a new process. The basic idea for one process was introduced in section

4.2. In this case, the self-organizing map is trained using information from a similar

process, while the map is also used to detect and visualize the process drifts of a new

process. Figure 15 a) shows the U matrix of a self-organizing map of the restored

process, while Figure b) shows a labeled map where the labels demonstrate successful

(marked as 1) and unsuccessful (marked as 0) welding experiments. It can be seen that

the unsuccessful welds are located in the upper and lower left corners. The grey area

in Figure 15 b) is used to mark the optimal process state area where the highest quality

welding spots are situated.

55

Page 58: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

a) b) c)

Fig 15. Self-organizing map for process drift detection and visualization (Publica-

tion V).

The labels demonstrated in Figure 15 c) are labels of the quality of the new process,

which are assigned to the map elements of the restored process. The same optimal,

grey area is also used for process drift detection in this case. It can be seen that the

unsuccessful welding experiments of both processes are situated in approximately the

same areas. This means a previously trained map of a similar process can help improve

the quality of the welding spot of a new process.

4.4 Model maintenance

The application area where the functionalities of the methods introduced in the previous

sections were verified was spot welding. In the model maintenance case the results

are from the steel industry. Already now, development engineers control mechanical

properties such as yield strength, tensile strength, and elongation of the metal plates

beforehand on the basis of planned production settings. The solution is achieved using

regression models introduced by Juutilainen & Röning (2006). However, acquisition of

56

Page 59: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

new data and the passing of time decrease the reliability of the models, which can bring

economic losses to the plant. Thus, maintenance of the models emerges as an important

step in improving modeling in the long run. The next subsections describe how to find

the correct moment for a model update and how to select suitable instances for model

fitting when all the available data cannot be used.

In model maintenance, similarity information is used in two ways. In a model

update information from previous similar data points is used to decide the optimal time

for model retraining. Information on the estimation accuracy of the models for similar

data points and on the number of similar data points is combined to form a limit, called

the exception limit. This limit is then used in deciding on the need for model retraining.

In the training data selection, similarity information is used to reduce the amount of data

needed to train the models without losing interesting pieces of information. In practice

this is done in the selection by weighting fewer data points having plenty of similar

observations in the data space. The similar data points are discovered using clustering,

which divides the data points into clusters based on their similarity.

4.4.1 Data set and regression models

The data for this study were collected from the production database of Ruukki steel

works between July 2001 and April 2006. The whole data set consisted of approxi-

mately 250,000 observations. Information was gathered on element concentrations in

actual ladle analyses, normalization indicators, rolling variables, steel plate thicknesses,

and other process-related variables (Juutilainen et al. 2003). The observations were

gathered during actual product manufacturing. The volumes of the products varied, but

if there were more than 500 observations from one product, the product was considered

a common product. Products with fewer than 50 observations were categorized as rare

products.

In the studied prediction model, the response variable used in the regression mod-

eling was the Box-Cox-transformed yield strength of the steel plates. The Box-Cox

transformation was selected to produce a Gaussian-distributed error term. The devi-

ation in yield strength also depended strongly on input variables. Thus, the studied

prediction model included separate link-linear models for both mean and variance

yi ∼ N(µi,σ 2i )

µi = f (x′iβ )

57

Page 60: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

σi = g(z′iτ). (4)

The length of the parameter vector of mean model β was 130 and the length of the

parameter vector of variance model τ was 30. The input vectors xi and zi included 100

carefully chosen non-linear transformations of the 30 original input variables; for exam-

ple, many of these transformations were products of two or three original inputs. The

link functions f and g that were used were power transformations selected to maximize

the fit with data. The results are presented in the original (nontransformed) scale of the

response variable (Juutilainen & Röning 2006).

Goodness measures for regression models

The data set is constituted of different amounts of observations from different products;

there could be thousands of observation from the most common products and only a

few from the rarest products. Thus, a normal average of absolute prediction error would

skew the regression models toward even small improvements in the prediction accuracy

of the most common products. To prevent skewness, two different goodness measures

were introduced where the prediction accuracy of the observations of the rarest products

was emphasized.

The accuracy of the models was measured in the property prediction case using the

weighted mean absolute prediction error

MAE =1

∑Ni=1 w(i)

N

∑i=1

w(i)|yi− µi|. (5)

In the variation model case, a robustified negative log-likelihood was employed to take

into account the variance

robL =1

∑Ni=1 w(i) ∑

iw(i)

(log(σ 2

i )+ρ[(yi− µi)

2

σ 2i

]). (6)

Here, the function ρ(·) is a robust function; this study employed

ρ(t) =

{t, when t ≤ 25

b2, when t > 25,(7)

which truncates the squared standardized residuals if the standardized residual is below

-5 or above +5.

Two different methods were used to define the weights, w(i). They were chosen to

reflect the usability value of the models. In the first goodness criterion (Goodness 1)

58

Page 61: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

the weights w(i) were defined by product, meaning the weight of the observations of a

product could be at most as much as the weight of T observations. Let Ti be the number

of observations that belong to the same product as the ith observation. Then the weight

of observation i is

w(i) =

{1, when Ti ≤ T

T/Ti, when Ti > T.(8)

Here the value of T = 50, meaning if there are more than 50 observations of a product,

the weight is scaled down. The second goodness criterion (Goodness 2) was formed

to take only rare observations into account. Thus, only the cases for which there were

only fewer than 30 previous observations within a distance 0.9 or a distance 1.8 were

included, but the weight of the latter was double (equation 9)

w(i) =

1, when {#x j|||xi− x j||< 0.9& j < i}< 30

2, when {#x j|||xi− x j||< 1.8& j < i}< 30

0, else.

(9)

4.4.2 Finding a correct moment for model update

In this study the correct moment for a model update was approached using informa-

tion from previous cases. The update was based on the average prediction errors and

the average of the squared standardized residuals of deviations of similar past cases.

The similar past cases inside a certain distance limit constituted a neighborhood and

the average values inside the neighborhood were obtained using equations 10 and 11.

Koskimäki et al. (2007) reported a preliminary study where the update was considered

only from the point of view of model prediction errors.

The average prediction error of the neighborhood (= APEN) was calculated as the

distance-weighted mean of the prediction errors of observations belonging to the neigh-

borhood:

APEN =∑n

i=1[(1−di

max(d) ) · εi]

∑ni=1(1−

dimax(d) )

, (10)

59

Page 62: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

where

n = number of observations in a neighborhood,

ε(i) = the prediction error of the ith observation of the neighbourhood,

di = the Euclidian distance from the new observation to the ith observation of

the neighborhood,

max(d) = the maximum allowed Euclidian distance between the new observation

and the previous observations in the neighborhood (= 3.5).

The average of the squared standardized residuals in the neighbourhood (= ASSRN)

was obtained using:

ASSRN =∑n

i=1[(1−di

max(d) ) ·εi

2

σ2 ]

∑ni=1(1−

dimax(d) )

, (11)

where

σ = deviation achieved from the regression model.

The optimal time for a model update was obtained by using two different limits: the

exception limit and the update limit. The exception limit was exceeded if the value of

APEN was too high in relation to the size of neighborhood or the value of ASSRN was

too high or low compared with the average. The update limit was formed to merge the

information of exceptions. The limit was defined as being exceeded if the sum of the

exceptions within a certain time interval was too high. Whenever the update limit was

exceeded the property prediction and deviation models were re-fitted.

Tables 3 and 4 show the goodness criteria when the models are updated using this

approach. To compare the results, the criteria are presented in three different cases:

predicted with a newly updated model, with a model updated in the previous iteration

step, and with a model that was not updated at all. The results show that, although

the differences between the new model and the model from the previous iteration step

are not in every case very big, the update improves the prediction in each of the steps.

In addition, in some steps it can be seen that even a small addition of training data

improves accuracy significantly. Naturally, the benefit of the model update is obvious

when the results of the updated model and the model without an update are compared.

60

Page 63: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Table 3. Means of absolute property prediction errors with the goodness criteria.

Step size With new

model

goodness 1

With previous

model

goodness 1

Without

update

goodness 1

With new

model

goodness 2

With previous

model

goodness 2

Without

update

goodness 2

12228 - - - - - -

36703 11.59 11.74 11.74 14.40 14.62 14.62

18932 10.48 11.01 10.70 13.20 14.15 14.03

5626 10.93 11.07 11.25 12.94 13.11 13.10

17636 12.48 12.57 12.97 16.48 16.60 18.23

6104 11.47 12.39 13.35 12.56 13.87 14.67

743 19.94 20.01 25.72 57.18 57.48 71.18

3772 12.77 14.75 17.77 21.10 36.00 49.63

13533 12.47 12.62 13.68 16.34 17.03 22.37

43338 11.78 11.97 12.51 13.66 13.98 15.61

14831 12.78 13.26 13.77 21.05 20.98 25.55

21397 12.45 12.70 14.26 22.43 23.84 36.60

622 16.20 18.28 25.39 73.18 99.29 156.71

Mean 11.99 12.26 12.79 15.43 16.08 18.10

61

Page 64: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Table 4. Robustified negative log-likelihood values with the goodness criteria.

Step size With new

model

goodness 1

With previous

model

goodness 1

Without

update

goodness 1

With new

model

goodness 2

With previous

model

goodness 2

Without

update

goodness 2

12228 - - - - - -

36703 6.29 6.30 6.30 6.87 6.92 6.92

18932 6.13 6.27 6.19 6.57 6.76 6.72

5626 6.22 6.23 6.29 6.52 6.52 6.56

17636 6.38 6.39 6.46 7.15 7.13 7.36

6104 6.32 6.42 6.53 6.62 6.90 6.84

743 8.02 8.06 10.66 15.84 16.11 29.56

3772 6.36 6.94 8.58 7.16 11.58 23.04

13533 6.43 6.45 6.84 7.15 7.20 10.13

43338 6.35 6.38 6.47 6.67 6.71 7.33

14831 6.55 6.65 6.94 7.47 7.52 11.13

21397 6.40 6.45 7.13 7.17 7.54 15.39

622 6.58 6.76 11.47 9.65 11.66 70.85

Mean 6.36 6.41 6.61 6.91 7.05 8.28

4.4.3 Selecting a suitable subset for training data

The previous section discussed how to recognize the optimal moment to update models

when the amount of available data increases. This section introduces a procedure for

selecting suitable training data instances for this update step.

Although computers are getting more efficient, there are still problems in training

data mining methods with all the data gathered. Memory or calculation capacities are

not sufficient enough because the sizes of data sets are also growing. To tackle this

problem, a five-step instance selection algorithm was developed for selecting suitable

instances from the whole data set for regression model fitting:

1. Perform coarse clustering (Algorithm 1)

2. Split coarse clusters whose size is greater than V using k-means clustering

3. Select all the instances from non-partitioned coarse clusters

62

Page 65: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

4. Select instances from k-means clusters using equation 13

5. Select the rest of the instances using random sampling

In coarse clustering the data set is divided into clusters using a certain predefined dis-

tance limit so that two different clusters cannot contain any observations within this

limit (Algorithm 1). After coarse clustering, k-means clustering is performed on the

largest clusters.

Algorithm 1 Coarse clustering algorithm

for each xi in data do

if ∃ one cluster: dist(x j,xi)≤ d &x j ∈ cluster then

cluster← xi

else

if ∃ several clusters: dist(x j,xi)≤ d &x j ∈ cluster then

merge these clusters← xi

else

create new cluster← xi

end if

end if

end for

After clustering has been performed, the actual data selection phase follows. The

instances are selected by calculating an information value for model fitting (measured

on the basis of the location and spread of the cluster) for each cluster and then the num-

ber of selected observations is determined in the proportion of the cluster’s information

values.

Let C1,C2, . . . ,Cq denote the clusters resulting from the application of coarse cluster-

ing and k-means algorithms and let c1,c2, . . . ,cq be the corresponding cluster centroids.

Let ni = #{x j|x j ∈ Ci} denote the number of observations included in the ith cluster.

Then the information contents of the ith k-means cluster can be measured by

Ji = exp(a1σi +a2mi) (12)

where σi =√

(1/ni)∑x j∈Ci||ci− x j|| is the within-cluster deviation, mi = max j||ci−c j||

is the distance to the closest cluster centroid and a1,a2 are tuning constants. Small

coarse clusters are valuable for model fitting because they are isolated from other input

space regions that contain data. Thus, all observations from coarse clusters that were

63

Page 66: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

not partitioned using k-means are included in the selected training data. The number of

selected instances, Si, of the partitioned clusters is obtained using the equation:

Si =N ∗ Ji

∑qj=1 Ji

, (13)

where q is the number of k-means clusters and N is the number of observations that still

need to be selected after the observations of the coarse clusters have been chosen.

The goodness of the training data selection was studied by comparing the predic-

tion accuracies of the models trained with the limited data chosen with our approach

(50,000/200,000, clust), with limited data chosen randomly (50,000/200,000, rs) and

with the whole data set (reference, 1-200,000, all). The prediction accuracy of the mod-

els was studied for the remaining observations (ca. 45,000). In addition, the results of

our method were studied for three different cluster structures (clust 1, 2 and 3). They

are introduced to show that although k-means clustering does not necessarily find the

optimal solution in every run, the use of our approach to instance selection minimizes

the problem.

Because the results varied during different instance selection runs using our method

and using random sampling, the selection phase was repeated ten times. Using goodness

criterion 1 for the property and prediction models, the results are slightly better than

those obtained from random sampling, while both ways were approximately as good

as the reference model. However, when considering the rarest cases (Goodness 2) and

comparing the results obtained from the worst instance selection run, our approach

outperformed random sampling (Table 5).

Table 5. Best and worst goodness 2 values and standard deviation for the prop-

erty and deviation models of different runs using our method and with random

sampling. Smaller values mean better results and references are 22.64 and 7.49.

MAE RobL

Min Max Std Min Max Std

clust 1 21.48 22.35 0.28 7.37 7.55 0.05

clust 2 21.55 22.15 0.20 7.41 7.54 0.04

clust 3 21.72 22.59 0.28 7.38 7.72 0.10

rs 22.00 24.52 0.79 7.36 8.55 0.37

64

Page 67: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

For example, when considering the application-specific working limits, which are a

combination of the property and deviation measures, it can be seen that the prediction

accuracy achieved using random sampling can be almost nine percent smaller than in

the reference case (goodness 2, MAE: 22.64 and goodness 2, RobL: 7.49). On the other

hand, when comparing the prediction accuracies achieved using our approach, better

results can be seen. In clust 1 and 2 the combined prediction capability of the property

and deviation models is always better than in the reference case. Although in clust 3

the prediction capability is weaker than in the reference case, the difference is only 0.5

percent, which is not considered a significant drop.

From the application point of view, the instance selection procedure should be con-

sistent in every run. In addition, it is desirable that it would not weaken the model’s

ability to predict rare steel grades, meaning the method is able to generalize at the

boundaries of the training data. Based on the results, it seems that these can be fulfilled

using our approach. Thus, our approach is suitable for the instance selection task.

65

Page 68: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

66

Page 69: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

5 Summary and conclusions

This thesis addresses the issue of how similarity information can be utilized in industrial

applications. Nevertheless, in general, every modeling task can be thought to utilize

similarity: the models are trained with data that is assumed to represent the data space

and thus the estimation capability of the models with new data points is related to the

existence of similar data points in the training data. However, most of these methods

do not preserve similarity information, and the information is lost in the training phase.

Because similarity information was the main interest in this study, the methods were

chosen to preserve similarity information and even to be based on the similarity of data.

On the other hand, similarity information was utilized in diverse, actual industrial

applications. The application areas were spot welding and steel manufacturing, includ-

ing not only low-level manufacturing problems like quality control of spot welding,

but also higher-level problems where the aim was to improve the whole manufacturing

process. Thus, it was shown that the similarity approach is usable when considering

different phases of manufacturing.

In this study, Chapter 4.1 introduced how Bayesian networks and self-organizing

maps were used to estimate quality inside a spot welding process. With these methods

the training data was divided into classes or nodes based on similarity information. In

addition, quality estimates were given based on the most similar events. It was shown

that Bayesian networks and self-organizing maps can be used to estimate the quality

of individual welding spots and that similarity information is valuable for industrial

applications. While with black box models only quality can be estimated, with SOM

and Bayesian networks the variable values constituting quality can also be seen, and by

avoiding these settings more high-quality products can be manufactured.

Chapter 4.2 also approached similarity information using self-organizing maps. The

different states of the spot welding manufacturing process were mapped into a map and

an area corresponding to the optimal process state was identified. For new events the

most similar node was searched from the map and, based on this information, it could

be noted that the ongoing process was drifting from the optimal. In addition, counter

actions were performed using the similarity information.

The study carried out in Chapter 4.3 aimed to make the overall manufacturing pro-

cess more efficient by saving time, money and effort from the beginning of production.

67

Page 70: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

This was done by utilizing information from similar process when setting up new pro-

cesses or when searching for suitable quality control methods. In practice, a database

of welding events from different manufacturers was built up. The most similar process

was sought from the database and its detailed information was retrieved for two dis-

tinct purposes: (1) to restore the process parameters leading to high-quality joints and

substantially reduce the time needed to set up a new process or (2) to seek the quality

prediction and improvement methods already developed for the similar process. In this

task also unsimilarity information was found to be valuable in giving information if the

most similar event was actually similar enough to be used.

In Chapter 4.4 similarity information was utilized in model maintenance tasks. The

study was not done to directly mine any specific information from the industrial point

of view, but the aim was to improve overall model efficiency and usability without

touching the actual model training procedure. This means the developed methods were

not directly modeling the industrial process, but they supported the model modeling the

process. In this task similarity information was used in two tasks: to decide the optimal

moment for model refitting and to select suitable training data for this fitting. In the

optimal-moment-seeking phase, the estimation accuracy of the models for similar data

points and the number of similar data points were utilized, while in instance selection

the selection was based on a cluster structure where similar data points constituted their

own clusters. In both cases similarity information proved its usefulness: an optimal

moment for model updating was discovered and the training data set could be reduced

to a one-fourth of the original without undermining the model’s accuracy.

As it was shown, research on data mining is spreading into industry. Statistical and

machine learning methods can benefit several steps of manufacturing. On the other

hand, when working with industry, new purposes for data mining methods can be dis-

covered. The use of real data requires methods that can adapt to changing situations

and are flexible and accurate. Thus, real-world applications are also a driving force for

the development of new methods and modification of existing ones. This thesis showed

some contributions that can be achieved when dealing with an extensive industrial area

and demonstrated that industry and data mining research can benefit highly from each

other.

68

Page 71: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

References

Aamodt A & Plaza E (1994) Case-based reasoning:foundational issues, methodological varia-tions, and system approaches. Artificial Intelligence Communications 7(1): 39–59.

Achiche S, Baron L & Balazinski M (2006) Predictive fuzzy control of paper quality. Proceedingsof Annual meeting of the North American Fuzzy Information Processing Society 31–34.

Aha D & Bankert R (1994) Feature selection for case-based classification of cloud types: Anempirical comparison. Case-Based Reasoning: Papers from the 1994 Workshop .

Alhoniemi E, Hollmén J, Simula O, & Vesanto J (1999) Process monitoring and modeling usingthe self-organizing map. Integrated Computer-Aided Engineering 6(1): 3–14.

Altinçay H & Ergün C (2004) Clustering based under-sampling for improving speaker verifica-tion decisions using adaboost. SSPR/SPR 698–706.

Androutsopoulos I, Paliouras G, Karkaletsis V, Sakkis G, Spyropoulos CD & Stamatopoulos P(2000) Learning to filter spam e-mail: A comparison of a naive bayesian and a memory-basedapproach. Proceedings of the workshop Machine Learning and Textual Information Access,4th European Conference on Principles and Practice of Knowledge Discovery in Databases1–13.

Aravinthan A, Sivayoganathan K, Al-Dabass D & Balendran V (2001) A neural network systemfor spot weld strength prediction. Proceedings of the UK Simulation Society, UKSIM2001156–160.

Bhattacharya AK, Debjani S, Roychowdhury A & Das J (2007) Optimization of continuous cast-ing mould oscillation parameters in steel manufacturing process using genetic algorithms.IEEE Congress on Evolutionary Computation 3998 – 4004.

Bishop C (2006) Pattern Recognition and Machine Learning. Springer.Boyle IM, Rong K & Brown DC (2004) CAFixD: A case-based reasoning fixture design method:

Framework and indexing mechanisms. Proceedings of the ASME Design Technical Confer-ences .

Brighton H & Mellish C (1999) On the consistency of information filters for lazy learning algo-rithms. Proceedings of the Third European Conference on Principles of Data Mining andKnowledge Discovery 283–288.

Brighton H & Mellish C (2002) Advances in instance selection for instance-based learning algo-rithms. Data Mining and Knowledge Discovery 6: 153–172.

Brown L & Schwaber J (2000) Identifying operating conditions from pre-weld information forresistance spot welding. Proceedings of the American Control Conference 1535–1539.

Cai L, Huang Y, Lii S & Chen J (2007) Using rbf neural network for fault diagnosis in satelliteads. Proceedings of the 1IEEE International Conference on Control and Automation 1052 –1055.

Cardie C (1993) Using decision trees to improve case–based learning. Proceedings of the TenthInternational Conference on Machine Learning 25–32.

Chamness KA (2006) Multivariate fault detection and visualization in the semiconductor industry.Ph.D. thesis, University of Texas.

Chen G, Chen J, Zhao Z & Ruan X (2005) An object-oriented hierarchical case representationof automotive panels in a computer-aided process planning system. International Journal ofAdvanced Manufacturing Technology 22(11-12): 1323–1330.

69

Page 72: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Cho Y & Rhee S (2002) Primary circuit dynamic resistance monitoring and its application toquality estimation during resistance spot welding. Welding Journal 104–111.

Cho Y & Rhee S (2004) Quality estimation of resistance spot welding by using pattern recog-nition with neural networks. IEEE Transactions on Instrumentation and Measurement 53:330–334.

Cover T & Hart P (1967) Nearest neighbor pattern classification. IEEE Transactions on Informa-tion Theory 13(1): 21–27.

Dash M & Liu H (1997) Feature selection for classification. Intelligent Data Analysis 1: 131–156.

Domingos P & Pazzani MJ (1997) On the optimality of the simple bayesian classifier under zero-one loss. Machine Learning 29(2-3): 103–130.

Doring A, Dangelmaier W & Danne C (2007) Using k-means for clustering in complex automo-tive production systems to support a q-learning-system. Proceedings of the IEEE Interna-tional Conference on Cognitive Informatics 487 – 497.

Duda RO, Hart PE & Stork DG (2001) Pattern Classification. John Wiley & Sons, Inc., NewYork.

Elsilä U & Röning J (2002) Knowledge discovery in steel industry measurements. Proceedingsof Starting Artificial Intelligence Researchers Symposium .

Eyheramendy S, Lewis D & Madigan D (2003) On the naive bayes model for text categorization.Proceedings of the 9th International Workshop on Artificial Intelligence and Statistics .

Fayyad U, Piatetsky-Shapiro G & Smyth P (1996) From data mining to knowledge discovery indatabases. Ai Magazine 17: 37–54.

Filev D (2002) Applied intelligent control - control of automotive paint process. Proceedings ofthe 2002 IEEE International Conference on Fuzzy Systems 1: 1–6.

Filipic B & Robic T (2004) A comparative study of coolant flow optimization on a steel castingmachine. Proceedings of the Congress on Evolutionary Computation 1: 569 – 573.

Fischer K, Jacobi S, Diehl C & Theis C (2004) Multiagent technologies for steel production andcontrol. Proceedings of the UIEEE/WIC/ACM International Conference on Intelligent AgentTechnology 555 – 558.

Fix E & Hodges Jr JL (1951) Discriminatory analysis - nonparametric discrimination: Consis-tency properties. Technical Report 4, U.S. Air Force, School of Aviation Medicine, RandolphField, TX .

Frawley WJ, Piatetsky-Shapiro G & Matheus CJ (1992) Knowledge discovery in databases - anoverview. Ai Magazine 13: 57–70.

Gabrys B, Leiviskä K & Strackeljan J (2005) Do Smart Adaptive Systems Exist, A Best-Practicefor Selection and Combination of Intelligent Methods. Springer-Verlag, Berlin, Heidelberg.

Garza F & Das M (2001) On real time monitoring and control of resistance spot welds usingdynamic resistance signatures. Proceedings of the 44th IEEE 2001 Midwest Symposium onCircuits and Systems, MWSCAS 2001 1: 41–44.

Goubanova O (2001) Using bayesian belief networks for model duration in text-to-speech sys-tems. Proceedings Theoretical & Applied Linguistics, Postgraduate Conference .

Gusikhin O, Rychtyckyj N & Filev D (2007) Intelligent systems in the automotive industry: ap-plications and trends. Journal Knowledge and Information Systems 12(2): 147–168.

Haapalainen E, Laurinen P, Junno H, Tuovinen L & Röning J (2005) Methods for classifying spotwelding processes: A comparative study of performance. Innovations in Applied ArtificialIntelligence 412–421.

70

Page 73: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Haapalainen E, Laurinen P, Junno H, Tuovinen L & Röning J (2006) Feature selection for iden-tification of spot welding processes. Proceedings of the 3rd International Conference onInformatics in Control, Automation and Robotics 40–46.

Haapamäki J, Tamminen S & Röning J (2005) Data mining methods in hot steel rolling for scaledefect prediction. Artificial Intelligence and Applications 90–94.

Hadid A & Pietikäinen M (2004) Selecting models from videos for appearance-based face recog-nition. Proceedings of the 17th International Conference on Pattern Recognition 1: 304–308.

Hand DJ & Yu K (2001) Idiot’s bayes U not so stupid after all? International Statistical Review69(3): 385U398.

Harding JA, Shahbaz M, Srinivas & Kusiak A (2006) Data mining in manufacturing: A review.Journal of Manufacturing Science and Engineering 128(4): 969–976.

Harms+Wende (www page) "http://www.harms-wende.de, homepage of the company, refer-enced: 13.1.2009".

Hastie T, Tibshirani R & Friedman J (2001) The Elements of Statistical Learning – Data Mining,Inference, and Prediction. Springer-Verlag, New York Berlin.

Haykin S (1999) Neural Networks, A Comprehensive Foundation. Prentice Hall, Upper SaddleRiver, New Jersey.

He Q & Jin W (2007) Fault detection using the k-nearest neighbor rule for semiconductor manu-facturing processes. IEEE Transactions on Semiconductor Manufacturing 20(4): 345 – 354.

Hechenbichler K & Schliep K (2004) Weighted k-nearest-neighbor techniques and ordinal clas-sification. Collaborative Research Center 386, Discussion Paper 399 .

Hongeng S, Bremond F & Nevatia R (2000) Bayesian framework for video surveillance applica-tion. Proceedings 15th International Conference on Pattern Recognition 1: 164 – 170.

Hou TH, Liu WL & Lin L (2003) Intelligent remote monitoring and diagnosis of manufactur-ing processes using an integrated approach of neural networks and rough sets. Journal ofIntelligent Manufacturing 14(2): 239–253.

Huang HP & Chen TY (2006) On-line rescheduling for semiconductor manufacturing. Proceed-ings of the IEEE International Conference on Automation Science and Engineering 106 –111.

Hunt M, Karnowski T, Kiest C & Villalobos L (2000) Optimizing automatic defect classificationfeature and classifier performance for post-fab yield analysis. Proceedings of the AdvancedSemiconductor Manufacturing Conference and Workshop 116 – 123.

Hur J, Lee H & Baek JG (2006) An intelligent manufacturing process diagnosis system usinghybrid data mining. Industrial Conference on Data Mining 561–575.

Ikeda H, Kato N, Kashimura H & Shimizu M (2002) Unsupervised learning for modulated view-based face detection system. Proceedings of the 9th International Conference on NeuralInformation Processing, ICONIP ’02 2: 670–674.

Ivezic N, Alien JD & Zacharia T (1999) Neural network-based resistance spot welding controland quality prediction. Proceedings of the Second International Conference on IntelligentProcessing and Manufacturing of Materials, IPMM ’99 989–994.

Jankowski N & Grochowski M (2004) Comparison of instances seletion algorithms i. algorithmssurvey. Lecture Notes in Computer Science, Artificial Intelligence and Soft Computing 598–603.

Jia H, Murphey Y, Shi J & Chang TS (2004) An intelligent real-time vision system for surfacedefect detection. Proceedings of the 17th International Conference on Pattern Recognition 3:239 – 242.

71

Page 74: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Junno H, Laurinen P, Haapalainen E, Tuovinen L & Roning J (2005) Resistance spot weldingprocess identification using an extended knn method. Proceedings of the IEEE InternationalSymposium on Industrial Electronics 7–12.

Juutilainen I & Röning J (2004) Modelling the probability of rejection in a qualification test basedon process data. 16th Symposium of IASC, COMPSTAT 1271–1278.

Juutilainen I & Röning J (2006) Planning of strength margins using joint modelling of mean anddispersion. Materials and Manufacturing Processes 21: 367–373.

Juutilainen I, Röning J & Myllykoski L (2003) Modelling the strength of steel plates using re-gression analysis and neural networks. Proceedings of International Conference on Compu-tational Intelligence for Modelling, Control and Automation 681–691.

Kaski S, Kangas J, & Kohonen T (1998) Bibliography of self-organizing map (som) papers:1981-1997. Neural Computing Surveys 1: 102–350.

Kim DM, su Kim K, Park KH, Lee JH & Lee KM (2007) A music recommendation system witha dynamic k-means clustering algorithm. Proceedings of the International Conference onMachine Learning and Applications 399 – 403.

Kim J, Seong D, Jung S & Park J (2004) Integrated cbr framework for quality designing andscheduling in steel industry. Lecture Notes in Computer Science 3155: 645–658.

Kim J, Sim E & Jung S (2005) An automated blowing control system using the hybrid conceptof case based reasoning and neural networks in steel industry. Lecture Notes in ComputerScience 3498: 807–812.

Kim KJ (2006) Artificial neural networks with evolutionary instance selection for financial fore-casting. Expert Systems with Applications 30: 519–526.

Kiviluoto K & Bergius P (1998) Two-level self-organizing maps for analysis of financial state-ments. Proceedings of the 1998 IEEE International Joint Conference on Neural Networks,IEEE World Congress on Computational Intelligence 1: 189 – 192.

Kohonen T (2001) Self-Organizing Maps. Springer-Verlag, Berlin, Heidelberg.Kohonen T, Oja E, Simula O, Visa A & Kangas J (1996) Engineering applications of the self-

organizing map. Proceedings of the IEEE 84(10): 1358–1384.Koskimäki H, Juutilainen I, Laurinen P & Röning J (2007) Detection of the need for a model up-

date in steel manufacturing. Proceedings of the 4th International Conference on Informaticsin Control, Automation and Robotics 55–59.

Kravis S & Irrgang R (2005) A case based system for oil and gas well design with risk assessment.Applied Intelligence 23(1): 39–53.

Kurgan LA & Musilek P (2006) A survey of knowledge discovery and data mining process mod-els. Knowledge Engineering Review 21(1): 1–24.

Laurinen P (2006) A top-down approach for creating and implementing data mining solutions.Ph.D. thesis. Acta Universitatis Ouluensis, Technica, C 246, ISBN 951-42-8125-X ISSN0355-3213.

Laurinen P, Röning J & Tuomela H (2001) Steel slab temperature modelling using neural andbayesian networks. Fourth International ICSC Symposium on Soft Computing and IntelligentSystems for Industry .

Law KH & Kwok LF (2005) False alarm filtering using knn classifier. Lecture Notes in ComputerScience: Information Security Applications 3325: 114–121.

Lee B (2001) Using bayes belief networks in industrial fmea modeling and analysis. Proceedingsof Annual Reliability and Maintainability Symposium 7–15.

Liu H & Motoda H (2002) On issues of instance selection. Data Mining and Knowledge Discov-

72

Page 75: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

ery 6: 115–130.Lopez de Mantaras R, McSherry D, Bridge D, Leake D, Smyth B, Craw S, Faltings B, Maher

M, Cox M, Forbus K, Keane M, Aamodt A & Watson I (2006) Retrieval, reuse, revise, andretention in case-based reasoning. Knowledge Engineering Review –. To be published.

Marimuthu A & Shanmugam A (2008) Intelligent progression for anomaly intrusion detection.Proceedings of the International Symposium on Applied Machine Intelligence and Informat-ics 261 – 265.

Markou M & Singh S (2003) Novelty detection: a review - part 1: statistical approaches. SignalProcessing 83(12): 2481–2497.

Masoud F, Shaikh M & Rabab’ah O (2004) Predicting c++ program quality by using bayesian be-lief networks. Proceedings of International Conference on Information and CommunicationTechnologies: From Theory to Applications 595 – 596.

Melin P & Castillo O (2002) A new approach for quality control of sound speakers combiningtype-2 fuzzy logic and fractal theory. Proceedings of the 2002 IEEE International Conferenceon Fuzzy Systems 2: 825–830.

Mitchell T (1997) Machine Learning. The McGraw-Hill Companies, Inc.Moczulski W & Szulim R (2004) On case-based control of dynamic industrial processes with

the use of fuzzy representation. Engineering Applications of Artificial Intelligence 17(4):371–381.

Moore AW & Lee MS (1994) Efficient algorithms for minimizing cross validation error. Proceed-ings of the International Conference on Machine Learning 190–198.

Munoz-Avila H & Hüllen J (1996) Feature weighting by explaining case-based planning episodes.Proceedings of the Third European Workshop on Advances in Case-Based Reasoning 280–294.

Nevalainen H (1970) Teräsopas.Niskanen M, Silvén O & Kauppinen H (2001) Experiments with som based inspection of

wood. Proceedings of 5th International Conference on Quality Control by Artificial Vision(QCAV’2001) 2: 311–316.

Oja M, Kaski S & Kohonen T (2003) Bibliography of self-organizing map (som) papers: 1998-2001 addendum. Neural Computing Surveys 3: 1–156.

Olsson E, Funk P & Xiong N (2004) Fault diagnosis in industry using sensor readings and case-based reasoning. Journal of Intelligent & Fuzzy Systems 15(1): 41–46.

Ondel O, Boutleux E & Clerc G (2006) A method to detect broken bars in induction machineusing pattern recognition techniques. IEEE Transactions on Industry Applications 42(4):916 – 923.

Osborne H & Bridge DG (1996) A case base similarity framework. EWCBR 309–323.Pirttikangas S, Fujinami K & Nakajima T (2006) Feature selection and activity recognition from

wearable sensors. Lecture Notes in Computer Science: Ubiquitous Computing Systems 4239:516–527.

Piuri V, Scotti F & Roveri M (2005) Computational intelligence in industrial quality control.IEEE International Workshop on Intelligent Signal Processing 4–9.

Pyle D (1999) Data Preparation for Data Mining. Morgan Kaufmann.Rautaruukki (2002) Vastushitsausopas.Rehorn AG, Jiang J & Orban PE (2005) State-of-the-art methods and results in tool condition

monitoring: a review. The International Journal of Advanced Manufacturing Technology26(7-8): 693–710.

73

Page 76: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Reinsch CH (1967) Smoothing by spline functions. Numerische Matematik 10: 177–83.Reinsch CH (1971) Smoothing by spline functions, ii. Numerische Matematik 16: 451–454.Rokach L, Maimon O & Lavi I (2003) Space decomposition in data mining: A clustering ap-

proach. Lecture Notes in Computer Science, Foundations of Intelligent Systems 2871: 24–31.

Sadoyan H, Zakarian A & Mohanty P (2006) Data mining algorithm for manufacturing processcontrol. The International Journal of Advanced Manufacturing Technology 28(3-4): 342–350.

Saward G (1999) Using cases for process modelling: An example from the water supply industry.Lecture Notes in Computer Science 1611: 382 – 387.

Silberschatz A, Stonebraker M & Ullman JD (1990) Database systems: achievements and oppor-tunities. SIGMOD Record 19(4): 6–22.

Silven O, Niskanen M & Kauppinen H (2003) Wood inspection with non-supervised clustering.Machine Vision and Applications 13(5-6): 275–285.

Skalak DB (1994) Prototype and feature selection by sampling and random mutation hill climbingalgorithms. International Conference on Machine Learning 293–301.

Stanzbiegetechnik (www page) http://www.stanzbiegetechnik.at, homepage of the company, ref-erenced: 13.1.2009.

Tamminen S, Juutilainen I & Röning J (2008) Product design model for impact toughness esti-mation in steel plate manufacturing. Proceedings of the International Joint Conference onNeural Networks .

Tamminen S, Pirttikangas S & Röning J (2000) Self-organizing maps in adaptive health moni-toring. Proceedings of International Joint Conference of Neural Networks (IJCNN2000) 4:259–266.

Tsao CC (2008) Prediction of thrust force of step drill in drilling composite material by taguchimethod and radial basis function network. The International Journal of Advanced Manufac-turing Technology 36(1-2): 11–18.

Tuovinen L, Laurinen P, Koskimäki H, Haapalainen E, & Röning J (2007) Building a databaseto support intelligent computational quality assurance of resistance spot welding joints. Pro-ceeding of the IEEE International Symposium on Industrial Electronics 1980–1985.

Wang K (2007) Applying data mining to manufacturing: the nature and implications. Journal ofIntelligent Manufacturing 18(4): 487–495.

Wang XZ (1999) Data mining and knowledge discovery for process monitoring and control.Springer-Verlag.

Wang Y & Wang ZO (2007) A fast knn algorithm for text categorization. Proceedings of Interna-tional Conference on Machine Learning and Cybernetics 6: 3436 – 3441.

Wettschereck D & Aha D (1995) Weighting features. Proceedings of the First InternationalConference on Case-Based Reasoning 347–358.

Wilson DR & Martinez TR (2000) Reduction techniques for instance-based learning algorithms.Machine Learning 38: 257–268.

Wiltschi K, Pinz A & Lindeberg T (2000) An automatic assessment scheme for steel qualityinspection. Machine Vision and Applications 12(3): 113–128.

Woon FL, Knight B, Petridis M & Patel M (2005) <i>cbe-conveyor</i>: A case-based reasoningsystem to assist engineers in designing conveyor systems. Lecture Notes in Computer Science3620: 640 – 651.

Wu X, Kumar V, Quinlan JR, Ghosh J, Yang Q, Motoda H, McLachlan GJ, Ng A, Liu B, Yu PS,

74

Page 77: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Zhou ZH, Steinbach M, Hand DJ, & Steinberg D (2008) Top 10 algorithms in data mining.Knowledge and Information Systems 14(1): 1–37.

Xinmin L, Xiaoyun Z, Yansong Z & Guanlong; C (2007) Weld quality inspection based on onlinemeasured indentation from servo encoder in resistance spot welding. IEEE Transactions onInstrumentation and Measurement 56(4): 1501 – 1505.

Yang M, Zhang H, Fu J & Yan F (2004) A framework for adaptive anomaly detection based onsupport vector data description. Lecture Notes in Computer Science, Network and ParallelComputing 443–450.

Yang T & Tsai TN (2002) Modeling and implementation of a neurofuzzy system for surfacemount assembly defect prediction and control. IIE Transactions 34(7): 637–646.

Zhang W, Tor S & Britton G (2006) Indexing and retrieval in case-based process planning formulti-stage non-axisymmetric deep drawing. The International Journal of Advanced Manu-facturing Technology 28(1-2): 25–32.

Zhou H & Sakane S (2002) Sensor planning for mobile robot localization using bayesian net-work representation and inference. Proceedings of IEEE/RSJ International Conference onIntelligent Robots and System 1: 440–446.

75

Page 78: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

76

Page 79: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

Original publications

I Laurinen P, Junno H, Tuovinen L & Röning J (2004) Studying the Quality of ResistanceSpot Welding Joints Using Bayesian Networks. Artificial Intelligence and Applications:705–711.

II Junno H, Laurinen P, Tuovinen L & Röning J (2004) Studying the Quality of ResistanceSpot Welding Joints Using Self-Organising Maps. Fourth International ICSC Symposiumon Engineering of Intelligent Systems.

III Zettel D, Sampaio D, Link N, Braun A, Peschl M & Junno H (2004) A Self OrganisingMap (SOM) Sensor for the Detection, Visualisation and Analysis of Process Drifts. PosterProceedings of the 27th Annual German Conference on Artificial Intelligence: 175–188.

IV Junno H, Laurinen P, Haapalainen E, Tuovinen L, Röning J , Zettel D, Sampaio D, Link N &Peschl M (2004) Resistance Spot Welding Process Identification and Initialization Based onSelf-Organising Maps. 1st International Conference on Informatics in Control, Automationand Robotics: 296–299.

V Koskimäki H, Laurinen P, Haapalainen E, Tuovinen L & Röning J (2007) Application of theExtended knn Method to Resistance Spot Welding Process Identification and the Benefits ofProcess Information. IEEE Transactions on Industrial Electronics 54(5): 2823–2830.

VI Koskimäki H, Juutilainen I, Laurinen P& Röning J (2008) Detection of Correct Moment toModel Update. Lecture Notes in Electrical Engineering - Informatics in Control, Automa-tion and Robotics: 87–94.

VII Koskimäki H, Juutilainen I, Laurinen P & Röning J (2008) Two-level Clustering Approachto Training Data Instance Selection: a Case Study For the Steel Industry. International JointConference on Neural Networks: 3044–3049.

Reprinted with permission from IASTED (I), ICSC (II), Springer (III and VI), INSTICC

(IV) and IEEE (V and VII).

Original publications are not included in the electronic version of the dissertation.

77

Page 80: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

78

Page 81: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

A C T A U N I V E R S I T A T I S O U L U E N S I S

Distributed byOULU UNIVERSITY LIBRARY

P.O. Box 7500, FI-90014University of Oulu, Finland

Book orders:OULU UNIVERSITY PRESSP.O. Box 8200, FI-90014University of Oulu, Finland

S E R I E S C T E C H N I C A

301. Mäkelä, Juha-Pekka (2008) Effects of handoff algorithms on the performance ofmultimedia wireless networks

302. Teräs, Jukka (2008) Regional science-based clusters. A case study of threeEuropean concentrations

303. Lahti, Markku (2008) Gravure offset printing for fabrication of electronic devicesand integrated components in LTCC modules

304. Popov, Alexey (2008) TiO2 nanoparticles as UV protectors in skin

305. Illikainen, Mirja (2008) Mechanisms of thermomechanical pulp refining

306. Borkowski, Maciej (2008) Digital Δ-Σ Modulation. Variable modulus and tonalbehaviour in a fixed-point digital environment

307. Kuismanen, Kimmo (2008) Climate-conscious architecture—design and windtesting method for climates in change

308. Kangasvieri, Tero (2008) Surface-mountable LTCC-SiP module approach forreliable RF and millimetre-wave packaging

309. Metsärinta, Maija-Leena (2008) Sinkkivälkkeen leijukerrospasutuksen stabiilisuus

310. Prokkola, Jarmo (2008) Enhancing the performance of ad hoc networking bylower layer design

311. Löytynoja, Mikko (2008) Digital rights management of audio distribution in mobilenetworks

312. El Harouny, Elisa (2008) Historiallinen puukaupunki suojelukohteena jaelinympäristönä. Esimerkkeinä Vanha Porvoo ja Vanha Raahe. Osa 1

312. El Harouny, Elisa (2008) Historiallinen puukaupunki suojelukohteena jaelinympäristönä. Esimerkkeinä Vanha Porvoo ja Vanha Raahe. Osa 2

313. Hannuksela, Jari (2008) Camera based motion estimation and recognition forhuman-computer interaction

314. Nieminen, Timo (2009) Detection of harmful microbes and their metaboliteswith novel methods in the agri-food production chain

315. Marjala, Pauliina (2009) Työhyvinvoinnin kokemukset kertomuksellisinaprosesseina–narratiivinen arviointitutkimus

316. Ahola, Juha (2009) Reaction kinetics and reactor modelling in the design ofcatalytic reactors for automotive exhaust gas abatement

C317etukansi.fm Page 2 Wednesday, February 11, 2009 1:43 PM

Page 82: Utilizing similarity information in industrial applicationscc.oulu.fi/~kamahei/y/casr/vk/koskimaki.pdf · technica acta universitatis ouluensis technica oulu 2009 c 317 heli koskimäki

ABCDEFG

UNIVERS ITY OF OULU P.O.B . 7500 F I -90014 UNIVERS ITY OF OULU F INLAND

A C T A U N I V E R S I T A T I S O U L U E N S I S

S E R I E S E D I T O R S

SCIENTIAE RERUM NATURALIUM

HUMANIORA

TECHNICA

MEDICA

SCIENTIAE RERUM SOCIALIUM

SCRIPTA ACADEMICA

OECONOMICA

EDITOR IN CHIEF

PUBLICATIONS EDITOR

Professor Mikko Siponen

University Lecturer Elise Kärkkäinen

Professor Hannu Heusala

Professor Olli Vuolteenaho

Senior Researcher Eila Estola

Information officer Tiina Pistokoski

University Lecturer Seppo Eriksson

Professor Olli Vuolteenaho

Publications Editor Kirsti Nurkkala

ISBN 978-951-42-9038-1 (Paperback)ISBN 978-951-42-9039-8 (PDF)ISSN 0355-3213 (Print)ISSN 1796-2226 (Online)

U N I V E R S I TAT I S O U L U E N S I SACTAC

TECHNICA

U N I V E R S I TAT I S O U L U E N S I SACTAC

TECHNICA

OULU 2009

C 317

Heli Koskimäki

UTILIZING SIMILARITY INFORMATION IN INDUSTRIAL APPLICATIONS

FACULTY OF TECHNOLOGY,DEPARTMENT OF ELECTRICAL AND INFORMATION ENGINEERING,UNIVERSITY OF OULU

C 317

ACTA

Heli K

oskimäki

C317etukansi.fm Page 1 Wednesday, February 11, 2009 1:43 PM