2 - bayesian decision theoryyqi/lect/patternrecognition3.pdf · decision boundary r 1 r 2 in an...

18
1 1 BAYESIAN DECISION THEORY 2 Introduction All the patterns to recognize belong to J different classes, j=1,2,..., J:

Upload: others

Post on 17-Jun-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

1

1

BAYESIAN DECISION THEORY

2

Introduction

All the patterns to recognize belong to J differentclasses, j=1,2,..., J:

Page 2: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

2

3

Prior probability

We assume that there is some a priori probability(prior)

Properties:

Decision rule using only the value of the prior probabilities:

Decide ifDecide if

(Knowledge we haveabout one class beforecarrying out anexperiment).

The probability of havingan error in classificationis the lower value of π1, π2.

4

Example of classification using prior probability

Example:

Classification problem: discriminate between healthypeople or people with anemia (Blutarmut)

Prior knowledge:

•90% of the people is healthy:

•10% of the people is ill:

If we have to classify a new patient, which is his/her class?

•Decision rule: THAT CLASS WITH GREATER PRIOR PROBABILTY

•Decide as

If we have no other information, we have to take this decision.

Page 3: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

3

5

Example of classification using prior probability

Example:

Classification problem: discriminate between healthy people or people with anemia (Blutarmut)

However, nobody will be satisfied if our family doctor would decide our state of health without looking at us or asking us a blood test.

It is necessary to use more information: we need measurements relative to the patterns.

6

Class-conditional probability density

Usually, there is additional information: the value of the observation to classify, x.

Considerations:

•Pattern values relative to a class must be essentially different to that of the other classes.

•Patterns of the same class are not exactly equal. (variability of class patterns)

Page 4: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

4

7

Class-conditional probability density

The variability of the measurements is expressed as a random variable x, and its probability density function depends on the class ωj.

p(x| ωj) is the class-conditional probability density function, the probability function for xgiven that the class is ωj.

For each class ωj: 1)|( =∫x

jxp ω

8

Example of classification using class-conditional probability

Example:

Classification problem: discriminate between healthypeople or people with anemia (Blutarmut)

•We have the results of a blood test, so we know theamount of red blood cells.

•The amount of red blood cells is the random variable (x) (We do not have the same number of red blod cells thanother people).

•This variable has a gaussian distribution.

Page 5: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

5

9

Example of classification using class-conditional probability

Example:

Classification problem: discriminate between healthypeople or people with anemia (Blutarmut)

x (# red blood cells)

Pro

bab

ility

Class 1: healthy

Class 2: ill

p(x|healthy)

p(x|ill)

10

Example of classification using class-conditional probability

Example:

Classification problem: discriminate between healthypeople or people with anemia (Blutarmut)

Blood test: 4.500.000 red blood cells

So, the patient is healthy.

p(x=4.500.000 | healthy) > p(x=4.500.000 | ill)

If we consider the patient is healty, the probaility he has 4.5 milion red blood cells is higher than if weconsider he is ill, with this number of red blood cells.

Page 6: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

6

11

Posterior probability

Suppose that we know both the prior probabilities and the conditional densities:

And suposse further that we measure an observation x

How we can use all this information together?

Bayes formula

12

Posterior probability

Bayes formula shows that by observing the value of x we can convert the prior probability πj to the a posteriori probability (or posterior) P(ωj|x ).

P(ωj|x ) : the probability of the pattern belonging to class ωj given that the feature value x has been measured.

We call p(x |ωj) the likelihood of ωj with respect to x; πj is the prior; and p(x) is called evidence.

Notice that it is the product of the likelihood and the prior probability that is most important in determining the posterior probability; the evidencefactor, p(x) , can be viewed as merely a scale factor that guarantees that the posterior probabilities sum to one, as all good probabilities must.

Page 7: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

7

13

Posterior probabilityEffect of the prob. a priori over the prob. a posteriori

Class 1

Class 2

14

Bayes decision rule

Decide ω1 if P(ω1|x)>P(ω2|x)

Decide ω2 if P(ω2|x)>P(ω1|x)

Page 8: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

8

15

Bayes decision rule

Decide ω1 if p(x|ω1)π1 > p(x|ω2)π2

Decide ω2 if p(x|ω2)π2 > p(x|ω1)π1

Or an equivalent rule:

16

Maximun likelihood rule

Decide ω1 if p(x|ω1) > p(x|ω2)

Decide ω2 if p(x|ω2) > p(x|ω1)

If π1 = π2 :

Page 9: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

9

17

Example of classification using the Bayes rule

Example:

Classification problem: discriminate between healthy people or people with anemia (Blutarmut)

Decide ω1 if P(ω1|x)>P(ω2|x)

Decide ω2 if P(ω2|x)>P(ω1|x)

Blood test: 3.500.000 red blood cells (x)

So, the patient is ill.

Prior knowledge:

•90% of the people is healthy:

•10% of the people is ill:

Bayes rule

18

Extension to multidimensional patterns

Bayes rule or Bayes classifier:

Select ωi if P(ωi|x)> P(ωj|x) for all j≠i

These concepts can be extended to study more complexand realistic situations:

Page 10: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

10

19

Decision boundary

R1 R2

In an uni-dimensional case, the decision boundary is just onepoint, and the decision regions are intervals in the x-axis.

20

Decision boundary

Class 2

Class 1

Different values of ‘prior probabilities’ make the decision boundary to move.

Page 11: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

11

21

Error in classification

In a two-category classification and onedimension patterns, the possible errors are:

•x is in R1 and its real class is ω2

•x is in R2 and its real class is ω1

22

Error in classification

Page 12: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

12

23

Error in classification

MINIMUM ERROR

24

Error in classification

ERROR NOT MINIMUM

If the decision boundary is not chosen correctly, the error isnot minimum.

Page 13: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

13

25

Error in classification

The Bayes classifier minimizes the average probability of error, so the best choice is touse the Bayes rule as the classifier of the pattern recognition system.

However, in most practical cases, the class-conditional probabilities are not known, and that fact makes impossible the use of the Bayes rule.

26

An example

PATTERN RECOGNITION SYSTEM:

CONSENT A MORTGAGE OR NOT

Variables:

•Age

•Wage bill

Classes:

•Yes, he/she will return the money (GOOD PAYER)

•No, he/she will NOT return the money (BAD PAYER)

We have 20 examples, 10 per each class

Page 14: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

14

27

An example

28

An example

0 20 40 60 800

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8

2

Age

# pe

ople

NO

0 20 40 60 800

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8

2

Age

# pe

ople

YES

Page 15: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

15

29

An example

0 1000 2000 30000

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8

2

Wage bill

# pe

ople

NO

0 1000 2000 30000

0.5

1

1.5

2

2.5

3

Wage bill

# pe

ople

YES

30

An example

10 20 30 40 50 60 70 800

500

1000

1500

2000

2500

Age

Wag

e bi

ll

MortatgeNOYES

Page 16: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

16

31

An example

Prior probability: 0.5 for both classes

We don´t know the class-conditional probability density functions of the variables, so we can´t use the Bayes rule.

We will use a simple classifier:

Euclidean distance to the mean of each class (nearest neighbour)

32

10 20 30 40 50 60 70 800

500

1000

1500

2000

2500

Age

Wag

e bi

ll

Mortatge NO YES

An example

(33, 589)

(62, 1745)

Class 1

Class 2

Page 17: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

17

33

An example

34

An example

Confusion Matrix

Recognition Rate

Page 18: 2 - Bayesian decision theoryyqi/lect/Patternrecognition3.pdf · Decision boundary R 1 R 2 In an uni-dimensional case, the decision boundary is just one point, and the decision regions

18

35

Performance of the classifier

Measuring the performance of the classifieron the training images is not fair (too optimistic):

•The goal is to know the expected performance of the classifier when it is fed with new data.

•Example: the goal is to know if a face recognition system trained with images of a certain person will recognize him/her the following day (using a previously unseen image).

•In brief, the performance of a classifier has to be tested with previously unseen examples (not used for training).

36

Performance of the classifier

Training set: the set of examples used for training the classifier.

Test set: the set of examples used to measure the expected classification accuracy.

Images (or examples) of the test set should never belong to the training set.