chapter 3: output: knowledge...

50
Chapter 3: Output: Knowledge Representation 1 Chapter 3: Output: Knowledge Representation

Upload: others

Post on 02-Jun-2020

23 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 1

Chapter 3: Output: Knowledge Representation

Page 2: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 2

Output: representing structural patterns

� Many different ways of representing patterns

– Decision trees, classification rules, …

� Also called “knowledge” representation

� Representation determines inference method

� Understanding the output is the key to

understanding the underlying learning

methods

� Different types of output for different learning

problems (e.g. classification, regression, …)

Page 3: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 3

Output: Knowledge representation

� Decision tables

� Decision trees

� Classification rules

� Association rules

� Rules with exceptions

� Rules involving relations

� Trees for numeric prediction

� Instance-based representation

� Clusters

Page 4: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 4

3.1 Decision tables

Page 5: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 5

Decision tables

� Simplest way of representing output

� Use the same format as input!

� Main problem: selecting the right attributes

Page 6: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 6

Decision table for the weather problem

Page 7: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 7

3.2 Decision trees

Page 8: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 8

Decision trees

� Nodes involve testing a particular attribute

� Usually, attribute value is compared to constant

� Other possibilities:

– Comparing values of two attributes

– Using a function of one or more attributes

� Unknown instance is routed down the tree

� Leaves assign classification, set of

classifications, or probability distribution to

instances

Page 9: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 9

Nominal and numeric attributes

� Nominal:

– Number of children usually equal to number values

– Attribute won’t get tested more than once

– Other possibility: division into two subsets

� Numeric:

– Test whether value is greater or less than constant

– Attribute may get tested several times

– Other possibility: three-way split (or multi-way split)

�Integer: less than, equal to, greater than

�Real: below, within, above

Page 10: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 10

Missing values

� Does absence of value have some

significance?

� Yes => “missing” is a separate value

� No => “missing” must be treated in a special

way

– E.g. assign instance to most popular branch

Page 11: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 11

Decision tree for the labor data

Page 12: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 12

3.3 Classification rules

Page 13: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 13

Classification rules

� Popular alternative to decision trees

� Antecedent (precondition): a series of tests (just like the tests at the nodes of a decision tree)

� Tests are usually logically ANDed together (but may also be general logical expressions)

� Consequent (conclusion): set of classes or probability distribution assigned by rule

� Individual rules are often logically ORedtogether

– Conflicts arise if different conclusions apply

Page 14: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 14

From trees to rules

� Easy: converting a tree into a set of rules

– One rule for each leaf:

� Antecedent contains a condition for every node on

the path from the root to the leaf

� Consequent is class assigned by the leaf

� Produces rules that are very clear

– Doesn’t matter in which order they are

executed

� But: resulting rules are unnecessarily complex

– It needs to remove redundant tests/rules

Page 15: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 15

From rules to trees

� More difficult: transforming a rule set into a

tree

– Tree cannot easily express disjunction

between rules

� Example: rules which test different attributes

� Corresponding tree contains identical subtrees

=> “replicated subtree problem”

Page 16: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 16

A tree for a simple disjunction

Page 17: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 17

The exclusive-or problem

Page 18: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 18

A tree with a replicated subtree

� there are four attributes, x, y, z, and w,

� each can be 1, 2, or 3

Page 19: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 19

A tree with a replicated subtree (II)

Page 20: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 20

Nuggets of knowledge

� Are rules independent pieces of knowledge?

(It seems easy to add a rule to an existing rule

base.)

� Problem: ignores how rules are executed

� Two ways of executing a rule set:

– Ordered set of rules (“decision list”)

� Order is important for interpretation

– Unordered set of rules

� Rules may overlap and lead to different

conclusions for the same instance

Page 21: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 21

Interpreting rules

� What if two or more rules conflict?

– Give no conclusion at all?

– Go with rule that is most popular on training

data?

– …

� What if no rule applies to a test instance?

– Give no conclusion at all?

– Go with class that is most frequent in training

data?

– …

Page 22: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 22

Special case: boolean class

� Assumption: if instance does not belong to class “yes”, it belongs to class “no”

� Trick: only learn rules for class “yes” and use default rule for “no”

� Order of rules is not important. No conflicts!

� Rule can be written in disjunctive normal form

Page 23: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 23

3.4 Association rules

Page 24: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 24

Association rules

� Association rules…

– … can predict any attribute and combinations

of attributes

– … are not intended to be used together as a

set

� Large number of possible associations

– Output needs to be restricted to show only the

most predictive associations

– only those with high support and high

confidence

Page 25: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 25

Support and confidence of a rule

� Support: number of instances predicted correctly

� Confidence: number of correct predictions, as

proportion of all instances that rule applies to

� Example: 4 cool days with normal humidity

� Support = 4, confidence = 100%

� Normally: minimum support and confidence prespecified (e.g. support >= 2 and confidence >=

95% for weather data)

Page 26: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 26

Interpreting association rules

� Interpretation is not obvious:

� is not the same as

� It means that the following also holds:

Page 27: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 27

3.5 Rules with exceptions

Page 28: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 28

Rules with exceptions

� Idea: allow rules to have exceptions

� Example: rule for iris data

� New instance:

� Modified rule:

Page 29: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 29

A more complex example

� Exceptions to exceptions to exceptions …

Page 30: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 30

Advantages of using exceptions

� Rules can be updated incrementally

– Easy to incorporate new data

– Easy to incorporate domain knowledge

� People often think in terms of exceptions

� Each conclusion can be considered just in the context of rules and exceptions that lead to it

– Locality property is important for

understanding large rule sets

– “Normal” rule sets don’t offer this advantage

Page 31: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 31

More on exceptions

� Default...except if...then...

is logically equivalent to

� if...then...else(where the else specifies what the default did)

� But: exceptions offer a psychological

advantage

– Assumption: defaults and tests early on apply

more widely than exceptions further down

– Exceptions reflect special cases

Page 32: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 32

3.6 Rules involving relations

Page 33: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 33

Rules involving relations

� So far: all rules involved comparing an

attribute value to a constant

(e.g. temperature < 45)

� These rules are called “propositional”

� What if problem involves relationships

between examples

(e.g. family tree problem from above)?

– Can’t be expressed with propositional rules

– More expressive representation required

Page 34: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 34

The shapes problem

Page 35: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 35

A propositional solution

Page 36: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 36

A relational solution

� Comparing attributes with each other

� Rules of this form is called relational

� Generalizes better to new data

� Standard relations: =, <, >

� Simple solution: add extra attributes

(e.g. a binary attribute is width < height?)

Page 37: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 37

3.7 Trees for numeric prediction

Page 38: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 38

Trees for numeric prediction

� Regression: the process of computing an

expression that predicts a numeric quantity

� Regression tree: “decision tree” where each leaf predicts a numeric quantity

– Predicted value is average value of training

instances that reach the leaf

� Model tree: “regression tree” with linear regression models at the leaf nodes

Page 39: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 39

Linear regression for the CPU data

Page 40: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 40

Regression tree for the CPU data

Page 41: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 41

Model tree for the CPU data

Page 42: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 42

3.8 Instance-based representation

Page 43: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 43

Instance-based representation

� Simplest form of learning: rote learning

– Training instances are searched for instance

that most closely similar to new instance

– The instances themselves represent the

knowledge

– Also called instance-based learning

� Instance-based learning is lazy learning

� Similarity function defines what’s “learned”

Page 44: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 44

Instance-based learning methods

� Nearest-neighbor method:

– each new instance is compared with existing

ones using a distance metric, and the closest

existing instance is used to assign the class to

the new one

� k-nearest-neighbor method:

– more than one nearest neighbor is used, and

the majority class of the closest k neighbors

Page 45: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 45

3.9 Clusters

Page 46: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 46

Clusters

� The output takes the form of a diagram that shows how the instances fall into clusters.

� Different cases:

– Simple 2D representation: involves associating a

cluster number with each instance

– Venn diagram: allow one instance to belong to more

than one cluster

– Probabilistic assignment: associate instances with

clusters probabilistically

– Dendrogram: produces a hierarchical structure of

clusters (dendron is the Greek word for tree)

Page 47: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 47

Representing clusters (I)

Page 48: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 48

Representing clusters (II)

Page 49: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 49

The end ofChapter 3: Output: Knowledge

Representation

Page 50: Chapter 3: Output: Knowledge Representationwebpages.iust.ac.ir/yaghini/Courses/Application_IT_Fall2007/Chapter… · Chapter 3: Output: Knowledge Representation 30 Advantages of using

Chapter 3: Output: Knowledge Representation 50

The distance function

� Simplest case: one numeric attribute

– Distance is the difference between the two attribute values involved

� Several numeric attributes: normally, Euclidean distance is used and attributes are normalized

� Nominal attributes: distance is set to 1 if values are different, 0 if they are equal

– E.g. distances between the values red, green, and blue?

– the distance between red and red is zero and between red and green is one.

� Are all attributes equally important?

– Weighting the attributes might be necessary