uva$cs$6316$$ –$fall$2015$graduate:$$ machine$learning ...€¦ · • k-nearest neighbor...

27
UVA CS 6316 – Fall 2015 Graduate: Machine Learning Lecture 16: K@nearest@neighbor Classifier / Bias@Variance Tradeoff Dr. Yanjun Qi University of Virginia Department of Computer Science 10/27/15 Dr. Yanjun Qi / UVA CS 6316 / f15 1 Announcements: Rough Plan HW3: due on Nov. 8th midnight Midphase Project Report : due on Nov. 4 th Late Midterm: Open note / Open lecture Nov. 18 th / conflicts with many students’ conference trips Nov. 23 rd ???? / conflicts ??? HW4: 20 samples quesWons for the preparaWon for exam Due depending on LateYMidterm Date ; If 23 rd , due on 20 th 10/27/15 2 Dr. Yanjun Qi / UVA CS 6316 / f15

Upload: others

Post on 30-Sep-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

UVA$CS$6316$$–$Fall$2015$Graduate:$$Machine$Learning$$

$$Lecture$16:$K@nearest@neighbor$Classifier$/$Bias@Variance$Tradeoff$

Dr.$Yanjun$Qi$$

University$of$Virginia$$$

Department$of$Computer$Science$$

$10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

1$

Announcements:$Rough$Plan$$•  HW3:$due$on$Nov.$8th$midnight$•  Midphase$Project$Report$:$due$on$Nov.$4th$$$$•  Late$Midterm:$$

–  Open$note$/$Open$lecture$–  Nov.$18th$/$conflicts$with$many$students’$conference$trips$$–  Nov.$23rd$????$/$conflicts$???$$$

•  HW4:$$–  20$samples$quesWons$for$the$preparaWon$for$exam$–  Due$depending$on$LateYMidterm$Date$;$If$23rd,$due$on$20th$$

10/27/15$ 2$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 2: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Where$are$we$?$!$$Five$major$secWons$of$this$course$

" $Regression$(supervised)$" $ClassificaWon$(supervised)$" $Unsupervised$models$$" $Learning$theory$$" $Graphical$models$$

10/27/15$ 3$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Where$are$we$?$!$$Three$major$secWons$for$classificaWon

•  We can divide the large variety of classification approaches into roughly three major types

1. Discriminative - directly estimate a decision rule/boundary - e.g., logistic regression, support vector machine, decisionTree 2. Generative: - build a generative statistical model - e.g., naïve bayes classifier, Bayesian networks 3. Instance based classifiers - Use observation directly (no models) - e.g. K nearest neighbors

10/27/15$ 4$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 3: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Today$:$$

#  KYnearest$neighbor$$# Model$SelecWon$/$Bias$Variance$Tradeoff$$

$$

10/27/15$ 5$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

•  Basic idea: –  If it walks like a duck, quacks like a duck, then

it�s probably a duck

Nearest neighbor classifiers

training samples

test sample

compute distance

choose k of the �nearest� samples

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

6$

Page 4: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Unknown recordNearest neighbor classifiers

Requires$three$inputs:$1.  The$set$of$stored$

training$samples$2.  Distance$metric$to$

compute$distance$between$samples$

3.  The$value$of$k,$i.e.,$the$number$of$nearest$neighbors$to$retrieve$

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

7$

Nearest neighbor classifiers

To$classify$unknown$sample:$1.  Compute$distance$to$

other$training$records$2.  IdenWfy$k$nearest$

neighbors$$3.  Use$class$labels$of$nearest$

neighbors$to$determine$the$class$label$of$unknown$record$(e.g.,$by$taking$majority$vote)$

Unknown record

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

8$

Page 5: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Definition of nearest neighbor

X X X

(a) 1-nearest neighbor (b) 2-nearest neighbor (c) 3-nearest neighbor

$$$$kYnearest$neighbors$of$a$sample$x$are$datapoints$that$have$the$k$smallest$distances$to$x$

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

9$

1-nearest neighbor

Voronoi$diagram:$$partitioning of a

plane into regions based on distance to points in a specific subset of the plane.

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

10$

Page 6: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

•  Compute distance between two points: – For instance, Euclidean distance

•  Options for determining the class from nearest neighbor list – Take majority vote of class labels among the

k-nearest neighbors – Weight the votes according to distance

•  example: weight factor w = 1 / d2

Nearest neighbor classification

∑ −=i

ii yxd 2)(),( yx

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

11$

•  Choosing the value of k: –  If k is too small, sensitive to noise points –  If k is too large, neighborhood may include points

from other classes

Nearest neighbor classification

X

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

12$

Page 7: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

•  Scaling issues – Attributes may have to be scaled to prevent

distance measures from being dominated by one of the attributes

– Example: •  height of a person may vary from 1.5 m to 1.8 m •  weight of a person may vary from 90 lb to 300 lb •  income of a person may vary from $10K to $1M

Nearest neighbor classification

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

13$

•  Problem with Euclidean measure: – High dimensional data

•  curse of dimensionality

– Can produce counter-intuitive results

Nearest neighbor classification…

1 1 1 1 1 1 1 1 1 1 1 0

0 1 1 1 1 1 1 1 1 1 1 1

1 0 0 0 0 0 0 0 0 0 0 0

0 0 0 0 0 0 0 0 0 0 0 1 vs$

d = 1.4142 d = 1.4142 $$$ $one$soluWon:$normalize$the$vectors$to$unit$length$

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

14$

Page 8: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

•  k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown samples is relatively

expensive.

•  k-Nearest neighbor classifier is a local model, vs. global model of linear classifiers.

Nearest neighbor classification

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

15$

K-Nearest Neighbor

classification

Local Smoothness

NA

NA

Training Samples

Task

Representation

Score Function

Search/Optimization

Models, Parameters

10/27/15$ 16$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 9: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Decision boundaries in global vs. local models

linear regression •  global •  stable •  can be inaccurate

15-nearest neighbor

1-nearest neighbor

•  local •  accurate •  unstable

What ultimately matters: GENERALIZATION

10/27/15$ 17$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

KYNearestYNeighbours$for$ClassificaWon$(Extra) $

• $K$acts$as$a$smother$• $For$$$$$$$$$$$$$$$$,$the$error$rate$of$the$1YnearestYneighbour$classifier$is$never$more$than$twice$the$opWmal$error$(obtained$from$the$true$condiWonal$class$distribuWons).$

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

18$

Page 10: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

19

KNN METHODS IN HIGH DIMENSIONS (Extra)

•  In high dimensions, all sample points are close to the edge of the sample

•  N data points uniformly distributed in a p-dimensional unit ball centered at the origin

•  Median distance from the closest point to the origin

•  d(10,500) = 0.52 –  More than half the way to the boundary (unit ball’s boundary

edge is 1 distance to the origin)

pN

Npd/1/1

211),( ⎟⎟⎠

⎞⎜⎜⎝

⎛−=

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

20$

Vs. Locally weighted regression

•  aka locally weighted regression, locally linear regression, LOESS, …

10/27/15$

linear_func(x)Y>y$$!$could$represent$$only$the$neighbor$region$of$x_0$

Use$RBF$funcWon$to$pick$out/emphasize$$the$neighbor$region$of$x_0$!$$

Kλ (xi, x0 )

Kλ (xi, x0 )

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 11: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

10/27/15

Vs. Locally weighted regression

•  Separate weighted least squares at each target point x0:

minα (x0 ),β (x0 )

Kλ (xi, x0 )[yi −α(x0 )−β(x0 )xi ]2

i=1

N

0000 )(ˆ)(ˆ)(ˆ xxxxf βα +=

x0$

Kτ (xi,x0) = exp −(xi − x0)

2

2τ 2"

#$

%

&'

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Today$:$$

#  KYnearest$neighbor$$# Model$SelecWon$/$Bias$Variance$Tradeoff$

#  EPE$#  DecomposiWon$of$MSE$$#  BiasYVariance$tradeoff$#  High$bias$?$High$variance$?$How$to$respond$?$$$$

10/27/15$ 22$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 12: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

e.g. Training Error from KNN, Lesson Learned

•  When k = 1, •  No misclassifications (on

training): Overtraining

•  Minimizing training error is not always good (e.g., 1-NN)

X1

X2

1-nearest neighbor averaging

5.0ˆ =Y

5.0ˆ <Y

5.0ˆ >Y

10/27/15$ 23$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Statistical Decision Theory

•  Random input vector: X •  Random output variable:Y •  Joint distribution: Pr(X,Y ) •  Loss function L(Y, f(X))

•  Expected prediction error (EPE): • 

EPE( f ) = E(L(Y, f (X))) = L(y, f (x))∫ Pr(dx,dy)

e.g. = (y − f (x))2∫ Pr(dx,dy)Consider

population distribution

e.g. Squared error loss (also called L2 loss ) 10/27/15$ 24$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 13: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Expected prediction error (EPE)

•  For L2 loss:

•  For 0-1 loss: L(k, �) = 1-�kl

EPE( f ) = E(L(Y, f (X))) = L(y, f (x))∫ Pr(dx,dy)Consider joint distribution

!!

f̂ (X )=C k !if!Pr(C k |X = x)=max

g∈CPr(g|X = x)

Bayes Classifier

e.g. = (y− f (x))2∫ Pr(dx,dy)

Conditional mean !! f̂ (x)=E(Y |X = x)

under L2 loss, best estimator for EPE (Theoretically) is :

e.g. KNN NN methods are the direct implementation (approximation )

10/27/15$ 25$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

LR VS. KNN FOR MINIMIZING EPE

•  We know under L2 loss, best estimator for EPE (Theoretically) is :

•  Two simple approaches using different approximations: –  Least squares assumes that f(x) is well approximated by a globally

linear function –  Nearest neighbors assumes that f(x) is well approximated by a locally

constant function.

26

Conditional mean )|(E)( xXYxf ==

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 14: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Review : WHEN EPE USES DIFFERENT LOSS

Loss Function Estimator

L2

L1

0-1

(Bayes classifier / MAP)

27 10/27/15$ Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Today$:$$

#  KYnearest$neighbor$$# Model$SelecWon$/$Bias$Variance$Tradeoff$

#  EPE$#  DecomposiWon$of$MSE$$#  BiasYVariance$tradeoff$#  High$bias$?$High$variance$?$How$to$respond$?$$$$

10/27/15$ 28$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 15: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Decomposition of EPE

– When additive error model: – Notations

•  Output random variable: •  Prediction function: •  Prediction estimator:

Irreducible / Bayes error

29

MSE component of f-hat in estimating f

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

BiasYVariance$TradeYoff$for$EPE: $

EPE (x_0) = noise2 + bias2 + variance

Unavoidable error

Error due to incorrect

assumptions

Error due to variance of training

samples

Slide$credit:$D.$Hoiem$10/27/15$ 30$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 16: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

BIAS AND VARIANCE TRADE-OFF for MSE (more general setting !!! ):

•  Bias –  measures accuracy or quality of the estimator –  low bias implies on average we will accurately

estimate true parameter or func from training data

•  Variance –  Measures precision or specificity of the estimator –  Low variance implies the estimator does not change

much as the training set varies

31

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

BIAS AND VARIANCE TRADE-OFF for MSE of parameter estimation:

32

Error due to incorrect

assumptions

Error due to variance of training

samples

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 17: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

e.g., BIAS AND VARIANCE IN KNN (Extra)

•  Prediction

•  Bias

•  Variance

•  When under data model:

33

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Today$:$$

#  KYnearest$neighbor$$# Model$SelecWon$/$Bias$Variance$Tradeoff$

#  EPE$#  DecomposiWon$of$MSE$$#  BiasYVariance$tradeoff$#  High$bias$?$High$variance$?$How$to$respond$?$$$$

10/27/15$ 34$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 18: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Bias-Variance Tradeoff / Model Selection

35

underfit region overfit region

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Which$is$the$best?$

Highest$Bias$Lowest$$variance$Model$complexity$=$low$

Medium$Bias$Medium$Variance$

Model$complexity$=$medium$

Smallest$Bias$Highest$variance$Model$complexity$=$high$

Why$not$choose$the$method$with$the$$best$fit$to$the$data?$

How$well$are$you$going$to$$predict$future$data?$

36$10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 19: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

BiasYVariance$TradeYoff$•  Models$with$too$few$parameters$are$inaccurate$because$of$a$large$bias$(not$enough$flexibility).$

•  Models$with$too$many$parameters$are$inaccurate$because$of$a$large$variance$(too$much$sensiWvity$to$the$sample$randomness).$

Slide$credit:$D.$Hoiem$10/27/15$ 37$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Training$vs$Test$Error$•  Training$error$

can$always$be$reduced$when$increasing$model$complexity,$$

•  But$risks$overYfinng$and$generalize$poorly.$

$$

Expected$Test$Error$

Expected$Training$Error$

38$

Page 20: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Mean$of$the$

Regression: Complexity versus Goodness of Fit

x

y

x

y

x

y

x

y

Too simple?

Too complex ? About right ?

Training data

What ultimately matters: GENERALIZATION

Low$Variance$/$High$Bias$

Low$Bias$$/$High$Variance$10/27/15$ 39$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Classification,

Decision boundaries in global vs. local models

linear regression •  global •  stable •  can be inaccurate

15-nearest neighbor

1-nearest neighbor

KNN •  local •  accurate •  unstable

What ultimately matters: GENERALIZATION 10/27/15$ 40$

Low$Variance$/$High$Bias$

Low$Bias$$/$High$Variance$

Low$Variance$/$High$Bias$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 21: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Model �bias� & Model �variance�

•  Middle RED: –  TRUE function

•  Error due to bias: –  How far off in general

from the middle red

•  Error due to variance: –  How wildly the blue

points spread

10/27/15$ 41$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

need$to$make$assumpWons$that$are$able$to$generalize$

•  Components of generalization error –  Bias: how much the average model over all training sets differ from

the true model? •  Error due to inaccurate assumptions/simplifications made by the

model –  Variance: how much models estimated from different training sets

differ from each other •  Underfitting: model is too “simple” to represent all the

relevant class characteristics –  High bias and low variance –  High training error and high test error

•  Overfitting: model is too “complex” and fits irrelevant characteristics (noise) in the data –  Low bias and high variance –  Low training error and high test error

Slide$credit:$L.$Lazebnik$10/27/15$ 42$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 22: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

MODEL COMPLEXITY CONTROL, EXAMPLES (Extra)

•  Regularization (Bayesian)

•  Kernel methods and local regression

•  Basis functions

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

43$

Today$:$$

#  KYnearest$neighbor$$# Model$SelecWon$/$Bias$Variance$Tradeoff$

#  EPE$#  DecomposiWon$of$MSE$$#  BiasYVariance$tradeoff$#  High$bias$?$High$variance$?$How$to$respond$?$$$$

10/27/15$ 44$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 23: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

High$variance$

10/27/15$ 45$

•  Test$error$sWll$decreasing$as$m$increases.$Suggests$larger$training$set$will$help.$

•  Large$gap$between$training$and$test$error.$$

•  Low training error and high test error Slide$credit:$A.$Ng$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Could$also$use$CrossV$Error$

How$to$reduce$variance?$

•  Choose$a$simpler$classifier$

•  Regularize$the$parameters$

•  Get$more$training$data$

•  Try$smaller$set$of$features$$

Slide$credit:$D.$Hoiem$10/27/15$ 46$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 24: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

High$bias$

10/27/15$ 47$

•$Even$training$error$is$unacceptably$high.$•$Small$gap$between$training$and$test$error.$

High training error and high test error Slide$credit:$A.$Ng$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Could$also$use$CrossV$Error$

How$to$reduce$Bias$?$$

48$

•  E.g.$$$Y  Get$addiWonal$features$

Y  Try$adding$basis$expansions,$e.g.$polynomial$

Y  Try$more$complex$learner$$$

10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 25: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

For$instance,$if$trying$to$solve$“spam$detecWon”$using$(Extra) $$

10/27/15$ 49$

L2$$Y$

Slide$credit:$A.$Ng$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

If$performance$is$not$as$desired$$

Model$SelecWon$and$Assessment$

•  Model$SelecWon$–  EsWmaWng$performances$of$different$models$to$choose$the$best$one$$

$

•  Model$Assessment$–  Having$chosen$a$model,$esWmaWng$the$predicWon$error$on$new$data$

50$10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 26: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

Model$SelecWon$and$Assessment$(Extra) $

•  Data$Rich$Scenario:$Split$the$dataset$

•  Insufficient$data$to$split$into$3$parts$–  Approximate$validaWon$step$analyWcally$

•  AIC,$BIC,$MDL,$SRM$–  Efficient$reuse$of$samples$

•  Cross$validaWon,$bootstrap$

Model Selection Model assessment

Train Validation Test

51$10/27/15$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Today$Recap:$$

#  KYnearest$neighbor$$# Model$SelecWon$/$Bias$Variance$Tradeoff$

#  EPE$#  DecomposiWon$of$MSE$$#  BiasYVariance$tradeoff$#  High$bias$?$High$variance$?$How$to$respond$?$$$$

10/27/15$ 52$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$

Page 27: UVA$CS$6316$$ –$Fall$2015$Graduate:$$ Machine$Learning ...€¦ · • k-Nearest neighbor classifier is a lazy learner – Does not build model explicitly. – Classifying unknown

References$$

" $Prof.$Tan,$Steinbach,$Kumar’s$“IntroducWon$to$Data$Mining”$slide$

" $Prof.$Andrew$Moore’s$slides$" $Prof.$Eric$Xing’s$slides$" $HasWe,$Trevor,$et$al.$The%elements%of%sta.s.cal%learning.$Vol.$2.$No.$1.$New$York:$Springer,$2009.$

$10/27/15$ 53$

Dr.$Yanjun$Qi$/$UVA$CS$6316$/$f15$