•url: .../publications/courses/ece_8423/lectures/current/lecture_02

13
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum Mean-Square Error Resources: ECE 8443: Chapters 2 and 3 Wiki: Bayes Estimator SP: ML Estimation Tutorial • URL: .../publications/courses/ece_8423/lectures/current/lectur e_02.ppt • MP3: .../publications/courses/ece_8423/lectures/current/lectur LECTURE 02: OPTIMAL ESTIMATION PROCEDURES

Upload: svea

Post on 19-Jan-2016

39 views

Category:

Documents


0 download

DESCRIPTION

LECTURE 02: OPTIMAL ESTIMATION PROCEDURES. Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum Mean-Square Error Resources: ECE 8443: Chapters 2 and 3 Wiki: Bayes Estimator SP: ML Estimation Tutorial. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8443 – Pattern RecognitionECE 8423 – Adaptive Signal Processing

• Objectives:Deterministic vs. RandomMaximum A PosterioriMaximum LikelihoodMinimum Mean-Square Error

• Resources:ECE 8443: Chapters 2 and 3Wiki: Bayes EstimatorSP: ML Estimation Tutorial

• URL: .../publications/courses/ece_8423/lectures/current/lecture_02.ppt• MP3: .../publications/courses/ece_8423/lectures/current/lecture_01.mp3

LECTURE 02: OPTIMAL ESTIMATION PROCEDURES

Page 2: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 2

• Estimation theory plays a major role in the design of statistical signal processing systems. Straightforward procedures such as mean-square error have given way to more powerful techniques that exploit all known statistics about the problem (e.g., Bayes). The goal of this lecture is to provide a high-level overview (Pattern Recognition 101) of parameter estimation from a series of measurements for which the exact generating distributions are unknown.

• Estimation problems are often delineated into deterministic (or classical) parameter estimation and random (or Bayesian) parameter estimation.

• In classical estimation, observations are random and parameters are regarded as unknown constant (deterministic) values.

• In Bayesian parameter estimation, the parameters are also viewed as random, and our prior knowledge of their behavior is expressed through an a priori density.

• Classical estimation operates without prior information; Bayes estimation exploits this information and attempts to estimate it as part of the process.

• Three estimation procedures have been widely used: Maximum A Posteriori (MAP) -- Bayesian Maximum Likelihood (ML) -- Classical Minimum Mean-Squared Error (MMSE) – actually Bayesian

Introduction

Page 3: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 3

Maximum A Posteriori (MAP) Estimation

• Suppose we are given a set of M samples from a discrete random variable, x(n), for n=0,1,…,M-1, and we need to estimate a parameter, , from this data.

• We assume the parameter is characterized by a probability density function, f(), that is called the a priori density.

• The probability density function f(|x) is the probability density of conditioned on the observed data, x, and is referred to as the a posteriori density for .

• The Maximum A Posteriori (MAP) estimate of is the peak value of this density.

• The maximum can generally be found by: . We refer to the value of that maximizes this equation as MAP .

• We often maximize the log of the posterior: .

• Determination of f(|x) is difficult. However, we use Bayes’ rule to write:

where is the prior density of .

• Applying a logarithm, we can write theMAP solution as:

0)|(

xf

0)|(ln

x

f

xx

xf

fff

)()|()|(

)(f

0ln)|(ln

ff x

Page 4: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 4

Maximum Likelihood (ML) Estimation

• The ML estimate is obtained by maximizing the density of the observed data, x, given . Let us define a likelihood function: . Note that we consider x a fixed set of observations. Also:

• The ML estimate is found by maximizing:

• The MAP and ML estimates are similar, though the MAP estimate takes into account prior information.

• ML estimates enjoy a number of important properties: Sufficiency: If a sufficient statistics for exists, then the ML estimate will be a

sufficient statistic. Efficiency: An estimator that attains the lower bound of the variance

attainable by an unbiased estimator is referred to as fully efficient. If such a statistic exists, the ML estimate will be that statistic. Gaussianity: ML estimates are asymptotically Gaussian.

• ML estimates are often a good start because no prior information is required. Once found, it can be used as an initial value for an MAP or Bayes solution.

2121 thanplausiblemoreisthen,;;if xx ff

;xfL

0ln

L

Page 5: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 5

Example – ML Estimate of a Noisy Measurement

• Consider a single measurement, x, consisting of the sum of a parameter, , and a random variable, w, which is zero-mean Gaussian with a variance :

• To construct the ML estimate, we form the likelihood function:

• The log likelihood function is:

• Differentiating with respect to and equating to zero:

• Thus, the best estimate in the ML sense is the value of the raw data. This is intuitively satisfying because in the absence of any other information, the best thing we can do is to trust the data.

• The variance can be shown to be:

Again, this makes sense, why?

2w

wx

22

)(2

1

2

1);(

x

w

wexfL

22

2 )(2

1)2ln(

2

1ln

xL

ww

xx MLMLw

0)(12

2}{ wMLVar

Page 6: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 6

Example – MAP Estimate

• The MAP estimate is obtained from:

• The log of the density is:

• The a priori density, , is given by:

• Taking the log:

• Substituting into the first equation, differentiating, and solving for :

0ln)|(ln

fxf

22

2 )(2

1)2ln(

2

1)|(ln

xxf

ww

f2

2)(

2

1

2

1)(

ef

22

2 )(2

1)2ln(

2

1)(ln

f

MAP

)/(1

)/(22

22

w

wMAP

x

Page 7: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 7

Example – MAP Estimate (Cont.)

• The significance of this result gives us insightinto the MAP estimate. The estimate is a weightedsum of the ML estimate and the prior mean.

• The ratio of the variances can be viewed as a measure of confidence in the a priori model of . The smaller the variance of , the less weight we give to the ML estimate, x.

• In the limit as , the MAP estimate is simply , the a priori mean.

• As the a priori variance increases, the MAP estimate converges to x.

• Exactly how we estimate the parameters of the a priori distribution, or how we deal with this information when it is unknown, will be discussed later in the course, and is discussed extensively in the Pattern Recognition course.

• This derivation can be extended to the important case of 1 parameter and M measurements:

)/(1

)/(22

22

w

wMAP

x

22 / w

1

0

)(1 M

iML ix

M

2

2

2

21

0

1

)(1

M

Mix

M

w

wM

i

MAP

Page 8: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 8

Cramer-Rao Lower Bound (CRLB)

• If is an (Lx1) unknown (deterministic) constant parameter vector, and if is

any unbiased estimator of , then .

• The inequality means that the difference of the two matrices is positive

definite (an alternate way of stating that the lower bound is J-1.

• The matrix, J, is known as the Fisher Information matrix:

• The CRLB provides a lower limit on the variance attainable with an unbiased

estimator for an unknown parameter vector. An estimate which has this

property is known as the Minimum Variance Unbiased Estimator.

• For the scalar parameter case, the CRLB

can be written as:

• For our single sample example, the CRLB

is which coincides with the ML estimate. The ML estimate achieves the

CRLB even for finite M.

);(ln

1)ˆ(

2

2

θxfE

Var

1)ˆ( JθCov

θ̂

2w

1...,,1,0,

);(ln),(

2

Ljiji

fEjiJ

)()(θx

Page 9: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 9

Minimum Mean-Squared Error (MMSE)

• MMSE estimators typically define a quadratic cost function:

• The cost function can be evaluated via the expected value:

• The joint density can be expanded:

• We can rewrite the cost function:

• Both the inner and outer integrals are positive everywhere, and the outer integral is unaffected by the choice of h(x). We can minimize the inner integral:

• This is done by differentiatingwith respect to h(x):

2)(xhEJ

xxx ddfhJ ),()( 2

)()|(),( xxx fff

xxxx ddfhfJ

)|()()( 2

dfh )|()( 2 xx

dfdfh

dfh

)|()|()(

or,

)|()(2

xxx

xx

Page 10: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 10

• But the marginal integral is equal to 1:

• Hence:

• The MMSE estimator is obtained when h(x) is chosen as the expectation of conditioned on the data.

• Note that unlike MAP and ML, this estimator requires the conditional mean value of the posterior density but does not rely on explicit knowledge of the density itself.

• The MMSE solution is often a nonlinear function of the data, though in certain cases, such as when the posterior is Gaussian, the equations are linear. Later we will invoke this to develop a technique called linear prediction used in adaptive filtering.

• Note that both MAP and MMSE produce estimates that are a function of the posterior density for .

• However, for MAP, the estimate is the peak of the posterior, while for MMSE, it is the mean value.

• The MMSE estimator can be extended to estimating a vector of parameters. It can also minimize something other than a quadratic form (L2 norm).

MMSE Estimation (Cont.)

1)|(

df x

xxx |)|()( EdfhMS

Page 11: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 11

• Suppose the estimator is constrained to be a linear function. (Why?)

• For a single parameter, M observation problem:

where the {hi’} are weights chosen to minimize the mean-squared error:

• The minimization is performed by differentiating with respect to each hi:

• We can define the estimation error as:

and write the minimization as:

• This equation states that the error is orthogonal to the data. This is known as the orthogonality principle and is a fundamental property of linear mean-squared error estimation (to be discussed later in greater detail).

Linear MMSE Estimation

1

0

)(M

ii ixh

})({}{21

0

2

M

ii ixhEeEJ

1...,,1,00)}()({1

0

MjjxixhEh

J M

ii

j

1

0

)(M

ii ixhe

0)}({ jxeE

Page 12: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 12

• Given some signal, x(n), how do we estimate another signal, d(n), that might represent for example a noise-reduced version of x(n)?

• The estimate of the desired signal can be written as:

where T{} represents a functional mapping of the data.

• We can invoke our four estimators:

MAP: minimize

ML: maximize

MMSE:

Linear MMSE:

• The last one is the least general, but easiest to implement, since it is essentially a linear filter.

• The others require explicit knowledge of the signal/parameter densities or conditional expectations. In the Pattern Recognition course, we learn about ways to estimate the distributions from the data as part of the overall estimation problem.

Estimation of Signals

)}({)(ˆ nxTnd

))(/)(( nxndf

))();(( ndnxf

)}(/)({ˆ nxndEnd

1

0

)(ˆN

ii ixhnd

Page 13: •URL:  .../publications/courses/ece_8423/lectures/current/lecture_02

ECE 8423: Lecture 02, Slide 13

• Adaptive signal processing requires estimating parameters from data.

• There are two classes of approaches to this problem, classical and Bayesian.

• We explored three approaches to parameter estimation: ML, MAP, MMSE, and MMSE.

• We briefly discussed the Cramer-Rao bound as a way to evaluate the quality of an estimator.

• We discussed two minimum mean-square error (MMSE) approaches: generalized and linear. We compared and contrasted these to MAP and ML.

• We also described a framework in which these estimators could be used to estimate a signal. We will expand on this later when we talk about adaptive filters and noise cancellation.

Summary