linear least squares problem - home | course web pages | jack

22
Math for CS Lecture 4 1 Linear Least Squares Problem Consider an equation for a stretched beam: Y = x 1 + x 2 T Where x 1 is the original length, T is the force applied and x 2 is the inverse coefficient of stiffness. Suppose that the following measurements where taken: T 10 15 20 Y 11.60 11.85 12.25 Corresponding to the overcomplete system: 11.60 = x1 + x2 10 11.85 = x1 + x2 15 - can not be satisfied exactly… 12.25 = x1 + x2 20

Upload: others

Post on 12-Sep-2021

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 1

Linear Least Squares ProblemConsider an equation for a stretched beam:

Y = x1 + x2 T

Where x1 is the original length, T is the force applied and x2

is the inverse coefficient of stiffness.

Suppose that the following measurements where taken:T 10 15 20Y 11.60 11.85 12.25

Corresponding to the overcomplete system:11.60 = x1 + x2 1011.85 = x1 + x2 15 - can not be satisfied exactly…12.25 = x1 + x2 20

Page 2: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 2

Linear Least Squares Problem

Problem:

Given A(m x n), m�n, b(m x 1) find x(n x 1) to minimize

||Ax-b||2.

• If m > n, we have more equations than the number of

unknowns, there is generally no x satisfying Ax=b

exactly.

• This is an overcomplete system.

Page 3: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 3

Linear Least Squares

There are three different algorithms for computing the least square minimum.

1. Normal Equations (Cheap, less Accurate).

2. QR decomposition.3. SVD (expensive, more reliable).

The first algorithm in the fastest and the least accurate among the three. On the other hand SVD is the slowest and most accurate.

Page 4: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 4

0LQLPL]H�WKH�VTXDUHG�(XFOLGHDQ�QRUP�RI�WKH�UHVLGXDO�YHFWRU�

7R�PLQLPL]H�ZH�WDNH�WKH�GHULYDWLYH�ZLWK�UHVSHFW�WR�[�DQG�VHW�LW�WR�]HUR�

:KLFK�UHGXFHV�WR�DQ��Q�[�Q��OLQHDU�V\VWHP�FRPPRQO\�NQRZQ�DV�1250$/�(48$7,216�

Normal Equations 1

Page 5: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 5

Normal Equations 2

11.60 = x1 + x2 1011.85 = x1 + x2 1512.25 = x1 + x2 20

A x b

min||Ax-b||2

Page 6: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 6

Normal Equations 3

»»»

¼

º

«««

¬

ª

201

151

101

A »¼

º«¬

ª �

72545

453AAT

»¼

º«¬

ª

201510

111TA»»»

¼

º

«««

¬

ª

25.12

85.11

60.11

B

»¼

º«¬

ª ��� �

650.0

925.10)( 1 bAAAx TT

We must solve the system

For the following values

(ATA)-1AT is called a Pseudo-inverse

Page 7: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 7

QR factorization 1

• A matrix Q is said to be orthogonal if its columns are orthonormal, i.e. QT·Q=I.

• Orthogonal transformations preserve the Euclidean norm since

• Orthogonal matrices can transform vectors in various ways, such as rotation or reflections but they do not change the Euclidean length of the vector. Hence, they preserve the solution to a linear least squares problem.

Page 8: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 8

QR factorization 2

Any matrix A(m·n) can be represented as

A = Q·R

,where Q(m·n) is orthonormal and R(n·n) is upper triangular:

> @ > @»»»»

¼

º

««««

¬

ª

nn

n

nn

r

r

rrr

qqaa

000

00

0|...||...| 11

11211

11!!

!!

!

Page 9: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 9

QR factorization 2• Given A , let its QR decomposition be given as A=Q·R, where

Q is an (m x n) orthonormal matrix and R is upper triangular. • QR factorization transform the linear least square problem into a

triangular least squares.

Q·R·x = bR·x = QT·bx=R-1·QT·b

Matlab Code:

Page 10: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 10

Singular Value Decomposition

• Normal equations and QR decomposition only work for fully-ranked matrices (i.e. rank( A) = n). If A is rank-deficient, that there are infinite number of solutions to the least squares problems and we can use algorithms based on SVD's.

• Given the SVD:U(m x m) , V(n x n) are orthogonal

is an (m x n) diagonal matrix (singular values of A)The minimal solution corresponds to:

Page 11: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 11

Singular Value DecompositionMatlab Code:

Page 12: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 12

/LQHDU�DOJHEUD�UHYLHZ�� 69'

( )

TT

TT

T

U UAA

V V AA

vuV UA

IVV

IUU

VU

AA

2

2

1

21

T

T

0 ),,...,,diag(

such that , ,exist there

,)rank( ,every for :Fact

=

=⇒

⋅==

>==

=

=

∑=

×××

×

p

i

Tiii

ip

pppnpm

nm p

σ

σσσσ

Page 13: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 13

$SSUR[LPDWLRQ�E\�D�ORZ�UDQN�PDWUL[

( )

( )

{ } ( ) 1222)(rank :

2

211

211

~~ min

:in ion toapproximat rank best theIs

),,...,,diag(~

,~

~

Then

... , of SVD thebe let

.rank have let :IIFact

+=

=

=

×

=−=−=−

<==⋅=

≥≥≥⋅==

rr

r

r

i

Tiii

p

p

i

Tiii

nm

r

pr

p

σ

σσσσ

σσσσ

T

XX

T

T

VUAAXA

A

V UvuA

AvuV UA

A

Page 14: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 14

*HRPHWULF�,QWHUSUHWDWLRQ�RI�WKH�

69'� � xAb nm � u

The image of the unit sphere under any mxn matrix is a hyperellipse

v1

v2

·v2

·v1

Page 15: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 15

/HIW�DQG�5LJKW�VLQJXODU�YHFWRUV

� � SAAS nm � u

We can define the properties of A in terms of the shape of AS

v1

v2

·u2

·u1S AS

Singular values of A are the lengths of principal axes of AS, usuallywritten in non-increasing order 1 � 2 � … � n

n left singular vectors of A are the unit vectors {u1,…, un}, oriented in the

directions of the principal semiaxes of AS numbered in correspondance with { i}

n right singular vectors of A are the unit vectors {v1,…, vn}, of S, which are the

preimages of the principal semiaxes of AS: Avi= iui

Page 16: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 16

6LQJXODU�9DOXH�'HFRPSRVLWLRQ

Avi= iui, 1 � i � n

),(

2

1

),(

21

),(

21

),(nnn

nm

n

nn

n

nm

uuuA

»»»»

¼

º

««««

¬

ª

»»»

¼

º

«««

¬

ª

»»»

¼

º

«««

¬

ª

»»»»»»

¼

º

««««««

¬

ª

V

V

V

QQQ!

!!

6 UAV

Matrices U,V are orthogonal and is diagonal

- Singular Value decomposition*VUA 6

Page 17: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 17

0DWULFHV�LQ�WKH�'LDJRQDO�)RUP

Every matrix is diagonal in appropriate basis:

Any vector b(m,1) can be expanded in the basis of left singular vectors of A {ui};

Any vector x(n,1) can be expanded in the basis of right singular vectors of A {vi};

Their coordinates in these new expansions are:

Then the relation b=Ax can be expressed in terms of b’ and x’:

� � xAb nm � u

xVxbUb � � ** ’;’

’**** xbxVUUxAUbUxAb 6 �6 ��

Page 18: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 18

5DQN�RI�$

Let p=min{m,n}, let r�p denote the number of nonzero

singlular values of A,

Then:

The rank of A equals to r, the number of nonzero

singular values

Proof:

The rank of a diagonal matrix equals to the number of its

nonzero entries, and in the decomposition A=U V* ,U and V

are of full rank

Page 19: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 19

'HWHUPLQDQW�RI�$

For A(m,m),

Proof:

The determinant of a product of square matrices is the

product of their determinants. The determinant of a Unitary

matrix is 1 in absolute value, since: U*U=I. Therefore,

m

iiA

1

|)det(| V

6 6 6 m

iiVUVUA

1

** |)det(||)det(||)det(||)det(||)det(||)det(| V

Page 20: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 20

For A(m,n), can be written as a sum of r rank-one matrices:

(1)

Proof:

If we write as a sum of i, where i=diag(0,.., i,..0), then

(1)

Follows from

(2)

$�LQ�WHUPV�RI�VLQJXODU�YHFWRUV

¦

r

jjjj vuA

1

*V

*VUA 6

Page 21: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 21

The L2 norm of the vector is defined as:

(1)

The L2 norm of the matrix is defined as:

Therefore

,where i are the eigenvalues

1RUP�RI�WKH�PDWUL[

¦

n

ii

T xxxx1

2

2

2

22

supx

xAA

)max(2 iA O

iii xxA O

Page 22: Linear Least Squares Problem - Home | Course Web Pages | Jack

Math for CS Lecture 4 22

For any with 0 � �r, define

(1)

If =p=min{m,n}, define v+1=0. Then

0DWUL[�$SSUR[LPDWLRQ�LQ�69'�EDVLV

¦

v

jjjj vuA

1

*V

Q

12

)(

2inf �

d

� � �

u Q

Q

QVBAAA

Brank

CB nm