linear algebra - part iimren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 linear...
TRANSCRIPT
![Page 1: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/1.jpg)
Linear Algebra - Part IIProjection, Eigendecomposition, SVD
(Adapted from Punit Shah’s slides)
2019
Linear Algebra, Part II 2019 1 / 22
![Page 2: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/2.jpg)
Brief Review from Part 1
Matrix Multiplication is a linear tranformation.
Symmetric Matrix:A = AT
Orthogonal Matrix:
ATA = AAT = I and A−1 = AT
L2 Norm:
||x||2 =
√∑i
x2i
Linear Algebra, Part II 2019 2 / 22
![Page 3: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/3.jpg)
Angle Between Vectors
Dot product of two vectors can be written in terms of their L2norms and the angle θ between them.
aTb = ||a||2||b||2 cos(θ)
Linear Algebra, Part II 2019 3 / 22
![Page 4: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/4.jpg)
Cosine Similarity
Cosine between two vectors is a measure of their similarity:
cos(θ) =a · b||a|| ||b||
Orthogonal Vectors: Two vectors a and b are orthogonal toeach other if a · b = 0.
Linear Algebra, Part II 2019 4 / 22
![Page 5: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/5.jpg)
Vector Projection
Given two vectors a and b, let b̂ = b||b|| be the unit vector in the
direction of b.
Then a1 = a1 · b̂ is the orthogonal projection of a onto astraight line parallel to b, where
a1 = ||a|| cos(θ) = a · b̂ = a · b
||b||
Image taken from wikipedia.
Linear Algebra, Part II 2019 5 / 22
![Page 6: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/6.jpg)
Diagonal Matrix
Diagonal matrix has mostly zeros with non-zero entries only inthe diagonal, e.g. identity matrix.
A square diagonal matrix with diagonal elements given by entriesof vector v is denoted:
diag(v)
Multiplying vector x by a diagonal matrix is efficient:
diag(v)x = v � x
� is the entrywise product.
Inverting a square diagonal matrix is efficient:
diag(v)−1 = diag(
[1
v1, . . . ,
1
vn]T)
Linear Algebra, Part II 2019 6 / 22
![Page 7: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/7.jpg)
Determinant
Determinant of a square matrix is a mapping to a scalar.
det(A) or |A|
Measures how much multiplication by the matrix expands orcontracts the space.
Determinant of product is the product of determinants:
det(AB) = det(A)det(B)
∣∣∣∣a bc d
∣∣∣∣ = ad − bc
Linear Algebra, Part II 2019 7 / 22
![Page 8: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/8.jpg)
List of Equivalencies
The following are all equivalent:
Ax = b has a unique solution (for every b with correctdimension).
Ax = 0 has a unique, trivial solution: x = 0.
Columns of A are linearly independent.
A is invertible, i.e. A−1 exists.
det(A) 6= 0
Linear Algebra, Part II 2019 8 / 22
![Page 9: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/9.jpg)
x‘
Linear Algebra, Part II 2019 9 / 22
![Page 10: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/10.jpg)
Zero Determinant
If det(A) = 0, then:
A is linearly dependent.
Ax = b has no solution or infinitely many solutions.
Ax = 0 has a non-zero solution.
Linear Algebra, Part II 2019 9 / 22
![Page 11: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/11.jpg)
Matrix Decomposition
We can decompose an integer into its prime factors, e.g.12 = 2× 2× 3.
Similarly, matrices can be decomposed into factors to learnuniversal properties:
A = Vdiag(λ)V−1
Unlike integers, matrix factorization is not unique:[0 00 1
]×[
1 11 1
]=
[1 −20 1
]×[
2 21 1
]=
[0 01 1
]
Linear Algebra, Part II 2019 10 / 22
![Page 12: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/12.jpg)
Eigenvectors
An eigenvector of a square matrix A is a nonzero vector v suchthat multiplication by A only changes the scale of v.
Av = λv
The scalar λ is known as the eigenvalue.
If v is an eigenvector of A, so is any rescaled vector sv.Moreover, sv still has the same eigenvalue. Thus, we constrainthe eigenvector to be of unit length:
||v|| = 1
Linear Algebra, Part II 2019 11 / 22
![Page 13: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/13.jpg)
Characteristic Polynomial(1)
Eigenvalue equation of matrix A:
Av = λv
λv − Av = 0
(λI− A)v = 0
If nonzero solution for v exists, then it must be the case that:
det(λI− A) = 0
Unpacking the determinant as a function of λ, we get:
PA(λ) = |λI− A| = 1× λn + cn−1 × λn−1 + . . . + c0
This is called the characterisitc polynomial of A.
Linear Algebra, Part II 2019 12 / 22
![Page 14: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/14.jpg)
Characteristic Polynomial(2)
If λ1, λ2, . . . , λn are roots of the characteristic polynomial, theyare eigenvalues of A and we have:
PA(λ) =n∏
i=1
(λ− λi)
cn−1 = −∑n
i=1 λi = −tr(A)
c0 = (−1)n∏n
i=1 λi = (−1)ndet(A)
Roots might be complex. If a root has multiplicity of rj > 1,then the dimension of eigenspace for that eigenvalue might beless than rj (or equal but never more). But one eigenvector isguaranteed.
Linear Algebra, Part II 2019 13 / 22
![Page 15: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/15.jpg)
Example
Consider the matrix:
A =
[2 11 2
]The characteristic polynomial is:
det(λI− A) = det
[λ− 2 −1−1 λ− 2
]= 3− 4λ + λ2 = 0
It has roots λ = 1 and λ = 3 which are the two eigenvalues of A.
We can then solve for eigenvectors using Av = λv:
vλ=1 = [1,−1]T and vλ=3 = [1, 1]T
Linear Algebra, Part II 2019 14 / 22
![Page 16: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/16.jpg)
Eigendecomposition
Suppose that n × n matrix A has n linearly independenteigenvectors {v(1), . . . , v(n)} with eigenvalues {λ1, . . . , λn}.
Concatenate eigenvectors (as columns) to form matrix V.
Concatenate eigenvalues to form vector λ = [λ1, . . . , λn]T .
The eigendecomposition of A is given by:
AV = Vdiag(λ) =⇒ A = Vdiag(λ)V−1
Linear Algebra, Part II 2019 15 / 22
![Page 17: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/17.jpg)
Symmetric Matrices
Every symmetric (hermitian) matrix of dimension n has a set of(not necessarily unique) n orthogonal eigenvectors. Furthermore,All eigenvalues are real.
Every real symmetric matrix A can be decomposed intoreal-valued eigenvectors and eigenvalues:
A = QΛQT
Q is an orthogonal matrix of the eigenvectors of A, and Λ is adiagonal matrix of eigenvalues.
We can think of A as scaling space by λi in direction v(i).
Linear Algebra, Part II 2019 16 / 22
![Page 18: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/18.jpg)
Eigendecomposition is not Unique
Decomposition is not unique when two eigenvalues are the same.
By convention, order entries of Λ in descending order. Then,eigendecomposition is unique if all eigenvalues are unique.
If any eigenvalue is zero, then the matrix is singular. Because ifv is the corresponding eigenvector we have: Av = 0v = 0.
Linear Algebra, Part II 2019 17 / 22
![Page 19: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/19.jpg)
Positive Definite Matrix
If a symmetric matrix A has the property:
xTAx > 0 for any nonzero vector x
Then A is called positive definite.
If the above inequality is not strict then A is called positivesemidefinite.
For positive (semi)definite matrices all eigenvalues arepositive(non negative).
Linear Algebra, Part II 2019 18 / 22
![Page 20: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/20.jpg)
Singular Value Decomposition (SVD)
If A is not square, eigendecomposition is undefined.
SVD is a decomposition of the form:
A = UDVT
SVD is more general than eigendecomposition.
Every real matrix has a SVD.
Linear Algebra, Part II 2019 19 / 22
![Page 21: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/21.jpg)
SVD Definition (1)
Write A as a product of three matrices: A = UDVT .
If A is m × n, then U is m ×m, D is m × n, and V is n × n.
U and V are orthogonal matrices, and D is a diagonal matrix(not necessarily square).
Diagonal entries of D are called singular values of A.
Columns of U are the left singular vectors, and columns of Vare the right singular vectors.
Linear Algebra, Part II 2019 20 / 22
![Page 22: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/22.jpg)
SVD Definition (2)
SVD can be interpreted in terms of eigendecompostion.
Left singular vectors of A are the eigenvectors of AAT .
Right singular vectors of A are the eigenvectors of ATA.
Nonzero singular values of A are square roots of eigenvalues ofATA and AAT .
Numbers on the diagonal of D are sorted largest to smallest andare positive (This makes SVD unique)
Linear Algebra, Part II 2019 21 / 22
![Page 23: Linear Algebra - Part IImren/teach/csc411_19s/tut/... · 2019-08-31 · 2 2 1 1 = 0 0 1 1 Linear Algebra, Part II 2019 10 / 22. Eigenvectors An eigenvector of a square matrix A is](https://reader033.vdocument.in/reader033/viewer/2022042418/5f33d5bf82f60e016f7da9c2/html5/thumbnails/23.jpg)
SVD Optimality
We can write A = UDVT in this form: A =∑n
i=1 diuivTi
Instead of n we can sum up to r: Ar =∑r
i=1 diuivTi
This is called a low rank approximation of A.
Ar is the best approximation of rank r by many norms:
When considering vector norm, it is optimal. Which means Ar
is a linear transformation that captures as much energy aspossible.When considering Frobenius norm, it is optimal which means Ar
is projection of A on the best(closest) r dimensional subspace.
Linear Algebra, Part II 2019 22 / 22