chapter v selected topics in numerical linear...

111
Chapter V Selected Topics in Numerical Linear Algebra

Upload: others

Post on 26-Jan-2021

2 views

Category:

Documents


1 download

TRANSCRIPT

  • Chapter V Selected Topics in Numerical LinearAlgebra

  • Chapter V Selected Topics in Numerical LinearAlgebra

    5.0 Review of Basic Concepts5.1 Matrix Eigenvalue Problems:

    Power Method5.2 Schur’s and Gershgorin’s Theorems5.3 Orthogonal Factorizations and

    Least-Squares Problems5.4 Singular-Value Decomposition and

    Pseudo-inverses5.5* QR-Algorithm of Francis for the

    Eigenvalue Problem

  • Chapter V Selected Topics in Numerical LinearAlgebra

    5.0 Review of Basic Concepts5.1 Matrix Eigenvalue Problems:

    Power Method5.2 Schur’s and Gershgorin’s Theorems5.3 Orthogonal Factorizations and

    Least-Squares Problems5.4 Singular-Value Decomposition and

    Pseudo-inverses

    5.5* QR-Algorithm of Francis for theEigenvalue Problem

  • Chapter V Selected Topics in Numerical LinearAlgebra

    5.0 Review of Basic Concepts5.1 Matrix Eigenvalue Problems:

    Power Method5.2 Schur’s and Gershgorin’s Theorems5.3 Orthogonal Factorizations and

    Least-Squares Problems5.4 Singular-Value Decomposition and

    Pseudo-inverses5.5* QR-Algorithm of Francis for the

    Eigenvalue Problem

  • 5.3 Orthogonal Factorizations and Least-SquaresProblems

    Recall the inner-product notation for complexvectors x , y ∈ Cn is defined as

    〈x , y〉 = ȳT x =n∑

    i=1

    xi ȳi

    and satisfies these axioms:1 〈x , x〉 > 0 if x 6= 02 〈αx + βy , z〉 = α〈x , z〉+ β〈y , z〉 ∀α, β ∈ C3 〈x , y〉 = 〈y , x〉

    If 〈x , y〉 = 0, then‖x + y‖22 = ‖x‖22 + ‖y‖22

    This is the so-called Pythagorean rule.

  • 5.3 Orthogonal Factorizations and Least-SquaresProblems

    Recall the inner-product notation for complexvectors x , y ∈ Cn is defined as

    〈x , y〉 = ȳT x =n∑

    i=1

    xi ȳi

    and satisfies these axioms:1 〈x , x〉 > 0 if x 6= 02 〈αx + βy , z〉 = α〈x , z〉+ β〈y , z〉 ∀α, β ∈ C3 〈x , y〉 = 〈y , x〉

    If 〈x , y〉 = 0, then‖x + y‖22 = ‖x‖22 + ‖y‖22

    This is the so-called Pythagorean rule.

  • 5.3 Orthogonal Factorizations and Least-SquaresProblems

    Recall the inner-product notation for complexvectors x , y ∈ Cn is defined as

    〈x , y〉 = ȳT x =n∑

    i=1

    xi ȳi

    and satisfies these axioms:1 〈x , x〉 > 0 if x 6= 02 〈αx + βy , z〉 = α〈x , z〉+ β〈y , z〉 ∀α, β ∈ C3 〈x , y〉 = 〈y , x〉

    If 〈x , y〉 = 0, then‖x + y‖22 = ‖x‖22 + ‖y‖22

    This is the so-called Pythagorean rule.

  • 5.3 Orthogonal Factorizations and Least-SquaresProblems

    Recall the inner-product notation for complexvectors x , y ∈ Cn is defined as

    〈x , y〉 = ȳT x =n∑

    i=1

    xi ȳi

    and satisfies these axioms:1 〈x , x〉 > 0 if x 6= 02 〈αx + βy , z〉 = α〈x , z〉+ β〈y , z〉 ∀α, β ∈ C3 〈x , y〉 = 〈y , x〉

    If 〈x , y〉 = 0, then‖x + y‖22 = ‖x‖22 + ‖y‖22

    This is the so-called Pythagorean rule.

  • 5.3 Orthogonal Factorizations and Least-SquaresProblems

    Recall the inner-product notation for complexvectors x , y ∈ Cn is defined as

    〈x , y〉 = ȳT x =n∑

    i=1

    xi ȳi

    and satisfies these axioms:1 〈x , x〉 > 0 if x 6= 02 〈αx + βy , z〉 = α〈x , z〉+ β〈y , z〉 ∀α, β ∈ C3 〈x , y〉 = 〈y , x〉

    If 〈x , y〉 = 0, then‖x + y‖22 = ‖x‖22 + ‖y‖22

    This is the so-called Pythagorean rule.

  • Suppose that {v1, v2, · · · , vn} is an orthonormalbasis for Cn.

    Then, each element x ∈ Cn has aunique representation in the form

    x =n∑

    i=1

    civi ∈ Cn

    where ci = 〈x , vi〉 (1 ≤ i ≤ n). Thus,

    x =n∑

    i=1

    〈x , vi〉vi ∈ Cn

  • Suppose that {v1, v2, · · · , vn} is an orthonormalbasis for Cn. Then, each element x ∈ Cn has aunique representation in the form

    x =n∑

    i=1

    civi ∈ Cn

    where ci = 〈x , vi〉 (1 ≤ i ≤ n). Thus,

    x =n∑

    i=1

    〈x , vi〉vi ∈ Cn

  • Suppose that {v1, v2, · · · , vn} is an orthonormalbasis for Cn. Then, each element x ∈ Cn has aunique representation in the form

    x =n∑

    i=1

    civi ∈ Cn

    where ci = 〈x , vi〉 (1 ≤ i ≤ n).

    Thus,

    x =n∑

    i=1

    〈x , vi〉vi ∈ Cn

  • Suppose that {v1, v2, · · · , vn} is an orthonormalbasis for Cn. Then, each element x ∈ Cn has aunique representation in the form

    x =n∑

    i=1

    civi ∈ Cn

    where ci = 〈x , vi〉 (1 ≤ i ≤ n). Thus,

    x =n∑

    i=1

    〈x , vi〉vi ∈ Cn

  • Gram-Schmidt Process

    It can be used to obtain orthonormal systems inan inner-product space. Suppose that [x1, x2, · · · ]is a linearly independent sequence of vectorsin an inner-product space (The sequence can befinite or infinite). We can generate an orthonormalsequence {u1, u2, · · · , } by the formula

    uk =xk −

    ∑k−1i=1 〈xk , ui〉ui

    ‖xk −∑k−1

    i=1 〈xk , ui〉ui‖2(k ≥ 1)

  • Gram-Schmidt ProcessIt can be used to obtain orthonormal systems inan inner-product space. Suppose that [x1, x2, · · · ]is a linearly independent sequence of vectorsin an inner-product space (The sequence can befinite or infinite).

    We can generate an orthonormalsequence {u1, u2, · · · , } by the formula

    uk =xk −

    ∑k−1i=1 〈xk , ui〉ui

    ‖xk −∑k−1

    i=1 〈xk , ui〉ui‖2(k ≥ 1)

  • Gram-Schmidt ProcessIt can be used to obtain orthonormal systems inan inner-product space. Suppose that [x1, x2, · · · ]is a linearly independent sequence of vectorsin an inner-product space (The sequence can befinite or infinite). We can generate an orthonormalsequence {u1, u2, · · · , } by the formula

    uk =xk −

    ∑k−1i=1 〈xk , ui〉ui

    ‖xk −∑k−1

    i=1 〈xk , ui〉ui‖2(k ≥ 1)

  • Gram-Schmidt ProcessIt can be used to obtain orthonormal systems inan inner-product space. Suppose that [x1, x2, · · · ]is a linearly independent sequence of vectorsin an inner-product space (The sequence can befinite or infinite). We can generate an orthonormalsequence {u1, u2, · · · , } by the formula

    uk =xk −

    ∑k−1i=1 〈xk , ui〉ui

    ‖xk −∑k−1

    i=1 〈xk , ui〉ui‖2(k ≥ 1)

  • Theorem (on Gram-Schmidt Sequence)The Gram-Schmidt Sequence [u1,u2, · · · , ]has the property that {u1,u2, · · · ,uk} is anorthonormal base for the linear span of for{x1, x2, · · · , xk} for k ≥ 1.

    Proof.Hint: by induction on k . See also P275.

    RemarkThe theorem tells us that

    span{

    x1, x2, · · · , xk}

    = span{

    u1,u2, · · · ,uk}

    for k ≥ 1.

  • Theorem (on Gram-Schmidt Sequence)The Gram-Schmidt Sequence [u1,u2, · · · , ]has the property that {u1,u2, · · · ,uk} is anorthonormal base for the linear span of for{x1, x2, · · · , xk} for k ≥ 1.

    Proof.Hint: by induction on k . See also P275.

    RemarkThe theorem tells us that

    span{

    x1, x2, · · · , xk}

    = span{

    u1,u2, · · · ,uk}

    for k ≥ 1.

  • Theorem (on Gram-Schmidt Sequence)The Gram-Schmidt Sequence [u1,u2, · · · , ]has the property that {u1,u2, · · · ,uk} is anorthonormal base for the linear span of for{x1, x2, · · · , xk} for k ≥ 1.

    Proof.Hint: by induction on k . See also P275.

    RemarkThe theorem tells us that

    span{

    x1, x2, · · · , xk}

    = span{

    u1,u2, · · · ,uk}

    for k ≥ 1.

  • For example, apply the Gram-Schmidtprocess to the columns A1,A2, · · · ,An ofAm×n (m ≥ n), we arrive after n steps at anm × n matrix B whose columns form anorthonormal set (if {A1,A2, · · · ,An} is alinearly independent set).

    RemarkNote: the Gram-Schmidt process whenapplied to the columns of a matrix can beinterpreted as a matrix factorization.

  • For example, apply the Gram-Schmidtprocess to the columns A1,A2, · · · ,An ofAm×n (m ≥ n), we arrive after n steps at anm × n matrix B whose columns form anorthonormal set (if {A1,A2, · · · ,An} is alinearly independent set).

    RemarkNote: the Gram-Schmidt process whenapplied to the columns of a matrix can beinterpreted as a matrix factorization.

  • Gram-Schmidt Algorithm

    for j = 1 to n dofor i = 1 to j − 1 do

    tij ← 〈Aj ,Bi〉end do

    Cj ← Aj −∑i

  • Gram-Schmidt Algorithm

    for j = 1 to n dofor i = 1 to j − 1 do

    tij ← 〈Aj ,Bi〉end do

    Cj ← Aj −∑i

  • Theorem (on Gram-Schmidt Factorization)The Gram-Schmidt process, when appliedto the columns of an m × n (m ≥ n) matrix Aof rank n, produces a factorization

    Am×n = Bm×nTn×nin which B is an m × n matrix withorthonormal columns and T is an n × nupper triangular matrix with positivediagonal.

  • Proof.

    We complete the Definition of the T matrix bysetting tij = 0 when i > j . By the precedingtheorem, the columns of B form an orthonormalset of n vectors in Rm, and each Aj is a linearcombination of B1,B2, · · · ,Bj . In fact,

    Aj =j∑

    i=1

    〈Aj ,Bi〉Bi

    =

    j−1∑i=1

    〈Aj ,Bi〉Bi + 〈Aj ,Bj〉Bj

    =

    j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj

  • Proof.We complete the Definition of the T matrix bysetting tij = 0 when i > j .

    By the precedingtheorem, the columns of B form an orthonormalset of n vectors in Rm, and each Aj is a linearcombination of B1,B2, · · · ,Bj . In fact,

    Aj =j∑

    i=1

    〈Aj ,Bi〉Bi

    =

    j−1∑i=1

    〈Aj ,Bi〉Bi + 〈Aj ,Bj〉Bj

    =

    j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj

  • Proof.We complete the Definition of the T matrix bysetting tij = 0 when i > j . By the precedingtheorem, the columns of B form an orthonormalset of n vectors in Rm, and each Aj is a linearcombination of B1,B2, · · · ,Bj .

    In fact,

    Aj =j∑

    i=1

    〈Aj ,Bi〉Bi

    =

    j−1∑i=1

    〈Aj ,Bi〉Bi + 〈Aj ,Bj〉Bj

    =

    j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj

  • Proof.We complete the Definition of the T matrix bysetting tij = 0 when i > j . By the precedingtheorem, the columns of B form an orthonormalset of n vectors in Rm, and each Aj is a linearcombination of B1,B2, · · · ,Bj . In fact,

    Aj =j∑

    i=1

    〈Aj ,Bi〉Bi

    =

    j−1∑i=1

    〈Aj ,Bi〉Bi + 〈Aj ,Bj〉Bj

    =

    j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj

  • Proof.We complete the Definition of the T matrix bysetting tij = 0 when i > j . By the precedingtheorem, the columns of B form an orthonormalset of n vectors in Rm, and each Aj is a linearcombination of B1,B2, · · · ,Bj . In fact,

    Aj =j∑

    i=1

    〈Aj ,Bi〉Bi

    =

    j−1∑i=1

    〈Aj ,Bi〉Bi + 〈Aj ,Bj〉Bj

    =

    j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj

  • Proof.We complete the Definition of the T matrix bysetting tij = 0 when i > j . By the precedingtheorem, the columns of B form an orthonormalset of n vectors in Rm, and each Aj is a linearcombination of B1,B2, · · · ,Bj . In fact,

    Aj =j∑

    i=1

    〈Aj ,Bi〉Bi

    =

    j−1∑i=1

    〈Aj ,Bi〉Bi + 〈Aj ,Bj〉Bj

    =

    j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi =n∑

    i=1

    tijBi

    i.e., A = BT .

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi =n∑

    i=1

    tijBi

    i.e., A = BT .

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi =n∑

    i=1

    tijBi

    i.e., A = BT .

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi =n∑

    i=1

    tijBi

    i.e., A = BT .

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi

    =n∑

    i=1

    tijBi

    i.e., A = BT .

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi =n∑

    i=1

    tijBi

    i.e., A = BT .

  • Proof.Next,

    〈Aj ,Bj〉 = 〈Cj +∑i 0

    So

    Aj =j−1∑i=1

    tijBi + 〈Aj ,Bj〉Bj =j∑

    i=1

    tijBi =n∑

    i=1

    tijBi

    i.e., A = BT .

  • To avoid the square roots involving in computing‖ · ‖2, the Gram-Schmidt Algorithm can bemodified as follows:

    Modified Gram-Schmidt Algorithmfor k = 1 to n do

    dk ← ‖Ak‖22tkk ← 1

    for j = k + 1 to n do

    tkj ← 〈Aj ,Ak〉/dkAj ← Aj − tkjAk

    end do

    end do

  • To avoid the square roots involving in computing‖ · ‖2, the Gram-Schmidt Algorithm can bemodified as follows:

    Modified Gram-Schmidt Algorithm

    for k = 1 to n do

    dk ← ‖Ak‖22tkk ← 1

    for j = k + 1 to n do

    tkj ← 〈Aj ,Ak〉/dkAj ← Aj − tkjAk

    end do

    end do

  • To avoid the square roots involving in computing‖ · ‖2, the Gram-Schmidt Algorithm can bemodified as follows:

    Modified Gram-Schmidt Algorithmfor k = 1 to n do

    dk ← ‖Ak‖22tkk ← 1

    for j = k + 1 to n do

    tkj ← 〈Aj ,Ak〉/dkAj ← Aj − tkjAk

    end do

    end do

  • Theorem (on Modified Gram-SchmidtFactorization)If the modified Gram-Schmidt process isapplied to the columns of an m × n (m ≥ n)matrix A of rank n, the transformed m × nmatrix B has an orthogonal set of columnsand satisfies

    Am×n = Bm×nTn×nwhere T is a unit n × n upper triangularmatrix whose elements tkj (j > k) aregenerated in the algorithm.

    Proof.Hint: by induction on k . See also P277-278.

  • Theorem (on Modified Gram-SchmidtFactorization)If the modified Gram-Schmidt process isapplied to the columns of an m × n (m ≥ n)matrix A of rank n, the transformed m × nmatrix B has an orthogonal set of columnsand satisfies

    Am×n = Bm×nTn×nwhere T is a unit n × n upper triangularmatrix whose elements tkj (j > k) aregenerated in the algorithm.

    Proof.Hint: by induction on k . See also P277-278.

  • Least-Squares Problems

    An important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations.

    ConsiderAx = b (1)

    where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1.

    Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n.

    Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.

    In such cases, it is often required to find an xthat minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax .

    The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum.

    (Ifrank(A) = n, then this x will be unique.)

  • Least-Squares ProblemsAn important application of the orthogonalfactorizations being discussed is theLeast-Squares Problem for a linear system ofequations. Consider

    Ax = b (1)where A is m × n, x is n × 1, b is m × 1. Assumethe rank of A is n; hence, m ≥ n. Usually, system(1) will have no solution if

    b /∈ span{A1, · · · ,An} ⊂ Cm.In such cases, it is often required to find an x

    that minimizes the norm of the residual vector,b − Ax . The least-squares “solution" of (1) isthe vector x that makes ‖b − Ax‖2 a minimum. (Ifrank(A) = n, then this x will be unique.)

  • Some Motivations of Least-Square Problems

    Assume y = αx + β. Then

    A4×2 z ≡

    x1 1x2 1x3 1x4 1

    ( αβ ) = y1y2

    y3y4

    ≡ b

  • Some Motivations of Least-Square Problems

    Assume y = αx + β. Then

    A4×2 z ≡

    x1 1x2 1x3 1x4 1

    ( αβ ) = y1y2

    y3y4

    ≡ b

  • Some Motivations of Least-Square Problems

    Assume y = αx + β. Then

    A4×2 z ≡

    x1 1x2 1x3 1x4 1

    ( αβ ) = y1y2

    y3y4

    ≡ b

  • Some Motivations of Least-Square Problems

    Assume y = αx + β.

    Then

    A4×2 z ≡

    x1 1x2 1x3 1x4 1

    ( αβ ) = y1y2

    y3y4

    ≡ b

  • Some Motivations of Least-Square Problems

    Assume y = αx + β. Then

    A4×2 z ≡

    x1 1x2 1x3 1x4 1

    ( αβ ) = y1y2

    y3y4

    ≡ b

  • Lemma (on the Least-Square Problems)If x is a point s.t. A∗(Ax − b) = 0, then x solvesthe least-square problems.

    Proof.Note that A∗(Ax − b) = 0 implies that b − Ax isorthogonal to the column space of A. For anyvector y , since A(x − y) is in the column space ofA, we have 〈b − Ax ,A(x − y)〉 = 0. Then, thePythagorean rule gives

    ‖b − Ay‖22 = ‖b − Ax + A(x − y)‖22= ‖b − Ax‖22 + ‖A(x − y)‖22≥ ‖b − Ax‖22, ∀y

  • Lemma (on the Least-Square Problems)If x is a point s.t. A∗(Ax − b) = 0, then x solvesthe least-square problems.

    Proof.Note that A∗(Ax − b) = 0 implies that b − Ax isorthogonal to the column space of A.

    For anyvector y , since A(x − y) is in the column space ofA, we have 〈b − Ax ,A(x − y)〉 = 0. Then, thePythagorean rule gives

    ‖b − Ay‖22 = ‖b − Ax + A(x − y)‖22= ‖b − Ax‖22 + ‖A(x − y)‖22≥ ‖b − Ax‖22, ∀y

  • Lemma (on the Least-Square Problems)If x is a point s.t. A∗(Ax − b) = 0, then x solvesthe least-square problems.

    Proof.Note that A∗(Ax − b) = 0 implies that b − Ax isorthogonal to the column space of A. For anyvector y , since A(x − y) is in the column space ofA, we have 〈b − Ax ,A(x − y)〉 = 0.

    Then, thePythagorean rule gives

    ‖b − Ay‖22 = ‖b − Ax + A(x − y)‖22= ‖b − Ax‖22 + ‖A(x − y)‖22≥ ‖b − Ax‖22, ∀y

  • Lemma (on the Least-Square Problems)If x is a point s.t. A∗(Ax − b) = 0, then x solvesthe least-square problems.

    Proof.Note that A∗(Ax − b) = 0 implies that b − Ax isorthogonal to the column space of A. For anyvector y , since A(x − y) is in the column space ofA, we have 〈b − Ax ,A(x − y)〉 = 0. Then, thePythagorean rule gives

    ‖b − Ay‖22 = ‖b − Ax + A(x − y)‖22= ‖b − Ax‖22 + ‖A(x − y)‖22≥ ‖b − Ax‖22, ∀y

  • If Am×n has been factored in the form A = BT asdescribed in the preceding theorem, then theleast-squares solution of Ax = b will be the exactsolution of the n × n system

    Tx = (B∗B)−1B∗b

    This can be verified by the above Lemma:

    A∗Ax = (BT )∗BTx = T ∗B∗B(B∗B)−1B∗b = T ∗B∗b = A∗b

    The matrix

    (B∗B)−1 = diag{d−11 , · · · , d−1n }

    the numbers di being those computed in themodified Gram-Schimdt algorithm.

  • If Am×n has been factored in the form A = BT asdescribed in the preceding theorem, then theleast-squares solution of Ax = b will be the exactsolution of the n × n system

    Tx = (B∗B)−1B∗b

    This can be verified by the above Lemma:

    A∗Ax = (BT )∗BTx

    = T ∗B∗B(B∗B)−1B∗b = T ∗B∗b = A∗b

    The matrix

    (B∗B)−1 = diag{d−11 , · · · , d−1n }

    the numbers di being those computed in themodified Gram-Schimdt algorithm.

  • If Am×n has been factored in the form A = BT asdescribed in the preceding theorem, then theleast-squares solution of Ax = b will be the exactsolution of the n × n system

    Tx = (B∗B)−1B∗b

    This can be verified by the above Lemma:

    A∗Ax = (BT )∗BTx = T ∗B∗B(B∗B)−1B∗b

    = T ∗B∗b = A∗b

    The matrix

    (B∗B)−1 = diag{d−11 , · · · , d−1n }

    the numbers di being those computed in themodified Gram-Schimdt algorithm.

  • If Am×n has been factored in the form A = BT asdescribed in the preceding theorem, then theleast-squares solution of Ax = b will be the exactsolution of the n × n system

    Tx = (B∗B)−1B∗b

    This can be verified by the above Lemma:

    A∗Ax = (BT )∗BTx = T ∗B∗B(B∗B)−1B∗b = T ∗B∗b = A∗b

    The matrix

    (B∗B)−1 = diag{d−11 , · · · , d−1n }

    the numbers di being those computed in themodified Gram-Schimdt algorithm.

  • If Am×n has been factored in the form A = BT asdescribed in the preceding theorem, then theleast-squares solution of Ax = b will be the exactsolution of the n × n system

    Tx = (B∗B)−1B∗b

    This can be verified by the above Lemma:

    A∗Ax = (BT )∗BTx = T ∗B∗B(B∗B)−1B∗b = T ∗B∗b = A∗b

    The matrix

    (B∗B)−1 = diag{d−11 , · · · , d−1n }

    the numbers di being those computed in themodified Gram-Schimdt algorithm.

  • OTOH, by the preceding lemma, ‖b − Ax‖2 willbe a minimum if

    A∗(Ax − b) = 0

    If rank(A) = n, then the n × n square matrix A∗Ais nonsingular, and in this case there is exactlyone least-squares solutions; it is determineduniquely by solving the n × n system of so-callednormal equations (���555§§§|||):

    A∗Ax = A∗b

    It is easy to see that A∗A is Hermitian andpositive definite. (So, Cholesky factorization maybe used.)

  • OTOH, by the preceding lemma, ‖b − Ax‖2 willbe a minimum if

    A∗(Ax − b) = 0

    If rank(A) = n, then the n × n square matrix A∗Ais nonsingular, and in this case there is exactlyone least-squares solutions;

    it is determineduniquely by solving the n × n system of so-callednormal equations (���555§§§|||):

    A∗Ax = A∗b

    It is easy to see that A∗A is Hermitian andpositive definite. (So, Cholesky factorization maybe used.)

  • OTOH, by the preceding lemma, ‖b − Ax‖2 willbe a minimum if

    A∗(Ax − b) = 0

    If rank(A) = n, then the n × n square matrix A∗Ais nonsingular, and in this case there is exactlyone least-squares solutions; it is determineduniquely by solving the n × n system of so-callednormal equations (���555§§§|||):

    A∗Ax = A∗b

    It is easy to see that A∗A is Hermitian andpositive definite. (So, Cholesky factorization maybe used.)

  • OTOH, by the preceding lemma, ‖b − Ax‖2 willbe a minimum if

    A∗(Ax − b) = 0

    If rank(A) = n, then the n × n square matrix A∗Ais nonsingular, and in this case there is exactlyone least-squares solutions; it is determineduniquely by solving the n × n system of so-callednormal equations (���555§§§|||):

    A∗Ax = A∗b

    It is easy to see that A∗A is Hermitian andpositive definite.

    (So, Cholesky factorization maybe used.)

  • OTOH, by the preceding lemma, ‖b − Ax‖2 willbe a minimum if

    A∗(Ax − b) = 0

    If rank(A) = n, then the n × n square matrix A∗Ais nonsingular, and in this case there is exactlyone least-squares solutions; it is determineduniquely by solving the n × n system of so-callednormal equations (���555§§§|||):

    A∗Ax = A∗b

    It is easy to see that A∗A is Hermitian andpositive definite. (So, Cholesky factorization maybe used.)

  • The direct use of the normal equations for solvinga least-squares problem seems very appealingbecause of its conceptual simplicity.

    However, it isregarded as one of the least satisfactory methodsto use on this problem. One reason is that thecondition number of A∗A may be considerablyworse than that of A.For example,

    A =

    1 1 1� 0 00 � 00 0 �

    A∗A = 1 + �2 1 11 1 + �2 1

    1 1 1 + �2

    For small �, in a computer one may have

    rank(A) = 3 but rank(A∗A) = 1

  • The direct use of the normal equations for solvinga least-squares problem seems very appealingbecause of its conceptual simplicity. However, it isregarded as one of the least satisfactory methodsto use on this problem.

    One reason is that thecondition number of A∗A may be considerablyworse than that of A.For example,

    A =

    1 1 1� 0 00 � 00 0 �

    A∗A = 1 + �2 1 11 1 + �2 1

    1 1 1 + �2

    For small �, in a computer one may have

    rank(A) = 3 but rank(A∗A) = 1

  • The direct use of the normal equations for solvinga least-squares problem seems very appealingbecause of its conceptual simplicity. However, it isregarded as one of the least satisfactory methodsto use on this problem. One reason is that thecondition number of A∗A may be considerablyworse than that of A.

    For example,

    A =

    1 1 1� 0 00 � 00 0 �

    A∗A = 1 + �2 1 11 1 + �2 1

    1 1 1 + �2

    For small �, in a computer one may have

    rank(A) = 3 but rank(A∗A) = 1

  • The direct use of the normal equations for solvinga least-squares problem seems very appealingbecause of its conceptual simplicity. However, it isregarded as one of the least satisfactory methodsto use on this problem. One reason is that thecondition number of A∗A may be considerablyworse than that of A.For example,

    A =

    1 1 1� 0 00 � 00 0 �

    A∗A = 1 + �2 1 11 1 + �2 1

    1 1 1 + �2

    For small �, in a computer one may have

    rank(A) = 3 but rank(A∗A) = 1

  • The direct use of the normal equations for solvinga least-squares problem seems very appealingbecause of its conceptual simplicity. However, it isregarded as one of the least satisfactory methodsto use on this problem. One reason is that thecondition number of A∗A may be considerablyworse than that of A.For example,

    A =

    1 1 1� 0 00 � 00 0 �

    A∗A = 1 + �2 1 11 1 + �2 1

    1 1 1 + �2

    For small �, in a computer one may have

    rank(A) = 3 but rank(A∗A) = 1

  • Householder’s QR-Factorization

    One of the most useful orthogonal factorizationsis called QR-Factorization. The objective is tofactor an m × n matrix A into a product

    Am×n = Qm×mRm×n

    where Q is a unitary matrix and R is an m × nupper triangular matrix. The factorizationalgorithm actually produces

    Q∗Am×n = Rm×n

  • Householder’s QR-FactorizationOne of the most useful orthogonal factorizationsis called QR-Factorization. The objective is tofactor an m × n matrix A into a product

    Am×n = Qm×mRm×n

    where Q is a unitary matrix and R is an m × nupper triangular matrix. The factorizationalgorithm actually produces

    Q∗Am×n = Rm×n

  • Householder’s QR-FactorizationOne of the most useful orthogonal factorizationsis called QR-Factorization. The objective is tofactor an m × n matrix A into a product

    Am×n = Qm×mRm×n

    where Q is a unitary matrix and R is an m × nupper triangular matrix.

    The factorizationalgorithm actually produces

    Q∗Am×n = Rm×n

  • Householder’s QR-FactorizationOne of the most useful orthogonal factorizationsis called QR-Factorization. The objective is tofactor an m × n matrix A into a product

    Am×n = Qm×mRm×n

    where Q is a unitary matrix and R is an m × nupper triangular matrix. The factorizationalgorithm actually produces

    Q∗Am×n = Rm×n

  • Q∗ is build up step-by-step as

    Q∗ = U(n−1)U(n−2) · · ·U(1)

    where

    U(k)m×m =

    (I(k−1)×(k−1) 0

    0 I(m−k+1)×(m−k+1) − vv∗

    )with v ∈ Cm−k+1, ‖v‖2 =

    √2.

    SoU(1) = Im×m − vv∗ and we want U(1)A1 = β1e(1)with e(1) = (1, 0, · · · , 0)T . Next, we want

    U(2)U(1)A1 = β1e(1),U(2)U(1)A2 = (∗, β2, 0, · · · , 0)T

    Finally, U(n−1)U(n−2) · · ·U(1)A = Rm×n.

  • Q∗ is build up step-by-step as

    Q∗ = U(n−1)U(n−2) · · ·U(1)

    where

    U(k)m×m =

    (I(k−1)×(k−1) 0

    0 I(m−k+1)×(m−k+1) − vv∗

    )with v ∈ Cm−k+1, ‖v‖2 =

    √2. So

    U(1) = Im×m − vv∗

    and we want U(1)A1 = β1e(1)

    with e(1) = (1, 0, · · · , 0)T . Next, we want

    U(2)U(1)A1 = β1e(1),U(2)U(1)A2 = (∗, β2, 0, · · · , 0)T

    Finally, U(n−1)U(n−2) · · ·U(1)A = Rm×n.

  • Q∗ is build up step-by-step as

    Q∗ = U(n−1)U(n−2) · · ·U(1)

    where

    U(k)m×m =

    (I(k−1)×(k−1) 0

    0 I(m−k+1)×(m−k+1) − vv∗

    )with v ∈ Cm−k+1, ‖v‖2 =

    √2. So

    U(1) = Im×m − vv∗ and we want U(1)A1 = β1e(1)with e(1) = (1, 0, · · · , 0)T .

    Next, we want

    U(2)U(1)A1 = β1e(1),U(2)U(1)A2 = (∗, β2, 0, · · · , 0)T

    Finally, U(n−1)U(n−2) · · ·U(1)A = Rm×n.

  • Q∗ is build up step-by-step as

    Q∗ = U(n−1)U(n−2) · · ·U(1)

    where

    U(k)m×m =

    (I(k−1)×(k−1) 0

    0 I(m−k+1)×(m−k+1) − vv∗

    )with v ∈ Cm−k+1, ‖v‖2 =

    √2. So

    U(1) = Im×m − vv∗ and we want U(1)A1 = β1e(1)with e(1) = (1, 0, · · · , 0)T . Next, we want

    U(2)U(1)A1 = β1e(1),U(2)U(1)A2 = (∗, β2, 0, · · · , 0)T

    Finally, U(n−1)U(n−2) · · ·U(1)A = Rm×n.

  • Q∗ is build up step-by-step as

    Q∗ = U(n−1)U(n−2) · · ·U(1)

    where

    U(k)m×m =

    (I(k−1)×(k−1) 0

    0 I(m−k+1)×(m−k+1) − vv∗

    )with v ∈ Cm−k+1, ‖v‖2 =

    √2. So

    U(1) = Im×m − vv∗ and we want U(1)A1 = β1e(1)with e(1) = (1, 0, · · · , 0)T . Next, we want

    U(2)U(1)A1 = β1e(1),U(2)U(1)A2 = (∗, β2, 0, · · · , 0)T

    Finally, U(n−1)U(n−2) · · ·U(1)A = Rm×n.

  • Recall (in Sect. 5.2)

    Lemma (1st Lemma on Unitary Matrix)For any vector v ∈ Cn, the matrix I − vv∗ isunitary if and only if ‖v‖2 =

    √2 or v = 0.

    Lemma (2nd Lemma on Unitary Matrix)Let x , y ∈ Cn s.t. ‖x‖2 = ‖y‖2 and〈x , y〉 = y∗x is real. Then there exists aunitary matrix U of the form I − vv∗ s.t.Ux = y.

    Here, v =√

    2(x−y)‖x−y‖2 .

  • Recall (in Sect. 5.2)

    Lemma (1st Lemma on Unitary Matrix)For any vector v ∈ Cn, the matrix I − vv∗ isunitary if and only if ‖v‖2 =

    √2 or v = 0.

    Lemma (2nd Lemma on Unitary Matrix)Let x , y ∈ Cn s.t. ‖x‖2 = ‖y‖2 and〈x , y〉 = y∗x is real. Then there exists aunitary matrix U of the form I − vv∗ s.t.Ux = y.

    Here, v =√

    2(x−y)‖x−y‖2 .

  • Recall (in Sect. 5.2)

    Lemma (1st Lemma on Unitary Matrix)For any vector v ∈ Cn, the matrix I − vv∗ isunitary if and only if ‖v‖2 =

    √2 or v = 0.

    Lemma (2nd Lemma on Unitary Matrix)Let x , y ∈ Cn s.t. ‖x‖2 = ‖y‖2 and〈x , y〉 = y∗x is real. Then there exists aunitary matrix U of the form I − vv∗ s.t.Ux = y.

    Here, v =√

    2(x−y)‖x−y‖2 .

  • For example in the 1st step of theQR-factorization, we can put

    v = α(A1 − β1e(1))with α =

    √2/‖A1 − β1e(1)‖2.

    If choose

    β1 ={ −‖A1‖2a11/|a11| if a11 6= 0−‖A1‖2 if a11 = 0

    then‖A1‖2 = ‖β1e(1)‖2, 〈A1, β1e(1)〉 = −|a11|‖A1‖2 ∈ RFrom the proof of Lemma 2,

    U(1)A1 ≡ (I − vv∗)A1 = β1e(1)

    Repeat the process for the (m− 1)× (n− 1)sub-matrix of U(1)A ≡ (I − vv∗)A and so on.

  • For example in the 1st step of theQR-factorization, we can put

    v = α(A1 − β1e(1))with α =

    √2/‖A1 − β1e(1)‖2. If choose

    β1 ={ −‖A1‖2a11/|a11| if a11 6= 0−‖A1‖2 if a11 = 0

    then‖A1‖2 = ‖β1e(1)‖2, 〈A1, β1e(1)〉 = −|a11|‖A1‖2 ∈ RFrom the proof of Lemma 2,

    U(1)A1 ≡ (I − vv∗)A1 = β1e(1)

    Repeat the process for the (m− 1)× (n− 1)sub-matrix of U(1)A ≡ (I − vv∗)A and so on.

  • For example in the 1st step of theQR-factorization, we can put

    v = α(A1 − β1e(1))with α =

    √2/‖A1 − β1e(1)‖2. If choose

    β1 ={ −‖A1‖2a11/|a11| if a11 6= 0−‖A1‖2 if a11 = 0

    then‖A1‖2 = ‖β1e(1)‖2, 〈A1, β1e(1)〉 = −|a11|‖A1‖2 ∈ R

    From the proof of Lemma 2,U(1)A1 ≡ (I − vv∗)A1 = β1e(1)

    Repeat the process for the (m− 1)× (n− 1)sub-matrix of U(1)A ≡ (I − vv∗)A and so on.

  • For example in the 1st step of theQR-factorization, we can put

    v = α(A1 − β1e(1))with α =

    √2/‖A1 − β1e(1)‖2. If choose

    β1 ={ −‖A1‖2a11/|a11| if a11 6= 0−‖A1‖2 if a11 = 0

    then‖A1‖2 = ‖β1e(1)‖2, 〈A1, β1e(1)〉 = −|a11|‖A1‖2 ∈ RFrom the proof of Lemma 2,

    U(1)A1 ≡ (I − vv∗)A1 = β1e(1)

    Repeat the process for the (m− 1)× (n− 1)sub-matrix of U(1)A ≡ (I − vv∗)A and so on.

  • For example in the 1st step of theQR-factorization, we can put

    v = α(A1 − β1e(1))with α =

    √2/‖A1 − β1e(1)‖2. If choose

    β1 ={ −‖A1‖2a11/|a11| if a11 6= 0−‖A1‖2 if a11 = 0

    then‖A1‖2 = ‖β1e(1)‖2, 〈A1, β1e(1)〉 = −|a11|‖A1‖2 ∈ RFrom the proof of Lemma 2,

    U(1)A1 ≡ (I − vv∗)A1 = β1e(1)

    Repeat the process for the (m− 1)× (n− 1)sub-matrix of U(1)A ≡ (I − vv∗)A and so on.

  • Example (See also P253)Find the QR-factorization of the matrix

    A =

    63 41 −8842 60 510 −28 56

    126 82 −71

  • U(1)A =1

    35

    −5145 −3675 29400 1078 29890 −980 19600 −196 1127

    U(2)U(1)A =

    −147 −105 840 −42 −210 0 96.92310 0 40.3846

    Q∗A = U(3)U(2)U(1)A = 21

    −7 −5 40 −2 −10 0 −50 0 0

    = R

  • U(1)A =1

    35

    −5145 −3675 29400 1078 29890 −980 19600 −196 1127

    U(2)U(1)A =

    −147 −105 840 −42 −210 0 96.92310 0 40.3846

    Q∗A = U(3)U(2)U(1)A = 21

    −7 −5 40 −2 −10 0 −50 0 0

    = R

  • U(1)A =1

    35

    −5145 −3675 29400 1078 29890 −980 19600 −196 1127

    U(2)U(1)A =

    −147 −105 840 −42 −210 0 96.92310 0 40.3846

    Q∗A = U(3)U(2)U(1)A = 21

    −7 −5 40 −2 −10 0 −50 0 0

    = R

  • U(1)A =1

    35

    −5145 −3675 29400 1078 29890 −980 19600 −196 1127

    U(2)U(1)A =

    −147 −105 840 −42 −210 0 96.92310 0 40.3846

    Q∗A = U(3)U(2)U(1)A = 21

    −7 −5 40 −2 −10 0 −50 0 0

    = R

  • Exercise (in class)Find the QR-factorization of the matrix

    A =[ 3 2 3

    4 5 6

    ]

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{.

    Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1.

    -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1

    =⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2.

    -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2

    =⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3.

    -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;

    � A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"

  • Some Applications of QR-Factorization

    d QR ©)½nµeAn�¢Ý§K3��

    QÚþn�R¦� A = QR.

    ïá3QR©)þ�QR{´¦Ý�ÜA��k�{. Ä�g´µ-A1 , A.

    éA1�QR©) A1 = Q1R1. -A2 , R1Q1=⇒ éA2�QR©) A2 = Q2R2. -A3 , R2Q2=⇒ éA3�QR©) A3 = Q3R3. -A4 , R3Q3=⇒ · · ·

    ´§þã�ÝS�{Ak}�Ñq;� A÷v½^§{Ak}Âñ�diag{λ1, · · · , λn}.k,��ÓÆ?Ú�Ö�5.5!"