chapter 4

10
Chapter 4 Linear Independence and Linear Transformations 4.1 Linear Independence, Subspaces and Dimen- sion Suppose we are given vectors v 1 , ··· , v m in R n . We would like to know whether these vectors are linearly independent. This amounts to asking whether x 1 v 1 + ··· + x m v m = 0 (4.1.1) has a non-trivial solution (the trivial solution is x 1 = ··· = x m = 0, anything else is a non-trivial solution). Let A be the n × m matrix whose column vectors are given by v . Then, the above question is equivalent to asking whether Ax = 0, x =(x 1 , ··· x m ) T (4.1.2) where has a non-trivial solution. This is nothing other than (3.5.1). Theorem 8 tells us exactly when there are non-trivial solutions. If the rank r is equal to m, then there is only the trivial solution, x = 0, and thus, the vectors v 1 , ··· , v m are linearly independent. Otherwise, the vectors are linearly dependent. We state the above observation as a proposition. Proposition 2. Let v 1 , ··· , v m be vectors in R n , and let A be the n × m matirx whose column vectors are given by v 1 , ··· , v m . Then, if the rank r of A is equal to m, then the vectors v 1 , ··· , v m are linearly independent. If not, the vectors are linearly dependent. 51

Upload: ursml12

Post on 17-Jul-2016

213 views

Category:

Documents


0 download

DESCRIPTION

chapter 4 and et all of physics

TRANSCRIPT

Page 1: chapter 4

Chapter 4

Linear Independence and

Linear Transformations

4.1 Linear Independence, Subspaces and Dimen-

sion

Suppose we are given vectors v1, · · · ,vm in Rn. We would like to know

whether these vectors are linearly independent. This amounts to askingwhether

x1v1 + · · · + xmvm = 0 (4.1.1)

has a non-trivial solution (the trivial solution is x1 = · · · = xm = 0, anythingelse is a non-trivial solution). Let A be the n × m matrix whose columnvectors are given by vℓ. Then, the above question is equivalent to askingwhether

Ax = 0, x = (x1, · · · xm)T (4.1.2)

where has a non-trivial solution. This is nothing other than (3.5.1). Theorem8 tells us exactly when there are non-trivial solutions. If the rank r is equalto m, then there is only the trivial solution, x = 0, and thus, the vectorsv1, · · · ,vm are linearly independent. Otherwise, the vectors are linearlydependent.

We state the above observation as a proposition.

Proposition 2. Let v1, · · · ,vm be vectors in Rn, and let A be the n × m

matirx whose column vectors are given by v1, · · · ,vm. Then, if the rank r

of A is equal to m, then the vectors v1, · · · ,vm are linearly independent. If

not, the vectors are linearly dependent.

51

Page 2: chapter 4

As an immediate consequence of the above, we have the following.

Proposition 3. It is impossible to have more than n linearly independent

vectors in Rn.

Proof. Suppose we have m > n linearly independent vectors. This impliesthat the n × m matrix A formed by these vectors must have rank m, byProposition 2. But the row-echelon matrix is a n×m matrix, and therefore,can only have at most n pivot columns. Therefore, its rank is at most n.Since m < n, this is a contradiction.

Another consequence of Proposition 2 is the following.

Theorem 9. For a n×n matrix A, the following statements are equivalent.

1. A is invertible.

2. The column vectors of A are linearly independent.

3. The row vectors of A are linearly independent.

Proof. Item (2) is equivalent to the statement that

Ax = 0 (4.1.3)

has only the trivial solution. The equivalence of item 1 and item 2 thusfollows from Theorem 6. Since the row vectors of A are the column vectorsof AT, item 3 is equivalent to AT being invertible. So we have only to showthat the invertibility of A is equivalent to the invertibility of AT. SupposeA is invertible. Then, there is a matrix B such that

AB = BA = I. (4.1.4)

Let us now take the transpose of the above, and use (1.4.3).

BTAT = ATBT = IT = I. (4.1.5)

We see therefore that BT is the inverse of AT and therefore, AT is invertible.Since the transpose of AT is A, we can repeat the same argument to showthat the invertibility of AT implies the invertibility of A.

Now, suppose the rank r of A in (4.1.2) is smaller than m. Let us lookat the situation a little further. As we did in Section 3.4, we label thecolumns with pivots as i1 < i2 < · · · < ir and the columns without pivots as

MATH 2574H 52 Yoichiro Mori

Page 3: chapter 4

j1 < j2 < · · · < jm−r. According to theorem 8, the solution to (4.1.2) canbe written as:

x = c1a1 + · · ·+ cm−ram−r. (4.1.6)

Take any aℓ, ℓ = 1, · · · ,m−r. x = aℓ is a solution to (4.1.2), and this impliesthat the vector vjℓ can be written as a linear combination of viq , iq < jℓ. Onthe other hand, the vectors vi1 , · · · ,vir are linearly independent. Indeed,if vi1 , · · · ,vir is linearly dependent, there will be a nontrivial solution to(4.1.2) such that xjℓ = 0 for all jℓ. But xjℓ = 0 implies cℓ = 0 in (4.1.6), acontradiction. Let us put this into a proposition.

Proposition 4. Suppose we have a n ×m matrix A whose rank is r, and

suppose we reduced A to row echelon form R. The r column vectors of A

corresponding to the pivot columns of R are linearly independent, and the

rest of the column vectors of A can be written as linear combinations of these

r vectors.

Example 8. Consider the column vectors of the 4× 5 matrix (3.5.5). The

rank of this matrix is 3, as can be seen by (3.5.6). Looking at the row echelon

form of this matrix (3.5.6), we see that the following column vectors 1, 2, 5are linearly independent:

v1 =

1−123

,v2 =

23−23

,v5 =

1320

. (4.1.7)

The column vectors v3 and v4 are expressed as:

v3 = v1 + v2, v4 = v1 − v2. (4.1.8)

To proceed further, let us introduce the notion of a subspace.

Definition 4 (Subspace of Rn). A subspace V of Rn is a subset of Rn with

the following two properties.

1. For a vector v ∈ V and an arbitrary scalar c, cv also belongs to V .

2. For two vectors v,w ∈ V , v+w is also in V .

Given m vectors v1, · · · ,vm, the set of all vectors:

c1v1 + · · ·+ cmvm (4.1.9)

forms a subspace. If all vectors in a subspace V can be written as linearcombinations of vectors v1, · · · ,vm, we say that the vectors v1, · · · ,vm spans

V .

MATH 2574H 53 Yoichiro Mori

Page 4: chapter 4

Proposition 5. Every subspace V of Rn is spanned by a finite number of

linearly independent vectors.

Proof. If the subspace consists of just the 0 vector, there is nothing to prove.Suppose otherwise. Pick a non-zero vector v1 that is in V . Consider thespan:

c1v1, c1 ∈ R. (4.1.10)

If this spans all of V , we are done. If not, there must be a vector v2 thatcannot be expressed in the above form. Therefore, v1 and v2 are linearlyindependent and

c1v1 + c2v2, c1, c2 ∈ R (4.1.11)

must belong to V . If this spans all of V , we are done. If not, we add anothervector v3 not expressible as above. This is thus linearly independent withrespect to the rest. This process has to stop before we add the n+1st vector,since there are at most n linearly vectors in R

n, according to Proposition3.

Definition 5 (Basis). Suppose a subspace V of Rn is spanned by linearly

independent vectors v1, · · · ,vm. We say that such vectors are a set of basisvectors of V .

Proposition 5 thus states that every subspace has a basis.

Example 9. Consider the following subset V of R3:

c1

124

+ c2

110

, (4.1.12)

where c1 and c2 are arbitrary constants. This is a subspace of R3 spanned

by (1, 2, 4)T and (1, 1, 0)T . The two vectors are linearly independent, and

therefore, the two vectors form a basis of V and the dimension of V is 2. It

is also possible to express the same subspace as:

c1

234

+ c2

110

. (4.1.13)

There are thus many different ways of expressing the same subspace. It is

also true that V can be expressed as:

c1

124

+ c2

234

+ c3

110

. (4.1.14)

MATH 2574H 54 Yoichiro Mori

Page 5: chapter 4

but in this case, the three vectors are not linearly independent.

As we have seen above, there are various choices for basis vectors of asubspace, but the number of basis vectors is always the same.

Proposition 6. Two sets of basis vectors always has the same number of

vectors.

Proof. Suppose otherwise. Then, there are basis vectors v1, · · · ,vm andw1, · · · ,wq with m 6= q. Suppose m < q. Then, each vector in wk can bewritten as:

wk = a1kv1 + a2kv2 + · · ·+ amkvm. (4.1.15)

where the ajk are scalar constants. To examine linear independence of wk,we must examine the expression

q∑

k=1

xkwk =

m∑

j=1

(

q∑

k=1

ajkxk

)

vj = 0, (4.1.16)

where xk are scalars. Since the vj are linearly independent, we have:

q∑

k=1

ajkxk = 0 for j = 1, · · · ,m (4.1.17)

This is a linear homogeneous equation with m equations in q unknownsx1, · · · , xq. Since m < q, by Proposition 1 there is a non-trivial solution.This contradicts the assumption that w1, · · · ,wq were linearly independent.The case q < m can be handled in exactly the same manner.

The above proposition allows us to define the dimension of a subspace.

Definition 6. The dimension of a subspace V in Rn is the number of basis

vectors of the subspace.

In particular, this means that subspaces of Rn can be classified by theirdimension. Subspaces of R2 are:

• Dimension 0: the origin.

• Dimension 1: lines through the origin.

• Dimension 2: the whole plane.

Subspaces of R3 are:

MATH 2574H 55 Yoichiro Mori

Page 6: chapter 4

• Dimension 0: the origin.

• Dimension 1: lines through the origin.

• Dimension 2: planes through the origin.

• Dimension 3: the whole space.

A similar classification is possible for Rn.

Example 10. Consider the vectors:

v1 =

123

, v2 =

456

, v3 =

789

. (4.1.18)

The span of these three vectors form a subspace in R3. To find the dimension

of the subspace, form the matrix consisting of these three vectors:

A =

1 4 72 5 83 6 9

(4.1.19)

The row echelon form is:

R =

1 0 −10 1 20 0 0

(4.1.20)

This shows that the space spanned by the three vectors is two dimensional,

and is spanned by vectors v1 and v2, with

v3 = 2v2 − v1. (4.1.21)

v1 and v2 are not the only vectors that form a basis of this subspace. Indeed,

one can find any number of bases. For example, v1 and v3 is also a basis of

the same subspace.

4.2 Linear Transformations

An n × n matrix A can be seen as a map from Rn to R

n. This is calleda linear transformation. We define two important concepts for a lineartransformation.

MATH 2574H 56 Yoichiro Mori

Page 7: chapter 4

Definition 7 (Kernel and Image). The kernel or nullspace of a matrix A

is set of vectors v ∈ Rn that satisfy:

Av = 0. (4.2.1)

The kernel of A is written as kerA. The image of a matrix A is the set of

vectors in v ∈ Rn for which

Ax = v (4.2.2)

has a solution x. The image of A is written as ImA.

Both the kernel and image are subspaces of Rn. This can be seen asfollows. Suppose v and w are in kerA. Then,

A(cv) = cA(v) = 0, A(v +w) = Av +Aw = 0. (4.2.3)

Therefore, cv and v +w are in the kernel of A. Take two vectors v and w

in the image of A. This means that there are vectors x and y such that

Ax = v, Ay = w. (4.2.4)

Therefore, we have

A(cx) = cAx = cv, A(x+ y) = Ax+Ay = v +w. (4.2.5)

Since the kernel and image are both subspaces of Rn, we can considertheir dimension. Let us now consider the dimension of the kernel and theimage.

Proposition 7. Let A be a n × n matrix. The dimension of the kernel is

equal to n− r, where r is the rank of the matrix.

Proof. Finding the kernel is the same as solving the equation:

Ax = 0, x ∈ Rn. (4.2.6)

We know from item 1 of Theorem 8 that the solution to the above is writtenas a linear combination of n−r linearly independent vectors. This is nothingother than the statement that the kernel has dimension n− r.

We now turn to the image.

Proposition 8. Let A be a n × n matrix. The dimension of the image is

equal to the rank r of the matrix A.

MATH 2574H 57 Yoichiro Mori

Page 8: chapter 4

Proof. The image of the matrix A consists of all vectors of the form:

Ax = x1v1 + · · ·+ cnvn (4.2.7)

where v1, · · · ,vn are the column vectors of A and x = (x1, · · · , xn). There-fore, the image is spanned by the column vectors. We know from Proposition4 that the r vectors that correspond to the pivots are linearly independentand that the rest are written as linear combinations of the others.

Thus follows the main result of this section.

Theorem 10. Suppose A is a n× n square matrix. Then,

rankA+ dimKerA = n, (4.2.8)

where dimKerA is the dimension of the kernel of A.

Example 11. Consider the matrix A and its row-reduced form R:

A =

1 2 31 3 50 1 2

, B =

1 0 −10 1 20 0 0

(4.2.9)

We see from this that the image is two-dimensional, where

v1 =

110

, v2 =

231

(4.2.10)

can be taken as a basis. The kernel of A is one-dimensional and is spanned

by:

1−21

. (4.2.11)

Consider the matrix and its row-reduced form

A =

1 4 3 22 7 5 33 4 1 −1−1 −1 0 1

, R =

1 0 −1 −20 1 1 10 0 0 00 0 0 0

. (4.2.12)

The image is 2-dimensional and is spanned by:

123−1

,

474−1

. (4.2.13)

MATH 2574H 58 Yoichiro Mori

Page 9: chapter 4

The kernel is also 2-dimensional and is spanned by:

1−110

,

2−101

. (4.2.14)

4.3 Exercises

1. Consider the linear dependence/independence of the following set ofvectors. If linearly dependent, find a set of linearly independent vectorsand express the other vectors in terms of them.

(a) (1, 3, 1)T, (1, 0, 1)T, (1, 0,−1)T , (3, 3, 1)T.

(b) (2, 1, 0)T, (0, 1,−2)T, (1, 0, 1)T , (1, 1,−1)T.

(c) (3, 0, 0, 3)T , (1, 0, 1, 0)T , (0, 1, 0, 0)T .

2. Consider the xy plane (the plane z = 0) in the three-dimensional spaceR3. Find two different sets of basis vectors for the xy plane.

3. Let x = (x, y, z, w)T ∈ R4. Consider the set of all vectors in R

4

consisting of vectors w = 0.

(a) Show that this set is a subspace of R4.

(b) Find a set of basis vectors for this subspace. What is its dimen-sion?

4. Argue why the following subsets of R2 are not subpsaces.

(a) The inside of a circle in R2 centered at the origin.

(b) A line in R2 that does not go through the origin.

(c) The first quadrant of R2.

(d) The first and third qudrants of R2 combined (including the x andy axes).

5. Consider two n× n matrices A and B. The matrix A has rank n andB has rank r. What is the rank of the matrix AB? What about BA?Can you say anything about the rank of A+B?

MATH 2574H 59 Yoichiro Mori

Page 10: chapter 4

6. Find the image and kernel of the following matrices.

2 −1 0−1 2 −10 −1 2

,

1 2 12 3 41 1 1

,

1 0 10 0 01 0 1

,

1 2 1 5−1 2 −3 21 6 −1 120 4 −2 7

,

0 −1 0 −11 1 1 01 2 3 02 2 4 −1

7. Consider the matrix:

A =1

3

1 1 11 1 11 1 1

(a) Find the image and kernel of A.

(b) Let v be a vector on the line spanned by (1, 1, 1)T. Where doesv get mapped to?

(c) Let w be a vector perpendicular to the vector (1, 1, 1)T. Wheredoes w get mapped to?

(d) Geometrically describe what kind of linear transformation A is.

(e) Show that A2 = A, and hence, (I −A)2 = (I −A).

(f) Geometrically describe what kind of linear transformation I −A

is.

MATH 2574H 60 Yoichiro Mori