• Tidak ada hasil yang ditemukan

Relations Between Vector Spaces and Their Dimensions

Dalam dokumen Jörg Liesen Volker Mehrmann (Halaman 132-138)

128 9 Vector Spaces Due to the uniqueness of the coordinates ofvwith respect to the basis{w1, . . . , wn} we obtain

⎢⎣ µ1

... µn

⎥⎦=P

⎢⎣ λ1

... λn

⎥⎦, or

⎢⎣ λ1

... λn

⎥⎦=P−1

⎢⎣ µ1

... µn

⎥⎦.

Hence a multiplication with the matrixPtransforms the coordinates ofvwith respect to the basis{v1, . . . , vn}into those with respect to the basis{w1, . . . , wn}; a multipli- cation withP−1yields the inverse transformation. Therefore,PandP−1are called coordinate transformation matrices.

We can summarize the results obtained above as follows.

Theorem 9.25 Let{v1, . . . , vn}and{w1, . . . , wn}be bases of a K -vector spaceV. Then the uniquely determined matrix PKn,n is (9.4) is invertible and yields the coordinate transformation from{v1, . . . , vn}to{w1, . . . , wn}: If

v=(v1, . . . , vn)

⎢⎣ λ1

... λn

⎥⎦=(v1, . . . , vn)

⎢⎣ µ1

... µn

⎥⎦,

then

⎢⎣ µ1

... µn

⎥⎦=P

⎢⎣ λ1

... λn

⎥⎦.

Example 9.26 Consider the vector spaceV =R2 = {(α12)|α12 ∈ R}with the entrywise addition and scalar multiplication. A basis ofV is given by the set {e1=(1,0),e2=(0,1)}, and we have(α12)=α1e12e2for all(α12)∈V. Another basis ofVis the set{v1=(1,1), v2=(1,2)}. The corresponding coordinate transformation matrices can be obtained from the defining equations (v1, v2) = (e1,e2)Pand(e1,e2)=(v1, v2)Qas

P= 1 1 1 2

!

, Q= P−1= 2−1

−1 1

! .

Proof LetU ⊆V and let {u1, . . . ,um}be a basis of U, where{u1, . . . ,um} = Ø for U = {0}. Using Theorem9.14we can extend this set to a basis ofV. IfU is a proper subset of V, then at least one basis vector needs to be added and hence dim(U) < dim(V). IfU =V, then every basis ofV is also a basis ofU, and thus

dim(U)=dim(V). ⊓⊔

IfU1andU2are subspaces of a vector spaceV, then theirintersectionis given by U1∩U2= {u ∈V |u ∈U1u∈U2}

(cp. Definition2.6). Thesumof the two subspaces is defined as U1+U2:= {u1+u2∈V |u1∈U1u2∈U2}.

Lemma 9.28 IfU1 andU2 are subspaces of a vector spaceV, then the following assertions hold:

(1) U1∩U2andU1+U2are subspaces ofV. (2) U1+U1=U1.

(3) U1+ {0} =U1.

(4) U1⊆U1+U2, with equality if and only ifU2⊆U1.

Proof Exercise. ⊓⊔

An important result is the followingdimension formula for subspaces.

Theorem 9.29 IfU1andU2are finite dimensional subspaces of a vector spaceV, then

dim(U1∩U2)+dim(U1+U2)=dim(U1)+dim(U2).

Proof Let {v1, . . . , vr} be a basis of U1 ∩ U2. We extend this set to a basis {v1, . . . , vr, w1, . . . , w}ofU1 and to a basis{v1, . . . , vr,x1, . . . ,xk}ofU2, where we assume thatr, ℓ,k≥1. (If one of the lists is empty, then the following argument is easily modified.)

If suffices to show that{v1, . . . , vr, w1, . . . , w,x1, . . . ,xk}is a basis ofU1+U2. Obviously,

span{v1, . . . , vr, w1, . . . , w,x1, . . . ,xk} =U1+U2,

and hence it suffices to show that v1, . . . , vr, w1, . . . , w,x1, . . . ,xk are linearly independent. Let

r i=1

λivi +

i=1

µiwi + k

i=1

γixi =0,

130 9 Vector Spaces

then

k i=1

γixi= − r

i=1

λivi +

i=1

µiwi

.

On the left hand side of this equation we have, by definition, a vector inU2; on the right hand side a vector in U1. Therefore,k

i=1γixi ∈ U1∩U2. By construction, however,{v1, . . . , vr}is a basis ofU1∩U2and the vectorsv1, . . . , vr, w1, . . . , ware linearly independent. Therefore,

i=1µiwi =0 implies thatµ1 = · · · = µ =0.

But then also

r i=1

λivi + k

i=1

γixi =0,

and henceλ1 = · · · =λr1 = · · · =γk =0 due to the linear independence of

v1, . . . , vr,x1, . . . ,xk. ⊓⊔

If at least one of the subspaces in Theorem9.29 is infinite dimensional, then the assertion is still formally correct, since in this case dim(U1+U2) = ∞and dim(U1)+dim(U2)= ∞.

Example 9.30 For the subspaces

U1= {[α12,0] |α12K}, U2= {[0,α23] |α23K} ⊂ K1,3 we have dim(U1)=dim(U2)=2,

U1∩U2 = {[0,α2,0] |α2K}, dim(U1∩U2)=1, U1+U2=K1,3, dim(U1+U2)=3.

The above definition of the sum can be extended to an arbitrary (but finite) number of subspaces: IfU1, . . . ,Uk,k ≥ 2, are subspaces of the vector spaceV, then we define

U1+. . .+Uk = k

j=1

Uj :=

k

j=1

uj |uj ∈Uj, j =1, . . . ,k

.

This sum is calleddirect, if

Uik

j=1 j=i

Uj= {0} for i=1, . . . ,k,

and in this case we write the (direct) sum as

U1⊕. . .⊕Uk=

"k j=1

Uj.

In particular, a sumU1+U2of two subspacesU1,U2⊆Vis direct ifU1∩U2 = {0}.

The following theorem presents two equivalent characterizations of the direct sum of subspaces.

Theorem 9.31 IfU =U1+. . .+Ukis a sum of k≥2subspaces of a vector space V, then the following assertions are equivalent:

(1) The sumUis direct, i.e.,Ui

j=iUj = {0}for i =1, . . . ,k.

(2) Every vector u ∈Uhas a representation of the form u=k

j=1ujwith uniquely determined uj ∈Uj for j=1, . . . ,k.

(3) k

j=1uj = 0 with uj ∈ Uj for j = 1, . . . ,k implies that uj = 0 for j = 1, . . . ,k.

Proof

(1) ⇒ (2): Letu = k

j=1uj = k

j=1#uj withuj,#uj ∈ Uj, j = 1, . . . ,k. For everyi =1, . . . ,kwe then have

ui−#ui = −

j=i

(uj−#uj) ∈ Ui

j=i

Uj.

NowUi

j=iUj = {0}implies thatui −#ui = 0, and henceui = #ui for i=1, . . . ,k.

(2) ⇒(3):This is obvious.

(3) ⇒ (1):For a giveni, letu ∈ Ui

j=iUj. Then u =

j=iuj for some uj ∈Uj, j =i, and hence−u+

j=iuj =0. In particular, this implies that u=0, and thusUi

j=iUj = {0}. ⊓⊔

Exercises

(In the following exercisesK is an arbitrary field.)

9.1. Which of the following sets (with the usual addition and scalar multiplication) areR-vector spaces?

12] ∈R1,212 ,

12] ∈R1,21222=1 ,

12] ∈R1,21≥α2

,

12] ∈R1,21−α2=0 and 2α12=0 . Determine, if possible, a basis and the dimension.

9.2. Determine a basis of theR-vector spaceCand dimR(C). Determine a basis of theC-vector spaceCand dimC(C).

9.3. Show thata1, . . . ,anKn,1are linearly independent if and only if det([a1,…, an])=0.

132 9 Vector Spaces 9.4. LetVbe aK-vector space,a nonempty set and Map(,V)the set of maps

fromtoV. Show that Map(,V)with the operations

+ :Map(,V)×Map(,V)→Map(,V), (f,g)→ f +g, with(f +g)(x):= f(x)+g(x) for all x∈,

· : K×Map(,V)→Map(,V), (λ,f)→λ· f, with(λ· f)(x):=λf(x) for allx ∈,

is aK-vector space.

9.5. Show that the functions sin and cos in Map(R,R)are linearly independent.

9.6. LetVbe a vector space withn =dim(V)∈Nand letv1, . . . , vn ∈V. Show that the following statements are equivalent:

(1) v1, . . . , vn are linearly independent.

(2) span{v1, . . . , vn} =V. (3) {v1, . . . , vn}is a basis ofV.

9.7. Show that(Kn,m,+,·)is a K-vector space (cp. (1) in Example9.2). Find a subspace of thisK-vector space.

9.8. Show that(K[t],+,·)is a K-vector space (cp. (2) in Example9.2). Show further thatK[t]nis a subspace ofK[t](cp. (3) in Example9.6) and determine dim(K[t]n).

9.9. Show that the polynomials p1 = t5+t4, p2 = t5 −7t3, p3 = t5 −1, p4 =t5+3t are linearly independent inQ[t]≤5 and extend{p1,p2,p3,p4} to a basis ofQ[t]≤5.

9.10. Letn∈Nand

K[t1,t2] :=

n

i,j=0

αi jt1it2j $$αi jK

.

An element ofK[t1,t2]is calledbivariate polynomialoverKin the unknowns t1 and t2. Define a scalar multiplication and an addition so that K[t1,t2] becomes a vector space. Determine a basis ofK[t1,t2].

9.11. Show Lemma9.5.

9.12. LetAKn,mandbKn,1. Is the solution setL(A,b)ofAx =ba subspace ofKm,1?

9.13. Let AKn,n and letλ∈ K be an eigenvalue of A. Show that the set{v ∈ Kn,1|Av=λv}is a subspace ofKn,1.

9.14. Let AKn,n and letλ12 be two eigenvalues of A. Show that any two associated eigenvectorsv1andv2are linearly independent.

9.15. Show thatB= {B1,B2,B3,B4}andC= {C1,C2,C3,C4}with

B1= 1 1 0 0

!

, B2= 1 0 0 0

!

, B3= 1 0 1 0

!

, B4= 1 1 0 1

!

and

C1= 1 0 0 1

!

, C2= 1 0 1 0

!

, C3= 1 0 0 0

!

, C4 = 0 1 1 0

!

are bases of the vector spaceK2,2, and determine corresponding coordinate transformation matrices.

9.16. Examine the elements of the following sets for linear independence in the vector spaceK[t]≤3:

U1= {t,t2+2t,t2+3t+1,t3}, U2= {1,t,t+t2,t2+t3}, U3= {1,t2t,t2+t,t3}.

Determine the dimensions of the subspaces spanned by the elements ofU1, U2,U3. Is one of these sets a basis ofK[t]≤3?

9.17. Show that the set of sequences{(α123, . . .)|αi∈Q,i ∈N}with entry- wise addition and scalar multiplication forms an infinite dimensional vector space, and determine a basis system.

9.18. Prove Lemma9.23.

9.19. Prove Lemma9.24.

9.20. Prove Lemma9.28.

9.21. LetU1,U2be finite dimensional subspaces of a vector spaceV. Show that the sumU1+U2is direct if dim(U1+U2)=dim(U1)+dim(U2).

9.22. LetU1, . . . ,Uk,k ≥ 3, be finite dimensional subspaces of a vector spaceV. Suppose thatUi∩Uj = {0}for alli = j. Is the sumU1+. . .+Ukdirect?

9.23. LetU be a subspace of a finite dimensional vector spaceV. Show that there exists another subspaceU#withU ⊕U# = V. (The subspace U#is called a complementofU.)

9.24. Determine three subspacesU1,U2,U3 ofV = R3,1 withU2 = U3 andV = U1⊕U2=U1⊕U3. Is there a subspaceU1ofVwith a uniquely determined complement?

Chapter 10

Linear Maps

In this chapter we study maps between vector spaces that are compatible with the two vector space operations, addition and scalar multiplication. These maps are called linear maps or homomorphisms. We first investigate their most important properties and then show that in the case of finite dimensional vector spaces every linear map can be represented by a matrix, when bases in the respective spaces have been chosen.

If the bases are chosen in a clever way, then we can read off important properties of a linear map from its matrix representation. This central idea will arise frequently in later chapters.

Dalam dokumen Jörg Liesen Volker Mehrmann (Halaman 132-138)