Linear Algebra
Sixth Edition
Seymour Lipschutz, PhD
Temple UniversityMarc Lars Lipson, PhD
University of VirginiaSchaum’s Outline Series
Copyright Act of 1976, no part of this publication may be reproduced or distributed in any form or by any means, or stored in a database or retrieval system, without the prior written permission of the publisher.
ISBN: 978-1-26-001145-6 MHID: 1-26-001145-3
The material in this eBook also appears in the print version of this title: ISBN: 978-1-26-001144-9, MHID: 1-26-001144-5.
eBook conversion by codeMantra Version 1.0
All trademarks are trademarks of their respective owners. Rather than put a trademark symbol after every
oc-currence of a trademarked name, we use names in an editorial fashion only, and to the benefit of the trademark
owner, with no intention of infringement of the trademark. Where such designations appear in this book, they have been printed with initial caps.
McGraw-Hill Education eBooks are available at special quantity discounts to use as premiums and sales promo-tions or for use in corporate training programs. To contact a representative, please visit the Contact Us page at www.mhprofessional.com.
SEYMOUR LIPSCHUTZ is on the faculty of Temple University and formally taught at the Polytechnic
In-stitute of Brooklyn. He received his PhD in 1960 at Courant InIn-stitute of Mathematical Sciences of New York
University. He is one of Schaum’s most prolific authors. In particular, he has written, among others, Beginning Linear Algebra, Probability, Discrete Mathematics, Set Theory, Finite Mathematics, and General Topology.
MARC LARS LIPSON is on the faculty of the University of Virginia and formerly taught at the University of Georgia, he received his PhD in finance in 1994 from the University of Michigan. He is also the coauthor of Discrete Mathematics and Probability with Seymour Lipschutz.
TERMS OF USE
This is a copyrighted work and McGraw-Hill Education and its licensors reserve all rights in and to the work. Use of this work is subject to these terms. Except as permitted under the Copyright Act of 1976 and the right to store and retrieve one copy of the work, you may not decompile, disassemble, reverse engineer, reproduce, modify, create derivative works based upon, transmit, distribute, disseminate, sell, publish or sublicense the work or any part of it without McGraw-Hill Education’s prior consent. You may use the work for your own noncommercial and personal use; any other use of the work is strictly prohibited. Your right to use the work may be terminated if you fail to comply with these terms.
Preface
Linear algebra has in recent years become an essential part of the mathematical background required by mathematicians and mathematics teachers, engineers, computer scientists, physicists, economists, and statis-ticians, among others. This requirement reflects the importance and wide applications of the subject matter.
This book is designed for use as a textbook for a formal course in linear algebra or as a supplement to all current standard texts. It aims to present an introduction to linear algebra which will be found helpful to all readers regardless of their fields of specification. More material has been included than can be covered in most first courses. This has been done to make the book more flexible, to provide a useful book of reference, and to stimulate further interest in the subject.
Each chapter begins with clear statements of pertinent definitions, principles, and theorems together with illustrative and other descriptive material. This is followed by graded sets of solved and supplementary problems. The solved problems serve to illustrate and amplify the theory, and to provide the repetition of basic principles so vital to effective learning. Numerous proofs, especially those of all essential theorems, are included among the solved problems. The supplementary problems serve as a complete review of the material of each chapter.
The first three chapters treat vectors in Euclidean space, matrix algebra, and systems of linear equations. These chapters provide the motivation and basic computational tools for the abstract investigations of vector spaces and linear mappings which follow. After chapters on inner product spaces and orthogonality and on determinants, there is a detailed discussion of eigenvalues and eigenvectors giving conditions for representing a linear operator by a diagonal matrix. This naturally leads to the study of various canonical forms, specifically, the triangular, Jordan, and rational canonical forms. Later chapters cover linear functions and the dual spaceV*, and bilinear, quadratic, and Hermitian forms. The last chapter treats linear operators on inner product spaces. The main changes in the sixth edition are that some parts in Appendix D have been added to the main part of the text, that is, Chapter Four and Chapter Eight. There are also many additional solved and supplementary problems.
Finally, we wish to thank the staff of the McGraw-Hill Schaum’s Outline Series, especially Diane Grayson, for their unfailing cooperation.
SEYMOURLIPSCHUTZ
MARCLARSLIPSON
List of Symbols
AH, conjugate transpose, 38
AT, transpose, 33
Aþ, Moore–Penrose inverse, 420
Aij, minor, 271 AðI;JÞ, minor, 275
AðVÞ, linear operators, 176 adjA, adjoint (classical), 273
AB, row equivalence, 72
A’B, congruence, 362
C, complex numbers, 11
Cn, complexn-space, 13
C½a;b, continuous functions, 230
CðfÞ, companion matrix, 306 colspðAÞ, column space, 120
dðu;vÞ, distance, 5, 243
HomðV;UÞ, homomorphisms, 176
i,j,k, 9
In, identity matrix, 33
Im F, image, 171
JðlÞ, Jordan block, 331
K, field of scalars, 112 KerF, kernel, 171
S?, orthogonal complement, 233 sgns, sign, parity, 269
spanðSÞ, linear span, 119 trðAÞ, trace, 33
½TS, matrix representation, 197
T*, adjoint, 379
T-invariant, 329
Tt, transpose, 353
kuk, norm, 5, 13, 229, 243
½uS, coordinate vector, 130
uv, dot product, 4, 13
V**, second dual space, 352
Vr
V, exterior product, 403
W0, annihilator, 353 z, complex conjugate, 12
Zðv;TÞ,T-cyclic subspace, 332
dij, Kronecker delta, 37
DðtÞ, characteristic polynomial, 296
l, eigenvalue, 298
P
Contents
List of Symbols iv
CHAPTER 1 Vectors in Rnand Cn, Spatial Vectors 1
1.1 Introduction 1.2 Vectors in Rn 1.3 Vector Addition and Scalar Multi-plication 1.4 Dot (Inner) Product 1.5 Located Vectors, Hyperplanes, Lines, Curves inRn 1.6 Vectors inR3(Spatial Vectors),ijkNotation 1.7 Complex Numbers 1.8 Vectors inCn
CHAPTER 2 Algebra of Matrices 27
2.1 Introduction 2.2 Matrices 2.3 Matrix Addition and Scalar Multiplica-tion 2.4 SummaMultiplica-tion Symbol 2.5 Matrix MultiplicaMultiplica-tion 2.6 Transpose of a Matrix 2.7 Square Matrices 2.8 Powers of Matrices, Polynomials in Matrices 2.9 Invertible (Nonsingular) Matrices 2.10 Special Types of Square Matrices 2.11 Complex Matrices 2.12 Block Matrices
CHAPTER 3 Systems of Linear Equations 57
3.1 Introduction 3.2 Basic Definitions, Solutions 3.3 Equivalent Systems, Elementary Operations 3.4 Small Square Systems of Linear Equations 3.5 Systems in Triangular and Echelon Forms 3.6 Gaussian Elimination 3.7 Echelon Matrices, Row Canonical Form, Row Equivalence 3.8 Gaussian Elimination, Matrix Formulation 3.9 Matrix Equation of a System of Linear Equations 3.10 Systems of Linear Equations and Linear Combinations of Vectors 3.11 Homogeneous Systems of Linear Equations 3.12 Elementary Matrices 3.13LU Decomposition
CHAPTER 4 Vector Spaces 112
4.1 Introduction 4.2 Vector Spaces 4.3 Examples of Vector Spaces 4.4 Linear Combinations, Spanning Sets 4.5 Subspaces 4.6 Linear Spans, Row Space of a Matrix 4.7 Linear Dependence and Independence 4.8 Basis and Dimension 4.9 Application to Matrices, Rank of a Matrix 4.10 Sums and Direct Sums 4.11 Coordinates 4.12 Isomorphism of V and Kn 4.13 Full Rank Factorization 4.14 Generalized (Moore–Penrose) Inverse 4.15 Least-Square Solution
CHAPTER 5 Linear Mappings 166
5.1 Introduction 5.2 Mappings, Functions 5.3 Linear Mappings (Linear Transformations) 5.4 Kernel and Image of a Linear Mapping 5.5 Singular and Nonsingular Linear Mappings, Isomorphisms 5.6 Operations with Linear Mappings 5.7 AlgebraA(V) of Linear Operators
CHAPTER 6 Linear Mappings and Matrices 197
6.1 Introduction 6.2 Matrix Representation of a Linear Operator 6.3 Change of Basis 6.4 Similarity 6.5 Matrices and General Linear Mappings
CHAPTER 7 Inner Product Spaces, Orthogonality 228
7.1 Introduction 7.2 Inner Product Spaces 7.3 Examples of Inner Product Spaces 7.4 Cauchy–Schwarz Inequality, Applications 7.5 Orthogonal-ity 7.6 Orthogonal Sets and Bases 7.7 Gram–Schmidt Orthogonalization
Process 7.8 Orthogonal and Positive Definite Matrices 7.9 Complex Inner Product Spaces 7.10 Normed Vector Spaces (Optional)
CHAPTER 8 Determinants 266
8.1 Introduction 8.2 Determinants of Orders 1 and 2 8.3 Determinants of Order 3 8.4 Permutations 8.5 Determinants of Arbitrary Order 8.6 Proper-ties of Determinants 8.7 Minors and Cofactors 8.8 Evaluation of Determi-nants 8.9 Classical Adjoint 8.10 Applications to Linear Equations, Cramer’s Rule 8.11 Submatrices, Minors, Principal Minors 8.12 Block Matrices and Determinants 8.13 Determinants and Volume 8.14 Determinant of a Linear Operator 8.15 Multilinearity and Determinants
CHAPTER 9 Diagonalization: Eigenvalues and Eigenvectors 294
9.1 Introduction 9.2 Polynomials of Matrices 9.3 Characteristic Polynomial, Cayley–Hamilton Theorem 9.4 Diagonalization, Eigenvalues and Eigenvec-tors 9.5 Computing Eigenvalues and EigenvecEigenvec-tors, Diagonalizing Matrices 9.6 Diagonalizing Real Symmetric Matrices and Quadratic Forms 9.7 Minimal Polynomial 9.8 Characteristic and Minimal Polynomials of Block Matrices
CHAPTER 10 Canonical Forms 327
10.1 Introduction 10.2 Triangular Form 10.3 Invariance 10.4 Invariant Direct-Sum Decompositions 10.5 Primary Decomposition 10.6 Nilpotent Operators 10.7 Jordan Canonical Form 10.8 Cyclic Subspaces 10.9 Rational Canonical Form 10.10 Quotient Spaces
CHAPTER 11 Linear Functionals and the Dual Space 351
11.1 Introduction 11.2 Linear Functionals and the Dual Space 11.3 Dual Basis 11.4 Second Dual Space 11.5 Annihilators 11.6 Transpose of a Linear Mapping
CHAPTER 12 Bilinear, Quadratic, and Hermitian Forms 361
12.1 Introduction 12.2 Bilinear Forms 12.3 Bilinear Forms and Matrices 12.4 Alternating Bilinear Forms 12.5 Symmetric Bilinear Forms, Quadratic Forms 12.6 Real Symmetric Bilinear Forms, Law of Inertia 12.7 Hermitian Forms
CHAPTER 13 Linear Operators on Inner Product Spaces 379
13.1 Introduction 13.2 Adjoint Operators 13.3 Analogy BetweenA(V) and
C, Special Linear Operators 13.4 Self-Adjoint Operators 13.5 Orthogonal
and Unitary Operators 13.6 Orthogonal and Unitary Matrices 13.7 Change of Orthonormal Basis 13.8 Positive Definite and Positive Operators 13.9 Diagonalization and Canonical Forms in Inner Product Spaces 13.10 Spec-tral Theorem
APPENDIX A Multilinear Products 398
APPENDIX B Algebraic Structures 405
APPENDIX C Polynomials over a Field 413
APPENDIX D Odds and Ends 417
Vectors in R
n
and C
n
,
Spatial Vectors
1.1
Introduction
There are two ways to motivate the notion of a vector: one is by means of lists of numbers and subscripts, and the other is by means of certain objects in physics. We discuss these two ways below.
Here we assume the reader is familiar with the elementary properties of the field of real numbers, denoted byR. On the other hand, we will review properties of the field of complex numbers, denoted by
C. In the context of vectors, the elements of our number fields are calledscalars.
Although we will restrict ourselves in this chapter to vectors whose elements come fromR and then fromC, many of our operations also apply to vectors whose entries come from some arbitrary field K.
Lists of Numbers
Suppose the weights (in pounds) of eight students are listed as follows:
156; 125; 145; 134; 178; 145; 162; 193
One can denote all the values in the list using only one symbol, sayw, but with different subscripts; that is,
w1; w2; w3; w4; w5; w6; w7; w8
Observe that each subscript denotes the position of the value in the list. For example,
w1 ¼156; the first number; w2¼125; the second number; . . .
Such a list of values,
w¼ ðw1;w2;w3;. . .;w8Þ
is called alinear array orvector.
Vectors in Physics
Many physical quantities, such as temperature and speed, possess only“magnitude.”These quantities can be represented by real numbers and are calledscalars. On the other hand, there are also quantities, such as force and velocity, that possess both “magnitude” and “direction.” These quantities, which can be represented by arrows having appropriate lengths and directions and emanating from some given reference pointO, are called vectors.
Now we assume the reader is familiar with the spaceR3 where all the points in space are represented by ordered triples of real numbers. Suppose the origin of the axes inR3 is chosen as the reference pointOfor the vectors discussed above. Then every vector is uniquely determined by the coordinates of its endpoint, and vice versa.
There are two important operations, vector addition and scalar multiplication, associated with vectors in physics. The definition of these operations and the relationship between these operations and the endpoints of the vectors are as follows.
(i) Vector Addition: The resultantuþvof two vectorsuandvis obtained by theparallelogram law; that is, uþv is the diagonal of the parallelogram formed byuand v. Furthermore, if ða;b;cÞ and
ða0;b0;c0Þare the endpoints of the vectorsuandv, thenðaþa0; bþb0; cþc0Þis the endpoint of the vectoruþv. These properties are pictured in Fig. 1-1(a).
(ii) Scalar Multiplication: The productkuof a vectoruby a real numberkis obtained by multiplying the magnitude ofubyk and retaining the same direction ifk>0 or the opposite direction ifk<0. Also, ifða;b;cÞis the endpoint of the vectoru, thenðka;kb;kcÞis the endpoint of the vectorku. These properties are pictured in Fig. 1-1(b).
Mathematically, we identify the vectoruwith itsða;b;cÞand writeu¼ ða;b;cÞ. Moreover, we call the ordered tripleða;b;cÞof real numbers a point or vector depending upon its interpretation. We generalize this notion and call ann-tupleða1;a2;. . .;anÞof real numbers a vector. However, special notation may be used for the vectors inR3 calledspatial vectors (Section 1.6).
1.2
Vectors in R
nThe set of alln-tuples of real numbers, denoted byRn, is calledn-space. A particularn-tuple inRn, say
u¼ ða1;a2;. . .;anÞ
is called apoint or vector. The numbers ai are called thecoordinates, components, entries, or elements
ofu. Moreover, when discussing the spaceRn, we use the termscalarfor the elements of R.
Two vectors,uandv, areequal, writtenu¼v, if they have the same number of components and if the
corresponding components are equal. Although the vectors ð1;2;3Þandð2;3;1Þcontain the same three numbers, these vectors are not equal because corresponding entries are not equal.
The vectorð0;0;. . .;0Þwhose entries are all 0 is called thezero vectorand is usually denoted by 0.
EXAMPLE 1.1
(a) The following are vectors:
ð2; 5Þ; ð7;9Þ; ð0;0;0Þ; ð3;4;5Þ
The first two vectors belong toR2, whereas the last two belong toR3. The third is the zero vector inR3. (b) Findx;y;zsuch thatðx y; xþy; z 1Þ ¼ ð4;2;3Þ.
By definition of equality of vectors, corresponding entries must be equal. Thus,
x y¼4; xþy¼2; z 1¼3
Solving the above system of equations yieldsx¼3,y¼ 1,z¼4.
Figure 1-1
z
y
x 0
u
ku
(a, b, c) (a, b, c)
(b) Scalar Multiplication
(ka, kb, kc) (a′, b′, c′)
z
y
x 0
v
u u +v
(a + a′, b + b′, c + c′)
Column Vectors
Sometimes a vector inn-spaceRn is written vertically rather than horizontally. Such a vector is called a
column vector, and, in this context, the horizontally written vectors in Example 1.1 are calledrow vectors. For example, the following are column vectors with 2;2;3, and 3 components, respectively:
1
We also note that any operation defined for row vectors is defined analogously for column vectors.
1.3
Vector Addition and Scalar Multiplication
Consider two vectorsuandv inRn, say
u¼ ða1;a2;. . .;anÞ and v¼ ðb1;b2;. . .;bnÞ
Theirsum, writtenuþv, is the vector obtained by adding corresponding components fromuandv. That is,
uþv¼ ða
1þb1; a2þb2; . . .; anþbnÞ
Theproduct, of the vectoruby a real number k, writtenku, is the vector obtained by multiplying each component ofubyk. That is,
ku¼kða1;a2;. . .;anÞ ¼ ðka1;ka2;. . .;kanÞ
Observe that uþv and ku are also vectors in Rn. The sum of vectors with different numbers of
components is not defined.
Negatives and subtraction are defined inRn as follows:
u¼ ð 1Þu and u v ¼uþ ð vÞ
The vector uis called thenegativeofu, andu v is called thedifference ofuandv.
Now suppose we are given vectorsu1;u2;. . .;uminRnand scalarsk1;k2;. . .;kminR. We can multiply the vectors by the corresponding scalars and then add the resultant scalar products to form the vector
v¼k
1u1þk2u2þk3u3þ þkmum
Such a vectorv is called alinear combinationof the vectors u
Basic properties of vectors under the operations of vector addition and scalar multiplication are described in the following theorem.
THEOREM1.1: For any vectorsu;v;win Rn and any scalars k;k0 in R,
(i) ðuþvÞ þw¼uþ ðvþwÞ, (v) kðuþvÞ ¼kuþkv,
(ii) uþ0¼u; (vi) ðkþk0Þu¼kuþk0u,
(iii) uþ ð uÞ ¼0; (vii) ðkk0Þu¼kðk0uÞ, (iv) uþv¼vþu, (viii) 1u¼u.
We postpone the proof of Theorem 1.1 until Chapter 2, where it appears in the context of matrices (Problem 2.3).
Supposeuandvare vectors inRn for whichu¼kvfor some nonzero scalarkinR. Thenuis called a
multipleofv. Also, uis said to be in thesameoropposite directionasv according to whetherk>0 or
k<0.
1.4
Dot (Inner) Product
Consider arbitrary vectorsuandv in Rn; say,
u¼ ða1;a2;. . .;anÞ and v¼ ðb1;b2;. . .;bnÞ
Thedot productorinner productofuandv is denoted and defined by
uv¼a
1b1þa2b2þ þanbn
That is, uv is obtained by multiplying corresponding components and adding the resulting products.
The vectorsuandv are said to beorthogonal (orperpendicular) if their dot product is zero—that is, if
uv¼0.
EXAMPLE 1.3
(a) Letu¼ ð1; 2;3Þ,v¼ ð4;5; 1Þ,w¼ ð2;7;4Þ. Then, uv¼1ð4Þ 2ð5Þ þ3ð 1Þ ¼4 10 3¼ 9
uw¼2 14þ12¼0; vw¼8þ35 4¼39 Thus,uand ware orthogonal.
(b) Letu¼
2 3 4
2
4 3
5andv¼
3 1 2
2
4 3
5. Thenuv¼6 3þ8¼11.
(c) Supposeu¼ ð1;2;3;4Þandv¼ ð6;k; 8;2Þ. Findkso thatuandvare orthogonal.
First obtainuv¼6þ2k 24þ8¼ 10þ2k. Then setuv¼0 and solve fork:
10þ2k¼0 or 2k¼10 or k¼5
Basic properties of the dot product inRn (proved in Problem 1.13) follow.
THEOREM1.2: For any vectorsu;v;win Rn and any scalar k inR:
(i) ðuþvÞ w¼uwþvw; (iii) uv¼vu,
(ii) ðkuÞ v¼kðuvÞ, (iv) uu0;anduu¼0 iffu¼0.
Note that (ii) says that we can“takekout”from the first position in an inner product. By (iii) and (ii),
That is, we can also“takek out”from the second position in an inner product.
The space Rn with the above operations of vector addition, scalar multiplication, and dot product is usually calledEuclidean n-space.
is the unique unit vector in the same direction asv. The process of findingv^fromvis callednormalizingv.
EXAMPLE 1.4
Thuswis a unit vector, butvis not a unit vector. However, we can normalizevas follows:
^
This is the unique unit vector in the same direction asv.
The following formula (proved in Problem 1.14) is known as the Schwarz inequality or Cauchy–
Schwarz inequality. It is used in many branches of mathematics.
THEOREM1.3 (Schwarz): For any vectors u;v in Rn,juvj kukkvk.
Using the above inequality, we also prove (Problem 1.15) the following result known as the“triangle inequality”or Minkowski’s inequality.
THEOREM1.4 (Minkowski): For any vectorsu;v inRn,kuþvk kuk þ kvk.
Theangleybetween nonzero vectorsu;v inRn is defined by
cosy¼ uv kukkvk
This definition is well defined, because, by the Schwarz inequality (Theorem 1.3),
1 uv
kukkvk1
Note that if uv¼0, then y¼90 (or y¼p=2). This then agrees with our previous definition of
orthogonality.
Theprojectionof a vectoruonto a nonzero vector v is the vector denoted and defined by
projðu;vÞ ¼ u
We show below that this agrees with the usual notion of vector projection in physics.
EXAMPLE 1.5
(b) Consider the vectorsuandvin Fig. 1-2(a) (with respective endpointsAandB). The (perpendicular) projection of uontovis the vectoru* with magnitude
ku*k ¼ kukcosy¼ kuk uv kukvk¼
uv
kvk
To obtainu*, we multiply its magnitude by the unit vector in the direction ofv, obtaining
u*¼ ku*k v
1.5
Located Vectors, Hyperplanes, Lines, Curves in R
nThis section distinguishes between ann-tuplePðaiÞ Pða1;a2;. . .;anÞ viewed as a point inRn and an
n-tupleu¼ ½c1;c2;. . .;cn viewed as a vector (arrow) from the originO to the pointCðc1;c2;. . .;cnÞ.
Located Vectors
Any pair of pointsAðaiÞandBðbiÞinRn defines thelocated vectorordirected line segmentfromAtoB, writtenAB!. We identify AB!with the vector
u¼B A¼ ½b1 a1; b2 a2; . . .; bn an
because AB! and u have the same magnitude and direction. This is pictured in Fig. 1-2(b) for the points Aða1;a2;a3Þ and Bðb1;b2;b3Þ in R3 and the vector u¼B A which has the endpoint
Pðb1 a1, b2 a2,b3 a3Þ.
Hyperplanes
Ahyperplane H inRn is the set of pointsðx1;x2;. . .;xnÞthat satisfy a linear equation
a1x1þa2x2þ þanxn ¼b
where the vectoru¼ ½a1;a2;. . .;anof coefficients is not zero. Thus a hyperplaneHinR2is a line, and a hyperplaneH inR3 is a plane. We show below, as pictured in Fig. 1-3(a) forR3, thatuis orthogonal to any directed line segmentPQ!, wherePðpiÞandQðqiÞare points inH:[For this reason, we say thatuis
normaltoH and thatH is normaltou:]
BecausePðpiÞ andQðqiÞbelong to H;they satisfy the above hyperplane equation—that is,
a1p1þa2p2þ þanpn¼b and a1q1þa2q2þ þanqn ¼b
v ¼PQ!¼Q P¼ ½q
1 p1;q2 p2;. . .;qn pn
Let
Then
uv¼a
1ðq1 p1Þ þa2ðq2 p2Þ þ þanðqn pnÞ
¼ ða1q1þa2q2þ þanqnÞ ða1p1þa2p2þ þanpnÞ ¼b b¼0
Thusv¼PQ!is orthogonal tou; as claimed.
Figure 1-3
(a) (b)
O
P Q
L P P + t1u
P − t2u u
u
Lines in R
nThe line L in Rn passing through the point Pðb1;b2;. . .;bnÞ and in the direction of a nonzero vector
u¼ ½a1;a2;. . .;anconsists of the pointsXðx1;x2;. . .;xnÞthat satisfy
X¼Pþtu or
x1¼a1tþb1 x2¼a2tþb2 :::::::::::::::::::: xn¼antþbn
or LðtÞ ¼ ðaitþbiÞ
8 > > < > > :
where theparameter t takes on all real values. Such a line Lin R3 is pictured in Fig. 1-3(b).
EXAMPLE 1.6
(a) LetH be the plane inR3 corresponding to the linear equation 2x 5yþ7z¼4. Observe that Pð1;1;1Þ and
Qð5;4;2Þare solutions of the equation. ThusPandQand the directed line segment
v¼PQ!¼Q P¼ ½5 1; 4 1; 2 1 ¼ ½4;3;1
lie on the planeH. The vectoru¼ ½2; 5;7is normal toH, and, as expected,
uv¼ ½2; 5;7 ½4;3;1 ¼8 15þ7¼0 That is,uis orthogonal tov.
(b) Find an equation of the hyperplaneH inR4 that passes through the pointPð1;3; 4;2Þ and is normal to the vectoru¼ ½4; 2;5;6.
The coefficients of the unknowns of an equation ofHare the components of the normal vectoru; hence, the equation ofHmust be of the form
4x1 2x2þ5x3þ6x4¼k
SubstitutingPinto this equation, we obtain
4ð1Þ 2ð3Þ þ5ð 4Þ þ6ð2Þ ¼k or 4 6 20þ12¼k or k¼ 10
Thus, 4x1 2x2þ5x3þ6x4¼ 10 is the equation ofH.
(c) Find the parametric representation of the lineLinR4passing through the pointPð1;2;3; 4Þand in the direction ofu¼ ½5;6; 7;8. Also, find the pointQonLwhent¼1.
Substitution in the above equation forLyields the following parametric representation:
x1 ¼5tþ1; x2¼6tþ2; x3 ¼ 7tþ3; x4 ¼8t 4
or, equivalently,
LðtÞ ¼ ð5tþ1;6tþ2; 7tþ3;8t 4Þ
Note thatt¼0 yields the pointPonL. Substitution oft¼1 yields the pointQð6;8; 4;4ÞonL.
Curves in R
nLetDbe an interval (finite or infinite) on the real lineR. A continuous functionF:D!Rn is acurvein
Rn. Thus, to each point t2Dthere is assigned the following point inRn:
FðtÞ ¼ ½F1ðtÞ;F2ðtÞ;. . .;FnðtÞ
Moreover, the derivative (if it exists) ofFðtÞ yields the vector
VðtÞ ¼dFðtÞ dt ¼
dF1ðtÞ dt ;
dF2ðtÞ dt ;. . .;
dFnðtÞ dt
which is tangent to the curve. NormalizingVðtÞyields
TðtÞ ¼ VðtÞ kVðtÞk
Thus, TðtÞis the unit tangent vector to the curve. (Unit vectors with geometrical significance are often presented in bold type.)
EXAMPLE 1.7 Consider the curveFðtÞ ¼ ½sint;cost;tinR3. Taking the derivative ofFðtÞ[or each component of
FðtÞ] yields
VðtÞ ¼ ½cost; sint;1
which is a vector tangent to the curve. We normalizeVðtÞ. First we obtain
kVðtÞk2 ¼cos2tþsin2tþ1¼1þ1¼2
Then the unit tangent vectionTðtÞ to the curve follows:
TðtÞ ¼ VðtÞ kVðtÞk¼
cost
ffiffiffi 2
p ; sinffiffiffit 2
p ; 1ffiffiffi 2
p
1.6
Vectors in R
3(Spatial Vectors), ijk Notation
Vectors inR3, calledspatial vectors, appear in many applications, especially in physics. In fact, a special notation is frequently used for such vectors as follows:
i¼ ½1;0;0denotes the unit vector in thexdirection:
j¼ ½0;1;0denotes the unit vector in theydirection:
k¼ ½0;0;1denotes the unit vector in thezdirection:
Then any vectoru¼ ½a;b;cin R3 can be expressed uniquely in the form
u¼ ½a;b;c ¼aiþbjþck
Because the vectors i;j;k are unit vectors and are mutually orthogonal, we obtain the following dot products:
ii¼1; jj¼1; kk¼1 and ij¼0; ik¼0; jk¼0
Furthermore, the vector operations discussed above may be expressed in the ijk notation as follows. Suppose
u¼a1iþa2jþa3k and v¼b
1iþb2jþb3k
Then
uþv¼ ða
1þb1Þiþ ða2þb2Þjþ ða3þb3Þk and cu¼ca1iþca2jþca3k
wherecis a scalar. Also,
uv¼a
1b1þa2b2þa3b3 and kuk ¼ ffiffiffiffiffiffiffiffiffi
uu p
¼
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
a2
1þa22þa23 q
EXAMPLE 1.8 Supposeu¼3iþ5j 2kandv¼4i 8jþ7k.
(a) To finduþv, add corresponding components, obtaininguþv¼7i 3jþ5k
(b) To find 3u 2v, first multiply by the scalars and then add:
(c) To finduv, multiply corresponding components and then add: uv¼12 40 14¼ 42
(d) To findkuk, take the square root of the sum of the squares of the components:
kuk ¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi9þ25þ4¼pffiffiffiffiffi38
Cross Product
There is a special operation for vectorsuandvinR3that is not defined inRnforn6¼3. This operation is
called thecross productand is denoted byuv. One way to easily remember the formula foruvis to
use the determinant (of order two) and its negative, which are denoted and defined as follows:
a b
Here a and d are called the diagonal elements and b and c are the nondiagonal elements. Thus, the determinant is the productadof the diagonal elements minus the productbcof the nondiagonal elements, but vice versa for the negative of the determinant.
Now supposeu¼a1iþa2jþa3kandv¼b That is, the three components ofuv are obtained from the array
a1 a2 a3 b1 b2 b3
(which contain the components ofuabove the component ofv) as follows:
(1) Cover the first column and take the determinant.
(2) Cover the second column and take the negative of the determinant. (3) Cover the third column and take the determinant.
Note that uv is a vector; hence, uv is also called the vector product or outer product of u
andv.
Remark: The cross products of the vectors i;j;kare as follows:
ij¼k; jk¼i; ki¼j
ji¼ k; kj¼ i; ik¼ j
Thus, if we view the tripleði;j;kÞas a cyclic permutation, whereifollowskand hencekprecedesi, then the product of two of them in the given direction is the third one, but the product of two of them in the opposite direction is the negative of the third one.
THEOREM1.5: Letu;v;w be vectors inR3.
(a) The vectoruv is orthogonal to bothuandv.
(b) The absolute value of the “triple product”
uvw
represents the volume of the parallelepiped formed by the vectors u;v, w.
[See Fig. 1-4(a).]
We note that the vectorsu;v,uvform a right-handed system, and that the following formula gives
the magnitude ofuv:
kuvk ¼ kukkvksiny
whereyis the angle between uandv.
1.7
Complex Numbers
The set of complex numbers is denoted byC. Formally, a complex number is an ordered pairða;bÞof real numbers where equality, addition, and multiplication are defined as follows:
ða;bÞ ¼ ðc;dÞ if and only ifa¼candb¼d ða;bÞ þ ðc;dÞ ¼ ðaþc; bþdÞ
ða;bÞ ðc;dÞ ¼ ðac bd; adþbcÞ
We identify the real numberawith the complex number ða;0Þ; that is,
a$ ða;0Þ
This is possible because the operations of addition and multiplication of real numbers are preserved under the correspondence; that is,
ða;0Þ þ ðb;0Þ ¼ ðaþb; 0Þ and ða;0Þ ðb;0Þ ¼ ðab;0Þ
Thus we viewRas a subset of C, and replace ða;0Þ byawhenever convenient and possible.
We note that the setCof complex numbers with the above operations of addition and multiplication is a
fieldof numbers, like the setR of real numbers and the setQof rational numbers.
Figure 1-4
(a) (b)
O
Volume = u.v × w Complex plane
y
x
z
v
O b
a z = a + bi
|z| u
The complex numberð0;1Þis denoted byi. It has the important property that
i2¼ii¼ ð0;1Þð0;1Þ ¼ ð 1;0Þ ¼ 1 or i¼pffiffiffiffiffiffiffi1
Accordingly, any complex numberz¼ ða;bÞ can be written in the form
z¼ ða;bÞ ¼ ða;0Þ þ ð0;bÞ ¼ ða;0Þ þ ðb;0Þ ð0;1Þ ¼aþbi
The above notation z¼aþbi, whereaRez andbImz are called, respectively, the real and
imaginary partsofz, is more convenient than ða;bÞ. In fact, the sum and product of complex numbers
z¼aþbi and w¼cþdi can be derived by simply using the commutative and distributive laws and
i2 ¼ 1:
zþw¼ ðaþbiÞ þ ðcþdiÞ ¼aþcþbiþdi¼ ðaþbÞ þ ðcþdÞi
zw¼ ðaþbiÞðcþdiÞ ¼acþbciþadiþbdi2 ¼ ðac bdÞ þ ðbcþadÞi
We also define thenegativeofzand subtraction inC by
z¼ 1z and w z¼wþ ð zÞ
Warning:The letteri representingpffiffiffiffiffiffiffi1has no relationship whatsoever to the vectori¼ ½1;0;0in Section 1.6.
Complex Conjugate, Absolute Value
Consider a complex numberz¼aþbi. Theconjugate ofzis denoted and defined by
z¼aþbi¼a bi
Thenzz¼ ðaþbiÞða biÞ ¼a2 b2i2 ¼a2þb2. Note thatz is real if and only ifz¼z.
Theabsolute valueofz, denoted by jzj, is defined to be the nonnegative square root ofzz. Namely,
jzj ¼pffiffiffiffizz¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffia2þb2
Note thatjzjis equal to the norm of the vectorða;bÞin R2.
Supposez6¼0. Then the inversez 1 ofzand division in Cofw byzare given, respectively, by
z 1 ¼ z zz¼
a a2þb2
b
a2þb2i and w
z wz
zz ¼wz 1
EXAMPLE 1.10 Supposez¼2þ3iandw¼5 2i. Then
zþw¼ ð2þ3iÞ þ ð5 2iÞ ¼2þ5þ3i 2i¼7þi
zw¼ ð2þ3iÞð5 2iÞ ¼10þ15i 4i 6i2¼16þ11i z¼2þ3i¼2 3i and w¼5 2i¼5þ2i w
z ¼
5 2i
2þ3i¼
ð5 2iÞð2 3iÞ ð2þ3iÞð2 3iÞ¼
4 19i
13 ¼
4 13
19 13i
jzj ¼pffiffiffiffiffiffiffiffiffiffiffi4þ9¼pffiffiffiffiffi13 and jwj ¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffi25þ4¼pffiffiffiffiffi29
Complex Plane
Recall that the real numbersRcan be represented by points on a line. Analogously, the complex numbers
1.8
Vectors in C
nThe set of alln-tuples of complex numbers, denoted byCn, is calledcomplexn-space. Just as in the real case, the elements ofCn are called points or vectors, the elements of C are called scalars, and vector addition inCn and scalar multiplication onCn are given by
½z1;z2;. . .;zn þ ½w1;w2;. . .;wn ¼ ½z1þw1; z2þw2; . . .; znþwn z½z1;z2;. . .;zn ¼ ½zz1;zz2;. . .;zzn
where thezi,wi, andzbelong toC.
EXAMPLE 1.11 Consider vectorsu¼ ½2þ3i; 4 i; 3andv¼ ½3 2i; 5i; 4 6iinC3. Then uþv ¼ ½2þ3i; 4 i; 3 þ ½3 2i; 5i; 4 6i ¼ ½5þi; 4þ4i; 7 6i
ð5 2iÞu ¼ ½ð5 2iÞð2þ3iÞ; ð5 2iÞð4 iÞ; ð5 2iÞð3Þ ¼ ½16þ11i; 18 13i; 15 6i
Dot (Inner) Product in C
nConsider vectorsu¼ ½z1;z2;. . .;znandv¼ ½w1;w2;. . .;wninCn. Thedotorinner productofuandvis
denoted and defined by
uv¼z
1w1þz2w2þ þznwn
This definition reduces to the real case becausewi¼wi whenwi is real. The norm of uis defined by
kuk ¼pffiffiffiffiffiffiffiffiffiuu¼ ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiz1z1þz2z2þ þznzn p
¼
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi jz1j
2 þ jz2j
2
þ þ jv
nj 2 q
We emphasize thatuuand sokukare real and positive when u6¼0 and 0 whenu¼0.
EXAMPLE 1.12 Consider vectorsu¼ ½2þ3i; 4 i; 3þ5iandv¼ ½3 4i; 5i; 4 2iinC3. Then uv¼ ð2þ3iÞð3 4iÞ þ ð4 iÞð5iÞ þ ð3þ5iÞð4 2iÞ
¼ ð2þ3iÞð3þ4iÞ þ ð4 iÞð 5iÞ þ ð3þ5iÞð4þ2iÞ ¼ ð 6þ13iÞ þ ð 5 20iÞ þ ð2þ26iÞ ¼ 9þ19i
uu¼ j2þ3ij2þ j4 ij2þ j3þ5ij2 ¼ 4þ9þ16þ1þ9þ25 ¼ 64
kuk ¼pffiffiffiffiffi64¼8
The spaceCn with the above operations of vector addition, scalar multiplication, and dot product, is calledcomplex Euclideann-space. Theorem 1.2 for Rn also holds forCn if we replaceuv ¼vuby
uv¼uv
On the other hand, the Schwarz inequality (Theorem 1.3) and Minkowski’s inequality (Theorem 1.4) are true forCn with no changes.
SOLVED PROBLEMS
Vectors in Rn
1.1. Determine which of the following vectors are equal:
u1 ¼ ð1;2;3Þ; u2¼ ð2;3;1Þ; u3¼ ð1;3;2Þ; u4¼ ð2;3;1Þ
1.2. Letu¼ ð2; 7;1Þ, v¼ ð 3;0;4Þ,w¼ ð0;5; 8Þ. Find:
(a) 3u 4v,
(b) 2uþ3v 5w.
First perform the scalar multiplication and then the vector addition.
(a) 3u 4v¼3ð2; 7;1Þ 4ð 3;0;4Þ ¼ ð6; 21;3Þ þ ð12;0; 16Þ ¼ ð18; 21; 13Þ
First perform the scalar multiplication and then the vector addition:
(a) 5u 2v¼5
(a) Because the vectors are equal, set the corresponding entries equal to each other, yielding
x¼2; 3¼xþy
(It is more convenient to write vectors as columns than as rows when forming linear combinations.) Set corresponding entries equal to each other to obtain
1.6. Writev¼ ð2; 5;3Þas a linear combination of
u1¼ ð1; 3;2Þ;u2¼ ð2; 4; 1Þ;u3¼ ð1; 5;7Þ:
Find the equivalent system of linear equations and then solve. First,
2 5 3
2 4
3 5¼x
1 3 2
2 4
3 5þy
2 4 1
2 4
3 5þz
1 5 7
2 4
3
5¼ 3xxþ24yyþ5zz
2x yþ7z 2
4
3 5
Set the corresponding entries equal to each other to obtain
xþ2yþ z¼ 2 3x 4y 5z¼ 5 2x yþ7z¼ 3
or
xþ2yþ z¼ 2 2y 2z¼ 1 5yþ5z¼ 1
or
xþ2yþ z¼2 2y 2z¼1 0¼3
The third equation, 0xþ0yþ0z¼3, indicates that the system has no solution. Thus,vcannot be written as a
linear combination of the vectorsu1,u2,u3.
Dot (Inner) Product, Orthogonality, Norm in Rn
1.7. Finduv where:
(a) u¼ ð2; 5;6Þ andv¼ ð8;2; 3Þ,
(b) u¼ ð4;2; 3;5; 1Þandv¼ ð2;6; 1; 4;8Þ.
Multiply the corresponding components and add:
(a) uv¼2ð8Þ 5ð2Þ þ6ð 3Þ ¼16 10 18¼ 12
(b) uv¼8þ12þ3 20 8¼ 5
1.8. Letu¼ ð5;4;1Þ, v¼ ð3; 4;1Þ,w¼ ð1; 2;3Þ. Which pair of vectors, if any, are perpendicular
(orthogonal)?
Find the dot product of each pair of vectors:
uv¼15 16þ1¼0; vw¼3þ8þ3¼14; uw¼5 8þ3¼0
Thus,uandvare orthogonal,uandware orthogonal, butvandware not.
1.9. Findk so that uandv are orthogonal, where: (a) u¼ ð1;k; 3Þandv¼ ð2; 5;4Þ,
(b) u¼ ð2;3k; 4;1;5Þandv¼ ð6; 1;3;7;2kÞ.
Computeuv, setuvequal to 0, and then solve fork:
(a) uv¼1ð2Þ þkð 5Þ 3ð4Þ ¼ 5k 10. Then 5k 10¼0, ork¼ 2.
(b) uv¼12 3k 12þ7þ10k¼7kþ7. Then 7kþ7¼0, ork¼ 1.
1.10. Findkuk, where: (a) u¼ ð3; 12; 4Þ, (b) u¼ ð2; 3;8; 7Þ.
First findkuk2¼uuby squaring the entries and adding. Thenkuk ¼ ffiffiffiffiffiffiffiffiffiffi kuk2 q
.
1.11. Recall thatnormalizing a nonzero vector v means finding the unique unit vector v^in the same
(c) Note thatwand any positive multiple ofwwill have the same normalized form. Hence, first multiplywby 12 to“clear fractions”—that is, first findw0¼12w¼ ð6;8; 3Þ. Then
i is nonnegative for eachi, and because the sum of nonnegative real numbers is nonnegative,
1.14. Prove Theorem 1.3 (Schwarz):juvj kukkvk. For any real numbert, and using Theorem 1.2, we have
0 ðtuþvÞ ðtuþvÞ ¼t2ðuuÞ þ2tðuvÞ þ ðvvÞ ¼ kuk2t2þ2ðuvÞtþ kvk2
Let a¼ kuk2, b¼2ðuvÞ,c¼ kvk2. Then, for every value of t, at2þbtþc0. This means that the
quadratic polynomial cannot have two real roots. This implies that the discriminantD¼b2 4ac0 or,
equivalently,b24ac. Thus,
4ðuvÞ24kuk2kvk2
Dividing by 4 gives us our result.
1.15. Prove Theorem 1.4 (Minkowski):kuþvk kuk þ kvk. By the Schwarz inequality and other properties of the dot product,
kuþvk2¼ ðuþvÞ ðuþvÞ ¼ ðuuÞ þ2ðuvÞ þ ðvvÞ kuk2þ2kukkvk þ kvk2¼ ðkuk þ kvkÞ2
Taking the square root of both sides yields the desired inequality.
Points, Lines, Hyperplanes in Rn
Here we distinguish between an n-tuple Pða1;a2;. . .;anÞ viewed as a point in Rn and an n-tuple
u¼ ½c1;c2;. . .;cn viewed as a vector (arrow) from the originOto the point Cðc1;c2;. . .;cnÞ.
1.16. Find the vectoruidentified with the directed line segment PQ!for the points:
(a) Pð1; 2;4ÞandQð6;1; 5ÞinR3, (b) Pð2;3; 6;5Þ andQð7;1;4; 8ÞinR4.
(a) u¼PQ!¼Q P¼ ½6 1; 1 ð 2Þ; 5 4 ¼ ½5;3; 9
(b) u¼PQ!¼Q P¼ ½7 2; 1 3; 4þ6; 8 5 ¼ ½5; 2;10; 13
1.17. Find an equation of the hyperplaneH inR4 that passes throughPð3; 4;1; 2Þand is normal to
u¼ ½2;5; 6; 3.
The coefficients of the unknowns of an equation ofHare the components of the normal vectoru. Thus, an equation ofH is of the form 2x1þ5x2 6x3 3x4¼k. SubstitutePinto this equation to obtaink¼ 26. Thus, an equation ofHis 2x1þ5x2 6x3 3x4¼ 26.
1.18. Find an equation of the planeH inR3 that contains Pð1; 3; 4Þand is parallel to the planeH0
determined by the equation 3x 6yþ5z¼2.
The planesH andH0are parallel if and only if their normal directions are parallel or antiparallel (opposite direction). Hence, an equation ofHis of the form 3x 6yþ5z¼k. SubstitutePinto this equation to obtain
k¼1. Then an equation ofHis 3x 6yþ5z¼1.
1.19. Find a parametric representation of the lineLinR4 passing throughPð4; 2;3;1Þin the direction ofu¼ ½2;5; 7;8.
HereLconsists of the pointsXðxiÞthat satisfy
X¼Pþtu or xi¼aitþbi or LðtÞ ¼ ðaitþbiÞ where the parameterttakes on all real values. Thus we obtain
1.20. LetC be the curveFðtÞ ¼ ðt2; 3t 2; t3; t2þ5ÞinR4, where 0t4.
(a) Find the pointPonC corresponding tot¼2. (b) Find the initial pointQand terminal point Q0 ofC. (c) Find the unit tangent vectorTto the curve C whent¼2.
(a) Substitutet¼2 intoFðtÞto getP¼fð2Þ ¼ ð4;4;8;9Þ.
(b) The parameter t ranges from t¼0 to t¼4. Hence, Q¼fð0Þ ¼ ð0; 2;0;5Þ and
Q0¼Fð4Þ ¼ ð16;10;64;21Þ.
(c) Take the derivative ofFðtÞ—that is, of each component ofFðtÞ—to obtain a vectorVthat is tangent to the curve:
VðtÞ ¼dFðtÞ
dt ¼ ½2t;3;3t
2;2t
Now find V when t¼2; that is, substitute t¼2 in the equation for VðtÞ to obtain
V¼Vð2Þ ¼ ½4;3;12;4. Then normalizeV to obtain the desired unit tangent vectorT. We have
kVk ¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi16þ9þ144þ16¼pffiffiffiffiffiffiffiffi185 and T¼ ffiffiffiffiffiffiffiffi4
185
p ; ffiffiffiffiffiffiffiffi3
185
p ; 12ffiffiffiffiffiffiffiffi
185
p ; ffiffiffiffiffiffiffiffi4
185
p
Spatial Vectors (Vectors in R3),ijk Notation, Cross Product
1.21. Letu¼2i 3jþ4k,v¼3iþj 2k,w¼iþ5jþ3k. Find:
(a) uþv, (b) 2u 3vþ4w, (c) uv anduw, (d) kukandkvk. Treat the coefficients ofi,j,kjust like the components of a vector in R3.
(a) Add corresponding coefficients to getuþv¼5i 2j 2k.
(b) First perform the scalar multiplication and then the vector addition:
2u 3vþ4w¼ ð4i 6jþ8kÞ þ ð 9i 3jþ6kÞ þ ð4iþ20jþ12kÞ ¼ iþ11jþ26k
(c) Multiply corresponding coefficients and then add:
uv¼6 3 8¼ 5 and uw¼2 15þ12¼ 1
(d) The norm is the square root of the sum of the squares of the coefficients:
kuk ¼p4ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiþ9þ16¼pffiffiffiffiffi29 and kvk ¼
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
9þ1þ4
p
¼pffiffiffiffiffi14
1.22. Find the (parametric) equation of the lineL:
(a) through the pointsPð1;3;2Þ andQð2;5; 6Þ;
(b) containing the point Pð1; 2;4Þ and perpendicular to the plane H given by the equation 3xþ5yþ7z¼15:
(a) First findv¼PQ!¼Q P¼ ½1;2; 8 ¼iþ2j 8k. Then
LðtÞ ¼ ðtþ1; 2tþ3; 8tþ2Þ ¼ ðtþ1Þiþ ð2tþ3Þjþ ð 8tþ2Þk
(b) BecauseLis perpendicular toH, the lineLis in the same direction as the normal vectorN¼3iþ5jþ7k
toH. Thus,
LðtÞ ¼ ð3tþ1; 5t 2; 7tþ4Þ ¼ ð3tþ1Þiþ ð5t 2Þjþ ð7tþ4Þk
1.23. LetSbe the surfacexy2þ2yz¼16 inR3.
(a) The formula for the normal vector to a surfaceFðx;y;zÞ ¼0 is
Nðx;y;zÞ ¼FxiþFyjþFzk
whereFx,Fy,Fz are the partial derivatives. UsingFðx;y;zÞ ¼xy2þ2yz 16, we obtain
Fx¼y2; Fy¼2xyþ2z; Fz¼2y
Thus,Nðx;y;zÞ ¼y2iþ ð2xyþ2zÞjþ2yk.
(b) The normal to the surfaceSat the pointPis
NðPÞ ¼Nð1;2;3Þ ¼4iþ10jþ4k
Hence,N¼2iþ5jþ2kis also normal toSatP. Thus an equation ofHhas the form 2xþ5yþ2z¼c. SubstitutePin this equation to obtainc¼18. Thus the tangent planeHtoSatPis 2xþ5yþ2z¼18.
1.24. Evaluate the following determinants and negative of determinants of order two:
(a) We have
u ðuvÞ ¼a1ða2b3 a3b2Þ þa2ða3b1 a1b3Þ þa3ða1b2 a2b1Þ ¼a1a2b3 a1a3b2þa2a3b1 a1a2b3þa1a3b2 a2a3b1¼0
Thus,uvis orthogonal tou. Similarly,uvis orthogonal tov.
(b) We have
kuvk2¼ ða2b3 a3b2Þ2þ ða3b1 a1b3Þ2þ ða1b2 a2b1Þ2 ð1Þ ðuuÞðvvÞ ðuvÞ2¼ ða2
1þa22þa23Þðb21þb22þb23Þ ða1b1þa2b2þa3b3Þ 2
ð2Þ
Expansion of the right-hand sides of (1) and (2) establishes the identity.
Complex Numbers, Vectors in Cn
1.29. Supposez¼5þ3i andw¼2 4i. Find: (a) zþw, (b) z w, (c) zw.
Use the ordinary rules of algebra together withi2¼ 1 to obtain a result in the standard formaþbi.
(a) zþw¼ ð5þ3iÞ þ ð2 4iÞ ¼7 i
(b) z w¼ ð5þ3iÞ ð2 4iÞ ¼5þ3i 2þ4i¼3þ7i
(c) zw¼ ð5þ3iÞð2 4iÞ ¼10 14i 12i2¼10 14iþ12¼22 14i
1.30. Simplify: (a) ð5þ3iÞð2 7iÞ, (b) ð4 3iÞ2, (c) ð1þ2iÞ3.
(a) ð5þ3iÞð2 7iÞ ¼10þ6i 35i 21i2¼31 29i
(b) ð4 3iÞ2¼16 24iþ9i2¼7 24i
(c) ð1þ2iÞ3¼1þ6iþ12i2þ8i3¼1þ6i 12 8i¼ 11 2i
1.31. Simplify: (a) i0;i3;i4, (b) i5;i6;i7;i8, (c) i39; i174,i252,i317:
(a) i0¼1, i3¼i2ðiÞ ¼ ð 1ÞðiÞ ¼ i; i4¼ ði2Þði2Þ ¼ ð 1Þð 1Þ ¼1
(b) i5¼ ði4ÞðiÞ ¼ ð1ÞðiÞ ¼i, i6¼ ði4Þði2Þ ¼ ð1Þði2Þ ¼i2¼ 1, i7¼i3¼ i, i8¼i4¼1
(c) Usingi4¼1 andin¼i4qþr¼ ði4Þqir¼1qir¼ir, divide the exponentnby 4 to obtain the remainderr:
i39¼i4ð9Þþ3¼ ði4Þ9i3¼19i3¼i3¼ i; i174¼i2¼ 1; i252¼i0¼1; i317¼i1¼i
1.32. Find the complex conjugate of each of the following:
(a) 6þ4i, 7 5i, 4þi, 3 i, (b) 6, 3, 4i, 9i.
(a) 6þ4i¼6 4i, 7 5i¼7þ5i, 4þi¼4 i, 3 i¼ 3þi
(b) 6¼6, 3¼ 3, 4i¼ 4i, 9i¼9i
(Note that the conjugate of a real number is the original number, but the conjugate of a pure imaginary number is the negative of the original number.)
1.33. Findzz andjzjwhenz¼3þ4i.
Forz¼aþbi, usezz¼a2þb2and z¼pffiffiffiffizz¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffia2þb2.
zz¼9þ16¼25; jzj ¼pffiffiffiffiffi25¼5
1.34. Simpify2 7i 5þ3i:
To simplify a fraction z=w of complex numbers, multiply both numerator and denominator by w, the conjugate of the denominator:
2 7i
5þ3i¼
ð2 7iÞð5 3iÞ ð5þ3iÞð5 3iÞ¼
11 41i
34 ¼
11 34
1.35. Prove: For any complex numbersz, w2C, (i)zþw¼zþw, (ii)zw¼zw, (iii)z¼z.
Supposez¼aþbiand w¼cþdiwherea;b;c;d2R. (i) zþw¼ ðaþbiÞ þ ðcþdiÞ ¼ ðaþcÞ þ ðbþdÞi
¼ ðaþcÞ ðbþdÞi ¼ aþc bi di ¼ ða biÞ þ ðc diÞ ¼ zþw
(ii) zw¼ ðaþbiÞðcþdiÞ ¼ ðac bdÞ þ ðadþbcÞi ¼ ðac bdÞ ðadþbcÞi ¼ ða biÞðc diÞ ¼ zw
(iii) z¼aþbi¼a bi¼a ð bÞi¼aþbi¼z
1.36. Prove: For any complex numbersz;w2C,jzwj ¼ jzjjwj.
By (ii) of Problem 1.35,
jzwj2 ¼ ðzwÞðzwÞ ¼ ðzwÞðzwÞ ¼ ðzzÞðwwÞ ¼ jzj2jwj2
The square root of both sides gives us the desired result.
1.37. Prove: For any complex numbersz;w2C,jzþwj jzj þ jwj.
Supposez¼aþbiand w¼cþdiwherea;b;c;d2R. Consider the vectorsu¼ ða;bÞ andv¼ ðc;dÞin R2. Note that
jzj ¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffia2þb2¼ kuk; jwj ¼pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffic2þd2¼ kvk
and
jzþwj ¼ jðaþcÞ þ ðbþdÞij ¼
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ðaþcÞ2þ ðbþdÞ2 q
¼ kðaþc;bþdÞk ¼ kuþvk
By Minkowski’s inequality (Problem 1.15),kuþvk kuk þ kvk, and so jzþwj ¼ kuþvk kuk þ kvk ¼ jzj þ jwj
1.38. Find the dot products uv and vu where: (a) u¼ ð1 2i; 3þiÞ, v ¼ ð4þ2i; 5 6iÞ;
(b) u¼ ð3 2i; 4i; 1þ6iÞ,v¼ ð5þi; 2 3i; 7þ2iÞ. Recall that conjugates of the second vector appear in the dot product
ðz1;. . .;znÞ ðw1;. . .;wnÞ ¼z1w1þ þznwn (a) uv¼ ð1 2iÞð4þ2iÞ þ ð3þiÞð5 6iÞ
¼ ð1 2iÞð4 2iÞ þ ð3þiÞð5þ6iÞ ¼ 10iþ9þ23i ¼ 9þ13i vu¼ ð4þ2iÞð1 2iÞ þ ð5 6iÞð3þiÞ
¼ ð4þ2iÞð1þ2iÞ þ ð5 6iÞð3 iÞ ¼ 10iþ9 23i ¼ 9 13i
(b) uv¼ ð3 2iÞð5þiÞ þ ð4iÞð2 3iÞ þ ð1þ6iÞð7þ2iÞ
¼ ð3 2iÞð5 iÞ þ ð4iÞð2þ3iÞ þ ð1þ6iÞð7 2iÞ ¼ 20þ35i vu¼ ð5þiÞð3 2iÞ þ ð2 3iÞð4iÞ þ ð7þ2iÞð1þ6iÞ
¼ ð5þiÞð3þ2iÞ þ ð2 3iÞð 4iÞ þ ð7þ2iÞð1 6iÞ ¼ 20 35i
In both cases,vu¼uv. This holds true in general, as seen in Problem 1.40.
1.39. Letu¼ ð7 2i; 2þ5iÞandv¼ ð1þi; 3 6iÞ. Find:
(a) uþv, (b) 2iu, (c) ð3 iÞv, (d) uv, (e) kukandkvk. (a) uþv¼ ð7 2iþ1þi; 2þ5i 3 6iÞ ¼ ð8 i; 1 iÞ
(b) 2iu¼ ð14i 4i2; 4iþ10i2Þ ¼ ð4þ14i; 10þ4iÞ