Professional Documents
Culture Documents
NEVER
Daniel Norman
LEARNING
Dan Wolczuk
Introduction to Linear
Algebra for Science and
Engineering
Student Cheap-ass Edition
Taken from:
Introduction to Linear Algebra for Science and Engineering, Second Edition
by Daniel Norman and Dan Wolczuk
Pearson Learning Solutions, 501 Boylston Street, Suite 900, Boston, MA 02116
A Pearson Education Company
www.pearsoned.com
PEARSON
Contents
A Note to Students
vi
A Note to Instructors
Independence
viii
91
Spanning Problems
Chapter 1
Bases of Subspaces
Subspaces
16
Chapter 3
18
of Matrices
31
Identity Matrix
40
Mappings
44
44
50
52
150
69
Solution Set of Ax
73
150
b 152
Rowspace of A
75
153
156
76
77
78
Matrix
167
83
Linear Systems
85
165
157
162
Rank of a Matrix
143
143
63
Rank T heorem
of Linear Equations
136
139
63
A Word Problem
134
Linear Mappings
Solutions
Linear Mappings
54
131
Elimination
131
Matrix Mappings
50
and Distances
127
43
Cross-Products
121
126
Block Multiplication
40
Minimum Distance
120
Chapter 2
115
34
115
28
115
28
107
and Inverses
24
102
105
Linear Programming
15
102
Equations
95
97
91
168
86
iii
170
iv
Contents
291
Matrix
187
185
6.2 Diagonalization
Chapter 4
193
197
Chapter 7
209
211
213
218
235
323
325
240
246
329
7.2 Projections and the Gram-Schmidt
Procedure 333
Projections onto a Subspace
Chapter 5
Determinants
255
Chapter 6
337
342
345
348
348
354
b
J f(x)g(x) dx
289
354
Fourier Series
Chapter 8
355
Quadratic Forms
363
Quadratic Forms
Overdetermined Systems
289
333
Eigenvectors and
Diagonalization
321
Orthonormal Bases
317
General Discussion
206
317
206
312
Eigenvalues
197
201
Subspaces
312
193
Polynomials
303
193
Vector Spaces
291
299
366
372
372
376
Contents
Chapter 9
388
388
390
395
395
397
Polar Form
398
399
402
404
407
408
2 Matrix
420
3 Matrix
422
413
Inequalities
415
426
429
Answers to Mid-Section
439
Answers to Practice Problems
425
426
Appendix B
411
The Case of a 3
399
The Case of a 2
418
395
417
Index
529
465
A Note to Students
Linear Algebra-What Is It?
Linear algebra is essentially the study of vectors, matrices, and linear mappings. Al
though many pieces of linear algebra have been studied for many centuries, it did not
take its current form until the mid-twentieth century. It is now an extremely important
topic in mathematics because of its application to many different areas.
Most people who have learned linear algebra and calculus believe that the ideas
of elementary calculus (such as limit and integral) are more difficult than those of in
troductory linear algebra and that most problems in calculus courses are harder than
those in linear algebra courses. So, at least by this comparison, linear algebra is not
hard. Still, some students find learning linear algebra difficult. I think two factors con
tribute to the difficulty students have.
First, students do not see what linear algebra is good for. This is why it is important
to read the applications in the text; even if you do not understand them completely, they
will give you some sense of where linear algebra fits into the broader picture.
Second, some students mistakenly see mathematics as a collection of recipes for
solving standard problems and are uncomfortable with the fact that linear algebra is
"abstract" and includes a lot of "theory." There will be no long-term payoff in simply
memorizing these recipes, however; computers carry them out far faster and more ac
curately than any human can. That being said, practising the procedures on specific
examples is often an important step toward much more important goals: understand
ing the concepts used in linear algebra to formulate and solve problems and learning to
interpret the results of calculations. Such understanding requires us to come to terms
with some theory. In this text, many of our examples will be small. However, as you
work through these examples, keep in mind that when you apply these ideas later, you
may very well have a million variables and a million equations. For instance, Google's
PageRank system uses a matrix that has 25 billion columns and 25 billion rows; you
don't want to do that by hand! When you are solving computational problems, al
ways try to observe how your work relates to the theory you have learned.
Mathematics is useful in so many areas because it is abstract: the same good idea
can unlock the problems of control engineers, civil engineers, physicists, social scien
tists, and mathematicians only because the idea has been abstracted from a particular
setting. One technique solves many problems only because someone has established a
theory of how to deal with these kinds of problems. We use definitions to try to capture
important ideas, and we use theorems to summarize useful general facts about the kind
of problems we are studying. Proofs not only show us that a statement is true; they can
help us understand the statement, give us practice using important ideas, and make it
easier to learn a given subject. In particular, proofs show us how ideas are tied together
so we do not have to memorize too many disconnected facts.
Many of the concepts introduced in linear algebra are natural and easy, but some
may seem unnatural and "technical" to beginners. Do not avoid these apparently more
difficult ideas; use examples and theorems to see how these ideas are an essential
part of the story of linear algebra. By learning the "vocabulary" and "grammar" of
linear algebra, you will be equipping yourself with concepts and techniques that math
ematicians, engineers, and scientists find invaluable for tackling an extraordinarily rich
variety of problems.
vi
A Note to Students
vii
Engineers
Suppose you become a control engineer and have to design or upgrade an automatic
control system. T he system may be controlling a manufacturing process or perhaps
an airplane landing system. You will probably start with a linear model of the sys
tem, requiring linear algebra for its solution. To include feedback control, your system
must take account of many measurements (for the example of the airplane, position,
velocity, pitch, etc.), and it will have to assess this information very rapidly in order to
determine the correct control responses. A standard part of such a control system is a
Kalman-Bucy filter, which is not so much a piece of hardware as a piece of mathemat
ical machinery for doing the required calculations. Linear algebra is an essential part
of the Kalman-Bucy filter.
If you become a structural engineer or a mechanical engineer, you may be con
cerned with the problem of vibrations in structures or machinery. To understand the
problem, you will have to know about eigenvalues and eigenvectors and how they de
termine the normal modes of oscillation. Eigenvalues and eigenvectors are some of the
central topics in linear algebra.
An electrical engineer will need linear algebra to analyze circuits and systems; a
civil engineer will need linear algebra to determine internal forces in static structures
and to understand principal axes of strain.
In addition to these fairly specific uses, engineers will also find that they need
to know linear algebra to understand systems of differential equations and some as
pects of the calculus of functions of two or more variables. Moreover, the ideas and
techniques of linear algebra are central to numerical techniques for solving problems
of heat and fluid flow, which are major concerns in mechanical engineering. And the
ideas of Jjnear algebra underjje advanced techniques such as Laplace transforms and
Fourier analysis.
Physicists
Linear algebra is important in physics, partly for the reasons described above. In addi
tion, it is essential in applications such as the inertia tensor in general rotating motion.
Linear algebra is an absolutely essential tool in quantum physics (where, for exam
ple, energy levels may be determined as eigenvalues of linear operators) and relativity
(where understanding change of coordinates is one of the central issues).
viii
A Note to Instructors
Managers
A manager in industry will have to make decisions about the best allocation of re
sources: enormous amounts of computer time around the world are devoted to linear
programming algorithms that solve such allocation problems. The same sorts of tech
niques used in these algorithms play a role in some areas of mine management. Linear
algebra is essential here as well.
So who needs linear algebra? Almost every mathematician, engineer, or scientist
will .find linear algebra an important and useful tool.
1.4, 1.5, 2.4, 5.4, 6.3, 6.4, 7.3, 7.5, 8.3, 8.4, and 9.2. You will get to see many more
applications of linear algebra in your future courses.
A Note to Instructors
Welcome to the second edition of Introduction to Linear Algebra for Science and
Engineering. It has been a pleasure to revise Daniel Norman's first edition for a new
generation of students and teachers. Over the past several years, I have read many
articles and spoken to many colleagues and students about the difficulties faced by
teachers and learners of linear algebra. In particular, it is well known that students typ
ically find the computational problems easy but have great difficulty in understanding
the abstract concepts and the theory. Inspired by this research, I developed a pedagog
ical approach that addresses the most common problems encountered when teaching
and learning linear algebra. I hope that you will find this approach to teaching linear
algebra as successful as I have.
Vectors in JR.11 are now always represented as column vectors and are denoted
with the normal vector symbol 1. Vectors in general vector spaces are still
denoted in boldface.
Some material has been reorganized to allow students to see important con
cepts early and often, while also giving greater flexibility to instructors. For
example, the concepts of linear independence, spanning, and bases are now
introduced in Chapter 1 in JR.11, and students use these concepts in Chapters 2
and 3 so that they are very comfortable with them before being taught general
vector spaces.
A Note to Instructors
ix
The material on complex numbers has been collected and placed in Chapter 9,
at the end of the text. However, if one desires, it can be distributed throughout
the text appropriately.
It prevents students from mistaking linear algebra as very easy and very com
putational early in the course and then becoming overwhelmed by abstract con
cepts and theories later.
One example of this approach is our treatment of the concepts of spanning and
linear independence. They are both introduced in Section 1.2 in JR.n, where they can be
motivated in a geometrical context. They are then used again for matrices in Section
3.1 and polynomials in Section 4.1, before they are finally extended to general vector
spaces in Section 4.2.
The following are some other features of the text's organization:
Topics are ordered to give students a chance to work with concepts in a simpler
setting before using them in a much more involved or abstract setting. For ex
ample, before reaching the definition of a vector space in Section 4.2, students
will have seen the 10 vector space axioms and the concepts of linear indepen
dence and spanning for three different vector spaces, and they will have had
some experience in working with bases and dimensions. Thus, instead of be
ing bombarded with new concepts at the introduction of general vector spaces,
students will j ust be generalizing concepts with which they are already familiar.
A Note to Instructors
Pedagogical Features
Since mathematics is best learned by doing, the following pedagogical elements are
included in the book.
Practice problems are provided for students at the end of each section. See "A
Note on the Exercises and Problems" below.
Examples, theorems, and definitions are called out in the margins for easy
reference.
Applications
One of the difficulties in any linear algebra course is that the applications of linear
algebra are not so immediate or so intuitively appealing as those of elementary cal
culus. Most convincing applications of linear algebra require a fairly lengthy buildup
of background that would be inappropriate in a linear algebra text. However, without
some of these applications, many students would find it difficult to remain motivated
to learn linear algebra. An additional difficulty is that the applications of linear alge
bra are so varied that there is very little agreement on which applications should be
covered.
In this text we briefly discuss a few applications to give students some easy sam
ples. Additional applications are provided on the Corripanion Website so that instruc
tors who wish to cover some of them can pick and choose at their leisure without
increasing the size (and hence the cost) of the book.
List of Applications
A Note to Instructors
xi
Computers
As explained in "A Note on the Exercises and Problems," which follows, some prob
lems in the book require access to appropriate computer software. Students should
realize that the theory of linear algebra does not apply only to matrices of small size
with integer entries. However, since there are many ideas to be learned in linear alge
bra, numerical methods are not discussed. Some numerical issues, such as accuracy
and efficiency, are addressed in notes and problems.
xii
A Note to Instructors
JR is covered in Chapters 1, 2, and 3.) Moreover, there is a natural flow from matrix
inverses and elementary matrices at the end of Chapter 3 to determinants in Chapter 5.
A Course Outline
The following table indicates the sections in each chapter that we consider to be "cen
tral material":
Chapter
Central Material
Optional Material
l , 2,3,4,5
1, 2, 3
1,2,3,4,5,6
l,2,3,4,5,6,7
1,2,3
1,2
3,4
1,2
3,4, 5
1,2
3,4
l , 2,3,4,5,6
Supplements
We are pleased to offer a variety of excellent supplements to students and instructors
using the Second Edition.
T he new Student Solutions Manual (ISBN: 978-0-321-80762-5), prepared by
the author of the second edition, contains full solutions to the Practice Problems and
Chapter Quizzes. It is available to students at low cost.
MyMathLab Online Course (access code required) delivers proven results
in helping individual students succeed. It provides engaging experiences that person
alize, stimulate, and measure learning for each student. And, it comes from a trusted
partner with educational expertise and an eye on the future. To learn more about how
A Note to Instructors
xiii
A Test Bank with a large selection of questions for every chapter of the text.
Finally,
the
second
edition
is
available
as
CourseSmart
eTextbook
Acknowledgments
T hanks are expressed to:
Agnieszka Wolczuk: for her support, encouragement, help with editing, and tasty
snacks.
Mike La Croix: for all of the amazing figures in the text and for his assistance on
editing, formatting, and LaTeX'ing.
Stephen New, Martin Pei, Barbara Csima, Emilio Paredes: for proofreading and
their many valuable comments and suggestions.
Conrad Hewitt, Robert Andre, Uldis Celmins, C. T. Ng, and many other of my
colleagues who have taught me things about linear algebra and how to teach
it as well as providing many helpful suggestions for the text.
To all of the reviewers of the text, whose comments, corrections, and recommen
dations have resulted in many positive improvements:
xiv
A Note to Instructors
Robert Andre
University of Waterloo
Luigi Bilotto
Vanier College
Dietrich Burbulla
University of Toronto
Dr. Alistair Carr
Monash University
Gerald Cliff
University of Alberta
Antoine Khalil
Manuele Santoprete
Wilfrid Laurier University
Alistair Savage
University of Ottawa
Denis Sevee
John Abbott College
Mark Solomonovich
Grant MacEwan University
CEGEP Vanier
Hadi Kharaghani
University of Lethbridge
Gregory Lewis
University of Ontario
Institute of Technology
Eduardo Martinez-Pedroza
McMaster University
Dorette Pronk
Dalhousie University
Brian Wetton
University of British Columbia
Dan Wolczuk
University of Waterloo
CHAPTER 1
"
Vectors in IR and JR?.3
1.2
11
Vectors in IR
1.3
1.4
1.5
Some of the material in this chapter will be familiar to many students, but some ideas
that are introduced here will be new to most. In this chapter we will look at operations
on and important concepts related to vectors. We will also look at some applications
of vectors in the familiar setting of Euclidean space. Most of these concepts will later
be extended to more general settings. A firm understanding of the material from this
chapter will help greatly in understanding the topics in the rest of this book.
P2
---
--
-- --
--
..
p =(pi, p2)
I
I
I
I
I
I
I
I
0
Figure 1.1.1
Pi
Coordinates in the plane
Chapter 1
Definition
JR2
[:l
where
and
xi x2
Remark
We shall use the notation 1 =
[ : ]
to denote vectors in
JR2
P(p , p ).
(pi, p2),i 2
(pi, p2)
JR2.
[]
as shown in Figure 1.1.2. Note, however, that the points between (0, 0)
and
vector as an arrow is particularly common in physics; force and acceleration are vector
quantities that can conveniently be represented by an arrow of suitable magnitude and
direction.
0 = (0, 0)
Figure 1.1.2
Definition
Addition and Scalar
Multiplication in :12
If 1
[: l [l t JR,
y =
and
X +y =
and the scalar multiplication of a vector by a factor oft, called a scalar, is defined by
tx =
t [Xzxi]= [txtxi2]
0
Figure 1.1.3
EXAMPLE I
Let x
[-]
and y
[n
(3, 4)
Then
(-2, 3)
X1
(1.S)J
X1
(-l)J
Figure 1.1.4
Chapter 1
EXAMPLE2
Let a=
[n [ ]
-
v=
.and w=
Solution: We get
a+v=
[-l
[ i ] [ ] [!]
[ ] [ ]
[-] [ ] [-] [] [-]
3w=3
Let a=
[ l [ ]
v=
.and w =
= -
2v - w = 2
EXERCISE 1
+ < -1)
rn
a sketch.
(a) a+ w
(c) (a+ w) - v
(b) -v
The vectors e1 =
[]
and e =
[]
will call the set {e1, e } the standard basis for IR.2. (We shall discuss the concept of
a basis fmther in Section 1.2.) The basis vectors e1 and e are important because any
vector v=
way:
[ ]
Remark
In physics and engineering, it is common to use the notation i
instead.
[]
and j =
[]
We will use the phrase linear combination to mean "sum of scalar multiples."
So, we have shown above that any vector x E IR.2 can be written as a unique linear
combination of the standard basis vectors.
One other vector in IR.2 deserves special mention: the zero vector,
0=
[ ]
.Some
important properties of the zero vector, which are easy to verify, are that for any
xEJR.2,
(1) 0 +x x
(2) x + c-1)x = o
(3) Ox= 0
=
Section 1.1
JR.2:
J creates
a line
{tJitEJR.}
Often we do not use formal set notation but simply write the vector equation of the
line:
rJ,
tEJR.
The vector
ff+ rJ.
tEJR.
direction vector of one line is a non-zero multiple of the direction vector of the other
line.
X2
.
line x =
rJ+ ff
Figure 1.1.5
EXAMPLE3
td + p.
A vector equation of the line through the point P(2, -3) with direction vector
[ ]
-
is
Chapter 1
EXAMPLE4
Write the vector equation of a line through P(l, 2) parallel to the line with vector
equation
x= t
[]
tEIR
Solution: Since they are parallel, we can choose the same direction vector. Hence, the
vector equation of the line is
EXERCISE 2
Write the vector equation of a line through P(O, 0) parallel to the line with vector
equation
x = jJ
tJ becomes
X1 = Pl
td 1
X2 = P2
td2,
t EIR
This is referred to as the parametric equation of the line. The familiar scalar form
of the equation of the line is obtained by eliminating the parameter t. Provided that
di* 0, d1 * 0,
X2 - P2
X1 - Pl
-r- --di
d1
- -
---
or
x2 = P2
d1
+
di
(xi - Pi)
EXAMPLES
Write the vector, parametric, and scalar equations of the line passing through the point
r-].
[!] r-l
+
XI = 3 - St
X2 = 4 + t,
- !Cx1
3).
tER
t ER
P to
point Q by
PQ.
P and
pointing
towards Q. We shall identify directed line segments from the origin 0 with the corre
sponding vectors; we write OP = fJ, OQ =if, and so on. A directed line segment that
starts at the origin is called the position vector of the point.
For many problems, we are interested only in the direction and length of the
directed line segment; we are not interested in the point where it is located. For
example, in Figure 1.1.3, we may wish to treat the line segment QR as if it were the
same as OP. Taking our cue from this example, for arbitrary points P, Q, R in JR.2, we
p
0
Figure 1.1.6
r-if =
fJ and t
r-if=i'-s
EXAMPLE6
-r - if=
[ ] - [] [ -] = [] - [ -!] =
_
S(-2,4)
r s
-
Chapter 1
PQ =
RS.
Remark
Writing
QR
ST
QR
= = [-l
ST
ST.
QR
is not really
classes" and more careful notation with directed line segments is not helpful at this
stage. By introducing directed line segments, we are encouraged to think about vectors
that are located at arbitrary points in space. This is helpful in solving some geometrical
problems, as we shall see below.
EXAMPLE 7
P(l,
1, 2) is
Observe in the example above that we would have the same line if we started at the
second point and "moved" toward the first point--0r even if we took a direction vector
in the opposite direction. Thus, the same line is described by the vector equations
x=[_iJ+r[-J. rE
x=[_iJ+s[_iJ sE
x=[;]+t[-], tE
In fact, there are infinitely many descriptions of a line: we may choose any point on
the line, and we may use any non-zero multiple of the direction vector.
EXERCISE 3
P(l,
Section 1.1
Figure 1.1.7
It should be noted that we are adopting the convention that the coordinate axes
form a right-handed system. One way to visualize a right-handed system is to spread
out the thumb, index finger, and middle finger of your right hand. The thumb is
the x1 -axis, the index finger is the x2-axis, and the middle finger is the x3-axis. See
Figure 1.1.8.
Figure 1.1.8
2
We now define JR.3 to be the three-dimensional analog of JR. .
Definition
:l3
Mathematically, we write
[:l
10
Chapter 1
Definition
If 1
[:n n
jl
Multiplication in J.3
[ l l l
x+y
X2 + Y2
X3
Xt + Yt
Xt
X2 + Y2
X3 + Y3
[l l
[
tx
Xl
t x2
X3
tX1
tx2
tX 3
Addition still follows the parallelogram rule. It may help you to visualize this
Figure 1.1.9
EXAMPLES
Let u
[ _i]. l-n
jl
Solution: We have
V +U
=
-w
-V + 2W-"
and w
=
[H
2W
u.
nHJ ni
-[] {]
l
- l 11+2 ll-l J l =r m l :1 r-l
=
11
It is useful to introduce the standard basis for JR.3 just as we did for JR.2. Define
[:]
Remark
In physics and engineering, it is common to use the notation i
e1, j
e1,
and k
e3
instead.
[]
Directed line segments are the same in three-dimensional space as in the two
dimensional case.
A line through the point P in JR.3 (corresponding to a vector {J) with direction
vector
p + tJ,
t E JR
It is important to realize that a line in JR.3 cannot be described by a single scalar linear
equation, as in JR.2. We shall see in Section 1.3 that such an equation describes a plane
in JR.3 .
EXAMPLE9
Find a vector equation of the line that passes through the points P(l, 5, -2) and
Q(4,-1,3).
if
- [-H
p
line is
EXERCISE4
-2
+St.
+ 3t, x2
6t,
and
Find a vector equation of the line that passes through the points P(l, 2, 2) and
Q(l,-2,3).
12
Chapter 1
PROBLEMS 1.1
Practice Problems
Al
[!]-[J
Ul
Hl
rn
Ut V = and W= Detenillne
(a) 2v- 3w
Cb) -3Cv +2w) +5v
(c) a such that w- 2a = 3v
(d) a such that a - 3v = 2a
AS Ut V =
and W = = Detennine
(a) v + !w
Cb) 2c v + w)- c2v - 3w)
(c) a such that w- a = 2V
(d) a such that !a + v = w
A4
A6
[-]
(d)
[
r1
Write a vector equation for the line that passes
(c) P(2,0,5),J=
-11
P(4,1,5),J =
AS
13
Homework Problems
B 1 Compute each of the following linear combinations
[-] + r-]
-3 [-]
(b)
(d)
[-]- []
-3 []- [;]
(b)
-t
(a)[]+ [=]
[]- []
2 [ =J
H [1 ] - ?a[]
[ + Y3 [- 12] - [ ]
tions.
(b)
(c)
(d
(e)
P(l,4,1), Q(4,3,-1),
R(-1,4,2), and S(8,6,-5). Determine PQ,
PR, PS, QR, and SR, and verify that PQ+QR=
PR= PS +SR.
Consider the points P(3,-2,1), Q(2, 7, -3),
R(3,1,5), and S(-2,4,-1). Determine PQ,
;1...,.
-+
-+
PK,
PS,
QR,
and SR, and verify that P Q+QR=
PR= PS +SR.
P(-3, 4),J=
(b)
P(O, 0). J=
(c)
P(2,3,-1), J=
(d)
P(3,1,2),J=
(e)
(f) (1 +)
B4 Ut V
(a)
(b)
(c)
(d)
(b)
(c)
(d)
0
-i
and W
Detecm;ne
2v- 3w
-2(v- w) - 3w
i1 such that w - 2i1 3v
i1 such that 2i1 3w= v
BS Ut V
(a)
1-
[ ]
-
and W
3v- 2w
-iv+ w
i1 such that v+i1= v
i1 such that 2i1 - w= 2v
[-]
[-]
P(3,1), Q(l,2)
P(l,-2,1), Q(O, 0, 0)
P(2,-6,3), Q(-1,5,2)
P(l,-1,i), Q(i, t. 1)
JR2, determine
x2= -2x1 3
Xi + 2X2= 3
AlO
(a)
to answer this.)
whether the points P(2,1,1),
Q(l,2, 3), and R(4,-1,-3) are collinear. Show
-H
[
[-]
(a)
tions.
(c)
-+
<{]-[-]
[=l
f;l Hl
[ 4J [-1
{ l {n
-+
(a) Determine
(b) Determine
14
Chapter 1
Computer Problems
[=u
[=m
Cl Let V,
v,
[ :1
-36
V2
and
171!1
+ 6691/4
Conceptual Problems
[ ]
[ l
[]
Dl Let i1
and w
[ ].
+t2 w
+t2w
PQ+ QR+RP
D3 Let
D2 Let
[ :2).
2
vectors in JR. . Prove that x
2
t E JR, is a line in JR. passing through the
and
J t= 0 be
D4 Let x and
vectors
fl
fl + t J,
+ t2i12
fl
is a scalar multiple of
Prove that
t(x
+ y)
tx
JR be a scalar.
+ t)!
the motion of a particle, an engineer needs to specify its position (3 variables) and its
velocity (3 more variables); the engineer therefore has 6 variables. A scientist working
in string theory works with 11 dimensional space-time variables. An economist seek
ing to model the Canadian economy uses many variables: one standard model has
more than 1500 variables. Of course, calculations in such huge models are carried
out by computer. Even so, understanding the ideas of geometry and linear algebra is
necessary to decide which calculations are required and what the results mean.
JR.1
15
Definition
If 1
Xi
Xn
y=
r1
1
[Xl
Xn
:
, where x; E R Mathematically,
x+.Y=
: + : =
:
+
tx = t
Theorem 1
1Xn1] ttX11]
=
(1) x+ y E JR.11
(closed under addition)
(2) x+ y = y + 1
(addition is commutative)
(3) c1 + y) + w 1 +CY+ w)
(addition is associative)
(4) There exists a vector 0 E
such that z+0 = z for all z E JR.ll
=
-+
JR.1
-+
-+
-+
-+
(zero vector)
(additive inverses)
(6) t1 E JR.11
(closed under scalar multiplication)
(7) s(t1) = (st)1
(scalar multiplication is associative)
(8) (s + t)x = s1 + tx
(a distributive law)
(9) t(1 + y) = t1 + tY
(another distributive law)
(10) lx = 1
Proof: We will prove properties (1) and (2) from Theorem 1 and leave the other proofs
to the reader.
,-
16
Chapter 1
X +y
[ Ytl
:
Xn
E !Rn
Yn
[ ] ti ]
+X1
XI +y1
x+st=
X11 +y,,
EXERCISE 1
11
=st+x
Xn
Observe that properties (2), (3), (7), (8), (9), and ( 10) from Theorem 1 refer only to
the operations of addition and scalar multiplication, while the other properties, ( 1), ( 4) ,
(5), and (6), are about the relationship between the operations and the set IR11 These
facts should be clear in the proof of Theorem 1. Moreover, we see that the zero vector
0
of JR11 is the vector
0
zero vector satisfies the same properties as the zero vector in JR2 and JR3.
Students often find properties (1) and (6) a little strange. At first glance, it seems
obvious that the sum of two vectors in !R11 or the scalar multiple of a vector in JR11 is
another vector in IR". However, these properties are in fact extremely important. We
now look at subsets of IR11 that have both of these properties.
Subspaces
Definition
A non-empty subset S of IR11 is called a subspace of IR11 if for all vectors x, y E S and
Subspace
t E IR:
0, then every
n
subspace of !R contains the zero vector. This fact provides an easy method for dis
=
qualifying any subsets that do not contain the zero vector as subspaces. For example,
a line in IR3 cannot be a subspace if it does not pass through the origin. Thus, when
_,
of IR11; this is called the trivial subspace. Additionally, IR" is a subspace of itself. We
will see throughout the text that other subspaces arise naturally in linear algebra.
Section 1.2
EXAMPLE 1
17
Vectors inJR"
0,
5!
0,
0.
0.
t X3
So, S is closed under scalar multiplication. Therefore, S is a subspace ofJR3.
EXAMPLE2
EXERCISE2
1(1)
0.
Show that S ={[] I 2x1 = x2} is a subspace ofJR2 and T = {[] I x1 +x2 = 2} is
not a subspace of R2.
Chapter 1
18
Theorem 2
If {v1,
, vk} is a set of vectors in JRn and S is the set of all possible linear combi
Definition
If S is the subspace of JR.11 consisting of all linear combinations of the vectors v1,
, vk
Span
E JR.11, then S is called the subspace spanned by the set of vectors 13 = {v1, ... , vk}, and
Spanning Set
we say that the set 13 spans S. The set 13 is called a spanning set for the subspace S.
We denote S by
S = Span{i11, ... , vk} =Span 13
EXAMPLE3
2
Let v E JR. with v *
0 and consider the line L with vector equation x = tV, t E JR. Then
L is the subspace spanned by {V}, and {V} is a spanning set for L. We write L =Span{V}.
2
Similarly, for v1, v2 E JR. , the set M with vector equation x = ti\11 + t2v2 is a
2
subspace oflR with spanning set {v1, v2}. That is, M =Span{i11, v2}.
If v E JR.
with v *
JR. that passes through the origin. However, we see that the geometrical interpretation
of Span{v1, v2} depends on the choices of v1 and v2. We demonstrate this with some
examples.
Section 1.2
EXAMPLE4
{[] []}
The set
S1
Hence,
S1
is a line in
The set
where
S2
S3
S3
2
JR. .
{ [] [-]}
,
Span
{[] []}
That is,
S3
2
{V1, v2} is a spanning set for JR. if and only
EXAMPLES
The set
Hence,
S1
"Span
19
2
JR. that passes
= Span
t1 - 2t2
The set
Hence,
Span
Vectors in JR.11
{[ =l m, [] }
20
Chapter 1
EXAMPLES
(continued)
{[ jl [ _; l [!]}
{] {i] [!]
+
t
3
[ jl [ _l [!] l!l
{[ 3] [!]}
t
= ,
So, S =Span
2
!
2
= (tI
1,,1 ,! ER
2 3
1 )
2
[ l
_
(1
2
1 )
3
[! ]
Theorem 3
Proof: We are assuming that there exists t1,...,tk-l E IR. such that
= s1v1
=
(s1
Skt1W1
Sk-1Vk-l
+
sk(t1v1
(sk-1
tk-1Vk-1)
sktk-1Wk-I
as required.
s;;
Span{\11,...,vk-d Clearly,
In fact, any vector which can be written as a linear combination of the other vectors
in the set can be removed without changing the spanned set. It is important in linear
algebra to identify when a spanning set can be simplified by removing a vector that can
be written as a linear combination of the other vectors. We will call such sets linearly
dependent. If a spanning set is as simple as possible, then we will call the set linearly
independent. To identify whether a set is linearly dependent or linearly independent,
we require a mathematical definition.
Section 1.2
Vectors in IR"
21
v;, is equal to a linear combination of some (or all) of the other vectors. Hence, we can
find scalars t1, ... tk E IR such that
has a solution where at least one of the coefficients is non-zero. On the other hand, if
the set is linearly independent, then the only solution to this equation must be when all
the coefficients are 0. For example, if any coefficient is non-zero, say t; f. 0, then we
can write
. .
Theorem 3.
We make this our mathematical definition.
Definition
A set of vectors {v1, ... , v k} is said to be linearly dependent if there exist coefficients
Linearly Dependent
t1,
Linearly Independent
is t1
Theorem 4
t2
tk
If a set of vectors {V1, ... , v k} contains the zero vector, then it is linearly dependent.
Proof: Assume V;
0. Then we have
t1v1 +
EXAMPLE6
show that the set
Solution: We consider
is lineady dependent
22
Chapter 1
EXAMPLE6
(continued)
Since vectors are equal only if their corresponding entries are equal, this gives us three
equations in three unknowns
7t1 - lOt2 - t3
-14t, + 15t2
6t,
15
14
t2
3t3
Solving using substitution and elimination, we find that there are in fact infinitely many
possible solutions. One is
EXERCISE 3
Determine whether
t1
, t2
[
m
ml m
,
, t3
Remark
Observe that determining whether a set
0.
t1v1
tkvk
However, this equation actually represents n equations (one for each entry of the
t1,
Definition
If
Basis
independent, then
{v,,...,vk}
EXAMPLE 7
Let
{V1,
[-H r \l [-H
=
v,
"
"'
S =Span {V1, i12, v3}. Find a basis for S.
Solution: Observe that
{V1,... ,vk}
is linearly
JR3
given by
EXAMPLE 7
(continued)
23
Theorem 3,
is t1
t2
0 since neither v 1 nor i12 is a scalar multiple of the other. Hence, {V1, i12} is
linearly independent.
Therefore, {v1, v2} is linearly independent and spans S and so it is a basis for S.
Bases (the plural of basis) will be extremely important throughout the remainder
of the book. At this point, however, we just define the following very important basis.
Definition
In !fll.1 , let e; represent the vector whose i-th component is 1 and all other components
Standard Basis
are 0. The set {e1' ... ' e,,} is called the standard basis for
for'.?."
!fli.11
Observe that this definition matches that of the standard basis for !fli. and !fll. given
in Section 1.1.
EXAMPLE 8
{ [] [!] [m
is 11
12
13
[:l
[:l [] [!] m
=
X1
X2
X3
Remark
Compare the result of Example 8 with the meaning of point notation P(a, b, c). When
we say P(a, b, c) we mean the point P having a amount in the x-direction, b amount
in they-direction, and c amount in the z-direction. So, observe that the standard basis
vectors represent our usual coordinate axes.
24
Chapter 1
EXERCISE4
JR5.
Definition
Let
Linein3i."
a line in
Definition
Let
Planein J."
vector equation x
jJ, v
Hyperplane in R"
JR" with
jJ + t1v, t1
JR
V1,V2,JJ
through
Definition
JR11
JR
JR11,
with
jJ
jJ.
JR11, with {V1, . ., v11_ i} being linearly independent. Then the set
11
with vector equation x
jJ + t1v1 +
+ t11_1v11_1, t; E JR is called a hyperplane in JR
that passes through jJ.
Let
v 1, ... , v11_ 1, jJ
EXAMPLE9
The set Span
{ l , I , : }
is linearly independent in
is a hyperplane since
-1
-2
EXAMPLE 10
JR4.
{[l r-ll rm
l : ' :}
1
-2
-1
[ l r-l l l H
+
Hence,
25
EXAMPLE 10
(continued )
{ll [- ]}
3
passing through the origin in IR. .
PROBLEMS 1.2
Practice Problems
Al Compute each of the following linear combina
tions.
( a)
(b )
2
-1
(el X
+2
-1
-1
3
1
-1
-3
+2
1
4
5
0
2
-2
{.x
(c)
(d)
-1
( b)
(c)
{[]I
(d )
{[::J
X
t
X3
=
X +X2
t
XtX2
x3
X3
11
(e)
(f )
{x
{x
{x
{x
0.
E JR.41Xt+2x3
E IR.4 I X1
E IR.4 I 2x1
3x4,X2 - Sx3
=
5,X[ -3X4
X3X4,X2 -X4
E JR.4 X[ +X2
1
-X4,X3
o
=
I xi - xl
{!}
2
3
+ 12
+1
1
{[:]
{[:J
[] [;] []
(D Span
1
2
-2
2
1 +2
1
(c) 2
0
2
-1
(a)
(b)
(c)
(d )
mrnl .[_:]}
{l-rrnrnJ}
{[i][l][]}
{[n.rJ.rn
Chapter 1
26
(b) Span
{r , r}
{ , ! }
_!
,
,
}
{
,
(c) Span
c d) Span
H' j'n
fl + tJ, t E JR is a
A6 Let fl, J E JR12 Prove that 1
12
JR
if and only if fl is a scalar multiple
subspace of
of J
=
Homework Problems
Bl Compute each of the following linear combinations.
(a)
(b)
(c)
1
-1
-2
-2
3
2
2
1
-2
3 +2 2
2
3
0
2
1
+
-1
-7
-2
3 0 -2 3 + 6
-4
0
-7
2
3
0
Cb)
{[]I
{[i]
X1
+x2
1 x, +x,
(c)
(d)
1
1
mirm
-1
CD Span
Ce)
I
I
4
{1E1R I
4
+
0,3x3 -x4}
X1 + X2 + X4
{1E JR I X1 + 2x3 0, Xt - 3x4 o}
=
un
CD Span
mirm
Cb)
WHW
{[-il Ul {:]}
mrnl' [i]}
{[][].[-;]}
Cc)
(d)
Section 1.2
Exercises
27
(a) Span
{ \ : }
pn
-
-2
(b) SpM
Computer Problems
1.23
Cl Let vi
4.21
4.16 .-t
, v2
_2_21
:t
-3.14
0
0.34
-9.6
1.01
' v3
2.02'
2.71
1.99
0.33
.v
and -t
4
2.12
_ _
.
3 23
0.89
Use computer software to evaluate each of the fol
lowing.
(a) 3vi - 2v2 + Sv3 - 3\14
Cb) 2Avi - I.3v2 +
Yiv3 - Y3v4
Conceptual Problems
Dl Prove property(8) from Theorem 1.
D6 Let 13
D7 Let v1,Vz
of JR.I!.
I i1
U, v
E V}. Prove
j3 +
Span{V1,sv1 + tV2}
a counterexample.
(a) If i12
is linearly dependent.
(b) If v1 is not a scalar multiple of v2, then {V1, 112)
is linearly independent.
(c) If {V1,v2,v3} is linearly dependent, then v1 can
be written as a linear combination of v2 and v3.
(d) If v1 can be written as a linear combination of
Vz and V3, then {V1,vz, v3} is linearly depen
{V1,
Span{V1, v2}
(c) Define U + V
DS Let 13
{V1,
dent.
(e) {vi} is not a subspace oflR.11
(f)
28
Chapter 1
3
2
Length and Dot Products in R , and R
The length of a vector in
Definition
If x
Length in ::2
[]
xf
0
Figure 1.3.10
For vectors in
JR3,
Length in JR2.
Definition
Length in:::.'
If it
[:l
xf
x +
One immediate application of this formula is to calculate the distance between two
Figure 1.3.11
EXAMPLE 1
Solution: We have
PQ=
Length in
[ - j [-]
=
-3
29
JR3.
5 , 1) in IR.3.
points is
IR.2
leads to the important idea of the dot product of two vectors. Consider
[ ]
[]
q
- qi
OP= jJ = P1 , OQ= q= 1 . PQ= jJ - q= Pi P2
q2
P2 q 2
simplifying gives
Observe that if
jJ
q, then e
PT + p + p= llf1112
(1 .1)
30
Chapter 1
Figure 1.3.12
llPQll2
This matches our definition of length in JR3 above. Thus, we see that the formula on the
left-hand side of these equations defines the angle between vectors and also the length
of a vector.
Thus, we define the dot product of two vectors
x y
x y
X1Y1
[::]
X1Y1
[]. y []
=
in IR2 by
X2Y2
and jl
X2Y2
in IR3 is defined by
x3y3
Thus, in IR2 and JR3, the cosine of the angle between vectors
xst
=
EXAMPLE2
Solution: We have
v. w
llvll
llwll
[ _] [-H
W
So e
:::::
1.966
radians.
../21
Y9
16
Y26
Hence,
and n.)
1r.
cose
(1.2)
1111111.Yll
-9
-9
Y26::::: -0. 3 85 1 6
W
EXAMPLE3
Solution: We have v
Hence, cos e
[ ]
-
1(2)
and w
(-2)(1)
0. Thus, e =
=
[n
=
w
0.
radians.
EXERCISE 1
llilll llwll
U]
and "1 =
31
[ :l
=
We now extend everything we did in JR2 and JR 3 to JRll. We begin by defining the dot
product.
Definition
Let
x=
Dot Product
Xi
Xn
and y =
rll
'
X1Y1
X2Y2
'
'
x and y is
XnYn
Remark
The dot product is also sometimes called the scalar product or the standard inner
product.
Theorem 1
Let
x, y, z E JR11
and
t E R Then,
Because of property (1), we can now define the length of a vector in JR n. The word
norm is often used as a synonym for length when we are speaking of vectors.
32
Chapter 1
Definition
Let x
[1 ].
Norm
Xn
EXAMPLE4
2
Let1
1 /3
andy
-2/3
=
-2/3
-1
Solution: We have
11111
11.Y11
EXERCISE2
Let X
[]
and let jl
....;1/9+4/9+o+4/9
11 11
Theorem 2
11
(2) llt111
...j(txi)2 + + (txn)2
-../ti xT + +x
Jtlll1JI.
0 or 1
t.Y. Ify is
11 .YI
11 Ct1)1
1rc1 1)1
1r1111112
11111 11r111
11111 11.Yll
>
0,
( 1.3)
Note that
1 1
>
0 since 1
f.
0.
33
>
is always positive if and only if the corresponding equation has no real roots. From
the quadratic formula, this is true if and only if
implies that
B2 - 4AC
<
<o
111+1112 Cll111+11111)2
The squared form will allow us to use the dot product conveniently. Thus, we consider
111+1112 - c11111+11111)2
by
(3)
Remark
Property
(3)
is called the
EXERCISE 3
Definition
Unit Vector
A vector 1
1.
We will see that unit vectors can be very useful. We often want to find a unit vector
that has the same direction as a given vector
that this is the vector
A
1.
(1.2).
we see
of
ffi..t
3,
1 and 1 in IR"
by matching equa
3,
ing definition.
Definition
Orthogonal
0.
34
Chapter 1
EXAMPLES
1
2
-1
3
0
Let v =
, w =
. Show that v is orthogonal to w but v is not
, and z... =
3
0
1
2
-2
orthogonal to Z.
Solution: We have v w = 1(2) + 0(3) + 3(0) + (-2)(1) = 0, so they are orthogonal.
v z = 1 (-1) + 0(-1) + 3(1) + (-2)(2) = -2, so they are not orthogonal.
...
ti vi
t2v2, ti, t2
jJ
more useful to have a scalar equation that represents the plane. We now look at how
to use the dot product to find such an equation.
Suppose that we want to find the equation of a plane that passes through the point
[:n
(That
equation of this plane, let X(x1, x2, x3) be any point on the plane. Then it is orthogonal
to
PX, so
This equation, which must be satisfied by the coordinates of a point X in the plane, can
be written in the form
This is the standard equation of this plane. For computational purposes, the form
it (x - jJ)
Xi
Figure 1.3.13
The normal it is orthogonal to every directed line segment lying in the plane.
35
EXAMPLE6
Find the scalar equation of the plane that passes through the point P(2, 3, -1) and has
nonnfil vectoc n
Hl
it (X
- p)
[l
[XJ - 2]
-4
X2 - 3
X3 + 1
or
X1 - 4x2 + X3
-11
It is important to note that we can reverse the reasoning that leads to the equation
of the plane in order to identify the set of points that satisfies an equation of the form
where n
[::l
This equation describes a plane through the point P(d In" 0, O) with
normal vector it. If if.2 * 0, we could combine the d with the x2 term and find that the
plane passes through the point P(O, d/n2, 0). In fact, we could find a point in the plane
in many ways, but the normal vector will always be a non-zero scalar multiple of it.
EXAMPLE 7
- p)
Describe the set of points in JR. that satisfies 5x1 - 6x2 + 7x3
or
( - l )-
5 x1
6(x2 - 0) + 7(x3 - 0)
x3
11.
=
1-H
0, we find that x2
Two planes are defined to be parallel if the normal vector to one plane is a non
zero scalar multiple of the normal vector of the other plane. Thus, for example, the
plane Xt + 2x2 - X3
7.
36
Chapter 1
Two planes are orthogonal to each other if their normal vectors are orthogonal.
For example, the plane x1
x2
x3 =
x2 - 2x3 = 0
since
EXAMPLE8
3x2 - 5x3
6.
the plane is
2x1
EXAMPLE9
(-5)(-1) = 21
3(4) +
4x3 = 2.
P,
[-il
so we pick
[H
Thus, an
2x2
EXERCISE4
X3 = 0(3) + 2(-1)
(1)(3) = 1
= -5.
can find the scalar equation of a hyperplane in IRn. In particular, if we have a vector
then for
o = m
PX
m. ex - ff)
O=mx-mff
mx=mff
m1X1
mnXn = m p
Thus, we see that a single scalar equation in JR.11 represents a hyperplane in IR".
EXAMPLE 10
. Find the scalar equation of the hyperplane in JR.4 that has normal vector m
37
2
=
3
_
and
2
3x2 - 2x3
x4
2(1)
3(0)
(-2)(2)
1(-1)
-3
PROBLEMS 1.3
Practice Problems
Al Calculate the lengths of the given vectors.
(a)
[ 1
-
{]
[I
(e )
(b )
[:;1
{]
[ l
(f)
l/Y3
l/Y3
-l/Y3
-1
(a) x
(b) ,
(a)
[-1
{]
-2
(e)
-2
1
0
(b)
rn
+J
(f)
and y
and
AS Determine
(a)
m [!]
H l nl
Schwarz inequality if
whether
each
pair
of
vectors
is
orthogonal.
(g)
0
0
-1
(c)
(e )
lH Hl
m nl
0
0
'
X3
4
(d)
-1
O'
4
3
-2
1/3
3/2
2/3
(f)
-1/3 ' -3/2
(c)
[ n [ ]
_
lH Hl
(d)
2
3
4
'
k
k
-k
0
38
Chapter 1
(b) P(2,5,4),fl=
[_]
[]
[-;]
[=l
0
1
2
(d) P(l,0,1,2,1),n=
-1
(c) P(l,-1,1),fl=
(d) P(2,l,1),fl=
1
-4
(c) P(O,0,0,0),n=
5
-2
[!]
0
1
(b) P(2,-2,0,l),n=
3
3
Homework Problems
Bl Calculate the lengths of the following vectors.
(a)
(c)
[l
[-]
-3
3/YW
1;Y25
Ce)
-3/YW
(b)
(d)
(f)
(c)
r-!l
[1il
2
2
(e)
0
[ ]
- 3
1
2/3
1
2
-1
1
3
-1;Y26
(a)
[ 1
(b)
r;1
]
<{
-1
(f) -1
-1
3
{]
[J
) x=
an<lY =
n<lY=
BS Determine whether
orthogonal.
(a)
(c)
rn r-1
mDl
nl
m
each pair
of
vectors
b
( )
(d)
is
r:l r-1
[ilHl
-3
(e)
1 '
-2
(f) -1 ' -1
-2
3
(a) P(2,l,1,5),it=
(c)
-5
1
(b) P(3,1,0,7),it=
vectors is orthogonal.
(a)
-2
39
-4
l
-3
[n []
[
m ]
0
O
0
0
[-1
[-]
[]
[: l
(d) P(l,2,0,1,1),it= -2
(a) P(-3,-3,1),it=
(c) P(0,0,0),it=
(d) P( I, l, I),it=
3
(b) -xi - 2x2 + 5x3 = 7 in JR.
4
(c) xi + 4x2 - X4 = 2 in JR.
4
(d) X1 + X2 + X3 - 2x4 = 5 in IR.
(e) xi - x5 = 0 in JR.5
(f) X1 + X2 - 2x3 + X4 - X5 = 1 in JR.5
BlO Find an equation for the plane through the given
3x3 = 7
Computer Problems
1.12
1.00
2.10
3.12
-3.13
1.21
3.31 , and
4.15
-2.21
1.14
6.13
2.00
-0.01
1.12
0
v4 = 2.13 .
3.15
3.40
Use computer software to evaluate each of the fol
lowing.
(a) i11 i12
(b) lli13ll
(c) i12 v4
2
(d) lli12 + v4[[
40
Chapter 1
Conceptual Problems
Dl (a) Using geometrical arguments, what can you
say about the vectors
1=
scalar equation it 1 = k have no point of inter
vector equation
S. That is,
S .L = { w
section?
111111-111111111-111.
(Hint:
11111=111-.Y +yll.)
and
Q.
Explain
llQXll.
llPXll
cos,B
cosy
cos a
[ l
06 Let it
S}
tor.
cosines.
1.
tity cos2 e
+ sin2 e = 1.
Projections
First, let us consider the case where
Yt e1
[]
1 = e1
[] =
CJ x)x. This will be called the projection of y onto x and is denoted proj1 y.
41
1
X1
0
X1
proj_.. y is a vector in the direction of x.
Figure 1.4.14
2
Next, consider the case where 1 E JR has arbitrary direction and is a unit vector.
First, draw the line through the origin with direction vector x. Now, draw the line
perpendicular to this line that passes through the point (y1, Y ) . This forms a right
2
triangle, as in Figure
The projection of y onto 1 is the portion of the triangle
that lies on the line with direction x. Thus, the resulting projection is a scalar multiple
of 1, that is proj_x y = k1. We need to determine the value of k. To do this, let tdenote
the vector from projxy toy. Then, by definition, tis orthogonal to x and we can write
1.4.14.
y = t + projx y = t + kx
We now employ a very useful and common trick-take the dot product of y with x:
EXAMPLE 1
[ ]
-
[1il
Solution: We have
i1
=
=
=
(-3 1 ) [1/.../2]
.../2 .../2 1;.../2
[1/.../2]
. .../2 1/.../2
[=]
+
[11.../21
1/.../2J
X1
-2
2
If x E JR is an arbitrary non-zero vector, then the unit vector in the direction of 1
is x = 11:11 Hence, we find that the projection of y onto x is
,
,"'t
"'t
,
( ,"'t
proJx Y = proJx Y = v
x)x =
A
( . jjiji ) jjiji
y 1
'
111112
To match this result, we make the following definition for vectors in JR11
Definition
Projection
proJx y =
yx
1
111112
42
Chapter 1
EXAMPLE2
Let V =
[ :J
_
and it =
...
Solution: We have
l-n
42 +32 +c-1)
11v1 1 2
[][ 1
[1[ 1
8/13
4
3 =
6113
26
-1
-2/13
uv
(-2)(4)+5(3)+3(-l) =
=
v
v=
ProJv l1
2
vl1
(4)(-2)+3(5) +(-1)(3)11
prOJu v = 11 11 =
=
2
38
(-2)2 +52 +32
11 1\
.
-2
5 =
3
-4/19
10/19
6/19
Remarks
proj" 1. Of course, we
y.
EXAMPLE3
Let v =
[jl
1/.../3
and 11 =
[ 1.
-2
Find proj;111.
11v
EXERCISE 1
Let V =
[ii
and it
l-n
V3
[ jJ []
l/
Section
43
We begin by restating the problem. In JR.n, given a fixed vector x and any other y,
express
write y
is,
c=
{i;I,
so in fact, w
= ex =
proj1
y,
bonus of approaching the problem this way is that it is now clear that this is the only
way to choose w to satisfy the problem.
Next, since y
to x? To check, calculate
x = CY
proj1 y)
y. x
x .x
- 111112
J. x
= J 1
x x)
- 111112 e
= y.x
( )
( )
= y. x -(fi;1) 111112
=Jx-Jx=o
So, zis or thogonal to x, as required. Since it is often useful to construct a vector zin
this way, we introduce a name for it.
Definition
Perpendicular of a Projection
to be
y= proj1 y + perp1 y.
0
Figure 1.4.15
perp_,(51)
y.
44
Chapter 1
EXAMPLE4
Let v =
and ii =
Solution:
[H
perpv it= i1
EXERCISE 2
Let v =
and ii=
[-H
projv i1 =
EXERCISE 3
1)
It follows that perp1 also satisfies the corresponding equations. We shall see that
proj1 and perp1 are just two cases amongst the many functions that satisfy the linearity
properties.
proj1 and perp1 also have a special property called the projection property. We
write it here for projx. but it is also true for perp1:
proj1 (proj1y) = proj1y,
Minimum Distance
What is the distance from a point Q(q1, q2) to the line with vector equation x = jJ + tJ?
In this and similar problems, distance always means the minimum distance. Geomet
rically, we see that the minimum distance is found along a line segment perpendicular
45
to the given line through a point Pon the line. A formal proof that minimum distance
requires perpendicularity can be given by using Pythagoras' Theorem. S
( ee Problem
D3.)
To answer the question, take any point on the line x jJ + tJ The obvious choice
is P(pi, p2) corresponding to jJ. From Figure 1.4.16, we see that the required distance
is the length perp;PQ.
x = (1,2) + t(-1, 1)
d=(-1,1)
0
X1
Figure 1.4.16
EXAMPLES
Q to the line 1
PQ
= [ = J = [n
II perp;PQll
2)
f1 + tJ is II perpJPQll.
= [J [-n
+ t
t ER
- Q
1 m-(-13 ) [- Jll
= 1 m l- Jll=1 l;J11=2
=
llPQ
proj;P ll
+ l
+l
V2
Notice that in this problem and similar problems, we take advantage of the fact
that the direction vector J can be thought of as "starting" at any point. W hen perp;AB
is calculated, both vectors are "located" at point P. W hen projections were originally
defined, it was implicitly assumed that all vectors were located at the origin. Now, it
is apparent that the definitions make sense as long as all vectors in the calculation are
located at the same point.
We now want to look at the similar problem of finding the distance from a point
Q(q1, q2, q3) to a plane in JR.3 with normal vector it. If Pis any point in the plane, then
proji1 PQ is the directed line segment from the plane to the point Q that is perpendicular
46
Chapter 1
PQ
Xi
Figure 1.4.17
EXAMPLE6
n3X3
d?
t=
0, we pick
P(d/ni, 0, 0)
with equation
nixi +n1x2 +
II proJ,1 PQll
...
I
I
l 11ni12 l 11ni1
l 1!1). l
l2
(q llrliff)
ei/
ei/
fJ) rt
rt
This is a standard formula for this distance problem. However, the lengths of pro
jections along or perpendicular to a suitable vector can be used for all of these prob
lems. It is better to learn to use this powerful and versatile idea, as illustrated in the
problems above, than to memorize complicated formulas.
47
OP+ PK
= OP+ perp11 PQ
-+
-+
-+
OR =
-+
O Q + QR = O Q + proj,1 QP
-+
-+
-+
-+
Notice that we need QP here instead of PQ. Problem D4 asks you to check that these
PROBLEMS 1.4
Practice Problems
Al For each given pair of vectors v and ii, determine
proj;1 i1 and perp;1 ii. Check your results by verify
ing that projil i1 + perp;1 i1 = i1 and v perp;1 i1 = 0.
[J [ ]
r J r-4J
.a= -
3/5 'u.... =
Cb) ....
v= 4
/5
(c) v =
[H {i]
[ ;] [ 1
a
Cd) v= -
2/3
,a=
(e ) v_-; =
....
ou =
-2
-1
-1
en v =
-3
-1
1
1
0,
Find the point on the plane x1 - 2x2 + 2x3 = 5 that is closest to the point Q(2, 1, 1).
EXAMPLE 7
(a) v=
[ i].
P=
2
a=
-3
[ J [ ;J
.a= _
U l Hl
{; ] H l
Ul. {ll
( b) V=
(c) v
( d) "=
-1
(e) v=
2
l
-3
2
-1
a=
and we
Chapter 1
48
[rn a [H
F=
and let
a.
a.
F onto
a.
perpendicular
[ !] a Ul
F=
and
X3 - X4= 4
[ J [ l
[] [-l
(a) Q(O,O),linex=
+r
+t
tEIR
t EIR
U] [ H
[J t[H
(c) Q( l,0,l),lineX=
+t -
tElR
t E JR
Homework Problems
a a. a a
[ ] [ ]
[ :]. a [ ]
=[]a= Hl
U J nl
a
....
(c) V
+ perpil
a=
(d) '=
o ,
(a) v=
(b ) v=
(c) V
-1
'u =
[]
[n a
[ =n
[]
a=
+; l a Ul
Ul a= [l
U l [ ]
,a
=
(d) ,=
a=
(e) v=
-4
v=
= 0.
(f)
and v . perpil
a=
(e) v=
X4
(f)
v=
-1
2
= -1
= [-q]
and let
nl
F perpendicular
(a)
= [ !] = HJ
and U
A2
but
(b)
(c )
(d)
( a)
(d)
(bl
(c )
(d)
line 1
JR
JR
Computer Problems
Cl Letv1
1.2.1102 1.3.0102
= 4.7.0135 'V2 = -0.-2.2415 'V3
6.13 2.00
V3
-1.1.1.000102 V4 = 1.2.01132 .
3.3.4105
-1.1.0034
, and
(a) proji11 v3
(b) perpit 1
(c) proj;:t3 v1
(d) proj;:t4 v2
plane
plane
(c )
(a)
-5x3= 5 = 7
2x1x 2x1+-x2x2-3x2-4x3
-X3
=
-xi + 2x2 -X3 =JR.54,
plane
plane
determine
the line.
Q(Q(2O,,0,-32),,-1),
Q(Q(2O,,0-1,, 0)2),,
(b)
49
Q(Q(2l,, 13,,00,,-1-1
3x4Q(3=,1,02,
X4Q(=5,23,3,7),
3x4 = 40
), hyperplane
2x12x1-x3- 2x2+3x4+=x30+
3x1 - x2 - X3 +
2x + x2 + 4x3 +
), hyperplane
-6), hyperplane
hyperplane
50
Chapter 1
Conceptual Problems
3
* 0 and
* 0,
JR.
---+
-+
DS (a) Let i1 =
D2 By
linearity
proja( -1)
property
(L2),
we
know
that
-+
and X =
[H
-+
Show that
0.
2
D3 (a) (Pythagoras' theorem) Use the fact that 11111 =
2
2
2
11 to prove that 111+111 =11111 +11111 if and
U]
proja(perpa(1)) =
only if 1
-+
0 for every 1.
1 =0.
td and
qll2
is obtained when
pendicular to
11.8'
d).
projjp) + projjp)
qll.)
is per-
qll
Cross-Products
3
In Chapter 2, we shall develop systematic methods for solving such equations for
EXERCISE 1
U2V3 - U3V2
U3V1 - U1V3
u1v2 - u2v1
51
Also notice from the form of the equations that any multiple of w would also be
orthogonal to both i1 and v.
Definition
Cross-Product
[:]
and V
[]
is defined by
U2 V3-U3V2
ilxv= U3V1-U1V3
U1V 2-U2 V1
EXAMPLE l
m [-il
[5] [-1 [-65--54] [-]
-
Solution:
and
(-3)
Remarks
1. Unlike the dot product of two vectors, which is a scalar, the cross-product of
two vectors is itself a new vector.
3
The formula for the cross-product is a little awkward to remember, but there are
many tricks for remembering it. One way is to write the components of i1 in a row
above the components of v:
Then, for the first entry in i1 xv, we cover the first column and calculate the difference
of the products of the cross-terms:
For the second entry in i1 xv, we cover the second column and take the negative of the
difference of the products of the cross-terms:
U1
U3
V1
V3
-(U1V3-U3V1)
Similarly, for the third entry, we cover the third column and calculate the difference of
the products of the cross-terms:
52
Chapter 1
Note carefully that the second term must be given a minus sign in order for this proce
dure to provide the correct answer. Since the formula can be difficult to remember, we
recommend checking the answer by verifying that it is orthogonal to both i1 and v.
EXERCISE2
[ ] m
and
EXERCISE 3
Let
but
Check that in every case, the three vectors taken in order form a right-handed system.
These simple examples also suggest some of the general properties of the cross
product.
Theorem 1
for
x,y,zE JR.3
and
t E JR., we have
(4) ( tx)xy=t(xxy)
Proof: These properties follow easily from the definition of the cross-product and are
left to the reader.
One rule we might expect does not in fact hold. In general,
exxy)x t
This means that the parentheses cannot be omitted in a cross-product. (There are for
mulas available for these triple-vector products, but we shall not need them. See Prob
lem F3 in Further Problems at the end of this chapter.)
il
and v, the direction of their cross-product is known. W hat is the length of the
Theorem 2
Let
it, v E JR.3
53
(U21
Thus,
lli1
U2
2
U23)(V21
V2
2
V2)
3 - (UtVt
U V
2 2
U3V3)2
To interpret this formula, consider Figure 1.5 .18. Assuming that i1 x v f:. 0, the
vectors z1 and v determine a parallelogram. Take the length of z1 to be the base of the
parallelogram; the altitude is the length of perpa v. From trigonometry, we know that
this length is IJVll sine, so that the area of the parallelogram is
(base)
Xt
Figure 1.5.18
EXAMPLE2
Find the area of the parallelogram detennined by i1=
Solution: By Theorem 2, we find that the area is
lli1
VII=
[]
= I
[]
and V =
[il
54
Chapter 1
EXAMPLE3
[-]
and ii =
[3l
and
EXERCISE4
[-: l
ii
LH
The cross-product allows us to answer many questions about lines, planes, and dis
tances in JR3.
EXAMPLE4
The lines 1
[i] []
+
and 1 =
m [- ]
+ t
have the point (1, 3, 2) in common. Find a scalar equation of the plane that contains
these lines.
Therefore, since the plane passes through P(l, 3, 2), we find that an equation of the
plane is
-4xi - 3x2
2x3 = (-4)(1)
(-3)(3)
2(2) = -9
55
EXAMPLES
Find a scalar equation of the plane that contains the three points
Q(2,-2,-1), and R(4,
Solution: Since
and
1, 1).
_,
PQ x PR
3 x3
(6)(1)
[j] [] HJ
x
1(3)
21
or
2x1 - 2x2
X3
EXAMPLE6
3x3
x2 - 2x3
6.
[ _l] [-H
and
x2
and 2x1 - x2
6. The solution is x1
EXERCISE 5
Find a vector equation of the line of intersection of the two planes -x1 - 2x2
and 2x1
x2 - 2x3
1.
x3
-2
56
Chapter 1
determined by
That is,
altltude
.
II proJ,1 w
...II =
.
lw . ill
liitil
lw . (it xv)I
11a xvii
Thus, to get the volume of the parallelepiped, multiply this altitude by the area of the
base to get
w (it xv) is
lw. ca xv)I
x11a xvii
Ila xvii
lw ca xv)I
w, it, and v.
Notice that
axv
altitude
base area
Figure 1.5.19
11 proj,1 wll
11axVil
lw. ca xv)I.
a, v, w
The sign of the scalar triple product also has an interpretation. Recall that the
ordered triple of vectors
{a, v, a xv}
vector
w is then "upwards,"
and
a xv as the
sa + tf!. Some other
{a, v, w}
the scalar triple product is positive. If the triple scalar product is negative, then
{a, v, w}
is a left-handed system.
It is often useful to note that
w . ca xv)
a . cv xw) = v . cw xa)
EXAMPLE 7
Find the volume of the pamllelepiped determined by the vectors
V=
m l[:H=rn [JHJ
=
=2
[H [H [ n
and
57
PROBLEMS 1.5
Practice Problems
Al Calculate the following cross-products.
+iJ {l]
{] +J
{Hl
[H=ii
(+I {!l
[!Hll
nl Ul [=H
A2 ld =
v =
and
w =
Check
hold.
axa = o
Cb) axv= -v xa
(c) ilx3w=3(ilxw)
(d) ilx(v +w) =ilxv+ilxw
c e) a. cv xw) = w . caxv)
Cf) a. cv xw) -v. ci1 xw)
(a)
[H Ul
[lHl
inR3)
(c)
X= -
+s
(d)
X=s
+1
(b)
(d)
[ii
(b) ,
[] Ul ,[l]
Ul [ii f11
[ i] H] []
][ {]
+
( c)
() ,
(a)
(a)
vector equation
(d)
co
[Hll
[-nr1
[-!]
and
P(-l,4,2),Q(3,l,-l),R(2,-3,-l)
P(l,O,1),Q(-1,0, 1),R(0,0,0)
- X3 = 5 and
2x1 - 5x2 +-.x3= 7
2x1 - 3x3 = 7 and x2 + 2x3 = 4
(a) Xi + 3x2
A 7 (b)
Find the volume of the parallelepiped determined
by each set of vectors.
()
(b)
(c)
(d)
(e)
[Hrrnl
u1rn m
nrnrn1
[_H urn1
n1nrn1
58
Chapter 1
Homework Problems
Bl Calculate the following cross-products.
Hl {]
m +l
[ ] {i]
[!] Ul
(a)
(c )
( el
+!Hl
<aflHil
{iHil
Hl
B2 Ut U =
V =
and W =
Ch k
(a) i1x i1 =
vector equation
(a) x
{i] {] ll
nl [!] ftl
[;] f1l nl
[] f l
+
1
(c )
+t
(d) X=
+i
(c ) i1 x 2w = 2(i1x w)
(c ) P(0,2,1), Q(3,-1,1),R(l,3,0)
lH Hl
{Jr:J
(a)
(b)
(d)
lrllfl
m. [-;]
m [-l
and
X3 = 5 and
7X2+- X3 = 6
(b) xi - 3x2 - 2x3 = 4 and
3x1 + 2x2 + x3 = 2
(a) x1 + 4x2
3X1
(b)
(c)
(d)
HlUlUI
liHH []
lH lH Ul
lH lH Ul
Conceptual Problems
Dl Show that if X is a point on the line through P and
Q, then x x (q -
p)
--7
....
p x if,
where x =
OX,
sv + tW.
59
(b) ax (v x w) *
(u xv) x w
(u xv) x w
CHAPTER REVIEW
Suggestions for Student Review
Organizing your own review is an important step to
v is defined
v. (Section 1.4)
10 Explain with a picture how projections help us to
(Section 1.2)
Chapter 1
60
Chapter Quiz
propriate level of difficulty for a test on this material.
both
{[], [ ]}
E4 Pro,e thatS
{[::J
3
is a subspace of JR. for any real numbers a1, az, a3
if and only if d
[- ]
0.
[H
t -
. ,
and
[] nl
proj.Y x.
(vi) For any vectors x and y, the set {proj1 y, perp1 y}
=
is linearly independent.
t ER
Further Problems
These problems are intended to be a little more chal
lenging than the problems at the end of each section.
Some explore topics beyond the material discussed in
the text.
F4 Prove that
(a) it v
11a + vll2 - llit - vll2
211a112 + 211v112
Cb) 11a + vli2 + 11a - v112
=
61
if)
cif
1'J
c1 x fJ) = o
skew.
(a) Observe that if two lines are skew, then they do
not lie in a common plane. Show that two skew
lines do lie in parallel planes.
(b) Find the distance between the skew lines
1=
MyMathlab
[Hl
,seRand1=
HHil
teR
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 2
Systems of Linear
Equations
CHAPTER OUTLINE
2.1
2.2
2.3
2.4
Linear Equation
form
(2.1)
The numbers a1, ..., a,, are called the coefficients of the equation, and b is usually
referred to as "the right-hand side," or "the constant term." The xi are the unknowns or
variables to be solved for.
64
Chapter 2
EXAMPLE 1
The equations
X1 +
x1 - 3x2 +
2x2
../3x3
(2.2)
JTX 4
(2.3)
are both linear equations since they both can be written in the form of equation (2. 1 ).
The equation
Definition
Solution
EXAMPLE2
A vector
x2
n
in !R is called a solution of equation (2.1) if the equation is satisfied
Sn
when we make the substitution x1
s1, x2
(2.2) are
s2, ... , Xn
[n [Os l [ ]
and
2 + 2(1)
3 + 2(0.5)
6 + 2(- 1 )
The vector
0
0
0
0
Sn.
since
4
4
4
.../3x3
JTX4.
a 11X1 + a12x2 +
a21X1 + a22X2 +
+ a1nXn
+ a2nXn
b1
b2
Note that for each coefficient, the first index indicates in which equation the coefficient
appears. The second index indicates which variable the coefficient multiplies. That
is, au is the coefficient of Xj in the i-th equation. The indices on the right-hand side
indicate which equation the constant appears in.
We want to establish a standard procedure for determining all the solutions of such
a system-if there are any solutions! It will be convenient to speak of the solution set
of a system to mean the set of all solutions of the system.
The standard procedure for solving a system of linear equations is elimination. By
multiplying and adding some of the original equations, we can eliminate some of the
Section 2.1
65
variables from some of the equations. The result will be a simpler system of equations
that has the same solution set as the original system, but is easier to solve.
We say that two systems of equations are equivalent if they have the same solution
set. In elimination, each elimination step must be reversible and must leave the solution
EXAMPLE 3
X1
X2 - 2X3
X1
3x2 - X3
2x1
x2 - Sx3
X1
2x1
X2 - 2x3
2x2
X3
x2 - Sx3
Note two important things about this step. First, if x1, x2, x3 satisfy the original system,
then they certainly satisfy the revised system after the step. This follows from the rule
of arithmetic that if P
Q and R
S, then
P + R
two equations and both are satisfied, the resulting sum equation is satisfied. Thus, the
revised system is equivalent to the original system.
Second, the step is reversible: to get back to the original system, we just add (1)
times the first equation to the revised second equation.
Add ( 2 ) times the first equation to the third equation.
-
X1
X2 - 2X3
2x2
+ X3
-X2 - X3
-1
Again, note that this step is reversible and does not change the solution set. Also note
that x1 has been eliminated from all equations except the first one, so now we leave
the first equation and turn our attention to x2.
Although we will not modify or use the first equation in the next several steps, we
keep writing the entire system after each step. This is important because it leads to a
good general procedure for dealing with large systems.
It is convenient, but not necessary, to work with an equation in which x2 has the
coefficient 1. We could multiply the second equation by 1 /2, but to avoid fractions,
follow the steps on the next page.
66
Chapter 2
Interchange the second and third equations. This is another reversible step that
EXAMPLE3
(continued)
X2 - 2X3
-X2 - X3
2X2
X3
4
-1
3
Multiply the second equation by (-1). This step is reversible and does not
X2 - 2X3
X2
X3
2X2
X3
X2 - 2x3
X2
X3
-X3
}
1
In the third equation, all variables except x3 have been eliminated; by elimination,
we have solved for X3. Using similar steps, we could continue and eliminate X3 from
the second and first equations and x2 from the first equation. However, it is often a
much simpler task to complete the solution process by back-substitution.
First, observe that x3
-1. Substitute this value into the second equation and find
that
X2
1 - X3
1 - ( -1)
Next, substitute these values back into the first equation to obtain
X1
4 - X2
2x3
4-2
2(-1)
[:] [ H
[ ]
=
lent to the original system, this solution is also the unique solution of the problem.
Observe that we can easily check that
tions:
0
0
2(0)
2 - 2( -1)
3(2) - (-1)
+
2 - 5(-1)
4
7
7
It is important to observe the form of the equations in our final system. The first
variable with a non-zero coefficient in each equation, called a leading variable, does
not appear in any equation below it. Also, the leading variable in the second equation
Section 2.1
67
is to the right of the leading variable in the first, and the leading variable in the thfrd is
to the right of the leading variable in the second.
The system solved in Example 3 is a particularly simple one. However, the solution
procedure introduces all the steps that are needed in the process of elimination. They
are worth reviewing.
not
combine steps of type ( 1) and type (3) into one step of the form
EXAMPLE4
X1
X1
2X3
3x3
X4
14
3x4
19
Remark
Notice that neither equation contains x2. This may seem peculiar, but it happens in
some applications that one of the variables of interest does not appear in the linear
equations. If it truly is one of the variables of the problem, ignoring it is incorrect.
Rewrite the equations to make it explicit:
x1
Xi
Ox2
Ox2
2x3
3x3
X4
14
3x4
19
Ox2
2x3
X3
X4
2X4
14
5
Observe that X2 is not shown in the second equation because the leading variable must
have a non-zero coefficient. Moreover, we have already finished our elimination pro
cedure as we have our desired form. The solution can now be completed by back
substitution.
Note that the equations do not completely determine both x3 and x4: one of them
can be chosen arbitrarily, and the equations can still be satisfied. For consistency, we
always choose the variables that do not appear as a leading variable in any equation to
be the ones that will be chosen arbitrarily. We will call these free variables.
68
Chapter 2
EXAMPLE4
(continued)
Thus, in the revised system, we see that neither x2 nor x4 appears as a leading
variable in any equation. Therefore, x2 and x4 are the free variables and may be chosen
arbitrarily (for example, x4
JR and x2
X3
5 - 2X4
5 - 2t
Now, solve the first equation for its leading variable x1:
Xt
14
2X3 - X4
14
2(5
2t) - t
3t
X1
3t
s, t
5 - 2t '
JR
In this case, there are infinitely many solutions because for each value of
and each
value oft that we choose, we get a different solution. We say that this equation is the
and
solution. For many purposes, it is useful to recognize that this solution can be split into
a constant part, a part in
t,
and a part in
XJ
X2
X3
X4
4
0
5
0
s:
0
1
0
0
3
+
0
-2
This will be the standard format for displaying general solutions. It is acceptable
s and x4 in the place of t, but then you must say x2, X4
to
E
JR. Observe that one immediate advantage of this form is that we can instantly see
the geometric interpretation of the solution. The intersection of the two hyperplanes
x1 + 2x3 + X4
P(4, 0, 5, 0).
14 and x1
3x3
3x4
4
4
19 in JR is a plane in JR that passes through
with back-substitution.
section.
EXERCISE 1
2x 1
X1
4x2
Ox3
2X2 - X3
12
4
Use the general solution to find three different solutions of the system.
Section 2.1
69
a12
alj
a111
b1
a21
a22
a2j
a211
b2
ail
ai2
aij
a;n
b;
am!
am2
amj
a,,111
bm
where the coefficient aij appears in the i-th row and }-th column of the coefficient
matrix. This is called the augmented matrix of the system; it is augmented because
it includes as its last column the right-hand side of the equations. The matrix without
this last column is called the coefficient matrix of the system:
au
a12
alj
a111
a21
a22
a2j
a211
a;1
a;2
aij
a;n
amt
am2
amj
amn
For convenience, we sometimes denote the augmented matrix of a system with coeffibi
cient matrix
by
bm
another way of representing a system of linear equations.
EXAMPLES
Write the coefficient matrix and augmented matrix for the following system:
3x1 + 8x2 - l 8x3 + X4
x1+ 2x2
x1+3x2
- 4x3
-
7 x3
+ X4
35
11
Solution: The coefficient matrix is formed by writing the coefficients of each equation
as the rows of the matrix. Thus, we get the matrix
3
A=
-1 8
l j
70
Chapter 2
EXAMPLES
For the augmented matrix, we just add the right-hand side as the last column. We get
(continued)
EXAMPLE6
8
2
3
[:
[
18
-4
-7
1
0
35
11
10
W1ite the system of linear equations that has the augmented matrix
1
0
0
0
-1
0
2
1
1
Solution: The rows of the matrix tell us the coefficients and right-hand side of each
equation. We get the system
x,
-X2
2x3
X3
X3
-2
Remark
Another way to view the coefficient matrix is to see that the }-th column of the matrix
is the vector containing all the coefficients of x1. This view will become very important
in Chapter 3 and beyond.
Since each row in the augmented matrix corresponds to an equation in the system
of linear equations, performing operations on the equations of the system corresponds
to performing the same operations on the rows of the matrix. Thus, the steps in elimi
nation correspond to the following elementary row operations.
Types of Elementary Row Operations
N, then N is row equivalent to M, so we may say that Mand N are row equivalent. It
also follows that if A is row equivalent to Band Bis row equivalent to C, then A is row
equivalent to C.
Let us see how the elimination in Example 3 appears in matrix notation. To do
this, we introduce notation to indicate this elementary row operation. We write Ri + cRJ
Section 2.1
cR;
R; ! R J
71
c.
Additionally,
or ::::} instead of
elementary row operations, one may omfr these indicators of which elementary row
operations were used. However, including them can make checking the steps easier,
and instructors may require them in work submitted for grading.
EXAMPLE 7
[ 11 31 -1-2 4 l
(-1)
2 1 -5
(-1)
[1
2
[;
1 -1-2
-5
-[
is
12 -21
1 -5 l
-12 -2-11
-! l
R3 - R1
(-l)R2
R2+(-l)R1
-[
-[
1 -2
2 1
-5
2-1 -2-1
j]l
-2
4
21 11 31
R2 ! R3
R3 - 2R2
1 11 -21
[ -1 n
3
0
0
3.
have
been performed. Observe that the final matrix is the augmented matrix for the final
system of linear equations that we obtained in Example
EXAMPLES
[ 11 32 31 11419
J
0
0
EXERCISE 2
R2 + (-l)R1
is
[ 1 21 21 11 4
5 J
0
0
0
1.
In the next example, we will solve a system of linear equations using Gaussian
elimination with back-substitution entirely in matrix form.
72
Chapter 2
EXAMPLE9
35
11
10
Solution: W rite the augmented matrix of the system and row reduce:
[:
[
[
8
2
3
-18
-4
-7
1
0
1
35
11
10
11
2
2
3
-4
-6
-7
2
1
2
-4
-3
-6
11
-1
2
l
l
Ri !R2
R3-R1
-[ i
-[[ I
R3 - 2R2
2
8
3
-4
-18
2
2
-4
-6
-3
-7
0
0
0
1
1
0
I
-1
1
0
-4
-3
0
11
35
10
-1
11
-1
4
R1 - 3R1
R1
t R3
To find the general solution, we now interpret the final matrix as the augmented matrix
of the equivalent system. We get the system
xi +2x2 - 4x3
11
X2 - 3X3+ X4
-1
-X4
to get
X4
-4
X2
-1 + 3X3 - X4
xi
11 - 2x2 +4x3
3 + 3t
5 - 2t
X1
X2
X3
X4
5 - 2t
3 + 3t
t
-4
-2
5
3
3
+t
1
0
-4
0
t E JR
Check this solution by substituting these values for x1, x2, x3, x4 into the original
equations.
Observe that there are many different ways that we could choose to row reduce
the augmented matrix in any of these examples. For instance, in Example 9 we could
interchange row 1 and row 3 instead of interchanging row 1 and row 2. Alternatively,
we could use the elementary row operations
R1 - R1
and
R3 - tR1
to eliminate the
Section 2.1
73
non-zero entries beneath the first leading variable. It is natural to ask if there is a way of
determining which elementary row operations will work the best. Unfortunately, there
is no such algorithm for doing these by hand. However, we will give a basic algorithm
for row reducing a matrix into the "proper" form. We start by defining this form.
Definition
Row Echelon Form (REF)
(2) When two non-zero rows are compared, the first non-zero entry, called the lead
ing entry, in the upper row is to the left of the leading entry in the lower row.
Remark
It follows from these properties that all entries in a column beneath a leading entry
must be 0. For otherwise, (1) or (2) would be violated.
EXAMPLE 10
Determine which of the following matrices are in row echelon form. For each matrix
that is not in row echelon form, explain why it is not in row echelon form.
i
0
0
[o o
(b) 0
2
3
(d)
0
0
-3
0
-1
4
2
1
0
-]
Solution: The matrices in (a) and (b) are both in REF. The matrix in (c) is not in REF
since the leading entry in the second row is to the right of the leading entry in the
third row. The matrix in (d) is not in REF since the leading entry in the second row is
beneath the leading entry in the first row.
74
Chapter 2
the procedure described for this submatrix to obtain the next pivot with zeros below
it. Keep repeating the procedure until we have "used up" all rows and columns of the
original matrix. The resulting matrix is in row echelon form.
EXAMPLE 11
Row reduce the augmented matrix of the following system to row echelon form and
use it to determine all solutions of the system:
X1
+ X2
X2 + X3
X1 + 2X2 + X3
-2
Solution: W rite the augmented matrix of the system and row reduce:
1
2
-2
R3 - R1
0
0
R3 - R2
Lil
Observe that when we write the system of linear equations represented by the aug
mented matrix, we get
X\ + X2
X2 + X3
0
-5
Clearly, the last equation is impossible. This means we cannot find values of x1, x2, X3
that satisfy all three equations. Hence, this system has no solution.
Remarks
1. Although the previous algorithm will always work, it is not necessarily the
fastest or easiest method to use for any particular matrix. In principle, it does
not matter which non-zero entry is chosen as the pivot in the procedure just
described. In practice, it can have considerable impact on the amount of work
required and on the accuracy of the result. The ability to row reduce a general
matrix to REF by hand quickly and efficiently comes only with a lot of practice.
Note that for hand calculations on simple integer examples, it is sensible to go
to some trouble to postpone fractions because avoiding fractions may reduce
both the effort required and the chance of making errors.
2. Observe that the row echelon form for a given matrix A is not unique. In partic
ular, every non-zero matrix has infinitely many row echelon forms that are all
row equivalent. However, it can be shown that any two row echelon forms for
the same matrix A must agree on the position of the leading entries. (This fact
may seem obvious, but it is not easy to prove. It follows from Problem F6 in
the Chapter 4 Further Problems.)
We have now seen that a system of linear equations may have exactly one solution,
infinitely many solutions, or no solution. We will now discuss this in greater detail.
Section 2.1
75
P3. Figure 2.1.1 illustrates an inconsistent system: there is no point common to all three
planes. Figure 2.1.2 illustrates a unique solution: all three planes intersect in exactly
one point. Figure 2.1.3 demonstrates a case where there are infinitely many solutions.
Figure 2.1.1
Two cases where three planes have no common point of intersection: the
corresponding system is inconsistent.
point of intersection
Figure 2.1.2
Figure 2.1.3
Three planes that meet in a common line: the corresponding system has
infinitely many solutions.
Row echelon form allows us to answer questions about consistency and uniqueness.
In particular, we have the following theorem.
76
Chapter 2
Theorem 1
equivalent to
[s I c] is of the
], with c * 0.
(2) If the system is consistent, there are two possibilities. Either the number of
pivots in S is equal to the number of varia
_ bles in the system and the system has
a unique solution, or the number of pivots is less than the number of variables
and the system has infinitely many solutions.
Proof: (1) If
...
]. where c * 0,
then this corresponds to the equation 0 = c, which clearly has no solution. Hence, the
system is inconsistent. On the other hand, if it contains no such row, then each row
must either be of the form
satisfied by any values of
x1,
0
.
, x11,
that consist entirely of zeros, leaving only rows with pivots. In the latter case, the
corresponding system can be solved by assigning arbitrary values to the free variables
and then determining the remaining variables by back-substitution. Thus, if there is no
row of the form
(2) Now consider the case of a consistent system. The number of leading variables
cannot be greater than the number of columns in the coefficient matrix; if it is equal,
then each variable is a leading variable and thus is determined uniquely by the system
corresponding to
[s I c].
free variables, and they may be chosen arbitrarily. Hence, there are infinitely many
solutions.
Remark
As we will see later in the text, sometimes we are only interested in whether a system
is consistent or inconsistent or in how many solutions a system has. We may not nec
essarily be interested in finding a particular solution. In these cases, Theorem
related theorem in Section
1 or the
1
1
2
1
3
1
-2
-1
-5
Choosing one particular row (in this case, the first row) and adding multiples of it to
several other rows is perfectly acceptable. There are other elementary row operations
that can be combined, but these should not be used until one is extremely comfortable
Section 2.1
77
with row reducing. This is because some combinations of steps do cause errors. For
example,
R1 -R1
R1 - R1
0
0
-1
1
-1
1
(WRONG!)
This is nonsense because the final matrix should have a leading 1 in the first column.
By performing one elementary row operation, we change one row; thereafter we must
use that row in its new changed form. Thus, when performing multiple elementary row
operations in one step, make sure that you are not modifying a row that you are using
in another elementary row operation.
A Word Problem
To illustrate the application of systems of equations to word problems, we give a simple
example. More interesting applications usually require a fair bit of background. In
Section 2.4 we discuss two applications from physics/engineering.
EXAMPLE 12
number of dimes is one-half of the total number of nickels and quarters. The value of
the coins is
and
q the
of quarters. Then
n+ d+q= 180
The second piece of information we are given is that
1
d= 2cn+q)
We rewrite this into standard form for a linear equation:
n -2d+q=O
Finally, we have the value of the coins, in cents:
+d
+q = 180
n -2d
+q=0
-2
1
1
10
2S
180
160
1
0
4
180
60
140
R1 -R1
R3 -SR1
-[[
-3
s
R3 -R1
0
0
1
0
20
1
0
4
180
-180
700
180
60
80
(-1/3)R2
(1 /S)R3
number
78
Chapter 2
EXAMPLE 12
(continued)
180
60
4q
80
20, d
60,
180
0.9990x2
1.000
O. lOOOx1
l.OOOx2
1.006
6.000, x1
were given to four digits. Suppose all entries are rounded to three digits. The system
becomes
O. lOOx1
0.999x2
1.00
0.100x1
l.OOx2
1.01
10, x1
only a small change in one term on the right-hand side, the resulting solution is not
close to the solution of the original problem. Geometrically, this can be understood
by observing that the solution is the intersection point of two nearly parallel lines;
therefore, a small displacement of one line causes a major shift of the intersection
point. Difficulties of this kind may arise in higher-dimensional systems of equations in
real applications.
Carefully choosing pivots in computer programs can reduce the error caused by
these sorts of problems. However, some matrices are ill conditioned; even with high
precision calculations, the solutions produced by computers with such matrices may
be unreliable. In applications, the entries in the matrices may be experimentally deter
mined, and small errors in the entries may result in large errors in calculated solutions,
no matter how much precision is used in computation. To understand this problem bet
ter, you need to know something about sources of error in numerical computation
and more linear algebra. We shall not discuss it further in this book, but you should be
aware of the difficulty if you use computers to solve systems of linear equations.
79
PROBLEMS 2.1
Practice Problems
substitution and write the general solution in stan
dard form.
(a)
(d)
x1 -3x2 = 5
X2
(e)
(b)
X1
(c)
x1
2X2 -X3
X3
2 3
3x2 - x
Sx3
X3
X2
2
3
9
13
0
2
4
6
(f)
4
2
0
4
16
20
2
-1
-4
3
1
6
1
0
0
-4
2
0
4
3
3
2
3
-2
11
4
1
3
8
4 4
x
X4
+X4
+X3 +
x1 - 2x2
X2
X3
-3
solution.
; -
n l
2
[ i _; 1
[ i 1
[ - -1
[ H :1
(b) B =
1
0 0
(d)D=
(b)
[ - ]
[i
-1
1
2
(c)
5
3
=i
-1
-1
0
4
0
-1
-2
0
5
- l
0 0 0 0
(c) C
-1
0 0
1 0 1
0 1 0
0 0 0
0 0 0
(e)
-2
0
0 0
0 0
0 0
-1
8
3
2
3x1 -Sx2
Xt
2X 2 = 4
80
Chapter 2
(b )
Xi +2X2
+X3
(d)
xi +2x2- 3x3
( a)
xi +3x2 - 5x3
11
19
-1
-2
!l
-3
cd
36
- 5x3
(b)
-11
-17
( e)
(f)
Xi +2x2-X3
10
4xi +9x2-x3
19
-5
+X4
-8
xi
2x1
+2x2
- 3x3
+ 4x2 -6x3
21
( g)
x, +2x2- 3x3
+X4
+X5
+4xs
+8xs
a,
Homework Problems
Bl Solve each of the following systems by back
(c)
(b )
x1 - 2x2 -x3
X2 +3x3
X3
X1 - 3X2
+X3
X2 +2X3
-2
1
-1
xi +3x2-x3 +2x4
-1
X2-X3 +2X4
-1
X 3 + 3X4
(d)
-2
-3
X3
+X4 - 2xs
[ - 1
[ i l
[l l
[ 1
(c)
(b)
(c) C
(d) D =
(d)
0
-1
0
0
0
0
0
1
0
0
-3
[ H l
[! r Lr l
solution.
(iv) If the system is consistent, write its general
solution in standard form.
(a)
3
6
]
2
3
2
(d)
1
2
4
0
-1
0
-1
2
5
7
6
1
2
4
2
7
2
12
25
3
6
9
6
solution.
(a)
-! -
1
1
0
0
2
0
X1+X2 +X3
-2
6
x1 - 2x2 - 2x3
X2 +X3
Xt +X2 +X3
2x1+3x2 + 3x3
-7
2x1+4x2 +X3
-16
X1+2X2 +X3
(d)
-1
-4
1
0
0
+X2 - X3
(c)
0
1
3
5
2x1+X2 +Sx3
2X1
(b)
-1
0
0
0
(c)
1
0
0
0
tions:
steps.
(a)
[ l
1
3
81
(e)
(f)
X1
+X4
X1
+X4
-21
1
-1
3
82
Chapter 2
the values of
a,
solution.
[ i : j]
-
(a)
(b)
a'
cd
c+d
-S
the cost for each item or explain why the sales slips
must be in error.
Computer Problems
Cl Use computer software to determine a matrix in
row echelon form that is row equivalent to each of
[
[
(b) B
3S
4S
18
13
l7
6S
-61
23
19
41
-2S
-36
37
41
22
SO
-38
49
13
4S
27
-23
-21
27
1.121
-2.0lS
2.131
2.SOl
3.214
4.130
3.11S
-1.639
-12.473
-1.827
8.430
4.612
Section 2.2 Reduced Row Echelon Form, Rank, and Homogeneous Systems
83
Conceptual Problems
Dl Consider the linear system in x, y, z, and w:
+w=b
+y
2x +3y +z +Sw
z
+w
(a) Inconsistent?
(b) Consistent with a unique solution?
3
D2 Recall that in JR , two planes n
2y +2z +aw= 1
x = c and m
EXAMPLE 1
In Example 2.1.7, we row reduced the augmented matrix for the original system to a
row equivalent matrix in row echelon form. That is, we found that
-2
-1
-5
4l [
7
-2
-1
4l
1
[
[
(- l)R3
Rt+ 2R3
[
-[
1
0
Rt - R1
1
x1
0,
x2
2, and x3
84
Chapter 2
This system has been solved by complete elimination. The leading variable in
the }-th equation has been eliminated from every other equation. This procedure is
often called Gauss-Jordan elimination to distinguish it from Gaussian elimination
with
back-substitution. Observe
that the
used
in
Example 1 are exactly the same as the operations performed in the back-substitution in
Example 2.1.7.
A matrix corresponding to a system on which Gauss-Jordan elimination has been
carried out is in a special kind of row echelon form.
Definition
Reduced Row
Echelon Form (RREF)
Theorem 1
For any given matrix A there is a unique matrix in reduced row echelon form that is
row equivalent to A.
Proof: You are asked to prove that there is only one matrix in reduced row echelon
form that is row equivalent to A in Problem F6 in the Chapter 4 Further Problem.
EXAMPLE2
Obtain the matrix in reduced row echelon form that is row equivalent to the matrix
A=
-2
-2
10
-6
-1
-4
R2 - 3R1
(- l )R2
-[
-2
-1
10
-6
- ]
R1
2R2
EXERCISE 1
Row reduce the matrices in Examples 2.1.9 and 2.1.12 into reduced row echelon form.
Because of the uniqueness of the reduced row echelon form, we often speak of
the reduced row echelon form of a matrix or of reducing a matrix to its reduced row
echelon form.
Section 2.2 Reduced Row Echelon Form, Rank, and Homogeneous Systems
85
Remarks
1. In general, reducing an augmented matrix to reduced row echelon form to solve
a system is not more efficient than the method used in Section 2.1. As previously
mentioned, both methods are essentially equivalent for solving small systems
by hand.
2. W hen row reducing to reduced row echelon form by hand, it seems more natural
not to obtain a row echelon form first. Instead, you might obtain zeros below
and above any leading 1 before moving on to the next leading 1. However, for
programming a computer to row reduce a matrix, this is a poor strategy because
it requires more multiplications and additions than the previous strategy. See
Problem F2 at the end of the chapter.
Rank of a Matrix
We saw in Theorem 2.1.1 that the number of leading entries in a row echelon form
of the augmented matrix of a system determines whether the system is consistent or
inconsistent. It also determines how many solutions (one or infinitely many) the system
has if it is consistent. Thus, we make the following definition.
Definition
The rank of a matrix A is the number of leading 1s in its reduced row echelon form
Rank
EXAMPLE3
The rank of the matrix in Example 1 is 3 since the RREF of the matrix has three
leading 1s. The rank of the matrix in Example 2 is 2 as the RREF of the matrix has
two leading ls.
EXERCISE 2
(a) A=
Theorem 2
!]
[ b] be a system of
Let A I
(b) B=
(1) The system is consistent if and only if the rank of the coefficient matrix A is
[ b].
86
Chapter 2
Theorem 2
(continued)
(2) If the system is consistent, then the number of parameters in the general solu
tion is the number of variables minus the rank of the matrix:
#of parameters
n - rank(A)
Proof: Notice that the first n columns of the reduced row echelon form of [A
tq
consists of the reduced row echelon form of A. By Theorem 2.1.1, the system is in
form [ 0
the rank of A.
If the system is consistent, then the free variables are the variables that are not
leading variables of any equation in a row echelon form of the matrix. Thus, by def
inition, there are n - rank(A) free variables and hence n - rank(A) parameters in the
general solution.
Corollary 3
Let [A I
= m.
Definition
A linear equation is
Homogeneous
A system of linear
EXAMPLE4
2X1
X2
X1
X2 - X3
-X2
2X3
Section 2.2 Reduced Row Echelon Form, Rank, and Homogeneous Systems
EXAMPLE4
87
1
[ !; ]l
[
(continued)
-1
1
-1
-1
R1 ! R2
- [ - l
-[
I
-[ - I
R2 - 2 R1
-1
R1+R2
R3 - R2
0
-1
0
0
1
0
(- l )R2
X1
+X3
X2 - 2 X3
t ER Then x1
-x3
-t, x2
x
2 3
2 t,
will
throughout the text, when dealing with homogeneous systems, we are often mostly
interested in how many parameters are in the general solution. Of course, for this we
can apply Theorem .
2
EXAMPLES
[1
2 x1+5x2 + 5x3 + x
2 4 +9xs
[- l
1 ]
3
2
2
7
5
2
7
5
3
2
4
13 R1 - 3R1
9 R3 - 2R1
0
0
2
1
1
2
1
1
0
0
4 RI - 2R2
l
1 R3 - R2
1
- o
0
0
1
0
0
1
0
1
0
0
.2
1
0
The rank of the coefficient matrix is 2 and the number of variables is 5. Thus, by
Theorem 2, there are 5 - 2
EXERCISE 3
88
Chapter 2
PROBLEMS 2.2
Practice Problems
Al Determine the RREF and rank of the following
matrices.
(a)
[! -i]
[ I
[i !]
[ ! -1
1-1 -222 33
2 4 3
[2: di
[31 -2 3 []
21 312 13 442
31 83 22 3
2 7
0
(b)
(c)
(d)
(e)
(D
(g )
(h )
1
0
(i)
1
0
(b)
(c)
(d)
(e)
(f)
0
0
0
0
1
0
0
0
1
0
0
1
0
0
-]
0
0
-5
0
0
0
1
0
0
0
0
2
-S
x
3
x
2
x1x1 4x22x2 -3xx3 33
x3x11 x2x2 -9x3
2x1 x2 -S-7x3x3
-3x4
xx3 11 -3-X2x2 8x2x33 -Sx4
23xx11 -3- x2x22 7Sx3X3 -4x4
-7x4
x2
2
x
3
2
x
4
x12x1 xx22 2 Sx3Sx3 x3 X44 -3-xsxs
xi x2 4x3 2x4 - 2xs
=
(b)
(c)
(d)
0
0
0
0
0
[ - 11
0
0
0
( a)
0
0
0
1
0
0
0
[ -3 2 !I
21
1
[ 4
[1
0
0
1
0
0
0
0
89
(g)
3x1 - x
S 2
X1
(b)
X1
2x2
2x2
2x1 - 3x2
(c)
(d)
-3x1
2x2 - 3x3
x1
3x2 - x
S 3
2x1
Sx2 - 8x3
6x2
16x3
x1 - 2x2
-x
S 3
2x1 - 3x2
- 8x3
2x2 - X3
2x1
Sx2 +x3
4x1
9x2 - X3
2x2
- 3x3
2x1
4x2
- 6x3
13x2 - l7x3
2x1
4x2 - x
S 3
2x1
Sx2 - x
7 3
[A b]
I
(b)
(c)
[ -ll
[f
[:
3x4
8xs
3x4
lOxs
[A ].
A=[: ; nb=[]
A = u J. b= HI
A= [-10 -1-1 ] -- [-14 ]
A= u _! i =H b= m
I0
(a)
19
(b)
(c)
(d)
-4
19
== -21
-5
X4
4x4
-8
(a)
Homework Problems
matrices.
4xs
geneous
11
X4
-17
2 x 2 - 3x3
==
=
3 6
Xi
==
==
=
== -11
=
==10
=
x,
6x1
2x3
X3
x1
( e)
(f)
==
x1
X5
(d)
(e )
-1
1 4
-1
-3
-2
[i d ;]
2
(f)
1
3
0 -1 0
-2
-1
'b
...
90
Chapter 2
(g)
0
3
-2
-6
-4
2x1+7x2
-4
X1 +3X2
(c)
X1 + X2 + X3 - 2X4 = 0
- 14x4=0
X1+4x2
-
]
(d)
(a)
(b)
(c)
[
[
-2
-3
-5
geneous system
(a)
A=
[A I 0].
[ : ]. [-]
b=
-4
-2
1
-4
-1
-5
-5
(b)
A=
(c)
X1 +X2 - 3X3 = 0
x1 +4x2 - 2x3 =0
2x1
- 3x3 =0
9
-4
-1
x1 +5x2 - 3x3 =0
3x1 +5x2 - 9x3 = 0
-4
-5
(b)
=0
(a)
- X5 = 0
2x2 +X3
A= -4
-2
-1
-1
(d)
A=
-1
10
_,
'
-1
1
8
4 -1
5 -2
2 'b
4
_,
5
2
-8
-2
-4
5
-4
_,
1 'b
-4
-6
6
Computer Problems
Cl Determine the RREF of the augmented matrix of
the system of linear equations
2.0lx +3.45y+2.23z=4.13
l.57x +2.03y- 3.1 lz=6.11
2.23x +7.lOy- 4.28z = 0.47
91
Conceptual Problems
0 in JR3 that is simul
taneously orthogonal to given vectors a, b, c E JR3.
i-
the matrix A =
[:: : :i
C1
C2
C3
-n
is the coefficient ma
is 4?
is 3?
01igin.
is 4?
1?
trix
Spanning Problems
Recall that a vector v E Rn is in the span of a set {V1, .. , vd of vectors in JR11 if and
.
Observe that this vector equation actually represents n equations (one for each compo
nent of the vectors) in the k unknowns t1, ... , tk. Thus, it is easy to establish whether
a vector is in the span of a set; we just need to determine whether the corresponding
92
Chapter 2
EXAMPLE 1
Detennine whethec the vectorV
[ ]
[
" rn Hl m =il [ 1
12
13
14
ti
t1
- t2
ti
t2
5t2
2t3 -t4 = -2
+
+
t3 - 3t4 = -3
4t3
[:
[
-1
-2
-3
-3
-1
5
I
-1
-2
-2
-1
-2
-1
R2-Ri
R3 - Ri
3t4
-1
-2
-2
-1
-2
-1
R3
2R2
By Theorem 2.1.1, the system is inconsistent. Hence, there do not exist values of
t1, t2, t3, and t4 that satisfy the system of equations, so v is not in the span of the
vectors.
EXAMPLE2
Ut v,
[H = nl
v2, andV3
v,
andV3
W riteV
[l]
EXAMPLE2
93
(continued)
ti - 2t2
2ti
t1
ti
t3
-1
t3
t3
-1
The solution is t1
2, t2
EXERCISE 1
Detennine whethec v
EXAMPLE3
Considec Span
0, and t3
- I-[ I
0 1
-1
-3
{ t ; , }
Ov2 - 3v3
v.
{[ il Hl [m
=
Xi
Solution: A vector x
x2
X3
is in this set if and only if for some t1, t2, and t3,
X4
1
ti
1
+ t1 3
5
t3 5
3
Simplifying the left-hand side gives us a system of equations with augmented matrix
X1
X2
X4
X3
Xi
X3
X2
X4
3
1
0
+
Xi
2X1 - X2
-Sxi
2x2
-X1
X3
2x2
+
X4
x3
0 and -xi
X4
0. Thus,
94
Chapter 2
EXAMPLE4
Show that every vector v E JR3 can be written as a linear combination of the vectors
V,
[!] Ul
Y2
andV3
=
Hl
Solution: To show that every vector v E JR3 can be written as a linear combination of
the vectors v1, v2, and v3, we need to show that the system
[-31 13
-3 -:1
[-13 31 -1 ] [10 1o 0ol
1 001
3,
-2
-2
Hence, by Theorem 2.2.2, the system is consistent for all v E JR3, as required.
Lemma 1
A set of k vectors { v 1, ... , vk} in JR11 spans JR11 if and only if the rank of the coefficient
matrix of the system t1v1 +
+ tkvk
v is n.
combination of the vectors {V1, ... , vk}. That is, the system of linear equations
coefficient matrix of the system equals n (the number of equations). On the other hand,
if the rank of the coefficient matrix of the system t1v1 +
+ tkvk
v is n, then the
system is consistent for all v E JR11 by Theorem 2.2.2. Hence, Span{V 1, . .., vk}
JR".
Theorem 2
Let {V1,
11
, vk) be a set of k vectors in JR. If Span{V1, ... , vk)
95
is n. But, if we haven leading ls, then there must be at leastn columns in the matrix
to contain the leading ls. Hence, the number of columns, k, must be greater than or
equal ton.
11
, vk} in JR. is said to be linearly independent if and
is the solution t;
0 for 1
EXAMPLES
Detenn;ne whether the set
{ [: l [-il m [ m
,
dent.
Solution: Consider
" m Hl r!l l l ll
]
[;
+
+ ,,
,,
,,
Simplifying as above, this gives the homogeneous system with coefficient matrix
1
-1
5
2
1
4
-1
-3
3
Notice that we do not even need to row reduce this matrix. By Theorem 2.2.2, the
number of parameters in the general solution equals the number of variables minus
the rank of the matrix. There are four variables, but the maximum the rank can be is 3
since there are only three rows. Hence, the number of parameters is at least one, so the
system has infinitely many solutions. Therefore, the set is linearly dependent.
EXAMPLE6
Let v I
rn. [-]
i12
and v,
independent or dependent.
[: l
96
Chapter 2
EXAMPLE6
(continued)
of the corresponding system is
as in Example 2, we get
[ - : l
=
o
1
0
o
0
1
Therefore, the set is linearly independent since the system has a unique solution (the
trivial solution).
EXERCISE 2
Determine whether the set
Again, we can generalize the method used in Examples 5 and 6 to prove some
important results.
Lemma 3
A set of vectors {V 1,
Proof: If {V1,
+ tkvk
0 is k.
has a unique solution. Thus, the rank of the coefficient matrix equals the number of
unknowns k by Theorem 2.2.2.
On the other hand, if the rank of the coefficient matrix equals k, then the ho
mogeneous system has k - k
t1
Theorem 4
If {v1 ,
tk
Section 2.3
97
Bases of Subspaces
Recall from Section 1 .2 that we defined a basis of a subspace S of JR.11 to be a linearly
independent set that spans S. Thus, with our previous tools, we can now easily identify
a basis for a subspace. In particular, to show that a set 13 of vectors in JR.11 is a basis for
a subspace S, we just need to show that Span 13
EXAMPLE 7
11 s
{[il Hl 'nl}
[" il Hl [-l
+
+ 1,
JR. 3.
1,
Row reducing the coefficient matrix that corresponds to this system to RREF gives
_;2
[
1 1
1
The rank of the matrix is 3, which equals the number of rows (equations). Hence, by
Theorem 2.2.2, the system is consistent for all v
JR.3, as required.
EXAMPLES
Show that S
{[ l [l l}
_
2x2
x3
O.
Solution: We first observe that 13 is clearly linearly independent since neither vector
is a scalar multiple of the other. Thus, we need to show that every vector in the plane
can be written as a linear combination of the vectors in 13. To do this, observe that any
vector x in the plane must satisfy the condition of the plane. Hence, every vector in
the plane has the form
x
since X3
3x1
3x1 - 2x2
98
Chapter 2
EXAMPLE 8
(continued)
So, the system is consistent and hence :B is a basis for the plane.
Theorem 5
A set of vectors {V 1, . .. , v } is a basis for ]Rn if and only if the rank of the coefficient
n
matrix of tiV1++t Vn =vis n.
Proof: If {V1, ... , Vn} is a basis for JRll, then it is linearly independent. Hence, by
Lemma 3, the rank of the coefficient matrix is n.
If the rank of the coefficient matrix is n, then the set is linearly independent and
Lemma6
Suppose that S is a non-trivial subspace of JR11 and Span {vi, ... , ve} = S. If
{ z11, ... , z1d is a linearly independent set of vectors in S, then k ::; e.
Proof: Since each z1i, 1 ::; i ::; k, is a vector in S, by definition of spanning, it can be
written as a linear combination of the vectors v1, ... , Ve. We get
_,
_,
Since {V 1,
99
This gives a homogeneous system of e equations in the k unknowns t1, ..., tk. If k > e,
then this system would have a non-trivial solution, which would imply that {it 1,
,uk}
is linearly dependent. But we assumed that {it 1, ..., uk} is linearly independent, so we
must have k e.
Theorem 7
. .
k= e.
S, so e k. Therefore, e k, as required.
,
Definition
If S is a non-trivial subspace of JR.I! with a basis containing k vectors, then we say that
Dimension
So that we can talk about the basis of any subspace of IR.n, we define the empty set
...,
to be a basis for the trivial subspace {O} of JR.11 and thus say that the dimension of the
trivial vector space is 0.
EXAMPLE9
By definition, a plane in JR.11 that passes through the origin is spanned by a set of two
linearly independent vectors. Thus, such a plane is 2-dimensional since every basis of
the plane will have two vectors. This result fits with our geometric understanding of a
plane.
100
Chapter 2
PROBLEMS 2.3
Practice Problems
Al Let B
{ 1 ! , - }
or hyperplane.
(a)
-3
Span B. (a)
)
(b
A2 Let B
(c) -
{ -: , -i , _; }
0
-1
(a) _1
-7
3
(b)
0
(c)
(b
(c)
(a) Span
(b) Span
(c) Span
(d) Span
{[] [!]}
{[]}
{[i].[-!]}
{[J. [-m
for 2x,
for x1
3x2 + x, = 0
x2 + 2x3 - 2x4 = 0
(a)
{ ' ' - }
{ I : 1 }
;
-1
( b)
'
0
r xi + x,- x,=
fo
-1
{[i], []}
{[:] .[j]} {i . , !} -
(c)
1
1
u,:, =!}
{ 1 - . -n
linearly independent.
(a)
-}
4
-3
(b)
Exercises
(b)
{[iH=:J.l:J}
WH-m
(c)
(d)
101
{[Jnrni1n
{[-: ].[J []}
Homework Problems
Bl Let B
{ l. , j }
1
{:
(e) Span
B2 Let B
6
0
(b)
0
3
3
-1
(c)
2
1
{ , -: }
,
(f) Span
-1
0
( a)
(b)
0
1
(a)
4
(c )
SpanB.
2
3
(c)
-1
{ [ i J lm
{[ i] }
{ [J nJ .U] }
j i . :)
l
(a) span
(b) Span
cci span
(d) Sp
(d)
{r-;] li]}
{HrnJ}
+ X2 + X3
fo
'2X>
for 4x1+2,, - x, = a
p 1n
{- -q
,
for 2x1+3x, - 54
for X1 + 2x,
(a)
u, , n
1
0
-2
0
0
0
0
1 ' 0
-1
0
3
-1
10
3
0
1
2
-5
0
1
0
or hyperplane.
(b)
-1
-4
-2
(a)
2
-6
-1
3
3 ' -5
1
2
(b)
1
0
'
1
2
0
0
0
1
2
-2
1
-3
1
0.
102
Chapter 2
3
B7 Determine whether the given set is a basis for JR. .
linearly independent.
(a)
(b)
p , , ; }
h -1. ; }
(a)
(b)
(c)
(d)
{[-illlUl}
minim[J}
WlHHm
rnHm
Computer Problems
1
-2
-1
-1
0
6
2
3
0
-4
3
6
3
2
-1
Cl Let B
I
2
4
-2
-1
8
2
-2
and v
0
1
dependent.
0.
0
3
5
0
0
Conceptual Problems
Dl Let B
pendent.
D2 Let B
dependent.
(c) Prove that if k
n, then SpanB
JR.11 if and
Ohm's Law
103
a resistor with resistance R ohms, then the drop in the voltage across the resistor is
V = IR, measured in volts. The filament in a light bulb and the heating element of an
electrical heater are familiar examples of electrical resistors. (See Figure 2.4.4.)
VVv--
Figure 2.4.4
V =IR
IR.
signed sum of the currents entering the node is zero. (See Figure 2.4.5.)
11
Figure 2.4.5
/2
]3 - ]4 = 0
14
0.
(2) In a closed loop consisting of only resistors and an electromotive force E (for
example, E might be due to a battery), the sum of the voltage drops across resistors is
equal to E. (See Figure 2.4.6.)
Figure 2.4.6
R11
R21.
Note that we adopt the convention of drawing an arrow to show the direction of I or
of E. These arrows can be assigned arbitrarily, and then the circuit laws will determine
whether the quantity has a positive or negative sign. It is important to be consistent in
using these assigned directions when you write down Kirchhoff's law for loops.
Sometimes it is necessary to determine the current flowing in each of the loops
of a network of loops, as shown in Figure 2.4.7. (If the sources of electromotive force
are distributed in various places, it will not be sufficient to deal with the problems as a
collection of resistors "in parallel and/or in series.") In such problems, it is convenient
to introduce the idea of the "current in the loop," which will be denoted i. The true
current across any circuit element is given as the algebraic (signed) sum of the "loop
currents" flowing through that circuit element. For example, in Figure 2.4.7, the circuit
consists of four loops, and a loop current has been indicated in each loop. Across the
resistor R1 in the figure, the true current is simply the loop current i1; however, across
the resistor R2, the true current (directed from top to bottom) is i1 -i2. Similarly, across
104
Chapter 2
A resistor circuit.
Figure 2.4.7
The reason for introducing these loop currents for our present problem is that there
are fewer loop currents than there are currents through individual elements. Moreover,
Kirchhoff's law at the nodes is automatically satisfied, so we do not have to write
nearly so many equations.
To determine the currents in the loops, it is necessary to use Kirchhoff's second
law with Ohm's law describing the voltage drops across the resistors. For Figure 2.4.7,
the resulting equations for each loop are:
E,
=
E2
=
0
=
-2
Multiply out and collect terms to display the equations as a system in the variables
i1, i1, i3, and i4. The augmented matrix of the system is
R, + R1 + R4
-R2
-R4
E,
-Rs
E2
-R2
R1+R3+Rs
- R4
R4+R6+R1
-R1
-Rs
-R1
Rs+R1 +Rs
0
-E2
To determine the loop currents, this augmented matrix must be reduced to row echelon
form. There is no particular purpose in finding an explicit solution for this general
problem, and in a linear algebra course, there is no particular value in plugging in
particular values for 1, 2, and the seven resistors. Instead, the point of this example
is to show that even for a fairly simple electrical circuit with the most basic elements
(resistors), the analysis requires you to be competent in dealing with large systems of
linear equations. Systematic, efficient methods of solution are essential.
Obviously, as the number of loops in the network grows, so does the number of
variables and so does the number of equations. For larger systems, it is important to
know whether you have the correct number of equations to determine the unknowns.
Thus, the theorems in Sections 2.1 and 2.2, the idea of rank, and the idea of linear
independence are all important.
with large systems of linear equations that may arise in dealing with circuits; really
interesting examples cannot be given without assuming greater knowledge of electrical
circuits and their components.
105
Planar Trusses
It is common to use trusses, such as the one shown in Figure 2.4.8, in construction.
For example, many bridges employ some variation of this design. When designing
such structures, it is necessary to determine the axial forces in each member of the
structure (that is, the force along the long axis of the member). To keep this simple,
only two-dimensional trusses with hinged joints will be considered; it will be assumed
that any displacements of the joints under loading are small enough to be negligible.
Fv
Figure 2.4.8
s.
The external loads (such as vehicles on a bridge, or wind or waves) are assumed
to be given. T he reaction forces at the supports (shown as R1, R2, and RH in the figure)
are also external forces; these forces must have values such that the total external force
on the structure is zero. To get enough information to design a truss for a particular
application, we must determine the forces in the members under various loadings. To
illustrate the kinds of equations that arise, we shall consider only the very simple case
of a vertical force Fv acting at C and a horizontal force FH acting at E. Notice that
in this figure, the right-hand end of the truss is allowed to undergo small horizontal
displacements; it turns out that if a reaction force were applied here as well, the equa
tions would not uniquely determine all the unknown forces (the structure would be
"statically indeterminate"), and other considerations would have to be introduced.
The geometry of the truss is assumed given: here it will be assumed that the trian
gles are equilateral, with all sides equal to s metres.
First consider the equations that indicate that the total force on the structure is zero
and that the total moment about some convenient point due to those forces is zero. Note
that the axial force along the members does not appear in this first set of equations.
0, so R2
Fv
R1
Next, we consider the system of equations obtained from the fact that the sum of
the forces at each joint must be zero. T he moments are automatically zero because the
forces along the members act through the joints.
At a joint, each member at that joint exerts a force in the direction of the axis of the
member. It will be assumed that each member is in tension, so it is "pulling" away from
the joint; if it were compressed, it would be "pushing" at the joint. As indicated in the
figure, the force exerted on this joint A by the upper-left-hand member has magnitude
N1; with the conventions that forces to the right are positive and forces up are positive,
106
Chapter 2
[ ]
2
1 12
[ J.
12
On the joint B,
For each of the joints A, B, C, D, and E, there are two equations-the first for the
sum of horizontal forces and the second for the sum of the vertical forces:
Al
N1/2
A2
...f3N1/2
Bl
- Ni/2
B2
-...f3Ni/2
Cl
C2
+N2
+ N3/2 +N4
-.../3N3/2
-N2 -N3/2
.../3N3/2
Dl
D2
+ Ns/2
+ N6
+R1=
= Fv
+ ...f3Ns/2
-N4 -Ns/2
-...f3Ns/2
El
+RH=
+ N1/2
-...f3N1/2
-N6 -N1/2
E2
...f3 N1/2
=-FH
+R2=
Notice that if the reaction forces are treated as unknowns, this is a system of 10
equations in 10 unknowns. The geometry of the truss and its supports determines the
coefficient matrix of this system, and it could be shown that the system is necessarily
consistent with a unique solution. Notice also that if the horizontal force equations (Al,
Bl, Cl, D l , and El) are added together, the sum is the total horizontal force equation,
and similarly the sum of the vertical force equations is the total vertical force equation.
A suitable combination of the equations would also produce the moment equation, so
if those three equations are solved as above, then the 10 joint equations will still be a
consistent system for the remaining 7 axial force variables.
For this particular truss, the system of equations is quite easy to solve, since some
of the variables are already leading variables. For example, if FH = 0, from A2 and
!rs Fv and then B2 , C2, and D2 give NJ= Ns= !;:s Fv;
!rs Fv.
then Al and E l imply that N2 = N6 = 1v-s Fv, and Bl implies that N4 =
2
E2 it follows that N1= N1= -
Note that the members AC, BC, CD, and CE are under tension, and AB, BD, and DE
experience compression, which makes intuitive sense.
This is a particularly simple truss. In the real world, trusses often involve many
more members and use more complicated geometry; trusses may also be three
dimensional. Therefore, the systems of equations that arise may be considerably larger
and more complicated. It is also sometimes essential to introduce considerations other
than the equations of equilibrium of forces in statics. To study these questions, you
need to know the basic facts of linear algebra.
It is worth noting that in the system of equations above, each of the quantities N1,
N2, ... , N1 appears with a non-zero coefficient in only some of the equations. Since
each member touches only two joints, this sort of special structure will often occur
in the equations that arise in the analysis of trusses. A deeper knowledge of linear
algebra is important in understanding how such special features of linear equations
may be exploited to produce efficient solution methods.
Section 2.4
107
Linear Programming
Linear programming is a procedure for deciding the best way to allocate resources.
"Best" may mean fastest, most profitable, least expensive, or best by whatever criterion
is appropriate. For linear programming to be applicable, the problem must have some
special features. These will be illustrated by an example.
In a primitive economy, a man decides to earn a living by making hinges and gate
latches. He is able to obtain a supply of 25 kilograms a week of suitable metal at a
price of 2 cowrie shells per kilogram. His design requires 500 grams to make a hinge
and 250 grams to make a gate latch. With his primitive tools, he finds that he can make
a hinge in 1 hour, and it takes 3/4 hour to make a gate latch. He is willing to work
60 hours a week. The going price is 3 cowrie shells for a hinge and 2 cowrie shells
for a gate latch. How many hinges and how many gate latches should he produce each
week in order to maximize his net income?
To analyze the problem, let x be the number of hinges produced per week and
let y be the number of gate latches. Then the amount of metal used is (O.Sx + 0.25y)
kilograms. Clearly, this must be less than or equal to 25 kilograms:
O.Sx
0.25y
25
or
2x
100
lx + 0.75y
60
or
4x
3y
240
3x
2y - 2(25) cowrie shells. This is called the objective function for the problem.
The mathematical problem can now be stated as follows: Find the point (x, y) that max
imizes the objective function R(x, y)
0, y
0, 2x
100, and 4x
3y
This is a linear programming problem because it asks for the maximum (or
minimum) of a linear objective function, subject to linear constraints. It is useful to
introduce one piece of special vocabulary: the feasible set for the problem is the set
of (x, y) satisfying all of the constraints. The solution procedure relies on the fact that
the feasible set for a linear programming problem has a special kind of shape. (See
Figure 2.4.9 for the feasible set for this particular problem.) Any line that meets the
feasible set either meets the set in a single line segment or only touches the set on its
boundary. In particular, because of the way the feasible set is defined in terms of linear
inequalities, it turns out that it is impossible for one line to meet the feasible set in two
separate pieces.
For example, the shaded region if Figure 2.4.10 cannot possibly be the feasible
set for a linear programming problem, because some lines meet the region in two line
segments. (This property of feasible sets is not difficult to prove, but since this is only
a brief illustration, the proof is omitted.)
108
Chapter 2
lines
Figure 2.4.9
R(x, y)
50
constant
"-- 2x + y
x
=
100
The feasible region for the linear programming example. The grey lines
are level sets of the objective function
R.
R(x, y)
R.
2.4.9.
R(20,20)
50
x
Figure 2.4.10
The shaded region cannot be the feasible region for a linear program
ming problem because it meets a line in two segments.
Section 2.4
109
meets the feasible set in a line segment. (30,30) is also a feasible point (check),
and
R(x,y)
R(30,30)
100
also meets the feasible set in a line segment. You can tell that (30,30) is not a boundary
point of the feasible set because it satisfies all the constraints with strict inequality;
boundary points must satisfy one of the constraints with equality.
As we move further from the origin into the first quadrant, R(x,y) increases. The
biggest possible value for R(x,y) will occur at a point where the set R(x, y)
k (for
some constant k to be determined) just touches the feasible set. For larger values of
R(x,y), the set R(x,y)
k does not meet the feasible set at all, so there are no feasible
points that give such bigger values of R. The touching must occur at a vertex-that is,
at an intersection point of two of the boundary lines. (In general, the line R(x,y)
for the largest possible constant could touch the feasible set along a line segment that
makes up part of the boundary. But such a line segment has two vertices as endpoints,
so it is correct to say that the touching occurs at a vertex.)
For this particular problem, the vertices of the feasible set are easily found to be
(0,0), (50,0), and (0, 80), and the solution of the system of equations is
2x + y
4x + 3y
100
240
For this particular problem, the vertices of the feasible set are (0,0),(50, 0),(0, 80),
and (30, 80). Now compare the values of R(x,y) at all of these vertices: R(O,0),R(50,0),
R(O, 80)
the producer should make 30 hinges and 40 gate latches each week.
General Remarks
business and government. Problems such as scheduling ship transits through a canal
can be analyzed this way. Oil companies must make choices about the grades of crude
oil to use in their refineries, and about the amounts of various refined products to pro
duce. Such problems often involve tens or even hundreds of variables-and similar
numbers of constraints. The boundaries of the feasible set are hyperplanes in some
IR", where
is large. Although the basic principles of the solution method remain the
same as in this example (look for the best vertex), the problem is much more com
plicated because there are so many vertices. In fact, it is a challenge to find vertices;
simply solving all possible combinations of systems of boundary equations is not good
enough. Note in the simple two-dimensional example that the point (60,0) is the inter
section point of two of the lines (y
0 and 4x + 3y
but it is not a vertex of the feasible region because it fails to satisfy the constraint
2x + y $ 100. For higher-dimension problems, drawing pictures is not good enough,
and an organized approach is called for.
The standard method for solving linear programming problems has been the sim
plex method, which finds an initial vertex and then prescribes a method (very similar
to row reduction) for moving to another vertex, improving the value of the objective
function with each step.
Again, it has been possible to hint at major application areas for linear algebra,
but to pursue one of these would require the development of specialized mathematical
tools and information from specialized disciplines.
110
Chapter 2
PROBLEMS 2.4
Practice Problems
Al Determine the system of equations for the reaction
E1R1E>
l1
R1 E>
R3 'ji) R4
R6 'Ji) R1 RsE2
CHAPTER REVIEW
Suggestions for Student Review
1 Explain why elimination works as a method for solv
ing systems of linear equations. (Section 2.1)
2 When you row reduce an augmented matrix
[A I b] to
Chapter Review
111
_,
+ tk vk = b
(Section 2.3)
Chapter Quiz
El Determine whether the following system is consis
tent by row reducing its augmented matrix:
X2 - 2x3
+ X4 = 2
X1
+ X3
2
to 3
1
4
and 8 .
0
0
5
9
ES Detenn;ne whether
0
2
-2
3
1
3
3
4
-4
9
-6
0
3
6
-3
-1
for JR3
{[!].[i]. [!]}
;s a bas;s
-1
E3 The matrix A =
A=
b+2
2
-3
c -1
c +
tion.
IS
112
Chapter 2
Further Problems
These problems are intended to be challenging. They
may not be of interest to all students.
[A I b]
the
[A I 0]. This
echelon form.
=
(a) Let R
is
r1 3
. Show that the general so
r23
[01
[RI c]
[ 0] is
where
jJ
nents of
c,
c]
is
system
jJ
where
is expressed in terms of
as above.
=
(b) Let R
r 2
0 0 0
feneral solution
lRI O ) is
c,
and x H is
r3s
[RI c]
can be written as
where
nents
[A I b]
has n equations
Chapter Review
lA I b] to a form [c I J] such
a21
a11
-one division.
lA
cations.
IL
(3) Note that i
i=l
n(n+l)(2n+l)
deleted.
a
zi,
a11
[A I 6]
but
augmented matrix C I
substitution.
Hints
113
II
1!(1L+J)
- and i2
2
i=1
f.
procedure is roughly
requires approximatel
when n is large.
MyMathlab
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 3
Matrices, Linear
Mappings, and Inverses
CHAPTER OUTLINE
3.1
Operations on Matrices
3.2
3.3
Geometrical Transformations
3.4
3.5
3.6
Elementary Matrices
3.7
LU-Decomposition
A=
a11
a12
a lj
a111
a21
a22
a2j
a2,,
a;1
a;2
aij
a;,,
aml
am2
amj
amn
116
Chapter 3
We say that A is an m x n matrix when A has m rows and n columns. Two matrices A
and B are equal if and only if they have the same size and their corresponding entries
are equal. That is, if aiJ
biJ for 1 i
m, 1
n.
For now, we will consider only matrices whose entries aiJ are real numbers. We
will look at matrices whose entries are complex numbers in Chapter 9.
Remark
We sometimes denote the ij-th entry of a matrix A by
(A)iJ
This may seem pointless for a single matrix, but it is useful when dealing with multiple
matrices.
Several special types of matrices arise frequently in linear algebra.
Definition
An n x n matrix (where the number of rows of the matrix is equal to the number of
Square Matrix
Definition
A square matrix U is said to be upper triangular if the entries beneath the main
Upper Triangular
Lower Triangular
be lower triangular if the entries above the main diagonal are all zero-in particular,
lij
0 whenever i
EXAMPLE l
The mauices
H j ]
0 whenever i
L is said to
< j.
[ ;]
nd
[ ]
[ ]
[-; ]
md
triangular.
Definition
A matrix D that is both upper and lower triangular is called a diagonal matrix-that
Diagonal Matrix
is, diJ
0 for all i
EXAMPLE2
We
denote
the
diagonal
diag(d11, d22,
matrix
, d 1111)
[ : ]
_
[ ]
0
by
while
Section
matrix, called a
117
column matrix.
For this reason, it makes sense to define operations on matrices to match those with
vectors in JR11
Definition
(A
and the
scalar multiplication
B)iJ
of matrices by
(B)iJ
of matrices by
(tA)iJ
EXAMPLE3
(A)iJ
addition
t(A)iJ
()a [ ]+[_; ]
Solution:[]+[_; ]-[4!2) :]=[ :]
[13 -OJ5 [-21 -0]1
(b)
5 [ i]
Solution:[5 24 ]31 = [5(5(42)) 5(5(1)3)] =[0210 1 5 5]
c
( )
Note that matrix addition is defined only if the mattices are the same size.
We
now look at the properties of addition and scalar multiplication of matrices. It is very
important to notice that these are the exact same ten properties discussed in Section
1.2 for addition and scalar multiplication of vectors in JR11
118
Chapter 3
Theorem 1
(1)
(2)
(3)
(4)
A+ B is an m x n matrix
(closed under addition)
A+ B = B + A
(addition is commutative)
(A+ B) + C =A + (B + C)
(addition is associative)
There exists a matrix, denoted by Om,n , such that A+ Om,n = A; in particular,
Om,n is the m x n matrix with all entries as zero and is called the zero matrix
(zero vector)
(5) For each matrix A, there exists an m x n matrix (-A), with the property that
A +(-A) = Om,n; in particular, (-A) is defined by (-A)iJ = -(A)iJ
(additive
inverses)
(6)
(7)
(8)
(9)
(10)
sA is an m x n matrix
s(tA) (st)A
(scalar multiplication is associative)
(s+ t)A =sA + tA
(a distributive law)
s(A + B) =sA + sB
(another distributive law)
lA =A
(scalar multiplicative identity)
=
These properties follow easily from the definitions of addition and multiplication
by scalars. The proofs are left to the reader.
Since we can now compute linear combinations of matrices, it makes sense to look
at the set of all possible linear combinations of a set of matrices. And, as with vectors
in JR'\ this goes hand-in-hand with the concept of linear independence. We mimic the
n
definitions we had for vectors in ]R .
Definition
Let :B = {A 1,
, Ak} be a set of
span of :B is defined as
Span
Definition
Linearly Independent
Let :B = {A1, ... , Ae} be a set of m x n matrices. Then :B is said to be linearly inde
pendent if the only solution to the equation
Linearly Dependent
t1A1 +
is t1 =
EXAMPLE4
+ teAe = Om,n
Determine if
[ ]
is in the span of
Solution: We want to find if there are t1, t2, t3, and t4 such that
Section 3.1
Operations on Matrices
119
EXAMPLE4
Since two matrices are equal if and only if their corresponding entries are equal, this
(continued)
ti + t3 + t4
t3
t1 + t4
1
1
-2
3
EXAMPLES
ti
-2,
t1
3,
t3
3, and
t4
1.
[ ]
{[ ] [ ] , [ ]}
independent.
Row reducing the coefficient matrix of the corresponding homogeneous system gives
2
2
-1
The only solution is
EXERCISE 1
Deterrune if:S
t1
t2
t3
0, so S is linearly independent.
{[ ][ n.[ ][- ]}
[- ]
-
early independent. Is X
120
Chapter 3
EXERCISE 2
Consider 13=
{[ ], [ ], [ ], rn ]}
and show that Span 13 is the set of all 2 x 2 matrices. Compare 13 with the standard
basis for IR.4.
Let
Transpose
EXAMPLE6
Solution:
AT=
1
3
6
5
-1
3
6
5
-4
and B=
2
[ : ]
-1
T
=
4
2
A. That is,
[ ll
AT,
J T
= [1 0
and BT= _
Observe that taking the transpose of a matrix turns its rows into columns and its
columns into rows.
Theorem 2
(1)
(2)
(3)
IR., we have
Proof:
2.
3.
((sA)T);;=(sA);;=s(A)Ji=s(AT);;=(sAT);;.
Section 3.1
EXERCISE 3
Let
[ a
_
Verify that
(ATl
and
121
Operations on Matrices
(3Al
3AT.
Remark
Since we always represent a vector in JR11 as a column matrix, to represent the row of a
matrix as a vector, we will write iJT. For now, this will be our main use of the transpose;
however, it will become much more important later in the book.
Y1 = a11x1
a12X2
Y2
a22x2
a21X1
[ai
a21
[J [].
y
a12
be the coefficient matrix. Then the change of variables equations
a22
Ax,
product of A and x
(3.1)
It is instructive to rewrite these entries in the right-hand matrix as dot products. Let
a1 =
A.
[:: ]
and a2 =
[: ]
so that a
[a11
a12 and
a{
[a21
rows
of
Thus, in order for the right-hand side of the original equations to be represented cor
rectly by the matrix product Ax, the entry in the first row of Ax must be the dot product
of the first row of
Z1
b11Y1
b12Y2
Z2
b21Y1
b22Y2
y to z:
122
Chapter 3
one after the other. The values for y1 and Y2 from the first change of variables are
substituted into the second pair of equations:
= b11(a11X1
a12X2) + b12(a21X1
a22X2)
z2 = b21(a11x1
a12x2)
a22x2)
ZJ
b22(a21X1
b11a11
b21a11
+
+
b12a21
b22a21
= By
b11a12 + b12a22 x
b21a12 + b22a22
BA
b11a12
b21a12
+
+
b12a22
b22a22
(3.2)
EXAMPLE7
(BA)11 is the dot product of the first row of Band the first column of A.
(BA)12 is the dot product of the first row of Band the second column of A.
(BA)ii is the dot product of the second row of Band the first column of A.
(BA)22 is the dot product of the second row of Band the second column of A.
[ ][
2
4
3
1
5
-2
] [
1
2(5)
7
4(5)
+
+
3(-2)
1(-2)
2(1)
4(1)
] [
4
3(7)
1(7)
18
+
+
23
11
b[
to represent the i-th row of B and a1 to represent the }-th column of A. Observe from
our work above that we want the ij-th entry of BA to be the dot product of the i-th row
number of entries as a1. Hence, the number of entries in the rows of the matrix B (that
is, the number of columns of B) must be equal to the number of entries in the columns
of A (that is, the number of rows of A). We can now make a precise definition.
Definition
Matrix Multiplication
a1,
bf,
(BA)iJ = b; ai
-;
Remark
If B is an m x n matrix and A is a p x q matrix, then BA is defined only if n = p.
Moreover, if n
More simply stated, multiplication of two matrices can be performed only if the
number of columns in the first matrix is equal the number of rows in the second.
EXAMPLES
(a)
2
4
3
-1
Solution:
(b)
0
2
2
1
l
-1
3
-1
3
1
1
2
1
-1
3
1
2
0
2
0
0
2
3
s
1
2
3
s
9
15
13
3
H - m !l
[- i [ i [ i
] [ ]
Solution:
EXAMPLE9
123
3
3
l
-0
-3
is not defined because the first matrix has two columns and the sec-
EXERCISE4
Let A
not defined.
(a) AB
EXAMPLE 10
LetA
-]
(b) BA
andB
Solution: ATB
[I
and B
[ ].
(c)ATA
m
2
(d) BET
CompuATB
[]
=
[1(6)
and'!
2(5)
3(4)]
[28]
28,
This matches the result in Example 10. This result should not be surprising since we
124
Chapter 3
have defined matrix multiplication in terms of the dot product. More generally, for any
x,y
JRn, we have
XTy
X y
where we interpret the 1 x 1 matrix on the right-hand side as a scalar. This formula
will be used frequently later in the book.
Defining matrix multiplication with the dot product fits our view that the rows of
the coefficient matrix of a system of linear equations are the coefficients from each
equation. We now look at how we could define matrix multiplication by using our
alternate view of the coefficient matrix; in that case, the columns of the coefficient
matrix are the coefficients of each variable.
Observe that we can write equations (3. 1 ) as
a11x1 +a12x2
a11X1 +a12X2
Ax=
] [ ] [ J
a11
a11
xi +
a12
a12
x2
That is, we can view Ax as giving a linear combination of the columns of A. So, for
an
Xt
:
Xn
JR", we have
BA
[Ba1
Ba2
(3.3)
Both interpretations of matrix multiplication will be very useful, so it is important
to know and understand both of them.
Remark
We now see that linear combinations of vectors (and hence concepts such as spanning
and linear independence), solving systems of Linear equations, and matrix multiplica
tion are all closely tied together. We will continue to see these connections later in this
chapter and throughout the book.
Some calculations
11
I ak
k=l
a1 +a1 ++a,,
blj
matrix with j-th column b1
:
bnj
(AB)ij
ai bj
_,
_,
->
a T bj
i
[ail
11
_,
A )ik( B)kj
L)
k=l
We can use this notation to help prove some properties of matrix multiplication.
Section 3.1
Theorem 3
Operations on Matrices
125
If A, B, and C are matrices of the correct size so that the required products are
defined, and t E JR, then
(2)
(A + B)C= AC+ BC
T
T T
(5) (AB) =B A
Each of these properties follows easily from the definition of matrix multiplica
tion and properties of summation notation. However, the proofs are not particularly
illuminating and so are omitted.
AB -:: BA. In fact, if BA is defined, it is not necessarily true that AB is even defined.
For example, if B is
and A is
2 x 3, then BA is defined,
even if both AB and BA are defined, they are usually not equal. AB = BA is true only
in very special circumstances.
EXAMPLE 11
[ 3]
[ -i H-;
[ ;][
Show that if A =
Solution: AB=
but
BA= _
[ 1]
] [ =] ,
i] [: -]
5
and B= _2
=
7 , then AB
-:: BA.
The cancellation law is almost never valid for matrix multiplication: Thus, if
EXAMPLE 12
Let A=
[ ] [; ]
[; n
[ ][; ] [ ] [ ] [; ]
.B =
AB=
.and C =
Then,
=AC
Remark
The fact that we do not have the cancellation law for matrix multiplication comes from
the fact that we do not have division for matrices.
We must distinguish carefully between a general cancellation law and the follow
ing theorem, which we will use many times.
126
Chapter 3
Theorem4
If
A and B are m xn
!Rn, then
A=B.
Note that it is the assumption that equality holds for every x that distinguishes this
from a cancellation law.
Identity Matrix
We have seen that the zero matrix Om,n is the additive identity for addition of m x n
matrices. However, since we also have multiplication of matrices, it is important to
determine if we have a multiplicative identity. If we do, we need to determine what
A and a
I such that Al=A=IA, both A and I must ben xn matrices. The multiplicative
identity I is then x n matrix that has this property for alln xn matrices A.
the multiplicative identity is. First, we observe that for there to exist a matrix
matrix
I=
[; {]
I,
such that
Al=A.
[; ].
We want to
A=
af + bh
cf + dh
ae
a=ae + bg, b=af + bh, c=ce + dg, and d=cf+ dh. Although
this system of equations is not linear, it is still easy to solve. We find that we must have
e = 1 =h and
f=g=0. Thus,
I=
It is easy to verify that
I also
diag(l , 1)
[ ]=
satisfies
IA = A.
Hence,
Definition
Then xn matrix
Identity Matrix
EXAMPLE 13
I=diag(l, 1, 1)=
I=diag(l, 1, 1,
1)
[ ]
0
1
0
0
0
0
0
0
0
0
0
0
0
Section 3.1
Operations on Matrices
127
Remarks
1. In general, the size of I (the value of n) is clear from context. However, in some
cases, we stress the size of the identity matrix by denoting it by
and
Im
2. The columns of the identity matrix should seem familiar. If {ei, ... , en} is the
standard basis for
Theorem 5
If
A is any m x n matrix,
n
,
then
then
I,.,zA =A=Ain.
You are asked to prove this theorem in Problem D2. Note that it immediately
implies that
In
Block Multiplication
Observe that in our second interpretation of matrix multiplication, equation (3.3), we
simple example of block multiplication. Observe that we could also regard the rows
of
which have been partitioned into blocks. In addition to clarifying the meaning of some
calculations, block multiplication is used in organizing calculations with very large
matrices.
Roughly speaking, as long as the sizes of the blocks are chosen so that the products
of the blocks are defined and fit together as required, block multiplication is defined
by an extension of the usual rules of matrix multiplication. We will demonstrate this
with an example.
EXAMPLE 14
Suppose that
A is
anm x n matrix,
is an n x p matrix, and
A and B
are partitioned
A=
Say that
A1
is r x n so that
A2
[1],
is (m
r) x n, while Bi is n x q and B2 is n x (p
q).
128
Chapter 3
EXAMPLE 14
(continued)
Observe that all the products are defined and the size of the resulting matrix is
m x
p,
as desired.
PROBLEMS 3.1
Practice Problems
3] [-3 ]
(-3)[; -i
[ ;] 3 [! -J
r- !H- ; -;J
[ 3 i] [ i
H -!l [ : ; _;J
[ ; ;J [i j]
[
(b)
2
4
-2
1
(c)
-1
-4
-5
A=
(a)
-2
(c)
AS
(b)
A=
C=
B=
-
H -]
[-i l C=[ =;
n B=[=
[- 3;].
1
[l j].
3
!J
AB
DC
BA
(c) AC
CD
(f) CTD
(g) Verify that A(BC)=(AB)C
(h) Verify that (ABf=BTAT
(d)
[- n
D=
; -i] .c=
do not exist.
(a)
A3 Check that
[ -i ; ] .B=[ -;
2
that
B=
UtA=
and
(d)
check whether
B=
A=
(b)
-1
B,
[ ], 1-3 =]
[; - ; ] n =]
-2
(b)
and
(a)
(b)
(e)
A6 Let
[ -i !] [;], [ i],
[-:]
A=
-1
andZ=
(a) Determine
x=
y=
-1
DTC
Section 3.1
1 13 -1ol.
-1 1
Ul
U
(d) =
(c)b= HI
" Ul
Calculate the following products. (b) [2 6] [ ]
(a)[-] [2 6]
(d) [s 4 -31[-]
(c) Hl [s 4 -3]
A9
AlO
All
A12
Homework Problems
Bl
B2
-9
6
3
-2
4
4
5
3
[ -4 1 1-2 1 ] 3
-3 2
= [-42 31 ][-26 34]+ [-42 1 ] [-31 32]
Determine if[ -]is in the span of
129
()
A8
Exercises
B3
a1
a11
(d) [ -][-: -]
Check that A(B+ =AB+ AC and that A(3B) =
3(AB) for the given matrices.
(a) A=[_; _nB=[ - -n
= r-; -J
(b) A=u ] B [; -n [- -1
For the given matrices A and B, check whether
B and
If so, check that
(AA ++ B/
=ATAB+ BareT anddefined.
(AB/=BTAT.
(a) A=[_; : :].s=H j]
C)
B4
a;.
IR.11
Chapter 3
130
H ! J.B=[! ! J
Let = H =H B= [! -].
1 3 -5
[=; land - ; 1
(b)
BS
A=
BS
-1
B9
AC
(f) CDT
(BC)A
T AT
DA
CA
(ABl
Der
B6
.x
(a) Determine
and
(b) Use the resultof(a) to determine [-2 -2-1 -3=]
Let ; - =] and consider as a lin[-2 -4 -3
ear combination of columns of to determine if
b where
(a)b=[=!l
(b)b=nl
Ax, Ay,
BlO
At
B7
Ax
Ax=
Computer Problems
(c)b=rn
Calculate the following products.
(b) [-2 4] [;J
(a) [;J [-2 4]
(d) [-3 2) [!]
(c) [!] [-3 2]
Verify
the following
multiplication
calculating
both sidescaseof oftheblock
equation
and comparby
ing.
43
25
61
-1 3
= [l3 -1]2 [42 53] [-2 4]3 [-16 1]3
Bll
Determine if[
-] is in the span of
-1
-4.94
2.29J
4.25
3.38
-3.73]
131
Conceptual Problems
Dl Prove T heorem 4, using the following hints. To
prove A
B, prove that A
= 0111,11; note
that Ax = Bx for every x E JR" if and only if
1
(A - B)x = 0 for every x E JR. 1 Now, suppose that
-
AB
BA + B2 = 02 2.
Matrix Mappings
Using the rule for matrix multiplication introduced in the preceding section, observe
that for any m x n matrix A and vector x E JR", the product Ax is a vector in JRm. We
see that this is behaving like a function whose domain is JR.11 and whose codomain is
JR.11!.
Definition
Matrix Mapping
For any mxn matrix A, we define a function fA : JR.11 JR.111 called the matrix mapping,
corresponding to A by
fACx) =Ax
Remark
Although a matrix mapping sends vectors to vectors, it is much more common to view
functions as mapping points to points. Thus, when dealing with mappings in this text,
132
Chapter 3
EXAMPLE 1
LetA
[-i ]
[[] tJ
=
Solution: We have
EXERCISE 1
LetA
[ n
FindfA(l,o),JAco,1),andfAC2,3).
_
W hat is the relationship between the value of fA(2,3) and the values of fA(1, 0) and
=
fA(O,1)?
EXERCISE 2
[ ; ]
-
LetA
EXAMPLE2
Let A
a1 i
a21
a12
and find the values of fA(l,0), fA(O, 1),andfA(X1,x2).
a22
Section 3.2
EXAMPLE2
133
Solution: We have
(continued)
Then we get
We can now clearly see the relationship between the image of the standard basis
Theorem 1
m x n
A.
'en be the standard basis vectors of IR.11' let A be an m x n matrix, and let
:
11
IR.
-t
IR.m
be the corresponding matrix mapping. Then, for any vector 1 E IR.11,
fA
we have
A = [a'.1 a'.2
a'.11 ] . Since e;
fA(e;) =Ae; =a'.;. Thus, we have
Proof: Let
we get
as required.
Since the images of the standard basis vectors are just the columns of
that the image of any vector
1 E IR."
A,
we see
A.
This should not be surprising as this is one of our interpretations of matrix multiplica
tion. However, it does make us think about how a matrix mapping will affect a linear
combination of vectors in IR.11 For simplicity, we look at how it affects these separately.
Theorem 2
for any
(L l )
(L2)
E IR.11
and
t ER
/A : IR.11 -t IR.m.
Then,
we have
134
Chapter 3
Linear Mappings
Definition
Linear !\lapping
A function L : JRn --+ JRm is called a linear mapping (or linear transformation) if for
E JRn and t E JR it satisfies the following properties:
every x, y
(Ll ) L(x
(L2) L( tx)
Definition
Linear Operator
y)
=
L(x)
L(Y)
tL(x)
A linear operator is a linear mapping whose domain and codomain are the same.
Theorem 2 shows that every matrix mapping is a linear mapping. In Section 1.4,
we showed that projil and perpil are linear operators from JRn to JRn.
Remarks
1. L inear transformation and linear mapping mean exactly the same thing. Some
people prefer one or the other, but we shall use both.
2. Since a linear operator L has the same domain and codomain, we often speak
of a linear operator Lon JR11 to indicate that Lis a linear mapping from JR11 to
JR".
3. For the time being, we have defined only linear mappings whose domain is JR"
and whose codomain is JRm. In Chapter 4, we will look at other sets that can be
the domain and/or codomain of linear mappings.
EXAMPLE3
2
Show that the mapping f : IR
linear.
--+
( 2x1
x2, -3x1
5x2) is
Solution: To prove that it is linear, we must show that it preserves both addition and
f(Y
ZJ
f(y1
[
[
Z1,Y2
Z2)
2y1 + Y2
-3y1 + Sy2
f(Y)
f(ZJ
] [
+
2 z1
-3z1
Z2
Sz2
EXAMPLE3
Thus,
(continued)
f(tY)
Hence,
135
f(ty1' ty2)
[ 2(ty1) +(ty2) ]
-3(ty1)+S(ty2)
[ 2y1 + Y2 ]
t
-3y1 + 5y2
tfCY)
In the previous solution, notice that the proofs for addition and scalar multipli
cation are very similar. A natural question to ask is whether we can combine these
into one step. The answer is
yes!
statement:
L :
scalar
JR.I!
E
JR.,
L(t1 + y)
tL(1) + LCY)
The proof that the definitions are equivalent is left for Problem D 1.
EXAMPLE4
f:
JR.3
JC1+Sf)
JR. defined by
111+5'11
/(1)
11111 is linear.
and
JC1) +!CY)
11111 +115'11
f(X + j1)
f( l, l, 0)
[l
=
and j1
..f5.
[n
then
136
Chapter 3
EXAMPLE4
but
(continued)
1
/( )
Thus,
f(x + y)
f(J)
[i] [ ! ]
+
I +
f(x) + f(Y) for any pair of vectors , y in JR.3, hence f is not linear.
1
Note that one counterexample is enough to disqualify a mapping from being linear.
EXERCISE 3
JR.2
---7
g : JR.2
---7
(a) f
(b)
(XT, X1
x2)
(x2, x1 - x2)
f(x1, x2)
f from
Example
3,
If this is to be a matrix mapping, then from our work with matrix mappings, we know
that the columns of the matrix must be the images of the standard basis vectors. We
have
A
f(l, 0)
[ ; ]
_
[ ;l f(O,
_
1)
[n
Theorem 3
If L : JR.11
---7
L(1)
JR.11 we have
+ + ..
+ + ... +
L(x1e1
X1L(e1)
137
[LCe1)
x2e2
Xne11)
x2L(e2)
L(e2)
...
X11L(e11)
L(e11)]
[1]
Xn
[L]1
as required.
Remarks
1.
2.
2,
The matrix [L] determined in this theorem depends on the standard basis. Con
sequently, the resulting matrix is often called the standard matrix of the linear
mapping. Later, we shall see that we can use other bases and get different ma
trices corresponding to the same linear mapping.
EXAMPLES
Let v
[!]
and i1
rn
JR.2
column of the matrix is the image of the first standard basis vector under prok Using
the methods of Section
1.4,
proJv e
we get
=
e1 v
v
llVll2
Similarly, the second column is the image of the second basis vector:
.
rO v
P J e2
e2. v
v
11v112
[proJv]
[proJv. -;
ei
. -; ]
proJv e2
[9/25 12/25]
12/25 16/25
Therefore, we have
. -;
proJv u
.]
[ proJv
-;
u =
138
Chapter 3
EXAMPLE6
Let
G : JR3
JR
--+
be defined by
(xi, x2).
Prove that
G(ry
tJ
Hence,
G.
G(tyi
Zi, ty2
Z2, ty3
JR,
we have
Z3)
[ : ] [] []
[] []
=
tG(Y)
G(tJ
is linear. Thus, we can apply Theorem 3 to find its standard matrix. The
G(l, 0,0)
[] ,
G(O, 1,0)
[],
G(O, 0, 1)
[]
Thus, we have
[G]
G(O,
[co. 0, 0)
[G]
1, 0)
G(O, 0,
1) =
[ ]
0
using the image of the standard basis vectors and then said that because
is a matrix mapping, it is linear? No! You must always check the conditions of the
can be constructed
from the images of the standard basis vectors. The converse is not true!
For example, consider the mapping
dard basis vectors are
[ l
f(l, 1)
f(l, 0)
[]
and
f(xi, x2)
f(O,
1) =
(xix2, 0).
ml
But this matrix does not represent the mapping! In particular, observe that
=
[] [ ] [ ] []
but
the images of the standard basis vectors, it does not imply that the matrix will represent
that mapping, unless we already know the mapping is linear.
EXERCISE4
2
JR be defined by H(xi, x2, x3, x4)
and find the standard matrix of H.
Let
H : JR
--+
(x3
x4, xi).
Prove that
H is linear
139
Definition
Operations on Linear
"
Let L and M be linear mappings from JR to JRm and let t
(L + M) to be the mapping from JR" to JRm such that
JR be a scalar. We define
Mappings
(L + M)(x)
L(x) + M(x)
1
We define (tL) to be the mapping from JR 1 to JRm such that
(tL)(x)
Theorem 4
tL(x)
"
If Land M are linear mappings from JR to JRm and t
linear mappings.
Proof: Let 1, y
(tL)(sx + y)
JR11 and s
R Then,
tL(sx + y)
t[sL(x) + L(J)]
stL(x) + tL(J)
s(tL)(x) + (tL)(J)
The properties we proved in Theorem 4 should seem familiar. They are properties
(1) and (6) of Theorem 1.2.1 and Theorem 3.1.1. That is, the set of all possible linear
mappings from JR" to JRm is closed under scalar multiplication and addition. It can be
shown that the other eight properties in these theorems also hold.
Definition
Composition of Linear
Mappings
1
Let L : JR1 4 ]Rm and M : ]Rm
11
JR JRP is defined by
for all x
(M 0 L)(x)
L :
M (L(x))
JR".
Note that the definition makes sense only if the domain of the second map M
contains the codomain of the first map L, as we are evaluating M at L(x). Moreover,
observe that the order of the mappings in the definition is important.
EXERCISE 5
1
Prove that if L : JR 1
linear mapping.
L is a
Since compositions and linear combinations of linear mappings are linear map
pings, it is natural to ask about the standard matrix of these new linear mappings.
140
Chapter 3
Theorem 5
Let
JR11 - Rm, M
JR111
Then,
[tL]= t[L],
[NoL]= [N][L]
Proof: We will prove [tL] = t[L] and [No L] = [N][L]. You are asked to prove that
[L + M]= [L] + [M] in Problem D2.
Observe that for any x E JR11, we have
[tLJx= (tL)(x)= tL(x)= t[LJx
[tL]= t[L] by Theorem 3.1.4.
[NJ is p xm and [L] ism x n, the matrix product [N][L]
is defined. For any 1 E JR11,
Hence,
[N
Thus,
EXAMPLE 7
Let Land
Determine [MoL].
Solution: We find that
L(l,0)=
[]
L(0,1)=
Thus,
[ ]
M(l,0)=
[ ]
-
M(0,1)=
[ ]
-
[ -][ ] [ ]
[MoL]= -
In Example 7,
JR".
Moreover, the standard matrix of the mapping is the identity matrix. We thus make the
following definition.
Definition
Identity Mapping
Id(x) =
is called the
identity mapping.
141
Remark
Since the mappings L and M in Example 7 satisfy Mo L= Id, they are called inverses,
as are the matrices [M] and [L]. We will look at inverse mappings and matrices in
Section 3.5.
PROBLEMS 3.2
Practice Problems
-2
3
3
0
Al Let A=
and let fA be the corresponding
5
4 -6
matrix mapping.
(a) Determine the domain and codomain of fA.
(b) Determine /A(2,-5) and /A(-3, 4).
(c) Find the images of the standard basis vectors
for the domain under fA.
(d) Determine fA(x).
(e) Check your answers in (c) and (d) by calculat
ing [/A(x)] using Theorem 3.
A2 Let A=
[S]=[
-
;],
l
[T]=[
2
2
2
-1
]
3
(b)
(c)
(d)
(e)
(f)
[S]=[
;],[T]=
-l
-4
[L]
HH
2
[N] =
[M]
[i - -l
and
142
Chapter 3
(b) MoL
(d) NoL
(f) NoM
(c) Lo N
(e) Mo N
AS Let v=
[-n
A9 Let v =
A 10 Let V
[! l
[ jl
Homework Problems
Bl Let A =
[ =
-1
0 2 3
B2 Let A =
7 9 and letfA be the corresponding
5
2 4 8
matrix mapping.
(a) Determine the domain and codomain offA
(b) DeterminefA(-4,2, 1) andfA(3,-3,2).
(c) Find the images of the standard basis vectors
for the domain underfA.
(d) Determine fA(x).
2x3
X4,3x2 + x3)
(c) M(x1,X2,x3) =(x3 -x1,0,Sx1
+
x2)
[S]=
[ -]
and
[ T]=
[! -]
[S]=
-3
- ,[ T]= [-
-2
un
[M]=
[i
and
Section 3.3
2
3
-3
[N] =
0
-4
(a) LoM
(c) Lo N
(e) Mo N
B8 Let v
(b) MoL
(d) NoL
(f) NoM
[-l
B9 Let v =
B 10 Let ii=
Geometrical Transformations
[ ].
_
[-l].
[-H
143
Conceptual Problems
Dl Let L
R.11
D2 Let L
R.11
R.m and M
R.11
R.m be linear
M]
D3 Let v
[L]+ [M].
E
DOTv by DOTv 1 = v
R2
through angle () to the image Re(1). See Figure 3.3.1. Is Re linear? Since a rotation
does not change lengths, we get the same result if we first multiply a vector 1 by
a scalar t and then rotate through angle (), or if we first rotate through () and then
multiply by t. Thus, Re(t1) = tRe(1) for any t E R. and any 1 E R2. Since the shape
of a parallelogram is not altered by a rotation, the picture for the parallelogram rule of
addition should be unchanged under a rotation, so Re(1 + y) = Re(1) + Re(Y). Thus,
144
Chapter 3
Re(1)
Figure 3.3.1
Re is linear.
a matrix mapping.
Assuming that
calculate
[Re].
Re
From Figure
3.3.2,
we see that
R e(1 O)
Re(O, l)
[ ]
[ ]
c se
sme
'
Hence,
[Re]
-sine
cose
c se
-sine
sme
cose
Re(1)
[XXit
[Re]1 that
c se -
sine
Sln 8 +
COS 8
XX22 ]
X2
Re(l, 0)
(0, 1)
Re(O, 1)
(cose, sin 8)
(-sine, cos 8)
(1, 0)
Figure 3.3.2
EXAMPLE 1
- and sin
[R2,,.13]
[
=
2rr/3?
f,
-1 /2
../3/2
v'3/2
-1/2
X1
EXERCISE 1
Determine
[Rrr14]
Rrr14(1, 1).
145
R, R(O,
noted
then
0, 1)
(0, 0, 1). Together with the previous case, this tells us that
the matrix of this rotation is
=
[R] [
=
OJ
cose
- sine
sine
0
cose
These ideas can be adapted to give rotations about the other coordinate axes. Is it
3
possible to determine the matrix of a rotation about an arbitrary axis in JR ? We shall
see how to do this in Chapter 7.
Figure 3.3.3
Stretches Imagine that all lengths in the x1 -direction in the plane are stretched by a
scalar factor t > 0, while lengths in the x2-direction are left unchanged (Figure 3.3.4).
This linear transformation, called a "stretch by factor tin the x1 -direction," has matrix
[ n
(If t < 1, you might prefer to call this a shrink.) It should be obvious that
stretches can also be defined in the x2-direction and in higher dimensions. Stretches
are important in understanding the deformation of solids.
[ l
vectors in all directions by the same factor. Thus, for example, a circle of radius 1
centred at the origin is mapped to a circle of radius t at the origin. If 0 < t < 1, such a
transformation is called a contraction; if t > l, it is a dilation.
146
Chapter 3
Figure 3.3.4
Shears
x1 -direction.
lelogram, as shown in Figure 3.3.5. The change can be described by the transformation
1
2
2
: IR IR , such that S (2, 0)
(2, 0) and S (0, 1) = ( s, ). Although the deformation
of a real solid may be more complicated, it is usual to assume that the transformation S
IS
[ ]
1
0
1 .
(5, 1)
(2 + 5, 1)
(2, 1)
...(0, 1) ----....,.-----.---""'7
(2,0)
Figure 3.3.5
x1
by amounts.
2
Let R: JR.2 JR. be a reflection in the x1 -axis (see Figure 3.3.6). Then each vector
corresponding to a point above the axis is mapped by R to the mirror image vector
below. Hence, R(xi, x2) = (x1, -x2), and it follows that this transformation is linear
[ OJ
1
.
. h matnx
wit
.
.
.
.
he x2-ax1s has the matnx
1 . smu'larly, a re ftect1on mt
[- OJ
l
0
1 .
Next, consider the reflection T : JR3 JR3 that reflects in the x1x2-plane (that is,
the plane x3 = 0). Points above the plane are reflected to points below the plane. The
o
1
0
o
0 .
-1
Figure 3.3.6
EXERCISE 2
x1
147
-axis.
W rite the matrices for the reflections in the other two coordinate planes in JR.3.
General Reflections
planes in JR.3 that pass through the origin. Reflections in lines or planes not containing
the origin involve translations (which are not linear) as well as linear mappings.
Consider the plane in JR.3 with equation it 1
proj,7, a reflection in the plane with normal vector it will be denoted refl;t. If a vector
corresponds to a point P that does not lie in the plane, its image under refl;t is the
vector that corresponds to the point on the opposite side of the plane, lying on a line
through P perpendicular to the plane of reflection, at the same distance from the plane
as P. Figure 3.3.7 shows reflection in a line. From the figure, we see that
reft;t a
Figure 3.3.7
148
Chapter 3
It is important to note that refl,1 is a reflection with normal vector it. The calcu
lations for reflection in a line in IR2 are similar to those for a plane, provided that the
equation of the line is given in scalar form it 1 0. If the vector equation of the line
is given as 1
tJ, then either we must find a normal vector it and proceed as above,
or, in terms of the direction vector J, the reflection will map jJ to (jJ 2 perpJ jJ).
=
EXAMPLE2
Consider a refiection refl,1
1!.3
[-i].
Hence,
Note that we could have also computed [refl,1] in the following way. The equation
for refl;t(p) can be written as
refl;t(P)
Id(jJ)
2 proj;t(P )
Thus,
PROBLEMS 3.3
Practice Problems
A 1 Determine the matrices of the rotations in the plane
through the following angles.
(b) 1T
can
(d) 6f'
(c) -
3x2
=
0.
x2.
S (x1,x2,x3) (x1,x2,X3
matrix of P o S .
=
(a) Xt
X2
X3
0
=
149
2
(c) Can you define a shear T : JR
To P
2
JR such that
(d) Let Q : JR
Homework Problems
Bl Determine the matrices of the rotations in the plane
through the following angles.
(c)
(a)-
(b)
(d)
(a) Xt - 3x2
-n
(b) 2X1
4x2
0.
=
- x3
X2- X3
0
3
inj(x1,x2,x3)
0.
(xt
Conceptual Problems
Dl Verify that for rotations in the plane [Ra o Re]
[Ra][Re] = [Ra+e].
D2 In Problem A3, [R]
4/5
_315
-3/5
_415 and [S]
-3/5 4/5
.
are refiect10n matrices. Calculate
415 315
[Ro S] and verify that it can be identified as the
matrix of a rotation. Determine the angle of the
rotation. Draw a picture illustrating how the com
position of these reflections is a rotation.
150
Chapter 3
S.
Definition
Solution Space
Section 3.4
EXAMPLE 1
151
or
EXERCISE 1
Let A=
[ ! ]
Notice that in both of these problems, the solution space is displayed automatically
as the span of a set of linearly independent vectors.
For the linear mapping L with standard matrix A = [L], we see that L(x) =Ax, by
definition of the standard matrix. Hence, the vectors x such that L(x) = 0 are exactly
the same as the vectors satisfying Ax = 0. Thus, the set of all vectors x such that
L(x) = 0 also forms a subspace of JR11
Definition
Nullspace
The nullspace of a linear mapping L is the set of all vectors whose image under L is
the zero vector 0. We write
Null(L) = {x E JR" I L(x) = O}
Remark
Let V=
[ l
JR" I
R3 .
Solution:
Null(proj,) =
{[:]
R3 I 2x1 - x,
3x3 = 0
152
Chapter 3
EXAMPLE3
Let L : JR.
--+
3
JR. be defined by L(x1, x2)
Solution: We have
[]
have
2x1 - X2
X1 + X2
0
0
0. Thus Null(L)
{0}.
EXERCISE 2
To match our work with linear mappings, we make the following definition for
matrices.
Definition
Nullspace
Null(A)
{x
e JR.11 J Ax
O}
Solution Set of Ax
--+
JR.m, Null(L)
Null([L]).
b, b t:- 0 and
compare this solution set with the solution space for the corresponding homogeneous
system Ax
EXAMPLE4
0 (that is, the system with the same coefficient matrix A).
EXERCISE3
Let A
[ ! l
and
+ 2x2 - 3x3
[n
5.
5 is
Section 3.4 Special Subspaces for Systems and Mappings : Rank Theorem
153
Observe that in Example 4 and Exercise 3, the general solution is obtained from
the solution space of the corresponding homogeneous problem (Example
cise
Theorem 2
and Exer
1,
jJ
b, b * 0.
v is any other solution of the same system, then A(jJ -v) = 0, so that jJ -v
is a solution of the corresponding homogeneous system Ax = 0.
(2) If h is any solution of the corresponding system Ax = 0, then jJ + h is a
solution of the system Ax = b.
Let
( 1) If
Proof:
Range
EXAMPLES
Let ii =
[:]
JR11}
multiples of v. On the other hand, the range of perpil is the set of all vectors orthogonal
to v. Note that in each of these cases, the range is a subset of the codomain.
EXAMPLE6
If Lis a rotation, reflection, contraction, or dilation in JR3, then, because of the geome
try of the mapping, it is easy to see that the range of Lis all of JR3.
154
Chapter 3
EXAMPLE 7
Let L
[ ]
2x1
L(x)=
x2
X1 + X2
EXERCISE4
Let L : R3
{[] nl}
Range(L).
It is natural to ask whether the range of L can easily be described in terms of the
matrix A of L. Observe that
ctn]
L(x)=Ax= a1
X1
:
Xn
= X1G1 +
"
X11Gn
Thus, the image of x under L is a linear combination of the columns of the matrix A.
Definition
Columnspace
The columnspace of an m x
Range(L)=Col(A).
EXAMPLE8
Let A= 1
2
3]
1
and B=
co1cAJ =span
[ -H
1
Then
and
coI(BJ =span
{[] [-l]}
Section 3.4 Special Subspaces for Systems and Mappings: Rank Theorem
EXAMPLE9
If L is the mapping with standa<d matrix A =
[ n
EXERCISE 5
155
then
{[] [i]}
Find the standard matrix A of L(x1, x2, x3) = (x1 - x2, -2x1 + 2x2 + x3) and show that
Range(L) = Col(A).
The range of a linear mapping L with standard matrix A is also related to the
system of equations Ax =
Theorem 3
b =L(x)
=Ax.
[ n
EXAMPLE 10
Suppose that L is a linear mapping with matrix A
C=
ul
and
J=
Determine whether
J is in the range of
matrix is the same for the two systems, we can answer both questions simultaneously
by row reducing the doubly augmented matrix
1
2
1
1
-1
][
I c I J ]:
1
-]
156
Chapter 3
EXAMPLE 10
(continued)
I J J shows that Ax
J is
Rowspace of A
The idea of the rowspace of a matrix A is similar to the idea of the columnspace.
Definition
Rows pace
EXAMPLE 11
Let A =
[l
2
1]
3
_
[1 ]
m] .[_m
3
and B =
Row(A) =Span
. Then
and
Row(B) =Span
{ [] [_i] . [])
.
We now prove an important result about the rowspaces of row equivalent matrices.
Theorem 4
Proof: We will show that applying each of the three elementary row operations does
_,T
bm .
= {c1a1 +
-+
-+
+ ci(tai) +
-+
cmam I ci
JR}
=Span{a1,
, am) = Row(A)
Section 3.4 Special Subspaces for Systems and Mappings: Rank Theorem
157
Now, suppose that B is obtained from A by adding t times the i-th row to the }-th
row. Then,
Row(B)
Span{a1, . .. , aj+ta;,
... , a111}
{c1a1 +...+c;a;+...+Cj(aj+ta;)+...+Cmam IC; E JR}
{c1a1+ +(c;+cjl)a;+ +cjaj+ +cmam I c; E JR}
Span{a1,
.. ,a111}
Row(A)
in the set as possible. We saw this in Section 1.2 when the set was linearly independent.
Thus, we defined a basis for a subspace S to be a linearly independent spanning set.
Moreover in Section 2.3, we defined the dimension of the subspace to be the number
of vectors in any basis.
Theorem 5
Let B be the reduced row echelon form of an m x n matrix A. Then the non-zero
rows of B form a basis for Row(A), and hence the dimension of Row(A) equals the
rank of A.
form a spanning set for the rowspace of A. Thus, we just need to prove that these non
zero rows are linearly independent. Suppose that B has r non-zero rows
consider
bf, .. ., b and
(3.-+)
Observe that if B has a leading I in the }-th row, then by definition of reduced row
echelon form, all other rows must have a zero as their }-th coordinate. Hence, we must
have
tj
0 in (3.4). It follows that the rows are linearly independent and thus form a
EXAMPLE 12
Let A
[l : ]
-
ll H ! -rl
l
158
Chapter 3
EXAMPLE 12
(continued)
EXERCISE 6
{[l .U]}
Let A
2
=
space of a matrix A? It is remarkable that the same row reduction that gives the basis
for the rowspace of A also indicates how to find a basis for the columnspace of A.
However, the method is more subtle and requires a little more attention.
Again, let B be the reduced row echelon form of A. Recall that the whole point of
the method of row reduction is that a vector x satisfies Ax
Bx
0.
b1,
columns of B, then
if and only if
So, any statement about the linear dependence of the columns of A is true if and only
if the same statement is true for the corresponding columns of B.
Theorem 6
Suppose that B is the reduced row echelon form of A. Then the columns of A that
correspond to the columns of B with leading ls form a basis of the columnspace of
A. Hence, the dimension of the columnspace equals the rank of A.
m x n
matrix B
[E1
set of columns containing leading ls is linearly independent as they are standard basis
vectors from JR"'. Moreover, if
then, by definition of the reduced row echelon form, it can be written as a linear combi
nation of the columns containing leading 1 s. Therefore, from our argument above, the
corresponding column a; in A can be written as a linear combination of the columns
in A that correspond to the columns containing leading l s in B. Thus, we can remove
this column from the spanning set without changing the set it spans by Theorem 1.2.3.
We continue to do this until we have removed all the columns of A that correspond
to columns of B that do not have leading l s, and we get a basis for the columnspace
A.
Section 3.4
EXAMPLE 13
2
LetA =
159
{ }.
The first, third, and fourth columns of the reduced row echelon form ofA are linearly
independent. Therefore, by Theorem 2, the first, third, and fourth columns of matrixA
Notice in Example 13 that every vector in the columnspace of the reduced row
echelon form of A has a last coordinate 0, which is not true in the columnspace of A,
so the two columnspaces are not equal. Thus, the first, third, and fourth columns of the
reduced row echelon form ofA do not form a basis for Col(A).
EXERCISE 7
LetA=
[
-1
2
-1
-3
-2
There is an alternative procedure for finding a basis for the columnspace of matrix
A, which uses the fact that the columnspace of a matrix A is equal to the rowspace of
AT. However, the basis obtained in this manner is sometimes not as useful, as it may
not consist of the columns ofA.
EXERCISE 8
Find a basis for the columnspace of the matrix A from Example 13 by finding a basis
for the rowspace ofAT.
r,
0 was
automatically expressed as a spanning set of n - r vectors. We can now show that these
spanning vectors are linearly independent, so that the dimension of the nullspace ofA
is n -
r.
Definition
LetA be an m x n matrix. We call the dimension of the nullspace ofA the nullity ofA
Nullity
160
Chapter 3
EXAMPLE 14
[1 2 0 4]
00 1
A=
is already in reduced row echelon form. By finding a basis for Null(A), determine the
nullity of A and relate it to rank(A).
-4
0
0
x =ti -6 + t2 -5 + t3
-2
1
0
0
0
-3
Thus,
-40 0 -21
0 01 000
-6
-5
linear independence.
x =ti
0
0
1
0
1
0
1
0
0
t3
+ t2
+ t3
t2
t1
t3 =
0,
so the
spanning vectors are linearly independent and hence form a basis for the nullspace of
A. It follows that nullity(A)
Theorem 7
14,
r.
Proof: Let {vi, ... , v11_,} be a spanning set for the general solution of Ax = 0 obtained
in the usual manner and consider
t1V1
+ + tn-rVn-r
=0
(3.5)
Section 3.4 Special Subspaces for Systems and Mappings: Rank Theorem
ti
161
are just the parameters to the free variables. Thus, the coordi
nate associated with the i-th parameter is non-zero only in the i-th vector. Hence, the
t1
t11_, =
independent spanning set for Null(A) and thus forms a basis for Null(A). Thus, the
nullity of A is
r,
as required.
Putting Theorem 6 and Theorem 7 together gives the following important result.
Theorem 8
[Rank Theorem]
If A is any
m x n
matrix, then
rank(A)
nullity(A)
= n
[-i ; ]
EXAMPLE 15
Find a basis for the rowspace, columnspace, and nullspace of A
1 [1 1 11
11
{[l []}-
2
3
3
2
and
of the reduced row echelon form of A have leading l s, so the corresponding columns
{[-il [I}.
from A form a basis for the columnspace of A. That is, a basis for the columnspace of A
is
rank(A)
{[ ;]}
=
nullity(A)
[=; l
t E
2+ 1
and
162
Chapter 3
EXERCISE9
Find a basis for the rowspace, columnspace, and nullspace of A=
verify the Rank Theorem in this case.
[ : = ]
and
PROBLEMS 3.4
Practice Problems
Al Let L be the linear mapping with matrix
1
0 -I
3
3
0 I
3
1
-5
.
1 'YI = 6 and Y2=
I
I 2
1
5
2 5
I
-t
3
JR whose nullspace is Span
'
(a) A=
-t
3
JR whose nullspace is Span
range is Span
{[]}
{[ ]}
and whose
and
{[: ]}
H-]}
[ _;]
[I ; ]
[ i -]
-2
(b) B=
(c) C=
(d) D
0
=
-5
-4
-1
A8 The matrix A
2
1
(b)
(c)
13
0
0
IR.12,
calculation,
give
basis
for
the
basis?
11.3, where V
-1
!Rm, what is m?
(d) Without
basis.
-1
has reduced
what is n?
-2
11
i
[ = l
1
0
0
163
11.3, where V
11.3, where V
[-]
[!]
[!]
0 and,
0 is
Homework Problems
Bl Let L be the linear mapping with matrix
1
-1
'y I =
'and h
-3
=
2 .
1
L(x)
L(x)
J1.
.Y'2.
(a) L(x1,x2)
(x1,X1 +2x2,3x2)
3x3)
164
Chapter 3
{[]}
{[]}
JR.2
range is Span
{[]}
{[-]}
and whose
BS The matrix A =
(a) A= 0
-1
(b) B =
(c) C
0
=
-3
-2
-1
-3
-1
-1
-2
R= 0
-2
1 .
-1
-2
calculation,
0 and,
Ax =
(c)
0
0
basis
for
the
basis?
(b)
give
[ l l
ln ! =:i
what is n?
(a)
4 has re-
(d) D =
[j]
[-i]
m
0 is
Section 3.5
165
Conceptual Problems
Dl Let L
JR11
dim(Range(l)) + dim(Nuil(l))
D2 Suppose that L
JR11
JRm and M
]Rm
D3 (a) If A is a 5
n
-
x7
4,
then
JRP are
(b) If A is a 5
linear mappings.
(a) Show that the range of Mo l is a subspace of
the range of M.
(b) Give an example such that the range of Mo l
is not equal to the range of M.
(c) Show that the nullspace of L is a subspace of
the nullspace of Mo l.
x4
4x5
3, then
D4 Let A be an
n x n
011,11
EXAMPLE 1
Let A be an
n x n matrix.
If there exists an
BA, then
A is said to be invertible, and B is called the inverse of A (and A is the inverse of B).
1
The inverse of A is denoted A-
The matrix
and
[ - ]
_
[ ]
because
[ H- -J=[ J=/
[- -][ ] =[ ]=/
Notice that in the definition, B is
the
166
Chapter 3
Theorem 1
Remark
Note that the proof uses less than the full assumptions of the theorem: we have proven
that if BA = I = AC, then B = C. Sometimes we say that if BA = I, then B is a
"left inverse" of A. Similarly, if AC = I, then C is a "right inverse" of A. The proof
shows that for a square matrix, any left inverse must be equal to any right inverse.
However, non-square matrices may have only a right inverse or a left inverse, but not
both (see Problem D4). We will now show that for square matrices, a right inverse is
automatically a left inverse.
Theorem 2
n x n
n.
Proof: We first show, by contradiction, that rank(B) = n. Suppose that B has rank
n. Then, by Theorem 2.2.2, the homogeneous system Bx=0 has non-trivial
less than
solutions. But this means that for some non-zero x, AB1 = A(Bx) = AO = 0. So,
AB is certainly not equal to I, which contradicts our assumption. Hence, B must have
rank
n.
n,
prove rank(B)=n.
Theorem 2 makes it very easy to prove some useful properties of the matrix in
verse. In particular, to show that A-1 =B, we only need to show that AB=I.
Theorem 3
Suppose that A and B are invertible matrices and that t is a non-zero real number.
Proof: We have
(tA)
u ) ()
A-1 =
AA-1 = 11 =I
(AB)(B-1A-1)=A(BB-1)A-1=AIA-1 =AA-1=I
(AT)(A-1l =(A-1Al = F =I
167
should be clear that it can be applied to any square matrix. Write the matrix equation
AX
I in the form
Hence, we have
So, it is necessary to solve three systems of equations, one for each column of X. Note
that each system has a different standard basis vector as its right-hand side, but all
have the same coefficient matrix. Since the solution procedure for systems of equations
requires that we row reduce the coefficient matrix, we might as well write out a "triple
augmented matrix" and solve all three systems at once. Therefore, write
Suppose that A is row equivalent to!, and call the resulting block on the right B
so that the reduction gives
Now, we must interpret the final matrix by breaking it up into three systems:
In particular, Ab1
e1, Ab2
desired matrix X is b1, the second column of X is b2, and so on. Thus, A is invertible
and B
A-1
n.
Hence, A is not
= n.
Algorithm 1
Finding A-1
168
Chapter 3
EXAMPLE2
Deternline whether A =
[ i
2
[l
1
-1
II
1 1 ol [1 11 -1 l
1 l [1
1
1
-1
j]
[-l : j].
-1
-2
-1
-2
-5
-1
You should check that the inverse has been correctly calculated by verifying that
AA-1 =I.
EXAMPLE3
Determine whether A =
[ ]
II
[ I J [ 1- J
EXERCISE 1
Determine whether A =
[ ]
Theorem 4
n xn
Theorem 4
(continued)
169
(1) A is invertible.
(2) rank(A) = n.
(3) The reduced row echelon form of A is/.
(4)
For all
P =>
in proving a theorem such as this to prove (1) => (2) => (3) ::::>
so that any statement implies any other.)
(1)
(2)
=>
=>
nx n.
(3)
(4)
::::>
(4):
=>
E IR." and so
Col(A) = IR.".
(6)
::::>
i n. Thus, A is
invertible.
Amongst other things, this theorem tells us that if a matrix A is invertible, then
the system Ax = b is consistent with a unique solution. However, the way we proved
the theorem does not immediately tell us how to find the unique solution. We now
demonstrate this.
Let A be an invertible square matrix and consider the system Ax=
b. Multiplying
l
both sides of the equation by A-1 gives A-1Ax= A-1b and hence x=A- b.
_,
EXAMPLE4
Let A=
[ n
EXERCISE 2
Let A=
1
2
[]
[ -n
_,
[)
is
3
I
.
and b= 14
7 . Determme A- and use 1t to find the solution of Ax=b.
1
_,
[ ]
_,
170
Chapter 3
It likely seems very inefficient to solve Exercise 2 by the method described. One
would think that simply row reducing the augmented matrix of the system would make
more sense. However, observe that if we wanted to solve many systems of equations
with the same coefficient matrix A, we would need to compute A-1 once, and then
each system can be solved by the problem of solving the system to simple matrix
multiplication.
Remark
It might seem surprising at first that we can solve a system of linear equations without
performing any elementary row operations and instead just using matrix multiplica
tion. Of course, with some thought, one realizes that the elementary row operations
are "contained" inside the inverse of the matrix (which we obtained by row reducing).
In the next section, we will see more of the connection between matrix multiplication
and row reducing.
Definition
Inverse Mapping
n
n
----?
If L : JR. ----? JR. is a linear mapping and there exists another linear mapping M : JR.11
JR.n such that M o L = Id = Lo M, then L is said to be invertible, and M is called the
inverse of L, usually denoted L-1
Observe that if M is the inverse of L and L(v)= w, then
M(w)= M(L(v))= (M
L)(v)= ld(v)
M)(w)= ld(w)= w
Theorem 5
n
Suppose that L : JR. ----? JR.11 is a linear mapping with standard matrix [L] = A and
n
that M : JR. ----? JR.11 is a linear mapping with standard matrix [M] = B. Then M is the
inverse of L if and only if B is the inverse of A.
EXAMPLES
171
For each of the following geometrical transformations, determine the inverse transfor
mation. Verify that the product of the standard matrix of the transformation and its
inverse is the identity matrix.
(a) The rotation Re of the plane
(b) In the plane, a stretch T by a factor oft in the x1-direction
Solution: (a) The inverse transformation is to just rotate by -e. That is, (Re t 1
We have
[R
since sin( -e)
[Re ][R_e ]
-e
cos( -e)
- sin( -e)
sin(-e)
cos(-B)
r
[
c se
- sine
sm e
cose
][
] [
cose
sine
- sine
cose
R_8
cose. Hence,
c se
sine
- sm e
cose
cos2 e + sin2e
cos2e+sin2e
1
(b) The inverse transformation r- is a stretch by a factor of
EXERCISE 3
] [ ]
=
0
1
? in the Xi-direction:
For each of the following geometrical transformations, determine the inverse transfor
mation. Verify that the product of the standard matrix of the transformation and its
inverse is the identity matrix.
(a) A reflection over the line x
x1 in the plane
Observe that if
range of the original L. Therefore, it follows that if L has an inverse, the range of L
must be all of the codomain JR.11 Moreover, if L(x1)
172
Chapter 3
Theorem 6
JR11 such
(7) (9): Assume that L is invertible. Hence, there is a linear mapping L-1 such that
L-1(L(x)) = x. If x E Null(L), then l(x) = o and then L-1L(x) = L-1(0). But, this
gives
x=L-1(0) =0
_,
_,
Remark
It is possible to give many alternate proofs of the equivalences in Theorem 4 and
Theorem 6. You should be able to start with any one of the nine statements and show
that it implies any of the other eight.
EXAMPLE6
EXAMPLE 7
Prove that the linear mapping projil is not invertible for any v E JR11, n 2.
Solution: By definition, projil(x) = tV, for some t E R Hence, any vector y
that is not a scalar multiple of v is not in the range of prok Thus, Range(projil)
hence projil is not invertible, by Theorem 6.
--+
E ]Rn
* IR11,
3
JR defined by
is invertible.
Solution: Assume that x is in the nullspace of L. Then L(x) = 0, so by definition of
L, we have
2 X1
X2 =0
X3 =0
X2 - 2X3 =0
The only solution to this system is x1 =x2 =x3 =0. Thus, Null(L)= {0} and hence L
is invertible, by Theorem 6.
173
Finally, recall that the matrix condition A B = BA = I implies that the matrix
inverse can be defined only for square matrices. Here is an example that illustrates for
linear mappings that the domain and codomain of L must be the same if it is to have
an inverse.
Consider the linear mappings :
P
JR.4 JR.3 defined by P(x ,x2,x3,x ) =(x ,x2,x3)
1
1
4
3
and inj : JR. JR.4 defined by inj(x1,x2,x3) =(xi,x2,X3,0).
EXAMPLE 8
It is easy to see that Poinj = Id but that injoP 1- Id. Thus, P is not an inverse for
inj. Notice that P satisfies the condition that its range is all of its codomain, but it fails
the condition that its nullspace is trivial. On the other hand, inj satisfies the condition
that its nullspace is trivial but fails the condition that its range is all of its codomain.
PROBLEMS 3.5
Practice Problems
Al For each of the following matrices, either show that
the matrix is not invertible or find its inverse. Check
by multiplication.
(a)
[; J
the following.
(a) B
(b)
(c)
[ ! i l
[i ]
[ : l
(e)
(f)
1
0
2
0
1
2
2
6
1
0
0
0
0
0
1
0
0
0
A2 Let B =
7
3
=
X
(c) B
A3 Let A =
(d)
(b) B
[;]
nl
[[ : ]{]]
[ ]
and B =
1
0
0
0
1
0
[ ! i]
and (ABt1
(ABt1 = s-1A-1
t A-1
1
0
AB
[ n
[ ]
[ -! ]
-
(d)
(c)
[ ]
174
Chapter 3
11
A6 Suppose that L
that
L(x).
Homework Problems
Bl For each of the following matrices, either show that
the matrix is not invertible or find its inverse. Check
by multiplication.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
U)
[- ]
[ : l
[ 3 ]
[ i -1
[f - ]
-
1
0
2
1
-1
0
0
1
3
2
1
3
6
2
0
1
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
3
1
-2
0
0
0
0
0
1
0
o
0
B2 Let A
0
1
o
0
1
0
0
1
0
0
0
o
1
1-
-1
0
0
o
0
-1
b if
01 b
=
2
0
5
3
(ii)
(iii)
B3 Let A
5
3
3
3
Ul
3[_:1
Hl
[ ]
and B
[ l
(b) Calculate AB
0
0
0
1
0
0
0
1
0
0
0
0
0
0
1
0
0
0
1
1
0
0
1
0
1
0
0
(AB)-1
[ ]
(c)
(d)
- 1/
[-
from JR.
2
BS The mappings in this problem are from JR.
(a) Determine the matrix of the stretch
-t
2
JR. .
by a fac
3
JR. , determine the matrices
-t
[W1 ],
(a)
-1
line x1 + x2
175
R-1
x3-direction.
(b)
in the x x3-plane.
in the
Computer Problems
1.23
2.01
Cl Let A=
1.11
2.14
3.11
-2.56
1.01
0.00
3.03
0.04
0.03
-5.11
-1 .9 0
4.05
.
2.56
1.88
of
Conceptual Problems
Dl Determine an expression in terms of A-1 and B-1
for
trix
(b)
((ABl)-1
D4 Let A
CA =
I. Hence,
verse.
A.
(1) A is invertible.
(2) Null(A) {O}.
=
AT
is invertible.
x n
176
Chapter 3
Definition
A matrix that can be obtained from the identity matrix by a single elementary row
Elementary Matrix
EXAMPLE 1
E1=
[ ]
is the elementary matrix obtained from I2 by adding the product oft times
the second row to the first-a single elementary row operation. Observe that 1 is the
matrix of a shear in the x1 -direction by a factor oft.
2 =
[ ]
[ ]
3 =
x1 in the plane.
Theorem 1
why this works by verifying the conclusion for some simple cases for 3 x 3 matrices.
Case 1. Consider the elementary row operation of adding k times row 3 to row 1. The
l
0 . Then,
a11
a12
a13
a21
a22
a23
a31
a32
a33
while
EA=
l[
R, + kR,
a11 + ka31
a12 + ka32
a13 + ka33
a21
a22
a23
a31
a32
a33
a11
a12
a13
a21
a22
a31
a32
a23 =
a33
[""
a12 + ka32
a1 3 + ka33
a21
a22
a23
a31
a32
a33
+ ka, ,
177
Case 2. Consider the elementary row operation of swapping row 2 with row 3, which
1 0
1 :
has the corresponding elementary matrix E= 0 0
0 1 0
a11
a21
a31
a12
a22
a32
a13
a23
a33
0
0
1
ol [
a11
az1
a31
while
EA=
1
0
[ OJ
[
][
a11
R2 l R3 -
a12
a22
a32
a31
a21
a12
a32
a22
a11
a13
a23 = a31
a21
a33
a13
a33
a23
a12
a32
a22
a13
a33
a23
EXERCISE 1
Theorem 2
Verify that Theorem 1 also holds for the elementary row operation of multiplying the
second row by a non-zero constant for 3 x 3 matrices.
E1, E2,
of A.
, Ek. such that Ek E2E1A is equal to the reduced row echelon form
Proof: From our work in Chapter 2, we know that there is a sequence of elemen
tary row operations to bring A into its reduced row echelon form. Call the elementary
matrix corresponding to the first operation E1, the elementary matrix corresponding
to the second operation E2 , and so on, until the final elementary row operation cor
responds to Ek. Then, by Theorem 1, E1A is the matrix obtained by performing the
first elementary row operation on A, E2E1A is the matrix obtained by performing the
second elementary row operation on E1A (that is, performing the first two elementary
row operations on A), and Ek E2E1A is the matrix obtained after performing all of
the elementary row operations on A, in the specified order.
EXAMPLE2
Let A =
[ ].
1
2
2
4
R2 - 2R1 -
1
0
R2
2
0
R2, so E2=
R1 - Rz
]
[ .
[ ]
1
1
1 2 .
1
0
2
0
178
Chapter 3
EXAMPLE2
(continued)
[ - ].
[ -][ 12] [ _ ][ !] [ l
Thus,E3E2E1A
Remark
We know that the elementary matrices in Example
must be 2 x
First, we had only two rows in A to perform elementary row operations on, so this
must be the same with the corresponding elementary matrices. Second, for the matrix
multiplication E1A to be defined, we know that the number of columns in E1 must be
equal to the number of rows in A. Also, E1 is square since it is elementary.
EXERCISE2
Let A
[ H
, E, such that
2,
In the special case where A is an invertible square matrix, the reduced row eche
lon form of A is I. Hence, by Theorem
operations such that Ek E1A
Ek E1 satisfies BA
I,
so B is the inverse of A. Observe that this result corresponds exactly to two facts we
observed in Section 3.5. First, it demonstrates our procedure for finding the inverse of
a matrix by row reducing
Finally, observe that elementary row operations are invertible since they are re
versible, and thus reflections, shears, and stretches are invertible. Moreover, since the
reverse operation of an elementary row operation is another elementary row operation,
the inverse of an elementary matrix is another elementary matrix. We use this to prove
the following important result.
Theorem 3
Proof: By Theorem 2, there exists a sequence of elementary row operations such that
Ek E1A = I. Since E k is invertible, we can multiply both sides on the left by (Ekt'
to get
1
(E k)- EkEk-l E1A
1
(Ek)- 1
or
Ek-I E1A
1
E"k
We continue to multiply by the inverse of the elementary matrix on the left until the
equation becomes
179
Remark
Observe that writing A as a product of simpler matrices is kind of like factoring a
composition. There are many very important matrix decompositions in linear algebra.
We will look at a useful matrix decomposition in the next section and a couple more
of them later in the book.
EXAMPLE3
Let A =
[ l
Solution: We row reduce A to I, keeping track of the elementary row operations used:
0 2
1 1
R2 ! Ri
-[
1 1
0 2
1 2
2R
-[
1 1
0 1
R1 - R1
-[
1 0
0 1
Hence, we have
Thus,
and
[o ] [1 OJ [ ]
1
1 1
1
A= e-lE-lE1 2 3 = 1 00 20 1
PROBLEMS 3.6
Practice Problems
Al Write a 3
[- !]
A =
20
(a) Add (-5) times the second row to the first row.
(b) Swap the second and third rows.
(c) Multiply the third row by (-1).
180
Chapter 3
(a) Add (-3) times the third row to the fourth row.
(b) Swap the second and fourth rows.
(i) Find
(a) A=
(c)
(e)
sequence
of
elementary
matrices
[i ]
[2 42 i
[--4 1 =]4
1 -2 4
(a)
[0 -4i1
[[o ! o]l
(b)
(d)
en
[-0 0 -1i
[ ! I
[ : ]
(b) A=
(c) A=
(d) A=
-1 3 -4 -1
0 1 2 0
-2 4 -8 -1
Homework Problems
Bl Write a 3
A=
[1 --2i
not elementary.
and verify that the productEA is
-3
(-2)
4 4
2.
(-2)
(-2).
0 1 0i
(a)[
[-0 1 0i
[0 0i1
(b)
[1 o ol
[1 00 1 i
[ l ! ]
(c)
(d)
(e)
cn -
sequence
of
elementary
matrices
[ - i
2 4 0
(b)
A=
[ 1
H _; -]
1
(c)
A=
3
(d) A=
-2
3
-2
4
4
-5
2
-1
1
181
1
2
-3
5
Conceptual Problems
2
IR.2
1 and 2 such
E2E1A= I.
Since A is invertible, we know that the system
1
Ax = b has the unique solution x = A- 6 =
E2E1 b. Instead of using matrix multiplication,
Dl (a) Let L
IR.
-t
A =
[ =l
By
(b)
-t
IR.11
First, compute
can be written as
matrix
tions.
D2 For 2
1 holds for
(c)
D3 Let A=
[ ;]
and
ated with
[;].
3.7 LU-Decomposition
One of the most basic and useful ideas in mathematics is the concept of a factorization
of an object. You have already seen that it can be very useful to factor a number into
primes or to factor a polynomial. Similarly, in many applications of linear algebra, we
may want to decompose a matrix into factors that have certain properties.
where the coefficient matrix A remains the same but the vector
b changes.
= b,
The goal of
Definition
Upper Triangular
Lower Triangular
be lower triangular if the entries above the main diagonal are all zero-in particular,
(L)iJ
= 0 whenever i
<
}.
= 0 whenever i
>
}. Ann
x n matrix L is said to
182
Chapter 3
Remark
By definition, a matrix in row echelon form is upper triangular. This fact is very im
portant for the LU-decomposition.
substitution. The only difference between the systems will then be the effect of the row
operations on
require are the row echelon form of A and the elementary row operation used.
For our purposes, we will assume that our n x n coefficient matrix A can be brought
into row echelon form using only elementary row operations of the form add a mul
tiple of one row to another. Since we can row reduce a matrix to a row echelon form
without multiplying a row by a non -zero constant, omitting this row operation is not
a problem. However, omitting row interchanges may seem rather serious: without row
interchanges, we cannot bring a matrix such as
[ ]
ever, we only omit row interchanges because it is difficult to keep track of them by
hand. A computer can keep track of row interchanges without physically moving en
tries from one location to another. At the end of the section, we will comment on the
case where swapping rows is required.
Thus, for such a matrix A, to row reduce A to a row echelon form, we will only use
row operations of the form R; + sRj where i > j. Each such row operation will have
a corresponding elementary matrix that is lower triangular and has ls along the main
diagonal. So, under our assumption, there are elementary matrices E1,
, Ek that are
A = LU, where U is upper triangular and Lis lower triangular. Moreover, L contains
the information about the row operations used to bring A to U.
Theorem 1
If A is an n x n matrix that can be row reduced to row echelon form without swap
ping rows, then there exists an upper triangular matrix U and lower triangular matrix
L such that A =LU.
Definition
LU-Decomposition
Writing an n x n matrix A as a product LU, where Lis lower triangular and U is upper
triangular, is called an LU-decomposition of A.
Our derivation has given an algorithm for finding the LU-decomposition of such
a matrix.
Section 3. 7
EXAMPLE 1
LU-Decomposition
183
2[ 4 -1-1 4
-1 -1 3]
2 -1-1 4 2 - 20 -11 -2
4
[ -1 -1 3 l
[ 0 -3/2 4 l
2
1
4
l
- [ 00 10 -22
[-102 0 1], [1/2 0 ],1 [0 3/2 ll
[ 100 Ol10 [-1/20 100 l01 00 -3/210
I
[ 100 ] [_lj -3/210 Ol01 - [-1/22 -3/210 100
6 .
Find an LU-decomposition of A=
R2 - R 1
R3+R1
R3+ R2
= u
E1 =
Hence, we let
L=
E; 1Ei1E;1 =
Observe from this example that the entries in Lare just the negative of the multi
pliers used to put a zero in the corresponding entry. To see why this is the case, observe
that if EkE1A= U , then
Hence, the same row operations that reduce A to U will reduce Lto /.
T his makes the LU-decomposition extremely easy to find.
EXAMPLE2
Find an LU -decomposition of B
-4[ 2 31 -1-33] .
6
184
Chapter 3
EXAMPLE2
(continued)
L=
H '. ]
H ]
0
Therefore. we have
H m i : i
]
[
0
B = LU =
EXAMPLE3
Find an LU-decomposition of C =
[ ;
[ _; - l
-4
-2
-11
Therefore, we have
-4
-3
3.
-2
i
; - l
-[
-3
L=
-1
R3
3R2
EXERCISE 1
Find an LU-decomposition of A=
1-
-1
-3
-3
16
L=
[ i
[ ll
-4
-4
-3
185
We now look at how to use the LU-decomposition to solve the system Ax =b. If
A =LU, the system can be written as
LUx =b
Letting y = Ux, we can write LUx =bas two systems:
Ly=b
and
Ux=y
which both have triangular coefficient matrices. This allows us to solve both sys
tems immediately, using substitution. In particular, since Lis lower triangular, we use
forward-substitution to solve y and then solve Ux =y for x using back-substitution.
Remark
Observe that the first system is really calculating how performing the row operations
on A would have affected
EXAMPLE4
-11
b.
[ ]
2 1
3
3 and b =- 13 Use an LU-decomposition of B to solve Bx=b.
Let B= -4 3
4
6 8 -3
.
b, we get
3
}'1 =
-2y1 + )'2
-13
[-;].
Using forward-substitution, we find that )'1 =3, so )'2 =-13 + 2(3) = -7 and
2X1 + X2 - X3 =3
Sx2 + x3 =-7
-X3 =2
Using back-substitution, we get x3
3- (-!) + (- 2)
=>
l=H
186
Chapter 3
EXAMPLES
+ : -4- + + l
[-: -4 H
[
:
+ [
+ -
]
:l
-[ l
[-: ]
-2
12
-2
[ _: -41 l
[ -1 48 l
[ H
-2
-2
3
6
R1
R3
-2
Thus, U =
-2
-1
-2
R1
2R1
I
0
0
R3 - 2R2
L=
-2
I
-I
0
L=
-2
YI =
-y1
+ Y2
-2y1 + 2y2 + y3
b.
This gives
= 6
= 12
1 + -14
[i].
Xt-X2+ X2++4X3X3
Ox3
8 -St. x3 t -x2 -4t8 -St, l x2 [ 8- + 4t x, 1 + 4t) -t
[ : 4t -] + t [-Si
So, y1
I,
y, =
6+I
= 7, and y, =
= O. Then we solve Ut =
which is
= 1
= 7
= 0
This gives
JR,
= 7
so
and
(7 -
Hence,
x = -7
EXERCISE 2
Let A =
[-14 -11 1]
2
-3 . Use the LU-decomposition of A that you found in Exercise 1
-3
-3
b=
b, where:
(b) b=
Ul
187
PA=
Then, to solve
LU
system
PAx=Pb
PROBLEMS 3.7
Practice Problems
[-2-
-2-2
[ 2 J
-2 2
-2 -22 2
2 2
(c)
-1
0
1
-]
[ ]
-2 -2
2
-3
-3
4
-3
(d )
1
-1
0
(f)
2.
-6
0
0
4
3
-1
3
3
-1
3
-1
-1
0
4
3
-1
0
0
1
4
A=
0
-4
-4
-1
(b)
3
-4
-3
A=
Homework Problems
matrices.
-[-1
[
(a)
(c)
1
5
-1
1
1
0
-1
-1
-4
-4
(b)
(d )
-4
0
3
( e)
-1
b1 =
b, =
-5
b1 =
-1
-8
-3
3
3
3
2 2 -2
3
-3
0
-1
-4
b 1 = = b, = =
-1
( d)
-1
(c)A= -
[ ]
l l l l
H -i. [=;]. [ ;]
[ i 2 -n l=H {l
2
-2 2
A= -
(b)
-4
5
-1
1
0
( e)
3
(a)
-4
0
0
0
1
-6
'bi=
_,
7
-4
5
'b2 =
-22 2
-1
(f)
2.
4
-3
5
4
3
-3
5
-5
5
-3
3
-5
-1
5
4
188
Chapter 3
[-4I
3[-2 -4 -], {H [l
; 2 b
6 =H nJ,c, ll
(a) A =
(b) A =
(c) A =
b, =
b,
-8
b, =
0
-5
(d) A =
b, =
,b1
Conceptual Problems
Dl (a) Prove that the inverse of a lower-triangular ele
CHAPTER REVIEW
Suggestions for Student Review
Try to answer all of these questions before checking an
swers at the suggested locations. In particular, try to in
3.3)
[ ]
under the ro
(Section
Check that
23
0 and
solutions for
b.
A is
3.1)
JR.2
and determine
3.2)
3.4)
3.4)
3.4)
Ax =
0. (Section
4 4;
3;
2.
structing examples of
(Section
rank
(ii) rank
and (iii)
and check
Chapter Review
3.4)
189
(Section
3.5)
7 For
3.5)
3.6)
Chapter Quiz
[3
-5
4
El Let A =
=J
and B
-1
0 2 .
3
1 -1 5
(a)AB
(b)BA
let
(b) Use
A
[n
=
Determine
the
result
fA
[_i]
of
part
fBCY)
(a) to
-1
3
2
calculate
0
A=
2
0
[ i
1
and b =
5
1 6 . Determine
18
[]
= v.
f8(x)
Ax =
-1
1 1 1
-7
and
-2
, and v =
E4 Let A =
-5
6
be the matrix
[ l
-1
4
-3
5
3
'I =
ES Let B =
(c)BAr
[- - ].
E2 (a) Let A =
-1 -1 -1
'i1
1
3
0
2 -1
0
1
-2
0
1
4
5
8
14
1 0
1 -1
0
0
0
0
0
1 1
3
0 1 2
0
[ i
2
1 1
190
Chapter 3
: JRll
--?
ElO Let
[ =]
MK for all 3 x 4
range
is
. .
nullspace is Span
Span
{[;]}
{[ ]}
and
whose
0 0
4
(a) Determine a sequence of elementary matrices
Ei, . .. , Ek> such that EkE i A = I .
(b) By inverting the elementary matrices in part
(a), write A as a product of elementary matri
ces.
E12 For each of the following, either give an example
whose
range
is
nullspace is Span
Span
{[]}
{ [ l l}
and
whose
{[ ;]}
-
Further Problems
These problems are intended to be challenging. They
may not be of interest to all students.
C commutes with matrix D if
DC. Show that the set of matrices that com-
CD
mute with A
[; 7]
the matrix
[ a2 :1
0
is nilpotent. Generalize.
x1 -axis.
JR.2).
Chapter Review
n x
matrices such
n x n
n x n
system
191
[ ! ]
AX+ BY= C
BX+ AY= D
MyMathlab
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 4
Vector Spaces
CHAPTER OUTLINE
4.1
Spaces of Polynomials
4.6
4.7
This chapter explores some of the most important ideas in linear algebra. Some of
these ideas have appeared as special cases before, but here we give definitions and
examine them in more general settings.
(p + q)(x)
and
EXAMPLE 1
(2+3x+4x2+x3)+(5+x-2x2+7x3) =2+5+(3+l)x+(4-2)x2+(1+7)x3
=7 + 4x + 2x2 + 8x3
(3+x2-5x3)-(l -x-2x2)
194
Chapter 4
EXAMPLE 1
Vector Spaces
(c)
5(2 + 3x
4x2 + x3)
(continued)
Solution: 5(2+3x+4x2+.x3)=5(2)+5(3)x+5(4)x2+5(1).x3=10+15x+20x2+5.x3
(d)
p(x), q(x) and r(x) be polynomials of degree at most n and let s, t E R Then
(1) p(x) + q(x) is a polynomial of degree at most n
(2) p(x) + q(x)=q(x) + p(x)
(3) (p(x) + q(x)) + r(x)=p(x) + (q(x) + r(x))
(4) The polynomial 0= 0 +Ox ++Ox't, called the zero polynomial, satisfies
p(x) + 0=p(x)=0 + p(x) for any polynomial p(x)
(5) For each polynomial p(x), there exists an additive inverse, denoted (-p)(x),
with the property that p(x) + (-p)(x)=O; in particular, (-p)(x) = -p(x)
(6) tp(x) is a polynomial of degree at most n
(7) s(tp(x))=( st)p(x)
(8) (s + t)p(x)=sp(x) + tp(x)
(9) t( p(x) + q(x))=tp(x) + tq(x)
(10) 1 p(x)= p(x)
Let
Remarks
1. These properties follow easily from the definitions of addition and scalar mul
tiplication and are very similar to those for vectors in !Rn. Thus, the proofs are
left to the reader.
2. Observe that these are the same 10 properties we had for addition and
scalar multiplication of vectors in JR.11 (Theorem 1.2.1) and of matrices
(Theorem 3 .1.1).
3. When we look at polynomials in this way, it is the coefficients of the polynomi
als that are important.
As with vectors in JR.11 and matrices, we can also consider linear combinations of
polynomials. We make the following definition.
Definition
Let 13= (p1 (x), . . . , Pk(x)} be a set of polynomials of degree at most n. Then the span
Span
of 13 is defined as
Section 4.1
The set 13
Linearly Independent
to the equation
t1 P1(x)+
EXAMPLE2
195
Definition
is t1
Spaces of Polynomials
tk
+tkPk(x)
{1+x,1 +x3,x+x2,x+x3}.
Solution: We want to determine if there are t1,t 2,t ,t4 such that
3
1+2x+3x2+4x3
By comparing the coefficients of the different powers of x on both sides of the equation,
we get the system of linear equations
t1 + t2
+ t
t1
t
t2
3
3
+ t4
+ t4
3
4
1
0
1
0
-2
3
3
We see that the system is consistent; therefore, 1 +2x+3x2+4x3 is in the span of 13.
In particular, we have t1
EXAMPLE3
-2, t2
3, t
3, and t4
1.
2x3} is linearly
Solution: Consider
0
2
2
-1
The only solution is t1
t2
2
2
Chapter 4
196
Vector Spaces
EXERCISE 1
Determine if :B
==
==
of :B?
EXERCISE 2
Consider :B
==
PROBLEMS 4.1
Practice Problems
Al Calculate the following.
(e)
(a)
A2
(f)
(b)
==
-1+7x+Sx2+4x3
2 +x+Sx3
A4
==
2.
Homework Problems
Bl Calculate the following.
(a )
(b )
B2
==
p(x)
p(x)
q(x)
q(x)
==
==
==
==
1
Sx+2x2+3x3
3 +x2 - 4x3
1+x3
{x2,x3,x2+x3+x4}
(b) {1+21
- 2 x+
6x - }
6
(c) {1 + x +x3,x+x3+ x5, 1 - x5 }
(d) (1 - 2x+x4,x - 2x2+ x5, 1 - 3x+x3}
(e) {l+2x+x2-x3, 2+3x-x2+x3+x4,1+x-2x2
+2x3+x4,1+2x+x2+x3 - 3x4,
4 + 6x - 2x2+Sx4)
Prove that the set 13
{1,x - 2,(x - 2 ) 2,( x - 2 ) 3}
(a)
B4
3.
197
Conceptual Problems
Dl
dependent.
Vector Spaces
Definition
Vector Space over
J.
A vector space over JR. is a set V together with an operation of addition, usually
y for any x, y E V, and an operation of scalar multiplication, usually
denoted sx for any x EV and s E JR., such that for any x, y, z EV and s, t E JR. we have
denoted x +
V1 x+
V2 x +
V3
x+0
0 +x
(additive identity)
x + (-x) = 0
V6 tx EV
V7 s(tx)
(additive inverse)
Remarks
1. We will call the elements of a vector space
very different objects than vectors in IR.11 Thus, we will always denote, as in the
definition above, a vector in a general vector space in boldface (for example, x).
However, in vector spaces such as JR.", matrix spaces, or polynomial spaces, we
will often use the notation we introduced earlier.
198
Chapter 4
Vector Spaces
2. Some people prefer to denote the operations of addition and scalar multiplica
tion in general vector spaces by E9 and 0, respectively, to stress the fact that
these do not need to be "standard" addition and scalar multiplication.
3. Since every vector space contains a zero vector by V3, the empty set cannot be
a vector space.
6. We define vector spaces to have the same structure as JRn. The study of vector
spaces is the study of this common structure. However, it is possible that vectors
in individual vector spaces have other aspects not common to all vector spaces,
such as matrix multiplication or factorization of polynomials.
EXAMPLE 1
JR11 is a vector space with addition and scalar multiplication defined in the usual way.
We call these standard addition and scalar multiplication of vectors in JR11
EXAMPLE2
EXAMPLE 3
M(m, n), the set of all m x n matrices, is a vector space with standard addition and
scalar multiplication of matrices.
EXAMPLE4
Consider the set of polynomials of degree n. Is this a vector space with standard addi
tion and scalar multiplication? No, since it does not contain the zero polynomial. Note
also that the sum of two polynomials of degree n may be of degree lower than n. For
example, (1 + x")
(1 - x'1)
under addition.
EXAMPLES
EXAMPLE6
Let C(a,b) denote the set of all functions that are continuous on the interval (a, b).
Since the sum of continuous functions is continuous and a scalar multiple of a contin
uous function is continuous, C(a,b) is a vector space. See Figure 4.2.1.
199
- ___. ,,
,
.
..
..
g
.
.
.
.
.
..
::
,
,
,
..
...
, .. ___
Figure 4.2.1
EXAMPLE?
,
,
,;
..
... . ....
"! +
x1+2x2
1,
2x1 + 3x2
standard addition and scalar multiplication? No. This set with these operations does
not satisfy many of the vector space axioms. For example, Vl does not hold since
[-3]
2
m
"TT"
1s
a so1ut1on
B ut
Jl ; 1t
1s
of th'1s system of l.mear equations.
EXAMPLE 8
EXAMPLE9
Let S
{[ ]
3) (3,
=
1 x1, x2 E JR
X1 + X2
2)
-:f.
(2,
3),
[ I
and y
r
1+Y2
X1 + X2
Vl We have
1 +y
[ ; l
X 1+ X2
be vectors ins.
r
l +Y2
X1+Y1 and z2
and hence
1+y
l[
; :
X1 + Yl + X2 +Y2
x2 + y2, then z1 + z2
[ IE
l
x1 + Yt +x2 + Y2
Z1 + Z2
since it satisfies the conditions of the set. Therefore, S is closed under addition.
V3 The vector 0
conditions of S.
[]
satisfies 1 + 0
200
EXAMPLE9
(continued)
[ ]
t[ l [tx1ttxxi2tx2l
is (-x)=
X1+X2
V6 tx
Xt
-X1- X2
which is in S since
X2
multiplication.
EXERCISE 1
{[] x1, x2
I
Z}
EXERCISE 2
Let S =
{ [1 2] a1, a2
I
2 diagonal
Again, one advantage of having the abstract concept of a vector space is that when
we prove a result about a general vector space, it instantly applies to all of the examples
of vector spaces. To demonstrate this, we give three additional properties that follow
easily from the vector space axioms.
Theorem 1
E JR
Proof: We will prove ( 1). You are asked to prove (2) and (3) in Problem Dl . For any
x EV we have
Ox=Ox+ 0
byV3
= Ox+ [x + (-x)]
byV4
byVlO
byV2
byV8
= lx + (-x)
operation of numbers in JR
= x + (-x)
byVlO
=0
byV4
201
Thus, if we know that V is a vector space, we can determine the zero vector of V
by finding Ox for any x E V. Similarly, we can determine the additive inverse of any
vector x EV by computing (-l)x.
EXAMPLE 10
Let V = {(a,b) I a,b E JR, b > O} and define addition by (a,b) EB (c, d)
(ad + be,bd)
and define scalar multiplication by t 0 (a,b) = (taY-1, b1). Use T heorem 1 to show
that axioms V3 and V4 hold for V with these operations. (Note that we are using EB
and 0 to represent the operations of addition and scalar multiplication in the vector
space to help distinguish the difference between these and the operations of addition
and multiplication of real numbers.)
(a,b) EB (0, 1) =(a(l) + b(O), b(l) ) =(a,b) =(O(b) + l(a), 1(b) ) =(0, 1) EB (a,b)
So, V satisfies V3 using 0 = (0, 1).
Similarly, if V is a vector space, then by Theorem 1 for any x = (a,b) E V we
must have
>
0. Also,
D2.
Subspaces
In Example 9 we showed that S is a vector space that is contained inside the vector
space JR3. Observe that, by the definition of a subspace of JR" in Section 1.2, S is
actually a subspace of JR3. We now generalize these ideas to general vector spaces.
Definition
Subspace
Equivalent Definition. If 1LJ is a subset of a vector space V and 1LJ is also a vector
space using the same operations as V, then 1LJ is a subspace of V.
To prove that both of these definitions are equivalent, we first observe that if 1LJ is a
vector space, then it satisfies properties S 1 and S2 as these are vector space axioms V 1
and V6. On the other hand, as in Example 4.2.9, we know the operations must satisfy
axioms V2, VS, V7, V8, V9, and VlO since V is a vector space. For the remaining
axioms we have
202
Chapter 4
Vector Spaces
V1
V3 Follows from Theorem 1 and property S2 because for any u E 11J we have
Oil E 11J
V4 Follows from Theorem 1 and property S2 because for any u E 11.J, the additive
inverse of u is (-u)
(-1 )u E 11J
Remarks
1. When proving that a set 11J is a subspace of a vector space V, it is important not
to forget to show that 11J is actually a subset of V.
2. As with subspaces of IRn in Section 1 .2, we typically show that the subset is
non-empty by showing that it contains the zero vector of V.
EXAMPLE 11
{[ ]
2
I a1, a2E
EXAMPLE 12
IR}
non-empty.
Let
Sl
(p + q)(3)
S2
(sp)(3)
p(3) + q(3)
sp(3)
sO
EXAMPLE 13
P3
Then
p(3)
=
0 and q(3)
0.
0, so sp(x) E 11J
Note that this also implies that 11J is itself a vector space.
.
2 matnx b y tr
([
a11
a11
a1 2
])
{A E M(2, 2) I tr(A)
O} is a subspace of M(2, 2).
Solution: By definition, is a subset of M(2, 2). The zero vector of M(2, 2) is
=
02,2
b11
[ l
Clearly, tr(02,2)
0, so02,2E.
S l tr(A+B)
tr(A)
0 and
A+BE
tr
([
aii +
a21 +
bii
b21
j)
b12
a22 + b 22
a12 +
0, so
EXAMPLE 13
S2 tr(sA)
(continued)
tr
([
])
sa12
sa22
sa11
sa2 1
sa1 i + sa22
s(a11 + a22)
s(O)
203
0, so sA ES
EXAMPLE 14
[]
JR2,
P2 I b + c
JR3,
since
JR2
is not a subset of
JR3,
JR3.
That is, if
since a vector in
JR3
has
three components.
a}
P2 .
EXERCISE 3
Prove that U
EXERCISE4
Let V be a vector space. Prove that {O} is also a vector space, called the trivial vector
space, under the same operations asV by proving it is a subspace of V.
is a subspace of
In the previous exercise you proved that {0} is a subspace of any vector space V.
Furthermore, by definition, V is a subspace of itself. We now prove that the set of all
possible linear combinations of a set of vectors in a vector space V is also a subspace.
Theorem 2
If {v1, ... , vk} is a set of vectors in a vector space V and Sis the set of all possible
linear combinations of these vectors,
0 for 1
+ tkvk
k, we get
t1V1 +
+ tkvk.
y ESsince
s1 v1 +
+ skvk
It follows that
(si + t;)
E JR,
by V7. Thus, Sis also closed under scalar multiplication. Therefore, Sis a subspace
ofV.
204
Chapter 4
Vector Spaces
To match what we did in Sections 1.2, 3.1, and 4.1, we make the following
definition.
Definition
If is the subspace of the vector space V consisting of all linear combinations of the
Span
vectors
Spanning Set
we say that the set 13 spans. The set 13 is called a spanning set for the subspace.
vi,... ,vkEV, then is called the subspace spanned by 13 ={vi,... ,vk}, and
We denote by
In Sections
1.2, 3.1 ,and 4.1, we saw that the concept of spanning is closely related
Definition
={vi,... ,vk}
If 13
is a set of vectors in a vector space V, then 13 is said to be linearly
independent if the only solution to the equation
Linearly Independent
Linearly Dependent
is
Remark
The procedure for determining if a vector is in a span or if a set is linearly independent
in a general vector space is exactly the same as we saw for
Pn in
Sections 2.3, 3.1, and 4.1. We will see further examples of this in the next section.
PROBLEMS 4.2
Practice Problems
Al Determine, with proof, which of the following sets
are subspaces of the given vector space.
(a)
(b)
{ i
of
R4
M(2,2) 2 3
(c) {ao+aix
a1x a3x I ao+2ai=0,
ao,ai,a1,a3EJR.} of ?3
(d) {[: :]lai,a2,a3,a4EZ}ofM(2,2)
(e) {[: :]I a1a4 - a2a3=O,a1,a2,a3,a4 EJR.}
of M(2,2)
(f) {[ ] I ai = a1,ai, a1EJR.}
of M(2,2).
of
symmetric if
AT
A or, equivalently, if
i and j.)
P5
Ps
205
EJR}
Homework Problems
Bl Determine, with proof, which of the following sets
are subspaces of the given vector space.
(a)
(b)
{ i X1
I
+ X3 =x,,x,,x,,x,,x, ER
A
of
1!.4
o f M(2, 2)
} of
(d)
(e)
(f)
of
of M(2, 2)
of M(2, 2).
m []
m []
=
aij -a1i
P5 5
JR
=
P
2
Pa4x2} 4
{ao3
a1a4 EJR
{x
Conceptual Problems
Dl Let V be a vector space.
(a) Prove that -x=(-l)x for every x EV.
(b) Prove that the zero vector in V is unique.
(c) Prove that tO 0 for every t R
=
bE
D2 Let V = {(a, b) I
b > 0} and define
addition by
b) B (c, d) =(ad+ be, bd) and scalar
multiplication by t 0
b) = (tab1-1, b1) for any
t
R Prove that V is a vector space with these
operations.
a, JR,
(a,
(a,
D3 Let V = {x
I x > O} and define addition by
xy and scalar multiplication by t 0 x = x1
x
E JR
By.=
JR"
206
Chapter 4
Vector Spaces
(u1
(tu1,tV1)
U2, V1
V2)
Bases
11
Recall from Section 1.2 that the two important properties of the standard basis in IR.
were that it spanned IR.11 and it was linearly independent. It is clear that we should want
a basis 13 for a vector space V to be a spanning set so that every vector in V can be
written as a linear combination of the vectors in 13. Why would it be important that the
set 13 be linearly independent? The following theorem answers this question.
Theorem 1
a linear combination of the vectors in 13. Assume that there are linear combinations
This gives
which implies
0, so a;
207
Ov I
... + Ovn
Thus, if 13 is a linearly independent spanning set for a vector space V, then every
vector in V can be written as a unique linear combination of the vectors in 13.
Definition
Basis
set for V.
Remark
According to this definition, the trivial vector space
{O}
any set of vectors containing the zero vector in a vector space Vis linearly dependent.
However, we would like every vector space to have a basis, so we define the empty set
to be a basis for the trivial vector space.
EXAMPLE 1
{[ ], rn ], [ ], [ ]}
{ x, x2, x3}
(1, x,
set
EXAMPLE2
ove that the set C
{[i], m, [m
JR.3,
[X2XJ] t1 [11] t2 [1]]
X3
1
is a basis for
In particular, the
R3.
JR.3
JR.3
t3 [1] [ti tt12 t2 tt32 t3]
1] [1 1 l
JR.3.
JR.3.
JR.3.
P3.
can be written as a
0
1
0 - 0
1
Observe that the rank of the coefficient matrix equals the number of rows, so by The
orem 2.2.2, the system is consistent for every 1 E
Hence, Span C
Moreover,
since the rank of the coefficient matrix equals the number of columns, there are no
parameters in the general solution. Therefore, we have a unique solution when we take
208
Chapter 4
EXAMPLE3
Vector Spaces
Is the set 23
], [ n, [ ]}
= {[-
Solution: Since 23 is a spanning set for Span 23, we just need to check if the vectors in
23 are linearly independent. Consider the equation
[o o] [ 2] [o
i
ti
0 = -1
+t2
1
1
+t3
[ ][
2
1
ti +2t3
-ti+3t2+t3
2t1 +t2+St3
ti+t2+3t3
-1
Observe that this implies that there are non-trivial solutions to the system. For example,
one non-trivial solution is given by t1 = -2, t2 = - 1 , and t3 = 1 , and you can verify
that
(-2) _
+(- 1 )
[ J
+< l )
[ J=[ J
Therefore, the given vectors are linearly dependent and do not form a basis for Span 23.
EXAMPLE4
Is the set C
2
2
2
{3+2x+2x , 1 +x , 1 +x+x } a basis for P2?
[ i l [ o ol
3
2
2
1
0
1
1
0
1
0
1
0
0
l
Observe that this implies that the system is consistent and has a unique solution for all
ao + aix+a2x
EXERCISE 1
209
EXAMPLES
{p(x) E P I p(l) O} of P .
2
2
Solution: We first find a spanning set for and then show that it is linearly indepen
dent. By the Factor Theorem, if p(l) = 0, then (x- 1) is a factor of p(x). That is, every
polynomial p(x) E can be written in the form
p(x)
(x - l)(ax +b)
a(x2 - x) +b(x
1)
Span{x2 - x, x - 1 }. Consider
t1 t
0. Hence, {x2 - x, x - l} is linearly independent. Thus,
2
{x2 - x, x - 1} is a linearly independent spanning set of and hence a basis.
{v1,
We want to choose a subset of'T that is a basis for V. If'T is linearly independent, then
Tis a basis for V, and we are done. If Tis linearly dependent, then t1 v1 +
t;
+tkvk 0
0. Then, we
v; to get
a1v1 +
+akvk
a1V1 + ... +a;-1V;-1 +a; - (t1V1 +... +t;-1Vi-J +t;+1V;+1 +... +tkvk)
+a;+1V;+1 +
akvk
T\{v;}. This
T\{v;} is a spanning set for V. If T\{vd is linearly independent, it is a
basis for V, and the procedure is finished. Otherwise, we repeat the procedure to omit
a second vector, say
v1, and get T\{v;, v1}, which still spans V. In this fashion, we
must eventually get a linearly independent set. (Certainly, if there is only one non-zero
vector left, it forms a linearly independent set.) Thus, we obtain a subset of Tthat is a
basis for V.
210
Chapter 4
EXAMPLE6
Vector Spaces
If T
1
[
[
[
[
]
1
=
=
[
]
]
]
[ol
Solution: Consider
0
ti
-2
1
+ t2 -1 + t3 -2 + t4 5
3
ti + 2t2 + t3 + t4
ti - t2 - 2t3 + 5t4
-2t1 + t2 + 3t3 + 3t4
-11
ti
The general solution is
t2
t3
t4
= 1
s
-1
s E R Taking s
.
1, we get
t1
t2
t3
=
l
Ul-Hl+H Hl -UJ+:I
[ ]
Tl{[-I}=
{fJHJ.rm
, which
gives
or
Now consider
The matrix is the same as above except that the third column is omitted, so the same
row operations give
[ - 1 r 11
===
-2
t2
t3
{[_iJ.l-:].[i]}
is
EXERCISE 2
Let 'B
{1
211
Span 'B.
Dimension
n
We saw in Section 2.3 that every basis of a subspace S of R contains the same number
of vectors. We now prove that this result holds for general vector spaces. Observe that
the proof of this result is essentially identical to that in Section 2.3.
Lemma2
, v,,}
V. If {u1 ,
. .
, uk} is a lin
n.
U1
U2
Since 0
t1u, + + tkuk
t1(a11V1 + az1V2 + + a111Vn) + + tk(a1kV1 + azkV2 + + a,,kvn)
(a11t1 + + alktk)v1 + + (a,,1t1 + + a,,ktk)v11
O v1 + + O v,, we can write this as
Theorem 3
If 'B
{u1,
n.
k = n.
Proof: On one hand, 'Bis a basis for V, so it is linearly independent. Also, C is a basis
for V, so Span C
k. Therefore,
k, as required.
k. Similarly, C is linearly
212
Chapter 4
Vector Spaces
As in Section 2.3, this theorem justifies the following definition of the dimension
of a vector space.
Definition
Dimension
If a vector space V has a basis with n vectors, then we say that the dimension of V is
n and write
dimV = n
0.
If a vector space V does not have a basis with finitely many elements, then V is called
Remark
EXAMPLE 7
x n vectors.
{ 1 , X, x2,
. .
, X1}.
EXAMPLES
Ut S =Span
2.
Solution: Consider
4
[1 -1 -li4 [0 01 6]
-1
0000
[-l r!l
{ li].[=]}
3
2
2
3
-2
and
Span
Moreom, B =
m].[=l}
is
213
EXAMPLE8
clearly linearly independent since neither vector is a scalar multiple of the other, hence
(continued)
EXAMPLE9
Let
{[: ]
Solution: Since d
2.
M(2, 2) I a+ b
=
Thus,
{[ ] [ n , [ ]}
,
EXERCISE 3
3.
P3 I a+ b + c + d
O}.
(4.1)
If tk+I * 0, then we have
Wk+I
t1
--V1 tk+I
' '
tk
- -Vk
tk+!
and so Wk+! can be written as a linear combination of the vectors in T, which cannot
be since Wk+! <t Span T. Therefore, we must have tk+I
tk 0. Thus, t1 v1 +
+
tk+I
0, and hence {v1,
, vb wk+!} is
linearly independent. Now, if Span{v1,
, vb Wk+d
V, then it is a basis for V. If
not, we repeat the procedure to add another vector Wk+2 to get {v1, ... , vb Wk+! Wk+2},
tkvk
tk+! Wk+I
0 implies that t1
which is linearly independent . In this fashion, we must eventually get a basis, since
according to Lemma 2, there cannot be more than n linearly independent vectors in an
n-dimensional vector space .
214
Chapter 4
EXAMPLE 10
Vector Spaces
Let C
{[ ], [- -]}
Solution: We first want to determine whether C is a spanning set for M(2, 2). Consider
-2
-1
1
1
0
1
b1
b2
b3
b4
1
0
0
0
-2
1
0
0
b1
b2 - b1
b1 - b2 + b3
2b1 - 3b2 + b4
[ ]
[ ]
Hence, by the procedure above, we should add this matrix to the set. We let
1
1
0
-2
-1
1
[: :]
0
0
1
0
0
0
-2
1
0
0
0
0
1
0
bi
b2 - bi
b1 - b2 + b3
2b1 - 3b2 + b4
is not in the span of '131 and thus '131 is not a basis for M(2, 2). Adding
get
rn ]
[ ]
to '131 we
M(2, 2).
EXERCISE 4
Extend the set 'T
{[; ]}
to a basis for
215
R3.
Knowing the dimension of a finite dimensional vector space Vis very useful when
trying to invent a basis for V, as the next theorem demonstrates.
Theorem 4
(3) A set with n elements of Vis a spanning set for V if and only if it is linearly
independent.
Proof:
< n
then it has a basis containing n vectors. This means we have a set of n linearly
independent vectors in a set spanned by k
<
(3) If 13 is a linearly independent set of n vectors that does not span V, then it can
be extended to a basis for V by the procedure above. But, this would give a
linearly independent set of more than n vectors in V, which contradicts (1).
Similarly, if 13 is a spanning set for V that is not linearly independent, then
by the procedure above, there is a linearly independent proper subset of 13 that
spans V. But, then V would be spanned by a set of fewer that n vectors, which
contradicts (2).
EXAMPLE 11
IR3
with equation
x1 +
2x2
0.
JR3.
Solution: (a) We know that a plane in JR3 has dimension 2. By Theorem 4, we just need
to pick two linearly independent vectors that lie in the plane. Observe that ii 1
and
v,
[I
both satisfy the equation of the plane and neither is a scalar multiple
plane P.
(b) From the procedure above, we need to add a vector that is not in the span of {v 1,
v2}.
But, Span{v 1, v2} is in the plane, so we need to pick any vector not in the plane. Observe that
V3
plane. Thus,
[j
does not satisfy the equation of the plane and hence is not in the
216
Chapter 4
Vector Spaces
EXERCISE 5
x1 - x2 + X - 2x4 = 0
3
and
PROBLEMS 4.3
Practice Problems
Al Determine whether each set is a basis for
(a)
{[il-l=: l [;]}
WlHl}
minrnlHl}
wirnrnJ}
{[-:].[J[]}
JR3.
(b)
(c)
(d)
(e)
{l 1}
(b)
(c)
A2 Ut
B=
' 1 '
3
(a)
(b)
B and determine
B=
B=
for
JR3.
Prove that
is a
{HJ.r!HH lm
{[l r=i flrnHm
(a)
[ -]}
11, determine a
x1 - x2 +X - X4 = 0
3
(a) S ={a+bx+cx2
(b) S =
(c) s =
sets
11, determine
2x1 - x2 - X = 0 in JR3.
3
2
3
B.
B={l+x,l+x+x2,l+x3}
B= (1+x,1 - x, 1+x3,1 - x3}
B = {1 + x + x2,1 - x3,1 - 2x + 2x2 - x3,
1 - x2+2x3,x2+x3}
polynomials spanned by
(a)
][ -][ =J.
{[ ]
P2 a= -c}
I
M(2,2) a,b,c
I
[{ :] [:l [i] o}
(d) S ={p(x)
(e) S
R3
P2 p(2) = O}
{[ !]
I
M(2,2) a= -c}
I
IR}
217
Homework Problems
Bl Determine whether each set is a basis forJR3.
(a)
{[-il Ul Ul}
(a) 3=
(b)
(c)
(d)
(e)
{[J .nrnrnJ}
{[] HJ.lm
mHm
{[J[J.Ul
}
-
(b) 3=
(a) 3= { +
1 x+x2,x+x2+x3, 1-x3}
x+ x x2+x3, 1 x2+x3}
11,
x1 3x2 4x3
(b) 3= {l+
(c) {
(d) {
(a)
(b)
=
B
{ [l .UJ rm
{[ :i [-irnl .ui Hll
_
determine a
=0 inJR3.
for JR3.
(b) {
{[ ][ !] [! -J.
M(2, 2).
[! -n.
P2.
+x+x2, 1 -x2}
11++x2,x2,21-xx x2,2x2,-x-2-x2,2x}1+2x2}
3+2x+2x2,1+x+x2,1 -x - x2}
(a) {1
x2,
B2 Let3=
inJR4.
determine a
11,
x1 x2+2x3+ =
+
X
4
S {[ ]EM(2,2) Ia,bE }
S { x2)p(x) p(x)EP2}
{[;;] R3 I[;;] Ul o}
S={[; :]EM(2,2)1a-b=O,
= = JR}
S {p(b -x) EPs I-p(a-x) Ep(x) x}
(a)
(b)
JR
= Cl+
(c) s =
(d)
(e)
0, c
for all
Conceptual Problems
Dl (a) It may be said that "a basis for a finite dimen
this section.
=V.
= n, then
218
Chapter 4
Vector Spaces
{(a, b) I
[]
However, it would be awkward to determine the standard matrix of this stretch and then
determine its effect on any other vector. It would be much better to have a basis that
takes advantage of the direction
[]
[-l
3
0 in JR . It is easy
lj] [ ]
[-] [}
to
and
and leaves
See Hg
ure 4.4.2. Describing this reflection in terms of these vectors gives more geometrical
information than describing it in terms of the standard basis vectors.
z
Figure 4.4.2
x1 + 2x2
3x3
0.
Notice that in these examples, the geometry itself provides us with a preferred
basis for the appropriate space. However, to make use of these preferred bases, we
219
Definition
Coordinate Vector
x1 v1
x2v2
basis <J3 is
[X]!B =
Xn
Remarks
1. This definition makes sense because of the Unique Representation Theorem
(Theorem 4.3.1).
2. Observe that the coordinate vector [x]!B depends on the order in which the basis
vectors appear. In this book, "basis" always means ordered basis; that is, it is
always assumed that a basis is specified in the order in which the basis vectors
are listed.
3. We often say "the coordinates of x with respect to <J3" or "the <J3-coordinates
of x" instead of the coordinate vector.
EXAMPLE 1
{[], []}
[;]
[] [] [;l
1 and a2
2. Thus,
+ a2
[ ] [ ] [-]
+
b2
is b1
3, b2
and
[-]
By inspection,
-2. Hence
Figure 4.4.3 shows JP?.2 with this basis and a. Notice that the use of the basis <J3 means
that the space is covered with two families of parallel coordinate lines, one with direction vector
[]
[n
relative to these two families. The axes of this new coordinate system are obviously
not orthogonal to each other. Such non-orthogonal coordinate systems arise naturally
in the study of some crystalline structures in material science.
220
Chapter 4
Vector Spaces
[]
/
Figure 4A.3
EXAMPLE2
{[; ], [ ], [ ]}
[ ]
[ ]
and B
Solution: We are required to determine whether there are numbers u1, u2, u3 and/or
v,, v2, v3, such that
UI
VI
[; ] [ OJ [ ] [ 1]
[; ] [ ] [ ] [ ]
+ Uz
+ U3
+ V3
+ Vz
Since the two systems have the same coefficient matrix, augment the coefficient matrix
twice by adjoining both right-hand sides and row reduce
2
2
-1
6
-3
-9
-8
It is now clear that the system with the second matrix B is not consistent, so that B is
not in the subspace spanned by '13. On the other hand, we see that the system with the
first matrix A has the unique solution u1
I, u2
1, and u3
-3. Thus,
Note that there are only three '13-coordinates because the basis '13 has only three vectors.
Also note that there is an immediate check available as we can verify that
221
[1
EXAMPLE2
(continued)
EXERCISE 1
Determine the coordinate vector of
Span 'B.
EXAMPLE3
=l
+l
{[_:l [ m
of
Suppose that you have written a computer program to perform certain operations with
polynomials in
3"3,-5,
{1,
x, x2} for P ,
2
5x +2x2, you would surely write your program in such a way that you would type
2",
{1 - 1
3-
On the other hand, for some problems in differential equations, you might prefer
t1, t
and
t3
x2, x,
x2}.
such that
5x +2x2, we must
[ 01 01 01 -531
-1 0 1 2
3-5x
[3-5x
+2x2].IB
"0.5, 5 2.5".
/2
Thus, if your computer program is written to work in the basis 'B, then you would input
- ,
We might need to input several polynomials into the computer program written
to work in the basis 'B. In this case, we would want a much faster way of converting
standard coordinates to coordinates with respect to the basis 'B. We now develop a
method for doing this.
Theorem 1
Let 'B be a basis for a finite dimensional vector space V. Then, for any
t E JR we have
[tx + Y]lB
=t[x]lB +[Y]lB
x, y E V and
222
Chapter 4
Vector Spaces
Proof: Let :B = {v1, ... , Vn}. Then, for any vectors x = X1V1 + + x11v11 and
y =Y1V1 + +y11v11 in V and any t E JR, we have tx+y = (tx1 +y1)V1 + +(tx11+y11)v11, so
tx1 + Y1
[tx + Y]!B =
tXn +Yn
[lt
Xl
=t:
+:
Xn
=t[x]!B+[ y]!B
11
Let :Bbe a basis for an n-dimensional vector space V and let C = {w1,
, w11} be
Xl
:
Xn
Since
[I]
= [x]c, we see that this equation gives a formula for calculating the
Definition
Change of Coordinates Matrix
[ [w1]!8
[x]!B = P[x]c
Of course, we could exchange the roles of :Band C to find the change of coordi
nates matrix Q from :B-coordinates to C-coordinates.
Theorem 2
is invertible and p-
C-coordinates.
EXAMPLE4
Let S = [ t,, l,, t3} be the standard basis for JR3 and let '1l =
{[_].[:l [;l}.
Find the
EXAMPLE4
(continued)
223
Solution: To find the change of coordinates matrix Q, we need to find the coordinates
of the vectors in 13 with respect to the standard basis S. We get
To find the change of coordinates matrix P, we need to find the coordinates of the
standard basis vectors with respect to the basis 13. To do this, we solve the augmented
systems
-1
10 l ' [
00 l , [
10 01 00 [ 01 01 00
00 l 00
1-11
[
-1
-1
0l
To make this easier, we row reduce the triple-augmented matrix for the system:
[_l
T hus,
3/5
3
4
3/5
P=
EXAMPLES
1 1
7/5
-4/5
7/5
-4/5
3/5
-1/5
-1
7/5
-4/5
3/5
-1
-4/5
i[
-1/5
-4/5
3/5
-1/5
-4/5
3/5
1
3
-1
1l
-1
-1
] 0 ]
=
Let 13
[a+bx+ex']=P[a +bx+cx']s =P
Hence, we just need to find the coordinates of the standard basis vectors with respect
to the basis 13. To do this, we solve the three systems of linear equations given by
224
Chapter 4
Vector Spaces
EXAMPLES
(continued)
Hence,
la+
2 le
2
EXERCISE 2
Let s
= {[i].[].[m.
Find the
/.
PROBLEMS 4.4
Practice Problems
Al In each case, check that the given vectors in B are
linearly independent (and therefore form a basis for
x y
{ }
dinates of
and
with respect to B.
= i : x= = i
2
1
0
=1
{ 1y=
x= 2
= {[ ] , [ ], [ ] } x = [ l
y= [- !]
B={[ 1 ],[ =]}
x=[ ! -ly=[- = ]
(a) B
(b) B
(c) B
(d)
=1
{ 2x4}, x1=2
y=1
= ml [J}
=0.
(e) B
plane 2x1 - x - 2x
[ l
[] [ ]
=1
{ 1
(ii)
(iii)
+ x2,
is a basis for P .
x+ 2x2, -1
x+ x2}
Section 4.4
:B
of the
following polynomials.
(i)
(ii)
(iii)
p(x)=1
q(x)=4 - 2x+7x2
r(x)=-2 - 2x+ 3x2
[2 -4x+ l0x2}8
;] ' [ -] ' [
[ -]
:B={ [
A=
(c) Determine
:B,
vector space.
(b)
Cc)
:B.
for
R3
=
:B={ 1 - 2x+5x2, 1 - 2x2,x+x2}
= {[ =n, [ =J, [ ]}
2
]}'
(b)
225
Exercises
:s
tor space of
for P
2 upper-triangular matrices
:B.
:B
- 2
[
]. 1 0]}'
:B
determine [A].
(a)
:B={[
[
-
A=
][-
]
and A
Span
:B,
Homework Problems
Bl In each case, check that the given vectors in
:B
are
{ 1i :1 } = j1 = !1
dinates of
(a)
(b)
(c)
(d)
:B=
and
with respect to
:B.
(e) :B=
x
{[] [W
(i)
(ii)
[i1
[1
= {UrnHm
(iii)
is a basis
for JR3.
(b) Determine the coordinates relative to :B of the
following vectors.
(i)
(ii)
[J
226
Vector Spaces
Chapter 4
BS
Ul. m. Ul.
A E
B4
P .
2
p(x)
3 +4x+ Sx2
q(x)
4+ Sx - 7x2
r(x)
1+x+x2
[4
B6
Sx + 6x2]2:1
(-1+2x2,1+x+x2,l-x-3x2)forP
Conceptual Problems
Dl
v;
D2
{v1, ,vk}
C
{w1,
,wk}
x E V, [x]2:1 [x]c.
w;
1 $ i $ k?
.
{v,,v ,V3,V4}.
2
V.
{v3,V ,V4,vi}
2
P[x]2:1
[x]c.
2.
227
Definition
Linear Mapping
Lx
(
y)
W is a linear mapping if it
L(x)
= + L(y)
+= tL(x)
L2 L(tx)
L(tx
y)
tL(x)
+ L(y)
EXAMPLE 1
Let L : M(2, 2)
P2 be defined by L
d)x
(b
([; ]) = + + +
2
ax . Prove that Lis a
linear mapping.
[;; :], [; ]
( [: :]+[ ])=L([:: ::])
= + + + + +
+ +
= + + +
+ + + +
L t
(td1
d2)
(ta1
t(d1
(tb1
2
a2)x
b2
td1
d2)X
2
a1x )
2
d2)x a2x )
d1)x
(b1
(d2
(b2
a1x
2
a1x )
So, L is linear.
EXAMPLE2
Let M : P3
P3 be defined by M(ao
linear operator.
a1x
=ao + +
+ =
=
=
M(tp(x)
q(x))
2
a2x , q(x)
M((tao
(ta1
t(a1
Hence, M is linear.
a1
+ + = +
=ho+ +
+ho)+ + + +
+ + +
+ + +
+
(ta1
2(ta2
b1)
2a2x)
tM(p(x))
b1x
(b1
M(q(x))
2
b2x , and t ER Then,
b1)x
b2)x
2b2x)
(ta2
2
b2)x )
228
Chapter 4
Vector Spaces
EXERCISE 1
Let
L R3
M(2, 2) be defined by
linear.
= Xi+XX22 + X3]
[[:ll [6
. Prove that
L.
1s
Remark
Since a linear mapping
L :V
scalar multiplication, and composition ) and the concept of invertibility on these more
general linear mappings in the same way as we did for linear mappings
]Rn
JR.111
Definition
Let
Range
and Wbe vector spaces over JR. The range of a linear mapping
L :V
Wis
={L(x) E I x EV}
Nulls pace
Range(L)
The nullspace of Lis the set of all vectors in V whose image under Lis the zero vector
Ow. We write
Null(L)
EXAMPLE3
Let
: M(2, 2)
P2
L([; ])=c +(b + d)x+ax2.
+x+x2
L,
A
L(A)= +x + x2.
a, b, d
l+x+x2=L([ !])=c+(b+d)x+ax2
c= b+d = a =
+x+x2 E
A
L(A)= +x+x2 A=[ ].
be the linear mapping defined by
Determine whether 1
is in the range of
1
Solution: We want to find
such that
1,
and
such that
1, and
such that
is
EXAMPLE4
Let
L : P2
Determine whether
[;]
is in the range of
Range(L)
L.
[a -bl
L(a + bx + cx2) b-c .
c-a
EXAMPLE4
229
(continued)
1,
1 l [ 1 -1 1
b-
= 1 , and c - a = 1 . Row
1
1
0
0
-1
1
Theorem 1
0
0
[;]
0
-1
0
1
0
1
3
( 1 ) L(Ov) = Ow
(2) Null(L) is a subspace of V
(3) Range(L) is a subspace of W
EXAMPLES
[ ]
Determine a basis for the range and nullspace of the linear mapping l : P1
defined by L(a+bx) =
a- 2b
Solution: If a+bx
]
[
l(a+bx) =
a- 2b
[ ]
IR.3
Hence, a = 0
and a- 2b = 0, which implies that b = 0. Thus, the only polynomial in the nullspace
of l is the zero polynomial. That is, Null(l) = {0}, and so a basis for Null(L) is the
empty set.
I l [l [ 1
{ [l [J]}.
=a
+b
a- 2b
-2
230
Chapter 4
EXAMPLE6
Vector Spaces
Determine a basis for the range and nullspace of the linear mapping L M(2, 2) P2
defined by L([; ]) =(b + c) + (c - d)x2.
Solution: If[: ] Null(L), then 0 =L([: ]) =(b + c) + (c - d)x2, sob + c =0
and c - d =0. Thus, b =-c and d =c, so every matrix in the nullspace of L has the
form
:
EXERCISE 2
Determine a basis for the range and nullspace of the linear mapping L : JR3
=[X Xi+ X3 X2 X1+ X3 ]
defined by L
[[ii
X3
M(2, 2)
Observe that in each of these examples, the dimension of the range of L plus the
dimension of the nullspace of L equals the dimension of the domain of L. This result
reminds us of the Rank Theorem (Theorem 3.4.8). Before we prove the similar result
for general linear mappings, we make some definitions to make this look even more
like the Rank Theorem.
Definition
Rank of
a Linear Mapping
Definition
Nullity of
a Linear Mapping
is
is
Theorem 2
231
[Rank-Nullity Theorem]
Let V and W be vector spaces over IR'. with dim V
n, and let L : V
-t
W be a linear
mapping. Then,
rank(L)
nullity(L)
Proof: The idea of the proof is to assume that a basis for the nullspace of L contains
k vectors and show that we can then construct a basis for the range of L that contains
n k vectors.
Let 13 = {v1,...,vd be a basis for Null(L), so that nullity(L) = k. Since V is
-
n-dimensional, we can use the procedure in Section 4.3. There exist vectors Uk+l ... ,u,,
such that {v1,...,vb Uk+l ... ,Un} is a basis for V.
Now consider any vector win the range of L. Then w = L(x) for some x
But any x
V.
t1V1
+ ... +
tkvk
W =
=
L(t1V1
t1L(v1)
tkVk + tk+IUk+I
+ tkL(vk)
tk+1L(uk+1)
=
tnU11)
+
tnL(un)
0, and so we have
independent? We consider
t;
nullity(L)
(n - k) + k
k and
232
Chapter 4
Vector Spaces
P3
([ :])=cx+(a+b)x3
EXAMPLE7
M(2,2)
defined by
[ :]
Null(L), then
= ([ :])=cx+(a+b)x3
a +b = c =
O
Hence, we have
Therefore, Null(L)
{[ -J, rn ]}
0 and
Span
{[ -] [ ]}
,
nullity(L) 2.
Clearly a basis for the range of L is
since this set is linealy independent
and spans the range. Thus rank(L) 2.
Then, as predicted by the Rank-Nullity Theorem, we have
=
{x, x3},
rank(L)
nullity(L)
=+ =
2
dim M(2,2)
PROBLEMS 4.5
Practice Problems
Al Prove that the following mappings are linear.
IR3 IR2
x x2, x3)
)
+
(X1 +X2, X\ X2+X3
P,
m J = (a+ b)+
(a+b+c)x
IR
([ :])=a+d
(a) L :
(b) L
R'
defined by L(
1,
defined by L
defined by tr
T : P3
cx2+dx3)= [ ]
T(a +bx+
linear.
IR
ad-be
P 2 P2 2
(a-b)+(b+c)x
T: IR2
[l :2]
M([ ])= [ ]
defined by
([ :]) =
L(a + bx + cx2) =
r([])=
defined by det
M(2,2) defined by
(d) M : M(2,2)
M(2,2) defined by
[-
([:ll
Xt +X3
y=
X2 - X4
[= -]
0
0
2 x2 - 2x3 - 2x4
-2Xt
2
0
y=
0
2
(b) L: P2 M(2,2) defined by L(a+bx+cx ) =
c
b
] [ ]
c .
([ll
(a+b+c)x
(c) tr : M(2,2) IR. defined by tr
y=
2
(c) L : P2 Pt defined by L(a + bx + cx ) =
(b+c)+ (- b - c)x,y = 1 +x
(d) L: JR4 M(2,2) defined by L
[]
233
= (a+b)+
([; ])
=a+d
[; ]
Homework Problems
Bl Prove that the following mappings are linear.
2
(a) L: P2 JR3 defined by /Aa+bx+cx ) =
([ !])
([
([;
a+ (a+b)x+bx
[;= =:]
])
!])
linear.
2
(a) L: P3 IR. defined by L(a+bx+cx +dx3) =
a
b
=0
matrices; L : JD P2 defined by L
([; ])
c
d
2
(b) M : P2 IR. defined by M(a + bx+ cx ) =
2
b - 4ac
(e)
N : JR3
Xl
X3 ]
mm
(d) L :
[ ;]
(e) T
M(2,2) defined by
M(2,2) P2 defined by T
2
a+ (ab)x+ (abc)x
([ ])
a
--
234
Vector Spaces
Chapter 4
R3
-
y.
L([::Jl =
l [ 1
:\ :
y=l+x- x2
---t
l [-1
a+b +c'
y=
[l l [ ]
(c) L:M(2,2)---tlR.defined by L
[:= =]
([])=
y=
([; :]) =
(a+b +c + d)x2
L([ ])= -
( a- c) +(a - 2b)x+
] [-
(g) T : M(2,2)
[ ;: :
(h) L: P
---t
-: = ]
(f) L
ex')= m
(
)=
([; :])=o
L([ ]) =
L([; :]) =
R3 defined by L(a+bx+
2
b + 2cx
(c) L : P
2
Theorem.
(b) L : P
y=
1
-2x1 + x - 2x3
2
(b) L : P ---t P defined by L(a + bx + cx2) =
2
2
(-a
2b) + (2a + c)x + (-2a + b - 2c)x2,
(a) L P,
R3 defined by
Conceptual Problems
Dl Prove Theorem 1.
( [ll
([ll
= x', L
([]l
= 2x
rank(L)= 2, and L
---t
inW.
Range(L)
{O} and
---t
{O}.
---t
L) $ rank(M).
L) $ rank(L).
mappings.
2,
in V but {L(v1),
let L : V
=
{[ J.rn J.rn ]}
=
=
([ ])
JR.4; nullity(L)
0
Range(L)=span
(c) V=M(2,2), W
in V.
W, where {v1,... , vd is linearly independent
= I +x+x2
then {v1,
of the ranks.
235
L(x1, x2,x3, ...) = (x2,x3, X4, ... ) and the right shift
=
(x1, x2,...,x11,). Define addition x + y and scalar
multiplication tx in the obvious way. Then S is a
infinite-dimensional.
{e1,..., e11} of IR.11 under L. We now look at the matrix of L with respect to other bases.
We shall use the subscript S to distinguish the standard mat1ix of L:
11
The standard coordinates of the image under L of a vector 1 E IR. are given by the
equation
[L(x)]s = [L]s[1]s
(4.2)
These equations are exactly the same as those in Theorem 3.2.3 except that the notation
is fancier so that we can compare this standard description with a description with
respect to some other basis. We follow the same method as in the proof of Theorem
Observe that
bn
tion (4.2) where the matrix
standard matrix of L.
[ [L(v1)]23
[L(v 1)J23
1
11
IR.11 is a linear
Definition
Suppose that 13 = {v1, ... 'v,,} is any basis for IR. and that L : IR.11
operator. Define the matrix of the linear operator L with respect to the basis 13 to be
the matrix
---+
236
Chapter 4
Vector Spaces
Note that the columns of [L].:B are the .3-coordinate vectors of the images of the
.3-basis vectors under L. The pattern is exactly the same as before, except that
everything is done in terms of the basis .3. It is important to emphasize again that
by "basis," we always mean ordered basis; the order of the basis elements determines
the order of the columns of the matrix [Lhi.
EXAMPLE 1
Let L : IR.
and let :B =
3
IR. be defined by L(x1,x2,X3) = (x1 +2x2 - 2x3,-x2 +2x3,X1 + 2x2)
{[ :J [: l [j]}
[H
where [1].=
Solution: By definition, the columns of [L].:B are the .3-coordinates of the images of
the vectors in .3 under L. So, we find these images and write them as a linear combi
nation of the vectors in .3:
L(2.-l.-l)=
L(l.l. l)=
L(0,0,-1) =
l-l
m
Hl
l=:l m (-1) ui
l=!l (1) m ui
(4/3) [=:l (-2/3) [:] Ul
+(0)
=(l)
=(0)
[L(l,
1,
1)].:B
[L(0,0,-l)J.:B]
[-1 -1
[-1 -4/2/33] [13] [-115]
-2
-2
[L(x)].:B=
1
0
2 =
-2
[]
+ (-2)
Thus,
+(-2)
Hence,
-2
We can verify that this answer is correct by calculating L(x) in two ways. First, if
[1].=
ilien
1=1
l :l l:J il-l Ul
=
+2
EXAMPLE 1
[ ]
(continued)
[L(x)]!B
L(4, 1, -2)
237
.then
-11
EXAMPLE2
Let v
[! ]
.In Example 3.2.5, the standard matrix of the linear operator projv
[pWJ ;t ]S
9/25
12/25
12/25
16/25
JR2
Find the matrix of projv with respect to a basis that shows the geometry of the trans
formation more clearly.
Solution: For this linear transformation, it is natural to use a basis for JR2 consisting
of the vector v, which is the direction vector for the projection, and a second vector
orthogonal to v, say w
[]
r-J.
Then, with 13
projv v
projv
proj" w
proj"
1 v +Ow
[!] [!]
[-J []
rnl
Ov +Ow
Thus,
JR2. We have
r 1 r1 r 1
=
[ll
[]
to
adapted to the geometry of the transformation (see Figure 4.6.4 ) This example will be
discussed further below.
238
Chapter 4
Vector Spaces
Vz
Figure 4.6.4
V - V on a vector
space V with respect to any basis '13 for V. To do this, we repeat our argument above
exactly.
Let '13 = {v1, ... , v,,} be a basis for a vector space V and let L: V - V be a linear
operator. Then, for any x EV, we can write x = biv1 +
+ b11v11 Therefore,
[L(x)]21 = [b1L(v1) +
= b1 [L(vi)]21+
= [L(v1)l
+ b11L(v11)]21
+ bn[L(v,,)]21
[L(v,)]
f]
Definition
Suppose that '13 = {v1, ... , v11} is any basis for a vector space V and that L: V - V is
a linear operator. Define the matrix of the linear operator L with respect to the basis
'13 to be the matrix
[L(x)]21 = [L]21[x]21
EXAMPLE3
Section 4.6
EXAMPLE3
239
Solution: We have
(continued)
L(l)=l+x2
L(x) =1 +x +x2
L(x2) =x2
Hence,
1 1
L
[ (a +bx+cx2)]1l = [L]1l[a+bx+ cx2]1l = 0 1 0
1 1 1
EXAMPLE4
a+b
b
and
a+b+c
a
a+b
b =
b
c
a+b+c
l[ ] [
Let L: P
P
be defined by L(a+bx+cx2) =(a+b)+bx+(a+b+c)x2 . Find the
2
2
matrixof L with respect to the basis '13 ={l+x+x2,1 - x,2}.
Solution: We have
EXERCISE 1
Let L
([ ]) [
] .
{[ ].[ ][ ][ ]}.
M2
( ,2) M2
( ,2) be defined by L
a
c
a+b
c
a-b
. Fmd the
a+b+d
Observe that in each of the cases above, we have used the same basis for the
domain and codomain of the linear mapping L. To make this as general as possible,
we would like to define the matrixcL
[ 1l
] of a linear mappingL: V
W, where '13 is
a basis for the vector space V and C is a basis for the vector space W. This is left as
Problem D3.
240
Chapter 4
Vector Spaces
2
2
IR. - IR. In some applications of these ideas,
the geometry does not provide such a simple way of determining [Ls
] . We need a gen
eral method for determining [L]s, given the standard matrix [Ls
] of a linear operator
L : IR.11 - IR.11 and a new basis 13.
Let L : IR.11 - IR.11 be a linear operator. Let Sdenote the standard basis for IR.11 and
let 13 be any other basis for IR.11 Denote the change of coordinates matrix from 13 to S
by P so that we have
P[xs
]
[x]s
and
If we apply the change of coordinates equation to the vector L(x) (which is in IR.11),
we get
[L(x)s
J
P-1 [L(x)s
J
But, P[x]s
[xs
] , so we have
Since this is true for every [x]s, we get, by Theorem 3.1.4, that
[Ls
]
p-I [L]sP
Thus, we now have a method of determining the 13-matrix of L, given the standard
matrix of Land a new basis 13.
We shall first apply this change of basis method to Example 2 to make sure that
things work out as we expect.
EXAMPLES
Let v
[ !]
metrically adapted basis 13. Let us verify that the change of basis method just described
does transform the standard matrix [projil ]s to the 13-matrix [projil ]s.
The matrix [projil s
]
[ {; } ;n
[! ]
2
The basis 13
3/25
_4125
{[!J, [-]}
so the change
4/25
. Hence, the .v-matnx
'
. given
.
by
. of proJ. v 1s
3125
-
p 1 [proJ,iJsP
3/25
-4/25
4/25
3/25
][
9/25
12/25
To make sure we understand exactly what this means, let us calculate the
EXAMPLES
(continued)
241
[]
two ways.
Method 1. Use the fact that [projv x]3
of 1:
[5]2
.
Hence,
[proJv 1]3
[proJi13
] [x3
]
[proJv x]s
Therefore,
[projv]3[x3
] . We need the B-coordinates
1
p-
p-1[projv x]s:
[proJv ]s
[proJv 113
The calculation is probably slightly easier if we use the first method, but that re
ally is not the point. W hat is really important is that it is easy to get a geometrical
understanding of what happens to vectors if you multiply by
[ ]
(the B-matrix); it
[ 1;;; ;;]
(the standard matrix). Using a non-standard basis may make it much easier to under
stand the geometry of a linear transformation.
EXAMPLE6
[; n
{ [] , [-]}
[ -n
Let B
be
a basis for IR.2. Find the matrix of L with respect to the basis B.
Solution: The change of coordinates matrix Pis P
[ ].
_
p-1AP
EXAMPLE 7
Let Lbe the linear mapping with standard matrix A
[L]3
'B
{[i], m, [:]}
-3
7
-7
95
-5
-5-3
1
242
Chapter 4
Vector Spaces
EXAMPLE 7
Solutiom The change of coordinates matrix P is P
(continued)
p-1
[i n
and we have
-1
[L]23
p-1AP
2
0
o
-3
0
o
0
Observe that the resulting matrix is diagonal. What does this mean in terms of the
geometry of the linear transformation? From the definition of [L]23 and the definition
of 13-coordinates of a vector, we see that the linear transformation stretches the first
basis vector
vector
vector
[:]
[:1
[i]
in the origin and stretches it by a factor of 3, and it stretches the third basis
by a factor of 4. This gives a very clear geometrical picture of how the linear
transformation maps vectors. This picture is not obvious from looking at the standard
matrix A.
At this point it is natural to ask whether for any linear mappingL : JR" JR" there
exists a basis 13 of JR" such that the 13-matrix ofLis in diagonal form, and how can we
find such a basis if it exists?
The answers to these questions are found in Chapter 6. However, in order to deal
with these questions, one more computational tool is needed, the determinant, which
is discussed in Chapter 5.
PROBLEMS 4.6
Practice Problems
Al Determine the matrix of the linear mappingL with
respect to the basis 13 in the following cases. Deter
mine [L(x)]23 for the given [1]23.
2
{v1,v2} andL(v1) v2,
(a) In JR , 13
=
L(v2)
[]
[XJ.
=
Ul
{ [ ] , [-;]}
2
of JR . In each
3
(b) In JR , 13
W1,v2, v3} and L(v1) 2v1 - \13,
L(v2) 2v1 - v3,L(v3) 4v2 + sv3;
=
.
[
-!
}
]
]
ml n
{v 1, v2,v3}
3
of R . In each of the follow-
ll: 11 l n l 11 m ll-rn ui
2
(a) L
(b) L
(c) L
mlHl[!lJinR'.
1, .
L(l,-1,0)=(0, L(l,0,l,2)1)==(2(,-2,0)
L(l,
[j]
s= {[j]. [] [m
L(l,0,-1) = (0,1,1),
basiss=
JR..3 is a linear
2, 4),
1,2),andL(O,
Determine the 13-matrix ofL.
(c) Use parts (a) and (b) to determine
in
basis
2, 4).
R'.
0) = 0,
L(112)
= [-]
1,1) =
L( l , 2,
(-2, 2), and L(O,
(5, 3, -5). Determine the 13-matrix ofL.
(c) Use parts (a) and (b) to determineL(5,3, -5).
(a)
(b)
(c)
(d)
(e
(fj
[ ;].13={[].[!]}
l-! =n ={[ ;1 r n
[ 6 -n- 2 0 ={[].[.]}
[ - 6 ].13={[ ] [ ]}
-
13
13
[ - Hs=mrnrnl}
! s=
l tH mrnrn11
=
s=mi. m. lll
1 -1
=
2cx, = {1,
M(2,2),
([ ]) = [ ! ; : ]
X3).
2
(b) L : P2 P2 defined by L(a+ bx+ cx )
2
2
2
a+ (b+ c )x , 13
{ + x , + x, 1 x+ x }
2
(c) D : P2 P2 defined by D(a+ bx+ cx )
2
b+
13
x, x }
by
c .
13={[ ][ ][ ]}
Homework Problems
Bl Determine the matrix of the linear mappingL with
respect to the basis 13 in the following cases. Deter
mine [L(1)]2l for the given [ 1)$,
2
111 + 3112,
{V1, 112} and L(111)
(a) In JR.. , 13
243
=
= Hl
=2
111 3v2,
{V1,112,v3}andL(111)
(b) InlR..3,13
LCv2) 3v1 + 4112 - 11 ,L(v ) =-vi + 2v2 +
3
3
=
6il3; [XJ.
244
Vector Spaces
Chapter 4
13= { [],[
]}
of JR2. In each
L
[L].:a.
L (l,2)= (1,-2) L (l,-2) (4, 8)
L (l,2)= (5,10) L (l,-2)= (-3,6)
L (l,2) (0,0) L (l,-2)= (1,2)
13 = {v1,v2,v }
:]
[
= ml [ ] l]}
[
L
is a linear
and
and
B3 Consider
the
basis
mine
L
[L (v1)].:a, [L (v2)].:a,
[L].:a.
[L (v3)].:a
determine
(b)
(c)
(d)
and
3)
transformation.
13
(as in Ex
L (-1,0,1)
(-2,0, 2).
L (2,1,0) = (3,3,0),
(1,4,4),
L (l,1,0)
L.
L (l,4,4).
and
13.
[136 n 13= {[!] ' [] }
[ 160l 13= {[] ' [ ]}
; s=
lO .
- -; - s=
-1 6 -6 .
basis
(c)
(d)
l= =! J mrnrnl}
[
!J mrni+m
13
x.
L
[L (x)].:a
given vector
for the
m
Wl L l:J [W
basisS=
inR3.
basis
(a)
basisS
and
(a)
with respect to the
inR3
JR3
.
JR3 is a linear
and
(b)
(c)
U = =n
[
XJ [il
=
s=
{[] [-;].[_:]}.
Section 4.6
(d)
li =: H {Ul r l lW
nl
=
[XJ. =
(a) L : IR.
3
IR. defined by L(x1,X ,x3) = (x1 +
2
{[i] 'm' lm
2
2
(a+b+c)+(a+2b)x+(a+c)x2,
13
= {1,x,x2}
(c) L: M(2,2)
245
Exercises
M(2,2) defined by
([ ]) = [ ; l
13= {[ ][ ][ ].
[ ]}
b
(d) D : P
2
b+2ax, 13
(e) T :
ID
r([ ]) [
13= { [ ][ ]}
a
2a
b .
Conceptual Problems
Dl Suppose that
13
13
to S and that Q is
(b) L: IR.2
2
b + 2cx,
x + x2}
v1 and v
1
in order for p- AP =
D
2
IR.? (Hint: Consider the equation
2
AP = PD, or A v1
[1 ] =
v = v1
2
v D.)
2
13
= {v1, ... ,v },
11
let W be a vector space with basis C , and let
L: V
(c) T :
{1,x}
IR.2
and C .
= {l,x,x2}, C
a1x2,
=
13
13 {[ ] [ ]},
IR.2. Let
{V1,v }
2
be a basis for IR.2 and let P denote the change of co
13
13
13=
and C .
13
C = { 1 + x2, 1 + x, -1 -
M(2,2) defined by
r([;])
[ ; ] { [ ] [ ]}
c= {[ ][ ][ ][ ]}
[; ]. 13
c= {[][]}
(d) L : P
:s=
b .
= {1 + x2, 1 + x, - 1 + x + x2},
246
Chapter 4
Vector Spaces
Definition
u2.
One-to-One
Lemma 1
EXAMPLE 1
--t
a transformation is one-to-one allows the definition of the inverse. The mapping inj :
JR3
--t
JR4 of Example 3.5.8 is a one-to-one mapping that is not invertible. The mapping
--t JR3 is not
P : JR4 --t JR3 of Example 3.5.8 is not one-to-one. For any n, proj11 : JR3
one-to-one, since many elements in the domain are mapped to the same vector in the
range.
EXAMPLE2
EXAMPLE3
Determine ifM : P2
--t
M(2,2) defined by La
( + bx+ cx2 ) =
2
[ ]
is one-to-one.
Solution: Let p x
( ) = l+x andq(x) = l+x+x .0bserve thatMp
( ) =
[ ]
= Mq
( ),
butp x
( ) * q(x), so Mis not one-to-one.
EXERCISE 1
Definition
L : U --t V is said to be onto if for every v E V there exists some u E U such that
Onto
EXAMPLE4
Invertible linear transformations of JR" are all onto mappings. The mapping P : JR4
--t
JR3 of Example 3.5.8 is onto, but the mapping inj : JR3 --t JR4 of Example 3.5.8 is not
onto.
EXAMPLES
2
Prove that L : JR. -t PI defined by L(yI, y2) =YI + (yI + y2)x is onto.
Solution: Let a+ bx be any polynomial in PI. We need to find a vector y =
24 7
[]
2
JR. ,
such that UJ) = a+ bx. For this to be true, we require that Yi+ (yI+ y2)x = L(yI, y2) =
a+ bx, so YI =a and b =YI + y2, which gives Y2 = b - Y1 = b - a. Therefore, we have
L(a, b - a) = a + bx, and so Lis onto.
EXAMPLE6
2
Determine if M : JR. -t JR.3 defined by M(xI, x2) = (xi, XI + Xz, x2) is onto.
Solution: If M is onto, then for every vector jl =
[;]
:]
[:]
11!.'
then we have
EXERCISE 2
Suppose that {ui, ... , uk} is a spanning set for U and Lis onto. Prove that a spanning
set for Vis {L(ui), ... , L( uk)}.
Theorem 2
I
The linear mapping L : U -t Vhas an inverse linear mapping L- : V -t U if and
only if Lis one-to-one and onto.
You are asked to prove Theorem 2 as problem D4.
Definition
Isomorphism
Isomorphic
If U and V are vector spaces over JR, and if L : U -t V is a linear, one-to-one, and
onto mapping, then Lis called an isomorphism (or a vector space isomorphism), and
U and Vare said to be isomorphic.
The word isomorphism comes from Greek words meaning "same form." The con
cept of an isomorphism is a very powe1ful and important one. It implies that the es
sential structure of the isomorphic vector spaces is the same, so that a vector space
statement that is true in one space is immediately true in any isomorphic space. Of
course, some vector spaces such as M(m, n) or P11 have some features that are not
purely vector space properties (such as matrix decomposition and polynomial factor
ization), and these particular features cannot automatically be transferred from these
spaces to spaces that are isomorphic as vector spaces.
248
Chapter 4
EXAMPLE7
Vector Spaces
3
Prove that P2 and JR are isomorphic by constructing an explicit isomorphism L.
Solution: We define L: P2
3
R by L(a, + aix + a2x')
[]
2
ao+a1x+a2x and q(x)
b0+b1x+b2x
2
2
L(t(ao + a1x + a1x ) + (bo + b1x + b1x ))
2
L(tao + bo + (ta1 + b1)x + (ta2 + b1)x )
tL(p) +Lq
( )
2
One-to-one: Let o0 + a1x + a2x
Hence, a0
a1
[:;]
[]
R3, we have La
( , + a1x + a2x')
{u1,
Theorem 3
. ,
2
L(a, + a1x + a2x )
[]
EXERCISE 3
[::J.
V is an isomorphism and
Suppose that 1I.J and V are finite-dimensional vector spaces over R Then 1I.J and V
are isomorphic if and only if they are of the same dimension.
You are asked to prove Theorem 3 as Problem DS.
EXAMPLE 8
249
If we know that two vector spaces over JR have the same dimension, then
Theorem 3 says that they are isomorphic. However, even if we already know that
two vector spaces are isomorphic, we may need to construct an explicit isomorphism
between the two vector spaces. The following theorem shows that if we have two
isomorphic vector spaces 1!J and V, then we only have to check if a linear mapping
L : 1!J
---t
spaces.
Theorem 4
If1!J and V are n-dimensional vector spaces over JR, then a linear mapping L : 1!J
---t
PROBLEMS 4.7
Practice Problems
Al For each of the following pairs of vector spaces,
1!J =
I
{[ ]
1
E JR
of 2 x 2 diagonal
matrices
Homework Problems
Bl For each of the following pairs of vector spaces,
define an explicit isomorphism to establish that the
{[] []}
I
{[ :]I
a1,a2,a3
E JR
of 2 x 2
upper-triangular matrices
Conceptual Problems
Dl Prove Lemma 1. (Hint: Suppose that L is one-to
one. What is the unique u E 1!J such that L(u) = O?
Conversely, suppose that Null (L) = {O}. If L(u1) =
L(u2), then what is L(u1 - u2)?)
D4 Prove Theorem 2.
DS Prove Theorem 3. To prove "isomorphic ::::} same
dimension," use Exercise 3. To prove "same dimen
sion ::::} isomorphic," take a basis {u1, ..., u11} for
1!J, and a basis {v1 ,
, v11} for V. Define an isomor
phism by taking L(u;) = v; for 1 i n, requiring
.
D6 Prove Theorem 4.
D7 Prove that any plane through the origin in JR3 is iso
2
morphic to JR .
250
Chapter 4
Vector Spaces
JR is isomorphic to JR3.
x ]Rm is isomorphic to JRn+m.
x
1l.J
CHAPTER REVIEW
Suggestions for Student Review
Remember that if you have understood the ideas of
Chapter 4, you should be able to give answers to these
questions without looking them up. Try hard to answer
them from your own understanding.
1 State the essential properties of a vector space over
[ !]
? (Section 4.4)
(c) Take the vector you found in (b) and carry out
[-!]
with re
Chapter Review
251
Chapter Quiz
El Determine whether the following sets are vector
spaces; explain briefly.
(a) The set of 4 x 3 matrices such that the sum of
the entries in the first row is zero (a11 + a12 +
a1 = 0) under standard addition and scalar
3
multiplication of matrices.
(b) The set of polynomials p(x) such that p(l) = 0
and p(2) = 0 under standard addition and scalar
multiplication of polynomials.
(c) The set of 2 x 2 matrices such that all entries
are integers under standard addition and scalar
multiplication of matrices.
[::J
such that x1 + x2 +
(a){[ ][ -nl
{[ ] ' [ -] ' [
{[_ !][ ][
(b)
(c)
].[; -][ ]}
] ' [; ]}
]}
-
1
0
1 ,
1
3
3
1
3
1 . Pick a
1 and V4
v2 = O, v
3
-2
0
1
2
0
subset of the given vectors that forms a basis 13
for. Determine the dimension of.
0
2
1 with
(b) Determine the coordinates of x
-3
-5
respect to 13.
,
tr{i - ]
ma
and let=
252
Chapter 4
Vector Spaces
Further Problems
These problems are intended to be ch allenging. They
sider
S.
A.
A,
A,
For example, if
[- -!].
3
square with
wt(A)
3.
(e)
(f)
A is a magic
f
and
1.
[ ;]1
2
with
3 x 3 magic
M(3, 3) consisting of magic
squares is denoted MS 3.
(a) Show that MS 3 is a subspace of M(3, 3).
(b) Observe that weight detenrunes a map wt
MS 3 lit Show that wt is linear.
(c) Compute the nullspace of wt. Suppose that
p, q E lit
Show that 15
{:l., K1, K2} is a basis for MS 3.
As an example, find all 3 x 3 magic squares of
weight
a
d '
g
for some
fine
S+T
wt(!...)
k
form a
s in the
A)
, v
a11K1 - a12K2.)
square with
S. W hat can
(d) Let!...
- 1 1 1l
[ 1
K2
a, b, c, . .
1 l
j
m
h
k
n
{ p s + q t I p, q E JR, s E S, t
T}
=
V and
{[ ]}
{ []}
to be Span
{[]}
Chapter Review
253
n-k.
dim( + T) + dim( n T)
MyMathlab
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 5
Determinants
CHAPTER OUTLINE
5.1 Determinants in Terms of Cofactors
5.2 Elementary Row Operations and the Determinant
5.3 Matrix Inverse by Cofactors and Cramer's Rule
5.4 Area, Volume, and the Determinant
For each square matrix A. we define a number called the determinant of A. Originally,
the determinant was used to "determine" whether a system of n linear equations in
n variables was consistent. Now the determinant also provides a second method for
finding the inverse of a matrix. It also plays an important role in the discussion of
volume. Finally, it is an important tool for finding eigenvalues in Chapter
6.
b1
b2
X1
X2
Definition
Determinant of
a
2 Matrix
detA
det
a11
a21
a11
a12
a21
a22
is defined by
256
Chapter 5
EXAMPLE 1
Determinants
[ !l [ -n [ n
Solution: We have
and
[ !]
[2 7]
[ ]
det
det
det
-5
1(4) - 3(2)
2(-5)
2(4) - 2(4)
7(8)
-2
10 - 56
-66
One risk with this notation is that one may fail to distinguish between a matrix and the
determinant of the matrix. This is a rather gross error.
EXERCISE 1
1; I
The 3
(b)
I _;1
(c)
I I
3 Case
Let A be a 3 x 3 matrix. We can show through elimination (with some effort) that the
system is consistent with a unique solution if and only if
Observe that the determfoant being multiplied by a11 is the determinant of the 2 x 2
matrix formed by removing the first row and first column of A. Similarly, a12 is being
multiplied by (-!) times the determinant of the matrix formed by removing the first
257
a13
matrix formed by removing the first row and third column of A. Hence, we make the
following definitions.
Definition
Cofactors of a
deleting the i-th row and }-th column. Define the cofactor of a 3 x 3 matrix
3 Matrix
CiJ
Definition
aiJ
to be
Determinant of a
3
3 Matrix
Remarks
1. This definition of the determinant of a 3 x 3 matrix is called the expansion of
the determinant along the first row. As we shall see below, a determinant can
be expanded along any row or column.
2. The signs attached to cofactors can cause trouble if you are not careful. One
helpful way to remember which sign to attach to which cofactor is to take a
blank matrix and put a+in the top-left corner and then alternate - and+both
across and down:
[ : i.
+
EXAMPLE2
Let A
[
1
CiJ
Calculate the cofactors of the first row of A and use them to find
the determinant of A.
Solution: By definition, the cofactor C11 is ( -1) 1+1 times the determinant of the
matrix obtained from A by deleting the first row and first column. Thus,
C11
(-1)1+ det
[ ]
3
3(6)- 5(0)
18
258
Chapter 5
Determinants
EXAMPLE2
The cofactor C12 is (-1) 1+z times the determinant of the matrix obtained from
(continued)
2
C12 =(-1)1+ det
[i ]
by
= -[2(6) - 5(1)] = -7
1 3
C13 =(-1) + det
[i ]
Hence,
EXERCISE2
Let
A = 0
3
-2 . Calculate the cofactors of the first row of A and use them to
-3
2
I
0
A.
EXAMPLE3
Calculate det
[- 1.
5 0 -1
Solution: By definition, we have
det -2
2
2
0
-1
= a11C11
1 2
=1(-1)1+
0
1
2 2
+2(-1)1+ -1
5
1
3 -2
+3(-1)1+
-1
5
2
0
Definition
Cofactors of an
11 x n
Matrix
Let
from
aij is defined to be
Section 5.1
Definition
259
Determinant of a
n x n
JVlatrix
Remark
This is a recursive definition. The result for the n x n case is defined in terms of the
(n
-1) -1)
x (n
EXAMPLE4
2 2
x
-2) -2)
x (n
* **
0.
25 63 07
3 20 43 a11 C11+a12C12+a13C13+a14C14
167
1 7
0(*)+2(-1)1+2 -2-5 20 43 + 3(-1)1+3 -2-5 53 43 +O(**)
(1(-1)1+1 I I+6(-1)1+2 I= I+7(-1)1+3 I= I)
+3(1(-1)1+11; 1+5(-1)1+21= 1+7(-1)1+31= ;1)
-2((0 -8) -6(-6+20)+7(-4 -0))
+3((9 -4) -5(-6+20)+7(-2+15))
-2(-8 -84 -28)+ 3(5 -70+91)
-2(-120)+ 3(26) 318
4 4
and
multiplied by
det
1-20
-5
det
-2
calculation, and things will get worse for larger matrices. In applications it is not un
common to have a matrix with thousands (or even millions) of columns, so it is very
important to have results that simplify the evaluation of the determinant. We look at a
few useful theorems here and some very helpful theorems in the next section.
Theorem 1
260
Chapter S
Determinants
We omit a proof here since there is no conceptually helpful proof, and it would be
a bit grim to verify the result in the general case.
Theorem
column along which we are going to expand. If one row or column has many zeros, it
is sensible to expand along it since we shall then not have to evaluate the cofactors of
the zero entries. This was demonstrated in Example 4, where we had to compute only
two cofactors.
EXAMPLES
2
3
2 -1
0
6
1
'B=
Calculate the detemtinant of A =
4 1
-1 5
3 -1
2
1
-1
4
0 2
2 2
C=
0 0 -1
3
0 0
0
4
0
0
2
-1
0
1
, and
a13C13
a23C23 + a33C33
= (-1)(-1)1+
1 I+
+
+
1 ; I) +
a21C21 + a22C22
=0
=o
+
+
a23C23
a24C24
3 0 -1
2 2
( 6)(-1) + 4 2 1
3 0
1
6
2 2
2(-1) +
= 12(3 - (-3)) = 72
We expanded the 3
=0
+
(
4 2 1
0 + 4( -1 ) 4+4 0 2 2
0 0 -1
3 3
= 4 (-1)(-1) +
I I )
we continuously
Section 5.1
EXERCISE 3
Calculate the determinant of A
-1
-1
-2
-4
261
by
Exercise 3 demonstrates the usefulness of expanding along the row or column with
the most zeros. Of course, if one row or column contains only zeros, then this is even
easier.
Theorem 2
n x n
0.
Proof: If the i-th row of A contains only zeros, then expanding the determinant along
the i-th row of A gives
detA
ailCil
ai2C;2
a;nCin
As we saw with the matrix C in Example 5, another useful special case is when
the matrix is upper or lower triangular.
Theorem 3
If A is an
n x n
Theorem 4
If A is an
n x n
detA7.
With the tools we have so far, evaluation of determinants is still a very tedious
business. Properties of the determinant with respect to elementary row operations make
the evaluation much easier. These properties are discussed in the next section.
262
Chapter 5
Determinants
PROBLEMS 5.1
Practice Problems
Al
.rr
(c)
:Ft:
following dete
-3
I I
1
(e) 0
0
(d)
-4
0
0
2
3
(f)
-4
5
2
1
2
-4
5
-1
2
0
0
0
0
0
-7
2
-5
0
-7 8
3 0
0 0
0 0
[ i
i
[-
-
-4
(b)
(c)
-1
2
0
-4
3
0
2
2
2
0
-5
0
(d)
2
-1
4
4
-3
2
-2
-1
-2
1
(d)
()
H ol
[- ]
0
0
(b )
-6
-4
-4
-3
(e)
0
0
3
(f)
0
0
0
0
0
0
0
3
-1
-1
-2
6
5
-5
6
-1
-3
-3
1
0
0
0
1
-6
-3
6
2
1
-5
-6
-5
2
1
-1
3
-3
0
0
7
3
0
8
[ ; ]
(a ) A =
1
0
0
1
1
-2
(b) A =
3
5
6
1
-2
-3
(a)
[ -i]
1
(c)
2
0
0
5
-2
) E 2 =
(c) 3 =
[ !]
[ ! ]
[- i
0
263
Homework Problems
Bl
:r:ir
1
1
(c)
1
-1
1
following dete
1
0
2
0
3
(e )
-1
0
0
";:;jf -:i
3
0
(d)
3
-5
4
0
(f)
0
0
2
3
0
0
(c)
2
0
0
-3
1
9
-1
2
9
2
0
[ -1
1
0
1
-1
-4
-2
(d)
2
-3
3
1
2
0
1
3
4
-1
(e)
-2
3
2
0
-1
0
1
3
2
5
[ 1
H - !J
(b)
-1
2
0
2
0
5
0
-6
-1
-3
( )
(b)
H j ll
[ - j]
-4
2
-4
-3
3
3
2
3
3
-1
3
0
4
1
4
0
0
-5
2
1
0
0
0
0
0
0
-1
choice.
-2
0
-4
0
1
0
5
-2
0
0
2
0
4
-2
0
2
0
-4
0
1
-2
0
(c)
4
0
(d )
(f)
-3
0
(a )A=
[ - -1
-1
3
1
(b)A
= -5
1
6
6
-3
6
3
0
4
2
5
0
0
(b) E
2=
(c)
3=
[ ]
[ ! ]
[ ! ]
_1
264
Chapter 5
Determinants
Computer Problems
0.5
0.5
0.5
0.5
0.5
0.5
-0.5
-0.5
0.5
0.5
0.5
-0.5
-0.5
-0.5
-0.5
0.5
following matrices.
(a)
(b)
l.09
2.13
1.72
4.83
-3.25
2.15
2.95
1.57
-0.89
(c)
1.23
-2.09
2.35
0.17
4.19
3.89
-1.28
22.1
0.78
1.58
2.15
-3.55
4.15
-2.59
1.01
0.00
Conceptual Problems
Dl Prove Theorem 3.
[: 1
b1
b2
EXAMPLE 1
Let A
a11
a21
a12
a22
a13
a23 and let B be the matrix obtained from A by multiplying the
a31
a32
a33
Solution: We have B
a11
a21
ra31
rdetA.
a12
a13
a22
ra32
ra33
along its third row. The cofactors for this row are
EXAMPLE 1
265
Observe that these are also the cofactors for the third row of A. Hence,
(continued)
Theorem 1
Let A be an n x n matrix and let B be the matrix obtained from A by multiplying the
i-th row of A by the real number r. Then, det B
r det A.
that the cofactors of the elements in this row are exactly the cofactors of the i-th row of
A since all the other rows of B are identical to the corresponding rows in A. Therefore,
Remark
It is important to be careful when using this theorem as it is not uncommon to incor
rectly use the reciprocal (1 / r) of the factor. One way to counter this error is to think
of "factoring out" the value of r from a row of the matrix. Keep this in mind when
reading the following example.
EXAMPLE2
By Theorem 1,
1
5
2
and
1
5 1
-1
-2
2
EXERCISE 1
4
15
0
[ :i
-2
det
3
10
-1
(-2)det
[i
4
3
0
1
2
-:1
r3 det A.
266
Chapter 5
EXAMPLE3
Determinants
The following calculations illustrate that swapping rows causes a change of sign of the
determinant. We have
[ ]=cb-da=-(ad-bc)=-det[ ]
=[aa2111 a12a22 a13a23]
a31 a32 a33
<let
Let A
and let Bbe the matrix obtained from A by swapping the first
and third rows of A. We expand the determinant of B along the second row (the row
that has not been swapped):
= [:: : :1
a11 a12 a13
=a2i(-1)2+1 la12a32 a33a131 + a22(-1)2+2 la11a31
=-a21 (-1: ::I)+ a22 (-1::
:I+ a22 I::
a12 a13 ]
a22a32 a23a33
<letB
Theorem 2
<let
Suppose that A is an
nxn
-det A .
an
nxn
(n - 1) x (n
n=
2 i n Example 3.
matrix obtained from A by swapping two rows. If the i-th row of A was not
x -1)
(n
matrices. We can
obtain these matrices from the cofactors of the i-th row of A by swapping the same two
rows. Hence, by the inductive hypothesis, the cofactors
c71
=-CiJ.
Hence,
c;1
of B and
CiJ
of A satisfy
detB
<let A
Section 5.2
Corollary 3
267
Proof: Let B be the matrix obtained from A by interchanging the two equal rows.
Obviously B =A, so detB = detA. But, by Theorem 2, detB= -detA, so detA =
-detA. This implies that detA = 0.
Finally, we show that the third type of elementary row operation is particularly
useful as it does not change the determinant.
EXAMPLE4
For any r
det
[c:
ra
Let A =
lit we have
: rb] =a(d+rb)-b(c+ra)
[:: : :i
a31
a32
= ad+arb-bc-arb =ad-be=det
[: ]
a33
row 2 to row 3. Expanding the determinant of Balong the first row and using the result
above gives
detB = det
[ ::
a31 + ra21
1 1
=a!l(-1) +
a32 + ra22
a12
a32 + ra22
a33 + ra23
a23
a33 + ra23
I
:1- 1:: :1
- 1+3
+ a13( 1)
a11
a31 + ra21
a12
Theorem 4
ai2
a13
a12
a32
a23
:
l
1 +2
+ an(-l)
I
I::
a12
a32 + ra22
+ a 13
a11
a31 + ra21
a33
268
Chapter 5
Determinants
the i-th row to the k-th row. Hence, by the inductive hypothesis, the cofactors
and
aeiCe1
ae11Ce11 = ae1Ce1 +
aenCe11
c;1 of B
= detA
EXAMPLES
LetA=
-3
-3
.Find detA.
2 11
6
Solution: By Theorem 4, performing the row operations
R2 - R1
and
R4 - R1 does not
detA=
-4
To get the matrix into upper-triangular form, we now swap row 2 and row 3. By
Theorem 2, this gives
3
detA= (-1)
By Theorem 1, factoring out the common factor of ( -4) from the third row gives
detA= ( -1 )( -4)
5
1
R4 - R2 to get
detA= 4
= 4 (1)(3)(1)(6)= 72
EXERCISE 2
LetA
2
-6
=
4
-6
1
6
-2
-2
3
269
6
5
. Find detA.
-1
5
-2
EXAMPLE6
Find the determinant of A
5
8
5
1
1
0
6
7
6
4
7
9
10 .
-2
Solution: By Theorem 4,
1
0
0
0
5
3
0
6
1
0
4
7
2
3
-2
3
1
1
(-1) + 0
1
0
4
2
3 +O
-2
detA
detA
detA
(1)(3)(-1)2+3
EXERCISE 3
-6
Find the determinant of A
3
-6
-3
-2
2
4
2
[i ![
4
-4
0
-3
(-3)(12 - 1)
-5
3
o
-4
-33
270
Chapter 5
Determinants
Theorem 5
(3) A is invertible
Proof:
In Theorem 3.5.4, we proved that (2) if and only if (3), so it is only necessary
Remark
Theorem 5 shows that detA :f:. 0 is equivalent to all of the statements in Theorem 3.5.4
and Theorem 3.5.6.
We shall see how to use the determinant in calculating the inverse in the next
section. It is worth noting that Theorem 5 implies that "almost all" square matrices are
invertible; a square matrix fails to be invertible only if it satisfies the special condition
detA
0.
Determinant of a Product
Often it is necessary to calculate the determinant of the product of two square matrices
A and B. When you remember that each entry of AB is the dot product of a row from
A and a column from B, and that the rule for calculating determinants is quite compli
cated, you might expect a very complicated rule. The following theorem should be a
welcome surprise. But first we prove a useful lemma.
Lemma6
(det E)(det C)
Proof: Observe that if Eis an elementary matrix, then since Eis obtained by perform
ing a single row operation on the identity matrix, we get by Theorem 1, Theorem 2,
and Theorem 4 that det E
a, where a is
1, -1, or
r,
row operation used. Moreover, since EC is the matrix obtained by performing that row
operation on C, we get
det(EC)
a det C
det Edet C
Section 5.2
Theorem 7
271
Theorem 3.5.4. If B is invertible, then y = Bx has a solution for every y E JRn, and
hence there are infinitely many x such that ABx =
On the other hand, if detA * 0, then by Theorem 3.5.4 the RREF of A is/. Thus,
EXAMPLE?
0
-1
2
Solution: By Theorem 4 we get
and B =
detA = -10
-1
detB
-1
15
-10
5
2
-2
1 =
-15
28 = -105
7
detAB = det
Ek such that
[=9 - i
12
20
(-5)
= (-5) 1
-4 = (-5)(-1)2+1
56
= (5)(315)= 1575
4
12
20
56
(detA)(detB)
272
Determinants
Chapter 5
PROBLEMS 5.2
Practice Problems
Al Use row operations and triangular form to compute
-2
-6
-6
-4
-1
-3
-4
-2
-1
-3
[ i
[ 1
trix is invertible.
(a) A=
-1
(b) A=
-1
(a)A=
-1
-2
(b)A=
-2
-4
-1
16
10
-5
27
-1
-2
(a)
-2
(c) A=
A4 Verify Theorem 7 if
(a)A=
(c)
-2
-1
(b)
(e) A=
[! : =il
invertible.
(c) A=
(d)A=
(d)
2
1
(b) A
[ -n
[- l
H ]
-
B=
B=
AS Suppose that A is an
matrix is invertible.
n x n
matrix.
Homework Problems
the determinants of each of the following matri
[ ]
(a) A=
[ -l 1]
de;A.
(b) A=
[ ; - -1
[; - 1
(c) A=
(d) A=
-1
-4
-1
invertible.
(a) A=
-9
7
-2
-3
-3
-4
(b) A=
1
1
2
2
1
8
(c) A=
-1
[! ]
[ i l
O
-1
-3
-2
1
-
(b)
-4
4
1
27
16
2
p
(d) A=
5
2
-1
-3
(c)
(a)
10
-2
-1
-2
-7
-6
(e) A=
(f) A=
-4
-3
(d)
4
4
273
3
5
-2
-1
-1
-2
Conceptual Problems
Dl A square matrix A is skew-symmetric if AT = -A.
If A is an n x n skew-symmetric matrix, with n odd,
prove that detA= 0.
n x n matrices. If
D3 Suppose that A is a 3
D4 Let A be an
[
l
[: {l [ : a
= det
a+p
b+q
+ det
c1+kr
274
Chapter 5
Determinants
a +p
b+q
c+r
e +y
f +z
a31
vl [
p+q
u+
det b+c
q +r
v +w
c+a
r+p
w+u
a+b
[:
: :
a32
(b) Let A be an n
D6 Prove that
[::
r det
a31
a32
a33
ra33
i [:: : :i .
=
2 det b
related?)
(c) Let A be an n
1 +a
2
(1 +a)
det B.
Theorem l
Proof: Let B be the matrix obtained from A by replacing (not swapping) the k-th row
of A by the i-th row of A. Then the i-th row of B is identical to the k-th row of B, hence
det B
0 by Corollary 5.2.3. Since the cofactors c;j of B are equal to the cofactors
CkJ of A, and the coefficients bkJ of the k-th row of B are equal to the coefficients aiJ
as required.
Definition
Cofactor Matrix
Let A be an n
(cof A)iJ
CiJ
EXAMPLE 1
Let A=
[02 4-2 -1 1
1 . Determine cof A.
5
6
Solution: The nine cofactors of A are
=Cl)1- I 17
C21 14_2 -11= -18
-11=7
=
ell
= (-1)
275
12 I I=
C22= 12 -11=
C32=(-1)012 -11 =-2
c
= c-1)
(1)
(-1)
16
=6
Hence,
= [-1
EXERCISE 1
Observe that the cofactors of the i-th row of A form the i-th row of cof A, so they
form the i-th column of (cof Al. Thus, the dot product of the i-th row of A and the
i-th column of (cof Al equals the determinant of A. Moreover, by the False Expansion
Theorem, the dot product of the i-th row of A and the }-th column of cof A equals
if i * j. Hence, if af represents the i-th row of A and c1 represents the }-th column of
(cof A)\ then
A(cof Al =
I1
[. .
/
[c1
c,!]
-+T
a
i11 en
Ct
a11
C1
detA
=0
0
0
=
0
0
an. en
detA
(detA)/
detA
A-1
=( )
de A
ccofA)7
276
Chapter 5
Determinants
If detA
method).
EXAMPLE2
Find the inverse of the matrix A
Solution:
-1
-2
detA
-1
-2
2
=
-1
-14
2(24 + 14)
76
1
=
[
76
--
detA
(cof A)
1
J_
-18
28
EXERCISE 2
Use the cofacto< method to find the inverse of A
]6
[-i n
Cramer's Rule
Consider the system of
linear equations in
variables, Ax
X1
X;
Xn
detA
1
7
-- (cof A) b
detA
C11
C 21
C111
b1
C1;
C2;
c,li
b;
C111
C211
Cnn
bn
b.
If detA * 0 so that
277
This is
de;A
1
-- (b1C1 +b2C2 +.. +bn C)
Ill
I
detA
But the i-th row of (cof A f is the i-th column of the original cofactor matrix cof A. So
Now, let N; be the matrix obtained from A by replacing the i-th column of A by
b.
Then the cofactors of the i-th column of N; will equal the cofactors of the i-th column
of A, and hence we get
det N;
b1Ci;+b2C2;+
+b11C11;
b is
det N;
-- detA
x 1
This is called Cramer's Rule (or Method). We now demonstrate Cramer's Rule with a
couple of examples.
EXAMPLE3
Xt +Xz - X3
b1
b2
X1+X2+ 2X3
b3
[ ]
-
detA
-1
5
so
-1
7
3
( 1 ) (2 )(3)
Hence,
Xt
det N1
=
x2
x3
detA
det N2
detA
det N3
detA
b2
b3
b1
2
6 l
2
6 l
!
=
b1
b2
b3
1
-1
5
2
-1
5
2
b1
b2
b3
b1
!
=
=-
b2+Sb1
6
b3 +2b1
1
1
- 7
6
3
1
9
3
b1
b2+ Sbi
b3 + 2b1
1
bi
b2 - 2b1
b3 -bi
-1
0
-1
b1+3bz-7b3
6
0
=
-2b1 + 2b3
6
Chapter 5
278
EXAMPLE4
Determinants
2
3
2
-xi
5
3
5
1
-x2
3
--X1 + -X2
. .s A
.
matnx
S oI ution: The coe ffi c1ent
1
.
detA
Hence,
Xi
x
2
r-22/315
1
5
1
2
- - - -
-2 -1
3
3
3/5
11 3
3 2
5 5
1
1
1
1
, so
-4
225
165
-225 . -11
1
3/5
1/5
1/2
-1/3
4
30
8
-4/225
1
-225
. 93 93
-2/3 1/5
-4/225 2/5 1/2
4
4
4
To solve a system of
equations in
n +
n x n.
Thus, solving a system by using Cramer's Rule requires far more calculation than
elimination, so Cramer's Rule is not considered a computationally useful solution
method. However, like the cofactor method above, Cramer's Rule is sometimes used
to write a formula for the solution of a problem.
PROBLEMS 5.3
Practice Problems
Al Determine the inverse of each of the following ma
trices by using the cofactor method. Verify your an
swer by using multiplication.
(a)
(b)
(c)
[! l]
[; =1
U - Il
(d)
[
-2
A2 Let A
2 -1
[-; - ]
-
A(cof Af
and
determine
detA
(c)
279
7x1+
x2 - x
4 3 =3
-x
6 1 - 4x2+
x3 =0
3x1 +x
5 2=7
4x1 -x2 - x
2 3 =6
(b) x
3 1+x
3 2=2
(d)
x
2 1-x
3 2=5
x
2 1+3x2 - 5x3 =2
3x1 -x2 +2x3 =1
5x1+4x2 - x
6 3 =3
Homework Problems
[ 1.
(c)
(d)
[- ]
[- ]
-2
3
(e)
-1
-7
-2
-3
0
2
(a)
(b)
-2
-3
detA
2
x1 - x
7 2=3
3x1+5x2=-2
X1 +x
3 2=-3
-4
x
5 1+x
4 2=-17
(c) x1 - 5x2 - x
2 3 =-2
2
x1
+x
3 3 =3
4X1 +
X2
-4
(f)
4
(a) Determine the cofactor matrix cof A.
(b) Calculate A(cof A)7 and determine
and A-1
-2
-2
H -1
1
[ -]
0
B2 Let A=
(d)
2
-3
X1
-X3 =1
+2X3 =-2
3x1 -x2+x
3 3 =5
=0
Conceptual Problems
Dl Suppose that A = i11
n x n
an]
is an invertible
matrix.
D2 Let A =
-1
-1
2
0
0
3
1
1
to calculate (A- )23 and (A- )42. (If you calcu
0
280
Chapter 5
Determinants
ma
trix is invertible. In particular, we have only been interested in whether the determinant
of a matrix is zero or non-zero. We will now see that we are sometimes interested in the
specific value of the determinant, as it can have an important geometric interpretation.
[]
and
[ l
Ut UJ + V) X1
0
Figure 5.4.1
We will calculate the area of this parallelogram by calculating the area of the
rectangle with sides of length u1 + v1 and u2 + v2 and subtracting the area inside the
rectangle and outside the parallelogram, as indicated in Figure 5.4.2.
This gives
Area(it, v)
V1V2 - U2V1 -
u1u2 -
V1V2 - U2V1 -
U1U2
[ ]
[it v].
Remark
At this time, you might be tempted to say that the area of the parallelogram induced
by any two vectors it and
v as a right-handed system.
281
Area(a, V)
AS
A6
u2
0
Figure 5.4.2
EXERCISE 1
Show that 1f a
[UU1]2
and v
[VVJ2]
a and v.
I [UUj2 VV1JI2
We have now shown that the area of the parallelogram induced by it and v is
Area(it, v)
EXAMPLE 1
l [ :]I
det
Draw the parallelogram induced by the following vectors and determine its area.
(a) it=
(b) it
[l [-]
[ [
l -n
v
"
Area (it, v)
l [- ;JI
det
2(2) - 3(-2)
10
282
Chapter 5
EXAMPLE 1
Determinants
(continued)
l [ JI
Area(it,v) = det
= 1(-2)-1(1)= -3
Now suppose that the 2 x 2 matrix A is the standard matrix of a linear transforma
2
IR. . Then the images of i1 and v under L are L(il) =Ail and L(v) =Av.
tion L: IR.2
l [
]I = ldet (A [a v])I
l ( [
(5.1)
In words: the absolute value of the determinant of the standard matrix A of a linear
transformation is the factor by which area is changed under the linear transformation
L. The result is illustrated in Figure 5.4.3.
Figure 5.4.3
EXAMPLE2
Let A =
i1 =
[ ]
[ ]
and v =
in two ways.
[-]
EXAMPLE2
283
(continued)
= l [ ;11 = - =
det
41
18
Area(L(i1), L(v))
EXERCISE2
Let A
= [ ]
=I I
= [ [ ;J[ [ [- JI= =
det
2(2)
det
JR.2 JR.2
in the
x1
e1 ez
and
direction
under S
and compute the area determined by the image vectors in two ways. Illustrate with a
sketch.
JR.3,
lw(i1xv)I
[::l = [:l
V
and W
Jw (it x
= [::I
v)I
then
= lw1(U2V3-U3V2)-w2(U1V3-U3V1)+w3(U1V2-U2V1)I = lU3UUzJ
det
Volume(i1,v,w)= det i1
EXAMPLE3
Ut A
= [;
-n = l-ll m
U
and W
= nl
wJI
Calculate the volume of the
parallelepiped induced by i1, v, and wand the volume of the parallelepiped induced by
284
Chapter 5
Determinants
EXAMPLE 3
(continued)
H : -]
Volume(ii, V, W) = det
= I
- 71 = 7
l [
Av
Aw
H :J
JI
= det
Moreover, det A =
= I
3851
385
-55, so
Volume (Ail.Av,Aw) = I det AIVolume (il,v,w)
2 case.
In general, if v1, ..., vn are n vectors in 11, then we say that they induce an
n-dimensional parallelotope (the n-dimensional version of a parallelogram or
parallelepiped). Then-volume of the parallelotope is
n, then
PROBLEMS 5.4
Practice Problems
Al (a) Calculate the area of the parallelogram induced
by il=
[; ]
and v=
[]
in 2.
[ n
-
A2 Let A =
tion R
[ ]
the image of il =
[]
and v =
[-]
under R and
lH [=n
jl =
and w =
[! - ].
0
2 s
0
2
and V4
v3 =
3
0
[H
[ - -H
HJ m
.w=
.andA=
, i12 =
0
2
5
2
5
standard matrixA=
0
0
A6 Let {i11, .
..
3
4
0
0
1
3
7
0
1
0
.
3
V=
285
1
3
Homework Problems
Bl (a) Calculate the area of the parallelogram induced
by il =
[]
and v =
[ ]
in JR.2 .
A=
[ n
B2 LetA =
H
[ l
: JR.2 JR.2
v =
Hl nl
w
[-H [ H
V=
[H
[ - H
and W=
andA=
dured by U=
[_il
u -r H
V3 =
2
0
, and V 4 =
.
3
5
0
7
2
2
standard matrixA=
-1
0
3 1
3
3 7
2
1
0
o
1
286
Determinants
Chapter 5
Conceptual Problems
Dl Let {ii\, ... , V,1} be vectors in lR.11 Prove that the
D2 Suppose that L, M
JR3
---7
L is I det BAI.
CHAPTER REVIEW
Suggestions for Student Review
Try to answer all of these questions before checking an
Chapter Quiz
El By cofactor expansion along some column, evaluate
-2
4 0 0
det
1
-3
-2
-1
E3
!]
-4
is invertible.
7.
12
0
Evaluate det 0
1 .
0
5
[ ]
Chapter Review
(b) If A =
2x1 + 3x + x3 = 1
2
X1 + Xz - X3
-1
+ 2x3 =
-4
-2x1
[ _;],
287
[ J [-H
'
and W=
[!]1
Further Problems
These exercises are intended to be challenging. They
a2
b2
3
a
3
b
1
1
c2
c3
d2
I!
det
. By
c= b cosA + a cosB
b2 + c2 - a2
cosA= ---2 bc
[ ] .
1
c2
MyMathlab
[fy]
0 3
2,
0 3
03
,2
B
I o.2 j
= detA detB.
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 6
Eigenvectors and
Diagonalization
CHAPTER OUTLINE
6.1
6.2
6.3
6.4
Definition
Eigenvector
that L(V)
Eigenvalue
-7
A non-zero vector v E
JR11 such
Remarks
1. The pairing of eigenvalues and eigenvectors is not one-to-one. In particular, we
will see that each eigenvector of L will have a distinct eigenvalue, while each
involving eigenvectors make sense only for non-zero vectors. In particular, we will
see that we often want to look for a basis of JR11 that contains eigenvectors of a linear
transformation.
290
Chapter 6
EXAMPLE 1
[ =i]
Let A =
and let L
JR2
L(x) = Ax. Determine which of the following vectors are eigenvectors of Land give
the corresponding eigenvalues:
Solution:
So,
[ ]
To test whether
[ ]
rn [=;J. rn [!]
and
[=]
= (-2)
[ ]
L(-2,-2) = (-2)L(l,
[=]
[ i]
[ ] 2[=]
=
[]
[17 -15J r J r J r J
l
20
-18
-28
-34 f. ;i 3
is not an eigenvector of L.
L(3 4)=
.
20
EXAMPLE2
= (-2)(2)
L(l ,3) =
[!]
1)
So,
1 ):
Since
So,
-18
-12
[3]
= 3
- 4
vector of proji1 with corresponding eigenvalue 0. Observe that this means there
is a whole plane of eigenvectors corresponding to the eigenvalue 0 as the set of
vectors orthogonal to it is a plane in JR3. For an arbitrary vector a
not
JR3, proj,1(a)
Section 6.1
EXAMPLE2
291
(continued)
eigenvalue 0. Since
eigenvalue 1.
EXAMPLE3
1r.
ln)
11'.2
---7
.
cose
.
lnl
.
11'.2 wit h matnx
- sine
cose
sme
.
, where e is not an
/lv for
eigenvalues or real eigenvectors. In Chapter 9 we will see that it does have complex
eigenvalues and complex eigenvectors.
EXERCISE 1
Let Re : IR3
cose
---7
- sine
cose
Derermine
the matrix A.
Definition
Suppose that A is an n x n matrix. A non-zero vector v E IRn such that Av= /lv is called
Eigenvector
Eigenvalue
an eigenpair.
In Example 1, we saw that
/l is called
[ ] [!]
and
an eigenvalue of
A.
[ = ]
Av= AV.
Av-M=o
(A
- ,i)v
0,
a matrix and /l is a number, so their difference is not defined. To get around this, we
292
Chapter 6
use the fact that v= IV, where I is the appropriately sized identity matrix. Then the
eigenvector condition can be rewritten
(A - ,U)v=
The eigenvector vis thus any non-trivial solution (since it cannot be the zero vector)
of the homogeneous system of linear equations with coefficient matrix (A -tl/). By
the Invertible Matrix Theorem, we know that a homogeneous system of
n equations in
Hence, for tl to be an eigenvalue, we must have det(A-tl/) = 0. This is the key result in
the procedure for finding the eigenvalues and eigenvectors, so it is worth summarizing
as a theorem.
Theorem 1
Suppose that A is an
n x n matrix.
(A -,l/)v=
are eigenvectors of A that correspond totl.
Definition
Let tl be an eigenvalue of an
Eigenspace
n x n matrix
Remark
From our work preceding the theorem, we see that the eigenspace of any eigenvalue tl
must contain at least one non-zero vector. Hence, the dimension of the eigenspace
must be at least 1 .
EXAMPLE4
Solution: We have
A-tl/=
17
20
] [ ] [
l
-15
-tl
-18
0
of Example 1.
17-tl
-15
20
-18 -,{
(You should set up your calculations like this: you will need A - tll later when you.find
the eigenvectors.) Then
det(A -tl/)
17-,{
20
-15
_
18
tl
EXAMPLE4
(continued)
-3,
(-3)I [ ] [ -30/4]
0
[3i4].
[3i4]
-3
{[3i4]}.
=
of A corresponding to il
eigenspace for il
-3
-15
-15
20
20
is v
is Span
A_ 21
{[ ]}
Span
15
20
0 is v
- 15
-20
2:
] [
-1
0
[ ],
multiples of
293
2 is
[n
Observe in Example
Definition
Characteristic Polynomial
polynomial of A.
(1) il1 is a root of C(il) if and only if (il - il1) is a factor of C(il).
(2) The total number of roots (real and complex, counting repetitions) is n.
(3)
(4)
Complex roots of the equation occur in "conjugate pairs," so that the total num
ber of complex roots must be even.
If n is odd, there must be at least one real root.
(5) If the entries of A are integers, since the leading coefficient of the characteristic
polynomial is 1, any rational root must in fact be an integer.
294
Chapter 6
EXAMPLES
[b l
So, A. = 1 is a double root (that is, (A.- 1) appears as a factor of C(A.) twice) so A. = 1
is the only eigenvalue of A.
For A.= 1, we have
A-A.I=
EXERCISE2
{[b]}
[ b ].
rn b]
lit Thus, the eigenspace for A = 1 is
t E
[ ].
EXAMPLE6
5 -5
-3
7 -3
-7
Solution: We have
C(,1)
det(A-A.I)=
-3-A.
-7
-7
5
-5
9-A.
-5
7
-3- ,i
Expanding this determinant along some row or column will involve a fair number of
calculations. Also, we will end up with a degree 3 polynomial, which may not be easy
to factor. But this is just a determinant, so we can use properties of determinants to
make it easier. Since adding a multiple of one row to another does not change the
determinant, we get by subtracting row 2 from row 3,
C(A.)=
-3-A.
-7
9- ,i
-2+,1
-5
-5
2-,1
C(A.)
(-2 + A.)(-1)((-3
,1)(-5)- (-5)(-7))
Section 6.1
EXAMPLE6
(continued)
[-5 5 -51 [
-5
ol
[n
295
1 -1
A- ..t1/= -7
-7
-5
eigenspace of;, is
For ..t2 = 4,
ml}
A- ..t2/ =
-7
{[11}
For A3 = -3,
is V =
is V =
-7
{[; l}
[ =1[ =1
m
A-..t3/= -
eigenspa of ,!3 is
[= =1 -1
[H
-7
eigenspace of A2 is
EXAMPLE 7
Find the eigenva!ues and eigenvectors of A=
is V
[: ; l
Solution: We have
1 -..t
C(..t) =
1 - ,t
1
1- ,l
1 - ,t
1
,t
,t
= -,l(-,l +
2
,t - 2A) = -..t c..t - 3)
2
1
-..t
l[ l
Therefore, the eigenvalues of Aare ..t1 = 0 (which occurs twice) and ..t2 = 3.
For ..t1 = 0,
A- ..t1/=
1 1 l
1 1 1
1 1 1
0
0
0
0
0
0
296
Chapter 6
EXAMPLE 7
(continued)
/l.2
3,
A-,l,/
A2
Definition
Algebraic Multiplicity
Geometric Multiplicity
EXAMPLES
/l.
n x n
/l
/l
multiplicity
algebraic multiplicity
/l.
InpolExampl
e
t
h
e
ei
g
enval
u
e
has
al
g
ebrai
c
mul
t
i
p
l
i
c
i
t
y
2
si
n
ce
t
h
e
charact
e
ri
s
t
i
c
ynomial is - - and has geometric multiplicity since a basis for
its eigenspace is{[]}.
IInn Exampl
ee eachthe eieiggenval
envaluuee has0 hasalgalebraigebraic mulc andtipligeomet
city andricgeomet
rilcicmulity 2,tiplandicitythe
Exampl
mul
t
i
p
eigenvalue has algebraic and geometric multiplicity
5,
(/l
1)(/1.
1),
/l
/l
6,
7,
3
EXERCISE 3
Theorem 2
/l
geometric
1.
1.
Let [00 -20 22]. Show that and -2 are both eigenvalues of and
determine the algebraic and geometric multiplicity of both of these eigenvalues.
sectiThese
on. definitions lead to some theorems that wil be very important in the next
Let be an eigenvalue of an matrix Then
geometric multiplicity algebraic multiplicity
A
-3
-4
/l
/l.1
n x n
/l.2
A.
297
A is an n x n matrix
with
plicity equals their algebraic multiplicity, then the sum of the geometric multiplicities
of all eigenvalues equals the sum of the algebraic multiplicities, which equals n (since
an n-th degree polynomials has exactly n roots). Hence, if we collect the basis vectors
from the eigenspaces of all k eigenvalues, we will end up with n vectors in JR.I!. We now
prove that eigenvectors from eigenspaces of different eigenvalues are necessarily lin
early independent, and hence this collection of n eigenvectors will form a basis for JRn.
Theorem 3
Suppose that Ai, ..., Ak are distinct (Ai * A1) eigenvalues of an n x n matrix
A,
with corresponding eigenvectors v1, ... , vb respectively. Then {V1, ... , vk} is linearly
independent.
0.
2'.
Avi
Aivi,
we have
(A
AJ)Vi
0 and
Ak+I I gives
{v1,
vk} is linearly independent; thus, all the coeffi
AJ; hence, we must have c1
ck
0. Thus (6.1)
becomes
0 + Ck+JVk+I
But Vk+1 *
independent.
Remark
In this book, most eigenvalues tum out to be integers. This is somewhat unrealistic; in
real world applications, eigenvalues are often not rational numbers. Effective computer
methods for finding eigenvalues depend on the theory of eigenvectors and eigenvalues.
298
Chapter 6
PROBLEMS 6.1
Practice Problems
Al Let A
[= 2 3 ]
-1
(e)
[ ]
[ ]
(b)
(d)
[ ; ]
[-26 29]
-75
(f)
--63]
[36
[! ]
(e)
[ -J
(b)
(d)
(f)
[ -J
[-2- = 63 ]
-7
[; 3i
1
10
Homework Problems
Bl Let A
[- - -1.6
8
-1
23
[ 53 1
1
(e) 0
[-2 236 61
-
(f)
-1
[[-; =J
-]
;]
[ 2
[=
!]
_;]
(c)
(e)
[; ]
[ ]
[ ]
(b)
(d)
(t)
[= =; j]
299
Computer Problems
Cl Use a computer to determine the eigenvalues
(b)
(c)
[
[l -i]
-
-5
-2
1
3
2
2
2
2
-1.28185
0.76414
-0.83198
2.42918
-0.67401
2.34270
[[ ] [ ] [ ]
C2 Let A
1.21
Verify that
-5
-2
-2
2
2
4
4
2.89316
-0.70562
1.67682
-0.34 ,
0.87
1.31
2.15
-0.21
-1.85
and
are
0.67
2.10
3
-4
4
4
Conceptual Problems
Dl Suppose that vis an eigenvector of both the matrix
A and the matrix B, with corresponding eigenvalue
geometric multiplicity.
DS Suppose that A is an
f a;k =
k=l
is
6.2 Diagonalization
At the end of the last section, we showed that if the k distinct eigenvalues ,11,
, Ak
of an n x n matrix A all had the property that their geometric multiplicity equalled
their algebraic multiplicity, then we could find a basis for JR.11 of eigenvectors of A by
collecting the basis vectors from the eigenspaces of each of the k eigenvalues. We now
see that this basis of eigenvectors is extremely useful.
Suppose that A is an n x n matrix for which there is a basis {V1, . . , v11} of eigen
300
Chapter 6
J, then we get
we let P =[v1
v1
1
AP=A[v1
vn
Avn
=[Av1
= [,,t1vi
AnVn
=[vi
vn
A1
A2
0
=PD
0
0
An
Definition
If there exists an invertible matrix P and diagonal matrix D such that p-l AP=D, then
Diagonalizable
we say that A is diagonalizable (some people prefer "diagonable")and that the matrix
P diagonalizes A to its diagonal form D.
It may be tempting to think that p-l AP = D implies that A = D since P and
1
p- are inverses. However, this isnot true in general since matrix multiplication is not
commutative. Not surprisingly, though, if A and B are matrices such that p-1AP = B
for some invertible matrix P, then A and B have many similarities.
Theorem 1
If A and B are n x n matrices such that p-1AP = B for some invertible matrix P,
then A and B have
11
(4) The same trace, where the trace of a matrix A is defined by tr A= .L:
a;;
i=l
(!)was proved as Problem D4 in Section 5.2. The proofs of (2), (3), and (4) are
left as Problems D 1, D2, and D3, respectively.
This theorem motivates the following definition.
Definition
If A and Bare n x n matrices such that p-1AP = Bfor some invertible matrix P, then
Similar Matrices
Theorem 2
301
[Diagonalization Theorem]
An n x n matrix A can be diagonalized if and only if there exists a basis for JR11 of
eigenvectors of A. If such a basis
{V1,
Ai
[v1
v11
is an eigenvalue
From the Diagonalization Theorem and our work above, we immediately get the
following two useful corollaries.
Corollary 3
Corollary 4
Remark
Observe that it is possible for a matrix A with real entries to have non-real eigenvalues,
which will lead to non-real eigenvectors. In this case, there cannot exist a basis for
JR11 of eigenvectors of A, and so we will say that A is not diagonalizable over JR. In
Chapter 9, we will examine the case where complex eigenvalues and eigenvectors are
allowed.
EXAMPLE 1
A=
[ ;].
t
p- AP =
D, where
A1 = 5,
At = 5 and Az = -1.
we get
] [
3 1
0
-3
So,
Vt =
For
So,
V'2 =
[ ]
A2
is an eigenvector for
r-]
Thus,
At = 5 and {V i }
-1
0
-1, we get
is an eigenvector for
{v1, v2}
A2= -1
and
[vt
v2 =
[ -].
we get
302
Chapter 6
EXAMPLE 1
(continued)
EXAMPLE2
Solution:
(-2
4)
1)
J1
-1
EXERCISE 1
EXAMPLE3
303
7
Is the matrix A
11
[=! =]
-4
diagonalizable?
-3
C(tl)
det(A
- -tl
-4
0
11 - tl
- 3 + tl
11 - tl
-6
8
-3-tl
1-1)
1.
3, we get A - tl1/
{[1[2]}
-6
3 - tl
28)
1 is an eigenvalue
[=: =] [ =12] .
-4
the eigenspace is
-1-tl
-4
-4
Thus, tl1
tll)
-6
EXERCISE 2
EXAMPLE4
Show that A
[ ]
is not diagonalizable.
[ -]
det(A - tll)
1-,,l 1
l
-1
-tl
tl 2 + 1
0 has no real solutions, the matrix A has no real eigenvalues and hence
[: ].
2bxy
cy2
easily recognize the graph as an ellipse, a hyperbola, or perhaps some degenerate case.
This problem will be discussed in Section 8 .3.
304
Chapter 6
[]
in each group. For some situations, the change from month to month can be described
by saying that the vector jJ changes according to the rule
jJ(n + 1) =AjJ(n)
whereA is some known 2 x 2 matrix. It follows that p(n)= N p(O). We are often inter
ested in understanding what happens to the population "in the long run." This requires
us to calculateA11 for n large. This problem is easy to deal with if we can diagonal
izeA. Particular examples of this kind are Markov processes, which are discussed in
Section 6.3.
One very important application of diagonalization and the related idea of eigen
vectors is the solution of systems of linear differential equations. This application is
discussed in Section 6.4.
In Section 4.6 we saw that if L : JR.11 JR.11 is a linear transformation, then its matrix
with respect to the basis '13 is determined from its standard matrix by the equation
[L21
] = P-1[L]sP
where P = v1
v11 is the change of basis matrix. Examples 5 and 7 in
Section 4.6 show that we can more easily give a geometrical interpretation of a lin
ear mapping L if there is a basis '13 such that [L]21 is in diagonal form. Hence, our
diagonalization process is a method for finding such a geometrically natural basis. In
particular, if the standard matrix of Lis diagonalizable, then the basis for JR.11 of eigen
vectors forms the geometrically natural basis.
PROBLEMS 6.2
Practice Problems
Al By checking whether columns of P are eigenvec
(a)A=
(b)A=
[ ]
[ -l
11
9
6
4
'
[ ]
[ n
P=
p=
(c) A=
(d) A=
-87]' [ ]
- ]
[: 2 2 n - :J
4 4
4
p=
4,
(f) A=
P=
(a) A=
(b) A=
(c) A=
(d) A=
(e) A=
[ ]
[-2 3]
[- -]
4
(g) A=
(a) A=
(b) A=
(d) A=
[1 1 11
[- -17- -8-1
(e) A=
(f) A=
-6
6
-4
-3
(c) A=
-3
-2- 72 1]
[ 21
-[ 1
12 81
(g) A=
305
[ ; ]
[: :]
[-; -]
_
A2.
[-2- 2: -=!]2
[2 2 1
[-1- -2 i
[ -:1
0
Homework Problems
[- n [-1 n
[ n [ - ]
[ n [ - ]
7 2
diagonal.
(a) A=
(b) A=
(c) A=
(d) A
p=
P=
p=
4,
[-2 ;]
[ ]
[ ;]
p=
-4
(b) A=
(c) A=
(d) A=
(e) A=
[ 3 -2 ]
[- -22 -]
6
6
4
-9
306
Chapter 6
<DA=
(g) A=
H -i
H -;1
-1
2
-3
-3
-4
-2
3
(e) A=
[= ]
[ =1
[ ; 1
[ - -]
2
(d) A=
[ 1
[ -1
[ ! -1
A=
(f)
A=
-1
-2
Conceptual Problems
Dl Prove that if A and B are similar, then A and B have
the same eigenvalues.
tr AB= tr BA.
(b) Use the result of part (a) to prove that if A and
B are similar, then tr A = tr B.
[] [].
[H
[
i].
Ul
and
<espec6 vel Y
[- - ]
-6
A2 (g).
12
[:
a b
a
a b
a
a+b
by considering p- AP.
tic polynomial is det A. (Hint: How do you find
the constant term in any polynomial p(A.)?)
(c) Without assuming that A is diagonalizable,
show that det A is equal to the product of the
roots of the characteristic equation of A (in
cluding any repeated roots and complex roots).
(Hint: Consider the constant term in the char
acteristic equation and the factored version of
A=
Theorem 1.
(b) Use the result of part (a) to determine, by in
respectively.
A=
vectors
that equation.)
DB Let A be an
Section 6.3
307
A=
is diagonalized by
PDP-1,
p -1 P
= A
Am=
=
(PDP-1r (PDP-1)(PDP-1)
(PDP-1)
PD(P-1P)D(P-1P)D (P-1P)DP PDmp-l
be valuable tools in these applications. One such application is the study of Markov
processes. After discussing Markov processes, we tum the question around and show
how the "power method" uses powers of a matrix
to determine an eigenvalue of
A.
EXAMPLE 1
Smith and Jones are the only competing suppliers of communication services in their
community. At present, they each have a
50%
has recently upgraded his service, and a survey indicates that from one month to the
next,
hand,
90%
70%
30%10%
for six months, how large are their market shares? If this goes on for a long time, how
big will Smith's share become?
Solution: Let
and let
lm
Sm
30%
Sm++ lm =
Sm+l= 0.90S m 0 3 m
lm+l = O.lSm 0 7lm
90%
100%
of his previous
Similarly,
0 9 0.7
0 3] [Sm]
[Sm+l] = [0.1
lm
lm+l
.
The matrix T
= [: :]
[:]
ym [0.0.55]
[::: ]
[:;]
at
and
form large.
To answer the first question, we might compute T6 directly. For the second ques
tion, this approach is not reasonable, and instead we diagonalize. We find that
il1
=1
308
Chapter 6
EXAMPLE 1
(continued)
is an eigenvalue of
with eigenvector
V2
with eigenvector
= [ l
_
v1
= []
.and
il.2
= 0.6
Thus,
It follows that
ym
PDmp-1
0 ] [11 - 1]
= [31 -11] [ l0m (0.6)"'
4 3
We could now answer our question directly, but we get a simpler calculation if we
observe that the eigenvectors form a basis, so we can write
Then,
Then, by linearity,
[] = c1Tmv1+c2Tmv2
= C1ilV1+c2il1v2
= (So+lo)[]+ (So - 3lo)(0.6)"' [ ]
= 6,
So= lo= 0.5.
[s ] = 4 [31]- 4 co.6)6 [-11]
1 [3- 0.0117
::::: 4 1+0.0117)
::::: [0.747]
0.253
74.7%
(0.6)"'
S = 0.75 loo = 0.25.
75%
75%.
(0.6)"' 0 So+lo= 1.
So lo
ym
Now
Whenm
6
16
oo
and
of the market.
Now look carefully: we get the same answer in the long run, no
and
are because
and
309
(1) Each column of T has sum l . This means that all of Smith's customers show
up a month later as customers of Smith or Jones; the same is true for Jones's
customers. No customers are lost from the system and none are added after the
process begins.
tij
t11 =
t21 =
0.9
tij
time m + 1 .
[l []
ym
Since
1 and
1, it follows that
Thus, it follows from (1) that each state vector has the same column sum. In our
example,
and
[n
[j:].
[j:J
[ ]=[ ]
3/4
1/4
3/4
1/4
[].
[]
Definition
Ann xn matrix Tis the Markov matrix (or transition matrix) of ann-state Markov
Markov Matrix
process if
Markov Process
(1)
tiJ
i=I tiJ =
n
1 for each j.
310
Chapter 6
s1
+ Sn
s;
;:::
s1
for
+ Sn = constant.
. [0.0.19 0.3] .
1
The matnx
O.S
.
1s not a Markov matnx
. smce
.
.
the sum of the entries
m the second
EXAMPLE3
such that
Sn
1.
Remark
EXAMPLE2
s[ :1
rn: :].
A_ I=
1,
1.
1 is
EXERCISE 1
Determine which of the following matrices is a Markov matrix. Find the fixed-state
vector of the Markov matrix.
(a) A
[0.50.4 0.0.651
(b) B =
[0.4 0.61
0.6 0.4
The goal with the Markov process is to establish the behaviour of a sequence with
states s, Ts, T2 S,..., T'n s. If possible, we want to say something about the limit of
Tms as
1,
the problem. It is beyond the scope of this book to establish all the properties of the
Markov process, but some of the prope1ties are easy to prove, and others are easy to
illustrate if we make extra assumptions.
PROPERTY 1.
311
1.
Proof: Since each column ofT has sum 1, each column of (T - 1/) has sum 0. Hence,
the sum of the rows of (T - 11) is the zero vector. Thus the rows are linearly dependent,
and (T - 1/) has rank less than n, so det(T - 1/)
0. Therefore, 1 is an eigenvalue
=
ofT.
PROPERTY 2.
1 has
sj
n.
This property is important because it means that the eigenvector S* is a real state of
the process. In fact, it is a fixed or invariant state:
PROPERTY 3.
IA;I
::;
1.
To see why we expect this, let us assume thatT is diagonalizable, with distinct eigen
values I, A2, ... , An and corresponding eigenvectors s*, s 2, . ..
state s can be written
, S,1
It follows that
If any
l,1;1
>
1, the term 1,i;nl would become much larger than the other terms when
is large; it would follow that T"' s has some coordinates with magnitude greater than
PROPERTY 4.
s;
::; 1,
so we must have
1 satisfy IA;I
state S, T"' s ---t S* as m ---t oo: all states tend to the invariant state S* under the process.
Notice that in the diagonalizable case, the fact that ym s
---t
EXERCISE 2
[ )
conclusion of Property 4. However, it also does not satisfy the extra assumption of
Property 4. It is worthwhile to explore this "bad" case.
Let s=
[ ; l
312
Chapter 6
AS(m)
difference equation.
Markov processes form a special class of this large class of systems of linear
difference equations, but there are applications that do not fit the Markov assumptions.
For example, in population models, we might wish to consider deaths (so that some
column sums of A would be less than 1) or births, or even multiple births (so that some
entries in A would be greater than 1). Similar considerations apply to some economic
models, which are represented by matrix models. A proper discussion of such models
requires more theory than is developed in this book.
T hen
and
A111 1
c1tlv1
CnA Vn
For m large, l..l'fl is much greater than all other terms. If we divide by c1..l'f, then all
terms on the right-hand side will be negligibly small except for v1, so we will be able
to identify Vt. By calculating Av1, we determine tl1
To make this into an effective procedure, we must control the size of the vectors: if
..t 1
>
1, then tl'1"
-t oo
-t
To avoid these problems, we normalize the vector at each step (that is, convert it to a
vector of length 1).
313
Algorithm 1
EXAMPLE4
Yo =
[ ]
y1
Yi
by
. Then
-12.02
xi
il1iii
12 =Ayi
y2
13 =Ah
y
3
X4 =Ah
'
.
.
.
,
-t
At this point, we JU dge that .rm
-t
.v1 =
, so -t
_1 [1] [0.707]
1 0.707
[-1213 -56] [0.707]
0.707 [ 13.44]
0.745]
[-0.667
[ 0.712]
[ 5.683]'
-0.702
-5.605
[-5.034
5.044]'
[ 0.7078]
-0.7063
[-4.9621
4.9636]
[-0.7070
0.7072]
0.707]
[_0.707]
0.707 7.[_0.707
11 =Ayo
=
[ ; -]
.
.
is an eigenvector of A,
Many questions arise with the power method. W hat if we poorly choose the initial
vector? If we choose x0 in the subspace spanned by all eigenvectors of A except vi,
the method will fail to give v1. How do we decide when to stop repeating the steps
of the procedure? For a computer version of the algorithm, it would be important to
have tests to decide that the procedure has converged-or that it will never converge.
Once we have determined the dominant eigenvalue of A, how can we determine
other eigenvalues? If A is invertible, the dominant eigenvalue of A-1 would give the
reciprocal of the eigenvalue of A with the smallest absolute value. Another approach
is to observe that if one eigenvalue /l.i is known, then eigenvalues of A - /l.il will give
us information about eigenvalues of A. (See Problem 6.2.D9.)
314
Chapter 6
PROBLEMS 6.3
Practice Problems
Al Determine which of the following matrices are
the eigenvalue ,1
(a)
(b)
(c)
(d)
[
[
[
[
]
]
0.2
0.6
0.8
0.3
0.3
0.6
0.7
0.4
0.7
0.3
0.0
0.1
0.2
0.6
0.2
0.1
0.9
0.9
0.0
0.1
0.1
0.9
0.0
0.0
0.1
0.9
1).
]
1
[ -J.10 [n
[ ]. 10 []
=
Homework Problems
Bl Determine which of the following matrices are
the eigenvalue ,1
(a)
(b)
(c)
(d)
[
[
[
[
0.4
0.7
0.5
0.3
]
]
1).
0.5
0.6
0.5
0.4
0.8
0.3
0.2
0.0
0.6
0.2
0.2
0.1
0.6
0.8
0.1
0.1
0.9
0.2
0.6
0.1
0.1
0.2
]
]
go to Johnson.
(a) Model this as a Markov process and determine
the steady-state distribution of customers.
(b) Determine a general expression for Johnson
and Thomson's shares of the customers, given
an initial state where Johnson has 25% and
Thomson has 75%.
Section 6.4
315
[]
Computer Problems
Cl Use the powec method whh initial vectoc
[]
to
3.5
4.5
to
4. 5
.
3.5
2.89316
-1.28185
-0.70562
0.76414
1.67682
-0.83198
2.42918
Conceptual Problems
Dl (a) Let T be the transition matrix for a two-state
Markov process. Show that the eigenvalue that
is not I is A.2
t11 + t22
1.
a!b
[].
I(Tx)k I xk.
=
k=I
k=I
value A. f. 1, then
I vk
0.
k=I
Equations
This section requires knowledge of the exponential function, its derivative, and first
order linear differential equations. The ideas are not used elsewhere in this book.
Consider two tanks, Y and Z, each containing 1000 litres of a salt solution. At a
initial time, t
concentration in tank Z. In each tank the solution is well stirred, so that the concentra
tion is constant throughout the tank. The two tanks are joined by pipes; through one
pipe, solution is pumped from Y to Z at a rate of 20 L/h; through the other, solution is
316
Chapter 6
pumped from Z to Y at the same rate. The problem is to determine the amount of salt
y(t) t.
(z/1000)
dy(20)(z/1000)
-0.02y 0.02z.
ydt z
Let
t.
(20)(y/1000)
t, z(t)
t (y/1000)
and let
be the
is
kg/L.
Similarly,
and
dy -0.02y 0.02z
dzdt 0.02y -0.02z
dt
d [y] [-0.02 0.02] [y]
dt z 0.02 _0.02 z
+
.
.
.
'
It 1s convernent to rewnte th 1s system m the form
How can we solve this system? Well, it might be easier if we could change vari
.
[l1].
[-0.0.0022 -0.02
0.202]2. ' 0 .
-0.04,
r- l
[-11 1]
[11 -11]
[:]
[]
4.6: [] [:l
d [y* [y*]
dt z*] - z*
1s llJ
value is il2
by p
-l
.
'With p
'
, wit h correspond mg eigenvector
l
2
Hence, A is diagonalized
1 .
.
The other eigen-
-P
AP
d y*] d [y*
dt [z* - dt z*]
-P-
-P
Multiply both sides of the system of equations (on the left) by p-i. Since Pdiagonal
izes A, we get
and
These equations are "decoupled," and we can easily solve each of them by using simple
one-variable calculus.
dy*
dt
x(t)
b is a constant.
dz*
dt
317
a.
The only
be-0041, where
Now we need to express the solution in terms of the original variables y and z:
-1
1
] [aa
y* - z*
y*
z*
y* + z* -
][ ] [
[] a [ ]
=
- be-0.04t
+ be-0.041
+be-0041
[ ]
-
oo,
a,
as we might expect.
:tce,i1 [:]
[:]
[]
=
ce,i1
A.ce,i1
[:].
[:l
is an eigenvector of
A.. We find the two eigenvalues A.1 and A.2 and the corresponding
eigenvectors v1 and v2, as above. Observe that since our problem is a linear homoge
A, with eigenvalue
neous problem, the general solution will be an arbitrary linear combination of the two
solutions e,i11v1 and e,i21v2. This matches the general solution we found above.
General Discussion
There are many other problems that give rise to systems of linear homogeneous or
dinary differential equations (for example, electrical circuits or a mechanical system
consisting of springs). Many of these systems are much larger than the example we
considered. Methods for solving these systems make extensive use of eigenvectors and
eigenvalues, and they require methods for dealing with cases where the characteristic
equation has complex roots.
318
Chapter 6
PROBLEMS 6.4
Practice Problems
Al Find the general solution of each of the following
systems of linear differential equations.
(a)
:t []
(b)
[! -J []
:!_
[y] [
dt z
0.2
0.1
0 .7
-0.4
] [y]
z
Homework Problems
Bl
:t []
[-: -:;][]
(b)
(c)
:t []
:
t
[- _:] []
[;] [- - i [;]
z
11
-5
-8
CHAPTER REVIEW
Suggestions for Student Review
1
(b) Is there any case where you can tell from the
/I.I)
0. (Section 6.1)
plicity).
(a) What conditions on these eigenvalues guaran
tees that A is diagonalizable over JR.? (Sec
(Section 6.2)
4 Use the idea suggested in Problem 6.2.D 4 to create
:t
x = Ax? (Sec
tion 6.4)
tion 6.2)
Chapter Quiz
El Let A =
[ - ]
8
-2
(a)[!]
(b)
[ ;
-11
(c)
m c{:J
- - i
5
Chapter Review
[- -11
A - 21?
onalizable?
Is A diag-
319
E6 Let A
ES Suppose that A is a 3
det A =
0,
det(A + 2/)
0,
det(A - 3/)
i=l
Xi =
:t [] [: :] []
Ax=
o?
Further Problems
Fl (a) Suppose that A and B are square matrices such
that AB
then
F2 If det B
0,
eigenvalues.
F3 Suppose that A is an
eigenvalues ;i1,
vectors v1,
, Vn, respectively. By representing 1
with respect to the basis of eigenvectors, show that
;i,J)x
for every
MyMathlab
F4 For an invertible
-1
to n
in A (that is, a linear combination of
2
11 1
{A - , ... 'A , A,/}.
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 7
Orthonormal Bas es
CHAPTER OUTLINE
7.1 Orthonormal Bases and Orthogonal Matrices
7.2 Projections and the Gram-Schmidt Procedure
7.3 Method of Least Squares
7.4 Inner Product Spaces
7.5 Fourier Series
In Section 1.4 we saw that we can use a projection to find a point P in a plane that is
closest to some other point Q that is not in the plane. We can view this as finding the
point in the plane that best approximates Q. In many applications, we want to find a
best approximation. Thus, it is very useful to generalize our work with projections from
Section 1.4 to not only general subspaces of JR.11 but also to general vector spaces. You
may find it helpful to review Sections 1.3 and 1.4 carefully before proceeding with
this chapter.
Orthonormal Bases
Definition
O whenever i
j.
Orthogonal
EXAMPLE l
The set
{ \ : -! }
{ ! =: n
,
If the zero vector is excluded, orthogonal sets have one very nice property.
322
Chapter 7
Theorem 1
Orthonormal Bases
If {V 1,
independent.
0.
0 unless i
j. Moreover, v; f.
0,
Ov;
0. Since
Remark
The trick used in this proof of taking the dot product of each side with one of the
vectors v; is an amazingly useful trick. Many of the things we do with orthogonal sets
depend on it.
In addition to being mutually orthogonal, we want the vectors to be unit vectors.
Definition
A set {V1,
Orthonormal
Notice that an orthonormal set of vectors does not contain the zero vector, since all
vectors have length 1. It follows from Theorem 1 that orthonormal sets are necessarily
linearly independent.
EXAMPLE2
Any subset of the standard basis vectors in JR11 is an orthonormal set. For example, in
JR6' {e,' e2, es, e6} is an orthonormal set of four vectors (where, as usual, e; is the i-th
standard basis vector).
EXAMPLE3
The set
{ ; j -n
l
'l
' },
of the vectors in Example 1, so they are certainly mutually orthogonal. They have been
normalized so that each vector has length 1.
EXERCISE 1
Verify that the set
{ [il rl [-]}
323
Many arguments based on orthonormal sets could be given for orthogonal sets of
non-zero vectors. However, the general arguments are slightly simpler for orthonor
mal sets since llvill
orthogonal sets and postpone the normalization that often introduces square roots until
the end. (Compare Examples 1 and 3.) The arguments here will usually be given for
orthonormal sets.
basis for JR.11 and that 1 is any vector in JR.12 To find the :B-coordinates of 1, we must
find b1,... , b such that
11
(7.1)
If :B were an arbitrary basis, the procedure would be to solve the resulting system of
n equations in n variables. However, since :B is an orthonormal basis, we can use our
amazingly useful trick: take the dot product of (7.1) with vi to get
x v;
because vi v1
0 for i
. + o + biCvi vi)+ o + + o
.
* j and
vi V;
bi
Theorem 2
If :B
W1,. . ., Vn}
b;
It follows that 1 can be written as
x v;
324
Chapter 7
Orthonormal Bases
EXAMPLE4
13
3
4
{ 1
1 1 -1
1'-../2 1 , -../2
1 -1
2 1 '2
-1
bi
b2
v1
1 . v2
b3
v3
b4
v4
12"(1+2 + +
-(1-2 +
2
(-1+ + +
1
-CO
-../2 +2 + O
3
4)
4)
-1
0)
4)
-../2
-Yz
-1
-../2 .
- -../2
EXERCISE 2
Ld
{ rn ul H]}
[!]
},
'
},
Another technical advantage of using orthonormal bases is related to the first one.
Often it is necessary to calculate the lengths and dot products of vectors whose co
ordinates are given with respect to some basis other than the standard basis. If the
basis is not orthonormal, the calculations are a little ugly, but they are quite simple
{v1,
Y1V1 +
X1V1 +
+ X11Vn
and
Section 7.1
that
vi Vj
1 Y
0 for i * j and vi v;
=
1 gives
+ ''' +
Y11V11)
X1Y1
325
X2Y2
+ ... +
111112
and
X11Y11
11
x + +
Thus, the formulas in the new coordinates look exactly like the formulas in standard
..
EXAMPLES
Ut B
]
jl o
[
Solution:
=
},
Detennine
111112
and
[i'] . [i'
,
J.
i' , jlJ.
[ ] [
1 y
i'
,Y
R3 such that
andt jl.
111112
EXERCISE 3
and let
Ul. Ul
Ul Hl
[i']
Ul
and
-2
111112
and
orthonormal basis.
11
inner
To keep the writing short, we give the argument in JR.3, but the corresponding argument
a change of coordinates matrix between the standard basis and an orthonormal basis.
works in any dimension.
326
Chapter 7
Orthonormal Bases
Let <J3
[vt v2 v3J.
From Section 4.4, P is the change of coordinates matrix from <J3-coordinates to coordi
nates with respect to the standard basis S. Now, consider the product of the transpose
of P with P: the rows of
viv2 viv3
viv2 viv3
vrv2 vrv3
Vt v2 Vt v3
v2 v2 v2 v3
v3 v2 v3 v3
Remark
We have used the fact that the matrix multiplication xT y equals the dot product x y.
We will use this very impo1tant fact many times throughout the rest of the book.
It follows that if P is a square matrix whose columns are orthonormal, then P is
invertible and
Definition
Orthogonal Matrix
p-t pT. Matrices with this property are given a special name.
=
pTP
pTP.
Theorem 3
(1) P is orthogonal.
(2) The columns of P form an orthonormal set.
(3) The rows of P form an orthonormal set.
Proof: Let P
[vt
vf VJ v;
Hence , ppT
I if and only if v; v;
1 and v; v1
=
VJ
ppT
I.
Section 7.1
327
Remark
Observe that such matrices should probably be called orthonormal matrices, but the
name orthogonal matrix is the name everybody uses. Be sure that you remember that
an orthogonal matrix has orthonormal columns and rows.
EXAMPLE6
The set
Cose
. e
sm
EXAMPLE 7
The set
{[ ;] [-:: :]}
]
,
-[
-sine .
cose
_
T
.
Hence, p-l - p IS an orthogonal matnx.
cose
sine
- sm
. e
cose
1/.../2
1/Y3
1/\/6
1/.../2 -1/Y3 -1/\/6
2/\/6
0
-1/Y3
-l
p
T
p
1/.../2
1 /.../2
l/-fJ -1/Y3 -1 Y3
11\16 -11\16 2/\/6
EXAMPLES
0
1/2
1/2 -1/.../2
0
1/2 -1/2
1/.../2
0
1/2
1
1/2
1.../2
0
1/2 -1/2
-1/.../2
EXERCISE4
Verify that P
0
1/.../2
1/.../2
j-{3
1
-1/Y3 1/Y3
1/\/6 2/\/6
-1/\/6
/.
328
Chapter 7
Orthonormal Bases
matrix is used to find the standard matrix of a rotation transformation about an axis
that is not a coordinate axis. (This was one question we could not answer in
Chapter 3.)
EXAMPLE9
ii=
l [I
0
0
1 0
0
R1 = 0 cosn/3 - sinn/3 = 0 1/2 - Y312
0 sinn/3 cosn/3
0 Y312 1/2
This will also be the 13-matrix of the rotation in this problem if there exists a basis
13
(3) 13 is right-handed (so that we can correctly include the counterclockwise sense
of the rotation with respect to a right-handed basis).
Let us find such a basis.
To start, let
/. = 11:11 = },
[; l
/. and
0 = f1
-+
-+
= 1 (X1 + X2 + X3)
Y3
[-i]
is orthogonal to
/..
many other choices for this vector; this is just one simple choice.) To form a right
handed system, we can now take the third vector to be
and
and the
orthogonal change of coordinates matrix from this basis to the standard basis is
=[A
Ii
[
1/vf3 1/-Y'i.
A]= 11Y3 - 11-Y'i.
1/Y3
1/../6
11../6
0 -2/../6
Section 7.1
EXAMPLE9
(continued)
Since [L]1l
[L]s
P[L]1lP-
2/3
2/3
-1/3
-1/3
2/3
2/3
2/3
-1/3
2/3
329
should be since it defines the axis of the rotation represented by the matrix. Notice
also that the matrix L
[ ]s is itself an orthogonal matrix. Since a rotation transformation
always maps the standard basis to a new orthonormal basis, its matrix can always be
taken as a change of coordinates matrix, and it must be orthogonal.
[
{ [;] ,[-:]}
. p
The matnx
'13
cose
Stne
- sine .
.
. from the bas1s
ts the c hange of coord'mates matnx
COS e
described as a "rotation of axes through angle 8" because each of the basis vectors
in '13 is obtained from the corresponding standard basis vector by a rotation through
angle e.
Treatments of rotation of axes often emphasize the change of coordinates equation.
4.4.2,
Recall Theorem
from '13 to S, then p-l is the change of coordinates matrix from S to '13. Then, for any
x E JR2 with [x]1l
[:J
p-1 [x]s. If the change of basis is a rotation of axes, then Pis an orthogonal
1
matrix, so p- = pr. Thus, the change of coordinates equation for this rotation of axes
[x]1l
can be written as
[:] -[ :
_
][]
Or it can be written as two equations for the "new" coordinates in terms of the "old":
b1
b2
X1 COS 8 + X2 sin 8
These equations could also be derived using a fairly simple trigonometric argument.
.
cose - sine
.
1
.
The matnx
.
a SO appeared .tn s ect10n 3 . 3 as the Standard matnx
sm B
cos B
R
[ e] of the linear transformation of JR that rotates vectors counterclockwise through
angle e. Conceptually, this is quite different from a rotation of axes.
It can be confusing that the matrix for a rotation through e as a linear transfor
mation is the transpose of the change of coordinates matrix for a rotation of axes
330
Chapter 7
Orthonormal Bases
through e. In fact, what may seem even more confusing is the fact that if y ou re
place e with ( -e), one matrix turns into the other (because cos ( -e)
sin (-B)
cos e and
- sin B). One way to understand this is to imagine what happens to the
=
vector e1 under the two different scenarios. First, consider R to be the transformation
that rotates each vector by angle e, with 0 < e <
first quadrant that makes an angle e with the x1 -axis. Next, consider a rotation of axes
through (-B); denote the new axes by y1 and y2, respectively. Then e1 has not moved
but the axes have, and with respect to the new axes, e1 is in the new first quadrant and
makes an angle of e with respect to the Y1 -axis. Therefore, the new coordinates of e1
relative to the rotated axes are exactly the same as the standard coordinates of R(e1).
Compare Figures 7.1.1 and 7.1.2.
Figure 7.1.1
Figure 7.1.2
--+
The standard basis vector e1 is shown relative to axes obtained from the
standard axis by rotation through -e.
Exercises
Section 7.1
331
PROBLEMS 7.1
Practice Problems
Al Determine which of the following sets are orthog
(b)
(c)
(d)
{[ ] [ ] }
{[-: rni +m
{[l].[J HJ}
-21
{ 1 _-1: ' --01 ' 01}
{ Hl [] {!]}
(a) A=
(b) A=
(c) A=
(d) A=
A2 t S =
(e) A=
Find the coordi-
(c)Y=
A3 Let
:B
m
Hl
=
(b)t=
(d) Z=
1
-1
2 11 -1
{' :
1
'2
' Y2
Hl
m
-10
10 -1!}
-14
3-5
-13
-23
1
' Y2
(c) w=
24
_35
103
(b) y=
(d) z=
AS Let L
12/13]
[15/13
2/[ 3/513 -5/13
-4/5 -3/54/5]
[1/52/5 -1/5]
1/3[2/3 -2/32/32/5 -2/31/3]
2/31/3 1/32/3 2/32/3
[2/32/3 -2/31/3 -2/31/3]
JR.3
JR.3
[H
ih =
is orthogonal to
Define
-]
[
22. 1
gi/l g;l 1, 2, 3,
nents of
(b) Let
:B
-+
f;
=
g,
g3
-+
for
-+
{/i, h.f;}
is an
g,
so that
orthonormal basis.
].
..
L, [L]2J.
[L]2l=
}i[".].
[L]s.
Chapter 7
332
Orthonormal Bases
{[ -; l[ ; l [ ]}
l/
1/../6
1/../3
-1/Y2
1/../6
1/../3
1;Y2
orthonormal.
Homework Problems
gonal. For each orthogonal set, produce the cor
responding orthonormal set and the orthogonal
-2
(b) 1=
-4
0
4
(b)
( c)
{[][-i]}
(c ) y =
{[-:J.Ul[J}
{[ll {l HJ}
B2 Ut =
]
u
{ [ l n l }
}o
},
(b) A=
Find the
(c ) A=
m
Ul
(c )
B 3 Let '13 =
(b) x=
(d)Z=
{d
, 1 , -2
1 0
-2
2
2
-2
},
(d ) z=
-1 2
0
5
0
: !
-1
},
nJ
nl}
Find
-1
(d ) A=
(e) A=
YS -1/ Ysl
[2/1;-VS
-21-VSJ
2/YS -1/YS]
[-11-VS -21-VS
1/2 1/2 ]
[ 1/2 1/2
[
[
BS (a) Ut W, =
[i]
[ ]
-
and W, =
Determine a thfrd
Ul. ni.
and
Section 7.2
333
Conceptual Problems
Dl Verify that the product of two orthogonal matrices
is an orthogonal matrix.
n x n
Procedure
Projections onto a Subspace
The projection of a vector y onto another vector 1 was defined in Chapter 1 by find
ing a scalar multiple of 1, denoted proj1 y, and a vector perpendicular to 1, denoted
perp1 y, such that
y = proj1 y + perp1 y
Since we were just trying to find a scalar multiple of 1, the projection of y onto 1
could be viewed as projecting y onto the subspace spanned by 1. Similarly, we saw
how to find the projection of y onto a plane, which is just a 2-dimensional subspace. It
is natural and useful to define the projection of vectors onto more general subspaces.
Let y E JRn and let S be a subspace of JR11 To match what we did in Chapter 1, we
want to write y as
y = proj5 y + perp5 y
where proj5 y is a vector in S and perp5 y is a vector orthogonal to S. To do this, we
first observe that we need to define precisely what we mean by a vector orthogonal to
a subspace.
Definition
Orthogonal
1 s= o
Orthogonal Complement
for alls ES
We call the set of all vectors orthogonal to S the orthogonal complement of S and
denote it .L. That is,
s.l
Remark
s.l
{1E]Rn11. V; =0 for
1 i
k}
334
Chapter 7
EXAMPLE 1
Orthonormal Bases
vector in the plane, so we say that n is orthogonal to the plane. On the other hand,
we saw in Chapter 1 that the plane is the set of all vectors orthogonal to n (or any
scalar multiple of
plane.
EXAMPLE2
Let W =Span
{ , n
Find W" in R4
V1
= 0 and v
V4
= 0. This
g -r}
EXERCISE 1
Lets= Span
{i ;}
{ -r}
FindS"
Theorem 1
(3) If {v 1
. ,
You are asked to prove these facts in Problems D 1, D2, and D4.
We are now able to return to our goal of defining the projection of a vector 1
onto a subspaces of
. '
Section 7.2
335
We get
Observe that this is exactly what we have been looking for. In particular, we have
(x v1)111 + + (x vk)Vb
(x vk+t)Vk+l + + (x v11 )1111 , which is a vector inj_.
written x as a sum of
definition.
Definition
Projection onto
a Subspace
The projection of 1
perpendicular to is defined to
be
Remark
Observe that a key component for this definition is that we have an orthonormal basis
for the subspace. We could, of course, make a similar definition for the projection if
we have only an orthogonal basis. See Problem D6.
We have defined perp8 x so that we do not require an orthonormal basis for j_.
However, we have to ensure that this is a valid equation by verifying that perp8 x E j_.
For any 1 i k, we have
= v;
=
x - ( 0 + + 0 + (1 v;)(v; v;) + 0 + + 0)
v; x -v; .1
=0
since '13 is an orthonormal basis and the dot product is symmetric. Hence, perp81
is orthogonal to every vector in the orthonormal basis
of . Hence, it is
EXAMPLE3
Let= Span
{l : : }
,l
Solution:
-1
and let X=
W1. i12}
{d : }
,
-1
Thus,
336
Chapter 7
EXAMPLE3
Orthonormal Bases
we get
1
-1
G) i + (-3 ) i
1
4 1
4
(continued)
projs
perps 1
EXERCISE2
Id
(1
1 - projs 1
mJr m
and perps
i11)111 + (1 i12)112
1
-1
-5/2
2
5
7
4
4
-5/2
9/2
-5/2
-5/2
-9/2
-]
Deterntlne proj,1
1.
1 E JR3 onto
3
a plane in JR is the vector in the plane that is closest to 1. We now prove that the
projection of 1 E JR" onto a subspace of lR.11 is the vector in that is closest to 1.
Recall that we showed in Chapter 1 that the projection of a vector
Theorem 2
Approximation Theorem
1
projs 1.
Proof: Let
111 - Sll
is s
E that
{i11, ... , vd be an orthonormal basis for and let lVk+I ... , v,,} be an or
s1v1 +
111 - 511 2
(x1
- sd +
(xk - sd + x+i
s;
x; for 1
::;
::;
337
Theorem 3
JR.11 Then
Algorithm l
338
Chapter 7
Orthonormal Bases
Remarks
1. It is an important feature of the construction that ;_ 1 is a subspace of the
next;.
3. Notice that the order of the vectors in the original basis has an effect on the
calculations because each step takes the perpendicular part of the next vector. If
the original vectors were given in a different order, the procedure might produce
a different orthonormal basis.
4. Observe that the procedure does not actually require that we start with a basis
; only a spanning set is required. The procedure will actually detect a linearly
dependent vector by returning the zero vector when we take the perpendicular
part. This is demonstrated in Example 5.
EXAMPLE4
Use the Gram-Schmidt Procedure to find an orthonormal basis for the subspace of JR5
defined by
-1
Span
1
2
Solution: Call the vectors in the basis w1, Wz, and w3, respectively.
1
1
First step: Let v1
w1
0 and 1
Span{Vi}.
perp51 Wz
0.)
As mentioned above, we can take any scalar multiple of perp51 w2, so we take
-3
3
v2
2 and S2
-1
Span{v1, vz}.
Section 7.2
EXAMPLE4
339
(continued)
-1/4
-3/4
1/2
1/4
3/4
(Again, it is wise to check that perps2 w3 is orthogonal to V1 and Vz.)
We now see that {v1, Vz, V3} is an orthogonal basis for. To obtain an orthonormal
basis for , we divide each vector in this basis by its length. Thus, we find that an
orthonormal basis for is
1
2
EXAMPLES
'
1
1
-1
-3
3
-3
1
1
2
2
2.../6 -1 2.../6 1
1
3
'
Use the Gram-Schmidt Procedure to find an orthogonal basis for the subspace
S
Span
ml m [i]}
3
ofR .
Solution: Call the vectors in the spanning set w1, w2, and w3, respectively.
W1
and S1
Span(Vi}.
We take v2
-2
and S2
Span{V1, v2}.
Span{v1, v2}
340
Orthonormal Bases
Chapter 7
PROBLEMS 7 .2
Practice Problems
Al Each of the following sets is orthogonal but not
2
3
orthonormal. Determine the projection of x = 5
{ =!
u
{ }
{ : = - }
Cb ) '13 =
( c) C =
( c)
(d )
1 ' 0 ' -1
(c) s =Span
each set.
(b ) S=Span
{[][i] [;]}
{[;].[i][]}
P -I - 1 }
{ - }
( b)
(a
{U]}
{ [ ; ]fi] }
{! . -I}
(a)
( b)
( c)
{[i][;].[] }
{[J.[:J.[=:J}
{ }
0
'
each set.
(d)
0
1
' -1
-1
-1
1
JPi.11
Prove that perps
AS Let be a subspace of
JPi.11
E
x
x = projgJ. x for any
Homework Problems
Bl Each of the following sets is orthogonal but not
orthonormal. Determine the projection of x = _
onto the subspace spanned by each set.
4
3
2
5
(a) .'.ll=
{ j _;}
1
-2
(b) =
(c) C =
p -; )
{I - )
{ : ' - ' _: )
,
(c)
(d)
(d) v
-1
{[]}
{[j] fi]}
{[J .mrm
) S=Span
(c) s =Span
(d) S=Span
{ I -n
-
(b)
{[l-l= :J.lm
{l:HiJ. Ul}
Conceptual Problems
11
Dl Prove that if is a k-dimensional subspace of JR. ,
11
D2 Prove that if is a k-dimensional subspace of JR. ,
k)-dimensional subspace.
thenJ_ is an (n
then n J_ = {0}.
-
'
(b)
{[J-l=:Hm
{[ll Ul lJ ll}
u : 'n
(c)
-1
(d)
0
l
7.
(a)
{-! ; ' n
P. . -n
341
BS Ut S=Span
0
2
{ 1 !)
and let X =
_: .
D4 Prove that if
and
then
342
Chapter 7
Orthonormal Bases
Prove that
Span{v1, ...,vk}
k 1,vk + t1v1
Spanv
{ 1, ... ,v-
be written
+ ... + tk-v
l k-d
D6 Suppose that is a k-dimensional subspace of JRll
and let 13 = v
{ 1, ... ,v}
k be an orthogonal basis of
. For any x E , find the coordinates of proj8 x
0
Figure 7 .3.3
343
One approach to finding the best-fitting curve might be to try to minimize the
ll
i=I
total error by having large positive errors cancelled by large negative errors. Thus, we
instead choose to minimize the sum of the squares of the errors
n
i=l
i=I
ues t1, . . .
, tn
and
y , ... , y11,
1
a, b,
and
Let
y=
r ;1 , l =
1,i'
1:
, and
i" =
'[
tn
t11
distance from y to (al+ bi'+ ci'2). Observe that the square of this distance is exactly
the sum of the squares of the errors
1
ll
= {l, t: i'2}.
t; are distinct,
Problem D2), so it is a basis for . Thus, the problem of finding the curve of best fit is
reduced to finding
a, b,
and
to
a, b,
and
Given what we know so far, it might seem that we should proceed by transforming
13 into an orthonormal basis for so that we can find the projection. However, we can
a, b, and c
y-al-bi'-ci'2 is equal to perps y.
vector in , so it is orthogonal to l, t: and
i'2.
Therefore,
-+
1 e-+ = -+
1 (y
a -+1
bt-+-ct72 )
-t - a -+
-+ = t (y
1 - bt-+- ct72 ) = 0
t-+ e
-+
...
t2 e= t2 (y-al-bi'-ci'2) = o
The required
a, b,
and
c are
x=
[1
i'
t2]
jl-Xii.
Since the three equations are obtained by taking dot products of e with the columns of
X,
XT(y-Xil) = 0
344
Chapter 7
Orthonormal Bases
The equations in this form are called the normal equations for the least squares fit.
Since the columns of X are linearly independent, the matrix xrX is a 3 x 3 invertible
matrix (see Problem D2), and the normal equations can be rewritten as
a= (XTx)-lXTy
This is consistent with a unique solution.
For a more general situation, we use a similar construction. The matrix
X,
called
the design matrix, depends on the desired model curve and the way the data are col
lected. This will be demonstrated in Example 2 below.
EXAMPLE 1
l.O
2.1
3.1
4.0
4.9
6.0
6.l
12.6
21.1
30.2
40.9
55.5
Solution: As in the earlier discussion, the experimenter wants to find the curve
= a+ bt + ct2 that best fits the data, in the sense of minimizing the sum of the squares
XT =
[f
6.1
(2r = 10 2\
1.0
4.41
3.1
4.0
4.9
6.0
9.61
16.0
24.01
36.0
12.6
21.1
y=
30.2
40.9
55.5
Using a computer, we can find that the solution for the system
[ ]
a = (XTx)-lXTy is
l.63175
a = 3.38382 .
0.93608
the best-fitting quadratic curve to bey = 1.63 + 3.38t + 0.94t . The results are shown
in Figure 7.3.4.
0
Figure 7.3.4
EXAMPLE2
345
a and b to obtain the best-fitting equation of the form y at2 +bt for the following
Find
data:
-1
4
0
1
Solution: Using the method above, we observe that we want the error vector
e
y - at2 - bt
t2 and t. Therefore,
=
must be orthogonal to
t2 . e = t2 . (y at2 bi') = 0
t e t (y at2 bi') = 0
-
Taking jl =
So, y =
then gives
Overdetermined Systems
The problem of finding the best-fitting curve can be viewed as a special case of the
problem of "solving" an overdetermined system. Suppose that Ax
b is
a system
variables, we expect the system to be inconsistent unless b has some special properties.
If the system is inconsistent, we say that the system is overdetermined-that is, there
are too many equations to be satisfied.
Note that the problem in Example 1 of finding the best-fitting quadratic curve was
of this form: we needed to solve Xa
were
n >
a,
bll.
b,
However, Ax = x1a1
+ xqaq,
which is a
we know that this vector is projs x. Thus, to find a vector x that minimizes the "error"
346
Chapter 7
Orthonormal Bases
EXAMPLE3
llAx - bll:
Ax = b
3x1 - X2
x1 + 2x2
2x1
Solution: Write the augmented matrix
x2
40
1
[ 1 -1 40 l [ 01 1 0 l
11 00
2
-2
-5
AT Ax = AT b.
Solving for
1
83
-3
87 /83
- -56/83
So,
[ ]
-
llAx - bll.
347
PROBLEMS 7 .3
Practice Problems
Al Find a and b to obtain the best-fitting equation of
the form y = a+ bt for the given data. Make a graph
showing the data and the best-fitting line.
(a)
(b)
-2
-1
2
(b) y =a+ bt for the data
t -2 -1 0 1
1 2 3 -2
Ax
are in
(a)
-1
1
(b)
llAx - bll.
X1 + 2Xz
2x1 - 3x2
X1 - 12x2
-4
2x1 + 3x2
-4
3x1 - 2x2
4
7
Xj - 6xz
Homework Problems
Bl Find a and b to obtain the best-fitting equation of
the form y =a+ bt for the given data. Make a graph
2
(a) y = at + bt for the data
-2
-]
-1
-1
t -2 - I
Ax
1 2
0
_
1 2
are in
vector
-2
-1
Xj - Xz
(a)
2
2
(b)
3x1 + 2x2
X1 - 6x2
10
X1 + Xz
X1 - Xz
4
14
Xj + 3x2
Computer Problems
Cl Find a, b, and c to obtain the best-fitting equation
2
of the form y = a+ bt + ct for the following data.
0.0
1.1
1.9
3.0
4.1
5.2
4.0
3.6
4.1
5.6
7.9
11.8
348
Chapter 7
Orthonormal Bases
Conceptual Problems
Dl Let X =
[f
.where
(=
l::J
and =
[:;]
D2 Let X = l
L t;
n
xrx=
i=l
II
Lt;
II
Lt
i=l
i=l
Lt2
Lt3
I!
i=l
i=l
ll
Lt2
i=l
I!
Lt3
I
ll
Lti
i=l
[]
and
f11
('n],
i' t2
t1
where t= :
[i =
[' ]
tll
i=l
-+
-+
-+
xr Xv= Ois
Definition
Inner Product
such that
-4
JR
Remark
Every non-trivial finite-dimensional vector space V has in fact infinitely many different
inner products. When we talk about an inner product space, we mean the vector space
and one particular inner product.
EXAMPLE 1
EXAMPLE2
349
0. Thus, it is positive
definite.
2
3. For any x,y,ZE lR. ands, t E JR,
(x, sw + tZ) = 2x1(sw1 + tz1) + 3x2(sw2 + tz2)
Remark
Although there are infinitely many inner products on IRn, it can be proven that for any
inner product on JR" there exists an orthonormal basis such that the inner product is
just the dot product on IR" with respect to this basis. See Problem Dl.
EXAMPLE3
Verify that (p, q) = p(O)q(O) + p(l)q(l) + p(2)q(2) defines an inner product on the
Solution: We first verify that ( , ) satisfies the three properties of an inner product:
2
2
2
(1) (p, p) = (p(0)) + (p(1)) + (p(2)) 0 for all p E P2. Moreover, (p, p) = 0 if
and only if p(O) = p(l) = p(2) = 0, and the only p E P2 that is zero for three
values of x is the zero polynomial, p(x) = 0. Thus ( , ) is positive definite.
(2) (p, q)= p(O)q(O)+p(l)q(l)+p(2)q(2) = q(O)p(O)+q(l)p(l)+q(2)p(2) = (q, p).
So, ( , ) is symmetric.
IR,
2
2
(1 + x + x , 2 - 3x )= (1 + 0 + 0)(2 - 0 ) + (1 + 1 + 1)(2 - 3) + (1 + 2 + 4)(2 - 12)
= 2 - 3 - 70 = -71
350
Chapter 7
EXAMPLE4
Orthonormal Bases
Let tr(C) represent the trace of a matrix (the sum of the diagonal entries). Then, M(2, 2)
is an inner product space under the inner product defined by (A, B) = tr(Br A). If
A =
[ ].
\[ -J. [ ]) ([ ][ -D
([ !])
and B =
=tr
= tr
EXERCISE 1
= 4 + 4= 8
Verify that (A, B)= tr(BT A) is an inner product for M(2, 2). Do you notice a relation
ship between this inner product and the dot product on JR.4?
Since these properties of the inner product mimic the properties of the dot product,
it makes sense to define the norm or length of a vector and the distance between vectors
in terms of the inner product.
Definition
Let V be an inner product space. Then, for any v EV, we define the norm (or length)
form
of v to be
Distance
llvll=
-J(v, v)
Definition
Unit Vector
EXAMPLES
[ ]
Solution: We have
llAll =
EXAMPLE6
351
p(O)q(O)
+ p(l)q(l) + p(2)q(2).
Solution: We have
Ill - 2x - x211 =
=
=
EXERCISE 2
EXERCISE 3
(p(0))2 + (p(1))2
(p(2))2
--)12 + (1 - 2 - 1)2 + (1
4 - 4)2
Ys4
(p, q)
--j(p, p)
In Sections 7.1 and 7 .2 we saw that the concept of orthogonality is very useful.
Hence, we extend this concept to general inner product spaces.
Definition
Let Vbe an inner product space with inner product ( , ). Then two vectors
Orthogonal
Orthonormal
be orthogonal if
have
(v;, v;)= 1
(v, w) = 0.
{v1, ... , vd
v, w
E V
in Vis said to
for all i.
With this definition, we can now repeat our arguments from Sections 7.1 and 7.2
for coordinates with respect to an orthonormal basis and projections. In particular, we
get that if '13 =
{w1,
, w11} for an inner
{v1, ... , v,i} defined by
VJ= VJ
is an orthogonal basis fo r V.
352
Chapter 7
Orthonormal Bases
EXAMPLE 7
Hence, we have
. x2 =
proJs
(x2,1)
(x2, x- 1)
(x - 1)
llx-1112
!ilii2
0(1)+1(1)+4(1)
0(-1)+1(0)+4(1)
1+
(x - 1)
1 2+12+12
(-1 )2 +02+12
1
5
= -1+2(x - 1) = 2x - 3
3
1+
PROBLEMS 7 .4
Practice Problems
Al On P , define the inner product (p,q) = p(O)q(O)+
2
p(1)q(1)+p(2)q(2) . Calculate the following.
(a) (x- 2x 2, 1+3x)
(b) (2 - x+3x2,4 - 3x2)
(d) 119 + 9x + 9x211
(c) 113 - 2x + x211
A2 In each of the following cases, determine whether
( , ) defines an inner product on P
2
(a) (p,q) = p(O)q(O)+p(l)q(l)
(b) (p, q) = lp(O)q(O)I + lp(l)q(l)I +lp(2)q(2)1
(c) (p, q) = p(- l )q(-1) + 2 p(O)q(O)+p(l)q(l)
(d) (p,q) = p(-l)q(l)+2p(O)q(O)+p(l)q(-1)
A3 On M(2, 2) define the inner product (A, B)
tr(BT A).
(i) Use the Gram-Schmidt Procedure to determine
an orthonormal basis for the following sub
spaces of M(2, 2).
.[ ]
{[ ],[ ],[ -]}
{[ i],[ -],[:1 ]}
.
to d eterrrune
proJs
ca) s=span
(b) =Span
4
_2
3
.
1
2X1Y1 + X2Y2 +
Span
{[i] .[-i][i]}
with re
353
Homework Problems
Bl On P , define the inner product (p,q) p(O)q(O) +
2
p(1)q(l) + p(2)q(2). Calculate the following.
(a) (1 -3x2,1+x+2x2)
(b) (3-x,-2 -x-x2)
(d) 1173x + 73x211
(c) Ill - Sx + 2x211
=
}
{
Span { 1 + x2,x - x2 }
Span l,x - x2
X1Y1 + 3x2 y2 +
S Span
p(-l)q(-1)
with re
Conceptual Problems
Dl (a) Let {e1, e } be the standard basis forlR.2 and sup
2
pose that ( ,) is an inner product on IR.2 Show
that if 1,y E JR.2,
(1, st>
(1,y)
I g;JXiYj
i,j=l
17 Gy
.Y2v2,
Conclusion. For an arbitrary inner product ( ,) on
JR.2, there exists a basis for JR.2 that is orthonormal
with respect to this inner product. Moreover, when
1 and y are expressed in terms of this basis, (1, Y>
looks just like the standard inner product in JR.2
354
Chapter 7
Orthonormal Bases
[i']8
[:;]
and
[jl]8
:l
then
G by
G).
(p , q )
7 .5 Fourier Series
b
J f(x)g(x) dx
product as follows.
The inner product ( , ) is defined on
C[a, b] by
b
(j, g)
l j x g x dx
( ) ( )
(1) (f, f)
(2)
(f, g)
(3)
Since an integral is the limit of sums, this inner product defined as the integral of
the product of the values off and g at each x is a fairly natural generalization of the
dot product in IR11 defined as a sum of the product of the i-th c omp onents ofx and y for
each i.
One interesting consequence is that the norm of a function f with respect to this
inner product is
11!11
) 1/2
j2(x) dx
355
Intuitively, this is quite satisfactory as a measure of how far the function is from the
zero function.
One of the most interesting and important applications of this inner product in
volves Fourier series.
Fourier Series
Let CP2rr denote the space of continuous real-valued functions of a real variable that
are periodic with period 2n. Such functions satisfy f(x+ 2n)
of such functions are f(x)
c,
that the function cos 2x is periodic with period 2n because cos(2(x + 2n))
cos 2x.
{:7!"
if
7!"-
7r/2
< x::;
7r/2
7r
7r
2
Figure 7.5.5
In the early nineteenth century, while studying the problem of the conduction of
heat, Fourier had the brilliant idea of trying to represent an arbitrary function in CP2rr
as a linear combination of the set of functions
{ 1, cos x, sin x, cos 2x, sin 2x, ... , cos nx, sin nx, ...}
This idea developed into Fourier analysis, which is now one of the essential tools in
quantum physics, communication engineering, and many other areas.
We formulate the questions and ideas as follows. (The proofs of the statements are
discussed below.)
(i) For any n, the set of functions { l, cos x, sin x, cos 2x, sin 2x, . .., cos nx, sin nx}
is an orthogonal set with respect to the inner product
(f, g)
1:
f(x)g(x) dx
The set is therefore an orthogonal basis for the subspace of CP2rr that it spans.
This subspace will be denoted CP2rr,n
356
Chapter 7
Orthonormal Bases
(ii) Given an arbitrary function f in CP2rr, how well can it be approximated by a
function in CP2rr,n? We expect from our experience with distance and subspaces
that the closest approximation to f in CP2rr,n is projcP:z..nf. The coefficients
for Fourier's representation off by a linear combination of { 1, cos x, sin x, ... ,
cosnx, sin nx, ...}, called Fourier coefficients, are found by considering this
projection.
(iii) We hope that the approximation improves as n gets larger. Since the distance
fromf to the n-th approximation prokP:z....f is II perpcP:z..n /II, to test if the ap
proximation improves, we must examine whether II perpCP:z.,n /II-+ 0 as n-+ oo.
Let us consider these statements in more detail.
(i) The orthogonality of constants, sines, and cosines with respect to the inner
rr
product (/, g) Jf(x)g(x) dx
-rr
=
rr
sinnx dx
rr
rr
cosnx dx
rr
1:
cos mx sinnx dx
and for m* n,
l
l
rr
cos mx cos nx dx
rr
rr
rr
sin mx sinnx dx
rr
1
- - cosnx
0
-rr
n
rr
1
- sinnx
0
-rr
n
1: (
l (
l 2(
rrl
- cos(m +n)x +cos(m -n)x) dx
-rr 2
rrl
cos(m -n)x - cos(m +n)x) dx
-rr
Hence, the set { 1, cos x, sin x, ..., cosnx, sinnx} is orthogonal. To use this as a basis
for projection arguments, it is necessary to calculate 111112, II cos mxll2, and II sin mxll2:
11111
II cos mxll2
II sin mxll2
1: 1
rr
I
l
rr
rr
rr
dx
2n
cos2 mx dx
sin2 mx dx
1 -(1
1
l 2(1
rr
-rr 2
rr 1
-rr
+cos 2mx) dx
- cos 2mx) dx
= 7r
= 7r
(ii) The Fourier coefficients off as coordinates of a projection with respect to the
orthogonal basis for CP2rr,n
The procedure for finding the closest approximation proj CP:z. nf in CP2rr,n to an
.
arbitrary function f in CP2rr is parallel to the procedure in Sections 7.2 and 7.4.
That is, we use the projection formula, given an orthogonal basis {v1, ... 'vd for a
subspace S:
(1, vk)
(1, v1)
.
vk
proJs 1 1iJJi2 v 1 +
+
llvkll2
=
357
ao
The factor in the coefficient of 1 appears here because 111112 is equal to 2rr, while the
other basis vectors have length squared equal to rr. Thus, we have
(f, 1)
ao = liij2 = ;
{rr_
J rr
f(x) dx
am =
_
rr
I-rr
rr .
I-rr f( )
as
---+
oo?
As n ---+ oo, the sum becomes an infinite series called the Fourier series for f. The
question being asked is a question about the convergence of series-and in fact, about
series of functions. Such questions are raised in calculus (or analysis) and are beyond
the scope of this book. (The short answer is "yes, the series converges to f provided
that f is continuous." The problem becomes more complicated if f is allowed to be
piecewise continuous.) Questions about convergence are important in physical and
engineering applications.
EXAMPLE 1
Determine prokPini f for the function f(x) defined by f(x) = lxl if -rr
f(x +2rr) = f(x) for all x.
Solution: We have
rr lxl dx
I-rr
1 rr
I-rr
rr lxl 2x dx
I-rr
1 rr
I-rrrr lxl 3x dx
1
I-rr lxl x dx
1 rr
I lxl
rr 3x dx
I
l
ao = a,= rr
7r
cos
a3 = -
cos
= --
bi = -
sin
7r
7r
7r
9rr
=0
sin2xdx = 0
- rr
b3 = Hence, prokPin.i f =
Figure 7.5.6.
rr and
a1= -
7r
lxlcos xdx = --
7r
= rr
7r
b1 = -
lxl sin
= 0
-rr
- ; cos x -
358
Chapter 7
Orthonormal Bases
y
-y=f(x)
--
-Tr
Figure 7 .5.6
EXAMPLE2
= prokP:u,.1 f(x)
( )
projCP:u,,3 fx
Y
Y
7r
{-Tr -
-Tr
f(x) = x
7r
Solution: We have
ao
a1
a1
a3
=
=
bi=
b2 =
b3 =
Hence, projcP:i,,.J f= ; sinx
7r
1
7r
1
7r
1
7r
1
7r
1
7r
1
rr
-
lf
Jlf
Jlf
Jlf
Jlf
Jlf
Jlf
J
if
7r x
/2
fdx = 0
-;r
fcosxdx = 0
-;r
fcos 2xdx = 0
-;r
fcos 3xdx = 0
-;r
-;r
fsinxdx =
7r
fsin 2xdx = 0
-;r
-;r
fsin 3xdx = -
9rr
Chapter Review
359
1r
--
-
Figure 7.5.7
Graphs of prokPi..i
y=f(x)
Y = projCPin.1 f(x)
Y projCPin,J f(x)
=
PROBLEMS 7 .5
Computer Problems
Cl Use a computer to calculate projCPin,,, f for
n =3,7, and 11 for each of the following functions.
Graph the function f and each of the projections
(b)
f(x) =eX,
(c)
f(x) =
f(x) =x2,
-;r
io
1
if
ifO
:5
7r
n :5
<
:5
7r
7r
PROBLEMS 7 5
.
360
Chapter 7
Orthonormal Bases
Chapter Quiz
El
{ t -> -i}
{ }
{}, -}
decide.
(a)
Cb)
(c)
E2
_l_
Y3
' Vs
E4
1
-2
'B
1
2
O,
0
1
0
-1
Span
1
1
, Y3 -1 , Y3 -1
1
0
{! H
ES
1
-2
.
1
0
2
5
1 is
(b) (A,B)
det(AB)
a11b11+2a12b12+2a21b21 + a22b22
-2
Further Problems
Fl
(lsometries of JR.3)
llL(x)ll
v and
orthonormal
[u
basis
for
pT AP
JR.3
1
021
and
012
A*
w}
let
is an
Chapter Review
(e) Use
Problem
3.F5
to
analyze
the
A*
of
S1
361
S2
Si
v. Prove that
i increases in the
sense that
F3 The sum S
MyMathlab
llv
proj.,
U!J+ I
vii
:=:;
llv
proj .,
1J11
vii
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 8
8.3
8.4
Symmetric matrices and quadratic forms arise naturally in many physical applica
tions. For example, the strain matrix describing the deformation of a solid and the
inertia tensor of a rotating body are symmetric (Section 8.4). We have also seen that
the matrix of a projection is symmetric since a real inner product is symmetric. We
now use our work with diagonalization and inner products to explore the theory of
symmetric matrices and quadratic forms.
8.1 Diagonalization of
Symmetric Matrices
Definition
Symmetric :\latrix
EXAMPLE 1
[ -n
that diagonalizes A?
Solution: We have
C(A) = det(A
Al) =
0-A
l
1
_
= A + 2A
Using the quadratic formula, we find that the roots of the characteristic polynomial are
A1 = -1
V2 and A2
-1
V2
-1 - V2
364
Chapter 8
EXAMPLE 1
For /l1 = -1 +
Y2,
we have
(continued)
A-/l1/=
-1 - Y2,
1 +
/l2/ =
-1 -
1
1
]
[
Y2
-
Y2
{ [ 1 4}.
1 +
- Y2
l
we have
Y2
1
1
-1 +
{[ 4}
[1 1 Y2
1-
-1 +
Y2l
J
.
1 -
Hence, A 1s
. d"iagonalized by p =
] [1
_
Y2
Y2lJ.
[ 1 [ -1 1
1+
= (1 +
= 0
thonormal columns.)
EXAMPLE2
Diagonali'e the symmetric A =
[ : -n
orthogonal matrix.
Solution: We have
0
1 - /l
-2
-2
1 - /l
4 - /l
C(/l) =
The eigenvalues are /l1 = 4, tl2 = 3, and tl3 = -1, each with algebraic multiplicity 1.
For /l1 = 4, we get
[ -] [ 1 !I
{ []}
0
A -Ail=
-3
-2
-3
[ -]- [ 1 !]
0
A-A2l =
-2
-2
-2
Section 8.1
EXAMPLE2
(continued)
{[-:]}
365
-1, we get
{[:]}
[H [-:l
1 [4 OJ
[- -2]
V2 =
and Vl =
[:]
[l
1/-,./2
1/../2
1/-,./2
1/../2
EXAMPLE3
to D
0 .
-1
-4
5
-2
an orthogonal matrix.
Solution: We have
5
C(A) =
-A
-2
-4
-2
-4
5-A
-2
-2
-2
8 -A
2
= -A(A - 9)
1.
For At = 9, we get
{[-il r]}
. However, observe
that these vectors are not orthogonal to each other. Since we require an orthonormal
basis of eigenvectors of A, we need to find an orthonormal basis for the eigenspace of
At. We can do this by applying the Gram-Schmidt Procedure to this set.
Pick Vt =Wt. Then St = Span{Vi} and
366
Chapter 8
EXAMPLE3
(continued)
Then,
-4 -21-2 [1
-2
{[]}
For
[i13) =
Observe that
find that
0
0
to D =
[ l
0
9
0
Definition
A matrix
Orthogonally
matrix
Diagonalizable
Pr AP= D
Remark
Two matrices A and Bare said to be orthogonally similar if there exists an orthogonal
matrix
Lemma 1
An n x n matrix
Section 8.1
x, y
367
JR!!,
Conversely, if x
lltn,
In Examples 1-3, we saw that the basis vectors of distinct eigenspaces are or
thogonal. This implies that every eigenvector in the eigenspace of one eigenvalue is
orthogonal to every eigenvector in the eigenspace of a different eigenvalue. We now
use Lemma 1 to prove that is always true.
Theorem 2
If
v1
and
corresponding to distinct
eigenvalues
Av1 = tl1v1
and
Av2 = tl2v2.
Hence,
Hence,
It follows that
Note that this theorem applies only to eigenvectors that correspond to different
eigenvalues. As we saw in Example 3, eigenvectors that correspond to the same eigen
value do not need to be orthogonal. Thus, as in Example 3, if an eigenvalue has al
gebraic multiplicity greater than 1, it may be necessary to apply the Gram-Schmidt
Procedure to find an orthogonal basis for its eigenspace.
Theorem 3
If
A is a symmetric matrix,
A are real.
The proof of this theorem requires properties of complex numbers and hence is
postponed until Chapter 9. See Problem 9.5.D6.
We can now prove that every symmetric matrix is orthogonally diagonalizable. We
begin with a lemma that will be the main step in the proof by induction in the theorem.
368
Chapter 8
Lemma4
where
A1 is an (n -
On-I,!
{vi}
n
to a basis for IR. , applying the Gram-Schmidt
P= [v1 w2
wn]
Then
PTAP=
VTl
->T
W
2
A [vi w2
wn]
->T,,
w
=
VTl
->T
W
2
[Av1 Aw2
Awn]
->T
w
n
=
vf Av1 vf
-> Aw2
wIAv1 W2TAW_,2
->
W2
wAv1 W11TA_,
i11 Av1 v1 Aw2
w2 Av1 w2 Aw2
Also,
->TA_,
w
n Wn
w11
First observe that
v->f Awn
T
Wn
W2 A_,
Av1
w,, A w2
vi Awn
w2 Awn
WnAWn
So, all other entries in the first column are 0. Hence, all other entries in the first row
1) x (n - 1) block
A1 is also
Theorem 5
369
hence is orthogonally diagonalizable with P = [l]. Now suppose the result is true
for (n - 1) x (n - 1) symmetric matrices, and consider an n x n symmetric matrix
A. Pick an eigenvalue tli of A (note that tli is real by Theorem 3) and find a corre
sponding unit eigenvector vi Then, by Lemma 4, there exists an orthogonal matrix
R= vi w2
wn such that
RTAR=
tli
011-i,i
PfA1P1=Di
where D i is an (n - 1) x (n - 1) diagonal matrix. Define
Oi,n-i
Pi
The columns of P2 form an orthonormal basis for Rn. Hence, P2 is orthogonal. Since
tli
tli
- 011-i,i
Oi,11-i
=D
Di
EXERCISE 1
Orthogonally diagonalize A=
EXERCISE2
Orthogonally diagonalize A=
Oi,11-i
p2
Ai
011-i,i
[- -n
[-
-1
-1
=2
370
Chapter 8
Remarks
1. The eigenvectors in an orthogonal matrix that diagonalizes a symmetric matrix
A are called the principal axes for A. We will see why this definition makes
sense in Section 8.3.
2. The converse of the Principal Axis Theorem is also true. That is, every or
thogonally diagonalizable matrix is symmetric. You are asked to prove this in
Problem D2. Hence, we can say that a matrix is orthogonally diagonalizable if
and only if it is symmetric.
PROBLEMS 8.1
Practice Problems
Al Determine which of the following matrices are
symmetric.
(a) A=
(b) B =
[ 7]
[ ]
_
[- ]
H -ll
2
(c) C =
(d) D
=
-2
-1
0
-1
(b) A=
3
-1
[ -71
[5 )
(c) A
=
(d) A=
(e) A=
3
-3
[ i
[ - =]
[! : -]
1
0
1
-2
-2
Homework Problems
Bl Determine which of the following matrices are
symmetric.
(a) A
=
(b) B =
[ ]
[ ]
[ :1
0
(c) c
=
(d) D
=
(e)
E=
[ !]
[ J
1
3
3
Section 8.1
P
pTAP=
A=[; ]
A = [ ]
A=[ -]
A= [1 1 1
A = [-1 1 -01
(d)
D.
-2
(g)
-4
-2
-2
-2
-1
-2
-2
-2
(i)
(e)
0
A=
(0 l-1 11 -ii4
2
[
A= -2 11
[
-1
A=
1
A= H -1-5 -1_;l
(h)
371
Exercises
Computer Problems
Cl Use a computer to determine the eigenvalues and a
basis of orthonormal eigenvectors of the following
4[ .1 1.9 0.51
1.0.95 0.6 -2.0.61
0.0.0.019555 0.0.0.102555 0.0.0.129555 0.0.0.290555
0.25 0.95 0.05 0.15
symmetric matrices.
(c)
1.2
By
A B
ATA ABA A2
A
and
be
n x n
(b)
(c)
D2 Show that if
(d)
A+ B
is symmetric.
calculating
the
eigenvalues
2t
t
-2t
2 t
-t
-t
and S
of
l ),
explore how
Conceptual Problems
Dl Let
1
(-0.05), (0), (0.05), (0.1), (-0.
S (t) =
[2 +
A-1 A
D3 Prove that if
then
is orthogonally diagonalizable.
372
Chapter 8
Quadratic Forms
Consl.der the symmetr1c matn'x A
x2
x2
[
][
]
= [ 2x1x ] .
:t A
a
b/2
b/2
C
b/2
. If
c
a
- b/2
-
then
][ ]
]
xi
X2
ax1 + bx2 /2
bx1/2 + CX2
[;
b 2
b 2
1 to be the coefficient of T
X ,
by choosing (A)1
EXAMPLE 1
A=
2
-2
-1
Notice that we could have written axT + bx1x2 +ex in terms of other asymmetric
matrices. For example,
Many choices are possible. However, we agree always to choose the symmetric matrix
for two reasons. First, it gives us a unique (symmetric) matrix corresponding to a
given quadratic form. Second, the choice of the symmetric matrix A allows us to apply
the special theory available for symmetric matrices. We now use this to extend the
definition of quadratic form ton variables.
Definition
Quadratic Form
373
JR.11
JR defined by
11
Q(x)
aijXiXj
.xr Ax
i,j=l
As above, given a quadratic form Q(x) on JR.11, we can easily construct the co1Te
sponding symmetric matrix A by taking (A);; to be the coefficient of x and (A);j to be
EXAMPLE2
Q(x)
matnx A =
Q(x)
3xi
xi
Sx1x2
3
512
4x
5/2
2 .
+
2x
4x1x2
xi x3
1/2
Q(x)
2xi
4x1x3 - x
2x2x3
EXERCISE 1
6x2x4
2
0
2
0
0
-
1
3
1
0
Find the quadratic form c01Tesponding to each of the following symmetric matrices.
(a)
EXERCISE2
1 2
1/2
l /2
-../2
Find the corresponding symmetric matrix for each of the following quadratic forms.
1. Q(x)
xi - 2x,x2 - 3
2. Q(x)
2xi
3. Q(x)
xi + 2x + 3x + 4x
3x1x2 - x1x3
4x
+ x
xi
2x
3x
Definition
Diagonal Form
A quadratic form Q(x) is in diagonal form if all the coefficients ajk with j * k
are equal to 0. Equivalently, Q(x) is in diagonal form if its corresponding symmetric
matrix is diagonal.
EXAMPLE3
Q(x)
2xi - 4x1x2
374
Chapter 8
EXAMPLE4
Let A = [ ]and let Q(x) = .XTAx = 2xi 2x1x2 2. Let x = Py, where
P= [ l/Yl 1/Yll
-fi.J ' Express Q(x) terms of y= [yYi2].
-l/Yl l/
Solution: We first observe that P diagonalizes A. In particular, we have
+
[l/Yl
pTAP= l/Yl
We have
][ ][
OJ
3
QCx)=xTAx
=(PylA(Py)
=yTp TAPy
=yT
2
[ ] y
2
Y1 3Y2
+
QCx)=xTAx
=(P[x]23)TA(P[x]23)
=[x]PTAP[x]23
=[xJD[xJ23
Let [X]B
l:J
375
Then we have
. .. Yn]
Q(x)= [Y1
A1
;l2
0
0
An
LJ
as required.
EXAMPLES
Let Q(x)
[12 21]
. HT
vve have
2
1
C(-1)= -2 ;l l--1 =(-1-3)(-1+1)
The eigenvalues are
with algebraic
multiplicity 1.
For
;J,1=3, we get
A
An eigenvector for
For
- '1i/ =
;J,1 is v1 =
A2 = -1, we get
[il
An eigenvector for
;J,2 is v2 =
[-22 -22] [ l -1 ]
- -12/=
[-i l
[; ;] [ ]
[ -n
Let 1
{v 1 }.
{v2}.
=
[ i -11 ]
to
EXERCISE 3
Let Q(x)
= 4x1x2 -3x.
376
Chapter 8
EXAMPLE6
4x1x2 + x .
4x , Q1(x)
2
2x1 +
Solution: Q1(x) is positive definite since Q(x) = 3xi + 4x >0 for all x * 0.
Q2(x) is indefinite since Q(l, 1) = 1>0 and Q(-1, 1 ) = - 1 < 0 .
Q3(x) is indefinite since Q(l, 1) = 8>0 and Q(l, -2)
-2 < 0.
=
Observe that classifying Q3(X) was a little more difficult than classifying Q1(X) or
Q2(X). A general quadratic form Q(x) on JR11 would be difficult to classify by inspec
tion. The following theorem gives us an easier way to classify a quadratic form.
Theorem 2
Let Q(x)
Proof: We prove (1) and leave (2) and (3) as Problems Dl and D2.
By Theorem 1, there exists an orthogonal matrix P such that
where x
Py and /l.1,
P is orthogonal, it is
Py. Thus we have shown that
0 if and only if y
0 since x
EXAMPLE7
4x + 8x1x2 + 3x
[: l
The character
istic polynomial of A is C(/l) = /l. - 7/l - 4. Using the quadratic formula, we find that
377
EXAMPLE 7
(b)
Q(x)
-2XT - 2x1X2
2X1X3 - 2x
2x2X3 - 2x
(continued)
EXERCISE4
2
-('1 + 1) ('1 + 4).
[= = ]
The
-2
Q(x)
(b) Q(x)
(a)
4xr
l6x1x2 + 4x
3x
4x2x3
3x
EXAMPLE8
[; ]
2
'1 - 6'1 + 5.
is positive definite.
(b)A =
[- - =]
-4
-2
2
4)(,1 - 28).
-4, 2 -../7,
-2 -../7, so A is indefinite.
PROBLEMS 8.2
Practice Problems
-2
(b) A
[ _i]
[ - j]
(c) A =
1
1
-1
-1
0
the
corresponding
symmetric
matrix A.
(ii) Express
Q(x)
378
Chapter 8
xT - 3x1 x2 + x
SxT - 4x1x2 + 2x
Q(x)= -2xT + 12x1x2 + 7x
Q(x)= xT 2x1x2 + 6x1x3 + x + 6x2x3 -3.S
Q(x)= -4xT + 2x1x2 - Sx - 2x2x3 - 4x
(a) Q(x)=
(c)A=
(c)
(d)
(e)
(b)A=
l- -!]
6 7
(d)A=
-1
[ - 1
[ - -1
[ 1 =1
[=; -; !]
0
(b) Q(x)=
(e)A=
-1 -2
[ 1
(fj A=
0 0 3
-3
7
Homework Problems
Bl Determine the quadratic form corresponding to the
[! ]
(b)A=
(c)A=
(e)
(a)A=
corresponding
(c)A=
symmetric
matrixA.
(d)A=
7xT + 4x1x2 + 4
Q(x)= 2x1 + 6x1x2 + 2x
(a) Q(x)=
(b)
_
l
(b)A= -
[ 4 -14]
[ -;J
n ; -l
[ j -!l
(i) Determine
xT + 3x1x2 +
3xT -2x1x2 - 2x1X3 + Sx + 2x2x3 + 3.S
Q(x)= 2xT-4x1x2 + 6x1x3 + 2 + 6x2x3 - 3x
(c) Q(x)=
(d) Q(X)=
(e)A=
n - j]
n -: J
[=r = =!I
Computer Problems
Cl Classify each of the following quadratic forms with
the help of a computer.
0.85(xT + x + x + x) - O.l.x1.x2 +
0.6X1X3 + 0.2X1X4 + 0.2X2.X3 + 0.6x2X4 -O.lx3X4
(c) Q(x) =
(a) Q(x)=
(b) Q(x)=
Conceptual Problems
Dl Let Q(x)= .xr
Ax, whereA is a symmetric matrix.
11
(a) Verify that for any 1,Y E JR ,
<.XS>=
11
I I xiY1<ei, e1>
i=l
j=l
(x,y) = A.1.iiY1 +
where A.1, .
particular,
(1,y) = 111112 =
+ Ani;zY"n
379
I Aixf
i=I
if i= k
if i * k
and that
(1,y) = xy +
xy
380
Chapter 8
JR.2,
JR.2
k,
one or two particular graphs, it might be sensible to simply use a computer to produce
these graphs. However, by applying diagonalization to the problem of determining
these graphs, we see a very clear interpretation of eigenvectors. We also consider a
concrete useful application of a change of coordinates. Moreover, this approach to
these graphs leads to a
the form
Q(x)
classification
= JR.2
k in
useful process because it allows us to say "I really only need to understand these few
standard cases." A classification of these graphs is given later in this section.
Observe that in general it is difficult to identify the shape of the graph of
bxJX2 + ex =
axi +
xi + =
to determine how diagonalizing the quadratic form will affect the graph.
Theorem 1
Let
Q(x)
matrix
a,
b,
with det
Proof: Let A
where
= [b b J.
2
JR.2
w = [Ww2J]
_,
of eigenvectors of A. Let v
= []
and
.
. smce .-t 1s a umt vector, we must have
v
VJ = - v2 =2]
w= [
JR.2.
we must have
VJ
and
cose and
_,
sine
cos e
= 1 111 2 = vi + v
w = + [- ]
]
=[
have
sine
_,
cose
c se
- sine
sme
cose
so that det
P=
1. Hence we
This corresponds to a rotation by e. Finally, from our work in Section 8.2, we know
that this change of coordinates matrix brings
Remark
If we picked w
reflection.
= [ ]
-
ne
e
381
v2
and taking y =
[]
[l
we get
gives
That is, the new y1-axis corresponds to the vector v1 in the x1x2-plane, and the y2-axis
corresponds the vector V2 in the x1x2-plane.
We demonstrate this with two examples.
EXAMPLE 1
[ )
3
2
. . po1 ynomia
. 1 1s
.
, so the charactenst1c
0
2
C(A) = det(A -Al) = A - 3,i -4 = (1l -4)(,i + 1)
4y -y =l6
This is an equation of a hyperbola, and we can sketch the graph in the y1y2-plane.
We observe that the y1-intercepts are (2, 0) and (-2, 0), and there are no intercepts on
the Yraxis. The asymptotes of the hyperbola are determined by the equation 4y -Y =
0. By factoring, we determine that the asymptotes are lines with equations 2y1 -y2 = 0
and 2y 1 + Y2 = 0. With this information, we obtain the graph in Figure 8.3.1.
Y2
However, we want a picture
[n
For A2 = -1,
a ymptotes
Figure 8.3.1
The graph of 4y
- y
16,
Y1-axis
382
Chapter 8
EXAMPLE 1
(continued)
[-]
[ ]
-
[ l
of 3x + 4x1x2
16 . In the
[]
[]
in-
of the hyperbola4y -y
16.
equation 3x + 4x1x2
16.
Figure 8.3.2
16 <=>
4y -y =l6.
}s
[ ].
[YI] 1 [ 1] [Xi]
Y -vs
2
since pT
p-
-1
X2
0 and x1
0.
EXAMPLE2
6xt + 4x1x2 + 3x
14.
[ l
2 and ;!2
[n
383
[-H
A basis
if2 is taken to define the Y2-axis, then the original equation is equivalent to
2yt + 7y
14
-fi.).
(-fl, 0)
and
(--fl, 0)
and
In Figure 8.3.3, the ellipse is shown relative to the y1 - and y2-axes. In Fig
ure 8.3.4, the new Yt - and y2-axes determined by the eigenvectors are shown rela
tive to the standard axes, and the ellipse from Figure 8.3.3 is rotated into place. The
resulting ellipse is the graph of
6xf + 4x1 x2 + 3
14.
14 in the Y1Y2-plane.
Y2
Figure 8.3.3
Figure 8.3.4
The graph of
The graph of
2yT ?y
4x1x2
3x
14
2yT ?y
+
14.
384
Chapter 8
A.1yf
A.2y = k.
Here,
Graphs of
tl1xT tl2x
+
k>O
k=O
k<O
ellipse
point (0,0)
empty set
parallel lines
line
x=0
empty set
intersecting lines
hyperbola
y= 0
empty set
line
empty set
point (0,0)
hyperbola
parallel lines
ellipse
but it
contains first-degree terms. Since a quadratic form contains only second-degree terms,
to consider how degenerate cases arise as limiting cases of nondegenerate cases. For
example, Figure 8.3.5 shows that the case of parallel lines (y
the family of ellipses
.,
::-:
. ..
-- ......-:
-- - .
,,
..
...
'
'
....
.
-.. ...
---- --
Figure
8.3.5
A.xf + x = 1 as A. tends to 0.
.. .-::
.:-:: .....
-..,
...
'
... -.
.
"
"
_.,
./
Xt
A family of ellipses
A.xf
,.
x =
k > 0 and k = 0.
k > 0, although they may be
oriented differently. For example, the graph of xf - x = - 1 is the same as the graph of
-xf + x = 1, and this hyperbola may be obtained from the hyperbola xf - x = 1 by
reflecting over the line x1 = x2 However, for purposes of illustration, it is convenient
Table 8.3. 1 could have been constructed using only the cases
k<
385
to include both k > 0 and k < 0. Figure 8.3.6 shows that the case of intersecting lines
(k
Figure 8.3.6
EXERCISE 1
Graphs of
xf 2x
-
k for
>
0) from
k E (-1, 0, 1}.
Diagonalize the quadratic form and sketch the graph of the equation
Show both the original axes and the new axes.
Q(x)
xi+2x1x2+x 2.
=
we did above. However, because there are three variables instead of two, there are more
possibilities. The nondegenerate cases give ellipsoids, hyperboloids of one sheet, and
hyperboloids of two sheets. These graphs are called quadric surfaces.
The usual standard form for the equation of an ellipsoid is
This is the case obtained by diagonalizing
Q(x)
x2
x2
x2
1.
k if the eigenvalues and k are all
c;
An ellipsoid is shown in Figure 8.3.7. The positive intercepts on the coordinate axes
are
Figure 8.3.7
386
Chapter 8
(a)
(b)
(b) k
(c)
Graphs of 4x + 4x
Figure 8.3.8
2
x
2
x
a + b
x2
Q are positive
and the third eigenvalue is negative. It is also obtained when k and two eigenvalues are
negative and the other eigenvalue is positive. Notice that if this is rewritten
i,
+ =
1it is clear that for every z there are values of x1 and x2 that satisfy the equation,
so that the surface is all one piece (or one sheet). A hyperboloid of one sheet is shown
if Figure 8.3.8 (a).
values are positive, or when k and one eigenvalue are positive and the other eigenvalues
are negative. Notice that if this is rewritten
lx3I <c, there are no values of x1 and x2 that satisfy the equation. Therefore, the graph
consists of two pieces (or two sheets), one with x3 c and the other with x3 :'.S: -c. A
hyperboloid of two sheets is shown in Figure 8.3.8 (c).
It is interesting to consider the family of surfaces obtained by varying k in the
x2 x2 x2
equation
b;
of one sheet; as k decreases toward 0, the "waist" of the hyperboloid shrinks until at
k = 0 it has "pinched in" to a single point and the hyperboloid of one sheet becomes
a cone. As k decreases towards -1, the waist has disappeared, and the graph is now a
hyperboloid of two sheets.
Table 8.3.2
Graphs of
-11xT '12x
+
-13
k>O
k=O
A1,A2,A3>0
ellipsoid
point (0, 0, 0)
elliptic cylinder
X3-axis
cone
hyperbolic cylinder
intersecting planes
cone
empty set
x1-axis
.11,.12,A3<0
empty set
point (0, 0, 0)
Section 8.3
Exercises
387
cases are the ellipsoids and hyperboloids. Note that the hyperboloid of two sheets
(a)
Figure 8.3.9
2
x
x
x
-t
- - --% = k, k > 0.
c
a
b
z
(b)
(c)
Figure 8.3.9 shows some degenerate quadric surfaces. Note that paraboloidal sur
faces do not appear as graphs of the form Q(x) = k in R3 for the same reason that
2
parabolas do not appear in Table 8.3.1 for R : their equations contain first-degree
terms.
PROBLEMS 8.3
Practice Problems
Al Sketch the graph of the equation 2xi + 4x1x2 - x =
6. Show both the original axes and the new axes.
A2 Sketch the graph of the equation 2xi + 6x1x2 +
new axes.
(a) A=
(b) A=
0
1
(d) A=
-2
15. Show both the original axes and the new axes.
AS For each of the following symmetric matrices, iden-
[1 ll
[ ]
[ -]
1
(c) A=
12. Show both the original axes and the new axes.
[ n
[ -]
-2
-1
-2
-2
(e) A=
-4
388
Chapter 8
Homework Problems
90. Show both the original axes and the new axes.
surface
(a)
(b)
(c)
+4x1X2+x = 8.
(d)
(e)
Q(x)
Q(x)
Q(x)
Q(x)
Q(x)
Q(x)
=x + 4x1x2 + x
x =
32. Show both the original axes and the new axes.
Computer Problems
Cl Identify the following surfaces by using a computer
(a)
Small Deformations
A small deformation of a solid body may be understood as the composition of three
stretches along the principal axes of a symmetric matrix together with a rigid rotation
of the body.
Consider a body of material that can be deformed when it is subjected to some
external forces. This might be, for example, a piece of steel under some load. Fix
an origin of coordinates
gin is left unchanged by the deformation. Suppose that a material point in the body,
which is at
1 before the forces are applied, is moved by the forces to the point f (1) =
0. The problem is to understand
this deformation f so that it can be related to the properties of the body. (Note that f
For many materials under reasonable forces, the deformation is small; this means
that the point
f(1)
1.
f(x) = 1
389
f3h(x)
3IR. IR.3
[ ; J,
(0)
is
so that
[; (0) J v.
[CO)]=
I+
f3G
where G
W=
W,
where
(G-G )
2
I+ = I+ + = (I+
f3G
f3(E
W)
Since f3 is assumed to be small, {32 is very small and may be ignored. (Such treatment
of terms like [32 can be justified by careful discussion of the limit at f3
0.)
The small deformation we started with is now described as the composition of two
linear transformations, one with matrix
be shown that
I+
f3W. It can
I+
f3W describes a small rigid rotation of the body; a rigid rotation does
I+
not alter the distance between any two points in the body. (The matrix f3W is called an
infinitesimal rotation.)
Finally, we have the linear transformation with matrix
I+
metric, so there exist principal axes such that the symmetric matrix is diagonalized to
[ 1 ]
I 1+3
Ei
E2
because
f3E
I,
f3 is small, it follows that the numbers EJ are small in magnitude, and therefore
1+
EJ
>
0. This diagonalized matrix can be written as the product of the three matrices:
l3 = [1+ o1 ol [1 1+
1+
1
0
0
fl
0
0
0
0
f2
It is now apparent that, excluding rotation, the small deformation can be represented
as the composition of three stretches along the principal axes of the matrix
quantities
E1, E2,
f3E.
The
and f3 are related to the external and internal forces in the material
by elastic properties of the material. (/3 is called the infinitesimal strain; this notation
is not quite the standard notation. This will be important if you read further about this
topic in a book on continuum mechanics.)
390
Chapter 8
(2) In general,
where
is the
instantaneous angular velocity vector. The vector lis a scalar multiple of w if and
only if
w is an eigenvector of N-that is, if and only if the axis of rotation is one of the
principal axes of the body. This is a beginning to an explanation of how the body wob
bles during rotation
(w
l is
a conserved quantity
fixed in the body (so that they rotate with the body). At any time
make certain angles with respect to the space axes; at a later time
have moved to a new position. Since (0, 0, 0) is fixed and the body is rigid, the body
axes have moved only by a rotation, and it is a fact that any rotation in
JR.3
is deter
'f*, the instantaneous rate of rotation about the axis. The instantaneous
angular velocity is defined to be the vector w = ( 'f*) u(t).
mined by its axis and an angle. Call the unit vector along this axis
-t
-t
-t
1(t)
v(t)
at
is determined by
v=wx1.)
at the point
x);
n33 .
The factor
(XT + x)
is simply the square of the distance of the mass from the x3-axis. There are similar
definitions of the moments of inertia about the
x2-axis (denoted by
n12).
For a general axis e through the origin with unit direction vector it, the moment of
inertia of the mass about e is defined to be
of
m from e.
Thus, if we let
1=
[:n
m multiplied
the 3 x 3 matrix
1 i1 = 1T i1,
m at 1,
391
(Vectors and matrices are special kinds of "tensors"; for our present purposes, we
simply treat N as a matrix.) With this definition, the moment of inertia about an axis
with unit direction u is
uT N u
It is easy to check that N is the matrix with components n11, n22, and n33 as given
above, and for i * j, niJ
xxT is a symmetric 3 x 3 matrix. (The term mx;x1 is called a product of inertia. This
name has no special meaning; the term is simply a product that appears as an entry in
the inertia tensor.)
It is easy to extend the definition of moments of inertia and the inertia tensor
to bodies that are more complicated than a single point mass. Consider a rigid body
that can be thought of as k masses joined to each other by weightless rigid rods. The
moment of inertia of the body about the x3-axis is determined by taking the moment of
inertia about the x3-axis of each mass and simply adding these moments; the moments
about the x1 - and x2-axes, and the products of the inertia are defined similarly. The
inertia tensor of this body is just the sum of the inertia tensors of the k masses; since
it is the sum of symmetric matrices, it is also symmetric. If the mass is distributed
continuously, the various moments and products of inertia are determined by definite
integrals. In any case, the inertia tensor N is still defined, and is still a symmetric
matrix.
Since N is a symmetric matrix, it can be brought into diagonal form by the Prin
cipal Axis Theorem. The diagonal entries are then the moments of inertia with respect
to the principal axes, and these are called the principal moments of inertia. Denote
these by N1, N2, and N3 Let 'P denote the orthonormal basis consisting of eigenvectors
[;:]
of N (which means these vectors are unit vectors along the principal axes). Suppose
an arbitrary axis t is determined by the unit vector ii such that [il]p
Then,
from the discussion of quadratic forms in Section 8.2, the moment of inertia about this
axis e is simply
This formula is greatly simplified because of the use of the principal axes.
It is important to get equations for rotating motion that corresponds to Newton's
equation:
The rate of change of momentum equals the applied force.
The appropriate equation is
The rate of change of angular momentum is the applied torque.
It turns out that the right way to define the angular momentum vector
body is
J'
l for
a general
N(t)w(t)
Note that in general N is a function oft since it depends on the positions at time t of
each of the masses making up the solid body. Understanding the possible motions of
392
Chapter 8
general, this is a very difficult problem, but there will often be important simplifications
if
N is diagonalized by the Principal Axis Theorem. Note that J(t) is parallel to w(t) if
w(t) is an eigenvector of N(t).
and only if
PROBLEM 8.4
Conceptual Problem
Dl Show that if P is an orthogonal matrix that di
agonalizes the symmetric matrix f3E to a matrix
with diagonal entries t:1, t:2, and t:3, then
P also
CHAPTER REVIEW
Suggestions for Student Review
1
8.2)
diagonalizing
the
corresponding
symme
8.3)
8.2)
Q(x) = _xT Ax = k?
8.3)
Chapter Quiz
El Let A =
[- ]
-
D is diagonal.
the
corresponding
symmetric
A.
Q(x) in diagonal form and give the or
(ii) Express
Q(x).
of the graph of
3 2
A=
31.
= (tl
- 3)4,
then
Chapter Review
393
Further Problems
Fl In Problem 7.F5, we saw the Q R-factorization: an
form A
QU. This is
QiR1, Ai
Ri Qi, A1
QzR 2, Az
RzQ 2, ...,
D.
(b) Let V
that A
QCQT.)
AAT,
MyMathlab
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
CHAPTER 9
Complex Numbers
9.5
9.6
:C
When they first encounter imaginary numbers, many students wonder why we look at
numbers that are not real. In fact, it was not until Rafael Bombelli showed in 1572 that
numbers involving square roots of negative numbers can be used to solve real-world
problems. Currently, complex numbers are used to solve problems in a wide variety of
areas. Some examples are electronics, control theory, quantum mechanics, and fluid
dynamics. Our goal in this chapter is to extend everything we did in Chapters 1-8 to
allow for the use of complex numbers instead of just real numbers.
1 The system of
.
numbers of the form x + yi where x, y E JR is called the complex numbers. Note that the
real numbers are included as those complex numbers with b = 0. As in the case with
all the previous extensions of our understanding of number, some people are initially
uncertain about the meaning of the "new" numbers. However, the complex numbers
have a consistent set of rules of arithmetic, and the extension to complex numbers is
justified by the fact that they allow us to solve important mathematical and physical
problems that we could not solve using only real numbers.
Complex Number
element such that i2 = -1. The set of all complex numbers is denoted by C.
396
Chapter 9
Addition
of complex. number'!.
Z.\ = X\ +
X2 + )l2i \ G.efmed b
EXAMPLE 1
EXERCISE 1
Remarks
1. Notice that for a complex number z = x + yi, we have that z = 0 if and only if
x= 0 and y = 0.
2. If z= x + yi, we say that the real part of z is x and write Re (z)= x. We say that
the
4. It is best not to use as a notation for the number i; doing so can lead to
confusion in some cases. In physics and engineering, it is common to use j in
place of i since the letter i is often used to denote electric current.
Section 9.1
Complex Numbers
397
4+l9i=-2+Si
3 - 2i
In order to give a systematic method for expressing the quotient of two complex num
bers as a complex number in standard form, it is useful to introduce the complex con
jugate.
Definition
The
Complex Conjugate
z = x - yi
EXAMPLE2
2+Si= 2 - Si
-3 - 2i = -3+2i
x = x,
Theorem l
for any x
IR
IR we have
(1) Zi = Zi
(2) z1 is purely real if and only if ZI= zi
(3) zi is purely imaginary if and only if ZI = -z
(4) zi+z2=Zl+z2
(S) ZiZ2 = ZiZ2
(6) t: = Zl1
(7) zi+ZI= 2 Re(zi) = 2x1
(8) zi - ZI = i2 Im(zi)= i2yi
(9) ziZI=+YT
EXERCISE2
ber in standard form by multiplying both the numerator and the denominator by the
complex conjugate of the denominator and simplifying. If zi = x i +Yii and z2 =
x2+Y2i * 0, then
398
Chapter 9
Notice that the quotient is defined for every pair of complex numbers z1, z2, provided
that the denominator is not zero.
2+Si
EXAMPLE3
3 -4i
--
EXERCISE 3
(2+5i)(3+4i)
(3-4i)(3+4i)
1+
l-t
(b)
l+t
(c)
(6-20)+(8+ 15)i
14 23.
=--+-t
25 25
9+ 16
4 -i
1+Si
Theorem 2
+ a1x+ ao.
where a; E JR for
1 ::; i ::;
n.
If z is a root of p(x),
0 = 0= a,,z11+
+a1z+ ao = a11zn +
+ a1z+ ao
EXAMPLE4
Solution: By the Rational Roots Theorem (or by observation), we see that x =-1 is
a root of p(x). Therefore, by the Factor Theorem, (x+ 1) is a factor of p(x). Thus,
x3 + 1 =(x+ l)(x2 - x+ 1)
Using the quadratic formula, we find that the other roots are
z=
1+ Y3i
and
z=
1 - Y3i
Section 9.1
Complex Numbers
399
z1 z2
Argand diagram.
y
imaginary
axis
(x.y)
real axis
Figure 9.1.1
(x, -y)
However, in the complex plane, we can also multiply by complex scalars. This has a
natural geometrical interpretation, which we shall see in the next section.
EXERCISE4
(c) (2+i)i
(b)2+i
Polar Form
Given a complex number z
lzl
Definition
;\1odulus
Argument
Polar Form
x2+ y2
r cos ()
and
y = r sin()
400
Chapter 9
z = r(cos()+ sin())
y
z = x+ iy = r cos()+ ir sin()
Figure 9.1.2
EXAMPLES
Polar representation of z.
Solution: We have
lzil=
-1 +
Y3i.
12 - 2il = V22+ 22 = 2 Y2
2 2 Y2 cos()
=
so cos()=
of z1 is
and
- 2 2 Y2 sin()
=
1zz1 =
Since
of z2 is
Remarks
1. An important consequence of the definition is
2
lzl = x2 + l = zZ
2.
The angles may be measured in radians or degrees. We will always use radians.
3. Notice that every complex number z has infinitely many arguments and hence
infinitely many polar forms.
Section 9.1
401
Complex Numbers
two trigonometric functions to locate the correct quadrant for z. Also note that
y/x is not defined if x
0.
EXERCISE 5
EXERCISE 6
Let z
-f3+i and z2
-1 - i.
r(cos() + i sin B). Prove that the modulus of z equals the modulus of z and an
argument of z is - B.
The polar form is particularly convenient for multiplication and division because
of the trigonometric identities
cos(B1+B2)
sin(B1+B2)
It follows that
z1z2
In words, the modulus of a product is the product of the moduli of the factors, while
an argument of a product is the sum of the arguments.
Theorem 3
z2 f. 0, we have
Corollary 4
EXERCISE 7
Let z
1
r(cos() +i sin B) with r f. 0. Then z-
( cos(-())+i sin(-B)).
402
Chapter 9
EXAMPLE6
r:;
Solution:
We have
2+2i
.
r:;
1 + v3i
,
us mg polar form.
( ( )
( G;)
2+2i
1 + -.../3i
+isin +isin
7r
+ isin 1
12
1.366+i(0.366 )
( ( )
EXERCISE 8
G;))
-0.732+i(2.732 )
( ))
= Y2 cos ;<;
2 - 2i
-1 +
Theorem 5
-:F
Proof:
Section 9.1
Complex Numbers
403
Therefore, the result is true for all non-negative integers n . Then, by Theorem 4, for
any positive integer m, we have
z
-m
()-1
= r
-m
( cos(-me) + isin(-m8))
EXAMPLE 7
Calculate
(2 + 2i)3.
Solution:
c2 +
2i)3
[2 Y2 ( G) G))r
(2 \12)3 (
( :) ( :))
cos
+ sin
cos 3
16
-
+ isin 3
( )
Y2 -
+i
16 + l6i
(e
Definition
ei!I
Euler's Formula
Definition
cose + isine
Remarks
1
1, e, i, and1r.
2. One area where Euler's Formula has important applications is ordinary differ
ential equations. There, one often uses the fact that
< +
e a bi)t
eateibt
eat (cos bt
isin bt)
404
Chapter 9
Observe that Euler's Formula allows us to write every complex number z in the
form
where r = lzl and e is any argument of z. Hence, in this form, de Moivre's Formula
becomes
EXAMPLES
3
Solution: (2 +2i)3 = 2 "2.ei1114 = (2 Y2.)3ei<3rr/4l = -1 6 + 1 6i
(b) (2i)3
cY3 + i)5
EXERCISE 9
n-th Roots
Using de Moivre's Formula for n-th powers is the key to finding n-th roots. Suppose
that we need to find then-th root of the non-zero complex number z = reil:i. That is, we
need a number w such that w'1 = z. Suppose that w =Rei. Then wn = z implies that
Then R is the realn-th root of the positive real number r. However, because arguments
of complex numbers are determined only up to the addition of 2nk, all we can say
about > is that
nr/> = 8 +2nk,
or
r/> =
EXAMPLE9
e +2nk
kEZ
kEZ
Section 9.1
Complex Numbers
405
0
If k = 0, we have the root w0 = 2e = 2.
2 /
If k= 1, we have the root w1 = 2ei rr 3 = -1 + -./3i.
/
If k = 2, we have the root w2 = 2ei 4rr 3 = -1 - -./3i.
If k = 3, we have the root 2ei2rr= 2 = wo.
EXAMPLE9
(continued)
By increasing k further, we simply repeat the roots we have already found. Simil
arly, consideration of negative k gives us no further roots. The number 8 has three third
roots, w0, w1, and w2. In particular, these are the roots of equation w3 - 8= 0.
Theorem 6
Let
EXAMPLE 10
+2rrk)/n
,
= rei6 are
k = 0, 1, ... 'n - 1
k= 0, 1,2,3
In Examples 9 and 10, we took roots of numbers that were purely real: we were
really solving x' - a= 0, where
a E
of the equations are real, the roots that are not real occur in complex conjugate pairs.
As a contrast, let us consider roots of a number that is not real.
EXAMPLE 11
Wo
W1
W2
=
=
51/3ei5JT/6
st /3
51/3
(f
( -f3
)
)
+i
+
-i 51/3
Examples 9, 10, and 11 all illustrate a general rule: the n-th roots of a complex
number
z=
rei6 all lie on the circle of radius r11", and they are separated by equal
angles of 2n In.
406
Chapter 9
PROBLEMS 9.1
Practice Problems
Al Determine the following sums or differences.
(a) (2+Si)+ (3+2i)
(b) (2-7i)+ (-S+3i)
(c) (-3+Si)-(4+3i)
(c)
3+2i
(d)
1+6i
.
4 l
-
z
i if
Z2
(b) 2+7i
(d) -4i
2- Si
Homework Problems
Bl Determine the following sums or differences.
(a) (3+4i)+ (1 +Si)
(b) (3- 2i)+ (-7+6i)
(c) (-S+7i)- (2+6i)
(d) (-7- 2i)- (-8 -9i)
B2 Express the following products in standard form.
(a) (2+ i)(S- 3i)
(1
VJi)4
(b) (8 li)1/5
2i)3
(c) ( -Y3 - i)4
(d) (-2 + 2 Y3i)5
(b) ( -2
407
(c) (-
Y3 + i)l/
3
(d) (4+i)1 1
Conceptual Problems
Dl Prove properties (3 ), (5), (6), (7), (8), and (9) of
Theorem 1.
(e;e+e-ie)
;
sine= (e e - e-ie)
(b) cos8=
(c)
D4 Prove Theorem 3.
9.2
EXAMPLE 1
(1
=i
i)z1 + z2+
(3 - i)z1 + 2z2+ z3
1 + 2i
Solution: The solution procedure is, as usual, to write the augmented matrix for the
system and row reduce the coefficient matrix to row echelon form:
3 -i
0
i
1 +2i
- l+i
1
0
0
1
0
1 +i
-2+i
1 + 2i
-1
-[
1 +i
1
1 - 2i
ll
l
R1 - (1 - i)R1
R3 - (3 - i)R1
R1 - R,
R3 + ( 1 - i)R2
R, +m,
R2 - (1 + i)R3
1 +;
-[
[
l l
-2+ i
1 - 2i
- 1+i
- 1+i
-2+i
-i
0
0
1 +i
-[
0
0
1
l
+; l
-1
0
0
-2 + i
1 - 2i
J l
-iR2
2;
-iR3
408
Chapter 9
EXAMPLE2
(1
i)z1
(1
ZI
2iz2 +
=1
1.
1
.
t)z2 + Z3 = - - -t
2
2
- Z3 = 0
[T
[
2i
0
i
0
0
1
2i
-1
-1
i
] [
][
j
1
0
0
+
2i
!;
-1
1
0
0
1
2i
-1
1-i
T
+
l
+H
21
l.
0
1
0
-1
1-i
T
0
I.
EXERCISE 1
iz1 + z2 + 3z3 = -1 - 2i
-i
2
j = -1, so that we can use i to denote current.
In Section 2.4, we discussed electrical circuits with resistors. We now also con
sider capacitors and inductors, as well as alternating current. A simple capacitor can be
thought of as two conducting plates separated by a vacuum or some dielectric. Charge
can be stored on these plates, and it is found that the voltage across a capacitor at
time tis proportional to the charge stored at that time:
V(t) =
Q(t)
where Q is the charge and the constant C is called the capacitance of the capacitor.
409
V(t)
d i(t)
dt
d i(t)
-
dt
1
+R i(t) + -Q(t)
c
E(t)
+
E
c
R
Figure 9.2.3
For our purposes, it is easier to work with the derivative of this equation and use
the fact that
dQ
dt
i:
d2 i(t)
--
dt2
+R
d i(t)
--
dt
d E(t)
-t
= --
1
c
(t)
dt
In general, the solution to such an equation will involve the superposition (sum)
of a steady-state solution and a transient solution. Here we will be looking only for the
steady-state solution, in the special case where the applied electromotive force, and
hence any current, is a single-frequency sinusoidal function. Thus, we can assume that
E(t)
BeJwt
i(t)
and
Aejwr
where A and B are complex numbers that determine the amplitudes and phases of
voltage and current, and w is 2rr multiplied by the frequency. Then
di
- =
dt
d2i
- =
dt2
.
jwAe1wr
(jw)2i
jwi
-w2i
. R.l +
+ JW
dE
-t = -
dt
410
Chapter 9
R1,
i1 - i2.)
that we have only one single frequency source, we may conclude that the steady-state
loop currents must be of the form
Figure 9.2.4
If we write the corresponding equations for the other two loops, reorganize each equa
tion, and divide out the non-zero common factor eiwr, we obtain the following system
and
Thus, we have a system of three linear equations with complex coefficients for the
three variables
and
emphasize that this example is for illustrative purposes only: we have constructed a
completely arbitrary network and provided the solution method for only part of the
problem, in a special case. A much more extensive discussion is required before a
reader will be ready to start examining realistic circuits to discover what they can do.
But even this limited example illustrates the general point that to analyze some electri
cal networks, we need to solve systems of linear equations with complex coefficients.
Section 9.3
411
PROBLEMS 9.2
Practice Problems
Al
(b)
(1
(1
z1 + iz2+ +i)z3 -i
-2z1 + - 2i)z2 - 2z3 2i
2iz1 - 2z2 - (2+3i)z3 + 3i
=
(1
z1 + + i)z2+2z3+Z4 - i
221 +(2 + i)z2 + 523 + (2+i)z4 4 -i
iz1 + + i)22+(1+2i)z3+2iZ4
(-1
-1
Homework Problems
Bl Determine whether each system is consistent, and
(c)
(1
z2 -i23 +3i
i21 - 22+(-1 + i)z3 + 2i
221 +2i22+ +2i)z3 4
21 + (2+i)22 + i23
+i
iz1 + (-1+2i)z2 +2iz4 -i
z1 +(2+i)z2+ + i)z3 + 2iz4 2 -i
(3
(b)
(3
(d)
(1
(-1
(1
4.2
real numbers. In fact, the definition makes sense when the scalars are taken from any
one system of numbers such that addition, subtraction, multiplication, and division
are defined for any pairs of numbers (excluding division by 0) and satisfy the usual
commutative, associative, and distributive rules for doing arithmetic. Thus, the vector
space axioms make sense if we allow the scalars to be the set of complex numbers. In
such cases, we say that we have a vector space over C, or a complex vector space.
EXERCISE 1
2 {[ ] Z1,Z2 c},
Le t C
+ wi]
[21]Z2 + [w']
[21
W2 Z2+W2
=
C defined by
a [zi]z2 [a2az21]
=
412
Chapter 9
It is instructive to look carefully at the ideas of basis and dimension for complex
vector spaces. We begin by considering the set of complex numbers C itself as a vector
space.
As a vector space over the complex numbers, C has a basis consisting of a single
element, {1 }. That is, every complex number can be written in the form a1, where a is a
complex number. Thus, with respect to this basis, the coordinate of the complex num
ber z is z itself. Alternatively, we could choose to use the basis {i}. Then the coordinate
of z would be -iz since
z=(-iz)i
In either case, we see that C has a basis consisting of one element, so IC is a one
dimensional complex vector space.
Another way of looking at this is to observe that when we use complex scalars, any
two non-zero elements of the space IC are linearly dependent. That is, given z1, z2 E IC,
there exist complex scalars, not both zero, such that
For example, we may take a1 = 1 and a2 = _fl, since we have assumed that z2 * 0.
z2
It follows that with respect to complex scalars, a basis for IC must have fewer than two
dimensions.
However, we could also view IC as a vector space over R Addition of complex
numbers is defined as usual, and multiplication of z = x + iy by a real scalar k gives
kz= kx + kyi
Observe that if we use real scalars, then the elements 1 and i in C are linearly in
dependent. Hence, viewed as a vector space over JR, the set of complex numbers is
two-dimensional, with "standard" basis {l , i}.
As we saw in Section 9.2, we sometimes write complex numbers in a way that
exhibits the property that IC is a two-dimensional real vector space: we write a complex
number z in the form
z = x + iy = (x,y) = x(l,0) + y(O, 1)
Note that (1, 0) denotes the complex number 1 and that (0, 1) denotes the complex
number i. With this notation, we see that the set of complex numbers is isomorphic to
2
JR. , which justifies our work with the complex plane in Section 9.1. However, notice
that this representation of the complex numbers as a real vector space does not include
multiplication by complex scalars.
2
Using arguments similar to those above, we see that IC is a two-dimensional
complex vector space, but it can be viewed as a real vector space of dimension four.
Definition
C"
Section 9.3
413
Remark
These vector spaces play an important role in much of modem mathematics.
Since the complex conjugate is so useful in C, we extend the definition of a com
plex conjugate to vectors in C'1
z1
Definition
Complex Conjugate
The
[]
:
complex conjugate of z=
en is defined to be l =
Zn
EXAMPLE 1
Let z=
1 +i
1 +i
-2i
-2i
. Then z=
3
1- 2i
[0l.
1-i
2i
3
1 +2i
1 - 2i
L :V
and W are both vector spaces over the complex numbers. We say that
the complex numbers if for any
L is linear over
EXAMPLE2
Let
L : C3
L(l,0,0) =
1+i
,
2
[ ]
L(O, 1, 0) =
[ L .]
1 -
L is
[L] =
The image of z=
-2i
i ,
[l ]
1 +i
2
-2i
1- i
1+2i
3+i
L(x) = [LJz=
1+i
2
-2i.
1-z
] [ ] = [8 8 ]
1+2.i
3+t
+2i
1-
The range of Lis the subspace of the codomain C2 spanned by the columns of
the nulls pace is the solution space of the system of linear equations Az =
that Zis a vector in C3 .)
[L], and
0. (Remember
414
Chapter 9
Similarly, the rowspace, columnspace, and nullspace of a matrix are also defined
as in the real case.
EXAMPLE3
Let A = 11.
nullspace of A.
1 +i
-i
-1 + 2i
2 +i
Solution: We row reduce and find that the reduced row echelon form of A is
R
[ 1
{ } }
0
As in the real case, a basis for Row(A) is the non-zero rows of the reduced row echelon
-1
-[
m l [j: Ul
that con-
0.
Using the
reduced row echelon form of A, we find that the homogeneous system is equivalent to
Zt
+ iz2
- Z4
Z3 - iz4
= -is+t and
z3
z2
and
z4,
so we let
Z2
Z3
Z4
Let A
1 +i
-[
nullspace of A.
=0
=s
C, and
Z4
=t
C. Then we get
EXERCISE 2
z2
= 0
-is+t
s
it
{- H
-1
1 +i
1 - 2i
-2 - 2i
1 -i
1
=s
+t
415
z=x+iy=(x,y)
Let us consider multiplication by i:
Mi : IR2
IR2
defined by
MJx, y) = (-y, x)
[M;] =
Observe that
angle of
[Md = [RI].
[o -1]
1
y
.::=(x,y)
iw = (
Figure 9.3.5
v,
u)
complex number
[a ]
Ma
of JR2 with
416
Chapter 9
PROBLEMS 9.3
Practice Problems
(b) Determine L(2 + 3i,1 - 4i).
(b)
[- ] [ ]
[ : I [ ! : I
[ ;]
+i
(c) 2i
(d) (-1
(a) A=
-3 - 4i
2 - Si
2i)
[ :I
(c) C =
[ ]
2i
;/
and
L(O, 1) =
(b) B= 1+i
2 - Si
L(l,0)=
[[
2i
-l+i
.
-1
1+ 2i
-2 +i
1+i
-1
-l+i
-2
-2 + 3i
-1 -i
[ ]
Homework Problems
[ ] [- ]
2 - i
1- l
-l
3 + 2;
(b)
l ;I
[ [
[ ;]
[I
4 - 3i
.
2 +
- 3 +
1+ 3t
1-
-2
(a)
(b)
(c) -3i
3i
-2 - 7i
_; i
mrnl' [J}
mrnHm
l
[
[ l
and
L(0,1)=
[ ;/]
1
(a) A=
l+i
l+i
-1-i
1 i
2 -i
.
-l
-l+i
2i
(b) B =
(c) C =
-1
[I Tl
i
4;
-4i
417
Conceptual Problems
Dl (a) Prove that multiplication by any complex num
ber a = a + bi can be represented as a lin
ear mapping Ma of IR2 with standard matrix
[ ]
-
tion
2
or dilation, and a rotation in the plane JR .
4.7
2 matrices with
e3
(taken as a com
L : en
en or,
Definition
Eigenvalue
Eigenvector
L : e11
e11 be a linear mapping. If for some A E e there exists a non-zero vector
z E e11 such that L(Z) AZ, then A is an eigenvalue of L and z is called an eigenvector
of L that corresponds to A. Similarly, a complex number A is an eigenvalue of an
n x n matrix A with complex entries with corresponding eigenvector z E C11, z f. 0, if
Let
Az= AZ.
Since the theory of solving systems of equations, inverting matrices, and finding
coordinates with respect to a basis is exactly the same for complex vector spaces as
the theory for real vector spaces, the basic results on diagonalization are unchanged
of A . Since the Fundamental Theorem of Algebra guarantees that every n-th degree
polynomial has exactly n roots over
over
multiplicity.
We do not often have to carry out the diagonalization procedure for complex
matrices. However, a simple example is given to illustrate the theory.
EXAMPLE I
. A
.
Determme
whether the matnx
3
1
Si
4i
2l
7i
6
+ i
. .
.
'
11zable. If 1t
1s diagona
1s,
418
Chapter 9
EXAMPLE 1
Solution: Consider
(continued)
/I.I
,t
-1
,ti)
.-t2
- -2i)tl+(3 - 3i) = 0
(1
/l = (1 -2i)
[(1
-------
Using the methods from Section 9.1, we find that the eigenvalues are
we get
/I.ii
,t
For .-t2
is
-1
IS
-1 - i]
0
-11
-2 -i]
0
-1
1
and
Hence,
that
we get
1t
/l1 = +i
and
all of the entries just happen to have zero imaginary part. In this context, if the real
matrix
of
A,
A over C.
EXAMPLE2
Let A=
[; =n
419
Solution: We have
A-A.I =
[5
- A.
3
-6
-1-A.
2
so det(A-A./) =A. -4A.+ 13, and the roots of the characteristic equation areA.1 =2+3i
andA.2 =2 - 3i=A.1.
ForA.1=2 + 3i,
] [
3- 3i
l
1
-6
A-A.i =
-3- 3i
0
3
-(l + i)
]
[ ]
1+ i
] [
3+ 3i
-6
l
1
A-A.2 =
.
3
-3+ 3t
0
]
[ l
] [;
- (1 - i)
3i 2 3i
by P
1- i
1
Definition
Complex Conjugate
Theorem 1
A by
a+
bi, b * 0 is
:1 is also an eigenvalue,
Proof: Suppose that Az =A.Z. Taking complex conjugates of both sides gives
Az =A.z::::} Al=
since
7it
vector z, as required.
420
Chapter 9
Now we note that the solution to Example 2 was not completely satisfying. The
point of diagonalization is to simplify the matrix of the linear mapping. However, in
Example 2, we have changed from a real matrix to a complex matrix. Given a square
matrix A with real entries, we would like to determine a similar matrix
AP that
also has real entries and that reveals information about the eigenvalues of A, even if
the eigenvalues are complex. The rest of this section is concerned with the problem of
finding such a real matrix.
If the eigenvalues of A are all real, then by diagonalizing A, we have our desired
p-l
similar matrix. Thus, we will just consider the case where A is an n x n real matrix
with at least one eigenvalue ;l = a + bi, where a, b E JR and b 'f. 0.
Since we are splitting the eigenvalue into real and imaginary parts, it makes sense
to also split the corresponding eigenvector z into real and imaginary parts:
z=
[Xl tll
Xn n
: +i : = x+iy'
x,y
JR.11
Ax= ax - by
and
Ay=bx+ay
(9.1)
Observe that equation (9 .1) shows that the image of any linear combination of
x and y under A will be a linear combination of x and y. That is, if v E Span{x,y},
then Av E Span{x,y}.
Definition
Invariant Subspace
ky,
Theorem 2
[_: !].
where P = x
coordinates.
p-l
AP,
Section 9.4
421
Thus, we have a real matrix that is similar to A and gives information about the
eigenvalues of A, as desired.
Definition
Real Canonical Form
EXAMPLE3
[: ]
[ =;J
b .
IS
Solution: We have
det(A - A.I) =
-5
_
= A.2 +
_2
A.
So, the eigenvalues of A are A.1 = 0 + i and A.2 = 0 - i = A.1. Thus, we have a= 0 and
[ ].
] [1
A_ A.ii=
-5
-2 - i
-2
so an eigenvector corresponding to A. is
P=
EXAMPLE4
[ ]
[; =l
EXERCISE 1
[ ;J.
-
1
p- AP=
[- ;J.
:i' =
2 - 3i.
422
Chapter 9
Remarks
1.
[ ]
can be rewritten as
c se - sine
sm e cose
va2 + b2 ?
where cose = a/Ya2 + b2 and sine= -b/Ya2 + b2. But, since the new basis
vectors 1 and y are not necessarily orthogonal, the matrix does not represent a
true rotation of Span{x,y}.
2. Observe in Example 3 that we could have taken ,i = -i and tl =
[ -]
Ti.
Thus,
(1
EXERCISE 2
- [ _10 0.1]
The Case of a 3
[-i ]
[ i] [ \:i]
[ ]
would also
3 Matrix
Av= v,
Ax=ax- by,
Ay =bx+ ay
W. X,j1) is
_: n
The matrix can be brought into this real canonical form by a change of coordinates
with matrix P = [v 1 y]. .
EXAMPLES
Find a real canonical form of the matrix C of A =
coordinates matrix P such that p-1AP = C.
[-1
-2
-4
-11
-4
4
EXAMPLES
Solution: We have
-4
-4
7-,1
-1 -,1
-1 -,1
(continued)
det(A- ,t/)=
4
0
-2
[3 o ol
1
3,
1
-2
2 .
we have
[
-
A-/=
-4
-4
-2
-4
[-2
- 2i
A-,11/=
6- 2i
-2
4
][
-
5)
Ti.
Thus, a real
1
0
[-H
-4 l
4
-2 - 2i
423
[1
P=
H -il
-l
corresponding eigenvector z= x +
iy,
bi, b * 0
and
other vectors being eigenvectors for other eigenvalues, then in a matrix similar to A,
there is a block
[ !].
-
ize the matrix over C. In other cases, further theory is required, leading to the Jordan
normal form.
424
Chapter 9
PROBLEMS 9.4
Practice Problems
Al For each of the following matrices, determine a
diagonal matrix D similar to the given matrix
over C. Also determine a real canonical form and
give a change of coordinates matrix P that brings
(b)
(c)
[- ]
(b)
(c)
2
0
(d)
-1
-1
1
-1
[= - =1
[ -1
[ -1
4
(e)
-5
-3
-2
2
-2
2
1
(d)
[l ]
[o -;J
2
-3
H
[
1
-1
Homework Problems
(a)
[-1 ]
(f)
-2
-11
-2
-1
-1
-5
Conceptual Problems
Dl Verify that if z is an eigenvector of a matrix A with
complex entries, then z is an eigenvector of A (the
matrix obtained from A by taking complex conju
gates of each entry of A).
D2 Suppose that A is an n x n real matrix and that
tl a+ bi is a complex eigenvalue of A with b * 0.
Let the corresponding eigenvector be 1 + iJ.
=
425
Z Z = Z1 + +
= ( +...+x - Yr - ... - y) +2i(X1Y1 +... +XnYn)
_,
_,
Z11
zz
Observe that z z does not even need to be a real number and so the condition ?:'. 0
does not even make sense. Thus we cannot use the dot product as a rule for de.fining
an inner product in en.
As in the real case, we want (z, tJ to be a non-negative real number so that we can
ZJ. We recall that if z E e, then zZ lzl2 ?:'. 0. Hence, it
define a vector by llZ11
makes sense to choose
<z.
= --./(z,
w>= z w
as this gives us
(z,Z'J
Definition
(z,
EXAMPLE 1
Let
il= v=
[ ]. [;:l
Determine
for W,ZE en
Solution:
a>= v. a= = -
[ : j [; J
[ : J [; : ]
+(3+ +
= (-2+
= +3i+ + = 3+
(il,v>= [; J [ = j
= - 3i+ - = 3
(v,(2 - i)il)= v. (2 - i)il = r-: ] (2 [;: ]
= r-:Jc2+i)[]
= (2+i)(3+
= +23i
<v,
i)(l - i)
4
-1
-1
2i)(2
7i
lOi
7i
- lOi
i
lOi)
-4
i)
- i)
426
Chapter 9
Observe that this does not satisfy the properties of the real inner product. In par
ticular, (it, v)* (v, it) and (v, ail)* a(v, it).
EXERCISE 1
Let it=
[i ]
2i
and v=
[i l
Definition
u, v,
w, z EV and a EC
a(z, w)
EXERCISE 2
Note that property (1) allows us to define the (standard) length by llz ll = (z, z)112,
as desired.
Property (2) is the Hermitian property of the inner product. Notice that if all the
vectors are real, the Hermitian property simplifies to symmetry.
Property (3) says that the complex inner product is not quite bilinear. However,
this property reduces to bilinearity when the scalars are all real.
Theorem 1
Let V be a complex inner product space with inner product (, ). Then, for all w,
z EV,
Proof:
427
(z,w)
a=
--
(w,w)
Then, we get
(z - aw, z - aw)
= llzll2 -
l(z,w)l2
(w,w)
l(z,w)l2
(w,w)
aa(w,w)
+
l(z,w)l2
(w,w)
l(z,w)l2
llwll2
Let C(m, n) denote the complex vector space of m x n matrices with complex en
tries. How should we define the standard complex inner product on this vector space?
We saw with real vector spaces that we defined (A, B) = tr(BTA) and found that this
inner product was equivalent to the dot product on IR.11111 Of course, we want to define
the standard inner product on C(m, n) in a similar way. However, because of the way
the complex inner product is defined on C", we see that we also need to take a complex
conjugate. Thus, we define the inner product (,)on C(m, n) by (A, B)= tr(i/ A).
b11
b111
a11
ain
In particular, let A
and B =
. Then
Since we want the trace of this, we just consider the diagonal entries in the product
and find that
tr(BT A) =
i=l
i=l
which corresponds to the standard inner product of the corresponding vectors under
the obvious isomorphism with C11111
428
Chapter 9
EXAMPLE2
Let A =
2 i
[ ; 1 i]
[ _;i
and B =
Solution:
We have
(A,B)
: l
2 i
1
a=
i
1 -i
and b=
3
2 - 3i
-2i .
1 2i
+
tr
tr
tr
::;
= ab
= (a,
b)
This matrix
Definition
Conjugate Transpose
Let A be an
Il
n x n
of A to be
-T
A* = A
EXAMPLE3
Theorem 2
for all Z, w
D2.
C11
a E
C. Then
conjugate transpose A
429
If we were to calculate (v 1, !) instead of (z, v1), we would likely get the wrong answer.
Notice that since (z, w) may be complex, there is no obvious way to define the
EXAMPLE4
Let v1
0
, v2
0
0
1
1
=
. , and v3
-1
ofe4.
(a) Use the Gram-Schmidt Procedure to find an orthogonal basis for S.
Solution: Let w1
0
Then
0
0
.
w3
v3 -
(v3, w1)
llw1112
w1 _,
0
+i
w 2
0
1
2
11w211
0
0
(v3, w2)
_,
2i
0
i/3
1/3
2/3
-1
0
and get that
2
{w1, w2, W3} is an orthogonal basis for S.
(b) Let z
: . Find proj8 Z.
-1
430
Chapter 9
EXAMPLE4
(continued)
.
_,
pro
Js z
<z. w1 >
llw1112
_,
0
-i
0
0
<z. w2>
wi
llw2ll2
+-
3 -l
...
w2
llw3ll2
1
0
0
-1
w3
W hen working with real inner products, we saw that orthogonal matrices are very
important. So, we now consider a complex version of these matrices.
Definition
Unitary Matrix
Theorem 3
(1)
Proof: Let U
[z1
(_,
_,)
Z;, Zj
Thus, since
_,
Z;
-T
( U U) ij - Z;
_,
we get that U* U
-;J
<.j
::t
<.j
->T-::;
Z;
Zj,
for
----=
_, ;f[ Zj
Z;
, Zn} is orthonormal
it= j
--
(Z;,
_, Zj
_, )
Observe that if the entries of A are all real, then A is unitary if and only if it is
orthogonal.
EXAMPLES
[ ]
-i
- [ (1
1 and V -
[]
thonormal.
+ i)
_Li
_l_(l + i)
Y6 2 .
unitary?
ygl
Y3
Hence,
1(1) + i(-i)
is not a unit vector. Thus, U is not unitary since its columns are not or
EXAMPLES
For v, we have v
(continued)
i) _Li
l
'1
= [_L(l
'{3 (1 - i) - i
-
so
_ [ !(2 + I) 3(2-2)] [1
(2-2) 1(2+4) - 0
3 Vi
V VV
'
Y6
Y6
Thus,
431
OJ
1
is unitary.
EXERCISE 3
A= [
2+i
Y6
1
- Y6
V6.
2+1
Y6
[ lY}+i
Y6
PROBLEMS 9.5
Practice Problems
Al Use the standard inner product in C" to calculate
(a)
i1
(d) D
+ i)/-f3
= [(-11;V3
(1
-i)/../6]
2/../6
(b)
(c)
(d)
(b) B
(c) C
Homework Problems
Bl Use the standard inner product in C" to calculate
(a)
(b)
ii= [i
(b) Determine the projection of w
it = []
= [ : 1
+
it 3
and
onto
and v.
(b) Give a
2 2
det U * l.
unitary
matrix
such
that
Chapter 9
432
B2
(b) B
(c)
Cd)
B3
C=
D=
[ ]
l (1+
(1+
l(1+
- 2i
B4
i)/--./7
5 /Y35
2i)/--./7 (3+ i)/Y35
i) I Y6 (1+ i) IY3
2i/.../6
i/Y3
-
[--1i].
] and
[-l+i
.
v2 =
[ i
[1- l .
Span
{v1, v2}.
1+ i
and v
[;1 ] onto
Conceptual Problems
Dl
D2
D3
detA = detA
D4
In Section 8.1, we saw that every real symmetric matrix is orthogonally diagonalizable.
It is natural to ask if there is a comparable result in the case of matrices with complex
entries.
First, we observe that if A is a real symmetric matrix, then the condition AT = A
is equivalent to A = A. Hence, the condition A* = A should take the place of the
condition Ar = A.
*
Definition
Hermitian Matrix
*
=
or, equivalently,
Section 9.6
EXAMPLE 1
433
A=
2
3
3-i
Solution: We have
1
-2i
-= [ 3 ]
B
c=
2i
[o
l
B=
'
2i
-2i
f. BT,so B.1s
3-i '
3
A= [3
: ] = AT,
-i
-i
so
A is Hermitian.
..
not Herm1tian.
-l
Observe that if A is Hermitian, then we have (A)ij = A1;, so the diagonal entries of
A must be real and for i f. j the ij-th entry must be the complex conjugate of the ji-th
entry.
Theorem 1
An n x n matrix
e11, we have
Proof: If
If (z,Aw) =
t,w E
Remark
434
Chapter 9
Theorem 2
Suppose that
A is an
(t,AZ)
But, since
A is
(t,,it)
(t,Z)
Hermitian, we have
(t,AZ)
(At,Z)
and
(At,Z}, so
(Al, Z)
;l.
T hus,
(t1,At2)
A2(l1,t2)
;l2(t1,l2)
(Al1,t2)
and
AJ <ti,l2).
Thus, since
A1
,i
;l must be real.
A with correspond
To prove (2), suppose that ,11 and ,12 are distinct eigenvalues of
ing eigenvectors
A1(l1,l2)
From this result, we expect to get something very similar to the principal axis
theorem for Hermitian matrices. Let's consider an example.
EXAMPLE2
Let
[ :.
1
; i] .
Solution: We have A*
Verify that
A, so A
A is
C(;l)
;l
1.
For ,i
A - ,if
is Hermitian. Consider
,12 - 5,i + 4
A.
=
; ].
4,
f
A_ ,i
-2.
1
A_ I)Lf
[I ;
il
t1
1
i
l
] r
[;]
+
l1
-1
- l
i .
If
I ] [
[ ]
+
1_ i
-2
0
;l
1
0
i .
1,
'
. Hence, A is diagonalized to
[ ]
Observe in Example 2 that since the columns of Qare orthogonal, we can make
Qunitary by normalizing the columns. Hence, we have that the Hermitian matrix
A is
diagonalized by a unitary matrix. We can prove that we can do this for any Hermitian
matrix.
435
Theorem 3
The proof is essentially the same as the proof of the Principal Axis Theorem, with
appropriate changes to allow for complex numbers. You are asked to prove the theorem
as Problem DS.
Remarks
1 . If A and Bare matrices such that B= U*AU for some U, we say that A and B
are unitarily similar. If B is diagonal, then we say that A is unitarily diago
nalizable.
2.
Unlike, the Principal Axis Theorem, the converse of the Spectral Theorem for
Hermitian Matrices is not true. That is, there exist matrices that are unitarily
diagonalizable but not Hermitian.
The matrix A =
EXAMPLE3
[: ;]
[0 +1.../6/.../6
1
d'iagona1
]' zed by u =
i)
(1
+ i)/ V3j
-l/.../3 J'
PROBLEMS 9.6
Practice Problems
Al For each of the following matrices,
(i) Determine whether it is Hermitian.
+i
2
(c) C
(d) F
(b) B
[ ..j2 +
y 24 L
_r,;
../2
../2
.../3 + i
-
]
]
[ V:+
1 +i
0
1 +i
436
Chapter 9
Homework Problems
Bl For each of the following matrices,
(b) B
-f2. + i
[
[ -f2.
+i
Y2 +
..f3
]
Y2 - ]
(c) C
(d) F"
[ ../3
../32+ ]
H '. -]
Conceptual Problems
Dl Suppose that A and B are
n x n
Hermitian matri
(c) A-1
D3 A general
(b) A2
b, c, and d if A is
(a) AB
a,
2 2
ci
: c ].
a,
b, c, d ER
a,
b, c, and d if A is
tian matrix.
CHAPTER REVIEW
Suggestions for Student Review
1 What is the complex conjugate of a complex num
ber? List some properties of the complex conjugate.
diagonalizing
tion 9.6)
Hermitian matrix.
(Sec
Chapter Review
437
Chapter Quiz
El Let z1
1 - Y3i and z2
.
Z1
determme z1z2 and -.
Z2
=
[ 3 ;]
and V
[4 J
lowing.
(b) it
(c) (it, v)
Cd) <v,it>
(e) llVll
(f) proja v
E4 Let A=
[ 1
-
13
4
ES Prove that U -
matrix.
E6 Let A=
[
3
...L
V3
ki
[1
l
-i
-l+i
is a unitary
Further Problems
Fl Suppose that A is a Hermitian matrix with all non
eigenvalues.
MyMathlab
Go to MyMathLab at www.mymathlab.com. You can practise many of this chapter's exercises as often as you
want. The guided solutions help you find an answer step by step. You'll find a personalized study plan available to
you, too!
APPENDIX A
Answers to Mid-Section
Exercises
CHAPTER!
Section 1.1
1.(a
[]
(b)
_,
2. 1 =
3. 1 =
4. 1=
[=J
(c) [ =i J
X2
Xt
t [-n t E IR
[ ] +t [-n t E IR
[]+1[-H IER
X2
ca+
w)
i1 + w
Xt
Section 1.2
1. (5)
X1
[
X
+(
x1)
l
o
]
[
X
t
t
l
1
since +
. Then
Let
X
X
X
n
+
(
-Xn
)
n
11
t
[
x
1
t1 tX1 ] E IR1 since tx; E IR for 1
t
x
1
l
s
x1
t
x
1
l
s(t1) s tXn StXnI st Xn (st)1.
1=
-1=
(-1) =
= ; = 0.
0
(6)
(7)
:5
: =
:5 n.
440
Appendix A
[]
[]
that x + y =
t1 =
[ ]
[ ]
and y =
[ ]
xi +Yi
E S since 2(x1 + Y1)
X2 +Y2
txi
ESsince 2(tx1)
tx2
[ ]
is not in T.
This gives c1 +c3 0, c2 +c3 = 0, and c1 +c2 0. The first two equations imply
that c1 = c2. Putting this into the third equation gives c1 = c2 = 0. Then, from
the first equation, we get c3
0. Hence, the only solution is c1 = c2 = c3 = 0,
so the set is linearly independent.
=
1
0
0
1
0 , 0
0
0
X1
1
0
X2
0
0
1
0
0
0
0
0
1
0
0
0
0
0
t1
t2
Consider X3 = t1 0 + t2 0 + t3 1 + t4 0 + t5 0
t3 . Thus, for every
X4
0
0
0
1
0
t4
X5
0
0
0
0
t5
1 E JR.5 we have a solution t; x; for 1 :::; i :::; 5. Hence, the set is a spanning set
for JR.5. Moreover, if we take 1 = 0, we get that the only solution is t; = 0 for
1 :::; i :::; 5, so the set is also linearly independent.
=
Section 1.3
1.
vw
cos()=
= 0, so ()= rads.
2. 11111 = v1 + 4 + 1 =
Y6, II.YI! =
+ i =1
(2), we get
II.XII=
I 1 II
11 11
1 =
11111 =
11 11
::;::
Section 1.4
1.
it. v
projil it =
v
llvl l2
[ ]
pro
ja
v=
[ ;1
8 v. it
=
17
ll ll2 it
it
=1
Appendix A
2.
proj il z1=
1
v=
4
1
v
ll ll
proJ1(ty )=
[-]-[ij:i [-j:i
=
441
2 /14
-1/7
<J+Z>1
y1+z1
:;.1
-z.1
.
1+
1 = proJx y+
2
2
2 1=
2 1=
11111
11111
11111
11111
proj1 z
[i].
(tY) 1
.
y. 1
=t 21.=tproJ1 y
1
"""jj1jj'2
11111
.
Section 1.5
1. w. il= (u2V3 - U3V2)(u1)+ (u3V1- U1V3)(u2)+(u1v2- U2V1)(u3)=0
w. v=(u2V3- U3V2)(v1) + (u3V1- U1V3)(v2)+ (u1v2- U2V1)(v3)=0
2
HHl [=rn
=
llil x VII=
[=l
-xi- 2x2= -2
and
= "9= 3
J=
[=HJ m
[ ! ] [H
=
2x1 + x2= 1.
Put
X3 = 0
and
ili en so Ive
+t
lit
CHAPTER2
Section 2.1
1. Add (-1/2) times the first equation to the second equation:
2x1+ x
4 2+ Ox3= 12
-X3=- 2
x2 does not appear as a leading l, it is
x2=t E R Then we rewrite the first equation as
Since
1
X1=2(12- 4x2)=6- 2t
Thus, the general solution is
[:H t] [] fl]
6
[7 1 ]
-
12
4
R2 - R1
E
I
12 ]
-
442
Appendix A
Section 2.2
11
-3
-1
-1
-4
11
-3
-4
l
l
(-l)R3
[
[
2. (a) rank(A)
180
60
80
160
60
1
0
=
20
R1 - Ri
-3
60
20
100
60
20
(b) rank(B)
R2 - R3
-4
180
-[
l
-1
R3
l
j]
11
-3
-4
R1 - 2R2
-[
[-
l
l
R1 - R,
3. We write the reduced row echelon form of the coefficient matrix as a homoge
neous system. We get
x1 + x4 + 2xs
X2 + X3 + X5
xs
t i, x4
X2
-t2 - 2t3
-t1 - t3
Xi
-t2
2t3
X2
-ti - t3
X3
ti
X4
X5
-1
-2
-1
-1
1 + t2
0 + t3
t2
t3
t1
Section 2.3
!. Considert1
[=l Hl {] m
+
,,
t1 + 2t2 - 2t3
-3ti - 2t2
-3t1
2t3
+ t2 - 3t3
t1 , t2, t3
IR
t2, and
Appendix A
443
Solving the system by row reducing the augmented matrix, we find that it is
consistent so that v is in the span. In particular, we find that
2
2
[
C-2) =;l+ I-l+ i--l= I l
l
19
4
13
4
2. Consider
0
0
O
Row reducing the coefficient matrix, we find that the rank of the coefficient
matrix is less than the number of variables. Thus, there is at least one param
eter. Hence, there are infinitely many solutions. Therefore, the set is linearly
dependent.
CHAPTER3
Section 3.1
1. consider
.simplifying
0
O
ti +t1+3t3=
2ti+t2+ St3 -t4=5
t1 +3t2+ St3 - 2t4= -5
t1 +t1+3t3=
Using the methods of Chapter
X is in the span of 'B.
2. For any
2,
2 [xX3i x2X4],
xi[ ]+x2[ ]+x3[ ]+x4[ ]=[ :]
22
xi =x2=x3 = x4=
we have
matrix
0,
then the only solution is the trivial solution, so the set is linearly independent.
444
Appendix A
3. AT=
[ -1
[ ;]
3AT=
and (A')T=
15
[-
=(3Al
4. (a) AB is not defined since A has three columns and B has two rows.
(b) BA =
[4
(c) ATA =
(d ) BET=
1
7 2 -1
[l i]
8
13
[;
Section 3.2
1. fA(l,0) =
[n
fA(O,1) =
[-n
fA(2,3) =
[]
[]
,f,(-3,1,0,1) =
[ ]
f(3,0).
2
(b) G is linear since for all 1,y E JR. and t E JR, we have
f.
X2
=t
1
X - X2
] [
+
tx2 + Y2
(txi +Yi ) - (tx2 + Y2)
Y2
=tg(x) + g(Y)
YI - Y2
[ ] [; ]
X3
X4
Y3
Y4
tx3 + y3 + tx4 + Y4
tx1 + Y1
= tH(x) + H(Y)
Appendix A
445
Section 3.3
[ o ol
[-1 o o
1l
-1
[ J
0.
0.
Section 3.4
-4
-2
2. If 1
[:l
-3
-5
-6
-3
-5
8
1 +t 0,
3. The general solution is x = 0 +s
9
0
-6
0
4. We have
span
-4
-2
{[i]}.
s,t E R
Range(L) = X1
[-J [-J [J
+X
+ X3
Thus,
Range(L) = Span
5. [L] =
[-
-1
2
So Col(L) =Span
{[l [m
1 o
0
0
= Range(L).
446
Appendix A
-8
Thus,
{ -i
and
and so rank(A)
nullity(A) =
and nullity(A) =
= 4, as required.
Section 3.5
[2
2' We have
'
b is x =
[- .
A-1 E =
[R]
[ l
_
so A 1 =
x2
x1.
[R][R]
[ l
as required.
[ l
as required.
-t.
We get
[S] rs-1]
[ ][ 7]
Appendix A
447
Section 3.6
[ n
l
[
l
j
[
[ ][""
0
!.
Ut
E=
Then,
a"
a21
a12
a13
a22
a23
a31
a32
a33
while
2.
a12
ka22
au
ka23
a31
a32
a33
a11
a12
a21
a22
a23 = ka21
ka22
a31
a32
a33
a31
a32
a13
ka23 , as
a33
R=
5432E1A = R, where
E1 =
[- ]. [
[ i
2 =
j
j
a11
ka21
A to
We row reduce
a13
a12
EA=
kR2
required.
[ n
We find that
-3
Es=
Section 3.7
1.
[-!
[-
-3
-J
-3
_;
-6
-5
l
l
H l
[ -\ j]
[ l
H j
]
ll
[
[ ! i [- ]
R, + 4R1
-\
0
R3 + 2R2
(a) We have
and take
A= LU=
y= Ux. W riting
-2
-2
L=
::::}
-2
A= LU= -4
3
2.
o
0
o
1
L=
::::}
-6
R3 - 3R1
Therefore, we have
Write
Ly=
Y1 = 3
-4y1 + Y2 = 2
3y1 - 2y2 + y3 = 6
Ax=
b, we get
bas
LU.X =
448
Appendix A
y1 = 3, Y2 = 2 + 4(3)= 14,
so
Hence 5'
and
[{ l
y is
-xi + x2 2x3 = 3
3x2 5x3 = 14
5x3 = 25
X3 = 5, 3x2 = 14 - 5(5)::::? X2 = -1,f-,
= P3
-x, = 3 !j- - 2(5) x, = !jl.
+
(b) Write
Ly=
Ax
=>
bas
b, we get
LUx
band take y
= Ux.
and
[ l
YI= 8
-4yI Y2 = 2
3yI - 2y2 + y3 = -9
YI= 8, Y2 = 2 4(8)= 34,
5' =
y3 = -9 - 3(8) 2(34) = 35.
Ux =
-xi x2 + 2x3 = 8
3x2 + 5x3 = 34
5x3 = 35
X3 = 7, 3x2 = 34 - 5(7) X2 = -t,
[17/3]
-Xi = 8 t -2(7)::::? XI = lj-.
x = -v3 .
+
and
::::?
and
[;l
Hence
so
y is
CHAPTER4
Section 4.1
1. Consider
3 0
2 1 5 -1
3 5 -2
3 0
0 2
1 1
0
1
0
0
0
1
0
Appendix A
449
p(x)
t1, t2, t3, t4
+Sx - 5x2+x3 = t1(1+2x+x2+x3)2 +t2(1+x+3x2+2x3)
+ t3(3+Sx+5x +3x3)+t4(-x - 2x )
To determine if
such that
E span
and
3
21 5
1 2 -3
5
3 5 -2 -5
1
3
p(x)2
a+bx+cx +dx3,
1
0
-1
1
0
0
0
0
0
0
0
1
0
1
0
0
t1 = t2 = t3 = t4 =
0
0
13.
we get
Section 4.2
1. The set S is not closed under scalar multiplication. For example,
[ ]
S, but
[]=[s.
2. Observe that axioms V2, VS, V7, V8, V9, and VlO must hold since we are
using the operations of the vector space
Let
A = [l 2]
and B
Vl We have
= [ i2J
M(2, 2).
be vectors in S.
02,2+A.
02,2 = [ ]
S (take
al = a2 =
0) satisfies
A+02,2 = A =
A = r-;i -a2].
tA = t[ 2] = [ti ta2]
is ( A)
-
which is clearly in S.
multiplication.
450
Appendix A
S l p + q = (a1 + a2) + (b1 + b2)x + (c1 + c2)x and (b1 + b2) + (c1 + c2) =
b1 + CJ + b2 + C2 = a1 + a2 sop+ q E 1U
2
Section 4.3
1. Consider the equation
[ ] - [ 1
0
Observe that this implies that the system is consistent and has a unique solution
-1
][
2 0
0 - 0
2
x3, x - x3, x - x3} spans S. Verify that 13 is also linearly
Appendix A
{[: l []}
451
{[: ]}
We pick
lH
Thus,
[1
X1
X2
X3
{[;]. [] [!]}
above with !I
[ !]
X1
Xt - X2
X3 - X2
[!].
row, and hence 13 also spans JR3. Thus, 13 is a basis for JR3. Note that there are
many possible correct answers.
5. A hyperplane in JR4 is three-dimensional; therefore, we need to pick three lin
{ , , }
early independent vectors that satisfy the equation of the hyperplane. We pick
13
forms a basis for the hyperplane. To extend this to a basis for JR4, we j ust need
to add one vector that does not lie in the hyperplane. We observe that
for JR4.
, 1 ' 0 ,
0
does
1s a basis
452
Appendix A
Section 4.4
{] [=!]=[H
[
= l [ l
m. [1]
+ t2
J. Considert
-1
Thus,
the standard basis vectors with respect to the basis '13. To do this, we solve the
Thus,
[ -4
-1
p
1
-1
[I
-4 I
l
[= --4 j]
0
1
-1
-1
-1
-1
j][i ]=[ ]
1
Section 4.5
Y1 + Y2 + Y3
Y2
Hence, L is linear.
]=
tL(x) + L(J)
[]
" [ !]
2. If 1
XJ
Hence,
X
x1
Appendix A
Null(L), then
0 and
x2
[Q ]
0
x3
L(x)
Since
X2
x3
{[ ] [ ]}
X2 +
X1
x2
x3
x3
xi
{[ :J}
x,
453
XJ
] [ ]
=
Xt
+ (X2 + X )
3
[ ]
range of L.
Section 4.6
1. We have
J + 1 [
] +o[ ]
] [ ] [ ]
+ 1
+0
Hence,
[L]23
-2
-2
Section 4.7
= t1L(u1) +
+ tkl(uk) = L(t1u1 +
tkuk), then t1u1 +
+ tuk E Null(L).
But, L is one-to-one, so Null(L) = {0}. Thus, t1u1 +
+ tuk
0 and hence t1
tk 0 since {u1,
ud is linearly independent. Thus, {L(u1 ),
, L(uk)}
is linearly independent.
1. If 0
. .
2. Let
v =
L(x)
L(t1u1 +
tkuk)
t1L(u1) +
tkl(uk)
V.
basis, it is a linearly independent spanning set. Thus, {L(u1), ..., L(u11)} is also
linearly independent by Exercise 1, and it is a spanning set for V by Exercise 2.
Thus, it is a basis for V.
454
Appendix A
CHAPTERS
Section 5.1
1. (a)
(b)
(c)
I I
1; 1
I I
=3(1) - 2(2)=-1
-2
=1(-2) - 3(0) = -2
=2(2)
4(1)=o
1
2. We have C11 = (-1) +
1 3
(-1) +
1 ll
-1
2
= 3, C12 = (-1)1+
I =I
= -8, a nd C13
Section 5.2
1. rA is ob tained by mul tiplying ea ch row of A by
2.
-8
detA = l
16
-14
-1 =
2 0
(-1)
-1
0
-8
-14
16
-]
9
2 0
=
(-1)
0
-14
16
-1
-2
3.
3 1
detA =(-6)(-1) +
=-6
r.
-2
=(-6)(-1)
-1
1
=(-1)
-4
-3
-1
5/3
-4
-3
-7
-1
-5
-4
2
-9 - 4(3)(-1) +1
-7
=6(4) + 12(11)=156
-4
-2
-1
-14
-3
-4
-9
2 0
-4
-1
-6
3 2
3 + 4(-1) +
-2 -4 3
-5
-5
-2
-1
1 -I
=20
-5
-4
Appendix A
455
Section 5.3
1. cofA =
[=; =;]
[-
2
-1
2. A-1 = dA(cofA)7 =
i i
-
-2
Section 5.4
1. We draw i1 and v to form a left-handed system and repeat the calculations for
the right-handed system:
Area(il, v) =Area of Square - Area 1 - Area 2
Area 3
Area 4 - Area 5
Area 6
1
1
1
1
= (v1 +U1)(v2 +u2) - -U1U2 - V2U1 - -V1V2 - -U1U2 - V2U1 - -V1V2
2
2
2
2
=Vt V2 +Vt U2 +V2U2 + V2U2 - U1 u2 - 2v2u1 - V2V1
=
V1U2 - V2U1
l [ JI
[]
-(u1v2 - U2V1)
2. We have L(e1) =
vectors is
Alternatively,
and L(e2) =
[].
456
Appendix A
CHAPTER6
Section 6.1
1. The only eigenvectors are multiples of e3 with eigenvalue 1.
2
2. C(..l)
det(A
,.1
so the eigenvalues are ..l1
..l2
For tl1
0, we have
=
=S.
0 and
0 is Span
For ,.12
we have
0 is Span
3. We have
C(..l)
det(A - ..l/)
=(S - I=
tl)
So tl1
=S
=S,
2. For ..l1
S
=
-,.l
0
0
-3
-tl
-2
Al =-(,.l -S)(,.l
=
2
-4 _
0
0
-2
2
2
-9
5 ;s
-2, we have
2
2)
[
o
]
[
o
ol
A -SI = -S -
2
-4
- tl
2 ;s
0
0
{[]}
1
0
0
1
0
{ [-q},
Appendix A
457
Section 6.2
-1.
1. We have C(,i) = det(A - ,if) = -,i(,i 2)(,1 + 1). Hence, the eigenvalues are
,11 = 0, -i2 = 2, and ,t3 =
For ,11 = 0, we have
- Of =
[ 111 -1] [1 11
{[:]}
[-111 -11 [1
]
{[=il}
0
-7
-4
3
-7
4
0
0
0
1
0
-1
0
A - 2f =
-7
-1,
0
-6
3
-7
0
0
0
1
0
we have
[ - 1 [ ;]
{[il}
r: =; il
[ ]
[ l
{[]}
- (-1)/ = 1
-7
and getP-1AP=D=
-1
. Hence,
the geometric multiplicity is less than the algebraic multiplicity. Thus, A is not
diagonalizable.
Section 6.3
1. (a) A is not a Markov matrix.
(b) Bis a Markov matrix. We have B- I =
[ ;l
[ l [l
fixed-state vector is
'f2 =
3
'f =
[-:
[].
0.6
-Q.6
1,
[1 -1]
Q
[n
. ThUS,
ltS
,i =
which is
458
Appendix A
CHAPTER 7
Section 7.1
[i] nJ [i] Hl
1 [ f] [ 6rl: [!!h]}
{
1. We have
s
go a
21-16
0,
di
1/Ys
0 and
,
nJ Hl
2 -../36
1 i12
-1 /{i, and b3
1 v3
5/-./6.
Hence,
b1
[1]11
1 i11
8/-.../3, b2
[!{]
5/-./6
c-v'lS - V2 + 1);-./6
c V2 + 2 Y3 - 1);-./6
y
.
and
( -v'18 - 2)/-./6
( .../2 + 2)/-./6
c V2 - 2 Y3 - 1);-./6
c-v'lS + m + 1);-./6
2
little effort, we find that 11111
6 and 1 y -2.
3. We have 1
With a
Section 7.2
Vt
i1
v2
V3
V4
JR4
such that
i1
1
1
1
-1
=
0,
i1
i1
-V1
V2
V3
V3
V4
V2 - V3 + V4
{ }
-
Hence, S'
Span
2. We have
perps
projs
[] [ ] [ ]
2
5/2
1 - 1
3
5/2
-1/2
1/2
{ -f } .
0, and
Appendix A
459
Section 7.4
a1
a2
] [
b'
, B=
c2 .
h , and C = ci
CHAPTERS
Section 8.1
1. We have C(,1. ) = (,1. - 8)(,1. - 2). Thus, the eigenvalues are ,1.1 = 8 and A2 = 2.
For A1 = 8, we get
-3 -3 - l l
A_ 81 =
-3 -3
0 0
[ ] [ ]
{[ ]}
A-8!= [ ] [ ]
{[ ]}
[ ] [-11{ ;1i].
3
-3
A is orthogonally diagonalized to D=
-3 - l
3
0
-1
0
2. We have C(,1.) = -,1.(,1. - 3)2. Thus, the eigenvalues are ,1.1= 0 and ,1.2 = 3. For
A1 = 0, we get
2 -1 -1
0
A Of = -1
2 -1 - 0 l -1
-1 -1
2
0 0
0
-
1 [1 -11
460
Appendix A
m ]}For = we get
-1 -1 1 [1 1 l
= [ -1 -
81
3,
,1,
-1
-1
-1
-1 -1 -1
Thus, a basis for its eigenspace is
But we need an orthonormal
basis for each eigenspace, 7pp t;Gram-Schmidt Procedure and norl
1
malize the vectors to get 1 / , - .
-:- J 2!ft
- 1 /.../6
Therefore,AisdiagonalizedtoD= byP=
1 - 1 /.../6 .
{[-il [-]}.
]}
{[
[
[ i [;
1;-Y3
2/.../6
Section 8.2
1.
2.
3.
_I
4.
[-
8.
Appendix A
461
Section 8.3
=
[ l
21 =
rn
{[ ]}
r-1 1] [ ]
1 -1
{[ ]}
r ]
-
-1
.We take
CHAPTER9
Section 9.1
1.
(a)3+i
2. (a) ZI
(b)-2 +2i
(c)5
x - yi = x + yi = z1
(b) x +iy =
z1
(c) Let z2
a+ib. Then
if
y =
5i
(d)
13
0.
462
Appendix A
1 +i 1 +i1+ i 2i .
=
=
=
i
1-i 1-i1 +i 2
2i
2i 1-i 2 + 2i
(b)
= 1 +i
=
=
2
l+i l+il- i
4-i
4-i 1-Si
1
21
(c)
= - i
=
1+Si 1+Si1 - Si
26
26
(a)
- - - --
4.
(2 +i (1+i)
--+-i}-i
x
0
=-
Zl = 2 COS
. n
6 +iSin 6
x
vf3 = 2 cos 8
z2 =
r,;
v L.
_
cos
Sn
. Sn
4 +ism 4
../2 cos 8
6. We have
l l = lzl
lzl =Ir cos 8 - r sin 81 = r2 cos2 8 + r2 sin2 8 = ...f7i = r
Using the trigonometric identities cos 8 = cos(-8) and - sin 8 = sin(-8) gives
z = rcos 8 - r sin 8
7. Theorem 3 says the modulus of a quotient is the quotient of the moduli of the
factors, while the argument of the quotient is the difference of the arguments.
Taking z1 = 1 = l(cos 0 +isin 0) and z2 = z in Theorem 3 gives
2_
z2
Appendix A
2 cos
+isin ) .Hence,
2 Y2 (cos
rr
rr
-4 +i sin -4
463
( ( ;) ( 2;))
Y2
Sn
. Sn
4 2(
l2 +ism l2 ) = 1.464+i( S.464)
n 2n
n 2n
2-2i
2 Y2
.
(
-4-3 ) + m (-4-3 ))
= -- (
2
-l+-./3i
-lln
---U- +i
---U- ) -1.366 - i( 0.366)
( -lln
cos -
(2-2i)(-1+ Y3i)=4 Y2
+ 2
+isin -
cos
cos
9. We have ( 1- i)
We have
_ ,.;;
v L cos
= 2512
cos
-Sn
-Sn
4 + i sin 4
sin
(1 - i)5
( 2n +
cos
-4 + 4i
so
isin
2 n
-16+16Y3i
Section 9.2
i
1
1
i
2 l+i
3 -1-2i
2+i
2
S- i
1 +2i
+i, z2
1 0 0
0 1 0
0 0 1
0, and Z3
- i.
Section 9.3
1. All 10 vector spaces axioms are easily verified.
2. The reduced row echelon form of A is
for Row(A) is
ml [!J}
1
0
0
0
0
1 -i
0 0
0 0
Hence, a basis
464
Appendix A
Section 9.4
= 2 + 3i,
1. For ..:l1
we haveA - ;J.1/
An eigenvector corresponding to ;i is
So, ..:l1 =
Hence, a real
An eigenvector correspond'mg to
.So,
)
/l
1s
So,
P=[ -.
Section 9.5
1. We have
6i
2. We have
Z1ZI
Thus,
Z>
for all
and equal to
ZnZn =
2
lzil
if and only if
lz111
0:
tT
ZT
3. The columns ofA are not orthogonal under the standard complex inner product,
soA is not unitary. We have B* B
APPENDIX B
Answers to Practice
Problems and Chapter
Quizzes
CHAPTERl
Section 1.1
A Problems
Al
X2
(a)
(b)
[;]
3[-!]
(c)
(d)
-2
Xi
A2 (
a)
A3 (
[ ]
{[ 1
--131
A4 (a)
(b)
(b
[=] [-] 4
2
[
r-10;[-1 ! 1 ;1
0
-22
(c)
10
(c) u
2[]-2[-]
[ ]
X2
[_]
Xi
[3[62] 3 [ 4 vl]
74//3] [ -rrl
c{
[-1/2]1 13/3
[=]
9/2
(d)
(c)
(b)
Xi
-[]
Xi
X2
[]
[72]
(e)
(f)
(e)
-7/2
(f)
(d) u
-Y2
-Y2 + 7f
466
Appendix B
AS (a)
[l
(b)
A6
[ :;]
(d)
-10
-1/2
- /3
-14/3
Ul-m [ il
m m [tl
ni-m [=]
[]-[_!] nJ
m-ni Ul
[=il nl [t] [=!] [j]
pQ = oQ-oP =
PR = oR -oh
PS = oS -oh
QR=
oR -oQ =
sR=
o R -oS =
pQ +QR=
[:l
PS + sR
[ ] [ J
tE
(b)x=
IER
(d}X=
(a)x= -
+t
[j] Hl
[;] [-;;],
-
(c}X=
(e)X=
+t
A9 (a) XI =
-1
-2/3
+t, X2 =
(b) X1 = l + 3t, X2 = l
AlO (a) Three points
(b) Since
-2PQ =
[J [ =J
+r
nl [H
+l
tE
IER
t E
-1 + 3t,
- 2t,
[= ] [].
UJ [ ;J
t E; X =
t E; x=
+t
+t
t E
t E
[-]
(c) The points S, T, and U are not collinear because SU* tST for any t.
Appendix B
Section 1.2
A Problems
0
0
(c) 1
1
3
10
-7
(b)
10
5
9
Al (a)
0
1
-5
[J o[l o [_:Hl
+!l- [] [Hl
lil [tl m ll
(a) l
(b)
(c) 1
(d) 1
1
+
l
-
[n [J [J [J
2
+1
{! i}
U }
,
467
468
Appendix B
{ -I }
{l }
.
-1
jJ + td =
t1J + tJ (t1 + t)J Hence, the set is Span{J} and thus is a subspace.
Assume that there is a non-empty subset 131
{V1, , ve} of 13 that is linearly
o=
c1v1 +
+ ceve
c1v1 +
ceve + Ove+1 +
Ovn
Section 1.3
A Problems
Al (a) Y29
(e) -fiSl/5
(b) 1
(f) 1
[ ;0]
[ ]
1
(b)
1/.../2
3/5
A2 (a) 4/5
c+!J
-2/3
-2/3
(e)
1/3
0
A3 (a) 2 VlO
A4 (a) 11111
(b) 5
(c) .../2
(g) -%
(d) Y17
(h) 1
(c)
2/Ys
1/.../2
0
(f)
0
-1/.../2
(d) 3-%
(c) vT75
-v26; 11.Yll -f35; 111 + .Yll 2 .../22; 11 .YI 16; the triangle inequality:
2 -v'22 ::::: 9.38 $ -v26 + -f35 ::::: 10.58; the Cauchy-Schwarz inequality: 1 6 $
-../26(30);:::: 27.93.
=
3
AS (a)
(b)
:::::
m [- l
[-!] [-:]
Appendix B
(c)
(d)
(e)
(f)
rn [-l
4 -1
0 34
-2 0
0 X1
0 X2
0 X3
0 X4
1/3 3/2
2/3 0
-1/3 -3/2
3
6
0
2x1 4x2 -X3
3x1 -4x2 x3 8
3x1
4x3 0
X1 -4x2x2 5x3
-2x4 0
m
=
A6 (a) k =
A7 (a)
(c)
AS (a)
(c)
(b) k =
or k =
(b)
(d)
(b)
(d)
(b) ii=
Hl1
(d) -12
-1
2
-3
-1
2x1 -3x2 5x3 6
X2 -2
xi - x2 3x3 2
it=
-3
3x1-4x1 -2x2
5x3 -2x3
26 -12
x2 3x3 3x4 1
X2 2X3
-X4 X5
(c) k
A9 (a) ii=
(c) ii
(d) any k E IR
{]
+
= 1
(e)n=
AlO (a)
(b)
469
(c)
[] [] 2 [] [ ]
A Problems
(b) proJvu =
.
_,
[-J.
r3648/2/25J
5
perpil i1 =
[]
,perpvu =
_,
_9
Section 1.4
Al (a) projil i1 =
r-136
102//25J
25
4 70
Appendix B
(d) projil a=
(e) projil a=
(f) projil a=
A3 (a) u =
-[ 4/98/9] [ 40/91/9 l
-8/0 9 -1 -19/9
00
-12
0 1/2 -1 5/2
-0
3
-5/22
-1/20
=
1111
(b) proja F =
(c) p erpa F =
A4 (a) u =
11 11
(b) proja F =
(c) p erpa F =
'perpil a=
'perpil a=
'perpil a=
2/76/7
[3/7220/49]
660/49]
[330/49
[ 270/49
222/49]
-624/49
3/1/M
Ml
[-224/71Yf4.
[-16/78/7]
-[ 693/7/7]
30/7
Appendix B
(b) 13/../38
(c) 4/Ys
(d) Y6
A7 (a) R(l/7,3/7,-3/7,4/7)
(b) R(l5/14,13/7,17/14,3)
(c) R(O,14/3,1/3,10/3)
(d) R(-12/7,11/7,9/7,-9/7)
Section 1.5
A Problems
[= ]
[]
-27
Al (a)
(b)
(c)
(d)
(e)
(0
31
- 4
Ul
Ul
[]
[]
A2 (a) ii x ii
(b) i1 xv
[]
[ -1
-13
-v x i1
471
472
Appendix B
(c) i1x 3w =
[ i
[ -:i
-15
Cd) ax cv+ w) =
= 3(i1x w)
-18
= axv+ax w
A3 (a) Y35
(b) -vTl
(c) 9
(d) 13
3x3 = -5
6x3 = 15
(d) X2 = 0
A6 (a) 1"
(b)
[6\} t;J
x{il+H
tER
tER
A 7 (a) 1
(b) 126
(c) 5
(d) 35
(e) 16
A9
Appendix B
Chapter
4 73
Quiz
E Problems
Et x
{] HJ
+
E2 8xi - x2+7 X3
E3 To show that
ieR
{[], [ ]}
-
linearly independent.
Consider
[] [] [-] [ : : i ]
=
This gives xi
we get ti
ti-t2 and x2
+t1
ti
t2
written as
that ti
for IR.
t2
x2
[]
can be
.
E4 If d * 0, then ai(0) +a2(0) +a3 (0) 0 * d, so 0 S. Thus, S 1s not a subspace
3
of IR. .
On the other hand, assume d
0. Observe that, by definition, S is a subset of
...,
[]
[:] :]
R3 and that
a1 Xi +a2x2 +a3 x3
Lett
aixi+a2x2+a3X3
S since taking x1
0, x,
0, and x3
0 satisfies
0.
0.
To show that S is closed under addition, we must show that x + y satisfies the
condition of S. We have x +y
[ : i
and
X3 +Y3
+ a3y3
Hence, x +y
IR, we have tx
0 +0
[l
tXi
tx2 and
tx3
474
Appendix B
e1
is
2
v e1
=
llVlllleill -v'14
e2 is
e3 is
cos Y =
v. e3
1iv1111e311 =
-v'14
1
[-il
E6 Since the origin 0(0, 0, 0) is on the line, we get that the point Q on the line closest
to Pis given by oQ = proj1oP, where
Hence,
OQ =
=
11di12JJ -18/11
Qp.
[ ;1
1
3
-;
-;
-;
-2
OR= OP+ proj11PQ =
O
2
2
4
5/2
-5/2
-1/2
3/2
Then the distance from the point to the line is the length of PR:
llPRll =
-1/2
-1/2
-1/2
-1/2
= 1
= proj,1 PQ.
Appendix B
475
[] r-:1 Hl
x
2)
(ii) True. This is the definition of a line, reworded in terms of a spanning set.
(iii) True. The set contains the zero vector and hence is linearly dependent.
(iv) False. The dot product of the zero vector with itself is 0.
(v) False. Let x =
[]
and y =
Ul
Then, proj1 y =
[l
while proj51 x =
[;.
(vi) False. If y = 0, then proj1 y = 0. Thus, {proj1 y, perp1 y} contains the zero
vector, so it is ljnearly dependent.
(vii) True. We have
CHAPTER2
Section 2.1
A Problems
Al (a) x =
(b) X=
(c) X =
[1;]
[] fn
[=2]
+
-1
(d) X=
-1
- +t -,
0
t ER
tEJR
476
Appendix B
A2 (a)
-1
0
0
2
1
0
(c)
5
0
0
0
0
-1
0
0
0
-2
1
0
(d)
2
0
0
0
0
2
0
0
2
2
4
0
0
4
(e)
1
0
0
0
2
1
0
0
2
7
0
1
5
2
(f)
1
0
0
0
0
1
0
0
3
-1
24
0
(a)
(b)
[i
8
0
0
2
1
0
1
11
A4 (a) Inconsistent.
(b) Consistent The solution is ii
m [!]
+t
-1
0
3
+t
19/2
0
5/2
-2
-1
0
1
0
ER
-1
-1
1'
0
E JR
-1
+t
+t
O'
0
1
0
O'
ER
s,t
ER
Appendix B
AS (a)
(b)
(c)
-5
24
/11
.
.
.
:;t
=
.
. cons1stent with so1ut10n
-t
l0
/1 l
_11
_10
f[ H
R
IE
-3
-5
11
-8
19
[ = i l
[ l f H
l [
[ -r
Hl
l
l[
[
1 [
[
x=
(d)
[ ]
1 ]
[ 1 ] [
[ - I H - 1 ;J
3
-17
-8
-3
-2
-11
-5
Consistent
IE R.
3 6
16
-2
477
-5
2
1
-11
5
with
[fl
solution
X=
(e)
-1
-1
13
19
-5
-17
-21
-6
(g)
-2
-3
0+
-5
-7
3
3
-3
1
0
4
-5
10
-1
-8
-7
solution x =
-3
(f)
10
Consistent
with
tER
'
0
2
0
-4
-3
-2
0
0
2
1
3/2
-1
t 3/2 '
+
2
-3/2
t ER
478
Appendix B
Section 2.2
A Problems
Al (a)
[i n
[ H
H
[
[i n
the rank is 2
(b)
the rank is 3.
0
0
(c)
the rank is 3.
0
0
(d)
the rank is 3.
(e)
(f)
(g)
(h)
(i)
[
[
3/2
1/2
17
23
; the rank is 2.
n
-H
the rank is 3.
the rank is 3.
-1/2
; the rank is 3.
-56
; the rank is 4.
-6
-2
1 ,
ER
0
+ t
-2
1 ,
-2
1
0
0
+ t
0
1 ,
0
s,tEJR.
s,tEJR.
Appendix B
4 79
-2
1 -1
+t
s, t
ER
0
0
0
-4
0
+ t
5,
s, t E IR
0
0
-1
t E JR.
0
0
A3 (a)
[1 -i [ 1];
[ 1 =;] [ =];
[H
-1
1 -1 1
4
-3
solution is 1
(b)
0.
-7
general solution is 1
(c)
t ER.
-7
-3
-3
-5
-2
-4
-3
-7
+t
s, t
ER
0
0
(d)
5
4
11 1 1
-11
-1
-3
-2
-2
-2
+t
s, t
ER
A4 (a)
[ - I ]- [ I l
[ l
480
Appendix B
(b)
-11
[ -2 -3 5 -1 l - [ ! - l
7
Hl[ i 9 -1 191 ]- [ 0 0 -0 1 ]
[ 1324 -1-6-3 014 -21 i [ 1 00 -510 00 -7 i .
7-7 5
10 -11 '
2
0
01 22 -2-3 0 4 21 10 10 00 00 -10 -40
22 54 --5 33 10 53 00 00 10 0 -3/23/2 -12
7
-40 01
-12 -3/2
3/2 '
1
0
01 00 -31 .
-13 04
[: 1 2 I [ 0 1 2i
[n
r-n
1 0 5 -2 i
- l - [ 0 1 0 1
2 -5 4 0 0 0 0
nl fH
fH
[ 12
(c)
+t
Consistent
with
solution
Consistent
with
solution
ER
-8
X=
(d)
IE R
-8
X=
(e)
(D
-5
-8
solution x
(g)
+t
ER
AS (a)
Consistent with
+t
ER
The
solution
to
X=
IER
is
[A I b]
is
. The solution to
[A I b]
Appendix B
(c)
-1 5 2 1 -1 ] [ 10 01 -59 2 -51
1 ]
[ - -5 -9 1
010 50 -2O'
-95 -2
0 O'1
01 00 -20 ]
0 -12
u 2 50 lH 0 1 1 -1
-11 21
0 1
02
is
-1
-4
+ s
_,
-1
-1
(d)
is x =
X=t
+t
-1
'
+t
-1
-4
3
-
-4
s,t
+t
homogeneous system is x = s
A Problems
(b)
(c)
2
-1
2 01 (-1) 01 12
5
+3
01
1 +
(-1 )
-1
-1
8
4
-1
1
(-2) -10
I
(b)
2
2
-2
0 (-1) 2
3
A2 (a)
4
is not in the span.
6
7
3
+
-1
1 +
The solution to
[A I b_,]
s,tER
The solution to
[A I b]
tER
Section 2.3
Al (a)
481
02 (-2) -1-1
1
-7
3
08
482
Appendix B
1
(c)
A3 (a) X3 = 0
(b)
X1 - 2x2 = 0, X3 = 0
(c) x1+3x2-2x3 =0
(d)
X1 +3x2 + Sx3 = 0
0
-
2t
-t
0
0
3
+t
6
3
0
0
(c) Linearly dependent. 2t 0 -t 1 +t
= 0 ,
1
1
3
0
1
1
3
0
2
0
0
O'
0
t E IR
t E IR
f. -3.
A 7 (a) It is a basis.
3
(b) Only two vectors, so it cannot span IR . T herefore, it is not a basis.
3
(c) It has four vectors in IR , so it is linearly dependent. T herefore, it is not a basis.
(d) It is linearly dependent, so it is not a basis.
Section 2.4
A Problems
Al To simplify writing, let a=
Appendix B
-N3 +aN6
-aN4 +N1
-N7 - aN6
0 and -aN4 - Ns
0 and -aN6 - Fv
R1 +R2
-R2
-R2
R2 +R3
A2
483
E1
-R3
-R3
R3 +R4 +Rs
-Rs
Rs+R6
-R6
-Rs
-R6
R6 +R1 +Rs
E2
A3
x =
Chapter 2
100
Quiz
E Problems
El
E2
-1
-2
-5
-1/3
. Inconsistent.
1/3
E3 (a) The system is inconsistent for all (a, b, c) of the form (a, b, 1) or (a, -2, c) and
is consistent for all (a, b, c) where b * -2 and c * I.
(b) The system has a unique solution if and only if b * -2, c * 1 and, c * -1.
E4 (a)
-2
11/4
-1
-11/2
1 +t
-5/4
(b) x . a
0, x . v
'
S,
JR
0, and x . w
linear equations with five variables. Hence, the rank of the matrix is at most
three and thus there are at least# of variables - rank
So, there are non-trivial solutions.
5-3
2 parameters.
484
Appendix B
=t1
ES Consider 51
matrix gives
m [1] [1].
[ ][
+t2
+ti
3
1
1
6
1
1 0
o
1
o
0
Thus, :Bis linearly independent and spans IR3. Hence, it is a basis for IR3.
(d) True. If there are more variables than equations and the system is consistent,
then there must be parameters and hence the system cannot have a unique
solution. Of course, the system may be inconsistent.
CHAPTER3
Section 3.1
A Problems
Al (a)
(b)
(c)
A2 (a)
(b)
[-
[
[
-6
1
6
-4
-3
-6
-12
-1
-11
-1
[I
-3
-1
-1
27
f l
n
12
-1
15
(c)
-3
15
l3 -4
-5
19
-1
(d) The product is not defined since the number of columns of the first matrix does
not equal the number of rows of the second matrix.
A3 (a)
(b)
A(B+C)=
[- ! l ] =AB+AC
r6
A(B+C)= l4
-16
14 =AB+AC
A(3B)= ; ; - =3(AB)
A(3B)= _2; =3(AB)
]
n
Appendix B
A4 (a)
(b)
AS (a)
A+
A+
12 2lO] .
AB= [ lO
13
31
(A+
Bl = [= ; ;]
= [- =]=Br
(ABl
485
=Ar+ Br.
Ar.
(b) Does not exist because the matrices are not of the correct size for this product
to be defined.
(c) Does not exist because the matrices are not of the correct size for this product
to be defined.
(d) Does not exist because the matrices are not of the correct size for this product
to be defined.
(e) Does not exist because the matrices are not of the correct size for this product
to be defined.
(f)
(g)
(h)
[l l]
5[ 2 l]
139
46
]
[ 21 12
62
13
3
10
7 7
15
A
AX=
=
Z
nl
{;].
n
[12 l:
17[
ll 1
10
(i)
A6 (a)
(b)
Dre= (CTDl =
AY
A7 (a)
(b)
(c)
(d)
11
[]
[]
[i l
8
4
-4
9
11
486
Appendix B
AS (a)
[-
-1
(b) [OJ
(c)
10
-6
-5
-4
12
3
-9
15
(d) [-3]
-l 3
.
_
A9 B oth s1'des give
27
16]
o
U ]
+( -2)
[- ]
+(-1)
[ -] [ -H
=
+ Oa;-1 +la;+Oa;+1 +
+Oan =a;
Section 3.2
A Problems
2
Al (a) Domain JR. , codomain JR.4
18
-19
(c) fA(l, Q) =
-9
6
,fA(-3, 4) =
17
-23
38
-36
-2
3
l
,fA(O, 1 ) =
3
0
5
-6
-2x1 + 3x2
(d) fA(x)
3x1 + Ox2
=
x1 +5x2
4x1 - 6x2
[-111
, fA(-3, 1, 4, 2)
fA (,) =
fA(t3) =
(c) fA(ti) =
[H
-2
3
3
0
1
4
5
-6
[ ]
l-n Ul
-13
[-H
fA (,) =
Appendix B
(d) fA(x)
487
ll+::: ;::]
X1
2X - X4
3
(e) The standard matrix of fA is
+
(b)
is linear.
is not linear.
[
[
-3
1
0
0
-1
-7
-3
-1
-1
-1
AS (a) The domain is IR.3 and codomain is IR.2 for both mappings.
(b) [S+ T]
3
1
[2S- 3T]
1
-8
-4
-6
9
-5
A6 (a) The domain of S is IR.4, and the codomain of Sis IR.2. The domain of Tis IR.2,
(b) [So T]
10
-19
_10
-3
, [To S]
-6
-9
5
8
16
-1 7
-16
-8
-4
0
-5
AlO [projv]
i l- -7]
[ ]
1
16
17 -4
-4
[ : : =]
-2
-2
488
Appendix B
Section 3.3
A Problems
[ -] [ -] [ ] rn:i -:;]
[S] [ ]
[R9oS] [; ;:]
[
( ) [s R
]
[R] [
] [S] [
]
Al ca)
(b)
A2 (a)
cos e
5 sin e
4/5 -3/5
=
-3/5 -4/5
A3 (a)
-2
-2
00 0 0
00 0
AS (a)
-2
-2
(b)
(c)There is no shear
-3/5
4/5
1
1
(b) - 8
(b)
-sin e
5 cos e
f0 0 1
A4 (a) - -2
(d)
(b)
cc)
4
-4
7
-4
[0 0 0]
4/5
3/5
T such that To P
PoS.
(d)
[ ]
0
Section 3.4
A Problems
3
(b)L(l,1,-2)=
3
-5
5
{[6].[m
0 0 0
0 0
00 ' 0 ' 0 ' 00
0 0
0
{[]}
-1
{0}.
[i 1
[ :1
-1
-2.
-3
Appendix B
489
AS (a) The number of variables is 4. The rank of A is 2. The dimension of the solution
space is 2.
(b) The number of variables is 5. The rank of A is 3. The dimension of the solution
space is 2.
(c) The number of variables is 5. The rank of A is 2. The dimension of the solution
space is 3.
(d) The number of variables is 6. The rank of A is 3. The dimension of the solution
space is 3.
{[] m, rm
[
J}.
{[: l m ,
Then, rank(A)+nullity(A)
3+0
{ j , _! }
{i}
{[l m, rm
Then, rank(A)+nullity(A)
3+ 1
2
(c) A basis for the rowspace is
r}
0 '
' 0
0
-2
Then, rank(A)+nullity(A)
vectorn orthogonal to
[-])
3+2
-4
{[ l [ ]}
-3
{ [- ] } .
{[-l l-m
[!]{ };
(c) A basis for the nullspace is the empty set; the range is
for JR.3
JR.3,
490
Appendix B
AS (a) n = 5
1
0
2
0
3
(b)
-2
(e) x
1
0
0
+ t
-3
-1
0
-1
1
-2
1
(f)
0
0
-3
-1
0
-1
1
0
0
1
0
-1 ' 0
0
1
1
s, t E
(c) m = 4
(d)
{t ' , }
1
3
1
2
-1
-2
-3
-1
0
0
0
-1
(g) The rank of A is 3 and a basis for the solution space has two vectors in it, so
the dimension of the solution space is 2. We have 3 + 2
5 is the number of
variables in the system.
=
Section 3.5
A Problems
Al (a)
(b)
23 -2
2
-1
-1
0
1
0
-1
-1
1
-1
1
0
H ]
10
2
-3
-3
(e)
(f)
A2 (a)
(b)
(c)
-2
0
0
0
0
0
0
]
0
0
0
Hl
ni
[=]
-1
0
1
0
0
-5/2
-1/2
1
0
-]
-1
1
0
-7/2
-1/2
1
-2
2
1
-2
1
Appendix B
A3 (a)
491
(AB)-1 =
(b) (AB) =
(c)
(d)
-1
-1
I
A4 (a) [Rrr;6] - = [R_rr/6]
(b)
(c)
(d)
0 1
[S-1] =
AS (a) [SJ=
=[WI)
(b) [R]
(c) [(RoSt1]=
[(SoR)-1)=
A6 Let v,y E IR11 and t ER Then there exists il,x E JR11 such that x = M(Y) and
i1 = M(v) . Then L(x)
y and L(il) = v. Since Lis linear L(tx + il) = tL(x) +
L(il)
tY
v. It follows that
M(ty
v) =tx +a
Section 3.6
[ 1 ol EA= [6-1 23 Tl
Ol
[
2
A
E
=
2
[
-1 3 ]
2
].EA
-:
=
[
[
-23 I
A Problems
(c)
Al (a)
-u
-5
0,
0 1
0
0 1
0
I
4
0
1
0 -1
-4
tMCY)
M(v)
492
Appendix B
=E [ 00 ].EA= H 1822 2]
01 A= 2
E = [ 0 ].E H 10 ]
00 001 001 000
00
00 000 00 00
0 00
10 01 00 00
00 00 0 0
021 00 001 000
000
00 001 00 000
000
001 001 00 000
000
1
1.
(d)
(e)
A2 (a)
-3
(b)
(c)
(d)
-3
(e)
(f)
-1.
0
0
l
[
0
[
1
0
Ol
E1 = 00 01 01 =] [00 01 1/20 = 00 0 ] = l 01 ]
-A 1 = [00 1-02 01
[
l
0
0
1
0
1
0
1
0
0
A= 0 1 0][0 0 rn 01 m 01 ]
-3
/2
Appendix B
E1 = [ 0I 00 n E2 = [ 00 1 J =[00I 001 ] = [ 0 ]
A-1 = [-7 0 ]
A= [0 00 m 00 rn 00 m 0 ]
[I
0
Ei = 01 HE2= [ 001 HE,=[ 00 H
0
0
E4 = [ 01 n E, = [ 01 I
A-1 = [ 0 !]
0
A=H 0 JU 001 rn 00 rn 00 001 ][100 00 10i
0 00 00 01 0 00 00 01 1 00 00
E i = 0 0 E2 = 0 0 1 0 = 0 0
0 0 0 0 0 1 0 0 0 0 01 0 0 0 00 0
= 00 0 00 'Es= 00 01 0 00 'E6 = 00 0 0 0
0 00 00 0 0 0 1 0 0 0
= 00 0 01 00
000
A-1 = 1 1 0 00
0
0
0 00 00 0 0 00 00 0 1 00 00 0 0 00 00
-1
A= 0 0 0 0 0 1 0 0 0 0 0 0
01 0 0 0 0 0 1 0 0 00 00 0 0 0 0 0 0
00 01 0 00 00 01 0 00 00 01 01 00
000100010001
O
(b)
-2
-3
-2
-2
1
'4
(c)
-2
-3 '3
-2
- 1 /4
- 1 /4
1 /2
4
-1/2
- 1 /4
-I
-4
(d)
,
O
'
O'
-4
-1
1 /2
O'
-1
-1
-2
- /2
- 1 /2
1 /2
-2
-2
493
494
Appendix
Section 3.7
A Problems
-:]
-1
- ]
-1201/210001
114
01 0000504 -7/2
53 -13/2
0 001 00 - 0 0
10 10 000001 -32-2 1
30-4/3-1 17/91 01 00 00-30-37
0-20-11 2220
--3/221 3/210000
1
0
0
0
4
0
-1 -221 0 000
A2(a)LU=[= ! rn _;];11=[=H12=Ul
0
H
LU= rn - i 11 = m 1, =r=i
LU=[= ! rn H 11= [] 1, {]
1 01 0000-10 -12 -33 01 -11 --53
0
LU= -3-1 2001 0 00 00-1201 -31 , 12= -20
(d)
(e)
(f)
(b)
(c )
O;xi=
(d)
Chapter 3 Quiz
E Problems
El (a)
1 =3179]
[-14-1 10
Appendix B
495
1- =1
-8
-42
[]
[-16 ]
- 1
E2 (a) fA(a) =
(b)
.1ACv) =
-11
0
17
(c) [Ro M] =
,-
2+ Y3
2 \(3 - 1
6
ol
2 -1
-Y312
1/ 2
0
[-]
1
[-7 - i
-l - 2Y3
2- 2
2 '\f3+ 2
-2
- '\f3 + 2
4
-2
E4 The solution space of Ax = 0 is x =
-1
5
-2
1
6
0
1 +t 0
of Ax= bis x = 0 + s
1
0
0
7
0
0
1
1
0
0
s, t E
-1
0
+t 0
1
0
s, t E
5
6
obtained from the solution space of Ax = 0 by translating by the vector 0 .
0
7
E5 (a) a is not in the columnspace of B. v is in the columnspace of B.
(b) x =
(cJY =
l=il
m
{ !}
1
of A is
1
0
0
0
0
1 , -1 , 0
0
0
-1
2
3
-1
1
0
0
1
-3
0
-2
496
Appendix B
2/3
1/6
0
0
1
0
-1/3 0
1/2
1/3
-1/6
0
0
1/3
f.
1. The inverse is
[-
1 p
p
1 - 2p
-1
-1
-;;
-p
p .
1
. .
Ell (a) E1
[ ], [ l [ 1. [
[ rn ! m ! -i [ ! +1
1 2
(b) A=
E12 (a)
E =
1/4
3 =
4 =
1 3/2
1
0 0
= /3
[1
[ =1
2
-4/3
(e) This contradicts the Rank Theorem, so there can be no such mapping L.
CHAPTER4
Section 4.1
A Problems
Al (a) -1 - 6x
4x2
6x3
(f) l3
2x - 5x2
x + llx2
3
3
Appendix B
A2 (a) 0
497
(b)
(c)
-x + 2x2 + x3
(d)
-4 - x + 3x2
(e)
-1 - 7 x + 5x2 + 4x3
2x2 + x3)
IR.
IR.
A4 Consider
1
0
0
-1
1
0
1
-2
1
a1
a .
2
a
3
Section 4.2
A Problems
Al (a) It is a subspace.
(b) It is a subspace.
(c) It is a subspace.
(d) It is not a subspace.
(e) It is not a subspace.
(f) It is a subspace.
A2 (a) It is a subspace.
(b) It is not a subspace.
,-
(c) It is a subspace.
(d) It is a subspace.
498
Appendix B
A3 (a) It is a subspace.
(b) It is a subspace.
(c) It is a subspace.
(d) It is not a subspace.
(e) It is a subspace.
A4 (a) It is a subspace.
(b) It is not a subspace.
(c) It is a subspace.
(d) It is not a subspace.
AS Let the set be {v1, ... , vk} and assume that vi is the zero vector. Then we have
0
Ov1
Ov;-1
v;
Ov;+1
+ +
Ovk
Section 4.3
A Problems
Al (a) It is a basis.
(b) Since it only has two vectors in IR.3, it cannot span IR. 3 and hence cannot be a
basis.
(c) Since it has four vectors in IR.3, it is linearly dependent and hence cannot be a
basis.
(d) It is not a basis.
(e) It is a basis.
{[-l [!]}
ml [ il [!]}
{[
{[
[ -n, [ = ]}
=J
U ] [ ] [ ]}
-
. Hence, the
Appendix B
499
(b)
mrnl}
mrnJHJ}
(a)
{ -! )
,
1
1
1
0
0 ' 0
0
0
0
-1
'
{[ ], [ ], rn ]}
{[ :l [ ]}
is 3.
{[ J.rn ][ ] }
Section 4.4
A Problems
) [xJ.
(c) [x].=
(d) [x]3 =
(e) [x] =
[_;J.
[y]3 =
[;]
UJ. Ul
nl {]
[y].=
[yJ.
[n
[y]3 =
[-]
Hl {]
l
Al (a) [x]3 =
500
Appendix B
(b)
m m
and
m.
[ ;J
_
(b) [p(x)]"
(c) [2- 4x
=[: l
[q(x)]" =
[H
!Ox']=
[4- 2x
[H
['(x)].=
We have
7x'J. + [-2 - 2x
Hl
= [2- 4x
3x'J.=
m +!l []
l0x2]13 = [(4 - 2)
(-2 - 2)x
(7
3)x2]!B
[-i j;l
3/4
[! =]
l
3I11
coordinates isP= -15/11
-1/11
o
1
0
2/11
1/11 .
3/11
[-
-2
2/9 -1/9
coordinates isP= 7/9
1/9
4/9
7/9
1/9
-1/9 .
2/9
[- - l
-1 -1
1/3
2/9
coordinates isP= 0 -1/3
1/3 -1/9
-8/9
1/3 .
4/9
501
Section 4.5
A Problems
Al Show that each mapping preserves addition and scalar multiplication.
A2 (a) det is not linear.
(b) L is linear.
(d) M is linear.
([\]
y.
y.
{[:J.lm
1
{[-i]}
We have
{[-l]}
We have
dim JR3.
+
dim JR3.
{[ -][ ][ ]}.
{[ ][ ][ ][ ]}
.Abasis forNull(L)
Section 4.6
A Problems
Al (a) [L]23 =
(b) [L]13 =
A2 (a) [L]13 =
(b) [L]13 =
[ -n
[
-1
-1
r- ]
[ ]
[L(x)]23
].
5
[]
[L(x)]23
[ l
-11
0 = dim P3
502
Appendix B
(b) [L(v,)J.
(c) [4v,)]
A4 (a) 23
(b)
(b) [LJ.
[L(v 2)],
[4v2ll
[L(v2)J.
{[-], []}.
H I
4 m
)
Ul. oUl ]
[10 00
[1]
=
(b) [L]B
-2
(c) 45,3,-5)
A7 (a) [L]B
(b) [L]B
(c) [L]B
(d) [L]B
[H
[]
[ll
[reflo,-2J]B
{[J. [].[:]}
(c) L(l, 2
A6 (a)
[!].
[H
[]
3
-3
[ ]
[ -]
[-1408 -11528]
[ ]
[L(vi)l
[4v1)]
[4vi)l
[- ]
[proj"' -ol
[H [! ]
[!] [ !]
l l [ l
=
[LJ.
[L],
[L]
Appendix B
503
[ ]
[ ]
[ _!]
[ - - ]
[ 1
[- =]
2
(e) [L], =
(0 [L]=
4
0
0
-2
AS (a) [L]1l =
2
0
-2
(b) [L],
1
-3
-1
(c) [DJ=
(d) [T]1l =
Section 4.7
A Problems
Al In each case, verify that the given mapping is linear, one-to-one, and onto.
a
b
=
c
d
(b) Define L
([; ])
d
(c) DefineL(a +bx+cx2 +dx3)=
[; l
[ O]
a2
Chapter 4 Quiz
E Problems
El (a) The given set is a subset of M(4, 3) and is non-empty since it clearly contains
the zero matrix. LetA and B be any two vectors in the set. Then a11 +a12+a13
0 and b11 +b12 +b13 0. Then the first row ofA + B satisfies
=
504
Appendix B
tA satisfies
ta11
ta12
ta13
t(au
a12
a13)
(p
q)( l )
0, p(2)
p(l)
0, q(l)
q(l)
0, and q(2)
(p
and
0. Hence, p
+
q)(2)
q satisfies
p(2) + q(2)
tp satisfies
(tp)(l)
tp(l)
(tp)(2)
and
tp(2)
[ ]
X1
Yl
X2
Y2
X3
Y3
X1
X2
X3 + Yl
Y2
Y3
JR., tx satisfies
E2 (a) A set of five vectors in M(2, 2) is linearly dependent, so the set cannot be a
basis.
(b) Consider
1
2
-
2
2
4
-2
1
0
0
0
0
1
0
0
0
0
0
2
-1
0
Thus, the system has infinitely many solutions, so the set is linearly dependent,
and hence it is not a basis.
(c) A set of three vectors in M(2, 2) cannot span M(2, 2), so the set cannot be a
basis.
Appendix B
E3 (a) Consider
t1
505
0
1
3
0 10 31 11 00
13
02 -20 00
10 33 11 01 01 00 -20
l
1
0
13 l 02 -20 000 000 001 001
3.
10 1 33 02
-1
1
0
13
02 -3
11 33 02 01 0 00 -2-1
01 01 -3-1 - 00 00 01 01
3 2 000 0
+
t2
t3
t4
Thus, '13
ten as a linear combination of the v1, v2, and v3, so '13 also spans. Hence, it
is a basis for and so dim
t1
t2
t3
-5
-5
Thus, [1]
[!]
[=:]
and il2
[H
[j]
v1, L(v2)
[Lhi
-v3, so
[01 o1 ol0
0 0 -1
[o 01 01]
0 1 -1
506
Appendix B
It follows that
[L]s
P[L]BP-1
[ ]
0
1
0
[i -\]
[-: ] [
0
Hence,
[L]
p -1
-1
0
-2
E6 If t i L(v i ) + + tkl(vk)
2/3
2/3
1/3
11/3
-10/3
1/3
0, then
Null(l). Thus,
and hence t1
tk
0 since {vi, ... , vk} is linearly independent. Thus,
{l(vi), ... , L(vk)} is linearly independent.
=
CHAPTERS
Section 5.1
A Problems
Al (a) 38
(d) 0
A2 (a) 3
(b) -5
(e) 0
(b) 0
(c) 0
(f) 48
(c) 196
(d) -136
Appendix B
A3 (a) O
(d)-90
(b) 20
(c) 18
(e) 76
(f) 420
A4 (a)-26
(b) 98
AS(a)-1
(b) 1
507
(c) -3
Section 5.2
A Problems
Al (a) detA= 30, so A is invertible.
(b) detA = 1, so A is invertible.
(c) detA=8,so A is invertible.
(d) detA=0,so A is not invertible.
(e) detA =-1120,so A is invertible.
A2 (a) 14
(b) -12
(c)-5
(d) 716
[ ]
26
=182
7
11
25
7 =-112
8
AS (a) Since rA is the matrix where each of then rows of A must been multiplied by
r ,we can use Theorem 5.2.1n times to get det(rA)=r" detA.
(b) We have AA-1 is /,so
1 = det/ = detAA-1
(detA)(detA-1)
deA.
(c) By Theorem 5.2.7, we have 1 = det/ = detA3 = (detA)3. Taking cube roots
of both sides gives detA = 1.
Section 5.3
A Problems
Al (a)
(c)
[
2
-]
-3
21
11
-1
-13
-7
(b)
[= ]
(d)
[=
-2
=:1
-8 -4
508
Appendix B
-11/15
-8/5
0
A-
t- 0.
3
A
Section 5.4
A Problems
Al
A2
A3
A4
AS
A6
(a) 11
Cb) Ail= [H Av=[]
(c) -26
(d) ldet u JI = 286 = 1(-26)1(11)
Ail=[;]. Av= [-n Area = Jctet [ -JI= 1-81= 8.
(a) 63
(b) 42
(c) 2646
(a) 41
(b) 78
(c) 3198
(a) 5
(b) 245
Then-volume of the parallelotope induced by
is ldet [v1 .. . vn]I
Since adding a multiple of one column to another does not change the determinant
(see Problem 5.2.D8), we get that
V1, ... 'Vn
tV
Appendix B
509
Chapter 5 Quiz
E Problems
El
E2
-2
-2
-3
-1
-2
9
=2(-1)2+3 -3
3
1
0
3 =(-2)(3)(-1)2+3
-1
-8
-6
-1
20
-9
21
-17
12
-17
1 = 5(2)(4)(3)(1)
(b) (-1)4(7)
E6 (A-1) 1
3
E7 Xz
-12
180
120
-* .
224
49
de A
ES (a) det
(e) 7(7)
-1
21
(d) de A
-8
(c) 25(7)
-2
E3 0
ES (a) 3(7)
-1
-1
-2
[ i
-
-2
=33
CHAPTER6
Section 6.1
A Problems
A1
m [-: l
U]
and
[ l
is an eigenvcctocwith eigenvalue ,!
eigenvalue ;i =4.
-6; and
[:]
O;
is an eigenvectoc with
510
Appendix B
2 and ..12
2 and ..12
A3 (a) ..11
{[- ]}.
{[ ]}.
{[ ]}.
{[ ]}.
so it
{[!]}
its eigenspace is
{[]}.
{[- ]}.
(b) ..11
{[]}.
0 and ..12
{[]}
{[]}
5 and ..12
{[]}.
-1 and ..12
{[]}
{[;]}
{[]}
. so it
{[ ]}.
so it
(d) A,
{[]}
so
{r-m,
{[; ]},
-2 has
so it has geometric
Appendix B
(e) ,! 1
{[:]}
( 0 ,! 1
511
{[;]},
{[-il r1;l}
1.
{ril r1;l}
1.
Section 6.2
A Problems
Al (a)
rn
1
with eigenvalue -7. P-
rn
[-il [-l
and
1
p- AP
p-
1 -
11 -1 ]
[
2
[- - 1
0 0 10
1 0.
and tl2
1.
1, [ ]
[1 O]
and
1
p- A p -
p-I
is
is an eigenvector of A
is an eigenvector of A
_3 .
1.
and
.
1ue
.
with eigenva
(d)
14, r- ]
[104 -l
[- n
[ l
p- I AP
is
{[ ]}
{[]}.
[=: - -n
rn
r- ]
and
512
Appendix B
-6 and '12
n-;n.
{[ ]}
=
[- l
fili.
[-; ]
and
-1, '12
2, and '13
{[-l]}
{[i]}
{[-]},
so the geometric
so the geometric
so the geomet
[- -1
0
and D
[- l
0
{[]}
{ [-l ] }
5 with alge
so the geometric
, so the geometric
-i.
overR
(g) The eigenvalues are ;l1
-1 with
{[] []}
{1-i]}
so the
so the
1 - i
0
and D
1 ]
0
-1
Appendix B
513
{[]}.
0 and A,2
{[-]}.
{[ ]}.
=
rn ].
3 and A2
{[]}.
{[- ]}.
and
[ -l
[- ]
[ -]
{I-:]}
{I:]},
and
-2 with
so the ge-
so the geo
{[-l ri]}
{I:]},
1- - i
1- - ]
4 with
, so
so the
and D
514
Appendix B
geometric multiplicity is
2.
[-0 1 1i
and D
2, 2
[0 0 1].
i and
62. 3. ,
1 with
{[-l []}
{[1]}
and ..12
so the
so the
A is diagonalizable
Section 6.3
A Problems
Al (a) A is not a Markov matrix.
25%
A3
67%
1 [ l
10 1 6
33%
Section 6.4
A Problems
[-!]
r-J
(b) ae-051
+ be41
[n
rn
+ be0.3t
a,b ER
a,b ER
7 5%
will be
60%
20%
5.
6.
Al (a) ae-51
[1/3]
[67 1113]
13 .
20%
of
Appendix B
515
Chapter 6 Quiz
E Problems
El (a)
(b)
(c)
(d)
m
m
m
[ f]
is not an eigenvector of A
[ i
-
-1
and D
[ - i
0
-2
1.
Thus, A is diagonalizable.
x.
ES (a) One-dimensional
(b) Zero-dimensional
(c) Rank(A ) =
[j:].
1/2
E7 ae-0.11
r-J
beo4t
[]
CHAPTER 7
Section 7.1
A Problems
[11-VS 21-VS ]
2/YS
l /YS
516
Appendix B
1[ 85/9//333l
4
-5/-1/...../2./2
1/3/5/..22 ./2
-0
(c) [jt]23 =
-5
A3 (a) [x)23 =
(c) [w ]23 =
1
l
3/
YlO
;
1/1/Yl
\!TiO
Yl
l
0
-10/
\!TiOI .
3/W -1;Y15 3/\!TiO
[22/2/4/333]
-5/3/22
6/7/1/....2 ./2./2
3/6/1/....2 ./2./2
(b) [.Y]s =
(d) [ZJ23 =
A4 (a) It is orthogonal.
(b) It is not orthogonal. The columns of the matrix are not orthogonal.
(c) It is not orthogonal. The columns are not unit vectors.
(d) It is not orthogonal. The third column is not orthogonal to the first or second
column.
(e) It is orthogonal.
AS (a)
fr 81
(b) p =
0. Hl
[-2 -2 -1i
82 =
[l -I ]
[-75+4.../2
-1+4
.../2.../2 -4-8-2.../2
1
5+4
+4
2..
.../2
-4 8-2 8+ ./2
[-2/1/../6../6 1;1;Y3Y3 1/..0 1./2
1/1../6 1;Y3 -1/.../2
.
:;1
Jfi
;
;
[
I
;
:
[ 1}
:::: ::: :J [
(c) [L] =
},
(d) [L]s=
-2
Y2
Y2
Y2
no
.s
11.../2
o{J J
-1;.../2
.""us
, w
is ortho-
Appendix B
517
Section 7.2
A Problems
5/2
5
Al (a)
512
5
AZ (
)
2
3
(c)
5
2
/2
9
(b)
5
9/2
{ }
{[-m
mi rm
mrnrnl}
{[:J lil HJ}
{ -: , -: } { i = }
{ll ;-fi.:-l [OJ , [ 6 ]}
{[ ;;], ['5]. [-:!rs]}
(c )
( b)
-1
-1
'
AJ (a)
(c)
A4 ( a )
Cc)
(d)
1/
'
1;Yi
-1 Yi
i;Y3
1;Y3
-2;Y15
2;Y16
1;Y15
-1;Y16
o
' 3/YIS ' 2;Y16
1;Y3
i;Y15
-1;YIO
'
(b)
-2/3
1;Y3
5/Y45
1;Y3
-1;Y15
3/YIS
1;Y3 , -1;Y3 , 1;Y15
o
1;Y3
2;Y15
0
1;Y3
0
Cct)
{v\, ... , vd be an orthonormal basis for and let lVk+l ... , v11} be an
orthonormal basis for 1-. Then {v1,
, v11} is an orthonormal basis for IR.11 Thus,
AS Let
any 1
Then
perp3(1)
1 - proj31
1- [( 1
(1
v1)v1
Vk+tWk+I
p roj3 1
.
.
(1
vk)Vk]
. + (1 VnWn
518
Appendix B
Section 7.3
A Problems
Al (a) y =
y
10.5- l.9t
y=
(b)y =
3.4 + 0.8t
10.5- 1.9r
0
A2 (a)y =
(b)y =
-t+ t2
383/98
A3 (a) x =
32/49
(b) x
- - t2
167/650
-401/325
Section 7.4
A Problems
Al (a) (x - 2x2, 1 +
A2 (a)
3x2 ) = -84
9 ../59
It does not define an inner product since (x - x2 , x - x2 ) 0.
3x) = -46
(c) 113 - 2x + x2 11
m
f.
-(p, q).
(x, x) = -2.
A4 (a)=
{[i] nrnl}
[ ] [-21 ;;]
[- 13] [ -7/3/3 4/33]
pro
j8 _
projs
(b)[X] =
[;l
= 11
Appendix B
519
AS We have
(Vt++vb Vt+ +vk) =(v1, Vt)+ +(Vt, vk) +(v2, Vt)+ (v2, v2)+
++(v2, vk) ++(vb v1) ++(vbvk)
Since {v1, ..., vk} is orthogonal, we have (v;, Vj)
Chapter
0 if i
j.
Hence,
Quiz
E Problems
El (a) Neither. The vectors are not of unit length, and the first and third vectors are
not orthogonal.
(b) Neither. The vectors are not orthogonal.
(c) The set is orthonormal.
v3x = -
v1x=2,
Y3
Hence, [x]B =
2
9/Y3 .
6/Y3
E3 (a) We have
1=
Thus, det P
1.
=
Thus, PR is orthogonal.
E4 (a) Denote the vectors in the spanning set for S by z1, z2, and z3. Let W1 = Zt.
Then
-1
-1
{ [i ,
1;V2
1I
1;V2
0
1;V2
-1/2
1/2
1/2
,-1/2
}
.
520
Appendix B
[ ].
Since det A =
0,
it follows that
0
0
(b) We verify that (,) satisfies the three properties of the inner product:
(A,A) = aT 1+2aT 2+2a 1+a 2 2::
and equals zero if and only if A = 02,2:
(A, B) =a11b11+2a12b12+2a21b21+a22b22
=b11a11+2b12a12+2b21a21+b22a22 = (B,A)
+2a21(sb21+tc21) +a22(sb22+tc22)
= s(a11b11+2a12b12 + 2a21b 21+a22b22)+
+t(a11C11+2a12c12+2a21C21+a22C22)
=s(A, B)+t(A, C)
Thus, it is an inner product.
CHAPTERS
Section 8.1
A Problems
Al (a) A is symmetric.
(c) C is not symmetric since c12 f. c21.
l /Y2l
[30 -1J
[-1/1/.../2.../2 1;Y2J'
1/YTO 3/YTOl
[-4o 6J
[-3/YTO
l /YTOJ'
D=
D=
[0 o ol0
0 0
0
[o0 o
[
0
0
j]
]
] [09 o9 ol0
[0
0 0 -9
2
1 ;-.../3 -1 ;-../2 -1;%
(c) P = 1/.../3
1/-../2 -1/../6, D =
1/.../3
2/../6
(d) p =
(e) P =
2/3
-2/3
1/3
2;3 1/3
1/3 2/3 , D =
-2/3 2/3
4/../4s -2/3
2/3 , D =
s;V4s
2/Vs -2/../4s 1/3
1/Vs
-1
-1
Appendix B
Section 8.2
A Problems
==
=
1
-
+
x7 + 6x1 x2 x
Q(x1, x2, x3)
X T - 2x + 6x2x3 x
Q(x1, X2, X3)
-2xi + 2X1X2
2x1X3 +
521
A - [-3/2 1 ] Q(x) _
-_ [-1/.../2
1/.../2 ! j1i]; Q(x)
A = [_2 -22] ; Q(x) = P = [1/21-YsYs -l2//YsYs]; Q(x) .
( ) A [- ] Q(:t) - 10 - [21//YsYs -l2//YsYSJ.J ' Q(:t)
l
A=[-3 3 -31;Q(x)=2yT+3y-6y,P= 11/0 1/../3
j -j
;
l
2
/.../6
Q(x)[-4 1 -11../3 -21;Y6 1;0.../2] ,
A= 10 -1 4ol i 1-1/../3
1/../3 1;Y6
1/.../2
;Y6
)
(c)
A2 (a)
3/2
I 2
5 2
- 2Y1 - 2Y2'
'
2x2X3
is indefi-
nite.
(b)
. .
1s pos1t1ve
2
2
y1 + 6y2,
definite.
,,1,
2
2
y1 - 5y2, p -
,,1,
(d)
is indefinite.
(e)
Q(x
-5
Q(1)
=-
3 y - 6y
4y
is negative definite.
(c) Indefinite.
(f) Indefinite.
Section 8.3
A Problems
Al
A2
.
. .
ismdefimte.
522
Appendix B
A4
A3
Y2
xT Ax =
-1
is a hyper
bola.
(c) The graph of xT Ax= 1 is a hyperboloid of two sheets. The graph of xT Ax=
-1 is a hyperboloid of one sheet.
XT Ax=
a hyperbolic cylinder.
(e) The graph of
xT Ax =
xT Ax= -1 is a hyperboloid
of two sheets.
Chapter 8 Quiz
E Problems
El P
1/../3 1;.../2
0
1/../3
l /v6 -1/../3 1/.../2
2/v6
[; ]. Q(x)
E2 (a) A =
definite.
(b)
l /v6
-
A=
Q(x) =
7yi + 3y,
-3
Q(x) =
= 0
-3
and P
1 is an ellipse, and
1- = -i.
-3
16 l
o
and D
o
0 .
4
[; ;.Ji2] Q(x)
.
is positive
-5y+5y-4y, and P=
...,12 1
-1
1/.../2
1;v6
Q(x)
;]
1/../3
0 is a
Appendix B
523
E3
Y2
n
For any x,y,z E IR. and s,t E IR., we have
A = P(3/)PT = 3ppT = 31
CHAPTER9
Section 9.1
A Problems
Al (a)5+7i
(b) -3 - 4i
(c) -7+ 2i
(d) -1 4+Si
A2 (a) 9 +7i
(c) 2+ 25i
(d) -2
A3 (a)3 +5i
(b)2 -7i
(c) 3
(d) 4i
6
21
(b) 53
+ 53
l
A6 (a)
z1z2
25
2
(d) -17
+ 171
524
Appendix B
(b)
(c)
( l?rr +i ?rr)
z z 2 Y2 ( +i )
z1z2 4 - 7i, -fJ - -hi
z1z2 -17+9i, -fi + fii
-4
-54 - 54i
-8 - 8 YJi
2 Y3+i)
5
( rr+krr)+ i ( rr+krr), 4.
2 [ (-z:2krr)+i (-%:2krr)]. 3.
2113 [ (-T;2krr)+ (-T;2krr)], 2.
4. 17 1 [ ( ikrr)+i ( ikrr)J 2,
1 2
cos llir.
12
sin llir.
12
'
z2
sin l12
Y2 cos 12
..1...
il
form.)
(d)
form.)
A7 (a)
(b)
(c)
sin
cos
isin
1 6 cos 0+
sin
cos
1 (
(d)
sin 0+
where
Section 9.2
A Problems
[ 1: ii]3
l
-5 - 51
+i -1+2i
-25+2i +t
-i
0
-t
EC.
Section 9.3
A Problems
Al (a)
57+- 8i3i
r-5; 3i]
-1 - 9il
[
[1+2i 3+]
1 1+ 3i, 1 - 4i) [1 --4i]
{[12i]}.
(b)
A2 (a) [l]
(b) l(2
-4 - i
(d) -1 -7i3
[ -12+il
m].[:]}
{[_:-!]}
{[- \ i]}
{[ 1_;/]. [L]}.
and a
Appendix B
525
ml, [m
{j I}
{ =1 }
{[ U [ 2;]}
l
Section 9.4
A Problems
Al (a) D=
(b) D=
(c )
[ J [ -].
r +i - l r
-2
0
0
2
-
p=
[- ]
p-'AP=
P=
r -]
-1 p-' A = -2
P
,
0
-1
-1
1
ol1 , [o o
o ol
1
+
[
[
- ]
ol
o
1
1
i
+
[
]
[
[
-1
i]
2i
0
1
2
p-1 AP= 0
0
P= 0 0
0
2 1 0
1 - 2i
,
(d) D =
p=
P-'AP=
3
s
Section 9.5
A Problems
Al (a) (il, v) = 2
(b)
(il, v) =
(c)
(il, v)
(v, il) =
(V, il) = 3
llilll
Yil, llVll = 2
11a11
(v, a> =
(b) B is unitary.
A3 (a) (il, v) = 0
A4 (a) We have
(b)
+[ ++iii l
(c) C is unitary.
f
[ ]
i.
526
Appendix B
Section 9.6
A Problems
Al (a)
-i)/Ys
= [(../3l/Ys
D=
Vs
[i o]
5
7
-i)/
_4 1-.fiO-.fiO] D= [ OJ2
( ..f3
0
0
[(1--i)/-{16
2/-VW
(3 - Ys)/b l
(3 + Ys)/a
(1 - i)(l + Ys)/a (1-i)(l - Ys)/b
-
Vs
2/a
2/-VW
where a 52. 361 and b 7.639.
2/b
Chapter 9 Quiz
E Problems
..e-711/12
= 4 ...f2e-irr/12 'z2 ...l.
Yi
The square roots ofi are e irr/4 and ei51114.
7i
(a) 3 +Si
9 + 3i
(c) 11 + 4i
(d) 11 - 4i
2
(e) ff?
(f) rs : :
22 - 8i
2 3i 23i andP-1AP=D= [ 2 3i
(a) P= [
]
2 3i]
il
z
El z 12
E2
(b) [J;l
[ ]
[ l
-
E3
E4
ES
(b)
E6 (a)
P= [
and C
= [ _ ;]
1 i
_;i] [ ; _/_i] =[ ]
A is Hermitian if A = A. Thus, we must have 3 +ki = 3 -i and 3 - ki = 3 +i,
which is only true when k = -1. Thus,
is Hermitian if and only if k = -1.
vu=
[ Ii
Appendix B
(b) If k
1 , then A
[3 i 3 i]
lo3 33 ii
and
det(A - ;l/)
;l I
I
- ;l
+i
-2 and A2
=
;l
(;l
5. For ;l1
2
0
{[3 i]}.
\f [3 i] lp i])
-
0.
5)
-2,
For ;l2
2)(,l
[3+i 3 i] [ 3 i]
{[3_i]}.
5,
527
Index
A
C",413
addition
subspaces, 97-99
eigenvalues, 419
matrices, 419--423
properties, 397
matrices, 117-120
polynomials, 193-196
repeated, 423
complex eigenvectors, 419
C vector space
basis, 412
mapping, 415
eigenvalues, 296--297
parallelogram, 280-282
Argand diagram, 399,405
arguments, 400--401
augmented matrix, 69-72
axial forces, 105
426--428
Cauchy-Schwarz Inequality, 33
determinants, 280-283
125-126
426--428
Hermitian property, 426
length, 426
space, 412
cancellation law and matrix multiplication,
angles, 29-31
area
addition, 396
arguments, 400--401
arithmetic of, 396--397
222,240,329
change of coordinates matrix, 222-224
division, 397-398
back-substitution, 66
orthogonality, 429--431
modulus, 399--401
bases (basis)
multiplication, 396
arbitrary, 321
C vector space, 412
complex vector spaces, 412
codomain, 131
linear mappings, 134
powers, 402--404
coefficients, 63
quotient, 397-398
definition, 22
eigenvectors, 299-300,308
scalars, 411
to, 213-216
linearly dependent, 206--208
cofactors
3
3 matrix, 257
basis, 412
nullspace, 229-230
of n
set, 209-211
ordered, 219,236
preferred, 218
n matrix, 258
dimension, 412
eigenvectors, 417--423
columnspace, 154-155,414
one-dimensional, 412
range, 229-230
complete elimination, 84
complex conjugates
subspaces, 413--414
529
530
Index
forming basis,308
linear transformation,289-290
373-375
mappings, 289-291
413-414
components of vectors, 2
matrices,291
315-317
non-real, 301
dimensions,99
compositions, I 39-141
non-zero vectors,289
orthogonality,367
vector spaces,211-213
principal axes,370
affecting accuracy,78
cone,386
conic sections, 384
three-dimensional space, 1 I
conjugate transpose,428
two-dimensional case, 11
consistent solutions,75-76
consistent systems,75-76
distance,44
division
constraint, I07
complex numbers,397-398
matrices,125
convergence, 357
polar forms,401
coordinate vector,2 I9
domain, 131
linear mappings,134
dominant eigenvalues,312
perpendicular part, 43
Cramer's Rule,276-278
projections, 40-44
cross-products,51-54
properties,348
R2,28-31
length,52-54
R",31-34,323,354
properties,52
symmetric, 335
current,102
E
eigenpair,289
eigenspaces,292-293
2 matrix,255-256, 259
3 matrix, 256-258
deficient, 297
area, 280-283
dominant,312
cofactors,255-261
eigenspaces and,292-293
finding,291-297
geometric multiplicity,296-297
integers,297
invertibility, 269-270
mappings, 289-291
matrices, 291
non-zero,270
n x n
products, 270-271
non-rational numbers,297
non-real,301
matrices,420
projections, 40-44
vectors and lines in R3,9-11
vectors in R2 and R3, 1-11
swapping rows,265
290-291
real and imaginary parts,420
diagonal matrices,114
diagonalization
electromotive force, I 03
3 matrix, 422-423
volume, 283-284
steady-state solution,409-410
electricity, resistor circuits in, 102-104
eigenvalues, 307-310,417
design matrix,344
expansion,269
orthogonality, 367
determinants
square matrices,269-270
defining, 354
290-291
real and imaginary parts, 420
space,152-153
deficient, 297
eigenvectors, 308-310
vectors in R",14--25
Euler's Formula,403-404
explicit isomorphism,248
exponential function,315
F
Factor Theorem,398
applications of,303
basis of,299-300
complex,419
3 matrix,422-423
eigenvectors,299-304
diagonalization,299-304
finding,291-297
factors,reciprocal,265
107-109
finite dimensional vector space
dimension, 2 I5
Index
kernel, 151
hyperplanes
Law of Cosines, 29
in R",24
functions, 131
addition, 134
domain of, 131
linear combinations, 134
linear mappings, 134-136
mapping points to points, 131-132
matrix mapping, 131-134
scalar multiplication, 134
Fundamental Theorem of Algebra, 417
G
Gauss-Jordan elimination, 84
Gaussian elimination with
back-substitution, 68
general linear mappings, 226-232
general reflections, 147-148
general solutions, 68
defining, 354
296-297
geometrical transformations
contractions, 145
cosines, 356
dialations, 145
properties, 349-350,354
R",348-349
instantaneous angular velocity vectors, 390
instantaneous axis of rotation at time, 390
instantaneous rate of rotation about the
axis, 390
integers, 396
eigenvalues, 297
integral, 354
intersecting planes, 387
invariant subspace, 420
invariant vectors, 309
inverse linear mappings, 170-173
inverse mappings, 165-173
procedure for finding, 167-168
inverse matrices, 165-173
invertibility
determinant, 269-270
H
Hermitian linear operators, 433
Hermitian matrix, 432-435
531
cross-products, 52-54
R3,28-31
R",31-34
level sets, 108
line of intersection of two planes, 55
linear combinations, 4
2 matrix, 420
eigenvectors, 289-290
standard matrix, 235-239,328
Index
532
linearity properties, 44
modulus, 399-401
basis, 206-208
matrices, 118-120
polynomials, 195-196
subspaces, 204
LU-decomposition, 181-187
basis, 206-207, 21 l
eigenvectors, 300
nullspace, 414
matrices, 118-120
polynomials, 195-196
orthogonal, 325-329
subspaces, 204
vectors, 323
lines
direction vector, 5
parametric equation, 6
in R3, 9
in R", 24
translated, 5
multiplication
multiplication, 121-126
117-120
mutually orthogonal vectors, 321
N
n-dimensional parallelotope, 284
11-dimensional space, 14
11-th roots, 404-405
11-volume, 284
11
representation of elimination, 71
equations, 69-73
eigenvalues, 420
Hermitian, 432-435
similar, 300
eigenvalues, 289-291
eigenvectors, 289-291
inverse, 165-173
nullspace, 150-152
special subspaces for, 150-162
Markov matrix, 309-313
fixed-state vector, 310
Markov process, 307-313
properties, 310-311
matrices, 69
addition, 117
addition and multiplication by scalars
properties, 117-120
basis of columnspace, 158-159
basis of nullspace, 159-161
basis of rowspace, 157
calculating determinant, 264-265
cofactor expansion, 259-261
column, 117
columnspace, 154-155, 414
complex conjugate, 419-423
decompositions, 179
design, 344
determinant, 268, 280
diagonal, 114, 300
division, 125
73-74
row reduction, 70
rowspace, 414
scalar multiplication, 117
shortcuts and bad moves in elementary
row operations, 76-78
skew-symmetric, 379
special types of, 114
square, 114
standard, 137-140
swapping rows, 187
symmetric, 363-370
trace of, 202-203
transition, 307
transpose of, 120-121
unitarily diagonalizable, 435
unitarily similar, 435
unitary, 430-43 l
upper triangular, 181-182
matrix mappings, 131-134
affecting linear combination of vectors in
R", 133
11 matrices, 296
(REF), 73-74
mappings
unitary, 430-431
natural numbers, 396
nearest point, finding, 47
negative definite quadratic forms,
376-377
negative semidefinite quadratic forms,
376-377
Newton's equation, 391
non-homogeneous system solution, 153
non-real eigenvalues, 301
non-real eigenvectors, 301
non-square matrices, 166
non-zero determinants, 270
non-zero vectors, 41
nondegenerate cases, 384
norm, 31-32
normal equations, 344
normal to planes, 54-55
normal vector, 34-35
normalizing vectors, 312
nullity, 159-161
linear mapping, 230-231
nullspace, 150-152,414
basis, 159-161, 229-230
300,326
eigenvalues, 291
eigenvectors, 291
elementary, 175-179
equality, 113-117
properties, 133-134
Hermitian, 432-435
identity, 126-127
ill conditioned, 78
inverse, 165-173
spaces, 412
one-to-one, 246
explicit isomorphism, 248
invertible linear transformation, 246
onto, 246
explicit isomorphism, 248
ordered basis, 219,236
Index
orthogonal, 33
points
calculating distance between,
28-29
planes, 36
position vector, 7
polynomials
addition, l 93-196
properties, l 94
span, 194-- l 96
323-325
position vector, 7
376-- 377
positive semidefinite quadratic forms,
376--377
power method of determining eigenvalues,
312-313
powers and complex numbers,
402--404
R",323,334
subspace S, 335
parabola, 384
probabilities, 309
problems
linear independence, 95-96
parallel planes, 35
parallelepiped, 55-56
volumes, 56,283-284
spanning, 91-95
products
determinants, 270-271
parallelogram
of inertia, 391
rotations, 291
length, 28-31
line through origin in, 5
reflections in coordinate axes, 146
rotation of axes, 329-330
standard basis for, 4
323,336
quotient, 397-398
multiplication, 401
polynomial equations, roots of, 398
division, 401
zero vector, 11
defining, 9
eigenvectors and eigenvalues of
projections and reflections, 290-29 l
lines in, 9-11
reflections in coordinate planes, 146
rotation through angle ( -) about x3 axis,
145-148
scalar triple product and volumes in,
55-56
vectors in, 1-1 l
zero vector, l l
range
basis, 229-230
linear mappings, 153-156, 228-230
rank
linear mapping, 230-231
matrices, 85-86
square matrices, 269-270
summary of, 162
Rank-Nullity Theorem, 231, 232
Rank Theorem, 150-162
rational numbers, 396
Rational Roots Theorem, 398
projection property, 44
projections, 40--44
eigenvectors and eigenvalues in R3,
parametric equation, 6
290-291
3 matrix, 422--423
perpendicular of projection, 43
properties of, 44
scalar multiple, 41
Pythagoras' Theorem, 44
matrices, 430
contractions, 145
dilations, 145
finding nearest point, 47
533
quadratic forms
diagonalizing, 373-375
orthogonal, 36
indefinite, 376-377
real scalars
parallel, 35
addition, 399
in R", 24
stretches, 145
Index
534
linear dependent,204
linearly independence,18-24
elementary matrix,176
linearly independent,204
similar,300
projections onto,333-336
simplex method, l 09
R", 16--1 7
sines,orthogonality of,356
290-291
elementary matrix,176
skew-symmetric matrices,379
small deformations,388-389
solids
resistance,102
deformation of,145
resistors,103
resources,allocating,107-109
spans,204
summation notation and matrix
multiplication,124
surfaces in higher dimensions,24-25
symmetric matrices
classifying,377
solution,64
diagonalization,327,363-370,373-375
back-substitution,66
eigenvalues,363-367
right-handed system,9
solution set,64
eigenvectors,363-367
right inverse,166
solution space,150-152
orthogonally diagonalizable,367
R"
addition and scalar multiplication of
vectors in,15
dot product,3 l-34,323,354
inner products,348-349
length,31-34
line in plane in hyperplane in,24
orthonormal basis,323,334
standard basis for,321
subset of standard basis vectors, 322
subspace, l 6--17
vectors in,14-25
zero vector,16
roots of polynomial equations,398
rotations
equations for,391
in plane,143-145
through angle(-) about x3-axis in R3,
145-148
row echelon form (REF), 73-74
consistency and uniqueness,75-76
number of leading entries in,85
row equivalent, 70,71
row reduction,70
rowspace,156--157, 414
quadratic forms,371-377
systems
systems,152-153
solution space,150-153
spanned,18
spanning problems,91-95
spanning sets,18-24
definition,18
subspaces,204
vector spaces,209
spans
polynomials,194-196
subspaces,204
word problems,77-78
systems of linear difference equations,
311-312
systems of linear equations,63-68,175-176
applications of,102-109
augmented matrix,69-70,71
coefficient matrix,69-70,71
complete elimination,84
435
square matrices,114
consistent solutions,75-76
determinant,269-271
consistent systems,75-76
diagonalization,300,363
elimination,64-66
facts about,168-170
invertible,269-270
equivalent,70
Gauss-Jordan elimination,84
lower triangular, 1 14
rank, 269-270
back-substitution,71-72
right inverse,166
general,64
upper triangular,114
general solution,68
homogeneous,86-87
inconsistent solutions,75-76
standard matrix,137-140
linear programming,107-109
scalar equation
hyperplanes,36-37
planes,34-36
scalar multiplication
stretches
elementary matrix,176
matrices,117
planes,145
real scalars,399
vector space,199-200
vector space over R, 197-20 l
vectors,2,3,10,15-16
scalar product,31
planar trusses,105-106
resistor circuits in electricity,102-104
state,307
scalar form,6
polynomials,193-196
linear operator,237
subspace S
spanning problems,91-95
unique solutions,75-76
systems of linear homogeneous ordinary
differential equations,317
systems with complex numbers,407-410
orthogonal basis,335
orthonormal basis, 335
subspace S or R", orthonormal or orthogonal
basis for,337-339
subspaces,201-204
T
three-dimensional space and directed line
segments,11
trace of matrices,202-203
bases of,97-99
transition,probability of,309
transition matrix,307,309
definition,16
transposition of matrices,120-121
dimensions,99
Triangle Inequality,33
invariant, 420
trivial solution,21, 87
scalars,2
multiplication by,117-120
Index
trivial subspace,16-17
trivial vector space,203
dimension,212
two-dimensional case and directed line
segments,11
linearly dependent,21,95-96
linearly independent,21,323
basis procedure,213-216
general linear mappings,226-232
mutually orthogonal,321
infinite-dimensional,212
norm,31-32,350-351
inner product,348-352
normalizing,312,322
orthogonal,33,321,333,351
isomorphisms of,245-249
orthonormal,322,351
projection of,335
complex numbers,412
u
multiple,198
polynomials,193-196
equations,75-76
spanning set,209
subspaces, 201-204
trivial,203
unitary matrices,430-431
upper triangular,114,181-182
vectors,197
zero vector, 198
vector-valued function,42
vectors,1
addition,197-201
addition of,10,15-16
scalar multiplication,197-201
zero vector,197
angular momentum,390
representation as arrow,2
reversing direction,3
scalars,41 l
unit vector,33,41,321-322,350-351
vector spaces
components of,2
controlling size of,312
in R", 14-25
scalar multiplication,2,3, 10,15-16
span of a set,91-95
standard basis for R2, 4
zero vector,4
vertices,109
voltage,102
volumes
determinants,283-284
image parallelogram,282
parallelepiped,56, 283-284
in R3, 55-56
w
word problems and systems of equations,
addition,199-200
cross-products, 51-54
basis,206-209
distance between,350-351
complex,396-435
dot products,29-31,323
fixed,309
zero vector,4,197-198
77-78
R2, 11
dimension,211-213
invariant,309
R3, 11
explicit isomorphism,249
length,323,350-351
R", 16-17
scalars,198
535