Professional Documents
Culture Documents
SCOTT ROME
C ONTENTS
1. Preliminaries
1.1. The Invertible Matrix Theorem
1.2. Computing the Rank
1.3. Other Definitions and Results
2. Determinants
2.1. Elementary Row Matrices (Row Operations) and Effects on Determinants
3. Eigenvalues
3.1. Basics
3.2. Eigenvalues after Addition
3.3. Relation to the Determinant and Trace
3.4. Tips on finding eigenvectors quickly
3.5. Questions to answer
4. Characteristic and Minimal Polynomial
4.1. Cayley-Hamilton Theorem
4.2. Using the Characteristic Polynomial to find Inverses and Powers
4.3. More On The Minimal Polynomial
4.4. Finding the Characteristic Polynomial using Principal Minors
4.5. Eigenvalues depend continuously on the entries of a matrix
4.6. AB and BA have the same characteristic polynomial
4.7. Lagrange Interpolating Polynomial
5. Similarity, Diagonalization and Commuting
5.1. Facts About Similarity
5.2. Diagonalizable
5.3. Simultaneously Diagonalizable and Commuting Families
6. Unitary Matrices
6.1. Definition and Main Theorem
6.2. Determining Unitary Equivalence
7. Schurs Unitary Triangulization Theorem
8. Hermitian and Normal Matrices
8.1. Normal Matrices
8.2. Definition and Characterizations of Hermitian matrices
8.3. Extra Spectral Theorem for Hermitian Matrices
8.4. Other Facts About Hermitian Matrices
9. The Jordan Canonical Form
9.1. Finding The Jordan Form
9.2. Remarks on Finding the Number of Jordan Blocks
9.3. Regarding the Characteristic and Minimal Polynomial of a Matrix A
1
3
3
3
3
3
3
4
4
5
5
5
5
5
5
6
6
6
7
7
7
7
7
8
8
8
8
9
9
9
9
10
10
10
11
11
11
12
12
12
13
13
13
13
14
14
14
14
15
15
15
16
16
16
17
17
18
18
19
19
19
19
20
20
20
20
21
21
21
22
22
22
22
22
22
23
23
The following material is based in part on Matrix Analysis by Horn and Johnson and the lectures
of Dr. Hugo Woerdeman. No originality is claimed or implied. The following document is made
available as a reference for Drexel Universitys Matrix Analysis qualifier.
1. P RELIMINARIES
1.1. The Invertible Matrix Theorem. Let A Mn . The following are equivalent:
(a) A is nonsingular
(b) A1 exists, i.e. A is invertible
(c) rankA = n
(d) The rows of A are linearly independent
(e) The columns of A are linearly independent
(f) detA 6= 0
(g) The dimension of the range of A is n
(h) The dimension of the null space of A is 0
(i) Ax = b is consistent for each b
(j) If Ax = b is consistent, then the solution is unique
(k) Ax = b has a unique solution for each b
(l) The only solution to Ax = 0 is x = 0
(m) 0 is not an eigenvalue of A
1.2. Computing the Rank. When computing the rank of a matrix by row reducing, the rank of a
matrix is invariant under the following operations:
permuting rows
permuting columns
adding a multiple of a row to another row
multiplying a row by a nonzero scalar
I include this because I occasionally forget about being able to permute columns and rows. This is
helpful for determining the Jordan Form.
The rank of a matrix can be deduced whenever a matrix is in any echelon (triangular-ish) form
by counting the pivots.
Relation to Eigenvalues In this light, the rank is also equal to the number of nonzero eigenvalues
counting multiplicity.
1.3. Other Definitions and Results.
A is invertible if there exists a matrix B such that AB = BA = I. This matrix is unique
and we say B = A1 .
The standard scalar product for x, y Cn is < x, y >= y x.
A set of mutually orthogonal vectors {vi } such that < vk , vj >= 0 for j 6= k is linearly
independent. (Dot c1 v1 + ... + cn vn = 0 with each vector to prove)
2. D ETERMINANTS
2.1. Elementary Row Matrices (Row Operations) and Effects on Determinants. We will begin with a few examples to build intuition and give a method of remembering the rules using 2x2
matrices. To convince yourself on what each matrix E does, multiply it by the identity (on the
right, so EI).
0 1
The permutation matrix
corresponds to the elementary row operation of switch1 0
ing the first row with the second. Notice this can be deduced from the first row containing
a 1 in the second column (i.e. this says switch the first row with the second row). The
determinate of this matrix is 1.
The matrix
c 0
0 1
multiplies the first row of a matrix by c. The determinant of this
matrix is c.
1 0
The matrix
adds the d (first row) to the second row. The determinant of this
d 1
matrix is 1.
Now let us explore their effects on the determinants by multiplying them by a matrix A. So we
have the following rules
0 1
det(
A) = det(A)
1 0
c 0
det(
A) = c det(A)
0 1
1 0
det(
A) = det(A)
d 1
So from this we have the following rules, I will do the rule for the 2x2 case for convenience.
Theorem Row Operations and the Determinant
(1) Interchange of two rows
a b
c d
= c d
a b
AB and BA have the same characteristic polynomial and eigenvalues. (if A is nonsingular,
they are similar using S = A. If not, for (AB), AB = x = BAB = Bx and so
is an eigenvalue of BA and similarly for the eigenvalues of BA being an eigenvalue of
AB )
If A is invertible, then AB BA, if not they will not be similar. (Example below)
3.2. Eigenvalues after Addition. Let A and B be in Mn
If (A) with eigenvector x, then for k constant, (k + ) (kI + A) with eigenvector
x.
Lemma For any matrix A, for all sufficiently small > 0, the matrix A + I is invertible.
The proof involves noting A + = U T U + U U whose entries on the diagonal are all
nonzero.
If A and B commute, (A + B) (A) + (B). (That is, if (A) = {1 , ..., n } and
(B) = {1 , ..., n }, there exists a permutation i1 , ..., in of indices 1, ..., n so that the
eigenvalues of A + B are 1 + i1 , ..., n + in .)
3.3. Relation to the Determinant and Trace. Assume A Mn and (A) = {1 , ..., n }. We
have the following identities:
n
X
trA =
i
detA =
i=1
n
Y
i=1
1 2 3
Let A = 4 5 6 , = {1, 3}. Then a 2-by-2 principal submatrix of A is A() =
7 8 9
1 3
7 9
Definition The
determinant of a principal submatrix is called a principal minor.
There are nk k-by-k principal minors of A = [aij ].
We denote the sum of
Pthe k-by-k principal minors of A as Ek (A).
Example E1 (A) = ni=1 aii = trA, En (A) = detA.
a b c
A= d e f
g h i
then
a b c
a b a c e f
+
+
t d e f
pA (t) = t3 (a + e + i)t2 +
d e
g i
h i
g h i
4.5. Eigenvalues depend continuously on the entries of a matrix. Facts The following facts
combine to yield the title of the section. Since this is mostly based on ideas involving the characteristic polynomial, I placed it in this section. The following only holds for square matrices.
The zeros of a polynomial depend continuously on the coefficients.
The coefficients of the characteristic polynomial are continuous functions of the entries of
the matrix.
The zeros of the characteristic polynomial are the eigenvalues.
This is discussed more thoroughly in Appendix D of Horn and Johnson. The moral is: sufficiently
small changes in the entries of A will cause small changes in the coefficients of pA (t) which will
result in small changes in the eigenvalues.
4.6. AB and BA have the same characteristic polynomial. AB and BA are similar in the case
the either A or B is invertible and the result is clear. Otherwise: Consider A = A + I which
is invertible for all sufficiently small . Then A B BA . Letting 0, similarity may fail,
but the characteristic polynomials will still be equal since the characteristic polynomials depend
continuously on the parameter . This holds from the above subsections discussion.
4.7. Lagrange Interpolating Polynomial. (0.9.11) add in later
5. S IMILARITY, D IAGONALIZATION AND C OMMUTING
A nice way to think about similarity is the following: Two matrices in Mn are similar if they
represent the same linear transformation T : Cn Cn in (possibly) two different basis. Therefore
similarity may be thought of as studying properties which are intrinsic to a linear transformation
itself, or properties that are common to all its various basis representations.
5.1. Facts About Similarity.
Similarity is an equivalence relation
If two matrices are similar, they have the same characteristic polynomial. Which implies
they have the same eigenvalues counting multiplicity.
The converse of the previous
is not true unless the matrices are normal. (Consider
statement
0 1
the zero matrix and A =
)
0 0
U is nonsingular and U 1 = U
UU = I
U is unitary
The columns of U form an orthonormal set, that is, its range is Rn
The row of U form an orthonormal set; and
For all x Cn , the Euclidean length of y U x is the same as that of x, that is,
y y = x x or again kU xk = kxk.
A real unitary matrix is called an orthogonal matrix.
|detU | = 1
(U ), || = 1.
A sequence of unitary matrices that converges, converges to a unitary matrix.
(b)
(c)
(d)
(e)
(f)
(g)
.
8. H ERMITIAN AND N ORMAL M ATRICES
8.1. Normal Matrices.
A matrix is normal if A A = AA
A Mn is normal if and only if every matrix unitarily equivalent to A is normal.
A real, normal matrix A M2 (R) is either symmetric (A = A ), or the sum of a scalar
matrix and some skew symmetric matrix (i.e. A = kI + B for k R and B = B ).
Unitary, hermitian, and skew hermitian matrices are normal.
For a normal matrix A, Ax = 0 if and only if A x = 0. So they have the same kernel. (Use
the fact below to prove).
For a normal matrix A, kAxk = x A Ax = x AA x = kA xk.
For C, A + I is normal.
A normal triangular matrix T is diagonal. Equate entries of T T and T T
If Ax = x for nonzero x, A x = x. From ||(A I)x|| = ||(A I) x|| since (A I)
is normal.
Spectral Theorem If A = [aij ] Mn has eigenvalues 1 , ..., n , the following are equivalent
(a) A is normal
(b) P
A is unitarily diagonalizable
(A is unitarily equivalent to a diagonal matrix)
P
(c) ni,j=1 |aij |2 = ni=1 |i |2 = tr(A A)
(d) There is an orthonormal set of n eigenvectors of A.
8.2. Definition and Characterizations of Hermitian matrices.
A matrix A Mn is Hermitian if A = A . It is skew-Hermitian if A = A .
A + A , AA , A A are always Hermitian
If A is Hermitian, Ak is Hermitian for all k N, and if A is nonsingular, A1 is Hermitian.
If A, B are Hermitian, aA + bB is hermitian for all real scalers a, b.
A A is skew-Hermitian
If A, B is skew-Hermitian, so is aA + bB for real scalers a, b.
If A is Hermitian, iA is skew-Hermitian. If A is skew-Hermitian, iA is Hermitian,
Any matrix A may be written A = 21 (A + A ) + 12 (A A ) where we have A split into a
Hermitian and Skew Hermitian part.
If A is Hermitian, the diagonal entries of A are real.
Theorem A is Hermitian if and only if at least one of the following holds:
x Ax is real for all x Cn
A is normal and all the eigenvalues of A are real
S AS is Hermitian for all S Mn .
If one of them holds, A is Hermitian and the others hold as well. If A is Hermitian all three hold.
The point of the theorem is that if you can find one of these hold, then you may conclude A is
Hermitian.
For example, if you can show A is unitarily diagonalizable by using Schurs theorem and finding
the non-diagonal entries are 0, then A is normal and if you can then show each diagonal entry is
real, then A is Hermitian.
Hermitian matrices are neccessarily normal and so all the theorems about normal matrices hold for
them.
8.3. Extra Spectral Theorem for Hermitian Matrices. If A is hermitian, then
(1) All eigenvalues of A are real and
(2) A is unitarily diagonalizable.
(3) If A Mn (R) is symmetric, then A is real orthogonally diagonalizable.
8.4. Other Facts About Hermitian Matrices.
For all x Cn the nxn matrix xx is hermitian. The only eigenvalues of this matrix are
0 and kxk22 = x x. Proof: Clearly by choosing a vector orthogonal to x, say p, xx p =
x(x p) = x0 = 0p. So notice xx x = x(x x) = xkxk22 = kxk22 x. If there was another
dim ker[(A I) ] =
k
X
j=1
For a J block of order k associated with (Jk ()), we have that k is also the smallest
number such that
dim ker[(Jk () I)k ] = 0
9.3. Regarding the Characteristic and Minimal Polynomial of a Matrix A. Let A Mn
whose
polynomial is given by pA (t) =
Qm distinctsieigenvalues are 1 , ..., m . Assume the characteristic
Qm
ri
(t
)
and
the
minimal
polynomial
by
q
(t)
=
(t
i
A
i) .
i=1
i=1
The exponent in the characteristic polynomial si (i.e. the algebraic multiplicity) is the sum
of the orders of all the J blocks of i .
The exponent in the minimal polynomial ri is the largest order J block corresponding to i .
9.4. Finding the S matrix. For the Jordan form we have A = SJS 1 for some S. If J is diagonal
then S is simply the eigenvectors of A arranged in a prescribed order,
but if not: To find S notice
1
AS = SJ. Lets look at the 2-by-2 case for S = [s1 , s2 ] and J =
, notice that AS = SJ
0
can be written:
As1 As2 = s1 s1 + s2
So you can solve for s1 as an eigenvector for . To find s2 , we solve (A I)s2 = s1 .
P
xk
9.5. Solutions to Systems of Differential Equations. As ex :=
k=0 k! this gives us the definition of eA = I + A + 2!1 A2 + ... which converges for all square matrices in Cnxn . If we have
two matrices A, B such that AB = BA then we have the identity eA+B = eA eB since eA , eB must
commute here. Using the Jordan form where J = D + T where D is the diagonal entries and T
1
are the other entries, we can see that etA = etSJS = Set(D+T ) S 1 = SetD etT S 1 , which will
allow us to solve systems of ODEs:
x~0 (t) = Ax(t) A Mn , ~x Cn
~
~x(0) = C
~ Another problem you may encounter is: Solve for x(t) where
Then our solution is ~x(t) = etA C.
x3 (t) x2 (t) 4x0 (t) + 4x(t) = 0
To do so, convert this to a system of ODEs.
x1 (t) = x(t) = x01 (t) = x2 (t)
x2 (t) = x0 (t) = x02 (t) = x3 (t)
x3 (t) = x2 (t) = x03 (t) = x3 (t) = x2 (t) + 4x0 (t) 4x(t) = x3 (t) + 4x2 (t) 4x1 (t)
If we define ~x(t) = (x1 (t), x2 (t), x4 (t), x4 (t))T
above as
0
x1 (t)
0
~x0 (t) = x02 (t) = 0
x03 (t)
4
1 0
x1 (t)
0 1 x2 (t) = A~x(t)
4 1
x3 (t)
You then solve this ODE for ~t, and the first component x1 (t) is the solution to the original question.
This construction also gives the general framework for Companion Matrices.
9.6. Companion Matrices and Finding Roots to Polynomials. Let p(x) = xn + an1 xn1 +
... + a1 x + a0 be a polynomial. The Companion Matrix C of p(x) is defined to be
0
1
0
0
...
0
0
0
1
0
...
0
..
C=
0
0
0
0
0
0
0 ...
1
0
a0 a1 a2 . . . an2 an1
Notice that this matrix has 1s along most of the super diagonal, and 0s on most of the diagonal,
and across the bottom are the negative coefficients. Now notice
1
x
x
1
x
x2
x
x2
2
.
2
.
x
x
.
.
C
=
=
=
x
.
.
.
.
..
xn1
..
xn1
n1
n1
n
x
a0 a1 x ... an1 x
x
xn1
if and only if x is a root of p(x) since then xn = an1 xn1 ... a1 x a0 . But this implies x
is also an eigenvalue of C and we also have an eigenvector. Therefore, the important thing about
this matrix is the eigenvalues of C are roots of p(x). Performing the QR algorithm on a matrix
like this is an efficient way to find roots of polynomials.
9.7. The Real Jordan Canonical Form. For = a + bi C we define
a b
C(a, b) :=
b a
Theorem If A Mn (R) then there exists an invertible real matrix S such that
Jn1 (1 ) 0
0
0
0
0
0
..
.
0
0
0
0
0
0
0
0 Jnr (r )
0
0
0
0
A=S
0
0
0
Cnq (a1 , b1 ) 0
0
0
.
.
.
0
0
0
0
0
0
0
0
0
0
0 Cnn (ak , bk ) 0
1
S
Where k = ak + bk i for k = nq , ..., nk and 1 , ..., r are the real eigenvalues of A. Furthermore,
C(a, b) I2
0
0
...
...
0
0
Ck (a, b) ==
0
0 C(a, b)
I2
0
0
0
C(a, b)
is a block matrix with 2k C(a, b) along the diagonal and I2 along the superdiagonal. This matrix is
where = a+bi.
similar to (and takes the place of in the Jordan form) D() = diag(Jk (), Jk ())
10. QR A ND A PPLICATIONS
10.1. Gram-Schmidt. Let x1 , ..., xm Cm be linearly independent. To find an orthonormal set
of vectors {y1 , ..., ym } whose span equals that of x1 , ..., xm use Gram-Schmidt. Algorithm
(1) Let v1 = x1
v x
2 ,v1 >
v = x2 v1 v11 v1
(2) v2 = x2 <x
<v1 ,v1 > 1
1
Notice the P
second term is the projection of x2 onto the subspace spanned by v1 .
<xk ,vk >
(3) vj = xj j1
k=1 <vk ,vk > vk
(4) Let yi = kvvii k = <vi ,vvii>1/2
x
x=1
x6=0 x x
12. C OURANT F ISCHER T HEOREM
Let A Mn be a Hermitian matrix with eigenvalues 1 2 ... n and let k be a given
integer with 1 k n. Then
min
w1 ,w2 ,...,wnk Cn
max
x6=0,xCn
xw1 ,w2 ,...,wnk
x Ax
= k
x x
and
max
w1 ,w2 ,...,wk1 Cn
min
x6=0,xCn
xw1 ,w2 ,...,wk1
x Ax
= k
x x
Remark: If k = 1, n then the theorem reduces to Rayleigh-Ritz. I would like to also put a word on
interpreting this result physically. The min/max statement was so intimidating it took me several
months to actually think about it. On the first equality, it says we first take the max of x Ax/x x
over all x 6= 0 that are perpendicular to a particular n k dimensional subspace of Cn . Then we
take the minimum value of the set of numbers formed for the previous condition over all possible
n k dimensional subspaces.
Corollary The singular values of A, 1 2 . . . q where q = min{m, n} and for 1 k leqn
may be given by
kAxk2
max n
k =
min
n
w1 ,w2 ,...,wk1 C
x6=0,xC
kxk2
xw1 ,w2 ,...,wk1
k =
max
w1 ,w2 ,...,wnk Cn
min
x6=0,xCn
xw1 ,w2 ,...,wnk
kAxk2
kxk2
max
kxk2 =kyk2 =1
|y Ax| =
max
kxk2 1;kyk2 1
|y Ax|
The previous identities can be used to show k|Ak|2 = k|A k|2 for all A Mn . Additionally, k|A Ak|2 = k|AA k|2 = k|Ak|22 from properties of matrix norms and that A A is
hermitian.
P
k|Ak| = max1in nj=1 |aij | This max row sum of the absolute values of the row entries.
P
k|Ak|1 = max1jn ni=1 |aij | This max column sum of the absolute values of the column
entries.
A matrix norm that is
pnot an induced norm (not determined by a vector norm) is the Frobenius norm kAk2 := tr(A A).
kAk := maxi,j=1,..,n |aij | is a norm on the vector space of matrices but NOT a matrix
norm.
13.3. Theorems, Definitions, and Sums.
(A) := max{|| : (A)}
The following lemmas build to the first theorem:
Property
k|Ak k| k|Ak|k
If k|Ak k| 0 for some norm, since all vector norms on the n2 dimensional space Mn are
equivalent, we have Ak 0 because the limit would also hold in the k k norm.
Lemma If there is a matrix norm k| k| such that k|Ak| < 1, then limk Ak = 0. Theorem If
A Mn , then limm Am = 0 if and only if all the eigenvalues of A have modulus < 1. That is,
(A) < 1 if and only if limn An = 0
This tells us that if Ak 0, then (A) < 1 so there exists a matrix norm k|Ak| < 1 in which case
k|Ak|k 0.
Theorem If k| k| is a matrix norm, then (A) k|Ak|.
Corollary Since kAk and kAk1 are matrix norms,
(A) min{max
i
n
X
j=1
|aij |, max j
n
X
|aij |}
i=1
Lemma Let A Mn and > 0. Then there exists a matrix norm k| k| such that k|Ak| (A) + .
k
Theorem If A Mn , then the series
k=0 ak A converges if there exists a matrix norm k| k| such
k
that the numerical series
k=0 |ak |k|Ak| converges, or even if the partial sums of this series are
bounded.
Important
P Theorem If there is a matrix norm such that k|I Ak| < 1, then A is invertible and
A1 = k=0 (I A)k .
P
k
1
Theorem If (A) < 1, then I A is invertible and
k=0 A = (I A) . This could also be
stated with the condition if there exists a matrix norm such that k|Ak| < 1.
13.4. Gelfands Formula for the Spectral Radius. If A Mn and k| k| is a matrix norm, then
(A) = limk k|Ak k|1/k
For the proof, consider (A)k = (Ak ) k|Ak k| for one inequality and the matrix A = ((A) +
)1 A for the other and use that the limit of Ak is 0.
14. W EYL S I NEQUALITIES AND C OROLLARIES
This is a consequence of the Courant Fischer Theorem.
Theorem Let A, B Mn be Hermitian and let the eigenvalues i (A), i (B) and i (A + B) be
arranged in increasing order (max = 1 ... n = max. For each k N we have
k (A) + 1 (B) k (A + B) k (A) + n (B)
Corollaries Many of these corollaries continue to stress intuition about the eigenvalues of Hermitian matrices. For example, the first corollary stresses that positive semidefinite matrices have
positive eigenvalues.
Assume B is positive semidefinite with the above assumptions, then k (A) k (A + B).
For a vector z Cn with the eigenvalues of A and A zz arranged in increasing order
(a) k (A zz ) k+1 (A) k+2 (A zz ) for k = 1, 2, ..., n 2
(b) k (A) k+1 (A zz ) k+2 (A) for k = 1, 2, ..., n 2.
Theorem Let A, B Mn be Hermitian and suppose that B has at most rank r. Then
(a) k (A + B) k+r (A) k+2r (A + B) k = 1, 2, ..., n 2r
(b) k (A) k+r (A + B) k+2r (A) k = 1, 2, ..., n 2r
(c) If A = U U with U = [u1 , ..., un ] unitary and = diag(1 , ..., n ) arranged in increasing
order, and if
B = n un un + n1 un1 un1 + ... + nr+1 unr+1 unr+1
then max (A B) = nr (A).
Corollary By applying the above theorem repeatedly we get: If B has at most rank r, then
kr (A) k (A + B) k+r (A)
This theorem intrinsically requires k r 1 and k + r n.
Theorem For A, B Hermitian,
(a) 1 j, k n, j + k n + 1 then
j+kn (A + B) j (A) + k (B)
(b) if j + k n + 1,
j (A) + k (B) j+k1 (A + B)
15. I NTERLACING E IGENVALUES
Let A Mn+1 be Hermitian. Define A Mn , y Cn and a R so that
A
y
A =
y a
where the eigenvalues of A and A are denoted {k } and {i } for k = 1, .., n + 1 and i = 1, .., n
(respectively) arranged in increasing order. Then
1 1 2 2 n1 n n n+1
Note: This Theorem sometimes specifies that A is Hermitian, but that is given because every principal submatrix of a Hermitian matrix is Hermitian. Further, notice a R, because the diagonal
entries of Hermitian matrices are real.
16. G ERSGORIN D ISC T HEOREM
This theorem gives us an idea of the location of the eigenvalues of a matrix. Furthermore it
allows us to say things about the eigenvalues of a matrix without computing them.
Theorem Let A = [aij ] Mn and let
n
X
Ri0 :=
|aij |, 1 i n
j=1
j6=i
denote the deleted absolute row sums of A. Then all the eigenvalues of A are located in the union
of n discs
n
[
{z C : |z aii | Ri0 } G(A)
i=1
in other words, (A) G(A). Furthermore, if a union of k of these n discs form a connected
region that is disjoint from all the remaining n k discs, then there are precisely k eigenvalues of
A in this region.
Corollary All the eigenvalues of A are located in the union of n discs
n
[
{z Cn : |z ajj | Cj0 } = G(AT )
j=1
where Cji :=
Pn
i=1
j6=i
n discs form a connected region that is disjoint from all the remaining n k discs, then there are
precisely k eigenvalues of A in this region.
Since similar matrices have the same eigenvalues, you can sometimes find a better estimate on the
location of a matrices eigenvalues by looking at a similar matrix. A convenient choice is a matrix
D = diag(p1 , ..., pn ) with pi > 0. One can then easily calculate D1 AD = [pj aij /pi ] and apply
Gersgorin to D1 AD and its transpose to yield:
Corollary All the eigenvalues of A lie in the region
n
n
[
1 X
n
z C : |z aii |
pj |aij | = G(D1 AD)
pi j=1
i=1
j6=i
n
X
1
z C : |z ajj | pj
|aij | = G[(D1 AD)T ]
p
j=1 i
n
j6=i
G(D1 AD)
where D is diagonal. Therefore you can use different Ds to attempt to trap eigenvalues by
taking intersections.
16.2. Strict Diagonal Dominance. Definition: A matrix is strictly diagonally dominant if:
n
X
|aii | >
|aij | = Ri0 for all i = 1, . . . , n
j=1
j6=i
tr(AA ) = i2
18.4. Computing the Singular Value Decomposition. For a nonsingular square matrix A Mn :
(a) Form the positive definite Hermitian matrix AA and compute the unitary diagonalization
AA = U U by finding the (positive) eigenvalues {i } of AA and a corresponding set
{ui } of normalized eigenvectors.
(b) Set = 1/2 and V = W = {u1 , ..., un }.
(c) Set W = A V 1
Notes:
If A were singular, then AA would be positive semi definite, and so would not be
invertible. Therefore it would be necessary to compute the eigenvectors of A A to find W .
The eigenvectors must be arranged in the same order as the singular values.
The singular values of a normal matrix are the absolute values of the eigenvalues. Besides
the usual way, since V, W are not unique, the columns of V can also be the eigenvectors
of A. In this case V is determined by the eigenvectors of A, but V 6= W . To find find W ,
notice for each k , k = |k |eik . So let D = diag(ei1 , ..., ein ). Then A = U U =
U ||DU = U ||(U D)
If A Mn is singular, that means it will have at least one zero singular value. When
this occurs, augment U and W as necessary by choosing orthogonal vectors to make them
square as well.
In fact, whenever you come up short, always augment U and W with orthogonal vectors
If the matrix is not square or singular, the process will not be fun.
18.5. Notes on Nonsquare matrices. When computing the singular value decomposition for nonsquare matrices, there are a few things to keep in mind:
Although you only use the nonzero eigenvalues for the singular value decomposition, you
still use the eigenvectors associated with any zero eigenvalues. AA = A A have the same
nonzero eigenvalues.
The matrix V should be the same size as AA and similarly for W and A A.
19. P OSITIVE M ATRICES AND P ERRON F ROBENIUS
1in
min
1jn
n
X
j=1
n
X
i=1
1in
1jn
n
X
aij
j=1
n
X
aij
i=1
This says (A) is between the biggest and smallest column and row sum of A.
1jn
n
X
aij
i=1
xi
(A) max xj
1jn
n
X
aij
i=1
xi
21. E XAMPLES
Having
is necessary but NOT sufficient for similarity. Consider
the same
eigenvalues
0 1
0 0
and
. These both have eigenvalue 0 with multiplicity 2, but they are
0 0
0 0
not similar.
0 1
A=
is not diagonalizable. If it were, it would be similar to the 0 matrixm which
0 0
is untrue since the only matrix similar to the 0 matrix is itself. Also, the dimKer(A0I) =
1 and so we know there is only one eigenvector associated with 0, but to be diagonalizable
there must be 2 since
A is 2x2.
0 1
Consider
and the 2x2 identity I. These two matrices commute but are not simul0 0
taneously diagonalizable. This is because the first matrix is not diagonalizable.
0 1
Consider two matrices where AB is diagonalizable but BA is not. A =
B =
0 0
1 1
0 0
3 1
1 1
and
are similar but not unitarily equivalent.
2 0
0 2
P
P
The following two matrices satisfy i,j |aij |2 = i,j |bij |2 but are not unitarily equivalent,
in fact they are not even similar!
0 1
1 0
0 0
0 0
A=
AA
1 1
0 0
and B =
0 1
0 0
1 1
A=
is a real 2-by-2 matrix that is normal but not symmetric, skew-symmetric
1 1
or orthogonal (real unitary).
1 0
3 0
A=
and B =
. A and B commute so (A + B) ((A) + (B)). But
0 2
0 4
1 + 4 = 5 ((A) + (B)) and (A + B) = {4, 6}. So (A + B) 6= ((A) + (B)) in
general.
0 1
0 0
A=
and B =
. A and B do not commute, and clearly ((A) + (B)) =
0 0
1 0
{0} but
(A +
B) = {1, 1}. So (A + B) 6 ((A) + (B)).
0 1
A=
the rank of A is bigger than the number of nonzero eigenvalues (This can
0 0
happen when A is not hermitian). Furthermore, the geometric multiplicity (1) is strictly
smaller than the number of eigenvalues counting multiplicity, and A only has one eigenvector(up to scaling).
1 i
i i
A =
and B =
are both complex and symmetric matrices, but one is
i 1
i 1
normal
and one
is not.
0 1
A=
0 0
0 0
A =
do not have the same null space. (For normal matrices they do).
1 0
1 1
A=
is positive semi definite but not positive definite.
1 1
0 0
A =
has all nonnegative leading principal minors but is not positive semi
0 1
definite (So the corresponding theorem for positive definite matrices does not apply to psd
matrices).
P
P
0 1
2
A =
is not normal but satisfies i,j ai,j = tr(A A) =
i |i (A)| which is
0 0
associated with normal matrices.
AB and BA can have the same eigenvalues and characteristic poly, but not the same minimal polynomial,
i.e.different jordan forms, i.e. not similar to each other:
1 0 0
A = 0 0 1 and
0 0 1
1 0 0
B= 0 1 0
0 0 0
For Hermitian matrices, the rank of A is equal to the number of nonzero eigenvalues. This
is not true for non-Hermitian matrices:
0 1
0 0
0 1
0 0
Weyls inequality (the first one) if A, B are not hermitian. Consider
and
0 0
1 0
Matrices with the same minimal polynomial and characteristic polynomial NEED NOT be
similar for matrices of order 4. Ex: 4x4 matrices with zeroes on the diagonals, where
one has two J blocks of order 2, and the other has only 1.