Académique Documents
Professionnel Documents
Culture Documents
+|Bq/?
Hence a* A* Aa =0 implies B; =0, for i=1,2,...,m; or. equivalently, Aa =0,
which, in turn, implies # =0 from the assumption of p(A)=n. Therefore we
conclude that p(A*A) =n.
2. This part can be similarly proved or directly deduced from the foregoing by
using the fact p(A) = p(A*). QED.
2-6 Eigenvectors, Generalized Eigenvectors, and Jordan-
Form Representations of a Linear Operator
With the background of Section 2-5, we are now ready to study the problem
posed at the end of Section 2-4. We discuss in this section only linear operators
that map (C", C) into itself with the understanding that the results are applicable
to any operator that maps a finite-dimensional linear space over C into itself.
The reason for restricting the field to the field of complex numbers will be seen
immediately.
Let A be an n xn matrix with coefficients in the field C. We have agreed to
consider A as a linear operator that maps (C”, C) into (C", C).
Definition 2-12
Let A be a linear operator that maps (C”, €) into itself. Then a scalar A in C
is called an eigenvalue of A if there exists a nonzero vector x in C" such that
Ax =4x. Any nonzero vector x satisfying Ax = Ax is called an eigenvector of A
associated with the eigenvalue /."' 1
In order to find an eigenvalue of A, we write Ax = Ax as
(A—ADx =0 (2-40)
where I is the unit matrix of order. We see that for any fixed 4 in C, Equation
(2-40) is a set of homogeneous linear equations. The matrix (A — AI) is an
nxn square matrix. From Corollary 2-5, we know that Equation (2-40) has a
nontrivial solution if and only if det (A —AI)=0. It follows that a scalar 2 is
an eigenvalue of A if and only if it isa solution of A(2) & det (AL—A)=0. A(A)
is a polynomial of degree n in 4 and is called the characteristic polynomial of A.
Since A(A) is of degree n, the n x n matrix A has n eigenvalues (not necessarily
all distinct),
'* [is also called a right eigenvector of A. Ifa 1 xn nonzero vector y exists such that yA = Ay, then
y is called a lefi eigenvector of A associated with 2.Re LINEAR SPACES AND LINEAR OPERATORS,
Example 1
Consider the matrix
1-1
a-[) “i (2-41)
which maps (IR?, ) into itself. We like to check whether Definition 2-12 can
be modified and applied to a linear operator that maps (R", R) into (R", R).
A modified version of Definition 2-12 reads as a scalar / in R is an eigenvalue of
A if there exists a nonzero vector x such that Ax = 4x. Clearly / is an eigenvalue
of A if and only if it is a solution of det (AI—A)=0. Now
; a-1 1 >
det (AI =A)=aet| A al=4 +1
which has no real-valued solution. Consequently, the matrix A has no eigen-
value in R.
Since the set of real numbers is a part of the field of complex numbers, there
is no reason that we cannot consider the matrix A in (2-41) as a linear operator
that maps (C?: C) into itself. In so doing, then the matrix A has eigenvalues
+iand —i where i? A -1. .
The constant matrices we shall encounter in this book are all real-valued.
However in order to ensure the existence of eigenvalues, we shall consider them
as linear operators under the field of complex numbers.
With these preliminaries, we are ready to introduce a set of basis vectors
such that a linear operator has a diagonal or almost diagonal representation.
We study first the case in which all the eigenvalues of A are distinct; the case
where A has repeated eigenvalues will then be studied.
Case 1: All the eigenvalues of A are distinct
Let 4,,43,...,4, be the eigenvalues of A, and let v, be an eigenvector of A
associated with 2,, for i=1,2,...,n; that is, Av; =4,¥;. We shall use the set of
vectors {V;,V2,...,¥,} aS a basis of (C", C). In order to do so, we have to show
that the set is linearly independent and qualifies as a basis.
Theorem 2-9
Let A,,42,...,4, be the distinct eigenvalues of A, and let v; be an eigenvector
of A associated with 4;, for i=1,2,...,n. Then the set {v,,¥2,...,¥,} is linearly
independent (over C).
Proof
We prove the theorem by contradiction. Suppose v,,¥5,...,¥, are linearly
dependent; then there exist «,,02,..., a, (not all zero) in € such that
OV, HQV2°°* +0G,V,=0 (2-42)
We assume a, #0. If «,=0, we may reorder A; in such a way that «, #0.REPRESENTATIONS OF A LINEAR OPERATOR 35
Equation (2-42) implies that
(A=ADMA 214200 ( F xm)=a (2-43)
Since (A-ADy=Gi— ay iti Fi
and (A-AD =
the left-hand side of (2-43) can be reduced to
(Ay —AgMAy Aa) Ay — AW =O
By assumption the 1's, for i=1,2,...,n, are all distinct; hence the equation
a; [] (41 —A)v, =0
implies 2, =0. This isa contradiction. Thus, the set of vectors {V¥1,V2,--..¥,}
Vn}
is linearly independent and qualifies as a basis. QE.D.
Let A be the representation of A with respect to the basis {¥,.¥2,....¥,}.
Recall from Figure 2-5 that a ith column of A is the representation of Av; =
Aw; with respect to {V,, V2, nj—that is.[0 --- 0 4; O --: OJ’, where
4, is located at the ith entry. “ence the representation of A with respect to
Wi, Vope es Voh iS
A, 0 0 0
, [o 4 0 0
A=|0
O ase 0 (2-44)
This can also be checked by using a similarity transformation. From
Figure 2-5, we have
Q= In v2 7 vs]
Since
Yoo) WJ=[Av, Av, «17 Avy]
dave dg] = QA
we have A=Q-'AQ
We conclude that if the eigenvalues of a linear operator A that maps (C”, C)
into itself are all distinct, then by choosing the set of eigenvectors as a basis.
the operator A has a diagonal matrix representation with the eigenvalues on
the diagonal.
Example 2
Consider36 LINEAR SPACES AND LINEAR OPERATORS
The characteristic polynomial of A is 4? +1. Hence the eigenvalues of A are
+iand —i. The eigenvector associated with 2, =i can be obtained by solving
the following homogeneous equation:
(A-A,Dy, -[' a = liz |-9
21
Clearly the vector v; =[1 1 —J]' is a solution. Similarly, the vector v, =
{t 1 + i] can be shown to be an eigenvector of 4, = —i. Hence the repre-
sentation of A with respect to {¥,, ¥2} is
ale |
The reader is advised to verify this by a similarity transformation. .
Case 2: The eigenvalues of A are not all distinct
Unlike the previous case, if an operator A has repeated eigenvalues, it is not
always possible to find a diagonal matrix representation. We shall use examples
to illustrate the difficulty that may arise for matrices with repeated eigenvalues.
Example 3
Consider
1 Oe!
A=|0 1 0
0 0 2
The eigenvalues of A are 1, = 1,4, =1,and4,;=2. The eigenvectors associated
with 4, can be obtained by solving the following homogeneous equations:
0 0 -1
(A-ADv=|0 0 O|v=0 (2-45)
0 Oo 2
Note that the matrix (A —/,]) has rank 1; therefore, two linearly independent
vector solutions can be fourid-foi' (2-43) (see Corollary 2-5). Clearly, v, =
[1 0 OJ andv,=[0 1 OJ’ are two linearly independent eigenvectors
associated with 14,=A,;=1. An eigenvector associated with 2,=2 can be
found as v;=[—-1 0 1]. Since the set of vectors {v,,V2,V} is linearly
independent, it qualifies as a basis. The representation of A with respect to
{¥1, V2. Va} is
100
A=|0 1 0
002
In this example, although A has repeated eigenvalues, it can still be diag-
onalized. However, this is not always the case, as can be seen from the following
example.REPRESENTATIONS OF A LINEAR OPERATOR 7
1 12
A=|0 1 3 (2-46)
0 0 2
The eigenvalues of Aare 2, =1, 4, =1,and 4,=2. The eigenvectors associated
with 2, =1 can be found by solving
012
(A—A,Dv=|0 0 3|v=0
001
Since the matrix (A — 4,1) has rank 2, the null space of (A — 4,1) has dimension 1.
Consequently, we can find only one linearly independent eigenvector, say
¥,=[1 0 OY, associated with 2;=/,=1. An eigenvector associated with
4, =2 can be found as v;=[5 3 1]. Clearly the two eigenvectors are not
~ sufficient to form a basis of (C3, C). 1
Example 4
Consider
From this example, we see that if an n xn matrix A has repeated eigen-
values, it is not always possible to find n linearly independent eigenvectors.
Consequently, the A cannot be transformed into a diagonal form. However,
it is possible to find a special set of basis vectors so that the new representation
is almost a diagonal form, called a Jordan canonical form. The form has the
eigenvalues of A on the diagonal and either 0 or 1 on the superdiagonal. For
example, if A has an eigenvalue 2, with multiplicity 4 and an eigenvalue />
with multiplicity 1, then the new representation will assume one of the following
(2-47)
Which form it will assume depends on the characteristics of A and will be
pdiscussed in the next subsection. The matrices in (2-47) are all of block-diagonal38 LINEAR SPACES AND LINEAR OPERATORS,
form. The blocks on the diagonal are of the form
A100 0
O41 00
000-10 ae
000-41
000-04
with the same eigenvalue on the main diagonal and I’s on the diagonal just
above the main diagonal. A matrix of this form is called a Jordan block associ-
ated with 4. A matrix is said to be in the Jordan canonical form, or the Jordan
Jorm, if its principal diagonal consists of Jordan blocks and the remaining
elements are zeros. The fourth matrix in (2-47) has two Jordan blocks associated
with 4, (one with order 3, the other with order 1) and one Jordan block associ-
ated with 4,. A diagonal matrix is clearly a special case of the Jordan form: all
of its Jordan blocks are of order 1.
Every matrix which maps (C”, C) into itself has a Jordan-form representa-
tion. The use of Jordan form is very convenient in developing a number of
concepts and results; hence it will be extensively used in the remainder of this
chapter. D
Derivation of a Jordan-form representation.'? In this subsection,
we discuss how to find a set of basis vectors so that the representation of A with
respect to this set of basis vectors is in a Jordan form. The basis vectors to be
used are called the generalized eigenvectors.
Definition 2-13
A vector v is said to be a generalized eigenvector of grade k of A associated with
A if and only if!*
(A—AD‘'v =0
and (A-ADK'v #0 :
Note that if k = 1, Definition 2-13 reduces to (A — Aly =0 and v #0, which
is the definition of an eigenvector. Hence the term “generalized eigenvector”
is well-justified.
Let v be a generalized eigenvector of grade k associated with the eigenvalue
'? This section may be skipped without loss of continuity. However, it is suggested that the reader
glances through it to gain a better feeling about the Jordan-form representation.
'3(A— 21 Q (A AKA — AD): (A — AD) terms), (A= 21) 21REPRESENTATIONS OF A LINEAR OPERATOR 39
4. Define
(2-43)
and
This set of vectors {v,,¥>....,¥q} is called a chain of generalized eigenvectors
of length k.
Let V7; denote the null space of (A — Al)’, that is, 4’; consists of all x such
that (A —AIx =0. It is clear that ifx is in .;, then it is in 7,.,. Hence V;
isa subspace of V;.,.denoted as Wjc W;,,. Clearly, the v defined in Defini-
tion 2-13 is in .V, but not in .,_;. In fact, for /=1,2,....k.¥; =(A— ADK 'v
defined in (2-49) is in V7; but not in V;_,. Indeed, we have
(A —AD'v, = (A —AD(A — ADE “'v =(A — ADV =0
and (A AD!" 'v, = (A = AD (A — AV =(A — ADV £0
hence vy, is in 7; but not in W;-y
Let A be an n xn matrix and have eigenvalue 4 with multiplicity m. We
discuss in the following how to find m linearly independent generalized eigen-
vectors of A associated with 2. This is achieved by searching chains of gener-
alized eigenvectors of various lengths. First we compute ranks of (A —AD)',
i=0,1,2,..., until rank (A — AI m. In order not to be overwhelmed by
notations, we assume n = 10,m = 8, k =4, and the ranks of (A — AI)‘, i =0, 1,2, 3,4,
are as shown in Table 2-2. The nullity v; is the dimension of the null space 1’;,
and is equal to, following -Theorem 2-5, n—p(A—Al). Because of
Moe Nic N20-**. we have 0 =v) SV; Sv) °° Se=m.
Now we are ready to find m=8 linearly independent eigenvectors of A
associated with 4. Because of 3c 44 and v4—v, =1, we can find one and
only one linearly independent vector u in 1", but not in 1’; such that
Bu=0 and BuO
where BA A—AI From this u, we can generate a chain of four generalized
Table 2-2 Chains of Generalized Eigenvectors. where B 2 A — Al
No. of independent
vectors in 4; but
not in _,
hte chains with
length 2
vi %o
v, = By
One chain with
length 440 LINEAR SPACES AND LINEAR OPERATORS
eigenvectors as
u, =B*u u, =B’u u; =Bu (2-50)
Because of Wc A’; and v;—v, =1, there is only one linearly independent
vector in M3 but not in 4. The uy in (2-50) is such a vector; therefore we
cannot find any other linearly independent vector in 43 but not in 4. Con-
sider now the vectors in 4 but not in W,. Because v,—v, =3, there are
three such linearly independent vectors. Since uy in (2-50) is one of them, we
can find two vectors v and w such that {u, v, w} are linearly independent and
By-0 Bv40
and Bw=0 Bw#0
From y¥ and w, we can generate two chains of generalized eigenvectors of length
2.as shown in Table 2-2. As can be seen from Table 2-2, the number of vectors
in NW, is equal to v, — vo =3, hence there is no need to search other vector in
WN. This completes the search of eight generalized eigenvectors of A associated
with 4.
Theorem 2-10
The generalized eigenvectors of A associated with A generated as in Table 2-2
are linearly independent.
Proof
First we show that if {u;, ¥, w} is linearly independent, then {u,, v,, w;} is linearly
independent. Suppose {u,,¥;,W,} is not linearly independent, then there
exist c;,i=1,2,3, not all zero, such that cyu; +¢2¥; +caw: =0. However,
we have
O0=cyu, +¢3¥, +¢,W, =c, Bu, +¢,Bv +c,Bw =Bic,u, +c,v +c,w) 4 By
Since y isa vector in A>, the only way to have By =Ois that y Since {u,, v, w}
is linearly independent by assumption, y=0 implies c,=0,i=1,2,3. This
is a contradiction. Hence if {u, v, w} is linearly independent, so is {u,, V1, W;}-
Now we show that the generalized eigenvectors {u;, i=1,2,3,4; ¥;, Wj =1,
2} are linearly independent. Consider
ed Fez +e3U3 +eqly +es¥, +c6V2 +c7W, +cgW2 =0 (2-51)
The application of B? =(A —2I)? to (2-51) yields
c,B°u,=0
which implies, because of B*u, #0,c4=0. Similarly, we can show c,=0 by
applying B? to (2-51). With cs =c4 =0, the application of B to (2-51) yields
2g +c6V. +cegw, =0REPRESENTATIONS OF A LINEAR OPERATOR 41
which implies, because of the linear independence of {uy, V2, W2}, C2 = C5 = Cg =0.
Finally, we have c, =cs =c, =0 following the linear independence of {u,,¥,, W1}-
This completes the proof of this theorem QED.
Theorem 2-11
The generalized eigenvectors of A associated with different eigenvalues are
linearly independent. '
This theorem can be proved as in Theorem 2-10 by applying repetitively
(A—AJ)(A ~A,D). The proof is left as an exercise.
Now we discuss the representation of A with respect to QO [u, wu) a5
Uv, V2 Ww; Wo X x]. The last two vectors are the eigenvectors of A
associated with other eigenvalues. The first four columns of the new representa-
tion A are the representations of At , 2, 3,4, with respect to {u,, WU, U3, U4,
Vi, V2,Wi,W2,X,X}. Because (A =0, (A —ADu, =u,, (A —ADu; =u,, and
(A — Alu, =u5. we have Au, = Au; “op 0 0 + Of, Au, =a, +42, =
Qti 4.0 0 +. Oy, Aus=u,+4u,=Qf0 1 4 0 --- OY, and
Au, =u; +du,=Q[0 0 t 2 0 --- OF, where the prime denotes the
transpose. Proceeding similarly, the new representation A can be obtained as
24100:0000
0410:'0000
0041:0000
, $9 0 0 419 000
A=|0 00 OFA 110 6 (2-82)
000 0:0 430 0
o000 00
00000 0:0 :!
This isa Jordan-form matrix, Note that the number of Jordan blocks associated
with 2 is equal to v, =3, the dimension of the null space of (A — AI)."*
Example 4 (Continued)
Consider
>
I
oo
one
Nw
*6This number is called the geometric multiplicity of 4 in Reference 86. In other words, geometric
multiplicity is the number of Jordan blocks, and the (algebraic) multiplicity is the sum of the
orders of all Jordan blocks associated with A.a2 LINEAR SPACES AND LINEAR OPERATORS
Its eigenvalues are 4; =1,4,=1,4;=2. An eigenvector associated with
23=2isv,=[5 3 1]. The rank of (A —A,I) is 2; hence we can find only one
eigenvector associated with A,. Consequently, we must use generalized eigen-
vectors. We compute
012
BA(A-AD=|0 0 3
oot
0 1 2Wfo 1 2) foo s
and (A-A,D?=/0 0 3]/0 0 3/=10 0 3
0 6 1fo 0 tf foo 1
Since pB?=1=n—m, we stop here. We search a v such that B’v=0 and
By #0. Ciearty,7=[0 1 OJ'issucha vector. It is a generalized eigenvector
of grade 2. Let
0] 0 1 2)fo] fi
v,Av=}1 vy, S(A-A v=] 0 allil=lo
4 - Lo o ijlo} Lo
Theorems 2-10 and 2-11 imply that v,,v2, and vs are linearly independent.
This can also be checked by computing the determinant of [¥, vz v3]. If
we use the set of vectors {¥,,¥2,¥3} as a basis, then the ith column of the new
representation A is the representation of Av; with respect to the basis {¥,, V2, ¥3}-
Since Av; =A,¥;,Av,=V, +4,¥,, and Av,=A3¥3, the representations of
Av,, Ay,, and Av, with respect to the basis {v,,¥2,¥3} are, respectively,
EH) Gl
where 2, =1,4;=2. Hence we have
, fi to
A=|0 110] (2-53)
This can also be obtained by using the similarity transformation
A=Q-'AQ
where Q=f), wh
oon
ono
RON
oe
-REPRESENTATIONS OF A LINEAR OPERATOR, 43
Example 5
Transform the following matrix into the Jordan form:
3-1 1 1 0 0
1 1-1 -1 0 0
{6 0 2 0 1 1 ,
A=lo 0 0 2 -1 -1 eae
0 0 0 0 1 41
o 0 0 0 1 1
1. Compute the eigenvalues of A.'*
det (A —Al) =[(3 — AI — 2) +1] — 2° [ — 2? - 1] =
Hence A has eigenvalue 2 with multiplicity 5 and samatue 0 with multi-
plicity 1.
2. Compute (A — 21), for i= 1,2,...,as follows:
‘ 1-1 1 1 0 0
1-1 -1 -1 0 0
/ 0 0 0 0 1 1 (A — 21) -
4(A-21= pl
BSA-W=)9 9 9 0 -1 -1 vy, =6-4=2
0 0 0 oO -1 1
0 0 0 0 1 -1
0 0 2 2 0 6
0 0 2 2 0 0
0 0 0 0 0 0 (A — 21)? =
(A — 21)? = pl
: Y=l0 6 0 0 0 0 v2 =4
0 0 0 0 2 -2
0 0 0 0-2 2
0 0 0 0 0 0
0 0 0 0 0 0
_yp_|9 0 0 0 0 0 p(A — 21) =
A =19 9 0 0 0 OO] a
0 0 0 0-4 4
0 0 0 0 4 ~4
Since p(A —21)* =n —m =1, we stop here. Because v; —v =1, we can find
a gerieralized eigenvector u of grade 3 such that B*u =0 and Bu #0. It is
45 We use the fact that
A BI
det =det A det C
0 Cc
where A and C are square matrices, not necessarily of the same order.44 LINEAR SPACES AND LINEAR OPERATORS
easy to verify thatu=[0 0 1 0 O OJ is sucha vector. Define
2 1 0
2 -1 0
0 0 1
a, 2Bu= 0 u, 4 Bu= a u,; Qu= 0
0 0 0
0 0. 0.
This is a chain of generalized eigenvectors of length 3. Because of v, — vy, =2,
there are two linearly independent vectors in AW but not in .W”,. The vector
u, is one of them. We search a vector v which is independent of #, and has
the property B?v=0 and Bv#0. It can be readily verified that v=
fo oO 1 1 1] is such a vector. Define
Now we have found five generalized eigenvectors of A associated with 1 = 2.
3. Compute an eigenvector associated with 2, =0. Let w be an eigenvector
of A associated with 4, =0; then
Clearly, w=(0 0 0 0 1 ~1Jisasolution.
4. With respect to the basis {u,,U2,U,¥;.¥2, W},A has the following Jordan-
form representation:
(2-55)
5. This may be checked by using
A=Q°'AQ or QA=AQFUNCTIONS OF A SQUARE MATRIX 45
where uoY¥, Vv. wi
ocoorce
mm cocoo
0
0
1
lll
1
1
oconnoso
In this example, if we reorder the basis {u,, U2, U3, ¥,,V¥2,w} and use
{W, Vo, Vy, U3, U2, Hy} as a new basis, then the representation will be
(2-56)
This is also called a Jordan-form representation. Comparing it with Equation
(2-55), we see that the new Jordan block in (2-56) has 1’s on the diagonal just
below the main diagonal as a result of the different ordering of the basis vectors.
In this book, we use mostly the Jordan block of the form in (2-55). Certainly
everything discussed for this form can be modified and be applied to the form
given in Equation (2-56).
A Jordan-form representation of any linear operator A that maps (C”, C)
into itself is unique up to the ordering of Jordan blocks. That is, the number of
Jordan blocks and the order of each Jordan block are uniquely determined
by A. However, because of different orderings of basis vectors, we may have
different Jordan-form representations of the same matrix.
2-7 Functions of a Square Matrix
In this section we shall study functions of a square matrix or a linear transforma-
tion that maps (C", C) into itself. We shall use the Jordan-form representation
extensively, because in terms of this representation almost all properties of a
function of a matrix can be visualized. We study first polynomials of a square
matrix, and then define functions of a matrix in terms of polynomials of the
matrix.
Polynomials of a square matrix. Let A be a square matrix that maps
(C*, C) into (C”, C). Mf k is a positive integer, we define
A‘ QAA---A — (k terms) (2-57a)
and ACSI (2-57b)46 LINEAR SPACES AND LINEAR OPERATORS:
where I is a unit matrix. Let f(A) be a polynomial in A of finite degree; then
f(A) can be defined in terms of (2-57). For example, if f(4)=43 +22? +6, then
s(A) & A? +24? +61
We have shown in the preceding section that every square matrix A that
maps (C’, C) into itself has a Jordan-form representation, or equivalently, there
exists a nonsingular constant matrix Q such that A=QAQ™' with A in a
Jordan canonical form. Since
AF =(QAQ™'YQAQ~!)---(QAQ™) = QA'Q™
we have ’
SIA=QFAQ™? or (A)=Q"' FAQ (2-58)
for any polynomial f(A).
One of the reasons to use the Jordan-form matrix is that if
_fA, 0
a-[4 2 (2-59)
where A, and A, are square matrices, then
_[fay 0
faye 0 jay]
This can be easily verified by observing that
a yat °
0 AS
The minimal polynomial of a matrix A is the monic polynomial'® (A) of least
degree such that y(A) =0. '
(2-60)
Definition 2-14
Note that the 0 in y(A)=0 is an n x n square matrix whose entries are all
zero. A direct consequence of (2-58) is that f(A) =0 if and only if f(A)=0.
Consequently, the matrices A and A have the same minimal polynomial, or
more generally, similar matrices have the same minimal polynomial. Computing
the minimal polynomial of a matrix is generally not a simple job; however, if
the Jordan-form representation of the matrix is available, its minimal poly-
nomial can be readily found.
Let A,,42,....4m be the distinct eigenvalues of A with multiplicities
Nj, N,-.-,MNm, respectively. It is the same as saying that the characteristic
‘© 4 monic polynomial is a polynomial the coefficient of whose highest power is 1. For example,
3x +1 and — x? +2x +4 are not monic polynomials, but x? —4x +7 is.FUNCTIONS OF A SQUARE MATRIX 41
polynomial of A is
(2-61)
A, 0 0
Az|® Ar 7 (2-62)
0 0 A,
where the n, x n; matrix A, denotes all the Jordan blocks associated with 1;
Definition 2-15
The largest order of the Jordan blocks associated with A; in A-is called the
index of A; in A. 1
The multiplicity of A; is denoted by n;, the index of A; is denoted by #. For
the matrix in (2-52), n, =8,/; =4; for the matrix in (2-53), n, =f, =2,n.=
fiz =1; for the matrix in (2-55), n, =5,f,=3,n2=f,=1. It is clear that
i.
ni
Theorem 2-12
The minimal polynomial of A is
WA = [T] a-
i=
where ji; is the index of 4; in A.
Proof
Since the matrices A and A have the same minimal polynomial, it is the same
as showing that (A) is the polynomial with least degtee such that y(A)=0.
We first show that the minimal polynomial of A; is pi(4)=(4 —4,)". Suppose
A, consists of r Jordan blocks associated with J; Then
Av=diag (Ai, Aig,..., As)
Wan) Q@ -- 0
and wiAj=| 9 Wan) 0
0 0 yay)
(2-63)48 LINEAR SPACES AND LINEAR OPERATORS
If the matrix (Ajj ~ A;l) has dimension n;;, then we have
010+ 0
a 001-0
(Ajy-AD=}i bt i (2-64a)
(nxn) [0 0 0 + 1
000+: 6
0010 0
oo0o0t- 0
A. -ayeli i i:
(Ai;— Ad) 0000: 1 (2-64b)
0000-0
0000 0.
000+ 01
(Ay—aaet =] 9 2-0 0 (2-64)
000.00
and (A,—2)=0 — for any integer k>n,; (2-64)
By definition, i, is the largest order of the Jordan blocks in A,, or equivalently,
=max (nj, i it). Hence (Ay— 24D =0 for j=1,2,....r. Con-
peer V(A))=0. Itis easy to see from (2-63) and (2-64) that if W(4) =(4—2)'
with either % #4, ork