6.4 Normal and Self-Adjoint Operators

We have seen the importance of diagonalizable operators in Chapter 5. For an operator on a vector space V to be diagonalizble, it is necessary and sufficient for V to contain a basis of eigenvectors for this operator. As V is an inner product space in this chapter, it is reasonable to seek conditions that guarantee that V has an orthonormal basis of eigenvectors. A very important result that helps achieve our goal is Schur’s theorem (Theorem 6.14). The formulation that follows is in terms of linear operators. The next section contains the more familiar matrix form. We begin with a lemma.

Lemma. Let T be a linear operator on a finite-dimensional inner product space V. If T has an eigenvector, then so does T*.

Proof.

Suppose that v is an eigenvector of T with corresponding eigenvalue λ. Then for any xVxV,

0=0, x=(TλI)(v), x=v, (TλI)*(x)=v, (T*ˉλI)(x),
0=0, x=(TλI)(v), x=v, (TλI)*(x)=v, (T*λ¯I)(x),

and hence v is orthogonal to the range of T*ˉλIT*λ¯I. So T*ˉλIT*λ¯I is not onto and hence is not one-to-one. Thus T*ˉλIT*λ¯I has a nonzero null space, and any nonzero vector in this null space is an eigenvector of T* with corresponding eigenvalue ˉλλ¯.

Recall (see the exercises of Section 2.1 and see Section 5.4) that a subspace W of V is said to be T-invariant if T(W) is contained in W. If W is T-invariant, we may define the restriction TW: WWTW: WW by TW(x)=T(x)TW(x)=T(x) for all xWxW. It is clear that TWTW is a linear operator on W. Recall from Section 5.2 that a polynomial is said to split if it factors into linear polynomials.

Theorem 6.14 (Schur).

Let T be a linear operator on a finite-dimensional inner product space V. Suppose that the characteristic polynomial of T splits. Then there exists an orthonormal basis γγ for V such that the matrix [T]γ[T]γ is upper triangular.

Proof.

By Exercise 12(a) of Section 5.2, there exists an ordered basis β={w1, w2, , wn}β={w1, w2, , wn} for V such that [T]β[T]β is upper triangular. Now apply the Gram-Schmidt process to ββ to obtain an orthogonal basis β={v1, v2, , vn}β={v1, v2, , vn} for V. For each k, 1kn1kn, let

Sk={w1, w2, , wk}andSk={v1, v2, , vk}.
Sk={w1, w2, , wk}andSk={v1, v2, , vk}.

As in the proof of Theorem 6.4, span(Sk)=span(Sk)span(Sk)=span(Sk) for all k. By Exercise 12 of Section 2.2, T(wk)span(Sk)T(wk)span(Sk) for all k. Hence T(vk)span(Sk)T(vk)span(Sk) for all k, and so [T]β[T]β is upper triangular by the same exercise. Finally, let zi=1||vi||vizi=1||vi||vi for all 1in1in and γ={z1, z2, , zn}γ={z1, z2, , zn}. Then γγ is an orthonormal basis for V, and [T]γ[T]γ is upper triangular.

We now return to our original goal of finding an orthonormal basis of eigenvectors of a linear operator T on a finite-dimensional inner product space V. Note that if such an orthonormal basis ββ exists, then [T]β[T]β is a diagonal matrix, and hence [T*]β=[T]*β[T*]β=[T]*β is also a diagonal matrix. Because diagonal matrices commute, we conclude that T and T* commute. Thus if V possesses an orthonormal basis of eigenvectors of T, then TT*=T*TTT*=T*T.

Definitions.

Let V be an inner product space, and let T be a linear operator on V. We say that T is normal if TT*=T*TTT*=T*T. An n×nn×n real or complex matrix A is normal if AA*=A*AAA*=A*A.

It follows immediately from Theorem 6.10 (p. 356) that T is normal if and only if [T]β[T]β is normal, where ββ is an orthonormal basis.

Example 1

Let T: R2R2T: R2R2 be rotation by θθ, where 0<θ<π0<θ<π. The matrix representation of T in the standard ordered basis is given by

A=(cos θsin θsin θcos θ).
A=(cos θsin θsin θcos θ).

Note that AA*=I=A*AAA*=I=A*A; so A, and hence T, is normal.

Example 2

Suppose that A is a real skew-symmetric matrix; that is, At=AAt=A. Then A is normal because both AAtAAt and AtAAtA are equal to A2A2.

Clearly, the operator T in Example 1 does not even possess one eigenvector. So in the case of a real inner product space, we see that normality is not sufficient to guarantee an orthonormal basis of eigenvectors. All is not lost, however. We show that normality suffices if V is a complex inner product space.

Before we prove the promised result for normal operators, we need some general properties of normal operators.

Theorem 6.15.

Let V be an inner product space, and let T be a normal operator on V. Then the following statements are true.

  1. (a) ||T(x)||=||T*(x)||||T(x)||=||T*(x)|| for all xVxV.

  2. (b) TcITcI is normal for every cFcF.

  3. (c) If x is an eigenvector of T corresponding to eigenvalue λλ, then x is also an eigenvector of T* corresponding to eigenvalue ˉλλ¯. That is, if T(x)=λxT(x)=λx, then T*(x)=ˉλxT*(x)=λ¯x.

  4. (d) If λ1λ1 and λ2λ2 are distinct eigenvalues of T with corresponding eigenvectors x1x1 and x2x2, then x1x1 and x2x2 are orthogonal.

Proof.

Proof. (a) For any xVxV, we have

||T(x)||2=T(x), T(x)=T*T(x), x=TT*(x), x=T*(x), T*(x)=||T*(x)||2.
||T(x)||2==T(x), T(x)=T*T(x), x=TT*(x), xT*(x), T*(x)=||T*(x)||2.

The proof of (b) is left as an exercise.

(c) Suppose that T(x)=λxT(x)=λx for some xVxV. Let U=TλIU=TλI. Then U(x)=0U(x)=0, and U is normal by (b). Thus (a) implies that

0=||U(x)||=||U*(x)||=||(T*ˉλI)(x)||=||(T*(x)ˉλx||.
0=||U(x)||=||U*(x)||=||(T*λ¯I)(x)||=||(T*(x)λ¯x||.

Hence T*(x)=ˉλxT*(x)=λ¯x. So x is an eigenvector of T*.

(d) Let λ1λ1 and λ2λ2 be distinct eigenvalues of T with corresponding eigenvectors x1x1 and x2x2. Then, using (c), we have

λ1x1, x2=λ1x1, x2=T(x1), x2=x1, T*(x2)=x1, ˉλ2x2=λ2x1, x2.
λ1x1, x2==λ1x1, x2=T(x1), x2=x1, T*(x2)x1, λ¯2x2=λ2x1, x2.

Since λ1λ2λ1λ2, we conclude that x1, x2=0x1, x2=0.

Theorem 6.16.

Let T be a linear operator on a finite-dimensional complex inner product space V. Then T is normal if and only if there exists an orthonormal basis for V consisting of eigenvectors of T.

Proof.

Suppose that T is normal. By the fundamental theorem of algebra (Theorem D.4), the characteristic polynomial of T splits. So we may apply Schur’s theorem to obtain an orthonormal basis β={v1, v2, , vn}β={v1, v2, , vn} for V such that [T]β=A[T]β=A is upper triangular. We know that v1v1 is an eigenvector of T because A is upper triangular. Assume that v1, v2, , vk1v1, v2, , vk1 are eigenvectors of T. We claim that vkvk is also an eigenvector of T. It then follows by mathematical induction on k that all of the vivi’s are eigenvectors of T. Consider any j<kj<k, and let λjλj denote the eigenvalue of T corresponding to vjvj. By Theorem 6.15, T*(vj)=ˉλjvjT*(vj)=λ¯jvj. Since A is upper triangular,

T(vk)=A1kv1+A2kv2++Ajkvj++Akkvk.
T(vk)=A1kv1+A2kv2++Ajkvj++Akkvk.

Furthermore, by the corollary to Theorem 6.5 (p. 345),

Ajk=T(vk), vj=vk, T*(vj)=vk, ˉλjvj=λjvk, vj=0.
Ajk=T(vk), vj=vk, T*(vj)=vk, λ¯jvj=λjvk, vj=0.

It follows that T(vk)=AkkvkT(vk)=Akkvk, and hence vkvk is an eigenvector of T. So by induction, all the vectors in ββ are eigenvectors of T.

The converse was already proved on page 367.

Interestingly, as the next example shows, Theorem 6.16 does not extend to infinite-dimensional complex inner product spaces.

Example 3

Consider the inner product space H with the orthonormal set S from Example 9 in Section 6.1. Let V=span(S)V=span(S), and let T and U be the linear operators on V defined by T(f)=f1fT(f)=f1f and U(f)=f1fU(f)=f1f. Then

T(fn)=fn+1andU(fn)=fn1
T(fn)=fn+1andU(fn)=fn1

for all integers n. Thus

T(fm), fn=fm+1, fn=δ(m+1), n=δm, (n1)=fm, fn1=fm, U(fn).
T(fm), fn=fm+1, fn=δ(m+1), n=δm, (n1)=fm, fn1=fm, U(fn).

It follows that U=T*U=T*. Furthermore, TT*=I=T*TTT*=I=T*T; so T is normal.

We show that T has no eigenvectors. Suppose that f is an eigenvector of T, say, T(f)=λfT(f)=λf for some λλ. Since V equals the span of S, we may write

f=mi=naifi,where am0.
f=i=nmaifi,where am0.

Hence

mi=naifi+1=T(f)=λf=mi=nλaifi.
i=nmaifi+1=T(f)=λf=i=nmλaifi.

Since am0am0, we can write fm+1fm+1 as a linear combination of fn, fn+1, , fmfn, fn+1, , fm. But this is a contradiction because S is linearly independent.

Example 1 illustrates that normality is not sufficient to guarantee the existence of an orthonormal basis of eigenvectors for real inner product spaces. For real inner product spaces, we must replace normality by the stronger condition that T=T*T=T* in order to guarantee such a basis.

Definitions.

Let T be a linear operator on an inner product space V. We say that T is self-adjoint (or Hermitian) if T=T*T=T*. An n×nn×n real or complex matrix A is self-adjoint (or Hermitian) if A=A*A=A*.

It follows immediately that if ββ is an orthonormal basis, then T is self-adjoint if and only if [T]β[T]β is self-adjoint. For real matrices, this condition reduces to the requirement that A be symmetric.

Before we state our main result for self-adjoint operators, we need some preliminary work.

By definition, a linear operator on a real inner product space has only real eigenvalues. The lemma that follows shows that the same can be said for self-adjoint operators on a complex inner product space. Similarly, the characteristic polynomial of every linear operator on a complex inner product space splits, and the same is true for self-adjoint operators on a real inner product space.

Lemma. Let T be a self-adjoint operator on a finite-dimensional inner product space V. Then

  1. (a) Every eigenvalue of T is real.

  2. (b) Suppose that V is a real inner product space. Then the characteristic polynomial of T splits.

Proof.

(a) Suppose that T(x)=λxT(x)=λx for x0x0. Because a self-adjoint operator is also normal, we can apply Theorem 6.15(c) to obtain

λx=T(x)=T*(x)=ˉλx.
λx=T(x)=T*(x)=λ¯x.

So λ=ˉλλ=λ¯ that is, λλ is real.

(b) Let n=dim(V), βn=dim(V), β be an orthonormal basis for V, and A=[T]βA=[T]β. Then A is self-adjoint. Let TATA be the linear operator on CnCn defined by TA(x)=AxTA(x)=Ax for all xCnxCn. Note that TATA is self-adjoint because [TA]γ=A[TA]γ=A, where γγ is the standard ordered (orthonormal) basis for CnCn. So, by (a), the eigenvalues of TATA are real. By the fundamental theorem of algebra, the characteristic polynomial of TATA splits into factors of the form tλtλ. Since each λλ is real, the characteristic polynomial splits over R. But TATA has the same characteristic polynomial as A, which has the same characteristic polynomial as T. Therefore the characteristic polynomial of T splits.

We are now able to establish one of the major results of this chapter.

Theorem 6.17.

Let T be a linear operator on a finite-dimensional real inner product space V. Then T is self-adjoint if and only if there exists an orthonormal basis ββ for V consisting of eigenvectors of T.

Proof.

Suppose that T is self-adjoint. By the lemma, we may apply Schur’s theorem to obtain an orthonormal basis ββ for V such that the matrix A=[T]βA=[T]β is upper triangular. But

A*=[T]*β=[T*]β=[T]β=A.
A*=[T]*β=[T*]β=[T]β=A.

So A and A* are both upper triangular, and therefore A is a diagonal matrix. Thus ββ must consist of eigenvectors of T.

The converse is left as an exercise.

We restate this theorem in matrix form in the next section (as Theorem 6.20 on p. 381).

Example 4

As we noted earlier, real symmetric matrices are self-adjoint, and self-adjoint matrices are normal. The following matrix A is complex and symmetric:

A=(iii1)andA*=(iii1).
A=(iii1)andA*=(iii1).

But A is not normal, because (AA*)12=1+i(AA*)12=1+i and (A*A)12=1i(A*A)12=1i. Therefore complex symmetric matrices need not be normal.

Exercises

  1. Label the following statements as true or false. Assume that the underlying inner product spaces are finite-dimensional.

    1. (a) Every self-adjoint operator is normal.

    2. (b) Operators and their adjoints have the same eigenvectors.

    3. (c) If T is an operator on an inner product space V, then T is normal if and only if [T]β[T]β is normal, where ββ is any ordered basis for V.

    4. (d) A real or complex matrix A is normal if and only if LALA is normal.

    5. (e) The eigenvalues of a self-adjoint operator must all be real.

    6. (f) The identity and zero operators are self-adjoint.

    7. (g) Every normal operator is diagonalizable.

    8. (h) Every self-adjoint operator is diagonalizable.

  2. For each linear operator T on an inner product space V, determine whether T is normal, self-adjoint, or neither. If possible, produce an orthonormal basis of eigenvectors of T for V and list the corresponding eigenvalues.

    1. (a) V=R2V=R2 and T is defined by T(a, b)=(2a2b, 2a+5b)T(a, b)=(2a2b, 2a+5b).

    2. (b) V=R3V=R3 and T is defined by T(a, b, c)=(a+b, 5b, 4a2b+5c)T(a, b, c)=(a+b, 5b, 4a2b+5c).

    3. (c) V=C2V=C2 and T is defined by T(a, b)=(2a+ib, a+2b)T(a, b)=(2a+ib, a+2b).

    4. (d) V=P2(R)V=P2(R) and T is defined by T(f)=fT(f)=f, where

      f(x), g(x)=10f(t)g(t) dt.
      f(x), g(x)=10f(t)g(t) dt.
    5. (e) V=M2×2(R)V=M2×2(R) and T is defined by T(A)=AtT(A)=At.

    6. (f) V=M2×2(R)V=M2×2(R) and T is defined by T(abcd)=(cdab)T(acbd)=(cadb).

  3. Give an example of a linear operator T on R2R2 and an ordered basis for R2R2 that provides a counterexample to the statement in Exercise 1(c).

  4. Let T and U be self-adjoint operators on an inner product space V. Prove that TU is self-adjoint if and only if TU=UTTU=UT.

  5. Prove (b) of Theorem 6.15.

  6. Let V be a complex inner product space, and let T be a linear operator on V. Define

    T1=12(T+T*)andT2=12i(TT*).
    T1=12(T+T*)andT2=12i(TT*).
    1. (a) Prove that T1T1 and T2T2 are self-adjoint and that T=T1+iT2T=T1+iT2.

    2. (b) Suppose also that T=U1+iU2T=U1+iU2, where U1U1 and U2U2 are self-adjoint. Prove that U1=T1U1=T1 and U2=T2U2=T2.

    3. (c) Prove that T is normal if and only if T1T2=T2T1T1T2=T2T1.

  7. Let T be a linear operator on an inner product space V, and let W be a T-invariant subspace of V. Prove the following results.

    1. (a) If T is self-adjoint, then TWTW is self-adjoint.

    2. (b) WW is T*-invariant.

    3. (c) If W is both T- and T*-invariant, then (TW)*=(T*)W(TW)*=(T*)W.

    4. (d) If W is both T- and T*-invariant and T is normal, then TWTW is normal.

  8. Let T be a normal operator on a finite-dimensional complex inner product space V, and let W be a subspace of V. Prove that if W is T-invariant, then W is also T*-invariant. Hint: Use Exercise 24 of Section 5.4.

  9. Let T be a normal operator on a finite-dimensional inner product space V. Prove that N(T)=N(T*)N(T)=N(T*) and R(T)=R(T*)R(T)=R(T*). Hint: Use Theorem 6.15 and Exercise 12 of Section 6.3.

  10. Let T be a self-adjoint operator on a finite-dimensional inner product space V. Prove that for all xVxV

    ||T(x)±ix||2=||T(x)||2+||x||2.
    ||T(x)±ix||2=||T(x)||2+||x||2.

    Deduce that TiITiI is invertible and that the adjoint of (TiI)1(TiI)1 is (TiI)1(TiI)1.

  11. Assume that T is a linear operator on a complex (not necessarily finite-dimensional) inner product space V with an adjoint T*. Prove the following results.

    1. (a) If T is self-adjoint, then T(x), xT(x), x is real for all xVxV.

    2. (b) If T satisfies T(x), x=0T(x), x=0 for all xVxV, then T=T0T=T0. Hint: Replace x by x+yx+y and then by x+iyx+iy, and expand the resulting inner products.

    3. (c) If T(x), xT(x), x is real for all xVxV, then T is self-adjoint.

  12. Let T be a normal operator on a finite-dimensional real inner product space V whose characteristic polynomial splits. Prove that V has an orthonormal basis of eigenvectors of T. Hence prove that T is self-adjoint.

  13. An n×nn×n real matrix A is said to be a Gramian matrix if there exists a real (square) matrix B such that A=BtBA=BtB. Prove that A is a Gramian matrix if and only if A is symmetric and all of its eigenvalues are non-negative. Hint: Apply Theorem 6.17 to T=LAT=LA to obtain an orthonormal basis {v1, v2, , vn}{v1, v2, , vn} of eigenvectors with the associated eigenvalues λ1, λ2, , λnλ1, λ2, , λn. Define the linear operator U by U(vi)=λiviU(vi)=λivi.

  14. Simultaneous Diagonalization. Let V be a finite-dimensional real inner product space, and let U and T be self-adjoint linear operators on V such that UT=TUUT=TU. Prove that there exists an orthonormal basis for V consisting of vectors that are eigenvectors of both U and T. (The complex version of this result appears as Exercise 10 of Section 6.6.) Hint: For any eigenspace W=EλW=Eλ of T, we have that W is both T- and U-invariant. By Exercise 7, we have that WW is both T- and U-invariant. Apply Theorem 6.17 and Theorem 6.6 (p. 347).

  15. Let A and B be symmetric n×nn×n matrices such that AB=BAAB=BA. Use Exercise 14 to prove that there exists an orthogonal matrix P such that PtAPPtAP and PtBPPtBP are both diagonal matrices.

  16. Prove the Cayley-Hamilton theorem for a complex n×nn×n matrix A. That is, if f(t) is the characteristic polynomial of A, prove that f(A)=Of(A)=O. Hint: Use Schur’s theorem to show that A may be assumed to be upper triangular, in which case

    f(t)=ni=1(Aiit).
    f(t)=i=1n(Aiit).

    Now if T=LAT=LA, we have (AjjIT)(ej)span({e1, e2, , ej1})(AjjIT)(ej)span({e1, e2, , ej1}) for j2j2, where {e1, e2, , en}{e1, e2, , en} is the standard ordered basis for CnCn. (The general case is proved in Section 5.4.)

The following definitions are used in Exercises 17 through 23.

Definitions.

A linear operator T on a finite-dimensional inner product space is called positive definite [positive semidefinite] if T is self-adjoint and T(x), x>0 [T(x), x0]T(x), x>0 [T(x), x0] for all x0x0.

An n×nn×n matrix A with entries from R or C is called positive definite [positive semidefinite] if LALA is positive definite [positive semidefinite].

  1. Let T and U be self-adjoint linear operators on an n-dimensional inner product space V, and let A=[T]βA=[T]β, where ββ is an orthonormal basis for V. Prove the following results.

    1. (a) T is positive definite [semidefinite] if and only if all of its eigenvalues are positive [nonnegative].

    2. (b) T is positive definite if and only if

      i, jAijajˉai>0 for all nonzero n-tuples (a1, a2, , an).
      i, jAijaja¯i>0 for all nonzero n-tuples (a1, a2, , an).
    3. (c) T is positive semidefinite if and only if A=B*BA=B*B for some square matrix B.

    4. (d) If T and U are positive semidefinite operators such that T2=U2T2=U2, then T=UT=U.

    5. (e) If T and U are positive definite operators such that TU=UTTU=UT, then TU is positive definite.

    6. (f) T is positive definite [semidefinite] if and only if A is positive definite [semidefinite].

    Because of (f), results analogous to items (a) through (d) hold for matrices as well as operators.

  2. Let T: VWT: VW be a linear transformation, where V and W are finite-dimensional inner product spaces. Prove the following results.

    1. (a) T*T and TT* are positive semidefinite. (See Exercise 15 of Section 6.3.)

    2. (b) rank(T*T)=rank(TT*)=rank(T).rank(T*T)=rank(TT*)=rank(T).

  3. Let T and U be positive definite operators on an inner product space V. Prove the following results.

    1. (a) T+UT+U is positive definite.

    2. (b) If c>0c>0, then cT is positive definite.

    3. (c) T1T1 is positive definite.

    Visit goo.gl/cQch7i for a solution.

  4. Let V be an inner product space with inner product , , , and let T be a positive definite linear operator on V. Prove that x, y=T(x), y defines another inner product on V.

  5. Let V be a finite-dimensional inner product space, and let T and U be self-adjoint operators on V such that T is positive definite. Prove that both TU and UT are diagonalizable linear operators that have only real eigenvalues. Hint: Show that UT is self-adjoint with respect to the inner product x, y=T(x), y. To show that TU is self-adjoint, repeat the argument with T1 in place of T.

  6. This exercise provides a converse to Exercise 20. Let V be a finite-dimensional inner product space with inner product , , and let ,  be any other inner product on V.

    1. (a) Prove that there exists a unique linear operator T on V such that x, y=T(x), y for all x and y in V. Hint: Let β={v1, v2, , vn} be an orthonormal basis for V with respect to , , and define a matrix A by Aij=vj, vi for all i and j. Let T be the unique linear operator on V such that [T]β=A.

    2. (b) Prove that the operator T of (a) is positive definite with respect to both inner products.

  7. Let U be a diagonalizable linear operator on a finite-dimensional inner product space V such that all of the eigenvalues of U are real. Prove that there exist positive definite linear operators T1 and T1 and self-adjoint linear operators T2 and T2 such that U=T2T1=T1T2. Hint: Let ,  be the inner product associated with V, β a basis of eigenvectors for U, ,  the inner product on V with respect to which β is orthonormal (see Exercise 22(a) of Section 6.1), and T1 the positive definite operator according to Exercise 22. Show that U is self-adjoint with respect to ,  and U=T11U*T1 (the adjoint is with respect to , ). Let T2=T11U*.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.144.36.141