6.6 Orthogonal Projections and the Spectral Theorem

In this section, we rely heavily on Theorems 6.16 (p. 369) and 6.17 (p. 371) to develop an elegant representation of a normal (if F=CF=C) or a self-adjoint (if F=RF=R) operator T on a finite-dimensional inner product space. We prove that T can be written in the form λ1T1+λ2T2++λkTkλ1T1+λ2T2++λkTk, where λ1, λ2, , λkλ1, λ2, , λk are the distinct eigenvalues of T T1, T2, , TkT1, T2, , Tk are orthogonal projections. We must first develop some results about these special projections.

We assume that the reader is familiar with the results about direct sums developed at the end of Section 5.2. The special case where V is a direct sum of two subspaces is considered in the exercises of Section 1.3.

Recall from the exercises of Section 2.1 that if V=W1W2V=W1W2, then a linear operator T on V is the projection on W1W1 along W2W2 if, whenever x=x1+x2x=x1+x2, with x1W1x1W1 and x2W2x2W2, we have T(x)=x1T(x)=x1. By Exercise 27 of Section 2.1, we have

R(T)=W1={xV: T(x)=x}andN(T)=W2.
R(T)=W1={xV: T(x)=x}andN(T)=W2.

So V=R(T)N(T)V=R(T)N(T). Thus there is no ambiguity if we refer to T as a “projection on W1W1” or simply as a “projection.” In fact, it can be shown (see Exercise 17 of Section 2.3) that T is a projection if and only if T=T2T=T2. Because V=W1W2=W1W3V=W1W2=W1W3 does not imply that W2=W3W2=W3, we see that W1W1 does not uniquely determine T. For an orthogonal projection T, however, T is uniquely determined by its range.

Definition.

Let V be an inner product space, and let T: VVT: VV be a projection. We say that T is an orthogonal projection if R(T)=N(T)R(T)=N(T) and N(T)=R(T)N(T)=R(T).

Note that by Exercise 13(c) of Section 6.2, if V is finite-dimensional, we need only assume that one of the equalities in this definition holds. For example, if R(T)=N(T)R(T)=N(T), then R(T)=R(T)=N(T)R(T)=R(T)=N(T).

An orthogonal projection is not the same as an orthogonal operator. In Figure 6.5, T is an orthogonal projection, but T is clearly not an orthogonal operator because ||T(v)||||v||||T(v)||||v||.

A diagram of the graph of 3 vectors and 2 triangles.

Figure 6.5

Now assume that W is a finite-dimensional subspace of an inner product space V. In the notation of Theorem 6.6 (p. 347), we can define a function T: VVT: VV by T(y)=uT(y)=u. It is easy to show that T is an orthogonal projection on W. We can say even more—there exists exactly one orthogonal projection on W. For if T and U are orthogonal projections on W, then R(T)=W=R(U)R(T)=W=R(U). Hence N(T)=R(T)=R(U)=N(U)N(T)=R(T)=R(U)=N(U), and since every projection is uniquely determined by its range and null space, we have T=UT=U. We call T the orthogonal projection of V on W.

To understand the geometric difference between an arbitrary projection on W and the orthogonal projection on W, let V=R2V=R2 and W=span{(1, 1)}W=span{(1, 1)}. Define U and T as in Figure 6.5, where T(v) is the foot of a perpendicular from v on the line y=xy=x and U(a1, a2)=(a1, a1)U(a1, a2)=(a1, a1). Then T is the orthogonal projection of V on W, and U is a different projection on W. Note that vT(v)WvT(v)W, whereas vU(v)WvU(v)W. From Figure 6.5, we see that T(v) is the “best approximation in W to v”; that is, if wWwW, then ||wv||||T(v)v||||wv||||T(v)v||. In fact, this approximation property characterizes T. These results follow immediately from the corollary to Theorem 6.6 (p. 348).

As an application to Fourier analysis, recall the inner product space H and the orthonormal set S in Example 9 of Section 6.1. Define a trigonometric polynomial of degree n to be a function gHgH of the form

g(t)=nj=najfj(t)=nj=najeijt,
g(t)=j=nnajfj(t)=j=nnajeijt,

where anan or anan is nonzero.

Let fHfH. We show that the best approximation to f by a trigonometric polynomial of degree less than or equal to n is the trigonometric polynomial whose coefficients are the Fourier coefficients of f relative to the orthonormal set S. For this result, let W=span({fj: |j|n})W=span({fj: |j|n}), and let T be the orthogonal projection of H on W. The corollary to Theorem 6.6 (p. 348) tells us that the best approximation to f by a function in W is

T(f)=nj=nf,  fjfj.
T(f)=j=nnf,  fjfj.

For an application of this material to electronic music, visit goo.gl/EN5Fai.

An algebraic characterization of orthogonal projections follows in the next theorem.

Theorem 6.24.

Let V be an inner product space, and let T be a linear operator on V. Then T is an orthogonal projection if and only if T has an adjoint T* and T2=T=T*T2=T=T*.

Proof.

Suppose that T is an orthogonal projection. Since T2=TT2=T because T is a projection, we need only show that T* exists and T=T*T=T*. Now V=R(T)N(T)V=R(T)N(T) and R(T)=N(T)R(T)=N(T). Let x, yVx, yV. Then we can write x=x1+x2x=x1+x2 and y=y1+y2y=y1+y2, where x1, y1R(T)x1, y1R(T) and x2, y2N(T)x2, y2N(T). Hence

x, T(y)=x1+x2, y1=x1, y1+x2, y1=x1, y1
x, T(y)=x1+x2, y1=x1, y1+x2, y1=x1, y1

and

T(x), y=x1, y1+y2=x1, y1+x1, y2=x1, y1.
T(x), y=x1, y1+y2=x1, y1+x1, y2=x1, y1.

So x, T(y)=T(x), yx, T(y)=T(x), y for all x, yVx, yV thus T* exists and T=T*T=T*.

Now suppose that T2=T=T*T2=T=T*. Then T is a projection by Exercise 17 of Section 2.3, and hence we must show that R(T)=N(T)R(T)=N(T) and R(T)=N(T)R(T)=N(T). Let xR(T)xR(T) and yN(T)yN(T). Then x=T(x)=T*(x)x=T(x)=T*(x), and so

x, y=T*(x), y=x, T(y)=x, 0=0.
x, y=T*(x), y=x, T(y)=x, 0=0.

Therefore xN(T)xN(T), from which it follows that R(T)N(T)R(T)N(T).

Let yN(T)yN(T). We must show that yR(T)yR(T), that is, T(y)=yT(y)=y. Now

||yT(y)||2=yT(y), yT(y)=y, yT(y)T(y), yT(y).
||yT(y)||2==yT(y), yT(y)y, yT(y)T(y), yT(y).

Since yT(y)N(T)yT(y)N(T), the first term must equal zero. But also

T(y), yT(y)=y, T*(yT(y))=y, T(yT(y))=y, 0=0.
T(y), yT(y)=y, T*(yT(y))=y, T(yT(y))=y, 0=0.

Thus yT(y)=0yT(y)=0; that is, y=T(y)R(T)y=T(y)R(T). Hence R(T)=N(T)R(T)=N(T).

Using the preceding results, we have R(T)=N(T)N(T)R(T)=N(T)N(T) by Exercise 13(b) of Section 6.2. Now suppose that xR(T)xR(T). For any yVyV, we have T(x), y=x, T*(y)=x, T(y)=0T(x), y=x, T*(y)=x, T(y)=0. So T(x)=0T(x)=0, and thus xN(T)xN(T). Hence R(T)=N(T)R(T)=N(T).

Let V be a finite-dimensional inner product space, W be a subspace of V, and T be the orthogonal projection of V on W. We may choose an orthonormal basis β={v1, v2, , vn}β={v1, v2, , vn} for V such that {v1, v2, , vk}{v1, v2, , vk} is a basis for W. Then [T]β[T]β is a diagonal matrix with ones as the first k diagonal entries and zeros elsewhere. In fact, [T]β[T]β has the form

(IkO1O2O3).
(IkO2O1O3).

If U is any projection on W, we may choose a basis γγ for V such that [U]γ[U]γ has the form above; however γγ is not necessarily orthonormal.

We are now ready for the principal theorem of this section.

Theorem 6.25 (The Spectral Theorem).

Suppose that T is a linear operator on a finite-dimensional inner product space V over F with the distinct eigenvalues λ1, λ2, , λkλ1, λ2, , λk. Assume that T is normal if F=CF=C and that T is self-adjoint if F=RF=R. For each i(1ik)i(1ik), let WiWi be the eigenspace of T corresponding to the eigenvalue λiλi, and let TiTi be the orthogonal projection of V on WiWi. Then the following statements are true.

  1. (a) V=W1W2Wk.V=W1W2Wk.

  2. (b) If WiWi denotes the direct sum of the subspaces WjWj for jiji, then Wi=WiWi=Wi.

  3. (c) TiTj=δijTiTiTj=δijTi for 1i, jk.1i, jk.

  4. (d) I=T1+T2++Tk.I=T1+T2++Tk.

  5. (e) T=λ1T1+λ2T2++λkTk.T=λ1T1+λ2T2++λkTk.

Proof.

(a) By Theorems 6.16 (p. 369) and 6.17 (p. 371), T is diagonalizable;

so

V=W1W2Wk
V=W1W2Wk

by Theorem 5.10 (p. 277).

(b) If xWixWi and yWjyWj for some ijij, then x, y=0x, y=0 by Theorem 6.15(d) (p. 368). It follows easily from this result that WiWiWiWi. From (a), we have

dim(Wi)=jidim(Wj)=dim(V)dim(Wi).
dim(Wi)=jidim(Wj)=dim(V)dim(Wi).

On the other hand, we have dim(Wi)=dim(V)dim(Wi)dim(Wi)=dim(V)dim(Wi) by Theorem 6.7(c) (p. 349). Hence Wi=WiWi=Wi, proving (b).

(c) The proof of (c) is left as an exercise.

(d) Since TiTi is the orthogonal projection of V on WiWi, it follows from (b) that N(Ti)=R(Ti)=Wi=WiN(Ti)=R(Ti)=Wi=Wi. Hence, for xVxV, we have x=x1+x2++xkx=x1+x2++xk, where Ti(x)=xiWiTi(x)=xiWi, proving (d).

(e) For xVxV, write x=x1+x2++xkx=x1+x2++xk, where xiWixiWi. Then

T(x)=T(x1)+T(x2)++T(xk)=λ1x1+λ2x2++λkxk=λ1T1(x)+λ2T2(x)++λkTk(x)=(λ1T1+λ2T2++λkTk)(x).
T(x)====T(x1)+T(x2)++T(xk)λ1x1+λ2x2++λkxkλ1T1(x)+λ2T2(x)++λkTk(x)(λ1T1+λ2T2++λkTk)(x).

The set {λ1, λ2, , λk}{λ1, λ2, , λk} of eigenvalues of T is called the spectrum of T, the sum I=T1+T2++TkI=T1+T2++Tk in (d) is called the resolution of the identity operator induced by T, and the sum T=λ1T1+λ2T2++λkTkT=λ1T1+λ2T2++λkTk in (e) is called the spectral decomposition of T. The spectral decomposition of T is unique up to the order of its eigenvalues.

With the preceding notation, let ββ be the union of orthonormal bases of the WiWi’s and let mi=dim(Wi)mi=dim(Wi). (Thus mimi is the multiplicity of λiλi.) Then [T]β[T]β has the form

(λ1Im1OOOλ2Im2OOOλkImk);
λ1Im1OOOλ2Im2OOOλkImk;

that is, [T]β[T]β is a diagonal matrix in which the diagonal entries are the eigenvalues λiλi of T, and each λiλi is repeated mimi times. If λ1T1+λ2T2++λkTkλ1T1+λ2T2++λkTk is the spectral decomposition of T, then it follows (from Exercise 7) that g(T)=g(λ1)T1+g(λ2)T2++g(λk)Tkg(T)=g(λ1)T1+g(λ2)T2++g(λk)Tk for any polynomial g. This fact is used below.

We now list several interesting corollaries of the spectral theorem; many more results are found in the exercises. For what follows, we assume that T is a linear operator on a finite-dimensional inner product space V over F.

Corollary 1.

If F=CF=C then T is normal if and only if T*=g(T)T*=g(T) for some polynomial g.

Proof.

Suppose first that T is normal. Let T=λ1T1+λ2T2++λkTkT=λ1T1+λ2T2++λkTk be the spectral decomposition of T. Taking the adjoint of both sides of the preceding equation, we have T*=ˉλ1T1+ˉλ2T2++ˉλkTkT*=λ¯1T1+λ¯2T2++λ¯kTk since each TiTi is self-adjoint. Using the Lagrange interpolation formula (see page 53), we may choose a polynomial g such that g(λi)=ˉλig(λi)=λ¯i for 1ik1ik. Then

g(T)=g(λ1)T1+g(λ2)T2++g(λk)Tk=ˉλ1T1+ˉλ2T2++ˉλkTk=T*.
g(T)=g(λ1)T1+g(λ2)T2++g(λk)Tk=λ¯1T1+λ¯2T2++λ¯kTk=T*.

Conversely, if T*=g(T)T*=g(T) for some polynomial g, then T commutes with T* since T commutes with every polynomial in T. So T is normal.

Corollary 2.

If F=CF=C, then T is unitary if and only if T is normal and |λ|=1|λ|=1 for every eigenvalue λλ of T.

Proof.

If T is unitary, then T is normal and every eigenvalue of T has absolute value 1 by Corollary 2 to Theorem 6.18 (p. 379).

Let T=λ1T1+λ2T2++λkTkT=λ1T1+λ2T2++λkTk be the spectral decomposition of T. If |λ|=1|λ|=1 for every eigenvalue λλ of T, then by (c) of the spectral theorem,

TT*=(λ1T1+λ2T2++λkTk)(ˉλ1T1+ˉλ2T2++ˉλkTk)=|λ1|2T1+|λ2|2T2++|λk|2Tk=T1+T2++Tk=I.
TT*====(λ1T1+λ2T2++λkTk)(λ¯1T1+λ¯2T2++λ¯kTk)|λ1|2T1+|λ2|2T2++|λk|2TkT1+T2++TkI.

Hence T is unitary.

Corollary 3.

If F=CF=C, then T is self-adjoint if and only if T is normal and every eigenvalue of T is real.

Proof.

Suppose that T is normal and that its eigenvalues are real. Let T=λ1T1+λ2T2++λkTkT=λ1T1+λ2T2++λkTk be the spectral decomposition of T. Then

T*=ˉλ1T1+ˉλ2T2++ˉλkTk=λ1T1+λ2T2++λkTk=T.
T*=λ¯1T1+λ¯2T2++λ¯kTk=λ1T1+λ2T2++λkTk=T.

Hence T is self-adjoint.

Now suppose that T is self-adjoint and hence normal. That its eigenvalues are real has been proved in the lemma to Theorem 6.17 (p. 371).

Corollary 4.

Let T be as in the spectral theorem with spectral decomposition T=λ1T1+λ2T2++λkTkT=λ1T1+λ2T2++λkTk. Then each TjTj is a polynomial in T.

Proof.

Choose a polynomial gj(1jk)gj(1jk) such that gj(λi)=δijgj(λi)=δij. Then

gj(T)=gj(λ1)T1+gj(λ2)T2++gj(λk)Tk=δ1jT1+δ2jT2++δkjTk=Tj.
gj(T)==gj(λ1)T1+gj(λ2)T2++gj(λk)Tkδ1jT1+δ2jT2++δkjTk=Tj.

Exercises

  1. Label the following statements as true or false. Assume that the underlying inner product spaces are finite-dimensional.

    1. (a) All projections are self-adjoint.

    2. (b) An orthogonal projection is uniquely determined by its range.

    3. (c) Every self-adjoint operator is a linear combination of orthogonal projections.

    4. (d) If T is a projection on W, then T(x) is the vector in W that is closest to x.

    5. (e) Every orthogonal projection is a unitary operator.

  2. Let V=R2, W=span({(1, 2)})V=R2, W=span({(1, 2)}), and ββ be the standard ordered basis for V. Compute [T]β[T]β, where T is the orthogonal projection of V on W. Do the same for V=R3V=R3 and W=span({(1, 0, 1)})W=span({(1, 0, 1)}).

  3. For each of the matrices A in Exercise 2 of Section 6.5:

    1. (1) Verify that LALA possesses a spectral decomposition.

    2. (2) For each eigenvalue of LALA, explicitly define the orthogonal projection on the corresponding eigenspace.

    3. (3) Verify your results using the spectral theorem.

  4. Let W be a finite-dimensional subspace of an inner product space V. Show that if T is the orthogonal projection of V on W, then ITIT is the orthogonal projection of V on WW.

  5. Let T be a linear operator on a finite-dimensional inner product space V.

    1. (a) If T is an orthogonal projection, prove that ||T(x)||||x||||T(x)||||x|| for all xVxV. Give an example of a projection for which this inequality does not hold. What can be concluded about a projection for which the inequality is actually an equality for all xVxV?

    2. (b) Suppose that T is a projection such that ||T(x)||||x||||T(x)||||x|| for all xVxV. Prove that T is an orthogonal projection.

  6. Let T be a normal operator on a finite-dimensional inner product space. Prove that if T is a projection, then T is also an orthogonal projection.

  7. Let T be a normal operator on a finite-dimensional complex inner product space V. Use the spectral decomposition λ1T1+λ2T2++λkTkλ1T1+λ2T2++λkTk of T to prove the following results.

    1. (a) If g is a polynomial, then

      g(T)=ki=1g(λi)Ti.
      g(T)=i=1kg(λi)Ti.
    2. (b) If Tn=T0Tn=T0 for some n, then T=T0T=T0.

    3. (c) Let U be a linear operator on V. Then U commutes with T if and only if U commutes with each TiTi.

    4. (d) There exists a normal operator U on V such that U2=TU2=T.

    5. (e) T is invertible if and only if λi0λi0 for 1ik1ik.

    6. (f) T is a projection if and only if every eigenvalue of T is 1 or 0.

    7. (g) T=T*T=T* if and only if every λiλi is an imaginary number.

  8. Use Corollary 1 of the spectral theorem to show that if T is a normal operator on a complex finite-dimensional inner product space and U is a linear operator that commutes with T, then U commutes with T*.

  9. Referring to Exercise 20 of Section 6.5, prove the following facts about a partial isometry U.

    1. (a) U*U is an orthogonal projection on W.

    2. (b) UU*U=U.UU*U=U.

  10. Simultaneous diagonalization. Let U and T be normal operators on a finite-dimensional complex inner product space V such that TU=UTTU=UT. Prove that there exists an orthonormal basis for V consisting of vectors that are eigenvectors of both T and U. Hint: Use the hint of Exercise 14 of Section 6.4 along with Exercise 8.

  11. Prove (c) of the spectral theorem. Visit goo.gl/utQ9Pb for a solution.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset