A motivation of adjoint operators
In the real inner product space Rn with the standard inner product ⟨v,w⟩=⟨[a1a2⋮an],[b1b2⋮bn]⟩=a1b1+a2b2+⋯+anbn=wtv,
However, in the complex inner product space Cn with the standard inner product ⟨v,w⟩=⟨[w1w2⋮wn],[z1z2⋮zn]⟩=w1¯z1+w2¯z2+⋯+wn¯zn=¯wtv,
In most of standard textbooks, like Friedberg, Insel and Spence's Linear Algebra, they gave a lemma first (Theorem 6.8, Riesz Representation Theorem), then proved the existence of T∗ by using the lemma (Theorem 6.9). However, the lemma seems to be clever for novices.
I am going to define adjoint of operators through adjoint of matrices here. A brief version see Treil's Linear Algebra Done Wrong (Section 5.5).
Let V be a finite dimensional inner product space V and let T be a linear operator on V. Given an "orthonormal" basis β for V, (the existence of such basis follows immediately from the Gram-Schmidt process,) we have the following commutative diagram VT⟶Vϕβ↓↓ϕβFn⟶[T]βFn
Note that we have (Theorem 2.14) T(v)=ϕ−1β([T]βϕβ(v)).
Obviously, T∗ is a linear operator. We show that ⟨T(v),w⟩=⟨v,T∗(w)⟩
Therefore, ⟨v,T∗(w)⟩=⟨v,ϕ−1β([T]∗βϕβ(w))⟩Lemma=⟨ϕβ(v),[T]∗βϕβ(w)⟩=⟨[T]βϕβ(v),ϕβ(w)⟩Theorem 2.14=⟨ϕβ(T(v)),ϕβ(w)⟩Lemma=⟨T(v),w⟩.
Remark. We defined T∗ by choosing a specific orthonormal basis. But the definition of T∗ is independent on the choice of the orthonormal basis since ⟨[v]β,[w]β⟩=⟨[v]α,[w]α⟩
Proof: Suppose that α={v1,v2,...,vn} and β={w1,w2,...,wn} are two orthonormal bases for Cn. Then [v]β=[I]βα[v]α. Note that ([I]βα)ij=⟨vj,wi⟩
As I mentioned before, Theorem 6.8 and Theorem 6.9 are tricky. But they appear in many exams. So you still have to know how to prove them. I come up with a method to help you remember these theorems.
We will need a very basic but important theorem (Theorem 6.5).
Theorem 6.5. Let V be a nonzero finite-dimensional inner product space. Then V has an orthonormal basis β. Furthermore, if β={v1,v2,...,vn} and x∈V}, then x=n∑i=1⟨x,vi⟩vi.
Now, let us discuss Theorem 6.8 and Theorem 6.9.
Theorem 6.8. Let V be a finite-dimensional inner product space over F, and let g:V→F be a linear transformation. Then there exists a unique vector y∈V such that g(x)=⟨x,y⟩ for all x∈V.
I am used to using the following identities to remind me the form of y. Let {v1,v2,...,vn} be an orthonormal basis for Fn. ⟨x,y⟩=⟨⟨x,v1⟩v1+⋯+⟨x,vn⟩vn,⟨y,v1⟩v1+⋯+⟨y,vn⟩vn⟩=⟨x,v1⟩¯⟨y,v1⟩+⋯+⟨x,vn⟩¯⟨y,vn⟩g(x)=g(⟨x,v1⟩v1+⋯+⟨x,vn⟩vn)=⟨x,v1⟩g(v1)+⋯+⟨x,vn⟩g(vn)
Theorem 6.9. Let V be a finite-dimensional inner product space, and let T be a linear operator on V. Then there exists a unique function T∗:V→V such that ⟨T(x),y⟩=⟨x,T∗(y)⟩ for all x,y∈V. Furthermore, T∗ is linear.
For applying the lemma, we have to construct a linear functional first and define T∗(v) to be the y induced by the lemma. VV↓?=↓⟨⋅,y⟩FF
No comments:
Post a Comment