Hilbert Spaces
Hilbert Spaces
Hilbert Spaces
Verner Vlačić
Sternwartstrasse 7
CH-8092 Zürich
Mathematics of Information
Hilbert spaces and linear operators
These notes are based on [1, Chap. 6, Chap. 8], [2, Chap. 1], [3, Chap. 4], [4, App. A] and [5].
1 Vector spaces
Let us consider a field F, which can be, for example, the field of real numbers R or the field of
complex numbers C. A vector space over F is a set whose elements are called vectors and in
which two operations, addition and multiplication by any of the elements of the field F (referred
to as scalars), are defined with some algebraic properties. More precisely, we have
Definition 1 (Vector space). A set X together with two operations (+, ·) is a vector space over
F if the following properties are satisfied:
(ii) there exists an element 0 ∈ X , called the zero vector, such that x + 0 = x for all x ∈ X
(iii) for every x ∈ X , there exists an element in X , denoted by −x, such that x + (−x) = 0.
• 1 · x = x for all x ∈ X
• α · (β · x) = (αβ) · x for all α, β ∈ F and x ∈ X (associativity for scalar multiplication)
• (α+β)·x = α·x+β ·x for all α, β ∈ F and x ∈ X (distributivity of scalar multiplication
with respect to field addition)
• α · (x + y) = α · x + α · y for all α ∈ F and x, y ∈ X (distributivity of scalar multiplication
with respect to vector addition).
We refer to X as a real vector space when F = R and as a complex vector space when F = C.
3. The set F[x] of all polynomials with coefficients in F is a vector space over F.
6. If X is an arbitrary set and Y an arbitrary vector space over F, the set F(X , Y)
of all functions X → Y is a vector space over F under pointwise addition and
multiplication.
A subspace of a vector space X over F is a subset of X which is itself a vector space over F
(with respect to the same operations). One can easily verify that a subset Y of X is a subspace
using the following proposition.
Proposition 1.1. Let X be a vector space over F and Y a nonempty subset of X . Then Y is a
subspace of X if it contains 0 and if it is stable under linear combinations, that is,
α·y+β·z ∈Y
3. The set `p (Z), p ∈ [1, ∞], of all complex-valued sequence {uk }k∈Z which satisfy
P+∞ p
k=−∞ |uk | < ∞, if p ∈ [1, ∞),
supk∈Z |uk | < ∞, if p = ∞,
is a subspace of CZ .
6. If (S, Σ, µ) is a measure space, we let Lp (S, Σ, µ) denote the space of all mea-
surable functions mapping S to C whose absolute value raised to the p-power is
µ-integrable, that is,
Z
p p
L (S, Σ, µ) = f : S → C measurable : |f | dµ < ∞ ,
S
2
Theorem 1 (Intersection of vector spaces). The intersection of any collection of subspaces of a
vector space X over F is again a subspace of X .
Definition 2. If S and T are linear subspaces of a vector space X with S ∩ T = {0}, then we
define the direct sum S ⊕ T by
S ⊕ T = {x + y | x ∈ S, y ∈ T }.
Note that the union of two subspaces is in general not a subspace. Take for example X = R2 .
The lines xR and yR, with x = (1, 0) and y = (0, 1), are both subspaces of R2 , but xR ∪ yR is a
not subspace, given that x + y = (1, 1) ∈
/ xR ∪ yR.
Definition 3 (Subspace spanned by S). Let S be a (possibly infinite) subset of a vector space
X over F. The subspace spanned by S is the intersection of all subspaces containing S. It is the
smallest subspace containing S. It is denoted by span(S) and may be written the set of all finite
combinations of elements of S, that is,
( k )
X
span(S) , λ` x` : k ∈ N, x` ∈ S, λ` ∈ F .
`=1
3. In F[x], the subspace spanned by {1, x, . . . , xN } is the space FN [x] of all polyno-
mials whose degree is less than N .
Definition 4 (Linear independence). Let X be a vector space over F. A finite set {x1 , x2 , . . . , xN }
of vectors of X is said to be linearly independent if for every λ1 , λ2 , . . . , λN ∈ F, the equality
λ1 x1 + λ2 x2 + . . . + λN xN = 0
Examples. 1. Any set of vectors containing the zero vector is linearly dependent.
4. The set of trigonometric functions {1, cos(x), sin(x), cos2 (x), cos(x) sin(x), sin2 (x)}
is linearly dependent.
1
Note, however, that there exist uncountable sets (e.g., the Cantor set) of Lebesgue measure zero.
3
Definition 5 (Dimension). A vector space X is N -dimensional if there exists N linearly
independent vectors in X and any N + 1 vectors in X are linearly dependent.
Definition 7 (Basis). A basis3 of a vector space X over F is a set of linearly independent and
spanning vectors of X .
Examples. 1. The set {ek }3k=1 , where e1 = [1 0 0]T , e2 = [0 1 0]T , and e3 = [0 0 1]T ,
forms a basis for R3 .
Theorem 2. Let X be a finite dimensional vector space. Any set of linearly independent vectors
can be extended to a basis of X .
Definition 8 (Norm). Let X be a vector space over F. A norm on X is a function which maps
X to R and satisfies the following properties:
(i) for all x ∈ X , we have kxk > 0 and kxk = 0 implies x = 0 (positivity)
(ii) for all x ∈ X and for all α ∈ F, we have kαxk = |α| kxk (homogeneity)
Definition 9. A vector space X over F is a normed vector space (or a pre-Banach space) if it is
equipped with a norm k·k.
3
This type of basis is sometimes referred to as an algebraic basis or a Hamel basis to highlight the difference
with Schauder bases in infinite-dimensional Banach spaces, and orthonormal bases in infinite-dimensional Hilbert
space. In finite dimensions, all these concepts coincide (cf. notes on bases in infinite-dimensional spaces for more
details).
4
Examples. 1. We can provide the space C 0 [a, b] with the norm
Z b
kf k = |f (t)|dt.
a
3. If we write u = {uk }k∈Z , then the following defines a norm on `p (Z), p ∈ [1, ∞],
Definition 10 (Inner product). Let X be a vector space over F. An inner product (or scalar
product) over F is a function X × X → F such that, for all x, y ∈ X and α ∈ F, the following
holds
(i) hy, xi = hx, yi (symmetry)
(ii) hαx, yi = α hx, yi (sesquilinearity if F = C, bilinearity if F = R)
(iii) hx + y, zi = hx, zi + hy, zi (sesquilinearity if F = C, bilinearity if F = R)
(iv) hx, xi ∈ R and hx, xi > 0 (positivity)
(v) hx, xi = 0 implies that x = 0 (positivity).
The inner product on a vector space X induces a norm on X . But note that not all norms
come from an inner product (see Problem 3 of Homework 1).
p
Theorem 3. Let X be a vector space. If h·, ·i is an inner product on X , then, kxk = hx, xi is
a norm on X fulfilling the parallelogram law:
On the other hand, every norm kxk on X fulfilling the parallelogram law is the induced norm of
exactly one inner product, which is defined as
1
hx, yi = (kx + yk2 − kx − yk2 ), F = R (1)
4
1
hx, yi = (kx + yk2 − kx − yk2 + i kx + iyk2 − i kx − iyk2 ), F = C. (2)
4
Formulas (1) and (2) are called polarization formulas.
5
Definition 11 (Inner product space). A vector space X over F is an inner product space or a
pre-Hilbert space if it is equipped with an inner product h·, ·i.
In what follows, we will always assume that for an inner product space, k · k is the norm
induced by the corresponding inner product.
4. The space C[0, 1] of all continuous functions on [0, 1] can be equipped with the
inner product Z 1
hf, gi = f (x)g(x)dx.
0
6
Theorem 6 (Inner product is continuous). Let X be an inner product space. Then, h·, ·i is
continuous, i.e., if two sequences {xn }n∈N and {yn }n∈N of elements of X converge to x ∈ X and
y ∈ X , respectively, then hxn , yn i converges to hx, yi.
Definition 13 (Cauchy sequence). Let X be a normed space equipped with the norm k·k. A
sequence {xn }n∈N of elements of X is called a Cauchy sequence if for every ε > 0 there exists an
integer N (ε) such that kxn − xm k 6 ε whenever n, m > N (ε).
Definition 17. A subset S ⊆ X of a normed space X is called a closed set if it contains all
limit points, i.e., if {xn }n∈N is a sequence of points xn ∈ S with limn→∞ xn = x ∈ X , then x ∈ S.
We write S for the smallest closed subset of X containing S in the following sense: if A is any
other closed subset of X such that A ⊃ S, then A ⊃ S. We say S is the closure of S in X .
Examples. 1. Consider R2 with norm k·k2 . The open disk D = {x ∈ R2 | kxk2 < 1}
is not a closed set (consider the sequence {(1 − 1/n, 0)}n∈N ). The closure of D
is the closed disk D = {x ∈ R2 | kxk2 6 1}.
3. To say that a set S is closed only makes sense relative to the space X that S is
a subset of. To see this, let X1 = C[0, 1] be the space of all continuous functions
from [0, 1] to C with the inner product (3), and let X2 = L2 [0, 1] be the space of
all square-integrable functions f : [0, 1] → C with the same inner product. Let
S = {f ∈ C[0, 1] : f (x) = 0 for x ∈ [0, 12 ]} and note that S is a subset of both X1
and X2 . However, S is closed in X1 , but it is not closed in X2 . Indeed, suppose
7
{fn }n∈N is a sequence in S that converges to some f ∈ X1 = C[0, 1]. Then
1 1
Z
2
Z
2
Z 1
2 2
|f (x)| dx = |f (x)−fn (x)| dx 6 |f (x)−fn (x)|2 dx = kf −fn k2 → 0,
0 0 0
4. Since the concept of closedness depends on the ambient space, so does the concept
of closure. In the above example, the closure of S in X1 is the set S itself, whereas
the closure of S in X2 is the set S = {f ∈ L2 [0, 1] : f (x) = 0 for x ∈ [0, 12 ]} ) S.
Theorem 8 (Completion of normed space). Let X be a normed space with norm k·k. Then,
there exists a Banach space Xb with norm k·k
c such that the following properties hold.
• X is a subspace of Xb;
• On X , k·k
c = k·k;
• Xb is the closure of X .
Theorem 9 (Completion of inner product spaces). Let X be an inner product space with
inner product h·, ·i . Then, there exists a Hilbert space Xb with inner product h·,
d·i such that the
following properties hold.
• X is a subspace of Xb;
• On X , h·,
d·i = h·, ·i;
• Xb is the closure of X .
8
Examples. 1. The set of rational numbers Q is an inner product space with inner
product equal to pq for p, q ∈ Q. This space is not complete.
2. The set of real numbers R is a Hilbert space space with inner product equal to
rs for r, s ∈ R.
3. CN is a Hilbert space.
Definition 20. Let S be a nonempty subset of the inner product space X . We define the
orthogonal complement S ⊥ of S as S ⊥ = {x ∈ X : x ⊥ S}.
Lemma 1. Let S be a nonempty subset of the inner product space X . Then S ⊥ is a closed
linear subspace of X .
Proof. Follows from the linearity and continuity of the inner product.
The following theorem expresses one of the fundamental geometrical properties of Hilbert
spaces. While the result may appear obvious (see Figure 1), the proof is not trivial.
kx − yk = min kx − zk.
z∈S
(b) The point y ∈ S closest to x ∈ X is the unique element of S such that the “error” (x−y) ∈ S ⊥ .
9
x
S
Figure 1: y ∈ S is the point in the closed subspace S closest to x and the unique point such that
the “error” (x − y) ∈ S ⊥ .
Lemma 2. Let S be a subset of a Hilbert space X . Then, the following properties hold.
(a) S ⊥ = span(S)⊥ ;
(b) S ⊥⊥ = span(S);
Proof. • Proof of (a): span(S)⊥ ⊆ S ⊥ follows from S ⊆ span(S). It remains to show that
S ⊆ span(S)⊥ , which follows from the linearity and continuity of the inner product.
⊥
Corollary 4.1 and Lemma 2 tell us that subspaces of Hilbert spaces behave analogously to
subspaces of finite-dimensional vector spaces familiar to us. In these results the completeness of
X is crucial, as illustrated by the following example:
Example.
Let X = C[0, 1] equipped with the inner product
Z 1
hf, gi = f (x)g(x)dx.
0
As we have already seen, X is an inner product space, but not a Hilbert space. Let
S = {f ∈ C[0, 1] : f (x) = 0 for x ∈ [0, 12 ]} and note that this is a closed subspace of
X . Its orthogonal complement is S ⊥ = {f ∈ C[0, 1] : f (x) = 0 for x ∈ [ 12 , 1]}, but
X 6= S ⊕ S ⊥ .
10
5 Orthonormal bases in Hilbert spaces
Definition 21 (Orthogonal/orthonormal set). A subset S of nonzero vectors of a Hilbert space
X is orthogonal if any two distinct elements in S are orthogonal. An orthogonal set S is called
orthonormal if kxk = 1 for all x ∈ S.
1 T
Z
hf, gi = lim f (t)g(t). (4)
T →∞ T −T
{eiωt | ω ∈ R} (5)
11
is orthonormal in X . Note that X is an inner product space but not complete.
The completion of X is the space of L2 -almost periodic functions and consists of
equivalence classes of functions of the form
∞
X
f (t) = ak eiωk t , ak ∈ C, ωk ∈ R,
k=1
P∞
where k=1 |ak |2 < ∞. The set in (5) is an uncountable orthogonal subset of
this Hilbert space.
We next introduce the concept of unconditional convergence, which allows us to define sums
of uncountable number of terms.
if for every ε > 0 there exists a finite subset Jε ⊆ I such that kSJ − xk < ε for all finite subsets
J ⊆ I containing Jε .
Example.
Let {xα ∈ X | α ∈ I} be an indexed set of non-negative real numbers xα > 0 and set
nX o
M = sup xα | J ⊆ I, J finite .
α∈J
P
• If
PM = ∞, then α∈J xα is arbitrarily large for sufficiently large J . Therefore,
α∈I xα does not converge unconditionally.
• Suppose that 0 6 M < ∞. Then, for each ε > 0 there exists a finite set Jε ⊆ J
such that
X
M −ε< xα 6 M.
α∈Jε
12
P
Definition 23. An unordered sum α∈I xα is Cauchy if for every ε > 0 there exists a finite
subset Jε ⊆ I such that kSK k < ε for all finite sets K ⊆ I \ Jε .
Proposition 5.1. An unordered sum in a Banach space converges if and only if it is Cauchy.
If an unordered sum in a Banach space converges then {α ∈ I | xα 6= 0} is at most countable.
Therefore, α∈I xα is Cauchy if and only if α∈I kxα k2 is Cauchy. Thus, one of the sums
P P
converges unconditionally if and only if the other does. Finally, (7) follows from the continuity of
the norm and the fact that the sum is the limit of a sequence of finite partial sums.
(c) x − xS ∈ S ⊥ .
Proof. Let J ⊆ I be a finite set. It follows immediately from the orthogonality that
X
X
2
| hx, xα i |2 = kxk2 −
x − hx, xα i xα
6 kxk2 . (8)
α∈J α∈J
Therefore, (see Example 5) (a) holds. Furthermore, (a) implies (b) because of Lemma 3. Finally,
the statement (c) means that hx − xS , xP
α i = 0 for all α ∈ I, which is true by the continuity of
the inner product and the fact that x − β∈K hx, xβ ixβ ⊥ xα for all finite subsets K of I and all
α ∈ A.
Definition 24 (Closed linear span). Given a subset S of a Hilbert space X , we define the closed
linear span [S] of S by
nX X o
[S] = cx x | cx ∈ F, cx x converges unconditionally .
x∈S x∈S
Lemma 3 implies that the closed linear span [S] of an orthonormal subset S = {xα ∈ X | α ∈ I}
of a Hilbert space X can be written as
nX X o
[S] = cα xα | cα ∈ F, |cα |2 < ∞ .
α∈I α∈I
13
Theorem 11. If S = {xα ∈ X | α ∈ I} is an orthonormal subset of a Hilbert space X , then
the following conditions are equivalent:
(a) hx, xα i = 0 for all α ∈ I implies that x = 0;
P
(b) x = α∈I hx, xα i xα for all x ∈ X ;
(c) kxk2 = α∈I | hx, xα i |2 for all x ∈ X ;
P
(d) [S] = X ;
(e) S is a maximal orthonormal set, i.e. it has the property that if S 0 ⊃ S is another orthonormal
set, then necessarily S 0 = S.
Proof. • (a) implies (b): (a) states that S ⊥ = {0}. Therefore, Proposition 5.2 implies (b);
• (b) implies (c): Follows from Lemma 3;
• (c) implies (d): (c) implies that S ⊥ = {0}. Therefore, we also have [S]⊥ = {0};
P
• (d) implies (e): (d) implies that x = α∈I cα xα for all x ∈ X . Therefore, if x ⊥ xα for all
α ∈ I, then x = 0.
• (e) implies (a): Suppose for a contradiction that there exists an x ∈ X such that hx, xα i = 0
for all α ∈ I, but x 6= 0. Then S ∪ {x/kxk} is an orthonormal set and is a strict superset
of S.
Condition (a) is the easiest to verify. Condition (b) is used most often. Condition (c) is called
Parseval’s identity. Condition (d) simply expresses completeness. Condition (e) can be used to
prove that every Hilbert space has an orthonormal basis using Zorn’s lemma.
If a Hilbert space X has an orthonormal basis S = {xα ∈ X | α ∈ I}, then it is isomorphic to
the sequence space l2 (I). In what follows, we are mostly interested in Hilbert spaces that have a
countable orthonormal bases, which are called separable.
2. The set {Hn }n∈N of Hermite polynomials forms an orthogonal basis for L
√w (R).
2
It can be transformed into an orthonormal basis by noting that kHn k = 2πn!
for all n ∈ N.
14
for all x, y ∈ X and α, β ∈ F. The set R(T) = {Tx : x ∈ X } is called the range of T. The
dimension of R(T) is called the rank. The set N (T) = {x ∈ X : Tx = 0} is called the null space
or kernel of T.
Example.
If X is N -dimensional with basis {xn }N M
n=1 and Y is M -dimensional with basis {ym }m=1 ,
then T is completely determined by its matrix representation U = {Un,m } 16n6N with
16m6M
respect to these two bases:
M
X
Txn = Un,m ym , n = 1, 2, . . . , N.
m=1
PN
Therefore, if x = n=1 αn xn and y = Tx, we have
M
X N
X
y= βm ym with βm = Um,n αn .
m=1 n=1
Theorem 12. Let X , Y be normed spaces. Then, B(X , Y) is a normed space with norm
This is called the operator norm of an operator between two normed vector spaces. Furthermore,
if Y is a Banach space, then B(X , Y) is a Banach space.
Proposition 6.2. Let X , Y be normed spaces and T : X → Y a linear operator. The following
properties are equivalent.
(a) T is continuous;
(c) T is bounded.
Proof.
15
(a) implies (b): Suppose that (b) does not hold. Then, there exists a sequence {xn }∞
n=1 of
elements xn ∈ X with kxn kX = 1 such that kTxn kY > n. Set yn = xn /n. Then, yn → 0, but
kTyn kY > 1. Therefore, T is not continuous as limn→∞ Tyn 6= T0 = 0;
(b) implies (c): Obvious for x = 0. For x 6= 0, we have
x
kTxkY = kxkX
T kxk
6 kTk kxkX ;
X Y
Lemma 5. If T1 and T2 are linear operators between normed spaces for which the range of T2
is contained in the domain of T1 , we can consider the composed operator T1 T2 . If T1 and T2 are
bounded, then also T1 T2 is bounded and
Proof.
= kT1 k kT2 k .
Definition 28. Let X , Y be a normed space with norms k·kX and k·kY , and let T : X → Y be
a linear operator. We say that T is a linear isometry if kTxkY = kxkX , for all x ∈ X .
(b) If X = S ⊕ T , then there exists a unique projection P : X → X such that S = R(P) and
T = N (P), called the projection onto S along T .
16
Proof. We first show that x ∈ R(P) if and only if x = Px. Clearly, Px ∈ R(P). Suppose that
x ∈ R(P). Then x = Pz for some z ∈ X and, therefore, Px = PPz = Pz = x.
Now we prove (a). Let x ∈ N (P) ∩ R(P). Then x = Px and Px = 0, which implies that x = 0.
Therefore, N (P) ∩ R(P) = {0}. Furthermore, we can decompose each x ∈ X as x = Px + (I − P)x
with Px ∈ R(P) and (I − P)x ∈ N (P). Therefore, X = R(P) ⊕ N (P).
Now we show that (b) holds. Each x ∈ X has a unique decomposition x = s + t with s ∈ S
and t ∈ T , and P (x) = s defines the required projection.
Example.
Let
n x o
S= | x ∈ R ⊆ R2
0
n x o
T = | x ∈ R ⊆ R2 .
x
P : R2 → R2
x x−y
7→ .
y 0
(a) If P : X → X is an orthogonal projection, then R(P) is closed and R(P)⊥ = N (P). Therefore,
X = R(P) ⊕ N (P) is an orthogonal direct sum;
(b) If S is a closed subspace, then there exists a unique orthogonal projection P : X → X such
that S = R(P) and S ⊥ = N (P).
Proof. We first prove (a). Let P : X → X be an orthogonal projection. According to Property (a)
in Theorem 13, X = R(P) ⊕ N (P). If x = Pz ∈ R(P) and y ∈ N (P), we have
17
Therefore, R(P) = N (P)⊥ and, in particular, a closed subspace.
Now we show that (b) holds. Let S be a closed subspace of X . Then, Corollary 4.1 implies
that X = S ⊕ S ⊥ . We define the projection P by
Px = s, where x = s + t with s ∈ S, t ∈ S ⊥ .
Then, R(P) = S and N (P) = S ⊥ . Now let x1 , x2 ∈ X and xi = si +ti be the unique decomposition
of xi with si ∈ S and ti ∈ S ⊥ , i = 1, 2. Then, we have
Example.
Let
n x o
S= | x ∈ R ⊆ R2 .
0
P : R2 → R2
x x
7→ .
y 0
Therefore, kPk 6 1. For Px 6= 0, we have kP(Px)k = kPxk, which implies that kPk > 1.
18
Proposition 6.3. Let S = {xα : α ∈ I} be an orthonormal system (but not necessarily an
ONB) in a Hilbert space X . Let V = [S] be the closed linear span of S. Then the orthogonal
projection PV is given explicitly by
X
PV x = hx, xα ixα .
α∈I
Proof. The set S can be extended to an ONB B of X (if X is separable, then this can be done by
applying the Gram-Schmidt orthogonalization procedure, or in the general case, by using Zorn’s
Lemma). Thus we can enumerate B = {xα : α ∈ J } by an index set J containing I. Now, given
an arbitrary x ∈ X , define s, t ∈ X according to
X X
s= hx, xα ixα , t= hx, xα ixα .
α∈I α∈J \I
Moreover, * +
X X
hs, ti = hx, xα ixα , hx, xβ ixβ
α∈I β∈J \I
X
= hx, xα ihx, xβ i hxα , xβ i = 0.
| {z }
α∈I
β∈J \I =0, as α6=β
Example.
Let X be a Hilbert space. For every y ∈ X , the mapping
X →F (11)
x 7→ hx, yi (12)
θX : X → B(X , F) (13)
y 7→ fy = h·, yi
19
is a bijection. In the case of F = C, this mapping is antilinear (as can be seen from θX (αy) =
h·, αyi = α h·, yi = θX (y)). Furthermore, the mapping θ is an isometry:
hfx , fy i = hy, xi .
Theorem 16. Let X , Y be Hilbert spaces and T ∈ B(X , Y) a bounded linear operator. Then,
there exists a unique operator T∗ ∈ B(Y, X ) such that
Proof. We first prove the existence. Let y ∈ Y be arbitrary but fixet. The mapping x 7→ hTx, yi
is a bounded linear functional on X . Therefore, according to Riesz’s representation theorem,
there exists a unique zy ∈ X such that
hTx, yi = hx, zy i , x ∈ X.
We set T∗ y = zy .
We now show that T∗ is unique. Suppose that
Then, we have
Therefore, (T∗1 − T∗2 )y ∈ X ⊥ = {0} for all y ∈ Y, which implies that T∗1 − T∗2 = 0 and, in turn,
that T∗1 = T∗2
The linearity of T∗ follows from the uniqueness and the linearity of the inner product.
Furthermore, we have
= sup kzy k
kyk=1
= kTk , (17)
where (15) follows from the fact that the mapping θ in (13) is an isometry and in (16) we applied
the Cauchy-Schwarz inequality. T∗∗ = T follows immediately from (14). Furthermore, we have
20
Definition 32. Let X , Y be Hilbert spaces and T ∈ B(X , Y). The operator T∗ ∈ B(Y, X ) in
Theorem 16 is called the adjoint of T ∈ B(X , Y).
1. self-adjoint if T = T∗ ;
3. normal if TT∗ = T∗ T.
The converse is not true, that is, a linear isometry may notP be unitary. To see this,
X {x ∈ 2
n∈N |xn | < ∞} with
consider the Hilbert space = = (x , x , . . . ) R N :
P 1 2
inner product hx, yi = n∈N xn yn , and let T(x1 , x2 , x3 , . . . ) = (0, x1 , x2 , . . . ) be
the right-shift operator. Then T is a linear isometry, but it is not invertible, so
it cannot be unitary.
Proof. Let
From |hTx, xi| 6 kTxk kxk 6 kTk kxk2 it follows that α 6 kTk. It remains to show that kTk 6 α.
It follows from the Cauchy Schwarz inequality that
with equality if and only if Tx and y are colinear. Therefore, we can rewrite
21
It follows from the polarization formula ((2) for F = C and (1) for F = R) in Theorem 3 for
hTx, yi ∈ R that
1
|hTx, yi| = kTx + yk2 − kTx − yk2
4
1
= |hTx + y, Tx + yi − hTx − y, Tx − yi|
4
1
= |hT(x + y), x + yi − hT(x − y), x − yi|
4
1 1
6 |hT(x + y), x + yi| + |hT(x − y), x − yi|
4 4
α
6 (kx + yk2 + kx − yk2 )
4
α
= (kxk2 + kyk2 ) , x, y ∈ X with hTx, yi ∈ R.
2
Using this result in (19) gives kTk 6 α.
Definition 35. Let X be a Hilbert space and T ∈ B(X , X ), T > 0. We call a self-adjoint
operator A ∈ B(X , X ) a square root of T if T = A2 .
Proposition 7.1. Let X be a Hilbert space and T ∈ B(X , X ), T > 0. Then, T has a unique
positive definite square root.
A Proof of Theorem 10
Proof of Theorem 10. We first prove the existence of y in (a). Set
From the definition of d, there exists a sequence {yn }n∈N of elements yn ∈ S such that
lim kx − yn k = d. (21)
n→∞
Suppose that this sequence is Cauchy. Since X is complete, there is a y ∈ X such that
limn→∞ yn = y, and since S is closed, we have y ∈ S. Finally, it follows from the continuity of
k · k that kx − yk = limn→∞ kx − yn k = d.
It remains to show that {yn }n∈N is Cauchy. The parallelogram law applied to (x − ym ) and
(x − yn ) implies that
22
where we used (20) and the fact that (ym + yn )/2 ∈ S in the last step. Furthermore, (21) implies
that for each ε > 0 there exists an N (ε) such that kx − yn k 6 d + ε. Combining this with (22),
we find that
kx − yk2 6 kx − y + λzk2
= kx − yk2 + |λ|2 kzk2 + hx − y, λzi + hλz, x − yi
= kx − yk2 + |λ|2 kzk2 + 2<(λ̄ hx − y, zi).
It follows that
We can write hx − y, zi = |hx − y, zi| eiϕz (with ϕz ∈ {0, π} in the case of F = R). Therefore, for
λ = −εeiϕz , we get
|hx − y, zi| = 0, z ∈ S.
Now we prove the uniqueness in (a). Suppose y1 and y2 are elements of S such that
kx − y1 k = kx − y2 k = d. Then since x − y1 ∈ S ⊥ and y1 − y2 ∈ S, we have
d2 = kx − y2 k2 = kx − y1 k2 + ky1 − y2 k2
= d2 + ky1 − y2 k2 ,
so y1 = y2 .
Finally, we prove the uniqueness in (b). Let y1 and y2 be points in S such that x − y1 ∈ S ⊥ ,
x − y2 ∈ S ⊥ . Since x − y1 ⊥ y1 − z for all z ∈ S, we have
23
B Proof of Theorem 15
Proof. If N (f ) = X we have yf = 0. Suppose that N (f ) 6= X . Since f is continuous, N (f ) is
closed. Therefore, according to Corollary 4.1, we can decompose X = N (f ) ⊕ N (f )⊥ . Since
N (f ) 6= X , there exists a non-zero z ∈ N (f )⊥ . We can rewrite any x ∈ X as
f (x) f (x)
x= x− z + z
f (z) f (z)
| {z }
∈N (f )
x = αz + n, α.
hx, zi
α= .
kzk2
Evaluating f on x = αz + n gives
f (z) hx, zi
f (x) = αf (z) = .
kzk2
In particular, we have
24
(b) if λ ∈
/ σ1 (T), then R(T − λI) = X ;
(c) if there exists an A > 0 such that k(T − λI)xk > Akxk for all x ∈ X , then λ ∈
/ σ(T).
Now let λ1 ∈ σ1 (T). Then there exists an x ∈ X , x 6= 0, such that Tx = λx. Then
λkxk2 = hTx, xi ∈ R, so λ ∈ R.
• Proof of (b): We show that R(T − λI)⊥ = {0}. Let x ∈ R(T − λI)⊥ . Then, we have
0 = h(T − λI)y, xi
= y, (T − λI)x , y ∈ X.
Therefore, {xn }n∈N is a Cauchy sequence in a Hilbert space and, therefore, converges to x.
By the continuity of T − λI, we have y = (T − λI)x ∈ R(T − λI).
where we used the Cauchy-Schwarz inequality in the first step. Therefore, Lemma 8 implies that
λ∈/ σ(T).
25
Theorem 17. Let T ∈ B(X , X ) self-adjoint and set
Suppose that λ = m − c with c > 0. Then, the Cauchy-Schwarz inequality implies that
Theorem 19. Let X , Y be Hilbert spaces and T ∈ B(X , Y). Then, we have
Proof. We first show (23). If y ∈ R(T), there exists a z ∈ X such that y = Tz. Then, we have
which implies that R(T) ⊆ N (T∗ )⊥ . Since N (T∗ )⊥ is closed, it follows that R(T) ⊆ N (T∗ )⊥ .
On the other hand, if y ∈ R(T)⊥ , we have
0 = hTx, yi = hx, T∗ yi , x ∈ X,
26
which implies that T∗ y = 0. This means that R(T)⊥ ⊆ N (T∗ ). By taking the complement of
this relation we get N (T∗ )⊥ ⊆ R(T)⊥⊥ = R(T). To prove (24), we first apply (23) to T∗ , use
T∗∗ = T, and apply complements.
An equivalent formulation of this theorem is that if T ∈ B(X , Y) for two Hilbert spaces X , Y,
then Y can be written as the orthogonal direct sum
Y = R(T) ⊕ N (T∗ ).
Theorem 20. Let X , Y be Hilbert spaces and T ∈ B(X , Y). Then, R(T) is closed if and only
if R(T∗ ) is closed.
Proof. We show that R(T∗ ) closed implies that R(T) is closed. We have the following situation:
T∗
X −→ (R(T) ⊕ R(T)⊥ ) −→ R(T∗ ) ⊆ Y.
T
U : R(T) → R(T∗ )
y 7→ T ∗ (y)
is a bounded invertible linear operator. Since R(R∗ ) is closed in X which is complete, R(T∗ ) is
itself complete and hence a Hilbert space in its own right. Therefore, the adjoint
U∗ : R(T∗ ) → R(T)
∗
exists and is a bounded invertible linear operator (with inverse U∗ −1 = U−1 ). We now show that
R(T) ⊆ R(T), which implies that R is closed. Let y ∈ R(T) and denote by P the orthogonal
projection onto R(T). Then, y = U∗ (x) for some x ∈ R(T∗ ), and we have
Therefore, y = Tx ∈ R(T).
The following lemma gives a condition that ensures the existence of a right-inverse.
Lemma 10. Let X , Y be Hilbert spaces and T ∈ B(X , Y) with closed range R(T). Then there
exists a unique operator T† ∈ B(X , Y) satisfying
27
Proof. We have the orthogonal direct sums
X = N (T) ⊕ N (T)⊥
Y = R(T) ⊕ R(T)⊥ .
It is clear that T
e is linear and bounded. T e = 0, it follows that
e is injective because if Tx
⊥
x ∈ N (T) ∩ N (T) = {0}. To show that T e is surjective, let y ∈ R(T). Then, there exists
an x ∈ X such that y = T(x). We can write x = u + v with u ∈ N (T) and v ∈ N (T)⊥ .
Therefore, y = T(u + v) = T(u) + T(v) = T(u) = T(u)
e ∈ R(T). e −1 exists and
e It follows that T
−1 ⊥
e ∈ B(R(T), N (T) ) by Theorem 18. We define the operator T as
T †
T† (y) = T
e −1 (y), y ∈ R(T) (28)
† ⊥
T (y) = 0, y ∈ R(T) . (29)
It follows immediately that T† has the desired properties. To prove uniqueness, suppose that T†1
and T†2 fulfill Properties (25)–(27) and let y ∈ Y. We can decompose y = w + z with w ∈ R(T)
and z ∈ R(T)⊥ . Then, we have
Proposition D.1 (Properties of T† ). Let X , Y be Hilbert spaces and T ∈ B(X , Y) with closed
range R(T). Then, the following properties hold.
(e) If T is surjective, then given y ∈ Y, the equation Tx = y has a unique solution of minimal
norm, namely x = T† y.
Proof. • (a) Follows from Theorem 14 and Properties (25) and (27);
• (b): Property (26) implies that T† Tx = 0 for all x ∈ R(T† )⊥ . Now let x ∈ R(T† ). Then,
x = T† y for some y ∈ Y. Decomposing y = u + v with u ∈ R(T) and v ∈ R(T)⊥ , Property
(25) implies that x = T† u. Therefore, Property (26) implies that T† Tx = T† TT† u = T† u = x.
Theorem 14 now shows that T† T is an orthogonal projection onto R(T† );
28
• (c): It follows from Theorem 20 that T∗ has closed range. This implies that T∗ † is well
defined and is the unique operator satisfying
N (T∗ † ) = R(T∗ )⊥
R(T∗ † ) = N (T∗ )⊥
T∗ T∗ † x = x, x ∈ R(T∗ ).
∗
Furthermore, it follows from (25)–(27) and (23) together with (24), that T† satisfies the
same properties. Therefore, (T∗ )† = (T† )∗ ;
• It follows from R(T) = N (T∗ )⊥ and N (T) = R(T∗ )⊥ that the operator U : R(T) → R(T),
y 7→ TT∗ y is invertible and the operator
T∗ (TT∗ )−1 on R(T) and 0 on R(T)⊥
fulfills (25)–(27);
• x = T† y is a solution of y = Tx, i.e., y = TT† y. Suppose that y = Tz for some z ∈ X and
set e = T† y − z. It follows that e ∈ N (T). Since T† y ∈ N (T)⊥ , we have e ⊥ T† y, which
2
2
implies that kzk2 =
T† y − e
=
T† y
+ kek2 , which is minimal for e = 0.
is bounded, invertible, self-adjoint, and positive, and that {S−1 fk }∞k=0 is a frame with frame
bounds 0 < B −1 6 A−1 < ∞. In fact, {Tfk }∞ k=0 is a frame for a larger class of operators T, as
stated in the following theorem.
Theorem 21. Let X be a Hilbert spaces and T ∈ B(X , X ) with closed range R(T) and {fk }∞
k=0
be a frame with
X∞
2
A kxk 6 |hx, fk i|2 6 B kxk2 , x ∈ X .
k=0
Then, {Tfk }∞
k=0 is a frame in span{Tfk }∞
k=0 with
∞
†
−2 2
X
A T
kyk 6 |hy, Tfk i|2 6 B kTk2 kyk2 , y ∈ span{Tfk }∞
k=0 . (30)
k=0
Proof. In order to prove Theorem 21, we make use of the following lemma, which shows that it is
enough to check the frame condition on a dense set.
29
The upper frame bound in (30) follows from
∞
X ∞
X
2
|hy, Tfk i| = |hT∗ y, fk i|2 6 B kT∗ yk2 6 B kTk2 kyk2 , y ∈ X.
k=0 k=0
Now we prove that the lower frame bound in (30) holdes for all y ∈ span{Tfk }∞ k=0 . We know
that the operator TT† is the orthogonal projection onto R(T) and, in particular, it is self-adjoint.
Let y ∈ span{Tfk }∞k=0 . Then, y ∈ R(T), and we obtain that
This gives
Furthermore, we have
∞ ∞
1 X 1 X
kT∗ yk2 6 |hT∗ y, fk i|2 = |hy, Tfk i|2 . (34)
A A
k=0 k=0
which shows that the lower frame condition is satisfied for all y ∈ span{Tfk }∞
k=0 . Using Lemma 11,
we can conclude that {Tfk }∞ k=0 forms a frame for span{Tf } ∞ .
k k=0
In the special case where T is a surjective bounded operator, T has closed range. Furthermore,
every y ∈ Y can be written as y = Tx, where x ∈ X . Since {fk }∞ k=0 is a frame, one has
∞
X
x, S−1 fk fk ,
x=
k=0
Corollary E.1. Assume that {fk }∞ k=0 is a frame for X with frame bounds 0 < A 6 B < ∞
and that T : X → Y is a bounded surjective operator. Then {Tfk }∞
k=0 is a frame for Y with frame
†
−2 2
bounds A T
, B kTk .
As a consequence, if we have at our disposal a frame, we can construct many other frames by
applying surjective operators to it.
30
References
[1] J. K. Hunter and B. Nachtergaele, Applied Analysis. Singapore: World Scientific Publishing,
2005.
[2] S. B. Damelin and W. Miller, The Mathematics of Signal Processing, ser. Cambridge texts in
applied mathematics. Cambridge, UK: Cambridge University Press, 2012.
[3] E. M. Stein and R. Sharachi, Real Analysis: Measure Theory, Integration, and Hilbert Spaces,
ser. Princeton lectures in analysis. Princeton, NJ, USA: Princeton University Press, 2005.
[4] O. Christensen, An Introduction to Frames and Riesz Bases, ser. Applied and Numerical
Harmonic Analysis. Boston, MA, USA: Birkhäuser, 2003.
[5] C. Heil, “Operators on Hilbert spaces,” Functional analysis lecture notes, February 2006.
31