Skip to content

Hilbert Space

We first introduce inner product in abstract space.

Inner Product

Assume \(\mathscr{U}\) is a linear space on number field \(\mathbb{K}\) (i.e. \(\mathbb{R}\) or \(\mathbb{C}\)). A mapping

\[ (\cdot, \cdot): \mathscr{U}\times \mathscr{U} \mapsto \mathbb{K} \]

is said to be a Inner Product, if it satisfies that, for all \(x,y,z\in \mathscr{U}\), for all \(\alpha \in \mathbb{K}\),

(i) Homogeneous on the first element. \((\alpha x,y)=\alpha(x,y)\).

(ii) Linearity on the first element. \((x+z,y)=(x,y)+(z,y)\).

(iii) Conjugate inverse. \((x,y)=\overline{(y,x)}\).

(iv) Positive-definite. \((x,x)\geq 0\) and \((x,x)=0\Leftrightarrow x=\theta\).

And A linear space \(\mathscr{U}\) with inner product is called Inner Product Space.

With the above definition, we have the following properties of inner product.

Properties of inner product

Assume \(\mathscr{U}\) is an inner product space on number field \(\mathbb{K}\). Then \(x,y,z\in \mathscr{U}\), for all \(\alpha \in \mathbb{K}\),

(i) Conjugate Homogeneous on the second element. \((x,\alpha y)=\overline{\alpha}(x, y)\).

(ii) Linearity on the second element. \((x,y+z)=(x,y)+(x,z)\).

(iii) Zero for null vector. \((x,\theta)=(x,0\cdot x)=0(x,x)=0\).

(iv) Normed linear space. Define \(\|x\|=\sqrt{(x,x)}\), then it is a norm.

(v) \((x,y)\) is a continuous function with respect to both \(x,y\).

Easy to show that (i), (ii), (iii).

We prove for (iv) and the triangle inequation. This is based on Schwarz Inequation.

\[ |(x,y)|^2\leq \|x\|^2 \|y\|^2 \]

We still make use of \(\forall \lambda\in \mathbb{K}\)

\[ \begin{align*} 0 &\leq \|x+\lambda y\|^2\\ & = (x+\lambda y,x+\lambda y)\\ &= \|x\|^2+ \overline{\lambda}(x,y)+ \lambda(y,x) + |\lambda|^2\|y\|^2\\ &= \|x\|^2 + \lambda(y,x) + \overline{\lambda}[ (x,y) + \lambda \|y\|^2] \end{align*} \]

Assume \((y,y)\neq 0\), and choose \(\lambda=-\frac{(x,y)}{\|y\|^2}\), we have

\[ \begin{align*} 0&\leq \|x\|^2 - \frac{(x,y)}{\|y\|^2} (y,x)\\ |(x,y)|^2&\leq \|x\|^2 \|y\|^2 \end{align*} \]

With Schwarz inequation, we have

\[ \begin{align*} \|x+y\|^2 &= |(x+y,x+y)|\\ &=|(x+y,y)+(x+y,x)|\\ &\leq |(x+y,y)|+|(x+y,x)|\\ &\leq \|x+y\|\|y\| + \|x+y\|\|x\|\quad \text{Schwarz Inequation} \end{align*} \]

and we are done.

Polarization Identity

(i) For \(\mathbb{K}=\mathbb{R}\), we have

\[ (x,y)=\frac{1}{4}(\|x+y\|^2-\|x-y\|^2). \]

(ii) For \(\mathbb{K}=\mathbb{C}\), we have

\[ (x,y)=\frac{1}{4}(\|x+y\|^2-\|x-y\|^2 + i \|x+iy\|^2 - i\|x-iy\|^2). \]

(i) Just expand the righthand item in inner product.

(ii) Still the same logic, but

\[ \|x+y\|^2-\|x-y\|^2=2(x,y)+2\overline{(x,y)}=4Re[(x,y)]. \]

We notice \(Re[(x,iy)]=Re[-i(x,y)]=Im[(x,y)]\), so apply \(y=iy\) in the above equation and we have

\[ 4Im[(x,y)]=4Re[(x,iy)]=\|x+iy\|^2-\|x-iy\|^2. \]

\(\square\)

Hilbert Space

If a inner product space \(\mathscr{U}\) is complete, then it is called Hilbert Space.

We have known that a inner product space must be a normed linear space, then how about the inverse? Generally speaking, a normed linear space could not be a inner product space, i.e. use the inner product to deduce a norm, unless we add the following condition.

Parallelogram Law

If a norm satisfies

\[ \|x+y\|^2+\|x-y\|^2=2(\|x\|^2+\|y\|^2) \]

then we could define a inner product using polarization identity.

Inversely speaking, a norm introduced by a norm satisfy the above condition, which is called Parallelogram Law.

  • Sufficient.

Just expand lefthand of the equation using inner product and we have the result.

  • Necessary.

We first show that the inner product defined by Polar equation satisfies linearity.

\[ \begin{align*} (x,z)+(y,z)&=\frac{1}{4}(\|x+z\|^2-\|x-z\|^2 + i \|x+iz\|^2 - i\|x-iz\|^2)\\ &+\frac{1}{4}(\|y+z\|^2-\|y-z\|^2 + i \|y+iz\|^2 - i\|y-iz\|^2)\\ &=\frac{1}{8}(\|x+y+2z\|^2-\|x+y-2z\|^2+i\|x+y+i2z\|^2-i \|x+y-i2z\|^2)\\ &=\frac{1}{2}(\|\frac{x+y}{2}+z\|^2-\|\frac{x+y}{2}-z\|^2+i\|\frac{x+y}{2}+iz\|^2-i \|\frac{x+y}{2}-iz\|^2)\\ &=2(\frac{x+y}{2},z)\\ &=\frac{1}{2}(x+y,2z)\quad\text{this is not good}. \end{align*} \]

Let \(y=\theta\), we have \((x,z)=2(\frac{x}{2},z)\). Choose \(x=x+y\), we have \((x+y,z)=2(\frac{x+y}{2}, z)\). Comparing with before and we have

\[ (x,z)+(y,z)=(x+y,z). \]

So let \(y=y-x\), we have

\[ (x,z)+(y-x,z)=(y,z)\Rightarrow (y-x,z)=(y,z)-(x,z) \]

which means "+" and "-" satisfy.

This is a little tricky, and we have to show a lemma.

For a continuous function \(f(\alpha)\) defined on \(\mathbb{R}\), which satisfies

\[ f(\alpha_1+\alpha_2)=f(\alpha_1)+f(\alpha_2),\quad \forall \alpha_1,\alpha_2\in \mathbb{R}, \]

then \(\forall \alpha\in \mathbb{R}\), we have

\[ f(\alpha)=\alpha f(1). \]

Let

\[ f(\alpha)=(\alpha x,z). \]

it is continuous because

\[ \begin{align*} |f(\alpha)-f(\alpha_0)|&=|((\alpha-\alpha_0)x,z)|\quad \text{subtraction}\\ \end{align*} \]

For \(\alpha\rightarrow \alpha_0\), \((\alpha-\alpha_0)x\rightarrow 0x=\theta\), so the lefthand of the above equation tends to \(0\).

And

\[ f(\alpha_1+\alpha_2)=((\alpha_1+\alpha_2)x, z)=(\alpha_1 x+\alpha_2 x, z)=(\alpha_1 x, z)+(\alpha_2 x, z), \]

so we have

\[ (\alpha x,z)=\alpha(x,z),\quad \forall \alpha\in \mathbb{R}, \]

which is homogeneity.

As for conjugate property and positive-definitivity, just follow the definition.

Orthogonal System

Orthogomal System

Assume \(\mathscr{U}\) is a inner product space, \(x,y\in \mathscr{U}\). If \((x,y)=0\), then we call \(x\) is orthogamal to \(y\), denoted by \(x\perp y\). If a subset \(M\subset \mathscr{U}\), \(\forall y\in M\), \(x\perp y\), then we call \(x\) is orthogonal to \(M\), denoted by \(x\perp M\). If \(N\subset \mathscr{U}\), \(\forall x\in N\), \(x\perp M\), then we call \(N\) is orthogonal to \(M\), denoted by \(N\perp M\).

An orthogonal complement of \(M\) is denoted by

\[ M^\perp=\{x\in \mathscr{U}: x\perp M\}. \]

Property of Orthogomal

(i) Pythagorean Theorem. For \(x_1,\cdots,x_n\) which are mutually orthogonal, then \(x:=x_1+\cdots+x_n\), we have

\[ \|x\|^2=\sum_{i=1}^n \|x_i\|^2. \]

(ii) Dense set with Zero Vector. Assume \(L\subset \mathscr{U}\) is a dense subset, if \(x\in \mathscr{U}\) is orthogonal to \(L\), then \(x=0\).

(iii) For \(M\subset \mathscr{U}\), \(M^\perp\) is a closed subset.

Orthogomal Projection

We introudce orthogonal projection using the best approximation element.

Definition of best Approximation element

Assume \(M\subset \mathscr{U}\), \(x\in \mathscr{U}\), \(y\in \mathscr{U}\) is called the best approximation element of \(x\) in \(M\), if

\[ \|x-y\|=\inf_{z\in M}\|x-z\| \]

Here we give a sufficient condition for the existence and uniqueness of best approximation element.

Theorem for existence and uniqueness of best approximation element

Assume \(M\subset \mathscr{U}\), \(x\in \mathscr{U}\). If \(M\) is a closed convex set, then there exists a unique best approximation element \(y\) of \(x\) in \(M\).

This proof leverages the convex property.

From the above theorem, A subspace is naturally convex, thus adding closedness could satisfy the above condition. So

Orthogomal decomposition and projection

Assume \(M\subset \mathscr{U}\) is a closed subspace, then \(\forall x\in \mathscr{U}\), \(\exists! y\in M, z\in M^\perp\), such that

\[ x=y+z. \]

the above equation is called orthogonal decomposition, and \(y\) is called the orthogonal projection of \(x\) in \(M\).

Normalized Orthogomal System

Definition of Normalized Orthogomal System

System \(\{e_n\}_{n\geq 1}\subset \mathscr{U}\) is called Normalized Orthogomal System, if it satisfies

\[ (e_n,e_m)=\begin{cases}1,\quad n=m,\\ 0,\quad n\neq m.\end{cases} \]

For a element \(x\in \mathscr{U}\), define

\[ c_n=(x,e_n) \]

to be the Fourier coefficient of \(x\). They also form a system \(\{c_n\}\).

Now we talk a little deeper into this one.

Explicite expression of orthogonal decomposition

Assume \(n\) is given, \(\{e_1,\cdots, e_n\}\) is a finite-number normalized orthogonal system. Let \(M=\text{span}(e_1,\cdots,e_n)\), then \(\forall x\in \mathscr{U}\), we have the orthogonal projection of \(x\) on \(M\)

\[ x=\sum_{i=1}^n (x,e_n)e_n \]

First use Orthogomal decomposition and projection to show that

\[ x=y+z,\quad y\in \text{space}(e_1,\cdots,e_n). \]

So by property of base, \(y=\sum_{i=1}^n a_n e_n\). Making a inner product with \(e_j\) on both sides gives

\[ a_j=(y,e_j) \]

So we are done.

Bessel Inequation

For a normalized orthogonal system \(\{e_n\}\subset \mathscr{U}\), \(x\in \mathscr{U}\) with its Fourier coefficients \(\{c_n\}\), we have

\[ \sum_{n=1}^\infty \|(x,e_n)\|^2\leq \|x\|^2. \]

or \(\{e_n\}\subset l^2\).

Use finite many elements and then let \(n\rightarrow\infty\).

Inversely speaking, given a sequence in \(\{c_n\}\subset l^2\), could we find a element in \(x\in \mathscr{U}\) such that \(c_n\) is its fourier coefficient? It turns out that we need to add another condition.

Riesz-Fischer Theorem

Assume \(\mathscr{U}\) is a Hilbert space, \(\{e_n\}\) are its normalized orthogonal system, number sequence \(\{c_n\}\subset l^2\). Then there exists \(x\in \mathscr{U}\), such that for each \(n\), we have

\[ c_n=(x, e_n) \]

and Parseval equation holds

\[ \sum_{i=1}^\infty \|c_n\|^2=\sum_{i=1}^\infty \|(x, e_n)\|^2= \|x\|^2 \]

\(\square\)

Now we pause here. It is interesting to see that Bessel inequation holds for all inner product space, but Parseval equation does not also hold. We dig deeper into this and this is actually related to the number of \(\{e_n\}\). For a Hilbert space, its number is enough to let "=" hold.

Completeness

Definition for completeness of orthogonal system

Assume \(\mathscr{U}\) is a inner product space. An normalized orthogonal system is said to be complete, if for all \(x\in\mathscr{U}\), Parseval equation holds; it is said to be total, if it is dense in \(\mathscr{U}\).

Relationship

Assume \(\mathscr{U}\) is a Hilbert space. \(\{e_n\}\) is a normalized orthogonal system. Then we have the following equivalence.

(i) \(\{e_n\}\) is complete.

(ii) \(\forall x\in \mathscr{U}\), series \(\sum_{i=1}^n (x,e_i)e_i\) converges to \(x\), i.e.

\[ x=\sum_{n=1}^\infty (x,e_n)e_n. \]

(iii) \(\forall x,y\in \mathscr{U}\),

\[ (x,y)=\sum_{n=1}^\infty (x,e_n)\overline{(y,e_n)}. \]

(iv) \(\{e_n\}\) is total.

(i) and (ii) are equivalent because

\[ \left\| x-\sum_{n=1}^N (x,e_n)e_n \right\|^2= \|x\|^2-\sum_{n=1}^N |(x,e_n)|^2 \]

Using equation from (ii) we only get infinite series. We start from finite-numbers. Define

\[ x_n=\sum_{i=1}^n (x,e_i)e_i,\quad y_n=\sum_{j=1}^n (y,e_j)e_j, \]

then

\[ \begin{align*} (x_n,y_n)&=\left( \sum_{i=1}^n (x,e_i)e_i, \sum_{j=1}^n (y,e_j)e_j\right)\\ &=\sum_{i=1}^n(x, e_n)\overline{(y,e_n)}. \end{align*} \]

Let \(n\rightarrow \infty\), since by hölder inequation

\[ \left(\sum_{i=1}^\infty |(x, e_n)\overline{(y,e_n)}|\right)^2\leq \left(\sum_{i=1}^\infty |(x,e_n)|^2 \right)\left(\sum_{i=1}^\infty |(y,e_n)|^2\right) \]

by (ii) the above converges. By continuity of inner product, we are done.

Assume \(\forall x\in \mathscr{U}\), \((x,e_n)=0\). So \(\forall y\in \mathscr{U}\)

\[ (x,y)=\sum_{n=1}^\infty (x,e_n)\overline{(y,e_n)}=0. \]

Choose \(y=x\) and \((x,x)=0\) implies \(x=\theta\).

\(\forall x\in \mathscr{U}\), by Bessel ineuqation, \(\sum_{n=1}^\infty |(x,e_n)|^2<\infty\). For number sequence \(\{(x,e_n)\}\subset l^2\), by Riesz-Fischer Theorem, \(\exists y\in \mathscr{U}\), such that \((x,e_n)=(y,e_n)\) for each \(n\). So \((x-y,e_n)=0\) for each \(n\), by (iv) assumption, we have \(x-y=\theta\), i.e. \(x=y\) and \(x\) satisfies

\[ \|x\|^2=\|y\|^2 =\sum_{i=1}^\infty |(x,e_n)|^2. \]