Skip to content

Approximation Theory

最佳平方逼近 | Best Square Approximation

引入 | Lead-in

We call a linear space with norm if there exists a function \(\|\cdot\|\) that satisfies 3 properties(check here).

So here we introduce

\[ \Delta(x,Y)=\inf_{y\in Y}\|x-y\| \]

to be the best square approximation of element \(x\), where \(x\in X\) is a linear space with norm 2 and \(Y\) is a subspace of \(X\).

We can also introduce norm with the definition of inner product. In Euclid space, \((\cdot,\cdot)\) is defined as

\[ (x,y)=x^Ty,\quad x,y\in \mathbb{R}^n \]

It is easy to see that \(\|x\|_2=\sqrt{(x,x)}\). So we can give a more specific problem of inner product space. Assume \(\varphi_i\), \((i=1,2\cdots,n)\) are \(n\) linearly irrelevant elements in inner product space \(X\), choose \(f\in X\), then subset

\[ \varPhi_n=\text{span}(\varphi_1,\varphi_2,\cdots,\varphi_n) \]

has a best approximation of \(f\), which is defined as

\[ \Delta(f,\varPhi_n)=\min_{\varphi\in \varPhi_n}\|f-\varphi\|_2 \]

we call the \(\varphi\) that enables the above equation to be the best square appromation element.

最佳平方逼近元 | Properties of Best Square Appromation Element

Sufficient and Necessary Condition for best square approximation element

Assume \(X\) is an inner product space, \(f\in X\), \(\varphi^*\in \varPhi_0\) is the best square approximation element, if and only if

\[ (f-\varphi^*,\varphi_i)=0, \quad, i=1,2\cdots,n. \]

"\(\Leftarrow\)".

"\(\Rightarrow\)".

The above theorem gives a general method to solve the best square approximation element. That is, we define

\[ G=\left[\begin{array}{cccc} (\varphi_1,\varphi_1) & (\varphi_1,\varphi_2) & \cdots & (\varphi_1,\varphi_n)\\ (\varphi_2,\varphi_1) & (\varphi_2,\varphi_2) & \cdots & (\varphi_2,\varphi_n)\\ \vdots & \vdots & \ddots & \vdots \\ (\varphi_n,\varphi_1) & (\varphi_n,\varphi_2) & \cdots & (\varphi_n,\varphi_n)\\ \end{array}\right], \quad \pmb{\alpha}^*=\left[\begin{array}{c} \alpha_1^*\\ \alpha_2^*\\ \vdots\\ \alpha_n^* \end{array}\right], \quad \pmb{\beta}=\left[\begin{array}{c} (\varphi_1,f)\\ (\varphi_2,f)\\ \vdots\\ (\varphi_n,f) \end{array}\right] \]

then we can solve \(G\pmb{\alpha}^*=\pmb{\beta}\) for the best square approximation element

\[ \varphi = \sum_{i=1}^n\alpha^*_i \varphi_i \]

The above theorem also tells us, if \(\{\varphi_i\}_{i=1}^n\) are group of linearly irrelevant elements, then \(G\) has rank \(n\) and the parameters \(\{\alpha_i\}_{i=1}^n\) is unique. And it is easy to give the error estimation form

\[ \begin{align*} \|f-\varphi^*\|_2^2&=(f-\varphi^*,f-\varphi^*)\\ &=(f-\varphi^*, f) - (f-\varphi^*, \varphi^*) \\ &=(f-\varphi^*, f) = (f,f) - (\varphi^*,f)\\ &=\|f\|_2^2-\sum_{i=1}^n\alpha^*_i (\varphi_i, f) \end{align*} \]

The geometrical meaning is also clear, if we define \(\Delta=f-\varphi^*\) and then

\[ \begin{align*} \|f-\varphi^*\|_2^2&=(\Delta+\varphi^*,\Delta+\varphi^*)\\ &=(\Delta, \Delta) +2 (\Delta, \varphi^*)+(\varphi^*,\varphi^*) \\ &=(\Delta, \Delta) +2 (f-\varphi^*, \varphi^*)+(\varphi^*,\varphi^*) \\ &= (\Delta, \Delta) - (\varphi^*,\varphi^*) \quad \text{(by property of BSPE)}\\ \end{align*} \]

which means \(\varphi^*\) is the orthogonal projection of \(f\) on \(\varPhi_n\).

But actually it is not so easy to solve the above linear system, so we consider another way in specific situation.

有限函数空间上的最佳平方逼近 | Best Square Approximation on \(L_\rho^2[a,b]\)

Weight Function

Assume \(\rho(x)\) is Lebesgue integrable on \([a,b]\) and is \(0\) on at most only one set of measure zero, then we call \(\rho(x)\) the weight function.

If \(\rho(x)\) is the weight function on \([a,b]\) then we denote all the measurable function \(\rho(x)f^2(x)\) that is Lebesgue integrable on \([a,b]\), by \(L^2_\rho[a,b]\), which is a linear space. The inner product is defined by

\[ (f,g)=\int_{a}^b\rho(x)f(x)g(x)dx, \quad f,g\in L^2_\rho[a,b] \]

which satisfies the 4 properties of inner product definition. So we can introduce natural norm

\[ \|f\|_2=\left[\int_a^b\rho(x)f^2(x)dx\right] \]

then \(L^2_\rho[a,b]\) is a linear space with norm, so we can consider the best approximation problem.

Because \(L^2_\rho[a,b]\) is a special inner product space \(X\), so the former discussion about the best square approximation element can be directly applied here. That is, if \(\{\varphi_i\}_{i=1}^n\) is a group of linearly irrelevant function in \(L^2_\rho[a,b]\), then the subset

\[ \varPhi_n=\text{span}\{\varphi_1,\varphi_2, \cdots, \varphi_n\} \]

is call the generalized polynomial space, whose element is called generalized polynomial, or polynomial for short. The best square approximation of \(f\) on \(\varPhi_n\) is defined by

\[ \Delta(f,\varPhi_n)=\min_{\varphi\in \Phi}\|f-\varphi\|_2 \]

and so does the best square approximation element \(\varphi^*\) is called the best square approximation polynomial(BSAP).

There is another way to get the BSAP.

If we choose

\[ E = (P-y,P-y)=\|P-y\|^2 \]

we can easily solve the approximation polynomial \(\sum\limits_{i=0}^\infty a_ix^i\) by letting \(\frac{\partial E}{\partial a_k}=0\), and get

\[ \sum_{j=0}^n(\varphi_k, \varphi_j)a_j=(\varphi_k,f),\quad k=0,1,\cdots,n \]

正交多项式 | Orthogonal Polynomials

If we limit the base function of the subspace, we can simplify the matrix \(G\) and simplify the solving process. Actually, we are narrawing down the condition number of \(G\) by using orthogonal polynomials. Readers can use exactly the same statement from abstrat inner product space and its corresponding statement of orthogonal elements.

Orthogonal Polynomials

If \(\{\varphi_i\}_{i=1}^n \subset L^2_\rho[a,b]\) satisfy

\[ (\varphi_i,\varphi_j)=\int_a^b\rho(x)\varphi_i(x)\varphi_j(x)dx=\begin{cases} 0, \quad &i=j,\\ \sigma, \quad & i\neq j. \end{cases} \]

where \(\sigma\) is a non-zero number, then we call \(\{\varphi_i\}_{i=1}^n\) are orthogonal on \([a,b]\) according to weight \(\rho(x)\). Furthermore, if we limit

\[ \int_a^b\rho(x)\varphi_i^2(x)dx=1 \]

then we call \(\{\varphi_i\}_{i=1}^n\) are normalized orthogonal system.

So from above definition we can get BSAP more easily with orthogonal functions \(\{\varphi_i\}_{i=1}^n\)

\[ \varphi^*=\sum_{i=1}^n\frac{(\varphi_i,f)}{(\varphi_i,\varphi_i)}\varphi_i(x) \]

and its corresponding error

\[ \|f-\varphi^*\|^2_2=\|f\|^2_2-\sum_{i=0}^n\frac{(\varphi_i,f)^2}{(\varphi_i,\varphi_i)} \]

Q1. Use \(y=2^{ax+b}\) to approximate the following 3 points.

\(x_i\) 0 1 4
\(f(x_i)\) 1 2 8
\(w\) 1 1 1

离散最小二乘法 | Discrete Least Squares Approximation

This problem can be discribed as the best square approximation on Euclid space. That is, if \(y\in \mathbb{R}^n\), \(\{\pmb{x}_i\}_{i=1}^m\subset \mathbb{R}^n\) is a group of linearly irrelevant vectors, then

\[ V=\text{span}\{\pmb{x}_1,\pmb{x}_2,\cdots,\pmb{x}_m\} \]

is a subspace of \(\mathbb{R}^n\).

So we can consider the best square approximation problem

\[ \Delta(\pmb{y},V)=\min_{\pmb{x}\in V}\|\pmb{y}-\pmb{x}\|_2 \]

where \(\|\cdot\|_2\) is the Euclid norm, i.e. \(\|\pmb{x}\|_2=\pmb{x}^T\pmb{x}\), \(x\in \mathbb{R}^n\)

That is, we have to solve

\[ \left[\begin{array}{cccc} (\pmb{x}_1,\pmb{x}_1) & (\pmb{x}_1,\pmb{x}_2) & \cdots & (\pmb{x}_1,\pmb{x}_n)\\ (\pmb{x}_2,\pmb{x}_1) & (\pmb{x}_2,\pmb{x}_2) & \cdots & (\pmb{x}_2,\pmb{x}_n)\\ \vdots & \vdots & \ddots & \vdots \\ (\pmb{x}_n,\pmb{x}_1) & (\pmb{x}_n,\pmb{x}_2) & \cdots & (\pmb{x}_n,\pmb{x}_n)\\ \end{array}\right] \left[\begin{array}{c} \alpha_1^*\\ \alpha_2^*\\ \vdots\\ \alpha_n^* \end{array}\right] = \left[\begin{array}{c} (\pmb{x}_1,\pmb{y})\\ (\pmb{x}_2,\pmb{y})\\ \vdots\\ (\pmb{x}_n,\pmb{y}) \end{array}\right] \]

If we denote \(\pmb{x}_i=(x_{1i},x_{2i},\cdots,x_{ni})^T\) and

\[ A=[\pmb{x}_1,\pmb{x}_2,\cdots,\pmb{x}_m]=(x_{ij})_{n\times m}=\left[\begin{array}{cccc} x_{11} & x_{12} & \cdots & x_{1m}\\ x_{21} & x_{22} & \cdots & x_{2m}\\ \vdots & \vdots & \ddots & \vdots \\ x_{n1} & x_{n2} & \cdots & x_{nm}\\ \end{array}\right] \]

then it is not so hard to detect that the above equation can be rewritten as

\[ A^TA\pmb{\alpha^*}=A^T\pmb{y} \]

Cause \(A^TA\) has full rank, we can get \(\pmb{\alpha^*} =(A^TA)^{-1}A^T\pmb{y}\).

最佳一致逼近 | Best Uniform Approximation

引入 | Lead-in

All the continuous function defined on \([a,b]\) compose a infinite-dimensional linear space, denoted by \(C_{[a,b]}\). To simplify our discription, we introduce norm \(\|\cdot\|_\infty\) to denote Chebyshev Norm

\[ \|\cdot\|_\infty = \max_{[a,b]}|f(x)|, \quad \forall f\in C_{[a,b]}. \]

With the following theorem explored by Weierstrass, we can find an algebraic polynomial to sufficiently approximate function \(f\) \(\in C_{[a,b]}\).

Weierstrass First Approximation Theorem

For all given function \(f(x)\in C_{[a,b]}\), \(\forall \varepsilon>0\), \(\exists p(x)\) which is an algebraic function, s.t.

\[ \|f(x)-p(x)\|< \varepsilon \]

Prove it by Bernstein Polynomials.

But what we care more about is whether we can find a polynomial of degree no more than \(n\) to approximate function \(f(x)\)?

The answer to the above question results in Chebyshev Approximation.

Consider set

\[ P_n(x)=\text{span}\{1,x,\cdots, x^n\} \]

It is not hard to see that \(P_n\) is a subspace of \(C_{[a,b]}\) with \(n+1\) dimension. And the goal is to get

\[ \Delta(f,P_n)=\min_{p\in P_n}\|f(x)-p(x)\|_\infty \]

which is called the best uniform approximation of \(f(x)\).

最佳一致逼近的特征 | Characteristics of Best Uniform Approximation

We can represent the best uniform approximation with the introduction of deviation points.

Deviation Point Set

we define

\[ \begin{cases} E^+(f)=\{x\in [a,b]: f(x)=\|f\|_\infty\} \\ E^-(f)=\{x\in [a,b]: f(x)=-\|f\|_\infty\} \end{cases} \]

to be positive and negative deviation point set, and define \(E(f)=E^+(f)\cup E^-(f)\) to be deviation point set.

Recall that \(f(x)\in C_{[a,b]}\), so \(E^+(f)\) and \(E^-(f)\) are both bounded closed set.

Alternating Point Set

If set \(\{x_1,x_2,\cdots,x_k\}\) \(\subset E(f)\) satisfies

\[ \begin{cases} a\leq x_1<x_2<\cdots <x_k\leq b \\ f(x_j)=-f(x_{j+1}), \quad j=1,2\cdots,k-1 \end{cases} \]

then we call the above set Alternating Point Set. We call them them maximal alternating point set, if there does not exist an alternating point set of size larger than \(k\).

We can construct maximal alternating point set by the following way.

Construction of Maximal Alternating Point Set

Let \(S_1=\inf E^+(f)\cup E^-(f)\), construct \(x_k\) recursively

\[ x_k=\inf S_k \]

where

\[ S_{k+1}=\begin{cases} E^-(f)\cap [x_k,b],\quad x_k\in E^+(f)\\ E^+(f)\cap [x_k,b],\quad x_k\in E^-(f) \end{cases} ,\quad k=1,2 \cdots \]

We assure \(x_k\in S_k\) when we let \(x_k=\inf S_k\) because \(S_k\) is bounded closed set. The recursion assures \(x_{k+1}\) is the minimal deviation point with opposite sign to \(x_k\), which of course assures \(\{x_i\}\) is monotonically increasing. We can prove that, if \(f\neq 0\), the above constructed set is maximal and finite(to be proved by readers).

Property of best uniform approximation on \(C_{[a,b]}\)

Assume \(f\in C_{[a,b]}\) and \(f\notin P_n\). If \(p(x)\) is the best uniform approximation polynomial of \(f(x)\) on \([a,b]\), then \(f-p\) has at least \(n+2\) alternating points.

By contradiction.

Vallée-Poisson Theorem

Assume \(f\in C_{[a,b]}\), if there exists polynomial \(p\in P_n\), such that \(f-p\) has at least \(n+2\) points \(x_1,x_2,\cdots,x_{n+2} \in [a,b]\) alternated with positive and negative, then

\[ \Delta(f,P_n)\geq \mu=\min_{1\leq i\leq n+2}|f(x_i)-g(x_i)| \]

By contradiction.

Chebyshev Theorem

For all \(f\in C_{[a,b]}\), \(f\notin P_n\), \(p\in P_n\) is the best uniform approximation of \(f\) if and only if, \(f-p\) has at least \(n+2\) alternating points on \([a,b]\).

By Sequence Theorem.

Uniqueness of best uniform approximation

If \(f\in C_{[a,b]}\), then there exists a unique polynomial \(p\in P_n\) that is the best uniform approximation of \(f\).

By prove two best uniform approximation polynomials \(p_1,p_2\) are the same.

Prove Uniqueness.

Assume \(p_1,p_2\in P_n\) are both the best uniform approximation polynomials of \(f\),i.e.

\[ \|f-p_1\|_\infty = \|f-p_2\|_\infty = \min_{p\in P_n}\|f-p\|_\infty = \Delta(f,P_n) \]

then define \(p_0=(p_1+p_2)/2\), then

\[ \|f-p_0\|_\infty\leq \frac{1}{2}(\|f-p_1\|_\infty+ \|f-p_2\|_\infty) =\Delta(f,P_n) \]

which means \(p_0\) is also the best uniform approximation polynomial of \(f\). By Chebyshev Theorem, there exists \(n+2\) alternating points \(x_0,x_1,\cdots,x_{n+1}\).

Note that

\[ \Delta(f,P_n)=|f(x_k)-p_0(x_k)|\leq \frac{1}{2}(|f(x_k)-p_1(x_k)|+|f(x_k)-p_2(x_k)|)\leq \Delta(f,P_n),\quad k=0,1,\cdots,n+1 \]

which means "=" must holds, i.e.

\[ |f(x_k)-p_1(x_k)| = |f(x_k)-p_2(x_k)| = \Delta(f,P_n),\quad k=0,1,\cdots,n+1 \]

and

\[ f(x_k)-p_1(x_k) = f(x_k)-p_2(x_k),\quad k=0,1,\cdots,n+1 \quad \text{(by Triangular inequation)} \]

which means \(p_1(x_k)=p_2(x_k)\), that is, \(p_1-p_2\) has \(n+2\) roots, so \(p_1=p_2\).

第一类Chebyshev多项式 | First Class Chebyshev Polynomials

Coonsider \(n+1\) extreme points of \(cosn\theta\) on \([0,\pi]\). If \(p(x)\) is the polynomial of best uniform approximation of \(f\) on \([a,b]\), then \(f-p\) has at least \(n+2\) alternating points.

We define \(x = \cos\theta\) and

\[ \begin{align*} T_n(x)&=\cos (n\cos^{-1}x)\\ &=\cos(n \theta)\\ &=\sum\limits_{k=1}^n a_k(\cos\theta)^i\\ &=\sum\limits_{k=1}^n a_k x^i \end{align*} \]

is a polynomial of degree \(n\), i.e. \(T_n(x)\in P_n\). It is easy to see the resursive definition

\[ \begin{cases} T_0(x)=1, T_1(x)=x,\\ \displaystyle T_{n+1}(x)=2xT_n(x)-T_{n-1}(x),\quad n=1,2,\cdots,n-1 \end{cases} \]

By definition, we have properties for first class Chebyshev Polynomials.

Properties of First Class Chebyshev Polynomials

(1) Prove the recursive form of definition.

(2) the coefficient of item with the highest degree of \(T_n(x)\) is \(1/2^{n-1}\).

(3) \(|T_n(x)|\leq 1\), \(\forall |x|\leq 1\).

(4) \(T_n(x)\) has \(n\) different real roots

\[ \cos\left[\frac{(2k-1)\pi}{2n}\right],\quad k=1,2,\cdots,n \]

(5) \(\{\cos(k\pi/n)\), \(k=0,1,\cdots,n\}\) are a maximal alternating point set of \(T_n(x)\) on \([-1,1]\).

(6) \(T_n(x)=(-1)^nT_n(-x)\).

(7)

\[ \int_{-1}^1\frac{T_m(x)T_n(x)}{\sqrt{1-x^2}}dx=\begin{cases}\pi, \quad &m=n=0;\\ \pi/2, \quad &m=n\neq 0;\\ 0,\quad &m\neq n.\end{cases} \]

(1) by combination property of triangular functions.

(7) by the orthogonal property of triangular functions.

With the above property, we can see the following theorem.

\(T_n(x)\) is the is the best uniform approximation of 0

\(T_n(x)\) is the best uniform approximation of function \(f\equiv 0\), that is, \(\forall\) Monicpolynomial \(p\in P_n\)

\[ \|p\|\geq |T_n|/2^{n-1} = 2^{1-n} \]

Use \(n\) Polynomial \(P_n(x)\) to approximate function \(f\) on region \([-1,1]\), its remainder

\[ \begin{align*} |P_n(x)-f(x)|=|R_n(x)|&=\left|\frac{f^{(n+1)}(\xi)}{(n+1)!}\prod_{i=0}^n(x-x_i)\right|\\ &\leq \max\limits_{x\in [-1,1]}\left|f^{(n+1)}(x)\right|\frac{1}{(n+1)!} \left| \prod_{i=0}^n(x-x_i)\right|\\ &\leq \max\limits_{x\in [-1,1]}\left|f^{(n+1)}(x)\right|\frac{1}{(n+1)!}\left| \frac{T_{n+1}(x)}{2^{n}} \right|\\ &\leq \frac{1}{(n+1)!} \frac{1}{2^{n}}\max\limits_{x\in [-1,1]}\left|f^{(n+1)}(x)\right| \end{align*} \]
  • Minimizing Approximation Error on Arbitrary Intervals

The technique for choosing points to minimize the interpolating error is extended to a general closed interval \([a, b]\) by using the change of variables

\[ \tilde{x} = \frac{1}{2}[(b-a)x+a+b] \]

Q. Find the best approximating polynomial of \(f (x) = e^x\) on \([0, 1]\) such that the absolute error is no larger than \(0.5\times 10^4\).

\(a=0\), \(b=1\), so

\[ x=\frac{a+b}{2}+\frac{b-a}{2}t = \frac{1}{2}(t+1), \quad t\in [-1,1] \]

So the actual function to be approximated is

\[ g(t) = f\left[\frac{1}{2}(t+1)\right] = e^{\frac{1}{2}(t+1)}, \quad t\in [-1,1] \]
\[ \max\limits_{x\in [-1,1]}\left|g^{(n+1)}(x)\right| = \max\limits_{x\in [-1,1]}\left|\frac{1}{2^{n+1}}e^{\frac{1}{2}(t+1)}\right|=\frac{e}{2^{n+1}} \]

So

\[ \begin{align*} |P_n(x)-f(x)|&\leq \frac{1}{(n+1)!} \frac{1}{2^{n}}\max\limits_{x\in [-1,1]}\left|g^{(n+1)}(x)\right| \\ &= \frac{1}{(n+1)!} \frac{1}{2^{n}}\frac{e}{2^{n+1}} \end{align*} \]

幂级数的降维 | Economization of Power series

This part is also called Reducing the Degree of Approximating Polynomials.

Consider approximating an arbitrary \(n\)th-degree polynomial

\[ P_n(x) = a_nx^n+a_{n-1}x^{n-1}+\cdots+a_1x+a_0, \quad x\in [−1, 1] \]

with a polynomial of degree at most \(n − 1\).

To let \(\max\limits_{x\in [-1,1]}|P_n(x)-P_{n-1}(x)|\) to be mininal, equals to let it be

\[ a_n\tilde{T}_n(x) = a_n T_n(x)/2^{n-1} \]

附录: 有限区间上的正交多项式 | Appendix: Orthogonal polynomials on \(L^2_\rho[a,b]\)

性质 | Properties

Properties

If \(\omega_0(x), \omega_1(x),\cdots\) are orthogonal polynomials on space \(L^2_\rho[a,b]\) by orthogonalizing power series, then is must follow

(i) \(\omega_n(x)\) is a \(n\)th algebraic polynomial.

(ii) \(\forall p \in P_k\), \(k\leq n\), \(p\) can be represented as

\[ p=\sum_{i=0}^na_i\omega_i(x) \]

(iii) \(\omega_n(x)\) is orthogonal to all polynomials whose degree is less than \(n\), that is,

\[ \int_{a}^b\rho(x)\omega_n(x)p_{n-1}(x)dx=0 \]

首一正交多项式 | Construction of Monic Orthogonal Polynomials

Construction of Monic Orthogonal Polynomials

Assume \(\{\overline{\omega}_i(x)\}_{i=0}^\infty\) are Monic Orthogonal Polynomials, then they satisfy the following recurrence relation

\[ \overline{\omega}_{n+1}(x)=(x-B_n)\overline{\omega}_n(x)-C_n\overline{\omega}_{n-1}, \quad n=1,2,\cdots \]

where

\[ \begin{align*} B_n &= \frac{(x\overline{\omega}_{n},\overline{\omega}_{n})}{(\overline{\omega}_{n},\overline{\omega}_{n})} \\ C_n&=\frac{( \overline{\omega}_{n}, \overline{\omega}_{n})}{(\overline{\omega}_{n-1},\overline{\omega}_{n-1})} \end{align*} \]

By using the property of orthogonal polynomials.

To simplify the notation, we temporarily use \(\omega_n(x)\) to replace \(\overline{\omega}_n(x)\).

We focus on \(x\omega_n(x)\), which is a \(n+1\)th polynomial, so it can be represented by \(\omega_0,\omega_1,\cdots,\omega_n\), i.e.

\[ \begin{equation} x\omega_n(x)=\omega_{n+1}(x)+\sum_{i=0}^nc_i\omega_i(x) \label{eq1} \end{equation} \]

where \(c_i\) are parameters to be determined.

Now we notice that \((\omega_n,\omega_s)=0\), \(s\leq n-1\), so we first employ inner product on both sides of \(\ref{eq1}\) with \(\omega_s\) (\(s=0,1,\cdots,n-2\))

\[ \begin{equation} (x\omega_n, \omega_s)=(\omega_{n+1},\omega_s)+\sum_{i=1}^nc_i(\omega_i,\omega_s)\label{eq2} \end{equation} \]

Because we have an exact meaning of inner product, that is, integral form, so we have \((x\omega_n, \omega_s)=(\omega_n, x\omega_s)\), then

for \(s=0,1,\cdots,n-2\), we get

\[ (x\omega_n, \omega_s)=0,\quad (\omega_{n+1},\omega_s)=0 \]

and

\[ (\omega_i,\omega_s)=0,\quad i\geq s+1 \text{ or }i\leq s-1 \]

So equation \(\ref{eq2}\) becomes

\[ 0=0+c_s(\omega_s,\omega_s) \]

which means \(c_s=0\), \(s=0,1,\cdots n-2\). Then we rewrite equation \(\ref{eq1}\)

\[ \begin{equation} x\omega_n(x)=\omega_{n+1}(x)+c_n\omega_n(x)+c_{n-1}\omega_{n-1}(x) \label{eq3} \end{equation} \]

In a similar way, we try employing inner product on both sides of the above equation \(\ref{eq3}\) with \(\omega_{n-1}\)

\[ (x\omega_n, \omega_{n-1})=0+c_{n-1}(\omega_{n-1},\omega_{n-1}) \]

which gives \(c_{n-1}=(x\omega_n, \omega_{n-1})/(\omega_{n-1},\omega_{n-1})\). Notice

\[ (x\omega_n, \omega_{n-1})=(\omega_n, x\omega_{n-1})=(\omega_n, \omega_{n}) \text{(by representing } x\omega_{n-1} \text{ again)} \]

so \(c_{n-1}=(\omega_n, \omega_{n})/(\omega_{n-1},\omega_{n-1})\).

In a similar way, employing inner product on both sides of the above equation \(\ref{eq3}\) with \(\omega_{n}\)

\[ (x\omega_n, \omega_{n})=0+c_{n}(\omega_{n},\omega_{n}) \]

which gives \(c_{n}=(x\omega_n, \omega_{n})/(\omega_{n},\omega_{n})\).

Rewrite equation \(\ref{eq3}\) and we prove the theorem.

零点分布 | Roots of Orthogonal Polynomials

Roots of Orthogonal Polynomials

\(n\)th Orthogonal Polynomial \(\omega_n(x)\) has \(n\) distinct roots on \([a,b]\).

By contradiction. First show that \(\omega_n(x)\) must have root, and then show it is not multiple root. Finally show the number of roots must be equal to \(n\). All the proof can be done by making use of properties.

常见的正交多项式 | Common Orthogonal Polynomials

  • Legendre Polynomial (on \(L^2[-1,1]\))

  • First class Chebyshev Polynomial (on \(L^2_\rho[-1,1]\) with \(\rho=1/\sqrt{1-x^2}\))

  • Second class Chebyshev Polynomial (on \(L^2_\rho[-1,1]\) with \(\rho=\sqrt{1-x^2}\))

  • Laguerre Polynomial (on \(L^2(0,\infty)\) with \(\rho=e^{-x}\))

  • Hermite Polynomial (on \(L^2(\infty,\infty)\) with \(\rho=e^{x^2}\))