Skip to content

Numerical Analysis

Basic ideas for Solving Equation

To find the solution of an equation, we hope to have an iteration method which takes good advantage of Computer resources like

\[ \begin{equation} \pmb{x}^{k} = f(\pmb{x}^{k-1}) \label{eq: iterative eq} \end{equation} \]

for \(k = 1, 2, \cdots n\). We use \(\pmb{x}\) instead of \(x\) because the above iteration method also applies to solving linear system.

Hopefully, if the above equation converges, that is, for \(k \rightarrow \infty\), \(\pmb{x}^{k-1}\rightarrow \pmb{x}^*\), \(\pmb{x}^{k}\rightarrow \pmb{x}^*\), and the equation becomes

\[ \pmb{x}^* = f(\pmb{x}^*) \]

where \(\pmb{x^*}\) is the sulution of the equation to be solved.

If the above thought gets right, then we can consider the converging speed of the iterative process, which makes great sense in practical applications. That is, in a given definition of distence,

\[ \frac{\|\pmb{x}^{k+1} - \pmb{x}^*\|}{\|\pmb{x}^{k}-\pmb{x}^*\|^\alpha} \]

to be small as much as possible for each \(k\).

Preliminary: Errors

If a real number \(x\) is denoted as \(0.d_1d_2d_3\cdots \times 10^{n}\), then

  • Truncation(截断) Error

is induced when

\[ \hat{x}=0.d_1d_2d_3\cdots d_k \times 10^{n} \]

for some definite \(k<\infty\)

  • Roundoff(舍入) Error

is induced when

\[ \hat{x}=0. \delta_1 \delta_2 \delta_3 \cdots \delta_k \times 10^{n} \]

for some definite \(k<\infty\)

where \(\delta_k >d_k\) if \(d_{k+1}>=5\).

t significant digits

The number \(p^*\) is said to approximate p to \(t\) significant digits(or figures) if \(t\) is the largest nonnegative integer for which the relative error

\[ e = \frac{\Delta p}{p}=\frac{\|p-p^*\|}{\|p\|}<5\times 10^{-t} \]

where \(p^*\) is the approximate number of the exact number \(p\).

  • for Chopping:
\[ \begin{align*} e &= \left|\frac{0.d_{k+1}d_{k+2}\cdots}{0.d_1d_2\cdots}\right| \times 10^{-k} \\ &\leq \frac{1}{0.1} \times 10^{-k} \quad \text{"=" for } d_{k+1}d_{k+2}\cdots\rightarrow\overline{9}\text{ and }d_{1}d_{2}\cdots\rightarrow 0 \\ &=10^{-k+1} \end{align*} \]
  • for rounding:
\[ \begin{align*} e &\leq \frac{0.5}{0.1} \times 10^{-k} \quad \text{"=" for } d_{k+1}d_{k+2}\cdots\rightarrow 5\overline{0}\text{ and }d_{1}d_{2}\cdots\rightarrow 0 \\ &=0.5\times 10^{-k+1} \end{align*} \]

Solutions of Equations in One Variables

the Bisection Method(二分法)

This method is quite intuitive. By choosing two end points \(a, b\), we get another point (Mid-point here)

\[ p = a+\frac{b-a}{2} \]

Then update \(a, b\) by evaluating whether \(f(p)>0\) or not to narrow down the interval.

What is interesting is the stopping procedure. Readers can see the following question if interested.

When we calculate the new point \(p\) based on \(a, b\), we need to judge whether \(p\) is an appropriate answer. Apart from \(f(p)=0\), which condition do you think is the best?

  1. \((b-a)/{|\min{(a, b)}|}<\epsilon\)
  2. \(|p-p_{prev}|=(b-a)/2 < \epsilon\)
  3. \(f(p)<\epsilon\)
  • 1
  • 2
  • 3

Choose 1, which is close to relative error, currently the best.

2: consider \(\{p_n\}=\sum\limits_{i=1}^{n}\frac{1}{k}\).

3: easy to see.

The converging speed can discribed as the following:

\[ |x_n- x^*| < \frac{(b-a)}{2^{n}} \]

Fixed-Point Iteration(不动点法)

As we said previously in Basic ideas for solving equation, we hope to find an iterative relation such that the converging point \(x^*\) is exactly what we want, which in this case, means that

\[ f(x^*) = 0 \]

So intuitively, we ask: Whether can we derive a relation from

\[ \begin{equation} f(x) = 0 \label{zero-equation} \end{equation} \]

to

\[ x = g(x) \]

for iterative method?

The answer is, of course, YES!

One of the easist way to transform is adding \(x\) to both sides of equation \(\ref{zero-equation}\), but in most cases this does not work. Because we rely on \(f(x)\) ifself for the convergence!

Thus, it is necessary to find the condition for \(x = g(x)\) to converge. The following theorem

gives a Sufficient condition.

不动点存在定理 | Fixed-Point Theorom

Let \(g \in C[a, b]\) be such that \(g(x) \in [a, b]\), for all \(x\) in \([a, b]\). Suppose, in addition, that \(g'\) exists on \((a, b)\) and that a constant \(0 < k < 1\) exists with

\[ |g'(x)|\leq k \quad \forall x \in (a, b) \]

Then for any initial number \(p_0 \in [a, b]\), the sequence \(\{p_n\}_{n=0}^{\infty}\) defined by

\[ p_n = g( p_{n−1}) \quad n \geq 1 \]

converges to the unique fixed point \(p \in [a, b]\).

Proving it is easily.

  • using the differential mean value theorem.

\(\forall n \geq 1, \exists \zeta_n \in (p_{n-1}, p) \subset (a, b)\), we have

\[ |p_n-p| = |g(p_{n-1}) - g(p)| = g'(\zeta_n)|p_{n-1}-p|\leq k|p_{n-1}-p| \]

by induction, we have

\[ |p_n-p|\leq k^{n}|p_0-p| \]

Let \(n \rightarrow \infty\), \(|p_n-p| \rightarrow 0\), that is, \(p_n\) converges to \(p\).

What we use in the proof will benefit us in identifying the speed of converging process.

Newton's Method(牛顿法)

This method is also a fixed-point method. There are two perspectives to get the inspirations.

  • version1: shrink the derivative of the iterative function \(g(x)\).
  • version2: using Taylor's expansion.

We can know that given a random function \(f(x)\), it may not be convergent to some point \(x^*\) for \(x^{k} = f(x^{k-1}) + x^{k-1}\) in a given interval. So the queation is, can we formulate a function \(g(x)\) such that \(x^{k} = g(x^{k-1})\) is convergent?

The answer is, again, YES!

The following content tells us we can formulate \(g(x)= x - f(x)/(f'(x))\) such that \(g'(x) < 1\) in a given interval.

Readers can easily see that

\[ \begin{align*} g'(x) &= 1 - \frac{f'^2(x)-f''(x)f(x)}{f'^2(x)}\\ &=\frac{f''(x)f(x)}{f'^2(x)} \end{align*} \]

if we add some constrictions, it can be easy to make \(g'(x)<1\).

Here we make use of the Taylor's expansion(or the derivatives) of the goal function.

Suppose that \(f \in C^2[a, b]\), Let \(x_0 \in [a, b]\) be an approximation to \(x^*\) such that \(f(x^*) \neq 0\) and \(|x_0-x^*|\) is "small". Consider the first Taylor polynomial for \(f(x)\) expanded at \(x_0\):

\[ f(x) = f(x_0) + (x-x_0)f'(x_0) + \frac{(x-x_0)^2}{2}f''(\zeta). \]

If we let \(x = x^*\), and according to \(f(x^*)=0\), we get

\[ 0 = f(x_0) + (x^*-x_0)f'(x_0) + \frac{(x^*-x_0)^2}{2}f''(\zeta) \]

neglecting the square item, we get

\[ 0 \approx f(x_0) + (x^*-x_0)f'(x_0) \]

to represent \(x^*\), we get

\[ x^* = x_0 - \frac{f(x_0)}{f'(x_0)} \]

Then we can define the iterative relation as

\[ x_n = x_{n-1} - \frac{f(x_{n-1})}{f'(x_{n-1})} \]

The following statement guarrantees the convergence of the above iterative method.

牛顿法收敛条件 | conditions for convergence of Newton's method

Let \(f \in C^2[a, b]\). If \(p \in (a, b)\) is such that \(f (p) = 0\) and \(f'( p) \neq 0\), then there exists a \(\delta > 0\) such that Newton’s method generates a sequence \(\{p_n\}_{n=1}^{\infty}\) converging to \(p\) for any initial approximation \(p_0 \in [p − \delta, p + \delta]\).

Prove it.

  • make use of the condition \(f(p) = 0\) and \(f'(p)\neq 0\)

for \(x \in (a, b)\), we aim to find a narrower interval \((x^*-\delta, x^*+\delta)\) to have \(g(x)\) map into itself. That is,

\[ g(x)\leq k, \forall k\in (0,1) \]

Firstly, \(f'(p)\neq 0\) implies that \(\exists \delta_1 >0\) such that \(f'(x)\neq 0, \forall x \in [x^* - \delta, x^*+\delta]\subset [a, b]\).

THus, we have

\[ g'(x) = \frac{f(x)f''(x)}{(f'^2(x))} \]

capable of dividing non-zero numbers.

Since \(f\in C^2[a,b]\), we have \(g' \in C^1[x^*-\delta_1, x^*+\delta_1]\)
for the exact solution \(x^*\), we have \(f(x^*)=0\), so

\[ g'(x^*) = 0 \]

which implies that \(\exists 0<\delta < \delta_1\), such that

\[ g'(x)\leq k, \forall k \in [x^*-delta, x^*+\delta] \]

By differential Mean Value Theorem, for \(x \in [x^*-delta, x^*+\delta], \exists \zeta \in [x, x^*]\) such that

\[ |g(x)-g(x^*)|=g'(\zeta)|x - x^*|\leq k|x-x^*|<|x-x^*| \]

which means that \(g\) maps into itself. By Fixed-Point Theorom, the sequence defined by Newton's method converges.

Secant Method

It may not be easy to find derivarive of function \(f\), so we can use difference instead. That is, we have to store two adjacent points for calculating differnce

\[ f'(x^{k}) \approx \frac{f(x^{k}) - f(x^{k-1})}{x^{k}-x^{k-1}} \]

generate \(p_{k+1}\) using the above approximation and iterate.

Order of Convergence

So how to identify the speed of convergence? The following definition gives a glimpse.

Suppose \(\{p_n\}_{n=1}^{\infty}\) is a sequence that converges to \(p\), with \(p_n \neq p (\forall n)\). If positive constants \(\lambda\) and \(\alpha\) exist with

\[ \lim_{n\rightarrow \infty}{\frac{|p_{n+1}-p|}{|p_n-p|^{\alpha}}}=\lambda \]

then \(\{p_n\}_{n=0}^{\infty}\) converges to \(p\) of order \(\alpha\), with asymptotic error constant \(\lambda\).

  • (i) If \(\alpha=1 (\lambda<1)\), the sequence is linearly convergent.
  • (ii) If \(\alpha=2\), the sequence is quadratically convergent.

The following theorem gives a sufficient condition for linear convergence.

线性收敛的充分条件 | sufficient condition of linear convergence

Let \(g \in C[a, b]\) be such that \(g(x) \in [a, b], \forall x \in [a, b]\). Suppose, in addition, that \(g\) is continuous on \((a, b)\) and a positive constant \(k < 1\) exists with

\[ |g'(x)|\leq k \quad \forall x \in (a, b) \]

If \(g'(p) \neq 0\), then for any number \(p_0=p\) in \([a, b]\), the sequence

\[ p_n=g(p_{n-1}) \quad \forall n \geq 1 \]

converges only linearly to the unique fixed point \(p\) in \([a, b]\).

Prove it.

Prove that linear convergence represents \(\exists alpha=1, lambda\), such that \(\lim\limits_{n\leftarrow \infty}\frac{|p_{n+1}-p^*|}{|p_{n}-p^*|} = \lambda\).

multiple roots

We see that the speed of convergence is limited by multiple roots.

Here we have modified Newton's Method:

\[ g(x)= x - \frac{f(x)f'(x)}{f'^2(x)-f(x)f''(x)} \]

Accelerating convergence

Aitken's \(\Delta^2\) Method

Suppose \(\{p_n\}_{n=0}^{\infty}\) is a linearly convergent sequence with limit \(p\). To motivate the construction of a sequence \(\{\hat{p}_n\}_{n=1}^{\infty}\) that converges more rapidly to \(p\) than does \(\{p_n\}_{n=0}^{\infty}\), let us first assume that the signs of \(p_n-p\), \(p_{n+1}-p\) and \(p_{n+2}-p\) agree and that \(n\) is sufficiently large that

\[ \frac{p_{n+1}-p}{p_n-p} \approx \frac{p_{n+2}-p}{p_{n+1}-p} \]

Then solving for \(p\) gives

\[ p \approx \frac{p_{n+2}p_n-p_{n+1}^2}{p_{n+2}-2p_{n+1}+p_n} \]

And to get \(p_n\) out gives

\[ \begin{align*} p &\approx p_n - \frac{(p_{n+1}-p_n)^2}{p_{n+2} - 2p_{n+1} + p_n}\\ \Rightarrow \hat{p}_n &= p_n - \frac{(\Delta p_n)^2}{\Delta p_{n+1}-\Delta p_{n}} \quad \text{(denote $\Delta p_n = p_{n+1} - p_n$)}\\ &= p_n - \frac{(\Delta p_n)^2}{\Delta^2 P_{n}} \end{align*} \]

Steffensen's Method

The following thought is based on that the generated sequence \(\hat{p}\) is a better approximation to true \(p^*\). We make use of the constructed sequence \(\{\hat{p}_n\}\) to update the original sequence \(\{p_n\}\). That is, after generating a new \(\hat{p}\), we can update \(p_0 \leftarrow p\).

Interpolation and Polynomial Approximation

Lagrange Interpolating Polynomial

Inspired by 加权平均.

There exists and only exists a \(n\)th Lagrange interpolating polynomial (拉格朗日基函数)

\[ L_n(x) = \sum_{i=0}^{n}l_i(x)y_i = \begin{bmatrix} l_0(x) & l_1(x) &l_2(x) & \cdots & l_n(x) \end{bmatrix} \begin{bmatrix} y_0 \\ y_1 \\ \vdots \\ y_n \end{bmatrix} = \Phi_n(x)\vec{y} \]

such that for each pair of given points \((x_i, y_i)\), \(i = 0, 1,2,\cdots n\), we have \(y_i = L_n(x_i)\).

Here we can consider \(l_i(x)\) as a base of a linear space \(\mathcal{P}_n(x)\), and it can be displayed by natural base \(1, x, x^2, \cdots x^n\). To be more specific,

\[ l_i(x) = \prod_{j=0 \atop j \neq i }^{n}\frac{(x - x_i)}{(x_i-x_j)} \quad i=0,1,\cdots n \]

readers can prove the above \(n+1\) polynomials are linearly irrelevant.

In fact, if we assume \(P_n(x) = \sum\limits_{i=0}^{n}a_ix^i\)(natural base), and to get the parameters \(\{a_i\}\) such that \(P_n(x_i) = y_i\), we have to solve the following linear system

\[ \begin{bmatrix} 1 & x_0 & x_0^2 &\cdots & x_0^n \\ 1 & x_1 & x_1^2 &\cdots & x_1^n \\ \vdots & \vdots & \vdots & &\vdots \\ 1 & x_n & x_n^2 &\cdots & x_n^n \end{bmatrix} \begin{bmatrix} a_0 \\ a_1 \\ \vdots \\ a_n \end{bmatrix}= \begin{bmatrix} y_0 \\ y_1 \\ \vdots \\ y_n \end{bmatrix} \]

which is a little tedious.

Error

The following theorem gives the error bound.

拉格朗日基函数的余项 | The remainder of Lagrange interpolating polynomial

Suppose \(x_0, x_1, \cdots , x_n\) are distinct numbers in the interval \([a, b]\) and \(f \in C^{n+1{[a, b]\). Then, for each \(x \in [a, b]\), a number \(\xi(x)\) (generally unknown) between \(x_0, x_1, \cdots , x_n\), and hence in \((a, b)\), exists with

\[ f(x) = P(x) + \frac{f^{n+1}(\xi(x))}{(n+1)!}\prod_{i=0}^{n}(x - x_i) \]

where \(P(x)\) is the Lagrange interpolating polynomial.

Prove it.

  • Using a function
\[ g(t) = f(t) - P(t) - [f(t)-g(t)]\prod_{i=0}^{n}\frac{(t-x_i)}{x-x_i} \]

which is \(C^{n+1}[a, b]\). Note that \(f(x_k)=0\) and \(f(x)=0\) for \(x \in [a, b]\). We can see that \(n+1+1=+\) zero points here.

  • Using generalized Rolle's Theorem: There exists \(\xi(x) \in [a, b]\) such that \(g^{n+1}(\xi(x))=0\)

And make \(n+1\) derivatives to \(g(t)\)(corresponding to \(t\)), and we can get the result.

Compared to the remainder of Taylor's extension

\[ R_n(x) = \frac{f^{n+1}(\xi(x))}{(n+1)!}(x-x_0)^{n+1} \]

Because \(\xi(x)\) is usually unknown, so we often use a number \(x' \in [a, b]\) such that

\[ |f^{n+1}(\xi(x))|\leq |f^{n+1}(x')| \]

Neville's Method

An interpolation polynomial for a set of points \(A =\{x_0, x_1, \cdots x_n \}\) can be expressed by two polynomials that interpolate \(A\) \ \(\{ x_i\}\) and \(A\) \ \(\{ x_j\}\). That is,

\[ P_n(x) = \frac{(x-x_i)P_{0,1,\cdots,i-1,i+1,\cdots,n}(x)-(x-x_j)P_{0,1,\cdots,j-1,j+1,\cdots,n}(x)}{(x_j-x_i)} \]

where \(P_{0,1,\cdots,i-1,i+1,\cdots,n}(x)\) and \((x-x_j)P_{0,1,\cdots,j-1,j+1,\cdots,n}(x)\) denotes the polynomial that interpolates \(A\) \ \(\{ x_i\}\) and \(A\) \ \(\{ x_j\}\) respectively.

Divided Differences

If we rewrite the \(n\)th Lagrange polynomial \(P_n(x)\) into another form:

\[ P_n(x) = a_0+a_1(x-x_0)+a_2(x-x_0)(x-x_1)+\cdots+a_n(x-x_0)\cdots(x-x_n) \]

By letting \(x =x_0, x_1,\cdots, x_n\), we get

\[ \begin{align*} P_n(x_0) &= a_0\\ P_n(x_1) &= a_0 + a_1(x_1-x_0)\\ P_n(x_2) &= a_0+a_1(x_2-x_0)+a_2(x_2-x_0)(x_2-x_1)\\ &\vdots\\ P_n(x_2) &= a_0+a_1(x_2-x_0)+a_2(x_2-x_0)(x_2-x_1)+\cdots+a_n(x_2-x_0)\cdots(x_2-x_n) \end{align*} \]

Then we can define:

\[ \begin{align*} f_n[x_0] &\overset{\Delta}{=} f(x_0) = a_0\\ f_n[x_0, x_1] &\overset{\Delta}{=} \frac{f(x_1) - f(x_0)}{x_1-x_0} = a_1\\ f_n[x_0, x_1, x_2] &\overset{\Delta}{=} \frac{f[x_1,x_2]-f[x_0,x_1]}{x_2-x_0} \\&= \frac{\frac{f(x_2)-f(x_1)}{x_2-x_1}-\frac{f(x_1)-f(x_0)}{x_1-x_0} }{x_2-x_0} \\&= \frac{\frac{f(x_2)-f(x_0)-(f(x_1)-f(x_0))}{x_2-x_1}-a_1}{x_2-x_0} \\ &= \frac{\frac{f(x_2)-f(x_0)}{x_2-x_1} - \frac{a_1(x_1-x_0)}{x_2-x_1}-\frac{a_1(x_2-x_1)}{x_2-x_1}}{x_2-x_0} \\ &= \frac{\frac{f(x_2)-f(x_0)}{x_2-x_1} - \frac{a_1(x_2-x_0)}{x_2-x_1}}{x_2-x_0}\\ &= \frac{f(x_2)-a_0 - a_1(x_2-x_0)}{(x_2-x_0)(x_2-x_1)}=a_2\\ &\vdots\\ f[x_1,x_2,\cdots, x_n] &\overset{\Delta}{=} \frac{f[x_1,x_2,\cdots,x_n] - f[x_0,x_1,\cdots,x_{n-1}]}{x_n-x_0} = a_n \end{align*} \]

We can prove the above definition \(f[x_0,x_1\cdots,x_n]\), which is called divided difference, to be equal to \(a_n\) by induction.

This iterative method is quite useful in determining the parameters of \(n\)th Lagrange polynomial.

\[ P_n(x) = f[x_0] + \sum_{i=1}^{n}\left(f[x_0,x_1,\cdots,x_i]\prod_{j=0}^{i-1}(x-x_j)\right) \]

Some properties are the followings.

差分与微分均值的关系 | Relationship of DD & Maen Value Differentials

Suppose that \(f \in C^n[a, b]\) and \(x_0, x_1, \cdots, x_n\) are distinct numbers in \([a, b]\). Then a number \(\xi\) exists in \((a, b)\) with

\[ f[x_0,x_1\cdots,x_n] = \frac{f^{(n)}(\xi)}{n!} \]

It is like

\[ f[x_0,x_1] = \frac{f(x_1)-f(x_0)}{x_1-x_0} = f'(\xi) \]

but add \(n!\) to the denominator.

The following relation gives a slightly quicker way to compute \(f[x_0,x_1,\cdots,x_n]\):

Another expression of calculating divided difference

\[ f[x_0,x_1,\cdots,x_n] = \sum_{k=0}^{n}\frac{f(x_k)}{\omega'_{n+1}(x_k)} \]

where

\[ \omega_{n+1}(x) = \prod_{i=0}^{n}(x-x_i),\quad \omega'_{n+1}(x_j) = \prod_{i=0 \atop i \neq j}^{n}(x_j-x_i) \]

Hermite Polynomials

We want to consider the smoothness of interpolating polynomials, so we need to consider its derivatives, especially the first order derivative.

Definition of Osculating Polynomial

The osculating polynomial approximating \(f\) is the polynomial \(P(x)\) of least degree such that

\[ \frac{d^kP(x_i)}{dx^k} = \frac{d^kf(x_i)}{dx^k},\quad, \forall i = 0,1,\cdots, n, \forall k = 0,1,\cdots, m_i \]

Where \(n\) is the total number of sampling points and \(m_i\) is the degree of smoothness at point \(x_i\).

If \(m_i=1\) for all \(i=0,1,\cdots, n\), then the above polynomial is called Hermite Polynomial.

Composition of Hermite Polynomial

If \(f \in C^1[a, b]\) and \(x_0, \cdots , x_n \in [a, b]\) are distinct, the unique polynomial of least degree agreeing with \(f\) and \(f'\) at \(x_0,\cdots , x_n\) is the Hermite polynomial of degree at most \(2n + 1\) given by

\[ H_{2n+1}(x) = \sum_{j=0}^{n}f(x_j)H_{n,j}(x) + \sum_{j=0}^{n}f'(x_j)\hat{H}_{n,j}(x) \]

where, for \(L_{n, j}(x)\) denoting the \(j\)th Lagrange coefficient polynomial of degree \(n\), we have

\[ H_{n,j}(x) = [1-2(x-x_j)L'_{n,j}(x_j)]L^2_{n,j}(x),\quad \hat{H}_{n,j}(x) = (x-x_j)L^2_{n,j}(x) \]

The composition is usually a little tidious.

Cubic Spline Interpolation

We change \(\frac{d^kP(x_i)}{dx^k} = \frac{d^kf(x_i)}{dx^k}\) into equations between adjacent curves. Because we ofte do not know the derivatives of the point.

  • Natural Spline

Let \(S''(a) = 0\) and \(S''(b) = 0\).

  • Clamped Cubic Spline

Let \(S'(x_0) = f'(x_0)\) and \(S'(x_n) = f'(x_n)\).

Approximation Theory

Discrete Least Squares Approximation

Orthogonal Polynomials and Least Squares Approximation

If we choose

\[ E = (P-y,P-y)=\|P-y\|^2 \]

we can easily solve the approximation polynomial \(\sum\limits_{i=0}^\inftya_ix^i\) by letting \(\frac{\partial E}{\partial a_k}=0\), and get

\[ \sum_{j=0}^n(\varphi_k, \varphi_j)a_j=(\varphi_k,f),\quad k=0,1,\cdots,n \]

Chebyshev Polynomials and Economization of Power series

Coonsider \(n+1\) extreme points of \(cosn\theta\) on \([0,\pi]\).

We still have \(cos(n \theta)=\sum\limits_{k=1}^n a_kcos(\theta)^i\).

Define \(T_0(x)=1\), \(T_1(x)=x\), \(T_{n+1}(x)=2xT_n(x)-T_{n-1}(x)\), \(n=1,2,\cdots,n-1\)