Command Palette

Search for a command to run...

Linear Methods of AI

Orthogonal Polynomials

Weighted Scalar Product

In numerical analysis, we often need to measure the difference between two functions using the maximum norm. This norm is defined as:

fg=maxx[1,1]f(x)g(x)\|f - g\|_\infty = \max_{x \in [-1,1]} |f(x) - g(x)|

In general, the maximum norm can become very large, especially near the endpoints of the interval where large errors often occur. Imagine measuring the height of a mountain only from its highest peak without considering the slope of its sides.

To address this problem, we can use a weighted scalar product that gives different emphasis to each part of the interval:

f,gω=11f(x)g(x)ω(x)dx\langle f, g \rangle_\omega = \int_{-1}^{1} f(x)g(x)\omega(x) \, dx

with the weight function:

ω(x)=11x2\omega(x) = \frac{1}{\sqrt{1-x^2}}

This weight function provides stronger emphasis on the interval endpoints, allowing us to suppress large errors that typically occur in those regions.

Orthogonalization and Chebyshev Polynomials

When we orthogonalize the monomial basis 1,x,,xn1, x, \ldots, x^n with respect to this weighted scalar product, we obtain orthogonal polynomials pkp_k that satisfy a two-level recurrence relation. This process is like rearranging building blocks so that each floor stands perfectly perpendicular to the others:

p0(x)=1p_0(x) = 1
p1(x)=xp_1(x) = x
pk+1(x)=4xpk(x)4pk1(x),k=1,,np_{k+1}(x) = 4xp_k(x) - 4p_{k-1}(x), \quad k = 1, \ldots, n

The norm of these polynomials is:

pkω={π,k=0π/2,k0\|p_k\|_\omega = \begin{cases} \sqrt{\pi}, & k = 0 \\ \sqrt{\pi/2}, & k \neq 0 \end{cases}
pk(1)=2k1p_k(1) = 2^{k-1}

If we perform normalization at x=1x = 1, we obtain the famous Chebyshev polynomials:

Tk(x)=21kpk(x)=cos(karccos(x)),k=0,1,2,T_k(x) = 2^{1-k}p_k(x) = \cos(k \arccos(x)), \quad k = 0, 1, 2, \ldots

This trigonometric form shows that Chebyshev polynomials are essentially transformations of cosine functions adapted for the interval [1,1][-1, 1].

There exists a theorem stating that two-way recurrence relations hold generally for polynomials derived from monomial bases in the space C([1,1])C([-1, 1]) with certain symmetry properties. The scalar product satisfies the relation:

p,xq=xp,q\langle p, xq \rangle = \langle xp, q \rangle

for all polynomials p,qPp, q \in P.

Gram-Schmidt Process

Through the Gram-Schmidt orthogonalization process from the basis 1,x,,xn1, x, \ldots, x^n, we obtain orthogonal polynomials p~k\tilde{p}_k with k=0,,nk = 0, \ldots, n. This process is like building a structural framework where each beam is positioned based on the previous beam with accurate calculations:

p~0(x)=1\tilde{p}_0(x) = 1
p~1(x)=xβ0\tilde{p}_1(x) = x - \beta_0
p~k+1(x)=(xβk)p~k(x)γkp~k1(x),k=1,,n\tilde{p}_{k+1}(x) = (x - \beta_k)\tilde{p}_k(x) - \gamma_k\tilde{p}_{k-1}(x), \quad k = 1, \ldots, n

with coefficients:

βk=xp~k,p~kp~k2,k=0,,n\beta_k = \frac{\langle x\tilde{p}_k, \tilde{p}_k \rangle}{\|\tilde{p}_k\|^2}, \quad k = 0, \ldots, n
γk=p~k2p~k12,k=1,,n\gamma_k = \frac{\|\tilde{p}_k\|^2}{\|\tilde{p}_{k-1}\|^2}, \quad k = 1, \ldots, n

The orthonormal polynomials are then obtained through:

pk=p~k1p~k,k=0,,np_k = \|\tilde{p}_k\|^{-1}\tilde{p}_k, \quad k = 0, \ldots, n

Legendre Polynomials

For orthonormalization of the basis 1,x,,xn1, x, \ldots, x^n with respect to the standard scalar product:

f,g=11f(x)g(x)dx\langle f, g \rangle = \int_{-1}^{1} f(x)g(x) \, dx

we obtain polynomials related to Legendre polynomials:

φk(x)=(2k2)!(k1)!22k12k1Lk1(x),k=1,,n+1\varphi_k(x) = \frac{(2k-2)!}{(k-1)!^2} \sqrt{\frac{2k-1}{2^{k-1}}} L_{k-1}(x), \quad k = 1, \ldots, n+1

where Lk(x)L_k(x) are Legendre polynomials defined through a two-level recurrence relation.

The Legendre polynomials themselves are defined as:

L0(x)=1L_0(x) = 1
L1(x)=xL_1(x) = x
Lk+1(x)=xLk(x)k24k21Lk1(x),k=1,,nL_{k+1}(x) = xL_k(x) - \frac{k^2}{4k^2-1}L_{k-1}(x), \quad k = 1, \ldots, n

With respect to the standard scalar product, Legendre polynomials produce an orthonormal system φi:[1,1]R\varphi_i : [-1, 1] \to \mathbb{R} with i=0,,ni = 0, \ldots, n that satisfies:

φi,φjt=δij\langle \varphi_i, \varphi_j \rangle_t = \delta_{ij}

Transformation to Different Intervals

When we have another approximation interval [a,b][a, b] different from the standard interval, we perform a linear variable substitution. This process is like changing the scale on a map from one region to another while maintaining the correct proportions:

x=a+ba2(t+1)[a,b]x = a + \frac{b-a}{2}(t+1) \in [a, b]
t=1+2ba(xa)[1,1]t = -1 + \frac{2}{b-a}(x-a) \in [-1, 1]

Then the differential becomes:

dx=ba2dtdx = \frac{b-a}{2}dt

On the interval [a,b][a, b] there exist polynomials φ~i:[a,b]R\tilde{\varphi}_i : [a, b] \to \mathbb{R} with i=0,,ni = 0, \ldots, n expressed as:

φ~i(x)=2baφi(1+2ba(xa))=2baφi(t)\tilde{\varphi}_i(x) = \sqrt{\frac{2}{b-a}}\varphi_i\left(-1 + \frac{2}{b-a}(x-a)\right) = \sqrt{\frac{2}{b-a}}\varphi_i(t)

Verification of Orthonormal Properties

We can verify that the orthonormal properties remain valid after transformation:

φ~i,φ~jx=abφ~i(x)φ~j(x)dx\langle \tilde{\varphi}_i, \tilde{\varphi}_j \rangle_x = \int_a^b \tilde{\varphi}_i(x)\tilde{\varphi}_j(x) \, dx
=ab2baφi(1+2ba(xa))= \int_a^b \sqrt{\frac{2}{b-a}}\varphi_i\left(-1 + \frac{2}{b-a}(x-a)\right)
×2baφj(1+2ba(xa))dx\times \sqrt{\frac{2}{b-a}}\varphi_j\left(-1 + \frac{2}{b-a}(x-a)\right) \, dx

With substitution, we obtain:

=112baφi(t)φj(t)ba2dt= \int_{-1}^1 \frac{2}{b-a}\varphi_i(t)\varphi_j(t) \frac{b-a}{2} \, dt
=11φi(t)φj(t)dt=φi,φjt=δij= \int_{-1}^1 \varphi_i(t)\varphi_j(t) \, dt = \langle \varphi_i, \varphi_j \rangle_t = \delta_{ij}

Thus:

f,φ~ix=abf(x)φi(1+2ba(xa))2badx\langle f, \tilde{\varphi}_i \rangle_x = \int_a^b f(x)\varphi_i\left(-1 + \frac{2}{b-a}(x-a)\right)\sqrt{\frac{2}{b-a}} \, dx

Gauss Approximation

In Gauss approximation with the scalar product:

f,g=11f(t)g(x)dx\langle f, g \rangle = \int_{-1}^1 f(t)g(x) \, dx

this method provides a very important foundation in numerical analysis. The orthogonal polynomials we have studied become the fundamental basis for developing efficient Gauss quadrature methods for various computational applications that require function approximation with high accuracy.