Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

The Minimax Problem

Recall the interpolation error formula:

f(x)p(x)=f(n)(ξ)n!(xx1)(xx2)(xxn)|f(x) - p(x)| = \left|\frac{f^{(n)}(\xi)}{n!}\right| \cdot |(x-x_1)(x-x_2) \cdots (x-x_n)|

We can’t control the first factor (it depends on ff), but we can minimize the second by choosing nodes wisely.

The Minimax Problem: Choose x1,,xn[1,1]x_1, \ldots, x_n \in [-1, 1] to minimize

maxx[1,1](xx1)(xx2)(xxn)\max_{x \in [-1,1]} |(x-x_1)(x-x_2) \cdots (x-x_n)|

The Solution: Chebyshev Roots

For n=10n = 10, the Chebyshev roots are:

xk=cos((2k1)π20),k=1,,10x_k = \cos\left(\frac{(2k-1)\pi}{20}\right), \quad k = 1, \ldots, 10

These points cluster near the endpoints of the interval—exactly where Runge’s phenomenon causes problems with equally spaced nodes!

Chebyshev Polynomials

The first six Chebyshev polynomials T_0(x) through T_5(x). Note that each T_k oscillates between -1 and +1 exactly k+1 times on [-1, 1].

The first six Chebyshev polynomials T0(x)T_0(x) through T5(x)T_5(x). Note that each TkT_k oscillates between -1 and +1 exactly k+1k+1 times on [1,1][-1, 1].

Definition via Complex Exponentials

The most illuminating definition connects Chebyshev polynomials to Fourier series and complex analysis.

This definition directly links Chebyshev series to Fourier series and Laurent series—three fundamental tools connected through the change of variables x=cosθx = \cos\theta.

Why They’re Polynomials

From the definition, we verify the first few:

Recurrence Relation

The recurrence follows from the complex exponential definition:

12(z+z1)(zk+zk)=12(zk+1+zk1)+12(zk1+zk+1)\frac{1}{2}(z + z^{-1})(z^k + z^{-k}) = \frac{1}{2}(z^{k+1} + z^{-k-1}) + \frac{1}{2}(z^{k-1} + z^{-k+1})

which gives:

By induction, each Tk(x)T_k(x) is a polynomial of degree exactly kk.

Key Properties

  1. Roots: TnT_n has nn roots at the Chebyshev nodes

  2. Extrema: TnT_n oscillates between -1 and +1 exactly n+1n+1 times on [1,1][-1, 1]

  3. Leading coefficient: The coefficient of xnx^n in TnT_n is 2n12^{n-1} (for n1n \geq 1)

  4. Minimax property: 12n1Tn(x)\frac{1}{2^{n-1}} T_n(x) is the monic polynomial of degree nn with smallest maximum on [1,1][-1, 1]

Proof of Optimality

The proof uses a contradiction argument based on the alternation property.

Proof 1

Let p(x)=(xx1)(xxn)p(x) = (x - x_1) \cdots (x - x_n) be any monic polynomial of degree nn, and suppose

maxx[1,1]p(x)<12n1\max_{x \in [-1,1]} |p(x)| < \frac{1}{2^{n-1}}

Let q(x)=12n1Tn(x)q(x) = \frac{1}{2^{n-1}} T_n(x), which is monic and alternates between ±12n1\pm\frac{1}{2^{n-1}} at n+1n+1 points y1,,yn+1y_1, \ldots, y_{n+1}.

At each yiy_i:

  • If q(yi)=+12n1q(y_i) = +\frac{1}{2^{n-1}}, then q(yi)p(yi)>0q(y_i) - p(y_i) > 0

  • If q(yi)=12n1q(y_i) = -\frac{1}{2^{n-1}}, then q(yi)p(yi)<0q(y_i) - p(y_i) < 0

So qpq - p alternates sign at least n+1n+1 times, meaning it has at least nn roots.

But qpq - p has degree at most n1n - 1 (both are monic of degree nn), so it can have at most n1n - 1 roots. Contradiction!

Transformation to Arbitrary Intervals

The Chebyshev nodes are defined for [1,1][-1, 1]. For an arbitrary interval [a,b][a, b], we transform:

xk=ba2cos((2k1)π2n)+a+b2x_k = \frac{b-a}{2} \cos\left(\frac{(2k-1)\pi}{2n}\right) + \frac{a+b}{2}

This:

  1. Scales the Chebyshev points by ba2\frac{b-a}{2} (the ratio of interval lengths)

  2. Shifts the center from 0 to a+b2\frac{a+b}{2}

The minimum value of the node polynomial on [a,b][a, b] becomes (ba)n22n1\frac{(b-a)^n}{2^{2n-1}}.

Chebyshev vs. Equally Spaced: Runge’s Function

Consider f(x)=11+25x2f(x) = \frac{1}{1 + 25x^2} on [1,1][-1, 1].

Chebyshev vs. equally spaced nodes: Convergence comparison for polynomial interpolation. Chebyshev nodes achieve exponential convergence for smooth functions, while equally spaced nodes fail for Runge’s function.

Chebyshev vs. equally spaced nodes: Convergence comparison for polynomial interpolation. Chebyshev nodes achieve exponential convergence for smooth functions, while equally spaced nodes fail for Runge’s function.

NodesError behavior as nn \to \infty
Equally spacedError grows without bound
ChebyshevError decreases exponentially

This is the power of optimal node placement!

When to Use Chebyshev Interpolation

Chebyshev interpolation is particularly valuable for:

  1. Spectral methods for differential equations

  2. Gaussian quadrature for numerical integration

  3. Approximating smooth functions with high accuracy

However, for most practical data fitting, piecewise polynomials (splines) are often preferred because:

Chebyshev Series

Beyond interpolation, Chebyshev polynomials provide a powerful series representation for functions, analogous to Taylor series but with superior convergence properties.

Lipschitz Continuity

Any function fC1[1,1]f \in C^1[-1, 1] is Lipschitz (by the mean value theorem), but not every continuous function is Lipschitz.

The Chebyshev Series Theorem

The coefficients are given by the integral formula:

Connection to Fourier Series

Under the change of variables x=cosθx = \cos\theta, we have Tk(x)=cos(kθ)T_k(x) = \cos(k\theta) and:

dx1x2=dθ\frac{dx}{\sqrt{1-x^2}} = -d\theta

So the Chebyshev coefficients are precisely Fourier cosine coefficients of f(cosθ)f(\cos\theta)!

Why Chebyshev Over Taylor?

AspectTaylor SeriesChebyshev Series
CenterSingle point x0x_0Entire interval [1,1][-1, 1]
ConvergenceDisk of convergenceWhole interval (uniform)
TruncationBest near x0x_0Near-best everywhere
ComplexityO(n3)O(n^3) via VandermondeO(nlogn)O(n \log n) via DCT

For functions on an interval, Chebyshev series are almost always preferable.

From Interpolation to Approximation

So far, we’ve treated interpolation as a data-fitting problem: given points (xi,fi)(x_i, f_i), find a polynomial passing through them. But there’s a more powerful perspective.

Key insight: If we have a function ff, we can sample it at n+1n+1 Chebyshev nodes to get data, then interpolate. The natural question becomes: how well does the interpolant approximate the original function?

This shift—from fitting given data to approximating a known function—is the foundation of spectral methods. The Chebyshev interpolant pnp_n becomes a computational stand-in for ff: we differentiate, integrate, or solve equations using pnp_n instead of ff.

Convergence Rates: A Preview

The smoothness of ff determines the convergence rate:

SmoothnessCoefficient DecayApproximation Error
fCkf \in C^kO(nk1)O(n^{-k-1})O(nk)O(n^{-k}) (algebraic)
ff analyticO(ρn)O(\rho^{-n})O(ρn)O(\rho^{-n}) (exponential)

The jump from algebraic to exponential convergence when ff becomes analytic is dramatic—this is what makes spectral methods so powerful for smooth problems.

See the Spectral Accuracy chapter for the precise theorems, including: