Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Fixed Points

The basic idea: reformulate a root-finding problem as a fixed-point problem.

Root finding: f(x)=0Fixed point: x=g(x)\text{Root finding: } f(x) = 0 \quad \longrightarrow \quad \text{Fixed point: } x = g(x)

There are many ways to do this! Given f(x)=0f(x) = 0, you could write:

The choice of gg matters enormously for convergence—as we’ll see.

The Algorithm

The fixed-point iteration is beautifully simple.

That’s it. But whether this converges—and how fast—depends entirely on properties of gg.

Why the Choice of gg Matters

Consider finding the root of f(x)=x23=0f(x) = x^2 - 3 = 0 (i.e., finding 3\sqrt{3}).

Here are three valid reformulations:

All three have 3\sqrt{3} as a fixed point. But their behavior is dramatically different:

Existence and Uniqueness

When does a fixed point exist? When is it unique?

Proof 1 (Existence)

Define h(x)=xg(x)h(x) = x - g(x).

If g(a)=ag(a) = a or g(b)=bg(b) = b, we’re done—we’ve found a fixed point.

Otherwise, since g([a,b])[a,b]g([a,b]) \subseteq [a,b]:

  • g(a)>ag(a) > a, so h(a)=ag(a)<0h(a) = a - g(a) < 0

  • g(b)<bg(b) < b, so h(b)=bg(b)>0h(b) = b - g(b) > 0

By the Intermediate Value Theorem, there exists c(a,b)c \in (a,b) with h(c)=0h(c) = 0, i.e., g(c)=cg(c) = c.

Proof 2 (Uniqueness)

Suppose two fixed points c1<c2c_1 < c_2 exist. By the Mean Value Theorem:

c1c2=g(c1)g(c2)=g(ξ)c1c2ρc1c2|c_1 - c_2| = |g(c_1) - g(c_2)| = |g'(\xi)||c_1 - c_2| \leq \rho|c_1 - c_2|

for some ξ(c1,c2)\xi \in (c_1, c_2).

This implies (1ρ)c1c20(1-\rho)|c_1 - c_2| \leq 0. But ρ<1\rho < 1 and c1c2c_1 \neq c_2, so (1ρ)c1c2>0(1-\rho)|c_1 - c_2| > 0—a contradiction.

Convergence

Proof 3

Since cc is a fixed point, g(c)=cg(c) = c. Using the Mean Value Theorem:

xn+1c=g(xn)g(c)=g(ξn)xncρxnc|x_{n+1} - c| = |g(x_n) - g(c)| = |g'(\xi_n)||x_n - c| \leq \rho|x_n - c|

Applying this recursively:

xncρxn1cρ2xn2cρnx0c|x_n - c| \leq \rho|x_{n-1} - c| \leq \rho^2|x_{n-2} - c| \leq \cdots \leq \rho^n|x_0 - c|

Since ρ<1\rho < 1, we have ρn0\rho^n \to 0, so xncx_n \to c.

The Derivative at the Fixed Point

The key insight: g(c)|g'(c)| determines everything.

For our three reformulations of x23=0x^2 - 3 = 0:

Order of Convergence

Proof 4

Taylor expand g(xn)g(x_n) around cc:

xn+1=g(xn)=g(c)+g(c)(xnc)++g(p)(ξ)p!(xnc)px_{n+1} = g(x_n) = g(c) + g'(c)(x_n - c) + \cdots + \frac{g^{(p)}(\xi)}{p!}(x_n - c)^p

Since g(c)=cg(c) = c and the first p1p-1 derivatives vanish:

xn+1c=g(p)(ξ)p!(xnc)px_{n+1} - c = \frac{g^{(p)}(\xi)}{p!}(x_n - c)^p

Thus xn+1cCxncp|x_{n+1} - c| \approx C|x_n - c|^p with C=g(p)(c)/p!C = |g^{(p)}(c)|/p!.

This explains why g3g_3 converges so fast: g3(3)=0g_3'(\sqrt{3}) = 0 means at least quadratic convergence.

The Banach Fixed Point Theorem

The convergence results above are special cases of a fundamental principle that appears throughout mathematics.

Remark 2 (Why the Banach FPT Matters)

The Banach FPT is not just about scalar equations. The same principle governs:

  • Newton’s method for systems: The iteration is a contraction near the solution

  • Picard iteration for ODEs: Proves existence and uniqueness for y=f(t,y)y' = f(t,y)

  • Iterative linear solvers: Jacobi and Gauss-Seidel converge when the iteration matrix is a contraction

Whenever you see an iterative method that “works,” there’s often a contraction hiding underneath.

Advantages and Disadvantages

Advantages:

Disadvantages:

Design principle: Make g(c)|g'(c)| small. If g(c)=0g'(c) = 0, you get quadratic convergence—this is Newton’s method.