Fixed Point Iteration
University of Massachusetts Amherst
Fixed Points¶
The basic idea: reformulate a root-finding problem as a fixed-point problem.
Root finding: f(x)=0⟶Fixed point: x=g(x) There are many ways to do this! Given f(x)=0, you could write:
g(x)=x−f(x)
g(x)=x−cf(x) for any nonzero c
g(x)=x−f(x)/f′(x) (this gives Newton’s method!)
The choice of g matters enormously for convergence—as we’ll see.
The Algorithm¶
The fixed-point iteration is beautifully simple.
That’s it. But whether this converges—and how fast—depends entirely on properties of g.
Why the Choice of g Matters¶
Consider finding the root of f(x)=x2−3=0 (i.e., finding 3).
Here are three valid reformulations:
G1: From x2=3, we get g1(x)=3/x
G2: Add x to both sides: g2(x)=x−(x2−3)/2
G3: Divide by 2x: g3(x)=(x2+3)/(2x)
All three have 3 as a fixed point. But their behavior is dramatically different:
G1 cycles forever, never converging
G2 converges slowly (linearly)
G3 converges rapidly (quadratically—this is Newton’s method!)
Existence and Uniqueness¶
When does a fixed point exist? When is it unique?
Proof 1 (Existence)
Define h(x)=x−g(x).
If g(a)=a or g(b)=b, we’re done—we’ve found a fixed point.
Otherwise, since g([a,b])⊆[a,b]:
g(a)>a, so h(a)=a−g(a)<0
g(b)<b, so h(b)=b−g(b)>0
By the Intermediate Value Theorem, there exists c∈(a,b) with h(c)=0, i.e., g(c)=c.
Proof 2 (Uniqueness)
Suppose two fixed points c1<c2 exist. By the Mean Value Theorem:
∣c1−c2∣=∣g(c1)−g(c2)∣=∣g′(ξ)∣∣c1−c2∣≤ρ∣c1−c2∣ for some ξ∈(c1,c2).
This implies (1−ρ)∣c1−c2∣≤0. But ρ<1 and c1=c2, so (1−ρ)∣c1−c2∣>0—a contradiction.
Convergence¶
Proof 3
Since c is a fixed point, g(c)=c. Using the Mean Value Theorem:
∣xn+1−c∣=∣g(xn)−g(c)∣=∣g′(ξn)∣∣xn−c∣≤ρ∣xn−c∣ Applying this recursively:
∣xn−c∣≤ρ∣xn−1−c∣≤ρ2∣xn−2−c∣≤⋯≤ρn∣x0−c∣ Since ρ<1, we have ρn→0, so xn→c.
The Derivative at the Fixed Point¶
The key insight: ∣g′(c)∣ determines everything.
For our three reformulations of x2−3=0:
g1(x)=3/x: We have g1′(x)=−3/x2, so ∣g1′(3)∣=1. Right on the boundary—no convergence guaranteed (and indeed, it fails).
g2(x)=x−(x2−3)/2: We have g2′(x)=1−x, so ∣g2′(3)∣=∣1−3∣≈0.73. Linear convergence with rate ρ≈0.73.
g3(x)=(x2+3)/(2x): We have g3′(x)=1/2−3/(2x2), so g3′(3)=0. The derivative vanishes—this signals faster-than-linear convergence.
Order of Convergence¶
Proof 4
Taylor expand g(xn) around c:
xn+1=g(xn)=g(c)+g′(c)(xn−c)+⋯+p!g(p)(ξ)(xn−c)p Since g(c)=c and the first p−1 derivatives vanish:
xn+1−c=p!g(p)(ξ)(xn−c)p Thus ∣xn+1−c∣≈C∣xn−c∣p with C=∣g(p)(c)∣/p!.
This explains why g3 converges so fast: g3′(3)=0 means at least quadratic convergence.
The Banach Fixed Point Theorem¶
The convergence results above are special cases of a fundamental principle that appears throughout mathematics.
Remark 2 (Why the Banach FPT Matters)
The Banach FPT is not just about scalar equations. The same principle governs:
Newton’s method for systems: The iteration is a contraction near the solution
Picard iteration for ODEs: Proves existence and uniqueness for y′=f(t,y)
Iterative linear solvers: Jacobi and Gauss-Seidel converge when the iteration matrix is a contraction
Whenever you see an iterative method that “works,” there’s often a contraction hiding underneath.
Advantages and Disadvantages¶
Advantages:
Simple — Just iterate xn+1=g(xn)
Unifying framework — Newton’s method is a fixed-point iteration in disguise
Flexible — Many choices for g; can design for fast convergence
Generalizes — Extends to systems, ODEs, infinite dimensions (Banach FPT)
Disadvantages:
Not guaranteed — Can diverge if ∣g′(c)∣≥1
Sensitive — Different reformulations give wildly different behavior
Slow — Linear convergence when ∣g′(c)∣ is close to 1
Design principle: Make ∣g′(c)∣ small. If g′(c)=0, you get quadratic convergence—this is Newton’s method.