2.2 Fixed Point Iteration

{y=cos(x)y=x

Pasted image 20250213160742.webp|224$${\left{\begin{array}{l l}{y=x^{2}}\ {y=x}\end{array}\right.}$$
Pasted image 20250213160918.webp|209

Clearly:
If h(x) is differentiable on (a,b) with |h(x)|<1 for all x(a,b), then whenever x1,x2(a,b),

|h(x2)h(x1)|<|x2x1|

Conceptually:
Moreover, a function whose derivative has magnitude less than 1 can only cross the line y=x one time. Once it has crossed, it can never “catch up” because that would require a slope greater than 1, the slope of the line y=x.

Suppose h(x) is continuous on [a,b], differentiable on (a,b) with |h(x)|<1 for all x(a,b), and h([a,b])[a,b] Then h has a unique fixed point in [a,b].

Fixed Point Convergence Theorem Given a function f(x) with continuous first derivative and fixed point x^, if |f(x^)|<1 then there exists a neighbourhood of x^ in which fixed point iteration converges to the fixed point for any initial value in the neighbourhood.

If |f(x^)|<1, then fixed point iteration converges to x^ for any initial value in some neighbourhood of x^.

If |f(x^)|>1, then fixed point iteration escapes some neighbourhood of x^ for any initial value in the neighbourhood other than x^.

Otherwise inconclusive.

We can find bounds on the largest such interval by solving the equation $$|f′(x)|= 1$$

Comparison To Root Finding

f(x)=x has exactly the same solutions as the equation f(x)x=0, so finding fixed points of f(x) is equivalent to finding roots of g(x)=f(x)x.

Algorithm

Input: A differentiable function f with fixed point x^ an initial point x0 in a neighbourhood (x^δ,x^+δ) where |f(x)|<1, a desired tolerance ε, max iterations N.
Output: An approximate x close to a fixed point of the function f , or failure.

*Key observation: |xk+1xk|=|f(xk)xk|, so xk is close to a fixed point if xk+1 is close to xk .

Outline:

Important note: Here we are using |xk+1xk|=|f(xk)xk| as a proxy for error (but the actual error may exceed this in reality)