Newton-Raphson Method |
Suppose \(f\) is a function with a unique root \(x^*\) in \([a,b]\). Assume \(f''\) is continuous in \([a,b]\). To find \(x^*\), let \(x_0\) be a "good" initial approximation (i.e., \(|x^*-x_0|^2 \approx 0\)) where \(f'(x_0)\neq 0\). Using Taylor's Theorem on \(f\) about \(x_0\), we get \[f(x)=f(x_0)+\frac{f'(x_0)}{1!}(x-x_0) +\frac{f''(\xi)}{2!}(x-x_0)^2,\] for some number \(\xi\) (depends on \(x\)) between \(x_0\) and \(x\). Plugging \(x=x^*\), we get \[\begin{align*} & 0=f(x^*)=f(x_0)+\frac{f'(x_0)}{1!}(x^*-x_0) +\frac{f''(\xi)}{2!}(x^*-x_0)^2\\ \implies & f(x_0)+\frac{f'(x_0)}{1!}(x^*-x_0) =-\frac{f''(\xi)}{2!}(x^*-x_0)^2 \approx 0 \;\;(\text{since }|x^*-x_0|^2\approx 0) \end{align*}\] Now solving for \(x^*\), we get \[x^*\approx x_0-\frac{f(x_0)}{f'(x_0)}.\] So \(x_1=x_0-\frac{f(x_0)}{f'(x_0)}\) is an approximation to \(x^*\). Using \(x_1\), similarly we get \(x_2=x_1-\frac{f(x_1)}{f'(x_1)}\). Continuing the process, we get a sequence \(\{x_n\}\) to approximate \(x^*\) where \[\boxed{x_{n+1}=x_n-\frac{f(x_n)}{f'(x_n)},\; n=0,1,2,\ldots}\]
Observation.
\(x_{n+1}\) is the \(x\)-intercept of the tangent line to \(y=f(x)\) at \((x_n,f(x_n))\).
The equation of the tangent line to \(y=f(x)\) at \((x_n,f(x_n))\) is
\[y=f(x_n)+f'(x_n)(x-x_n).\]
For the \(x\)-intercept, \(y=0\). So
\[0=f(x_n)+f'(x_n)(x-x_n)\implies x=x_n-\frac{f(x_n)}{f'(x_n)}.\]
Thus \(x_{n+1}\) is the \(x\)-intercept of the tangent line to \(y=f(x)\) at \((x_n,f(x_n))\).
Remark. We will show later that \(\{x_n\}\) converges to \(x^*\) for a good choice of an initial approximation \(x_0\). If \(x_0\) is far from \(x^*\), then \(\{x_n\}\) may not converge to \(x^*\). For example, \(f(x)=x^2-5\) has a unique root \(x^*=\sqrt{5}\) in \([-2,3]\). But the Newton-Raphson Method constructs a sequence \(\{x_n\}\) using \(x_0=-2\) that converges to \(-\sqrt{5}\neq x^*\).
Example.
Do four iterations by the Newton-Raphson Method to approximate the root \(x^*\) of \(f(x)=e^x-x-2\) on the interval \([0,2]\)
using \(x_0=1\).
Solution. \(f(x)=e^x-x-2 \implies f'(x)=e^x-1\) and \(f'(1)=e-1\neq 0\). Thus
\[x_{n+1}
=x_n-\frac{e^{x_n}-x_n-2}{e^{x_n}-1},\; n=0,1,2,\ldots\]
\[\begin{array}{|l|l|c|l|l|}
\hline n & x_n & f(x_n) & f'(x_n) & x_{n+1}=x_n-\frac{f(x_n)}{f'(x_n)}\\
\hline 0 & 1 & -0.2817 & 1.7182 & 1.1639 \\
\hline 1 & 1.1639 & 0.0384 & 2.2023 & 1.1464 \\
\hline 2 & 1.1464 & 0.0004 & 2.1468 & 1.1462 \\
\hline 3 & 1.1462 & 0.0000 & 2.1462 & 1.1462 \\
\hline
\end{array}\]
Since \(|x_4-x_3|<10^{-4}\), we can say that the root is \(x_4=1.1462\) roughly correct to four decimal places
(almost true as \(x^*=1.146193\)). Actually \(x_4\) is correct to 12 decimal places! Note that this sequence converges
to the root faster than that of other methods (why?).
Convergence: Suppose \(f\) is a function with a simple root \(x^*\) in \([a,b]\), i.e., \(f'(x^*)\neq 0\). Assume \(f''\) is continuous in \([a,b]\). Then there is a \(\delta>0\) such that the sequence \(\{x_n\}\) constructed by the Newton-Raphson Method converges to \(x^*\) for any choice of the initial approximation \(x_0\in (x^*-\delta,x^*+\delta)\).
The Secant Method: In the iteration formula by the Newton-Raphson Method \[x_{n+1}=x_n-\frac{f(x_n)}{f'(x_n)},\; n=0,1,2,\ldots,\] we need to calculate the derivative \(f'(x_n)\) which may be difficult sometimes. To avoid such calculations, the Secant Method modifies the above formula by replacing \(f'(x_n)\) with its approximation. Note that \[f'(x_n)=\lim_{x\to x_n}\frac{f(x)-f(x_n)}{x-x_n}.\] If \(x_{n-1}\) is close to \(x_n\), then \[f'(x_n)\approx\frac{f(x_{n-1})-f(x_n)}{x_{n-1}-x_n} =\frac{f(x_n)-f(x_{n-1})}{x_n-x_{n-1}}.\] Thus \[x_{n+1}=x_n-\frac{f(x_n)}{f'(x_n)}\approx x_n-\frac{f(x_n)}{\frac{f(x_n)-f(x_{n-1})}{x_n-x_{n-1}}}=x_n-\frac{(x_n-x_{n-1})f(x_n)}{f(x_n)-f(x_{n-1})}.\] Thus the iteration formula by the Secant Method is \[\boxed{x_{n+1}=x_n-\frac{(x_n-x_{n-1})f(x_n)}{f(x_n)-f(x_{n-1})},\; n=1,2,3,\ldots}\] Note that geometrically \(x_{n+1}\) is the \(x\)-intercept of the secant line joining \((x_n,f(x_n))\) and \((x_{n-1},f(x_{n-1}))\). Also note that the convergence of the sequence \(\{x_n\}\) by the Secant Method is slower than that of the Newton-Raphson Method.
Last edited