Sie sind auf Seite 1von 20

CHAPTER 2

SOLUTION OF NON-LINEAR EQUATIONS

One of the most common problem encountered in engineering analysis is that given a function f (x), find
the values of x for which f (x) = 0. The solution (values of x) are known as the rootsof the equation f (x) =
0, or the zeroesof the function f (x).
The roots of equations may be real or complex. In general, an equation may have any number of (real)
roots or no roots at all. For example, sin x – x = 0 has a single root, namely, x = 0, whereas tan x – x = 0
has infinite number of roots (x = 0, ± 4.493, ± 7.725,). There are two types of methods available to find
the roots of algebraic and transcendental equations of the form f (x) =0.
1. Direct Methods: Direct methods give the exact value of the roots in a finite number of steps. We
assume here that there is no round off errors. Direct methods determine all the roots at the same time.
2. Indirect or Iterative Methods: Indirect or iterative methods are based on the concept of successive
approximations. The general procedure is to start with one or more initial approximation to the root and
obtain a sequence of iterates (xk) which in the limit converges to the actual or true solution to the root.
Indirect or iterative methods determine one or two roots at a time.
The indirect or iterative methods are further divided into two categories: Bracketing and open methods.
The bracketing methods require the limits between which the root lies, whereas the open methods require
the initial estimation of the solution. Bisection and False position methods are two known examples of
the bracketing methods. Among the open methods, the Newton- Raphson and the method of successive
approximation are most commonly used. The most popular method for solving a non-linear equation is
the Newton-Raphson method and this method has a high rate of convergence to a solution.
In this chapter, we present the following indirect or iterative methods with illustrative examples:
1. Bisection Method
2. Method of False Position (Regular Falsi Method)
3. Secant (Successive Approximation) Method.
4. Newton-Raphson Method (Newton’s method)
2.1. Bisection Method of Solving a Nonlinear Equation
After reading this topic, you should be able to:
1. follow the algorithm of the bisection method of solving a nonlinear equation,
2. use the bisection method to solve examples of finding roots of a nonlinear equation, and
3. Enumerate the advantages and disadvantages of the bisection method.
What is the bisection method and what is it based on?
One of the first numerical methods developed to find the root of a nonlinear equation f ( x)  0 was the
bisection method (also called binary-search method). The method is based on the following theorem.
Theorem
An equation f ( x)  0 , where f (x) is a real continuous function,
 If f ( x ) f ( xu )  0 has at least one root between x  and xu (See Figure 1).
 If f ( x ) f ( xu )  0 , there may or may not be any root between x  and xu (Figures 2 and 3).
 If f ( x ) f ( xu )  0 , then there may be more than one root between x  and xu (Figure 4).
So the theorem only guarantees one root between x  and xu .

Pre. By Gebresilassie H. Page 1 of 20


f (x)

xℓ
xu x

Figure 1:- At least one root exists between the two points if the function is real, continuous, and changes sign.

f (x)

x
xℓ xu

Figure 2:-The function does not change sign between two points, roots of the equation may still exist
between the two points.

f (x) f (x)

xℓ xu
x x
xℓ xu

Pre. By Gebresilassie H. Page 2 of 20


Figure 3:-If the function f (x) does not change sign between two points, there may not be any roots for
the equation f ( x)  0 between the two points.

f (x)

xu
x
xℓ

Figure 4:-If the function f (x) changes sign between the two points, more than one root for the equation
f ( x)  0 may exist between the two points.

Since the method is based on finding the root between two points, the method falls under the category of
bracketing methods.
Since the root is bracketed between two points, x  andxu , one can find the mid-point, xm between x 
and xu . This gives us two new intervals
1. x and xm , and
2. xm and xu .
Is the root now between x  and xm or between xm and xu ? Well, one can find the sign of f ( x ) f ( xm ) ,
and if f ( x ) f ( xm )  0 then the new bracket is between x  and xm , otherwise, it is between xm and xu .
So, you can see that you are literally halving the interval. As one repeats this process, the width of the
interval x , xu  becomes smaller and smaller, and you can zero in to the root of the equation f ( x)  0 .
The algorithm for the bisection method is given as follows.

Algorithm (procedure) for the bisection method


The steps to apply the bisection method to find the root of the equation f ( x)  0 are
1. Choose x  and xu as two guesses for the root such that f ( x ) f ( xu )  0 , or in other words,
f (x) changes sign between x and xu .
2. Estimate the root, xm , of the equation f ( x)  0 as the mid-point between x  and xu as
x   xu
xm =
2
3. Now check the following

Pre. By Gebresilassie H. Page 3 of 20


a) If f ( x ) f ( xm )  0 , then the root lies between x  and xm ; then x  x and xu  xm .
b) If f ( x ) f ( xm )  0 , then the root lies between xm and xu ; then x  xm and xu  xu .
c) If f ( x ) f ( xm )  0 ; then the root is xm . Stop the algorithm if this is true.
4. Find the new estimate of the root
x   xu
xm =
2
Find the absolute relative approximate error as
xmnew - xmold
a =  100
xmnew
where
xmnew = estimated root from present iteration
xmold = estimated root from previous iteration
5. Compare the absolute relative approximate error a with the pre-specified relative error
tolerance s . If a s , then go to Step 3, else stop the algorithm. Note one should also check
whether the number of iterations is more than the maximum number of iterations allowed. If so,
one needs to terminate the algorithm and notify the user about it.
Example
You are working for a start-up computer assembly company and have been asked to determine the
minimum number of computers that the shop will have to sell to make a profit.
The equation that gives the minimum number of computers n to be sold after considering the total costs
and the total sales is
f (n)  40n1.5  875n  35000  0
Use the bisection method of finding roots of equations to find the minimum number of computers that
need to be sold to make a profit. Conduct three iterations to estimate the root of the above equation. Find
the absolute relative approximate error at the end of each iteration and the number of significant digits
at least correct at the end of each iteration.
Solution
Let us assume
n  50, nu  100
Check if the function changes sign between n  and nu .
f (n )  f 50  40(50)1.5  875(50)  35000  5392.1
f(nu )  f (100)  40(100)1.5  875(100)  35000  12500
Hence
f n  f nu   f 50 f 100  5392.1 12500  0
So there is at least one root between n and nu , that is, between 50 and 100.

Pre. By Gebresilassie H. Page 4 of 20


Iteration 1
The estimate of the root is
n  nu 50  100
nm    75
2 2
f nm   f 75  40(75)1.5  875(75)  35000  4.6442  103
f n  f nm   f 50 f 75  5392.1 4.6442  103   0
Hence the root is bracketed between n and nm , that is, between 50 and 75. So, the lower and upper limits of the
new bracket are
n  50, nu  75
At this point, the absolute relative approximate error a cannot be calculated, as we do not have a previous
approximation.

Iteration 2
The estimate of the root is
n  nu 50  75
nm    62.5
2 2
f nm   f 62.5  40(62.5)1.5  875(62.5)  35000  76.735
f n  f nm   f 50 f 62.5  5392.176.735  0
Hence, the root is bracketed between nm and nu , that is, between 62.5 and 75. So the lower and upper limits of
the new bracket are
n  62.5, nu  75
The absolute relative approximate error, a at the end of Iteration 2 is

nmnew  nmold
a   100
nmnew
62.5  75
  100  20%
62.5
None of the significant digits are at least correct in the estimated root nm  62.5 as the absolute relative
approximate error is greater than 5%.

Iteration 3
The estimate of the root is
n  nu 62.5  75
nm    68.75
2 2
f nm   f 68.75  40(68.75)1.5  875(68.75)  35000  2.3545 103
 
f n  f nm   f 62.5 f 68.75  76.735  2.3545 103  0
Hence, the root is bracketed between n  and nm , that is, between 62.5 and 68.75. So the lower and upper limits of
the new bracket are
n  62.5, nu  68.75
The absolute relative approximate error a at the end of Iteration 3 is

Pre. By Gebresilassie H. Page 5 of 20


nmnew  nmold
a   100
nmnew
68.75  62.5
  100  9.0909%
68.75
Still none of the significant digits are at least correct in the estimated root of the equation, as the absolute relative
approximate error is greater than 5%. The estimated minimum number of computers that need to be sold to break
even at the end of the third iteration is 69. Seven more iterations were conducted and these iterations are shown in
the Table 1.
Table 1 Root of f x   0 as a function of the number of iterations for bisection method.
Iteration n nu nm a % f nm 
1 50 100 75 ----------  4.6442 10 3
2 50 75 62.5 20 76.735
3 62.5 75 68.75 9.0909  2.3545 10 3
4 62.5 68.75 65.625 4.7619
 1.1569 10 3
5 62.5 65.625 64.063 2.4390
−544.68
6 62.5 64.063 63.281 1.2346 −235.12
7 62.5 63.281 62.891 0.62112
−79.483
8 62.5 62.891 62.695 0.31153 −1.4459
9 62.5 62.695 62.598 0.15601
37.627
10 62.598 62.695 62.646 0.077942
18.086

th
At the end of the 10 iteration,
a  0.077942%
Hence the number of significant digits at least correct is given by the largest value of m for which
a  0.5  10 2 m
0.077942  0.5  10 2m
0.15588  10 2 m
log0.15588  2  m
m  2  log0.15588  2.8072
So
m2
The number of significant digits at least correct in the estimated root 62.646 is 2.

2.2. FALSE POSITION METHOD


The method of False Position (also called the Regular Falsi method, and the linear interpolation method) is
another well-known bracketing method. It is very similar to Bisection method with the exception that it uses a
different strategy to end up with its new root estimate. Rather than bisecting the interval (a, b), it locates the root
by joining f (a1) and f (b1) with a straight line. The intersection of this line with the x-axis represents an improved
estimate of the root.

Pre. By Gebresilassie H. Page 6 of 20


f x 

f xU 

Exact root

xL
O xr xU x

f xL 
Figure 1 False-Position Method

Based on two similar triangles, shown in Figure 1, one gets


0  f ( x L ) 0  f ( xU )
 (1)
xr  x L xr  xU
From Equation (4), one obtains
xr  xL  f xU   xr  xU  f xL 
xU f x L   x L f xU   xr  f x L   f xU 
The above equation can be solved to obtain the next predicted root xm as
xU f x L   x L f xU 
xr  (2)
f x L   f xU 
The above equation, through simple algebraic manipulations, can also be expressed as
f xU 
xr  xU  (3)
 f x L   f xU 
 
 x L  xU 
or
f x L 
xr  x L  (4)
 f xU   f x L 
 
 xU  x L 
Observe the resemblance of Equations (6) and (7) to the secant method.
False-Position Algorithm
The steps to apply the false-position method to find the root of the equation f x   0 are as follows.
1. Choose x L and xU as two guesses for the root such that f x L  f xU   0 , or in other words, f x  changes sign
between x L and xU .
2. Estimate the root, x r of the equation f x   0 as
xU f x L   x L f xU 
xr 
f x L   f xU 
3. Now check the following

Pre. By Gebresilassie H. Page 7 of 20


If f x L  f xr   0 , then the root lies between x L and x r ; then x L  x L and xU  xr .
If f x L  f xr   0 , then the root lies between x r and xU ; then x L  x r and xU  xU .
If f x L  f xr   0 , then the root is x r . Stop the algorithm.
4. Find the new estimate of the root
xU f x L   x L f xU 
xr 
f x L   f xU 
Find the absolute relative approximate error as
x rnew  x rold
a   100
x rnew
where
xrnew = estimated root from present iteration
xrold = estimated root from previous iteration
5. Compare the absolute relative approximate error a with the pre-specified relative error tolerance s . If
a s , then go to step 3, else stop the algorithm. Note one should also check whether the number of iterations is
more than the maximum number of iterations allowed. If so, one needs to terminate the algorithm and notify the
user about it.
Note that the false-position and bisection algorithms are quite similar. The only difference is the formula used to
calculate the new estimate of the root x r as shown in steps #2 and #4!
Example 1
You are working for “DOWN THE TOILET COMPANY” that makes floats for ABC commodes. The floating ball
has a specific gravity of 0.6 and has a radius of 5.5cm. You are asked to find the depth to which the ball is
submerged when floating in water. The equation that gives the depth x to which the ball is submerged under
water is given by
x 3  0.165 x 2  3.993  10 4  0
Use the false-position method of finding roots of equations to find the depth x to which the ball is submerged
under water. Conduct three iterations to estimate the root of the above equation. Find the absolute relative
approximate error at the end of each iteration, and the number of significant digits at least correct at the end of
third iteration.

Figure 2 Floating ball problem.


Solution
From the physics of the problem, the ball would be submerged between x  0 and x  2 R ,

Pre. By Gebresilassie H. Page 8 of 20


Where
R  radius of the ball,
That is
0  x  2R
0  x  2(0.055)
0  x  0.11
Let us assume
x L  0, xU  0.11
Check if the function changes sign between x L and xU
f x L   f 0  0  0.1650  3.993  10 4  3.993  10 4
3 2

f xU   f 0.11  0.11  0.1650.11  3.993  10  4  2.662  10  4


3 2

Hence
f xL  f xU   f 0 f 0.11  3.993  104  2.662  104   0
Therefore, there is at least one root between x L and xU , that is between 0 and 0.11.
Iteration 1
The estimate of the root is
xU f x L   x L f xU 
xr 
f x L   f xU 



0.11 3.993 10 4  0   2.662  10 4
 0.0660


3.993  10 4   2.662 10 4 
f x r   f 0.0660
 0.0660  0.1650.0660  3.993  10  4
3 2
 
5
 3.1944  10
f x L  f xr   f 0 f 0.0660     0
Hence, the root is bracketed between x L and x r , that is, between 0 and 0.0660. So, the lower and upper limits of
the new bracket are x L  0, xU  0.0660 , respectively.
Iteration 2
The estimate of the root is
xU f x L   x L f xU 
xr 
f x L   f xU 


0.0660  3.993  10 4  0   3.1944  10 5    0.0611

3.993  10 4   3.1944  10 5 
The absolute relative approximate error for this iteration is
0.0611  0.0660
a   100  8%
0.0611

f xr   f 0.0611

 0.0611  0.1650.0611  3.993  10 4  1.1320  10 5
3 2

f x L  f xr   f 0 f 0.0611      0

Pre. By Gebresilassie H. Page 9 of 20


Hence, the lower and upper limits of the new bracket are x L  0.0611, xU  0.0660 , respectively.
Iteration 3
The estimate of the root is
xU f  x L   x L f  xU 
xr 
f x L   f xU 



0.0660  1.132  10 5  0.0611   3.1944  10 5 

1.132  10 5   3.1944  10 5 
 0.0624
The absolute relative approximate error for this iteration is
0.0624  0.0611
a   100  2.05%
0.0624
f xr   1.1313  10 7
f x L  f xr   f 0.0611 f 0.0624     0
Hence, the lower and upper limits of the new bracket are x L  0.0611, xU  0.0624
All iterations results are summarized in Table 1. To find how many significant digits are at least correct in the last
iterative value
a  0.5  102m
2.05  0.5  102m
m  1.387
The number of significant digits at least correct in the estimated root of 0.0624 at the end of 3rd iteration is 1.
Table 1 Root of f x   x 3  0.165x 2  3.993  10 4  0 for false-position method.
Iteration xL xU xr a % f  xm 
1 0.0000 0.1100 0.0660 ----  3.1944  10 5
2 0.0000 0.0660 0.0611 8.00  1.1320  10 5
3 0.0611 0.0660 0.0624 2.05  1.1313  10 7

Example 2
Find the root of f x   x  4 x  2  0 , using the initial guesses of x L  2.5 and xU  1.0, and a pre-
2

specified tolerance of s  0.1% .


Solution
The individual iterations are not shown for this example, but the results are summarized in Table 2. It takes five
iterations to meet the pre-specified tolerance.
Table 2 Root of f x   x  4 x  2  0 for false-position method.
2

Iteration xL xU f x L  f xU  xr a % f xm 


1 -2.5 -1 -21.13 25.00 -1.813 N/A 6.319
2 -2.5 -1.813 -21.13 6.319 -1.971 8.024 1.028
3 -2.5 -1.971 -21.13 1.028 -1.996 1.229 0.1542
4 -2.5 -1.996 -21.13 0.1542 -1.999 0.1828 0.02286
5 -2.5 -1.999 -21.13 0.02286 -2.000 0.02706 0.003383

Pre. By Gebresilassie H. Page 10 of 20


To find how many significant digits are at least correct in the last iterative answer,
a  0.5  102m
0.02706  0.5  102m
m  3.2666
Hence, at least 3 significant digits can be trusted to be accurate at the end of the fifth iteration.

2.3. Newton-Raphson Method


After reading this chapter, you should be able to:
1. derive the Newton-Raphson method formula,
2. develop the algorithm of the Newton-Raphson method,
3. use the Newton-Raphson method to solve a nonlinear equation, and
4. discuss the drawbacks of the Newton-Raphson method.
Introduction
Methods such as the bisection method and the false position method of finding roots of a nonlinear equation
f ( x)  0 require bracketing of the root by two guesses. Such methods are called bracketing methods. These
methods are always convergent since they are based on reducing the interval between the two guesses so as to zero
in on the root of the equation.

In the Newton-Raphson method, the root is not bracketed. In fact, only one initial guess of the root is
needed to get the iterative process started to find the root of an equation. The method hence falls in the category of
open methods. Convergence in open methods is not guaranteed but if the method does converge, it does so much
faster than the bracketing methods.

Derivation
The Newton-Raphson method is based on the principle that if the initial guess of the root of f ( x)  0 is at xi ,
then if one draws the tangent to the curve at f ( xi ) , the point xi 1 where the tangent crosses the x -axis is an
improved estimate of the root (Figure 1).
Using the definition of the slope of a function, at x  xi
f xi  = tan θ
f xi   0
= ,
xi  xi 1
This gives
f xi 
xi 1 = xi  (1)
f xi 
Equation (1) is called the Newton-Raphson formula for solving nonlinear equations of the form f x   0 . So
starting with an initial guess, xi , one can find the next guess, xi 1 , by using Equation (1). One can repeat this
process until one finds the root within a desirable tolerance.

Algorithm
The steps of the Newton-Raphson method to find the root of an equation f x   0 are
1. Evaluate f x  symbolically
2. Use an initial guess of the root, xi , to estimate the new value of the root, xi 1 , as

Pre. By Gebresilassie H. Page 11 of 20


f xi 
xi 1 = xi 
f xi 
3. Find the absolute relative approximate error a as

xi 1  xi
a =  100
xi 1
4. Compare the absolute relative approximate error with the pre-specified relative error tolerance, s . If
a >s , then go to Step 2, else stop the algorithm. Also, check if the number of iterations has
exceeded the maximum number of iterations allowed. If so, one needs to terminate the algorithm and
notify the user.

f (x)

f (xi) [xi, f (xi)]

f (xi+1 )

θ
x
xi+2 xi+1 xi

Figure 1Geometrical illustration of the Newton-Raphson method.


Example 1
You are working for a start-up computer assembly company and have been asked to determine the
minimum number of computers that the shop will have to sell to make a profit. The equation that
gives the minimum number of computers n to be sold after considering the total costs and the total
sales is
f n  40n1.5  875n  35000  0
Use the Newton-Raphson method of finding roots of equations to find the minimum number of
computers that need to be sold to make a profit. Conduct three iterations to estimate the root of the
above equations. Find the absolute relative approximate error at the end of each iteration and the
number of significant digits at least correct at the end of each iteration.

Solution
f n  40n1.5  875n  35000  0
f n  60n 0.5  875
Let us take the initial guess of the root of f n   0 as n0  50 .

Pre. By Gebresilassie H. Page 12 of 20


Iteration 1
The estimate of the root is
f n0 
n1  n0 
f n0 
4050  87550  35000
1.5
 50 
6050  875
0.5

 50   11.963  61.963
5392.1
 50 
 450.74
The absolute relative approximate error a at the end of Iteration 1 is

n1  n0
a   100  61.963  50  100  19.307%
n1 61.963
The number of significant digits at least correct is 0, as you need an absolute relative approximate
error of less than 5% for one significant digit to be correct in your result.
Iteration 2
The estimate of the root is
f n1  4061.963  87561.963  35000
1.5
n2  n1   61.963 
f n1  6061.963  875
0.5

292.45
 61.963 
 402.70
 61.963   0.72623  62.689
The absolute relative approximate error a at the end of Iteration 2 is

n2  n1
a   100
n2
62.689  61.963
  100  1.1585%
62.689
The number of significant digits at least correct is 1, because the absolute relative approximate error
is less than 5%
Iteration 3
The estimate of the root is
f n2 
n3  n 2 
f n2 
4062.689  87562.689  35000
1.5
 62.689 
6062.689  875
0.5

1.0031
 62.689 
 399.94
 
 62.689   2.5080  10 3  62.692
The absolute relative approximate error a at the end of Iteration 3 is

Pre. By Gebresilassie H. Page 13 of 20


n3  n 2
a   100
n3
62.692  62.689 3
a   100  4.0006  10 %
62.692
Hence the number of significant digits at least correct is given by the largest value of m for which
a  0.5  10 2 m
4.0006  10 3  0.5  10 2m
8.0011 10 3  10 2m
 
log 8.0011 10 3  2  m
 
m  2  log 8.0011 10 3  4.0968
So
m4
The number of significant digits at least correct in the estimated root 62.692 is 4.

Drawbacks of the Newton-Raphson Method


I. Divergence at inflection points
If the selection of the initial guess or an iterated value of the root turns out to be close to the inflection point (see
the definition in the appendix of this chapter) of the function f x  in the equation f x   0 , Newton-Raphson
method may start diverging away from the root. It may then start converging back to the root. For example, to
find the root of the equation
f x   x  1  0.512  0
3

The Newton-Raphson method reduces to


( xi  1)3  0.512
3
xi 1 = xi 
3( xi  1)2
Starting with an initial guess of x0  5.0 , Table 1 shows the iterated values of the root of the equation. As you
can observe, the root starts to diverge at Iteration 6 because the previous estimate of 0.92589 is close to the
inflection point of x  1 (the value of f '  x  is zero at the inflection point). Eventually, after 12 more iterations
the root converges to the exact value of x  0.2 .
Table 1 Divergence near inflection point.
Iteration 9 –8.2217
xi
Number 10 –5.1498
0 5.0000 11 –3.1044
1 3.6560 12 –1.7464
2 2.7465 13 –0.85356
3 2.1084 14 –0.28538
4 1.6000 15 0.039784
5 0.92589 16 0.17475
6 –30.119 17 0.19924
7 –19.746 18 0.2
8 –12.831

Pre. By Gebresilassie H. Page 14 of 20


Figure 2 Divergence at inflection point for f x   x  13  0 .

II. Division by zero


For the equation
f x   x 3  0.03x 2  2.4  106  0
the Newton-Raphson method reduces to
xi  0.03xi  2.4  106
3 2
xi 1 = xi 
3xi  0.06 xi
2

For x0  0 or x0  0.02 , division by zero occurs (Figure 4). For an initial guess close to 0.02 such as
x0  0.01999 , one may avoid division by zero, but then the denominator in the formula is a small
number. For this case, as given in Table 2, even after 9 iterations, the Newton-Raphson method does not
converge.
Table 2 Division by near zero in Newton-Raphson method
Iteration
xi f ( xi ) a %
Number
0.019990 1.00E-05

0  1.60000 10-6 f(x)


–2.6480 7.50E-06

1 18.778 100.75
–1.7620
5.00E-06

2
–1.1714 –5.5638 50.282 2.50E-06

x
3
–0.77765 –1.6485 50.422
-0.03 -0.02 -0.01
0.00E+00
0 0.01 0.02 0.03 0.04
4
–0.51518 –0.48842 50.632 -2.50E-06
0.02
5 –0.14470 50.946 -5.00E-06

–0.34025
6
–0.22369 –0.042862 51.413 -7.50E-06

–0.012692
-1.00E-05
7 52.107
–0.14608
8
– –0.0037553 53.127
9
0.094490 –0.0011091 54.602

Figure 3 Pitfall of division by zero or a near zero number.

Pre. By Gebresilassie H. Page 15 of 20


III. Oscillations near local maximum and minimum
Results obtained from the Newton-Raphson method may oscillate about the local maximum or minimum
without converging on a root but converging on the local maximum or minimum. Eventually, it may lead
to division by a number close to zero and may diverge.
For example, for
f x   x 2  2  0 the equation has no real roots (Figure 4 and Table 3).

Iteration
xi f ( xi ) a %
Number
0 –1.0000 3.00
1 0.5 2.25 300.00
2 –1.75 5.063 128.571
3 –0.30357 2.092 476.47
4 3.1423 11.874 109.66
5 1.2529 3.570 150.80
6 –0.17166 2.029 829.88
7 5.7395 34.942 102.99
8 2.6955 9.266 112.93
9 0.97678 2.954 175.96

Table 3 Oscillations near local maxima and minima in


Newton-Raphson method
Figure 4 Oscillations around local minima for
f x   x  2 .
2

IV. Root jumping


In some case where the function f (x) is oscillating and has a number of roots, one may choose an initial guess
close to a root. However, the guesses may jump and converge to some other root. For example for solving the
equation sin x  0 if you choose x0  2.4  7.539822 as an initial guess, it converges to the root of x  0 as
shown in Table 4 and Figure 6. However, one may have chosen this as an initial guess to converge to
x  2  6.2831853.

Table 4 Root jumping in Newton-Raphson method


Iteration
xi f ( xi ) a %
NO
0 7.539822 0.951
1 4.462 –0.969 68.973
2 0.5499 0.5226 711.44
3 –0.06307 –0.06303 971.91
4
4 8.376 10 8.375 10 5 7.54 10 4
5  1.95861 10 13  1.9586110 13 4.28 1010

Pre. By Gebresilassie H. Page 16 of 20


Figure 5 Root jumping from intended location of root for f x   sin x  0 .
What is an inflection point?
For a function f x  , the point where the concavity changes from up-to-down or down-to-up is
called its inflection point. For example, for the function f x   x  13 , the concavity changes at x  1
(see Figure 3), and hence (1,0) is an inflection point.
An inflection points may exist at a point where f ( x)  0 and where f ' ' ( x ) does not exist. The reason
we say that it MAY exist is because if f ( x)  0 , it only makes it a possible inflection point. For example,
for f ( x )  x 4  16 , f (0)  0 , but the concavity does not change at x  0 . Hence the point (0, –16) is
not an inflection point of f ( x )  x 4  16 .
For f x   x  13 , f (x) changes sign at x  1 ( f ( x)  0 for x  1 , and f ( x)  0 for
x  1), and thus brings up the Inflection Point Theorem for a function f (x) that states the following.
“If f ' (c) exists and f (c ) changes sign at x  c , then the point (c, f (c)) is an inflection point
of the graph of f .”

2.4. SECANT METHOD


After reading this topic, you should be able to:
1. derive the secant method to solve for the roots of a nonlinear equation,
2. use the secant method to numerically solve a nonlinear equation.
What is the secant method and why would we want to use it instead of the
Newton-Raphson method?
The Newton-Raphson method of solving a nonlinear equation f ( x)  0 is given by the iterative formula
f ( xi )
xi 1 = xi  (1)
f ( xi )
One of the drawbacks of the Newton-Raphson method is that you have to evaluate the derivative of the
function. With availability of symbolic manipulators such as Maple, MathCAD, MATHEMATICA and
MATLAB, this process has become more convenient. However, it still can be a laborious process, and

Pre. By Gebresilassie H. Page 17 of 20


even intractable if the function is derived as part of a numerical scheme. To overcome these drawbacks,
the derivative of the function, f (x) is approximated as
f ( xi )  f ( xi 1 )
f ( xi )  (2)
xi  xi 1
Substituting Equation (2) in Equation (1) gives
f ( xi )( xi  xi 1 )
xi 1  xi  (3)
f ( xi )  f ( xi 1 )
The above equation is called the secant method. This method now requires two initial guesses, but
unlike the bisection method, the two initial guesses do not need to bracket the root of the equation. The
secant method is an open method and may or may not converge. However, when secant method
converges, it will typically converge faster than the bisection method. However, since the derivative is
approximated as given by Equation (2), it typically converges slower than the Newton-Raphson method.
The secant method can also be derived from geometry, as shown in Figure 1. Taking two initial guesses,
xi 1 and xi , one draws a straight line between f ( xi ) and f ( xi 1 ) passing through the x -axis at xi 1 .
ABE and DCE are similar triangles.
Hence
AB DC

AE DE
f ( xi ) f ( xi 1 )
 On rearranging, the secant method is given as
xi  xi 1 xi 1  xi 1
f ( xi )( xi  xi 1 )
xi 1  xi 
f ( xi )  f ( xi 1 )
f (x)

f (x i) B

f (x i–1 ) C

E D A
x
xi+1 xi–1 xi

Figure 1:- Geometrical representation of the secant method

Pre. By Gebresilassie H. Page 18 of 20


Example 1
You are working for a start-up computer assembly company and have been asked to determine the
minimum number of computers that the shop will have to sell to make a profit.
The equation that gives the minimum number of computers n to be sold after considering the total costs
and the total sales is
f (n)  40n1.5  875n  35000  0
Use the secant method of finding roots of equations to find the minimum number of computers that need
to be sold to make a profit. Conduct three iterations to estimate the root of the above equation. Find the
absolute relative approximate error at the end of each iteration and the number of significant digits at least
correct at the end of each iteration.
Solution
Let us take the initial guesses of the root of f n   0 as n1  25 and n0  50 .
Iteration 1
The estimate of the root is
f n0 n0  n1 
n1  n0 
f n0   f n1 

 n0 
40n  875n0  35000 n0  n1 
1.5

40n   
0
1.5
0  875n0  35000  40n1.15  875n1  35000

 50 
40(50)  875(50)  35000 50  25
1.5

40(50) 1.5
 
 875(50)  35000  40(25)1.5  875(25)  35000 
 60.587
The absolute relative approximate error a at the end of Iteration 1 is

n1  n0
a   100
n1
60.587  50  17.474%
  100
60.587
The number of significant digits at least correct is 0, as you need an absolute relative approximate error of
less than 5% for one significant digit to be correct in your result.
Iteration 2
The estimate of the root is
f n1 n1  n0 
n2  n1 
f n1   f n0 

 n1 
40n
 875n1  35000 n1  n0 
1.5

40n   
1
1.5
1  875n1  35000  40n01.5  875n0  35000

 60.587 
4060.587  87560.587  3500060.587  50
1.5

4060.587   87560.587   35000


1.5

 
  4050  87550  35000 
1.5

 62.569
The absolute relative approximate error a at the end of Iteration 2 is

Pre. By Gebresilassie H. Page 19 of 20


n2  n1
a   100
n2
62.569  60.587
  100  3.1672%
62.569
The number of significant digits at least correct is 1, because the absolute relative approximate error is
less than 5% .
Iteration 3
The estimate of the root is
f n2 n2  n1 
n3  n 2 
f n2   f n1 

 n2 
40n 
 875n2  35000 n2  n1 
1.5

40n   
2
1.5
2  875n2  35000  40n11.5  875n1  35000

 62.569 
4062.569  87562.569  3500062.569  60.587
1.5

4062.569  87562.569  35000 


1.5

 
  4060.587   87560.587   35000
1.5

 62.690
The absolute relative approximate error a at the end of Iteration 3 is

n3  n 2
a   100
n3
62.690  62.569
  100  0.19425%
62.690
The number of significant digits at least correct is 2, because the absolute relative approximate error is
less than 0.5% .

Pre. By Gebresilassie H. Page 20 of 20

Das könnte Ihnen auch gefallen