American Journal of Computational and Applied Mathematics

p-ISSN: 2165-8935    e-ISSN: 2165-8943

2018;  8(3): 47-49

doi:10.5923/j.ajcam.20180803.01

 

A New Method for Solving Nonlinear Equations Based on Euler’s Differential Equation

Masoud Saravi

Emeritus Professor of Islamic Azad University of Iran, Iran

Correspondence to: Masoud Saravi, Emeritus Professor of Islamic Azad University of Iran, Iran.

Email:

Copyright © 2018 The Author(s). Published by Scientific & Academic Publishing.

This work is licensed under the Creative Commons Attribution International License (CC BY).
http://creativecommons.org/licenses/by/4.0/

Abstract

Usually the methods based on Taylor expansion series for have better convergence [1]. But, nearly, all of them contain one or more derivatives of . The purpose of this paper is to introduce a technique to obtain free from derivatives which works better than methods others that been considered in most text book for solving nonlinear equations by providing some numerical examples.

Keywords: Nonlinear equations, Order of convergence, Euler’s equation, Iteration formulae

Cite this paper: Masoud Saravi, A New Method for Solving Nonlinear Equations Based on Euler’s Differential Equation, American Journal of Computational and Applied Mathematics , Vol. 8 No. 3, 2018, pp. 47-49. doi: 10.5923/j.ajcam.20180803.01.

1. Introduction

The problem of finding the roots of a given equation
(1)
where function f is sufficiently smooth in a neighborhood of a simple root arise frequently in science and engineering. In most cases it is difficult to obtain an analytical solution of (1). Hence the exploitation of numerical methods for solving such equations becomes a main subject of considerable interests. Usually in all text books the methods split into two sections, namely methods without derivatives and methods with derivatives [2, 3, 4, 6, 8, 9, 10]. Probably the most well-known and widely used algorithm to find a root of without derivative is the fixed point iteration method. In next section, we introduce a new algorithm and by expressing weak and strong aspect of this method, it will be deduced that the order of convergence is more than other methods without derivatives if the equation (1) contains simple roots.

2. Procedure

Expanding in (1) by Taylor's series about the point , we get
By approximating this series we may write
That is,
(2)
This is an Euler’s equation with general solution given by
(3)
or
(4)
This is a nonlinear equation with degree two, hence
That is
(5)
This leads to the following iteration formulas which can be used to approximate a solution of
(6)
Obviously can be found by two choices for in (4). For example, let , then
and
Therefore equation (6) becomes
(7)
where
and
Remark 1: It should be noted that our starting value cannot be a or b, i.e., It would be better to start with where
Remark 2: The sign, of the square root term is chosen to agree with the sign to keep close to
Remark 3: To find the order of convergence of this method we need some difficult square root computations, hence we avoid these computations. But the following examples in next section show that the order of convergence of this method must nearly be quadratic.

3. Numerical Examples

This sections deals with some numerical test on some problems that been considered in several Numerical Analysis text books. We resolved them by this method and compare the results.
Example 1: Consider This equation has been taken from [5] and has a real root on (0,1). If we wish to approximate this root with accuracy 10-4 by the bisection method, we need 14 iterations to obtain an approximation accurate to 10-4. Because with and we get
But if we apply our method, we obtain so that .
Even we use chord (modified regula-falsi) method given by
starting with we come to following results:
Example 2: Equation has a root in (1/4, 1/3). This equation been considered in [2].
The correct value to (4D) is The authors used fixed point iteration formula showed that if we write and start with mid-point of [1/4, 1/3] then Since is continue there is an interval within [1/4, 1/3] over which But by our method with plus sign (since f(1/3)>0) we obtained:
Although if they used a new scheme given by , with but with the same starting vale this scheme requires fifteen iterations to converges to the root 0.2872.
Let's consider another example. This example was chosen from [3].
Example 3: Approximate a zero of
In this book only mentioned that this equation has not real root. The roots correct to (4d) are We used with starting value for fixed point iteration method and get But by (7) we obtained
It seems this equation has only two conjugate complex roots, because we examined several numbers and every time we reached to this result. This fact may be examined by considering complex equation
Example 4: Consider This equation has a root on (1, 2) and is given in [2,4, 5,6]. Let
By fixed point iteration method we obtained following results:
Of course in this example Now we apply our method. We have Since hence we use (7) with minus sign and we get We also used Newton method with and get
Note 1: Although by Newton method we had the same result on first iteration but this is not always true. See following example [4].
Example 5: Consider has a root in (1, 2). We used scheme given by (7) and get But by Newton’s method with we obtained But with initial value , in third iteration we get
In general, the Newton method works better, in particular when the equation has complex roots.
Let’s consider polynomial equations with all complex roots.
Example 6: Approximate all roots of equation given by
This example was chosen from [7]. In this book mentioned that this equation has not real roots and with starting value found It is clear that a second root will be The other two roots are
If we start with with after 8 iterations we obtain but by Newton method, with we need only two iterations.

4. Conclusions

Results of all examples in this paper show the efficiency of this method comparing with other methods without derivatives. Although several methods free from derivatives been considered, but they contain too much computations [11, 12, 13, 14, 15, 16]. This method is not better than Newton method but it is not far from this method. In particular, sometimes its convergence is better than Newton method. If we compare with other methods which contain more differentiations is a useful formula to ignore differentiations. Author hope to extend this method to a system of nonlinear equations. Research in this matter is one of my future goals.

References

[1]  M. Saravi, A new look at one of the Newton-like methods for solving nonlinear equations, International Journal of Numerical Methods and Applications Volume 5, Number 1, 2011.
[2]  J. Murphy, D. Ridout, B. Mc Shane, Numerical Analysis, Algorithms and Computation. Ellis Horwood Limited, England, 1988.
[3]  R. W. Hamming, Introduction to Applied Numerical Analysis, McGraw-hill, 1971.
[4]  Carel-Erik Froberg. Introduction to Numerical Analysis. Addison-Wesley publishing company, 1974.
[5]  M. K. Jain, S. R. K. Iyengar, R. K. Jain, Numerical Methods problems and solutions. New Age International Publishers, New Delhi, 2003.
[6]  A. K. KAW, Numerical Method with Algorithms, Holistis Numerical Method Institute, University of South Florida, 2009.
[7]  G. Shanker Rao, Numerical Analysis, New Age International Publishers, New Delhi, 2002.
[8]  L.V. Atkinson & P. J. Harley, An Introduction to Numerical Methods with Pascal. Addison-Wesley Publishers Limited, London, 1983.
[9]  R.L. Burden, J.D.Faires, Numerical Analysis. PWS-KENT Publishing Company, USA, 1989.
[10]  Joe D. Hoffman, Numerical Methods for Engineers and Scientists. McGraw-Hill, Inc. New York, 1992.
[11]  Alicia Corderoa, José L. Huesoa, Eulalia Martínezb, Juan R. Torregrosaa, A new technique to obtain derivative-free optimal iterative methods for solving nonlinear equations. Journal of Computational and Applied Mathematics 252(2013).
[12]  Jovana Džunić, Miodrag S. Petković, A cubically convergent Steffensen-like method for solving nonlinear equations. Applied Mathematics Letters 25(2012).
[13]  F. Soleymani, S. Karimi Vanani, Optimal Steffensen-type methods with eighth order of convergence. Computers and Mathematics with Applications 62(2011).
[14]  J. P. Jaiswa, An Improved R-order Convergence Derivative-Free Method for Solving Nonlinear Equations. Procedia Engineering 127(2015).
[15]  F. Soleymani, Optimal Fourth-order Iterative Methods Free from Derivatives. Miskolc Mathematical Notes 12(2011).
[16]  R. Thukral, New Eighth-order Iterative Derivative-Free Methods for solving nonlinear Equations, Volume 2012, Article ID 493456, 12 pages.