1. Introduction
Let
X and
Y stand for Banach spaces and
be a convex and nonempty subset of
X. A plethora of applications from diverse disciplines can be solved if reduced to a nonlinear equation of the form
This reduction takes place using Mathematical Modeling [
1,
2]. Then, a solution denoted by
is to be found that answers the application. The solution may be a number or a vector or a matrix or a function. This task is very challenging in general. Obviously, the solution
is desired in closed form. However, in practice, this is achievable only in rare cases. That is why researchers mostly develop iterative methods convergent to
under some conditions on the initial data.
A popular method is the Newton’s method [
2,
3,
4,
5] defined, respectively, for a starting point
and all
by
Here,
is the notation for the Fréchet derivative of the operator
F. The convergence rate of Newton’s method is quadratic. However, this method requires the calculation of the derivative of the operator
F [
1,
2,
3]. It is not always easy or impossible to do, in particular, in the case when the operator is not given analytically, but only the algorithm for its calculation on the computer is known. Then, Newton’s method (
2) and its modifications [
4,
5,
6,
7,
8] using derivatives are not suitable for solving (
1). In this case, we can use difference methods [
1,
3,
9,
10,
11]. The simplest of them is the Secant method [
2,
3,
6,
7]
for all
,
are starting points. The Secant method was extended for the solution of (
1) in Banach spaces by J.W. Schmidt [
9]. This method under different conditions was studied in many papers [
2,
7]. The convergence order of the method (
3) is equal to
. The method with a higher quadratic convergence is described for all
by the formula
This method is famous as the method of the linear interpolation or the Kurchatov’s method. It does not interfere with Newton’s method in the convergence order, and it does not require analytically given derivatives as the Secant method does. The method (
4) was proposed for the first time by V.A. Kurchatov in [
12] for the one-dimensional case. In the Banach space, the method (
4) was first presented in the works of S.M. Shakhno [
13,
14]. In addition, this method was studied by many authors I.K. Argyros, H. Ren, J.A. Ezquerro, and M.A. Hernández [
15,
16,
17]. The Kurchatov method uses only first-order divided differences in its iterative formula. However, often the studying of its convergence additionally requires conditions for the second-order divided differences. This ensures theoretically obtaining the second order of convergence. Kurchatov’s two-step methods were studied by I.K. Argyros, S. George, H. Kumar, P.K. Parida, and S.M. Shakhno [
18,
19].
In this article, we propose the following modification of the method (
4).
Let
. Define the two-step Kurchatov-type methods for all
by
and
It is known that multi-step methods converge faster than the corresponding one-step methods. Therefore, there is a growing interest in the development and theoretical studying of the convergence of such algorithms. It is worth noting that the method (
5) uses the same inverse operator in both steps. This helps to reduce the total number of calculations compared to the corresponding one-step method, especially for large scale problems.
We provide the local as well as the semi-local convergence analysis for these methods under generalized conditions. Moreover, these conditions include only operators that appear in methods. The local convergence is given in
Section 2. The semi-local convergence is presented in
Section 3, followed by the examples and the concluding remarks in
Section 4 and
Section 5, respectively.
2. Local Convergence
It is convenient for the study of the local convergence for the methods to introduce some parameters and real functions. Set .
Suppose:
There exists a function
which is continuous and nondecreasing in both variables such that the equation
has the smallest solution
.
Set .
There exists a function
, which is continuous and nondecreasing in both variables such that the equation
has a smallest solution
, where the function
is given by
Define the function
by
The equation
has a smallest solution
.
This parameter will be shown to be a radius of convergence in Theorem 1 for the method (
5).
Set
. Then, follows by definition (
7) that for all
and
Let stand for the open and closed ball in X, respectively, of center and radius . By we denote the space of bounded linear operators from X into Y.
The convergence analysis uses the conditions for both methods.
Suppose:
The equation has a simple solution such that .
for all .
Set .
for all .
.
Next, the local convergence is established for the method (
5).
Theorem 1. Suppose that the conditions hold. Moreover, if the starting points , then the sequence generated by Formula (5) exists in , stays in for all and is convergent to . Moreover, the following assertions holdandwhere the radius r is given by Formula (7) and the functions are as previously defined. Proof. By hypothesis
. Then, by applying conditions
, the definition of the radius
r and (
8), we have
It follows by (
12) and the Banach lemma on the invertible operator [
4] that
and
Moreover, the iterates
and
are well defined by the two substeps of the method (
5). In view of that, we can write in that
Using (
7) and (
9) (for
),
,
, (
13) and (
14) we obtain
where we also used
and
Similarly, by the second substep of the method (
5), we can write
so
Hence, the estimates (
10) and (
11) hold for
. By simply replacing the role of
by
in the preceding calculations the induction for the estimates (
10) and (
11) is terminated. It follows that
where
Thus, we conclude
. □
Next, a unique result is presented for the solution of the equation .
Proposition 1. Suppose:
There exists a solution of the equation for some .
The conditions and hold.
There exists such that Set .
Then, the equation is uniquely solvable by the element in the region .
Proof. Let
. If then follows by
–
in turn that
Hence, the operator
T is invertible. Then, by the identity
we conclude that
. □
Concerning the local convergence analysis of the method (
6), clearly the function
is the same, whereas the function
corresponding to
is given by
This is due to the similar computation
where
and
is the smallest solution of the equation
in the interval
, where
and
is the smallest positive solution of the equation
(if it exists). Hence, we arrived at the corresponding semi-local convergence result for the method (
6).
Theorem 2. Suppose that the conditions hold with replacing r. Then, the conclusions of Theorem 1 hold for the method (6) with the function . Clearly, the uniqueness of the solution results of Proposition 1 holds for the method (
6).
3. Semi-Local Convergence
The analysis is based on majorizing sequences. Let
and
be given parameters. Suppose that there exists a function
which is continuous and nondecreasing such that the equation
has a smallest solution
. Set
. Moreover, suppose that there exist a function
which is continuous and nondecreasing.
Define the sequence
for
and all
by
where
Next we present a convergence result for the sequence .
Lemma 1. Suppose that for all Then, the sequence given by Formula (23) is nondecreasing and convergent to its unique least upper bound . Proof. The sequence is nondecreasing and bounded from above by and as such it is convergent to . □
The condition
shall be used in the semi-local convergence analysis first of the method (
5).
Suppose:
There exist points , parameters such that and
for all
Set
for all
Conditions (
24) holds
and
Next, the semi-local convergence of the method (
5) is presented based on the conditions
and the preceding terminology.
Theorem 3. Suppose that the conditions hold. Then, the sequence generated by the method (5) is well defined in , remains in for all and is convergent to a solution of the equation . Moreover, the following error estimates hold Proof. It follows as in the proof of Theorem 1 but there are some small differences. Iterates
and
are well defined by the condition
and the first substep of the method (
5) for
. We also have
so the iterate
. Then, as in Theorem 1 but using the
instead of
, we obtain the estimates
so
and
leading to
and
where we also used
Moreover, we can write
so
Thus, we obtain by (
23) and the second substep of the method (
5) that
and
It follows from (
28) and (
30) that the sequence
is complete (since (
28) is also complete as convergent) in a Banach space
X and as such is convergence to some point
. Furthermore, by letting
in (
29) and using the continuity of
F we conclude that
. Then, from the estimate
and letting
we show (
25). □
Proposition 2. Suppose:
(1) There exists a solution of the equation for some .
(2) Conditions on and hold on .
(3) There exist such that Set
Then, the equation is uniquely solvable by in the region .
Proof. Let
with
. Define the linear operator
G by
. By applying (2) and (
31), we obtain
So, the linear operator
G is invertible. Therefore, from the identity
we conclude that
. □
The majorizing sequence
for the method (
6) is defined similarly by
Lemma 2. Suppose that for all Then, the sequence given by Formula (32) is nondecreasing and convergent to its unique least upper bound . Theorem 4. Suppose that the conditions hold with (32), replacing (24) and , respectively. Then, the conclusions of Theorem 3 hold for the method (6). The uniqueness of the solution is given in Proposition 2.
Remark 1. (1) Proposition 2 is shown without using all the conditions of the Theorem 3; however, if all conditions are used, we can set . In this case .
(2) If , then we have for all . Consequently, the conditions or can be replaced by for the method (5) or for the method (6) and similarly for the method (5) or for the method (6). (3) The parameter given in closed form can be replaced or in the condition or .
4. Numerical Examples
In this section, we provide examples to verify the theoretical result.
Example 1. Let and . Define the function F on Ω
byThen, Local case Let and . Then, , , , .
Semi-local case. Let
,
,
,
. Then,
. Majorizing sequences for method (
5) and (
6) are
respectively. So,
and
.
Example 2. Consider the system of m equations Here , and .
Local case. Let , . Then, , , .
Semi-local case. Let
,
,
and
. Then,
. Majorizing sequences for method (
5) and (
6) are
respectively. So,
and
.
Let us apply methods (
4)–(
6) for solving considered nonlinear problems under different initial approximations
. All these methods require addition approximation
. It is computed by the rule
. The stopping conditions for the iterative process are
Table 1 and
Table 2 show number of iterations that are needed for solving one equation and system of equations for
.
Figure 1 and
Figure 2 demonstrate that norms
and
for the two-step Kurchatov’s methods (
5) and (
6) decrease faster than for Kurchatov’s method (
4).
5. Conclusions
The objective in this work is to develop a process for studying the convergence of iterative methods containing inverses of linear operators under weak conditions. These conditions involve only operators appearing in the methods. In particular, a local and a semi-local convergence analysis of the two-step Kurchatov-type methods is provided under the generalized Lipschitz conditions for only divided differences of order one. Regions of convergence and uniqueness of the solution are established. The results of the numerical experiment are given. The developed technique does not rely on the studied methods. That is why it can also be used on other methods that contain inverses of divided differences or inverses of linear operators in general.
The future work involves the application of this process on other single step, multi-step iterative methods with inverses [
14,
15,
17,
20,
21]. We will also study the analogs of the studied methods when the Fréchet is replaced by the Gateaux derivative.
Author Contributions
Conceptualization, I.K.A., S.S., S.R. and H.Y.; methodology, I.K.A., S.S., S.R. and H.Y.; software, I.K.A., S.S., S.R. and H.Y.; validation, I.K.A., S.S., S.R. and H.Y.; formal analysis, I.K.A., S.S., S.R. and H.Y.; investigation, I.K.A., S.S., S.R. and H.Y.; resources, I.K.A., S.S., S.R. and H.Y.; data curation, I.K.A., S.S., S.R. and H.Y.; writing—original draft preparation, I.K.A., S.S., S.R. and H.Y.; writing—review and editing, I.K.A., S.S., S.R. and H.Y.; visualization, I.K.A., S.S., S.R. and H.Y.; supervision, I.K.A., S.S., S.R. and H.Y.; project administration, I.K.A., S.S., S.R. and H.Y.; funding acquisition, I.K.A., S.S., S.R. and H.Y. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Ortega, J.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equations in Several Variables; Academic Press: New York, NY, USA, 1970. [Google Scholar]
- Dennis, J.E., Jr.; Schnabel, R.B. Numerical Methods for Unconstrained Optimization and Nonlinear Equations; Prentice-Hall: Englewood Cliffs, NJ, USA, 1983. [Google Scholar]
- Traub, J.F. Iterative Methods for the Solution of Equations; Prentice Hall: Hoboken, NJ, USA, 1964. [Google Scholar]
- Kantorovich, L.V.; Akilov, G.P. Functional Analysis; Pergamon Press: Oxford, UK, 1982. [Google Scholar]
- Schwetlick, H. Numerische Lösung Nichtlinearer Gleichungen; VEB Deutscher Verlag der Wissenschaften: Berlin, Germany, 1979. [Google Scholar]
- Wang, X. Convergence of Newton’s method and uniqueness of the solution of equations in Banach space. IMA J. Numer. Anal. 2000, 20, 123–134. [Google Scholar] [CrossRef] [Green Version]
- Yamamoto, T. Historical developments in convergence analysis for Newton’s and Newton-like methods. J. Comput. Appl. Math. 2000, 124, 1–23. [Google Scholar] [CrossRef] [Green Version]
- Werner, W. Über ein Verfahren der Ordnung 1 + zur Nullstellenbestimmung. Numer. Math. 1979, 32, 333–342. [Google Scholar] [CrossRef]
- Schmidt, J.W. Untere Fehlerschranken für Regula–Falsi Verhahren. Period. Math. Hungar. 1978, 9, 241–247. [Google Scholar] [CrossRef]
- Argyros, I.K. On the Secant method. Publ. Math. 1993, 43, 223–238. [Google Scholar] [CrossRef]
- Potra, F.A. On an iterative algorithm of order 1.839…for solving nonlinear operator equations. Numer. Funct. Anal. Optimiz. 1985, 7, 75–106. [Google Scholar] [CrossRef]
- Kurchatov, V.A. On a method of linear interpolation for the solution of functional equations. Dokl. Akad. Nauk SSSR 1971, 198, 524–526. Translation in Soviet Math. Dokl. 1971, 12, 835–838. (In Russian) [Google Scholar]
- Shakhno, S.M. On a Kurchatov’s method of linear interpolation for solving nonlinear equations. PAMM—Proc. Appl. Math. Mech. 2004, 4, 650–651. [Google Scholar] [CrossRef]
- Shakhno, S.M. On the difference method with quadratic convergence for solving nonlinear operator equations. Mat. Stud. 2006, 26, 105–110. (In Ukrainian) [Google Scholar]
- Argyros, I.K. A Kantorovich-type analysis for a fast iterative method for solving nonlinear equations. J. Math. Anal. Appl. 2007, 332, 97–108. [Google Scholar] [CrossRef] [Green Version]
- Argyros, I.K.; Ren, H. On the Kurchatov method for solving equations under weak conditions. Appl. Math. Comput. 2016, 273, 98–113. [Google Scholar] [CrossRef]
- Ezquerro, J.A.; Grau, A.; Grau-Sánchez, M.; Ángel Hernández, M. On the efficiency of two variants of Kurchatov’s method for solving nonlinear systems. Numer. Algor. 2013, 64, 685–698. [Google Scholar] [CrossRef]
- Argyros, I.K.; Shakhno, S. Extended Two-Step-Kurchatov Method for Solving Banach Space Valued Nondifferentiable Equations. Int. J. Appl. Comput. Math. 2020, 6, 32. [Google Scholar] [CrossRef]
- Kumar, H.; Parida, P.K. On semilocal convergence of two-step Kurchatov method. Int. J. Comput. Math. 2019, 96, 1548–1566. [Google Scholar] [CrossRef]
- Dhiman, D.; Mishra, L.N.; Mishra, V.N. Solvability of some non-linear functional integral equations via measure of noncompactness. Adv. Stud. Contemp. Math. 2022, 32, 157–172. [Google Scholar]
- Sharma, N.; Mishra, L.N.; Mishra, V.N.; Pandey, S. Solution of Delay Differential equation via iteration algorithm. Eur. J. Pure Appl. Math. 2020, 13, 1110–1130. [Google Scholar] [CrossRef]
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).