Académique Documents
Professionnel Documents
Culture Documents
Ioannis K. Argyros
Department of Mathematical
Editor Sciences
Cameron University
Xu
Lawton, Hou USA
Oklahoma,
Harvard University
andand Applied Sciences
School of Engineering
Á.Cambridge, MA, USA
Alberto Magreñán
Universidad Internacional de La Rioja
Escuela Superior de Ingenierı́a y Tecnologı́a
Logroño, La Rioja, Spain
p,
A SCIENCE PUBLISHERS BOOK
CRC Press
Taylor & Francis Group
6000 Broken Sound Parkway NW, Suite 300
Boca Raton, FL 33487-2742
This book contains information obtained from authentic and highly regarded sources. Reasonable efforts
have been made to publish reliable data and information, but the author and publisher cannot assume
responsibility for the validity of all materials or the consequences of their use. The authors and publishers
have attempted to trace the copyright holders of all material reproduced in this publication and apologize to
copyright holders if permission to publish in this form has not been obtained. If any copyright material has
not been acknowledged please write and let us know so we may rectify in any future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced, transmit-
ted, or utilized in any form by any electronic, mechanical, or other means, now known or hereafter invented,
including photocopying, microfilming, and recording, or in any information storage or retrieval system,
without written permission from the publishers.
For permission to photocopy or use material electronically from this work, please access www.copyright.
com (http://www.copyright.com/) or contact the Copyright Clearance Center, Inc. (CCC), 222 Rosewood
Drive, Danvers, MA 01923, 978-750-8400. CCC is a not-for-profit organization that provides licenses and
registration for a variety of users. For organizations that have been granted a photocopy license by the CCC,
a separate system of payment has been arranged.
Trademark Notice: Product or corporate names may be trademarks or registered trademarks, and are used
only for identification and explanation without intent to infringe.
This book comprises a modern approach which combines recent and past results in
numerical methods with applications in fields such us Engineering, Chemistry, etc.
The book also provides a comparison between various investigations of recent years.
Each chapter contains several new theoretical results and important applications.
The applications appear in the form of examples or study cases or they are implied
since our results improve earlier ones that have already been applied in concrete
problems. Chapters and sections have been written independent of each other. Hence,
the interested reader can go directly to a certain chapter or section and understand
the material without having to go back and forth. Furthermore, several references to
recent studies appear along the whole book.
This monograph presents recent works of authors on Convergence and Dynamics
of iterative methods. It is the natural outgrowth of their related publications in these
areas. Chapters are self-contained and can be read independently. Moreover, an
extensive list of references is given in each chapter, in order to allow reader to use
the previous ideas. For these reasons, we think that several advanced courses can be
taught using this book.
The book’s results are expected to find applications in many areas of applied
mathematics, engineering, computer science and real problems. As such this
monograph is suitable to researchers, graduate students and seminars in the above
subjects, also to be in all science and engineering libraries.
The preparation of this book took place during the year 2016 in Lawton, Oklahoma,
USA and Logroño, La Rioja, Spain.
October 2016 Ioannis K. Argyros
Á. Alberto Magreñán
Contents
Dedication v
Preface vii
List of Figures xiii
List of Tables xv
Symbol Description xvii
1 Halley’s method 1
1.1 Introduction 1
1.2 Semilocal convergence of Halley’s method 3
1.3 Numerical examples 10
1.4 Basins of attraction 13
1.4.1 2 roots 13
1.4.2 3 roots 15
1.4.3 4 roots 15
References 23
2 Newton’s method for k-Fréchet differentiable operators 25
2.1 Introduction 26
2.2 Semilocal convergence analysis for Newton’s method 27
2.2.1 Uniqueness of solution 32
2.2.2 Special choices for function g 34
2.2.2.1 Choice 1 34
2.2.2.2 Choice 2 36
2.3 Numerical examples 39
References 43
3 Nonlinear Ill-posed equations 46
3.1 Introduction 46
3.2 Convergence analysis 50
3.3 Error bounds 55
3.4 Implementation of adaptive choice rule 57
3.4.1 Algorithm 57
3.5 Numerical examples 57
References 60
x < Iterative Methods and Their Dynamics with Applications: A Contemporary Study
Halley’s method
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Semilocal convergence of Halley’s method . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.4 Basins of attraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.4.1 2 roots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.4.2 3 roots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.4.3 4 roots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.1 Introduction
In this chapter we are concerned with the problem of approximating a locally
unique solution x of the nonlinear equation
F(x) = 0, (1.1)
where F is twice Fréchet-differentiable operator defined on a nonempty open and
convex subset of a Banach space X with values in a Banach space Y .
Many problems from computational sciences and other disciplines can be
1
2 Iterative methods and their dynamics with applications: a contemporary study
(C1 ) There exists x0 ∈ D such that F (x0 )−1 ∈ L(Y, X), the space of bounded
linear operator from Y into X;
(C2 ) F (x0 )−1 F(x0 ) ≤ γ, γ > 0;
(C3 ) F (x0 )−1 F (x) ≤ M for each x in D, M > 0;
(C4 ) F (x0 )−1 [F (x) − F (y)] ≤ Kx − y for each x and y in D, K > 0.
In the present chapter we show how to expand even further the applicability
of Halley’s method using (C1 ), (C2 ), (C3 ) and the well-known center-Lipschitz
condition, defined by
(C5 ) F (x0 )−1 [F (x) − F (x0 )] ≤ Lx − x0 f or each x ∈ D, L > 0.
We have that
L≤M (1.5)
M
holds in general. Moreover, as it was stated in [4] and [5] can be arbitrarily
L
large. In the literature (C3 ) is used to obtain
1
F (x)−1 F (x0 ) ≤ f or each x ∈ D. (1.6)
1 − Mx − x0
However, using condition (C5 ) and if L < M we obtain the tighter estimate
1
F (x)−1 F (x0 ) ≤ f or each x ∈ D. (1.7)
1 − Lx − x0
This idea, used in the study of many methods [4, 5], allows to obtain:
We refer to (C1 )−(C3 ), (C5 )−(C7 ) as the (C) conditions. Let B(x, R), B(x, R)
Halley’s method 5
stand, respectively, for the open and closed balls in X with center x and radius
R > 0.
Theorem 1.1
Let F : D ⊂ X → Y be continuously twice Fréchet differentiable, where X, Y are
Banach spaces and D is an open and convex subset of X. Suppose that the (C) condi-
tions hold and B(x0 , R0 ) ⊂ D. Then, the sequence {xn } generated by Halley’s method
(1.2) is well defined, remains in B(x0 , R0 ) for all n ≥ 0 and converges to a unique
solution x ∈ B(x0 , R0 ) of equation F(x) = 0. Moreover, the following estimate holds
for each n = 0, 1, 2, . . .
B(x0 , R ) ⊆ D (1.14)
and
L
(R0 + R ) < 1, (1.15)
2
then, x is the only solution of equation F(x) in B(x0 , R ).
x1 − x0 = (I − LF (x0 ))−1 F (x0 )−1 F(x0 ) ≤ (I − LF (x0 ))−1
F (x0 )−1 F(x0 )
γ
≤ = γ0 < R0 .
1−a
Now, using (C5 ) we get that
(iii)
dn+1 ≤ ddn ≤ d n+1 d0 ;
d0 = γ,
M M
dn+1 = dn xn+1 − xn + xn+1 − xn 2 n ≥ 0.
2(1 − β ) 2
The rest of the proof will be obtained using mathematical induction. Assum-
ing that (i)-(vi) are true for all natural integers n ≤ k, where k ≥ 0 is a fixed
integer. Then F (xk+2 ) exists, since xk+2 ∈ B(x0 , R0 ) and
Next, we shall estimate F (x0 )−1 F(xk+2 ). We have by Halley’s method that
M dγ
dk+2 ≤ ( + bγ0 )xk+2 − xk+1
2 1−β
M dγ dk+1
≤ ( + bγ0 )
2 1−β M
(1 − β )(1 − d k+1 γ)
2(1 − β )2
M dγ dk+1
≤ ( + bγ0 )
2 1−β M
(1 − β )(1 − dγ)
2(1 − β )2
M γ dk+1
≤ ( + γ0 ) = ddk+1 ≤ d k+2 d0
2 1−β M
(1 − β )(1 − γ)
2(1 − β )2
and
1
LF (xk+2 ) ≤ F (xk+2 )−1 F (x0 )2 F (x0 )−1 F(xk+2 )
2
F (x0 )−1 F (xk+2 )
M M2 γ
≤ d k+2 ≤ ( + γ0 )γ0 < 1.
2(1 − β )2 4(1 − β ) 1 − β
2
xk+3 − xk+2 ≤ (I − LF (xk+2 ))−1 F (xk+2 )−1 F (x0 )F (x0 )−1 F(xk+2 )
dk+2
≤
M
(1 − β )(1 − dk+2 )
2(1 − β )2
M γ
( + γ0 )xk+2 − xk+1
2 1−β
≤ = bxk+2 − xk+1
M2 γ
(1 − β ) 1 − ( + γ0 )γ0
4(1 − β )2 1 − β
≤ bk+2 x1 − x0 .
Hence, we deduce that xk+3 ∈ B(x0 , R0 ). The induction for (i)-(vi) is complete.
Let m be a natural integer. Then, we have that
bk
x − xk ≤ x1 − x0 .
1−b
In order to show the uniqueness part, let y be a solution of F(x) = 0 in
1
B(x0 , R ). Denoting Q = 0 F (x0 )−1 F (x + θ (y − x ))dθ . Using (C5 ) we have
in turn that
1
I − Q = 0 F (x0 )−1 [F (x + θ (y − x )) − F (x0 )]dθ
1 L
≤ L 0 [(1 − θ )x − x0 + θ y − x0 ]dθ ≤ (R0 + R ) < 1.
2
It follows that Q−1 exists. Now, using
we deduce y = x . 2
2|x| 2
|F (x0 )−1 F (x)| = 2
≤ 2 = 2.76816609 (1.19)
x0 x0
Halley’s method 11
and
|x + x0 |
|F (x0 )−1 [F (x) − F (x0 )]| = |x − x0 | ≤ 2.560553633|x − x0 |. (1.20)
x02
That is, we obtain M = 2.76816609, L = 2.560553633, a = 0.016522791,
γ0 = 0.012138274 and 1 − Lγ0 = 0.968919297.
Furthermore, we have that b = 0.035021656 by using the secant method.
Then we get R0 = 0.012578805, β = 0.032208705 and d = 0.035628901. As a
consequence, conditions (C6 ) and (C7 ) are satisfied.
On the other hand,
B(x0 , R0 ) = [0.837421195, 0.862578805] ⊂ D.
So, we can ensure the convergence of Halley’s method by Theorem 1.1.
Now, we want to compare our uniqueness ball with the ones obtained in [11].
Using (1.4) we get that
√
4− 6
β0 = Mγ = 0.03304558 < .
5
So, conditions of [11] are also satisfied.
The uniqueness ball is B(x0 , r0 γ), where
2 − 3β0
r0 = = 1.053745860.
2(1 − 3β0 + β02 )
Then, we get that R1 = r0 γ = 0.012579319 > R0 .
So, it is clear that we provide a better information on the location of the
solution. Moreover by (1.14) and (1.15) we can set R = 0.15, which extends the
uniqueness ball from B(x0 , R0 ) to B(x0 , R ).
Example 1.3.2 In this example we provide an application of our results to
a special nonlinear Hammerstein integral equation of the second kind. In this
example we will use the max-norm.
1
4
x(s) = 1 + G(s,t)x(t)3 dt, s ∈ [0, 1], (1.21)
5 0
where, G is the Green’s kernel on [0, 1] × [0, 1] defined by
t(1 − s), t ≤ s;
G(s,t) = (1.22)
s(1 − t), s ≤ t.
Let X = Y = C[0, 1] and D be a suitable open convex subset of D := {x ∈ X :
x(s) > 0, s ∈ [0, 1]}, which will be given below. Define F : D → Y by
4 1
[F(x)](s) = x(s) − 1 − G(s,t)x(t)3 dt, s ∈ [0, 1]. (1.23)
5 0
12 Iterative methods and their dynamics with applications: a contemporary study
We get that
1
12
[F (x)y](s) = y(s) − G(s,t)x(t)2 y(t)dt, s ∈ [0, 1], (1.24)
5 0
and
1
24
[F (x)yz](s) = − G(s,t)x(t)y(t)z(t)dt, s ∈ [0, 1]. (1.25)
5 0
Let x0 (s) = 1 for all s ∈ [0, 1], then, for any y ∈ D, we have
12 1
[(I − F (x0 ))y](s) = G(s,t)y(t)dt, s ∈ [0, 1], (1.26)
5 0
which means
12 1 12 3
I − F (x0 ) ≤ maxs∈[0,1] 0 G(s,t)dt = = < 1. (1.27)
5 5 × 8 10
It follows from the Banach lemma on invertible operators that F (x0 )−1 exists
and
1 10
F (x0 )−1 ≤ = . (1.28)
3 7
1−
10
4 1 1
Moreover, from (1.23) we obtain F(x0 ) = maxs∈[0,1] 0 G(s,t)dt = .
5 10
1
Then, we get γ = .
7
It is clear that F (x) is not bounded in X or its subset D. A solution x of
F(x) = 0 must satisfy
1
x − 1 − x 3 ≤ 0, (1.29)
10
i.e., x ≤ ρ1 = 1.153467305 and x ≥ ρ2 = 2.423622140, where ρ1 and ρ2
1
are the positive roots of the equation z − 1 − z3 = 0. Consequently, if we look
10
for a solution such that x < ρ1 ∈ X1 , we can consider D = {x : x ∈ X1 and x <
r}, with r ∈ (ρ1 , ρ2 ), as a nonempty open convex subset of X. For example,
choose r = 1.7.
Using (1.24) and (1.25), we have that for any x, y, z ∈ D
12 1
|[(F (x) − F (x0 ))y](s)| = | G(s,t)(x(t)2 − x0 (t)2 )y(t)dt|
5 0
12 1
≤ G(s,t)|x(t) + x0 (t)||x(t) − x0 (t)|y(t)dt
5 0
12 1
≤ G(s,t)(r + 1)|x(t) − x0 (t)|y(t)dt, s ∈ [0, 1]
5 0
(1.30)
Halley’s method 13
and
24 1
|[F (x)yz](s)| = G(s,t)x(t)y(t)z(t)dt, s ∈ [0, 1]. (1.31)
5 0
Then, we have
12 1 81
F (x) − F (x0 ) ≤ × (r + 1)x − x0 = x − x0 (1.32)
5 8 100
and
24 r 51
F (x) ≤ × = . (1.33)
5 8 50
Choosing
51 81 51 70 M 66
M= , L= , a= , γ0 = , (L + )γ = ,
35 70 490 439 2 245
358
1 − Lγ0 = , (1.34)
439
b = 0.4266517573, R0 = 0.2781089940, β = 0.3218118359,
d = 0.5138822165.
We get
B(x0 , R0 ) ⊂ D. (1.35)
So, all conditions in Theorem are satisfied and we can state the convergence of
Halley’s method (1.2) starting in x0 . Moreover, by (1.14) and (1.15) we can set
R = 0.7, which extends the uniqueness ball from B(x0 , R0 ) to B(x0 , R ).
1.4.1 2 roots
Throughout this subsection, we have used cyan color is related to the conver-
gence to r1 = −1 and magenta is related to the convergence to r2 = 1. Moreover,
black color is related to that pairs for which there is no convergence to any of the
roots. In Figures 1.1–1.3 some basins of attraction related to polynomials with
two different roots with different multiplicities are shown.
14 Iterative methods and their dynamics with applications: a contemporary study
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
1.4.2 3 roots
Throughout this subsection, we have used cyan color is related to the √conver-
gence to r1 = −1, magenta is related to the convergence to r2 = − 12 − 23 i and
√
yellow to r3 = − 12 + 23 i. Moreover, black color is related to that pairs for which
there is no convergence to any of the roots. In Figures 1.4–1.11 some basins of
attraction related to polynomials with three different roots with different multi-
plicities are shown.
1.4.3 4 roots
Throughout this subsection, we have used cyan color is related to the conver-
gence to r1 = −1, magenta is related to the convergence to r2 = 1, yellow to
r3 = i and red to r4 = −i. Moreover, black color is related to that pairs for which
there is no convergence to any of the roots. In Figures 1.12–1.16 some basins of
attraction related to polynomials with four different roots with different multi-
plicities are shown.
16 Iterative methods and their dynamics with applications: a contemporary study
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
Figure 1.5: Basins of attraction associated to the Halley’s method applied to
p(z) = (z3 − 1)(z − 1).
Halley’s method 17
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
Figure 1.13: Basins of attraction associated to the Halley’s method applied to
p(z) = (z4 − 1)(z − 1).
Halley’s method 21
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
10
-5
-10
-10 -5 0 5 10
23
24 References
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
2.2 Semilocal convergence analysis for Newton’s method . . . . . . . . . . . . . . . 27
2.2.1 Uniqueness of solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2.2.2 Special choices for function g . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
2.2.2.1 Choice 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
2.2.2.2 Choice 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
2.3 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
25
26 Iterative methods and their dynamics with applications: a contemporary study
2.1 Introduction
We determine a solution x∗ of the equation
F(x) = 0,
L(X,Y ) as the space of bounded linear operators from the Banach space
setting X into the Banach space setting Y .
U(x, r) as the open ball centered at the point x ∈ X of radius r > 0.
U(x, r) as the open ball centered at x ∈ X of radius r > 0.
(C1 ) There exists Γ0 = [F (x0 )]−1 ∈ L(Y, X) for some x0 ∈ D and Γ0 F(x0 ) ≤
η,
(C2 ) Γ0 F (x) ≤ l, for x ∈ D,
√
1 1 − 1 − 2lη
(C3 ) h = lη ≤ and U x0 , ⊂ D.
2 l
The corresponding majorizing sequence {un } for {xn } is defined by
⎧
⎪
⎨ u0 = 0,
p(un ) (2.2)
⎪
⎩ un+1 = un − , for each n = 0, 1, 2, . . .
p (un )
where
l
p(t) = t 2 − t + η. (2.3)
2
There exists a plethora of results studies in which authors uses Lipschitz-
type conditions [1, 2, 4–6, 8–11, 13, 14, 17, 20]. The main idea of these works is
to extend the results of the Theorem by means of relaxing second condition. In
particular, the condition weakening (C2 ) is given by
Then, it follows from the Banach lemma on invertible operators [20] and from
the preceding estimate that F (x)−1 ∈ L(Y, X) and F (x)−1 F (x0 ) is bounded
by
1
F (x)−1 F (x0 ) ≤ ·
1 − g(t)
2 ) is used instead, we obtain
However, if (D
1
F (x)−1 F (x0 ) ≤ − ·
f (t)
Hence, if g(t) ≤ 1 + f (t), the new estimate is at least as tight as the old one, so
it is easy to see the advantages of this idea.
and
Proof: As f (0) > 0, then t ∗ ≥ 0. Using the well-known mean value theorem,
we get that
that is t1 < t ∗ , since N (t) > 0 in the interval [θ0 ,t ∗ ) by (2.6). Moreover, by the
definition of the sequence, we obtain that
f (0)
t1 = − > 0,
g(0) − 1
since
g(0) < 1 (by (2.5))
and f (0) > 0.
Finally, using mathematical induction on n, we obtain tn < t ∗ and tn − tn−1 ≥
0, since (tn−1 ,t ∗ ) ⊂ (0,t ∗ ). As a consequence, we find that sequence (2.4) con-
verges to r ∈ [0,t ∗ ]. Furthermore, as t ∗ is the unique root of f (t) = 0 in [0,t ∗ ], it
clearly follows that r = t ∗ . 2.
The next step, is to prove that sequence {tn } is a majorizing sequence of
Newton’s method (2.1). We present this fact in the following result.
Lemma 2.1
conditions are satisfied; xn ∈ D, for all n ≥ 0, and f (β ) ≤
Suppose that the (D)
0, where β is a solution of equation f (t) = 0 in (0,t1 ), and {xn } is the sequence
generated by Newton’s method (2.1). Moreover, denote Γn = [F (xn )]−1 . Then, for
all n ≥ 1, we have
1
(in ) Γn F (x0 ) ≤ − ,
g(tn ) − 1
Proof: We will use again mathematical induction on n to proof this result. Case
n = 1 is clear so we omit writing it. We suppose that (in )–(ivn ) are true for n =
1, 2, . . . , m − 1 and we will prove it for n = m.
From the induction hypothesis (ivm−1 ) we obtain that
xm − x0 ≤ xm − xm−1 + xm−1 − xm−2 + · · · + x1 − x0
≤ (tm − tm−1 ) + (tm−1 − tm−2 ) + · · · + (t1 − t0 ) = tm − t0 = tm ,
as t0 = 0.
3 ) and (2.5) we get that
(im ) Using (D
Γ0 (F (xm ) − F (x0 )) ≤ g(xm − x0 ) ≤ g(tm ) < 1. (2.7)
It follows the Banach lemma on invertible operators [7, 10, 20] and from (2.7)
1
that, Γm exists and Γm F (x0 ) ≤ − .
g(tm ) − 1
30 Iterative methods and their dynamics with applications: a contemporary study
(iim ) For any i ∈ {2, 3, . . . , k −1} and Taylor’s series, and denoting x = xm−1 +
τ(xm − xm−1 ) we obtain
k−i
1
Γ0 F (i) (xm ) ≤ Γ0 F ( j+i−1) (xm−1 )xm − xm−1 j−1
j=1
( j − 1)!
1
1
+ Γ0 F (k) xm−1 + τ(xm − xm−1 )
(k − 1 − i)! 0
×(1 − τ)k−2 xm − xm−1 k−i dτ
k−i
1
≤ f ( j+i−1) (tm−1 )(tm − tm−1 ) j−1
j=1
( j − 1)!
tm
1
+ f (k) (t)(tm − t)k−i−1 dt
(k − 1 − i)! tm−1
= f (i) (tm ).
k−1
1 ( j)
F(xm ) = F (xm−1 )(xm − xm−1 ) j
j=2
j!
1
1
+ F (k) xm−1 + t(xm − xm−1 )
(k − 1)! 0
×(1 − t)k−1 (xm − xm−1 )k dt,
Newton’s method for k-Fréchet differentiable operators 31
k−1
1 ( j)
Γ0 F(xm ) ≤ f (tm−1 )(tm − tm−1 ) j
j=2
j!
1
1
+ f (k) tm−1 + t(tm − tm−1 )
(k − 1)! 0
×(1 − t)k−1 (tm − tm−1 )k dt
= f (tm ) − f (tm−1 ) − f (tm−1 )(tm − tm−1 )
≤ f (tm ),
since f (s) + f (s)(t − s) ≥ 0 for each s,t ∈ [0,t1 ] with s ≤ t. Indeed, let ϕ(u) =
f (u) + f (u)(t − u) for each fixed t and u ∈ [s,t]. Then ϕ (u) = f (u)(t − u) ≥ 0,
since f (u) ≥ 0 and t ≥ u. Hence, function ϕ is increasing for u ∈ [s,t] and
ϕ(u) > 0, since ϕ(s) = f (s) > 0 (see also Lemma 2.2).
f (tm )
(ivm ) xm+1 − xm ≤ Γm F (x0 )Γ0 F(xm ) ≤ − = tm+1 − tm .
g(tm ) − 1
And now the proof is complete. 2
Then, the main result for Newton’s method (2.1) is presented.
Proof: From (D
1 ) and (D
2 ),we get that x1 is well-defined and x1 −x0 = t1 < t ∗ ,
∗
so that x1 ∈ U(x0 ,t ) ⊂ D. We now suppose that xn is well-defined and xn ∈
U(x0 ,t ∗ ) for n = 1, 2, . . . , m − 1. It follows from Lemma 2.1, the operator Γm−1
1
exists and Γm−1 F (x0 ) ≤ − . In addition, xm is well-defined. Moreover,
g(tm−1 )
since xn+1 − xn ≤ tn+1 − tn for n = 1, 2, . . . , m − 1, we have
m−1
m−1
xm − x0 ≤ xi+1 − xi ≤ (ti+1 − ti ) = tm < t ∗ ,
i=0 i=0
xn ∈ U(x0 ,t ∗ ) for all n ≥ 0. By the last lemma, we also have that F(xn ) ≤ f (tn )
and
f (tn )
xn+1 − xn ≤ Γn F (x0 )Γ0 F(xn ) ≤ − = tn+1 − tn ,
g(tn ) − 1
for all n ≥ 0, so that {tn } is a majorizing sequence of {xn } and is convergent.
Moreover, as lim tn = t ∗ , if x∗ = lim xn , then x∗ − xn ≤ t ∗ − tn , for all
n→+∞ n→+∞
n = 0, 1, 2, . . . Furthermore, as Γ0 F(xn ) ≤ f (tn ), for all n = 0, 1, 2, . . ., by letting
n → +∞, it follows Γ0 F(x∗ ) = 0. Finally, by the continuity of F we get that
F(x∗ ) = 0, since Γ0 ∈ L(Y, X). 2
Lemma 2.2
Suppose the conditions of the Theorem 2.2.2 hold. Then, the scalar function f (t) has
two real zeros t ∗ and t ∗∗ such that 0 < t ∗ ≤ t ∗∗ and if x ∈ U(x0 ,t ∗∗ ) ∩ D, then
On the other hand, if f (t) has two real zeros t ∗ and t ∗∗ such that 0 < t ∗ ≤ t ∗∗ ,
then the uniqueness of solution follows from the next result.
Theorem 2.2.3 Under the conditions of Theorem 2.2.2, the scalar function
f (t ∗∗ )t ∗∗
f (t) has two real zeros t ∗ and t ∗∗ such that 0 < t ∗ ≤ t ∗∗ . Let μ ∗∗ = .
2(1 − g(t ∗∗ ))
Suppose that μ ∗∗ ∈ (0, 1). Then, the solution x∗ is unique in U(x0 ,t ∗∗ ) ∩ D if
t ∗ < t ∗∗ or in U(x0 ,t ∗ ) if t ∗∗ = t ∗ .
In either case, it follows from the Banach lemma on invertible operators that
A−1 ∈ L(Y, X). Using the identity
we deduce that x∗ = y∗ . 2
34 Iterative methods and their dynamics with applications: a contemporary study
2.2.2.1 Choice 1
3 ) can be dropped, since only
We choose g(t) = 1 + f (t). Then, in this case (D
(D2 ) is needed for the derivation of the upper bounds on Γm F (x0 ). Moreover,
our scalar sequence {tn } specializes to the sequence {vn } considered in [15] and
defined by ⎧
⎪
⎨ v0 = 0,
f (vn−1 ) (2.8)
⎪
⎩ vn = N(vn−1 ) := vn−1 − ·
f (vn−1 )
Indeed, we have in turn from Taylor’s series that
k−1
1
I − Γm−1 F (xm ) = I − Γm−1 F (i) (xm−1 )(xm − xm−1 )i−1
i=1
(i − 1)!
xm
1 (k)
+ F (x)(xm − x) dx .k−2
(k − 2)! xm−1
k−1
1 1
= − f (i) (vm−1 )(vm − vm−1 )i−1
f (vm−1 ) i=2 (i − 1)!
vm
1 (k)
+ f (v)(vm − v) dv k−2
(k − 2)! vm−1
f (vm )
= 1− < 1.
f (vm−1 )
Next we justify this choice for function g and then provide a comparison between
sequences {tn } and {vn }. It follows from the Banach lemma that
1
Γm F (x0 ) ≤ − ·
f (v m)
Notice also that (2.5) and (2.6) are satisfied and N (t) ≥ 0. Hence, we arrive at
the results obtained in [15] and the choice g(t) = 1 + f (t) for g is justified under
conditions. However, if
the (D)
and
f (s)(1 − g(t)) + f (t) f (s) ≤ 0 for each s,t ∈ [0,t1 ] with s≤t (2.10)
and
tn+1 − tn ≤ vn+1 − vn for each n = 1, 2, . . . . (2.12)
Indeed, using (2.9) we have that
f (0) f (0)
t1 = − ≥− = v1 ⇒ t1 ≥ v1 ,
g(0) − 1 f (0)
f (v1 ) f (v1 )
v2 = N(v1 ) = v1 −
≤ v1 − = N(v1 ) ≤
f (v1 ) g(v1 ) − 1
36 Iterative methods and their dynamics with applications: a contemporary study
f (t1 )
≤ t1 − = N(t1 ) = t2 ⇒ v2 ≤ t2 ,
g(t1 ) − 1
(since N is an increasing function) and by (2.10)
f (t1 ) f (v1 )
t 2 − t1 = − ≤− = v2 − v1 ⇒ t2 − t1 ≤ v2 − v1 .
g(t1 ) − 1 f (v1 )
Suppose that tm ≥ vm and tm −tm−1 ≤ vm −vm−1 for each m = 1, 2, . . . , n. Then,
in an analogous way we show that (2.11) and (2.12) hold for m = n + 1. The
induction is complete. Notice that (2.5), (2.6), (2.9) and (2.10) hold for g(t) =
f (t) + 1. Hence, the class of functions g that satisfy (2.5), (2.6), (2.9) and (2.10)
is non-empty. Moreover, if strict inequality holds in (2.9) or (2.10), then so does
in (2.11) and (2.12).
2.2.2.2 Choice 2
Define function g as follows:
k−1 (i) t (k)
f (0) i−1 f (u)
g(t) = t + (t − u)k−2 du.
i=2
(i − 1)! 0 (k − 2)!
Let us also define iteration {sn } as {tn } but using function g as above. Then, as
in Choice 1 we have in turn that
k−1
1
I − Γ0 F (xm ) = I − Γ0 F (i) (x0 )(xm − x0 )i−1
i=1
(i − 1)!
xm
1
+ F (k) (x)(xm − x)k−2 dx .
(k − 2)! x0
k−1
1
I − Γ0 F (xm ) ≤ Γ0 F (i) (x0 )xm − x0 i−1
i=2
(i − 1)!
1
1
+ Γ0 F (k) x0 + τ(xm − x0 )
(k − 2)! 0
×(1 − τ)k−2 xm − x0 k−1 dτ
k−1
1
= f (i) (0)(sm − s0 )i−1
i=2
(i − 1)!
sm
1
+ f (k) (s)(sm − s)k−2 ds
(k − 2)! 0
= g(sm ) < 1.
Newton’s method for k-Fréchet differentiable operators 37
In this, since f (i) (0) ≤ f (i) (t) for each i = 2, 3, . . . , k − 1, t ∈ [0,t1 ] we have that
(2.5) is satisfied. Moreover, notice that f (i) (0) ≤ f (i) (t) for each n = 2, 3, . . . Then
again, if (2.9) and (2.10) hold we have as in Choice 2.2.2.1 that
sn ≤ vn (2.13)
and
sn+1 − sn ≤ vn+1 − vn (2.14)
for each n = 1, 2, 3, . . .
Then, again (2.13) and (2.14) hold as strict inequalities if (2.9) or (2.10) hold
as strict inequalities. Other choices of function g are possible such as g is any
polynomial of degree m ∈ N satisfying (D
3 ), (2.5), (2.6) (preferably as strict
inequalities),(2.9) and (2.10). This way sequence {tn } will be tighter than {vn }
as it was already shown in Choices 2.2.2.1 and 2.2.2.2.
Remark 2.2.5 It follows from the proof of Lemma 2.1 and Theorem 2.2.2 that
scalar sequence {rn } defined for each n = 0, 1, 2, . . . by
⎧
⎨ f (0)
r0 = 0, r1 = − ,
g(0) − 1
⎩
rn+2 = rn+1 − βn+1 (rn+1 − rn )2 ,
where
βn+1
1
βn+1 = ,
g(rn+1 ) − 1
k−1 ( j)
f (rn )
βn+1
1
= (rn+1 − rn ) j−2
j=2
j!
1
1
+ f (k) (rn + t(rn+1 − rn ))(1 − t)k−1 (rn+1 − rn )k dt,
(k − 1)! 0
is the tightest majorizing for {xn } (from the ones mentioned above in this paper)
which is obviously converging under the hypotheses of theorem 2.2.2. These hy-
potheses depend on function f having a positive zero. However this may no be the
case (see also the numerical examples). Then, one can simply directly study the
convergence of {rn } hoping to obtain weaker sufficient convergence conditions.
For example, if
0 ≤ −βn+1 (rn+1 − rn ) ≤ γ < 1 (2.15)
38 Iterative methods and their dynamics with applications: a contemporary study
for each n = 0, 1, 2, . . . and some γ ∈ [0, 1) then, we obtain that sequence {rn }
is non-decreasing and converges to some r∗ ≥ r1 . Hypotheses (2.15) can then
replace the existence of the zero β of equation f (t) = 0. Note that we have pro-
vided sufficient convergence conditions which on the one hand imply (2.15); not
necessarily imply the existence of β and can be weaker [4, 5, 10]. For simplic-
ity we will only present the case when k = 2. Then, we have for g(t) = l0t and
l
f (t) = p(t) = t 2 − t + η that scalar sequence {un } defined by
2
⎧
⎪
⎪ l0 (u1 − u0 )2
⎨ u0 = 0, u1 = η, u2 = u1 − ,
2(l0 u1 − 1) (2.16)
⎪
⎪ l(un − un−1 ) 2
⎩ un+1 = un − , for each n = 2, 3, . . .
2(l0 un − 1)
l0 ≤ l
l
holds in general and can be arbitrarily large [6, 9, 11]. Then, {un } converges
l0
provided that
1
h1 = Lη ≤ , (2.17)
2
where
1
L= 4l0 + l0 l + l0 l + 8l0 .
2
8
Note that
1 1
h≤ ⇒ h1 ≤
2 2
h1 l0
but not necessarily vice versa unless if l0 = l and → 0 as → 0. The sequence
h l
{un } used in [13] (see also (2.2)) is given by
⎧
⎪
⎨ u0 = 0,
f (un−1 )
⎪
⎩ un = un−1 − , for each n =, 1, 2, . . .
f (un−1 )
It can then easily be seen that (2.5) and (2.6) are satisfied if 2l0 ≤ l.
1
1
x(s) = + G(s,t)x(t)3 dt, (2.20)
2 0
where x ∈ C([0, 1]), t ∈ [0, 1] and the kernel G is the Green’s function
(1 − s)t, t ≤ s,
G(s,t) = (2.21)
s(1 − t), s ≤ t.
where the nodes t j and the weights ω j are known. Moreover, we denote x(ti ) by
xi , i = 1, 2, . . . , 8, so that equation (2.20) is now transformed into the following
system of nonlinear equations:
1
8
xi = + ai j x3j ,
2 j=1
where
ω j t j (1 − ti ) if j ≤ i,
ai j =
ω j ti (1 − t j ) if j > i.
Then, we write the above system in the following matrix form:
F(x) = x − v − Aw = 0, (2.22)
T
1 1 1
where x = (x1 , x2 , . . . , x8 ) , v = T
, ,..., , A = (ai j )8i, j=1 and w =
2 2 2
(x13 , x23 , . . . , x83 )T . Besides,
n tn un un
2 0.0173040441616297 . . . 0.0174299241630703 . . . 0.0174946186637403 . . .
3 0.0173040450083055 . . . 0.0174299542337089 . . . 0.0174946733196340 . . .
4 0.0173040450083055 . . . 0.0174299542337097 . . . 0.0174946733196365 . . .
by (2.4), (2.16) and (2.8) respectively. Besides, the corresponding majorizing se-
quences {un } and {tn } provide better a priori error estimates than the majorizing
sequence {un } obtained by Kantorovich or in [13] using the polynomial p(s).
The a priori error estimates are shown in Table 2.3. Observe the improvement
obtained from the majorizing sequences {tn } and {un } constructed in this work.
43
44 References
Nonlinear Ill-posed
equations
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
3.2 Convergence analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
3.3 Error bounds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
3.4 Implementation of adaptive choice rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.4.1 Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.5 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.1 Introduction
In this chapter we provide an extended the analysis of the Lavrentiev regulariza-
tion for nonlinear ill-posed problems
F(x) = y, (3.1)
First of all, let us recall that F is a monotone operator if, for all x1 , x2 ∈ D(F)
it satisfies the relation
Throughout all the chapter we will denote by X as a real Hilbert space with
inner product ., . and the norm .. Let U(x, ρ) and U(x, ρ) stand respectively,
for the open and closed ball in the Hilbert space X with center x ∈ X and radius
ρ > 0. We also denote by L(X) the space of all bounded linear operators from X
into X.
Our first assumption is that x∗ is a solution. Moreover, we also assume that
F has a locally uniformly bounded Fréchet derivative F in a ball around the
solution.
In application, usually only noisy data yδ are available, such that
y − yδ ≤ δ . (3.3)
It is known, see Theorem 1.1 in [23], that the equation (3.5) has a unique solu-
tion xαδ for α > 0, provided F is Fréchet differentiable and monotone in the ball
Br (x∗ ) ⊂ D(F) with radius r = x∗ − x0 + δ /α. However the regularized equa-
tion (3.5) remains nonlinear and it could be complicated to solve it numerically.
In the last years, several authors [8–11, 17] have considered iterative regular-
ization methods for obtaining stable approximate solutions for equation (3.5).
As it appears in [18], an iterative method with iterations defined by
δ
xk+1 = Φ(x0δ , x1δ , · · · , xkδ ; yδ ),
where Aδk := F (xkδ ) and {αk } is a sequence of positive real numbers such that
limk→∞ αk = 0. Actually, the stopping index kδ in [8] was chosen according to
the discrepancy principle
for some τ > 1 and showed that xkδδ → x∗ as δ → 0 under the following well-
known assumptions:
There exists L > 0 such that for all x, y ∈ D(F)
In [18], Mahale and Nair showed that xkδδ → x∗ as δ → 0 and an optimal order
error estimate for xkδδ − x∗ was obtained under the following assumptions:
Assumption 1 There exists r > 0 such that U(x∗ , r) ⊆ D(F) and F is Fréchet
differentiable at all x ∈ U(x∗ , r).
3. supλ ≥0 αϕ(λ )
λ +α ≤ cϕ ϕ(α), ∀λ ∈ (0, a]
Nonlinear Ill-posed equations 49
Assumption 4 There exists a sequence {αk } of positive real numbers such that
limk→∞ αk = 0 and there exists μ > 1 such that
αk
1≤ ≤ μ, ∀k ∈ N. (3.9)
αk+1
It is worth noticing that (3.9) is weaker than (3.7).
In [14] motivated by iteratively regularized Lavrentiev method, we showed
the quadratic convergence of the method defined by
δ δ
xn+1,α = xn,α − (F (xn,α
δ
) + αI)−1 (F(xn,α
δ
) − yδ + α(xn,α
δ
− x0 )), (3.10)
δ
where x0,α := x0 is a starting point of the iteration. Let Rα (x) = F (x) + αI and
Assumption 5 There exists r > 0 such that U(x0 , r) ∪ U(x∗ , r)) ⊆ D(F) and F
is Fréchet differentiable at all x ∈ U(x0 , r) ∪U(x∗ , r)).
Assumption 6 There exists a constant k0 > 0 such that for every x, u ∈ U(x0 , r)∪
U(x∗ , r)) and v ∈ X there exists an element Φ(x, u, v) ∈ X satisfying [F (x) −
F (u)]v = F (u)Φ(x, u, v), Φ(x, u, v) ≤ k0 vx − u.
Assumption 7 There exists r > 0 such that U(x0 , r) ⊆ D(F) and F is Fréchet
differentiable at all x ∈ U(x0 , r).
Assumption 8 Let x0 ∈ X be fixed. There exists a constant L > 0 such that for
every x, u ∈ U(x0 , r) ⊆ D(F) and v ∈ X there exists an element Φ(x, u, v) ∈ X sat-
isfying [F (x) − F (u)]v = F (u)Φ(x, u, v), Φ(x, u, v) ≤ Lv(x − x0 + u −
x0 ).
50 Iterative methods and their dynamics with applications: a contemporary study
x − u ≤ x − x0 + x0 − u,
to obtain the same estimates in this chapter; where PQ is the metric projection
onto the set Q.
The chapter is organized as follows. In Section 3.2, we provide a convergence
result for the method. Then, in Section 3.3, we give error bounds. In Section
3.4, we deal with the starting point and the algorithm. Finally, some numerical
examples are presented in Section 3.5.
xn+1 − xn ≤ tn+1 − tn , ∀n ≥ 0.
Lemma 3.2.1 (cf. [13], Lemma 2.1) Let {tn } be a majorizing sequence for {xn }
in X. If limn→∞ tn = t ∗ , then x∗ = limn→∞ xn exists and
x∗ − xn ≤ t ∗ − tn , ∀n ≥ 0. (3.12)
16Lη ≤ 1. (3.13)
Let √
1 − 1 − 16Lη
q= . (3.14)
2
Nonlinear Ill-posed equations 51
η ≤ t ∗ ≤ t ∗∗ . (3.17)
and
qn
t ∗ − tn ≤ η. (3.19)
1−q
Proof: It is clear that q ∈ (0, 1). Now using induction we obtain
L
(5tm + 3tm−1 ) ≤ q. (3.20)
2
Estimate is true for m = 1 by the definition of sequence {tn } and (3.13). Then,
2
we have by (3.15) that t2 −t1 ≤ q(t1 −t0 ) and t2 ≤ η + qη = (1 + q)η = 1−q
1−q η <
η ∗∗
1−q = t . If we assume that condition (3.20) is satisfied for all integers k ≤ m.
As a consequence, we have
and
1 − qm+1
tm+1 ≤ η. (3.22)
1−q
Now, we must prove that (3.20) is satisfied for m + 1. Using (3.21) and (3.22),
(3.20) shall be true if
L 1 − qm 1 − qm−1
[5( ) + 3( )]η ≤ q. (3.23)
2 1−q 1−q
Estimate (3.23) motivates us to define recurrent functions fm on [0, 1) by
f∞ (t) ≤ 0, (3.28)
since
fm (q) ≤ fm+1 (q) ≤ · · · ≤ f∞ (q). (3.29)
But (3.28) is true by (3.23) and (3.24). The induction for (3.20) is complete.
Hence, sequence {tn } is nondecreasing, bounded from above by t ∗∗ and such it
converges to some t ∗ which satisfies (3.17). 2
Now we must have to take into account the following Lemma.
1
F(u) − F(v) − F (u)(u − v) = F (u) Φ(v + t(u − v), u, u − v)dt.
0
From now on we assume that for r > 0, L > 0 and q given in (3.14):
δ 1
< η ≤ min{ , r(1 − q)} (3.30)
α 16L
and x∗ − x0 ≤ ρ, where
1
ρ≤ ( 1 + 2L(η − δ /α) − 1) (3.31)
L
The following remark is also interesting for our study.
Remark 3.2.4 Note that (3.30) and (3.31) imply
L 2 δ 1
ρ + ρ + ≤ η ≤ min{ , r(1 − q)}. (3.32)
2 α 16L
Now, we present the main result for this chapter.
δ qn η
xn,α − xαδ ≤ t ∗ − tn ≤ (3.34)
1−q
and
δ δ L δ δ δ δ δ δ
xn+1,α − xn,α ≤ [5xn,α − x0,α + 3xn−1,α − x0,α ]xn,α − xn−1,α . (3.35)
2
Proof: We begin by proving that
δ δ L δ δ δ δ δ δ
xn+1,α − xn,α ≤ [5xn,α − x0,α + 3xn−1,α − x0,α ]xn,α − xn−1,α . (3.36)
2
With G as in (3.11), we have for u, v ∈ Bt ∗ (x0 ),
G(u) − G(v) = u − v − Rα (u)−1 [F(u) − yδ + α(u − x0 )] + Rα (v)−1
×[F(v) − yδ + α(v − x0 )]
= u − v − [Rα (u)−1 − Rα (v)−1 ](F(v) − yδ + α(v − x0 ))
−Rα (u)−1 (F(u) − F(v) + α(u − v))
= Rα (u)−1 [F (u)(u − v) − (F(u) − F(v))]
−Rα (u)−1 [F (v) − F (u)]Rα (v)−1 (F(v) − yδ + α(v − x0 ))
= Rα (u)−1 [F (u)(u − v) − (F(u) − F(v))]
−Rα (u)−1 [F (v) − F (u)](v − G(v))
1
−1
= Rα (u) [F (u)(u − v) + (F (u + t(v − u))(v − u))dt]
0
−Rα (u)−1 [F (v) − F (u)](v − G(v))
1
= Rα (u)−1 [(F (u + t(v − u)) − F (u))(v − u)dt]
0
−Rα (u)−1 [F (v) − F (u)](v − G(v)).
The last, but one step follows from the Lemma 3.2.1. So by Assumption 8
and the estimate Rα (u)−1 F (u) ≤ 1, we have
1
G(u) − G(v) ≤ L [u + t(v − u) − x0 + u − x0 ]dtv − u
0
+L[v − x0 + u − x0 ]v − u
L
≤ [3u − x0 + v − x0 ] + L[v − x0 + u − x0 ]
2
L
≤ [5u − x0 + 3v − x0 ]v − G(u). (3.37)
2
54 Iterative methods and their dynamics with applications: a contemporary study
δ δ
Now by taking u = xn,α and v = xn−1,α in (3.37), we obtain (3.36).
Next we shall prove that the sequence (tn ) defined in Lemma 3.2.1 is a ma-
δ
jorizing sequence of the sequence (xn,α ). Note that F(x∗ ) = y, so by Lemma
3.2.2,
δ
x1,α − x0 = Rα (x0 )−1 (F(x0 ) − yδ )
= Rα (x0 )−1 (F(x0 ) − y + y − yδ )
= Rα (x0 )−1 (F(x0 ) − F(x∗ ) − F (x0 )(x0 − x∗ )
+F (x0 )(x0 − x∗ ) + y − yδ )
≤ Rα (x0 )−1 (F(x0 ) − F(x∗ ) − F (x0 )(x0 − x∗ ))
+Rα (x0 )−1 F (x0 )(x0 − x∗ ) + Rα (x0 )−1 (y − yδ )
1
≤ Rα (x0 )−1 F (x0 ) Φ(x∗ + t(x0 − x∗ ), x0 , (x0 − x∗ ))dt
0
−1 δ
+Rα (x0 ) F (x0 ))(x0 − x∗ ) +
α
L δ
≤ x0 − x∗ 2 + x0 − x∗ +
2 α
L 2 δ
≤ ρ +ρ +
2 α
≤ η = t1 − t0 .
δ δ
Assume that xi+1,α − xi,α ≤ ti+1 − ti for all i ≤ k for some k. Then
δ δ δ δ δ δ
xk+1,α − x0 ≤ xk+1,α − xk,α + xk,α − xk−1,α + · · · + x1,α − x0
≤ tk+1 − tk + tk − tk−1 + · · · + t1 − t0
= tk+1 ≤ t ∗ .
δ
As a consequence, xi+1,α ∈ Bt ∗ (x0 ) for all i ≤ k, and therefore, by (3.36),
δ δ L δ δ δ δ δ δ
xk+2,α − xk+1,α ≤ [5xn,α − x0,α + 3xn−1,α − x0,α ]xn,α − xn−1,α
2
L
≤ (5tk+1 + 3tk−1 ) = tk+2 − tk+1 .
2
δ δ
Hence, by induction xn+1,α − xn,α ≤ tn+1 − tn for all n ≥ 0 and therefore
δ δ
{tn }, n ≥ 0 is a majorizing sequence of the sequence {xn,α }. In particular xn,α −
∗ δ ∗ δ
x0 ≤ tn ≤ t , i.e., xn,α ∈ U(x0 ,t ), for all n ≥ 0. So, {xn,α }, n ≥ 0 is a Cauchy
sequence and converges to some xαδ ∈ U(x0 ,t ∗ ) ⊂ U(x0 ,t ∗∗ ) and by Lemma 3.2.3
qn η
xαδ − xn,α
δ
≤ t ∗ − tn ≤ .
1−q
In order to prove (3.35), we observe that G(xαδ ) = xαδ , so (3.35) follows from
Nonlinear Ill-posed equations 55
δ
(3.37), by taking u = xn,α and v = xαδ in (3.37). Now letting n → ∞ in (3.10) we
obtain F(xαδ ) = yδ + α(x0 − xαδ ). 2
Remark 3.2.6 The convergence of the method is quadratic as it has been shown
in [14], under Assumption 6. Since the error bounds shown in Theorem 3.2.5
are too pessimistic we will use the famous computational order of convergence
(COC) ( [5]) defined by
xn+1 − xαδ xn − xαδ
ρ ≈ ln / ln .
xn − xαδ xn−1 − xαδ
Let
η
c̄ := max{ + 1, (L0 r + 2)}, (3.38)
1−q
and let
δ
}.
nδ := min{n : qn ≤ (3.39)
α
Theorem 3.3.4 Let c̄ and nδ be as in (3.38) and (3.39) respectively. Suppose
that hypotheses of Theorem 3.3.3 hold. Then, the following assertions hold
δ
xnδδ ,α − x∗ ≤ c̄(ϕ(α) + ). (3.40)
α
Note that the error estimate ϕ(α) + αδ in (3.32) is of optimal order if α := αδ
satisfies, ϕ(αδ )αδ = δ .
Now using the function ψ(λ ) := λ ϕ −1 (λ ), 0 < λ ≤ a we have δ =
αδ ϕ(αδ ) = ψ(ϕ(αδ )), so that αδ = ϕ −1 (ψ −1 (δ )).
Now, we present the following result.
Theorem 3.3.5 Let ψ(λ ) := λ ϕ −1 (λ ) for 0 < λ ≤ a, and the assumptions in
Theorem 3.3.4 hold. For δ > 0, let α := αδ = ϕ −1 (ψ −1 (δ )) and let nδ be as in
(3.39). Then
xnδδ ,α − x∗ = O(ψ −1 (δ )).
Now, we present a parameter choice rule based on the famous balancing prin-
ciple studied in [12, 19, 21]. In this method, the regularization parameter α is
chosen from some finite set
DM (α) := {αi = μ i α0 , i = 0, 1, · · · , M}
where μ > 1, α0 > 0 and let
δ
ni := min{n : e−γ0 n ≤ }.
αi
Then for i = 0, 1, · · · , M, we have
δ
xnδi ,αi − xαδ i ≤ c , ∀i = 0, 1, · · · M.
αi
Let xi := xnδi ,αi . The parameter choice strategy that we are going to consider in
this chapter, we select α = αi from DM (α) and operate only with corresponding
xi , i = 0, 1, · · · , M.
Theorem 3.3.6 ( [22, Theorem 3.1], see also the proof) Assume that there exists
i ∈ {0, 1, 2, · · · , M} such that ϕ(αi ) ≤ αδi . Suppose the hypotheses of Theorem
3.3.3 and Theorem 3.3.4 hold and let
δ
l := max{i : ϕ(αi ) ≤ } < M,
αi
Nonlinear Ill-posed equations 57
δ
k := max{i : xi − x j ≤ 4c̄ , j = 0, 1, 2, · · · , i}.
αj
Then l ≤ k and
x∗ − xk ≤ cψ −1 (δ )
where c = 6c̄μ.
On the other hand, the adaptive algorithm associated with the choice of the
parameter specified in Theorem 3.3.6 involves the following steps:
3.4.1 Algorithm
Set i ← 0.
Solve xi := xnδi ,αi by using the iteration (3.10).
√
δ
If xi − x j > 4c μj , j ≤ i, then take k = i − 1.
Set i = i + 1 and return to Step 2.
|x − x0 |
= i−1 i−1
x0 i + · · · + x i
so
F (x) − F (x0 ) ≤ L0 |x − x0 |.
Example 3.5.2 ( [7, Example 7.5]) We consider the integral equations
b
u(s) = f (s) + λ G(s,t)u(t)1+1/n dt, n ∈ N. (3.42)
a
Here, f is a given continuous function satisfying f (s) > 0, s ∈ [a, b], λ is a real
number, and the kernel G is continuous and positive in [a, b] × [a, b].
Equation of the form (3.42) generalize equations of the form
b
u(s) = G(s,t)u(t)n dt (3.43)
a
studied in [1]– [6]. Instead of (3.42) we can try to solve the equation F(u) = 0
where
F : Ω ⊆ C[a, b] → C[a, b], Ω = {u ∈ C[a, b] : u(s) ≥ 0, s ∈ [a, b]},
and b
F(u)(s) = u(s) − f (s) − λ G(s,t)u(t)1+1/n dt.
a
We consider the max-norm.
The derivative F is given by
b
1
F (u)v(s) = v(s) − λ (1 + ) G(s,t)u(t)1/n v(t)dt, v ∈ Ω.
n a
It is worth noticing that F does not satisfy a Lipschitz-type condition in Ω. Let
us choose [a, b] = [0, 1], G(s,t) = 1 and y(t) = 0. Then F (y)v(s) = v(s) and
b
1
F (x) − F (y) = |λ |(1 + ) x(t)1/n dt.
n a
Nonlinear Ill-posed equations 59
would hold for all x ∈ Ω and for a constant L2 . But this is not true. Consider, for
example, the functions
t
x j (t) = , j ≥ 1, t ∈ [0, 1].
j
If these are substituted into (3.44)
1 L2
≤ ⇔ j1−1/n ≤ L2 (1 + 1/n), ∀ j ≥ 1.
j1/n (1 + 1/n) j
This inequality is not true when j → ∞.
Therefore, Assumption 6 is not satisfied in this case. However, Assumption
8 holds. To show this, let x0 (t) = f (t) and γ = mins∈[a,b] f (s), α > 0. Then for
v ∈ Ω,
b
1
[F (x) − F (x0 )]v = |λ |(1 + ) max | G(s,t)(x(t)1/n − f (t)1/n )v(t)dt|
n s∈[a,b] a
1
≤ |λ |(1 + ) max Gn (s,t)
n s∈[a,b]
G(s,t)|x(t)− f (t)|
where Gn (s,t) = x(t)(n−1)/n +x(t)(n−2)/n f (t)1/n +···+ f (t)(n−1)/n
.
Hence,
|λ |(1 + 1/n) b
[F (x) − F (x0 )]v = max G(s,t)v(t)dtx − x0
γ (n−1)/n s∈[a,b] a
≤ L0 vx − x0 ,
|λ |(1+1/n) b
where L0 = γ (n−1)/n
N and N = maxs∈[a,b] a
G(s,t)dt. Then by the following
inequality;
[F (x)−F (u)]v ≤ [F (x)−F (x0 )]v+[F (x0 )−F (u)]v ≤ L0 v(x−x0 +u−x0 )
60
References 61
Sixth-order iterative
methods
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
4.2 Scheme for constructing sixth-order iterative methods . . . . . . . . . . . . . . . 65
4.3 Sixth-order iterative methods contained in family USS . . . . . . . . . . . . . . 69
4.4 Numerical work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
4.4.1 Solving nonlinear equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
4.5 Dynamics for method SG . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
4.5.1 Study of the fixed points and their stability . . . . . . . . . . . . . . . . . 73
4.5.2 Study of the critical points and parameter spaces . . . . . . . . . . . 74
4.1 Introduction
We develop sixth-order iterative methods in order to approximate zeros x∗ of
the function f defined on the real line. This method can be used to solve many
63
64 Iterative methods and their dynamics with applications: a contemporary study
problems from computational sciences and other disciplines [14, 17, 25, 34, 36,
37].
Sharma and Guha [35], through a modification of the well-known the Os-
trowski method [32], developed the following sixth-order convergent method
(SG) ⎧
⎪
⎪ f (xn )
⎪
⎪ yn = xn − ,
⎪
⎪ f (xn )
⎨ f (xn ) f (yn )
zn = yn − , (4.1)
⎪
⎪ f (x ) − 2 f (y ) f (x )
⎪
⎪
n n
f (xn ) + a f (yn )
n
f (zn )
⎪
⎪
⎩ xn+1 = zn − ,
f (xn ) + (a − 2) f (yn ) f (xn )
where a ∈ R.
Neta proposed a three-step uniparametric family of sixth-order (N)
⎧
⎪
⎪ f (xn )
⎪
⎪ yn = xn − ,
⎪
⎪ f (xn )
⎨ f (xn ) + α f (yn ) f (yn )
zn = yn − , (4.2)
⎪
⎪ f (xn ) + (α − 2) f (yn ) f (xn )
⎪
⎪ f (xn ) − f (yn ) f (zn )
⎪
⎪
⎩ xn+1 = zn − ,
f (xn ) − 3 f (yn ) f (xn )
[31].
Grau and Dı́az in [18] developed another sixth-order variant of the Os-
trowski’s method GD
⎧
⎪
⎪ f (xn )
⎪
⎪ yn = xn − ,
⎪
⎪ f (xn )
⎨ f (xn ) f (yn )
zn = yn − , (4.3)
⎪
⎪ f (xn ) − 2 f (yn ) f (xn )
⎪
⎪ f (xn ) f (zn )
⎪
⎪
⎩ xn+1 = zn − .
f (xn ) − 2 f (yn ) f (xn )
Chun and Ham [9] derived the following family of sixth-order variant of the
Ostrowski’s method CH
⎧
⎪
⎪ f (xn )
⎪
⎪ yn = xn − ,
⎪
⎪ f (xn )
⎨ f (xn ) f (yn )
zn = yn − , (4.4)
⎪
⎪ f (x ) − 2 f (y ) f (x )
⎪
⎪
n
f (zn )
n n
⎪
⎪
⎩ xn+1 = zn − H(un ) ,
f (xn )
Sixth-order iterative methods 65
where un = f (yn )/ f (xn ) and H(t) represents a real valued function satisfying:
H(0) = 1 and H (0) = 2.
1+β t
If H(t) = , then (CH)=(SG)
1 + (β − 2)t
1
If H(t) = , then (CH)=(GD)
1 − 2t
Chun and Neta [10] proposed the following sixth-order iterative method
based on a modification of the Kung and Traub method [24] CN
⎧
⎪
⎪ f (xn )
⎪
⎪ yn = xn − ,
⎪
⎪ f (xn )
⎪
⎪ f (yn )
⎪
⎨ zn = yn −
1
2 ,
f (xn ) f (yn ) (4.5)
⎪ 1 − f (xn )
⎪
⎪
⎪
⎪ f (zn ) 1
⎪
⎪ xn+1 = zn − 2 .
⎪
⎪ f (xn )
⎩ 1 − ff (yn)
− f (zn )
(xn ) f (xn )
f (x) c2 2 c3 c1 − c2 2 3 3 c4 c1 2 − 7 c2 c3 c1 + 4 c2 3 4
= en − en − 2 en − en + O en 5 .
f (x) c1 c1 2 c1 3
(4.10)
f (xn )
yn − x∗ = en − , (4.11)
f (xn )
c2 2 c3 c1 − c2 2 3 3 c4 c1 2 − 7 c2 c3 c1 + 4 c2 3 4
yn − x ∗ = en + 2 2
en + 3
e n + O en 5 .
c1 c1 c1
(4.12)
Using Taylor expansion on f (yn ) around the solution x∗ we obtain the following
and using f (x∗ ) = 0 we get
∞
f (yn ) = c j (yn − x∗ ),
j=1
Sixth-order iterative methods 67
c3 c1 − c2 2 3 3 c4 c1 2 − 7 c2 c3 c1 + 5 c2 3 4
f (yn ) = c2 en 2 + 2 en + 2
en + O e n 5
c1 c1
(4.13)
f (yn ) c2 2 c3 c1 − 3 c2 2 2 3 c4 c1 2 − 10 c2 c3 c1 + 8 c2 3 3
= en + en + en + O en 4
f (xn ) c1 c1 2 c1 3
(4.14)
substituting f (xn )/ f (xn ) from the equation (4.10), and f (yn )/ f (xn ) from the
equation (4.14), into the above equation we obtain
∗ c2 2 (a1 − 2) 3 c2 4 a1 c3 c1 − 7 a1 c2 2 + a2 c2 2 − 7 c3 c1 + 9 c2 2 4
∗
zn − x = x − en − en
c1 2 c1 3
+ O en 5 . (4.16)
2
f (zn ) f (yn ) f (zn ) f (yn ) f (zn )
xn+1 = zn − 1 + b 1 μ1 + μ2 + b2 μ1 + μ2 +··· ,
f (xn ) f (xn ) f (xn ) f (xn ) f (xn )
68 Iterative methods and their dynamics with applications: a contemporary study
substituting from the equations (4.9), (4.14), (4.18) into the above equation we
obtain the error relation
c2
− 2 a2 + 18 a1 − 26)c22 )e5n + 5 (−16 a1 − 16 b1 μ1 + 28 + 9 a1 b1 μ1 ) c1 2 c2 c4
c1
+ (12 a1 b1 μ1 − 20 b1 μ1 + 33 − 20 a1 ) c1 2 c3 2 + (131 b1 μ1 − 89 a1 b1 μ1 − 28 b1 μ2
− 15 b2 μ1 2 + 8 a1 b2 μ1 2 + 8 b1 μ1 a2 + 30 b1 μ2 a1 − 8 b1 μ2 a1 2 − 15 a2 + 125 a1
− 167)c1 c2 2 c3 + 86 a1 b1 μ1 − 15 b1 μ1 a2 − 58 b1 μ2 a1 + 17 b1 μ2 a1 2 − 15 a1 b2 μ1 2
+ 4 a2 b1 μ2 + a2 b2 μ1 2 − 8 b2 μ1 μ2 + a1 b3 μ1 3 + b1 μ1 a3 − 2 a2 b1 μ2 a1
+ 8 b2 μ1 μ2 a1 − 2 b2 μ1 μ2 a1 2 − a1 2 − 98 a1 + 24 a2 − 2 a3 + 108 + 48 b1 μ2
− 111 b1 μ1 + 25 b2 μ1 2 − 2 b3 μ1 3 )c2 4 )e6n + O(e7n ).
(4.19)
Three-step scheme (4.6) define a family of sixth order methods if the following
three equations are satisfied simultaneously
⎧
⎪
⎪ (a1 − 2)(b1 μ1 − 2) = 0,
⎨
−11 a1 − 11 b1 μ1 +6 b1 a1 μ1 + 20 = 0,
⎪
⎪ (−2 + a 1 ) μ 1
2
b2 + (−12 a 1 + a2 + 19) μ1 + −4 + 4 a1 − a1
2
μ 2
⎩
b1 − 2 a2 + 18 a1 − 26 = 0.
It is easy to get
2
a1 = 2 and b1 = .
μ1
Substituting a1 = 2 and b1 = 2/μ1 in the equation (4.19) produces the required
error equation (4.7).
Therefore, we present the following three-step sixth-order iterative scheme
USS
⎧
⎪
⎪ f (xn )
⎪
⎪ yn = xn − ,
⎪
⎪ f (xn )
⎪
⎪ j
⎪
⎪ f (y ) f (y ) m f (y )
⎪
⎪ = yn −
n
1+2
n
+ j=2 a j
n
,
⎨ zn f (xn ) f (x ) f (x )
n n (4.20)
⎪
⎪ f (zn ) 2 μ1 f (yn ) + μ2 f (zn )
⎪
⎪ xn+1 = z n − 1 +
⎪
⎪ f (xn ) μ1 f (xn )
⎪
⎪ k
⎪
⎪ l μ f (y ) + μ f (zn)
⎪
⎪ + k=2 bk
1 n 2
.
⎩ f (xn )
Sixth-order iterative methods 69
aj = 2j for j ≥ 2, m = ∞, μ1 = 1,
bk = 2 (2 − β ) k−1
for k ≥ 2, l = ∞, μ2 = 0.
aj = 2j for j ≥ 2, m = ∞, μ1 = 1,
b k = ωk for k ≥ 2, l = ∞, μ2 = 0.
un = f (yn )/ f (xn ) and H(t) is a real valued function satisfying H(0) = 1 and
H (0) = 2. Moreover, notice that function H(un ) can be expressed as
∞ k
f (yn ) f (yn ) f (yn )
H = 1+2 + ωk .
f (xn ) f (xn ) f (xn )
k=2
Hence, method (4.21), that comes from scheme (USS) is method CH.
Method (GC) can be obtained using
aj = 2j for j ≥ 2, m = ∞, μ1 = 1,
bk = 2 k
for k ≥ 2, l = ∞, μ2 = 0.
Where, a ∈ R.
Finally, to obtain CN, parameters are
a j = ( j + 1) for j ≥ 2, m = ∞, μ1 = 1,
bk = (k + 1) for k ≥ 2, l = ∞, μ2 = 1.
Using the equivalence 1 + 2r + 3r2 + 4r3 + · · · = 1/(1 − r)2 for |r| < 1 in the
second and third steps of preceding method we get method (CN).
then ξ is called the convergence order of the sequence and the constant C is called
the asymptotic error constant. From the preceding relation, the computational
order of convergence (COC) is approximated as follows
All the computations are performed in the programming language C++ . For
numerical precision, the C++ library ARPREC [1] is being used. For conver-
gence of the method, it is required that the distance of two consecutive iterates
(|xn+1 − xn |) and the absolute value of the function (| f (xn )|), also referred to as
residual, be less than 10−300 . The maximum allowed iterations are 200.
Table 4.1: The (number of function evaluations, COC) for various sixth order iterative
methods.
f (x) x0 SG N GD CH CN
f1 (x) 100 (36, 6) (36, 6) (28, 6) (36, 6) (28, 6)
f2 (x) 100 (172, 6) (884, 6) (160, 6) (172, 6) (108, 6)
f3 (x) −1 (20, 6) (20, 6) (20, 6) (20, 6) (32, 6)
f4 (x) 1 (20, 6) (20, 6) (20, 6) (20, 6) (20, 6)
f5 (x) 1 (24, 6) (24, 6) (24, 6) (24, 6) (24, 6)
f5 (x) 3 (24, 6) (28, 6) (20, 6) (24, 6) (20, 6)
f6 (x) −1.5 (24, 6) (28, 6) (20, 6) (24, 6) (20, 6)
Various free parameters are choosen as, in the method SG: a = 2, in the
method N: α = 5 and in the method CH: β = 3. Outcome of numerical experi-
mentation is presented in the Table 4.1. The Table 4.1 reports (number of function
evaluations, COC during the second last iterative step). COC is rounded to the
72 Iterative methods and their dynamics with applications: a contemporary study
nearest significant digits. In the Table 4.1, we see that the methods GD and CN
are showing better results.
The Fatou set of the rational function R, F (R) , is the set of points z ∈ Ĉ
whose orbits tend to an attractor (fixed point, periodic orbit or infinity). Its com-
plement in Ĉ is the Julia set, J (R). That means that the basin of attraction of
any fixed point belongs to the Fatou set and the boundaries of these basins of
attraction belong to the Julia set.
By applying this operator on a quadratic polynomial with two different roots
A and B, p(z) = (z − A)(z − B) and using the Möebius map h(z) = z−A z−B , which
Sixth-order iterative methods 73
Theorem 4.5.0
The character of the strange fixed point z = 1 is:
√
−975 (a)2 −3928 (a)−3952
i) If −(52/25) < Re(a) < − 39 and −
76 √ < (a) <
√ 5 39
−975 (a)2 −3928 (a)−3952
√
5 39
, then z = 1 is an attractor. Moreover, if a = −2
it is superattractor.
74 Iterative methods and their dynamics with applications: a contemporary study
√
−975 (a)2 −3928 (a)−3952
ii) When −(52/25) < Re(a) < − 76 and (a) = − √ ∨
√ 39 5 39
−975 (a) −3928 (a)−3952
2
(a) = √
5 39
z = 1 is a parabolic point.
−975(a)2 − 3928(a) − 3952
G (z, 1) = (a) = − √ ∨ (a)
5 39
−975(a)2 − 3928(a) − 3952
= √
5 39
In Figure 4.2 the region in which z = 1 is attractor is shown.
cr5 (a) = −1
cr6 (a) = i
It is easy to see that cr5 and cr6 are preimages of the extraneous fixed point
z = 1, so we will not consider them in the dynamical study, as the stability of this
fixed point has been already analyzed.
In Figure 4.3 the bifurcation diagram of the critical points is shown
So, there are four independent free critical points in which we are interested,
without loss of generality, we consider in this paper the free critical points cr1 (a)
and cr2 (a), since its behavior is representative. Now, we are going to look for the
best members of the family by means of using the parameter space associated to
the free critical points.
The study of the orbits of the critical points gives rise about the dynamical
behavior of an iterative method. In concrete, to determinate if there exists any
76 Iterative methods and their dynamics with applications: a contemporary study
attracting periodic orbit different to the roots of the polynomial p(z), the follow-
ing question must be answered: For which values of the parameter, the orbits
of the free critical points are attracting periodic orbits? One way of giving re-
sponse to that questions is drawing the parameter spaces associated to the free
critical points. There will exist open regions of the parameter space for which the
iteration of the free critical points does not converge to any of the roots of the
polynomial p(z), that is, in these regions the critical points converges to attracting
cycles or to an strange fixed point or even they diverge.
In Figure 4.4, the parameter spaces associated to cr1 (a) are shown in which
we observe that there exists a zone with no convergence to the roots, this zone can
be better seen in Figure 4.8, and in Figure 4.5 the parameter spaces associated to
cr2 (a) is shown. A point is painted in cyan if the iteration of the method starting
in z0 = cr1 (a) converges to the fixed point 0 (related to root A), in magenta if it
converges to ∞ (related to root B) and in yellow if the iteration converges to 1
(related to ∞). Moreover, it appears in red the convergence, after a maximum of
2000 iterations and with a tolerance of 10−6 , to any of the strange fixed points,
in orange the convergence to 2-cycles, in light green the convergence to 3-cycles,
in dark red to 4-cycles, in dark blue to 5-cycles, in dark green to 6-cycles, dark
yellow to 7-cycles, and in white the convergence to 8-cycles. The regions in
black correspond to zones of convergence to other cycles. As a consequence,
every point of the plane which is neither cyan nor magenta is not a good choice
of α in terms of numerical behavior.
Once the anomalies has been detected, the next step consist on describing
them in the dynamical planes. In these dynamical planes the convergence to 0
will appear in magenta, in cyan it appears the convergence to ∞ and in black
the zones with no convergence, after a maximum of 2000 iterations and with a
tolerance of 10−6 to the roots.
Then, focussing the attention in the region shown in Figure 4.8 it is evident
that there exist members of the family with complicated behavior. In Figure 4.9,
the dynamical planes of a member of the family with regions of convergence to
some of the strange fixed points is shown.
In Figures 4.10, 4.15 dynamical planes of members of the family with regions
of convergence to an attracting 2-cycle is shown, those regions with convergence
to 2-cycles are painted in black since there is no convergence to any of the roots.
Both values of parameter correspond to red zones in the parameter planes.
On the other hand, in Figure 4.12, the dynamical plane of a member of the
family with regions of convergence to z = 1, related to ∞ is presented, painted
in black. This value of parameter a is located in a yellow zone of the parameter
planes.
Other special cases are shown in Figures 4.13, 4.14, 4.15 and 4.16 where
there are no convergence problems, as the only basins of attraction correspond to
the roots of p(z).
Sixth-order iterative methods 77
Figure 4.4: Parameter space associated to the free critical point cr1 (a).
Figure 4.5: Parameter space associated to the free critical point cr2 (a).
78 Iterative methods and their dynamics with applications: a contemporary study
Figure 4.6: Parameter space associated to the free critical point cr3 (a).
Figure 4.7: Parameter space associated to the free critical point cr4 (a).
Sixth-order iterative methods 79
Figure 4.8: Detail of the parameter space associated to the free critical point cr1 (a).
As a consequence, we can state that this family is very stable and that the
existing anomalies are very rare in practice.
References
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
5.2 Local convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
5.3 Basins of attraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
5.3.1 Basins of F(x) = (x − 1)2 (x + 1) . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
88
Local convergence and basins of attraction of a two-step Newton-like method 89
5.1 Introduction
Let S = R or S = C, D ⊆ S be convex and let F : D → S be a differentiable
function. We shall approximate solutions x∗ of the equation
F(x) = 0, (5.1)
Many problems from Applied Sciences including engineering can be solved
finding the solutions of equations in a form like (5.1) [4, 5, 20, 25, 31, 39, 43–45].
We study the local convergence and the dynamics of the two-step Newton-
like method defined for each n = 0, 1, 2, . . . by
to obtain
1 (m − 1)2 F (x)2
φ (x) = −1 + m(m + 1) − .
g 2 F (x)F(x)
Other choices of φ can be found in [4, 5, 13, 29, 30, 33, 34, 36–41, 45, 46] and
the references therein.
The rest of the chapter is organized as follows: Section 5.2 contains the lo-
cal convergence analysis of method (5.2). The dynamics of method (5.2) are
discussed in Section 5.3. Finally, the numerical examples are presented in the
concluding Section 5.4.
since g0 (r0 ) = 1. It follows from the intermediate value theorem that function H1
has zeros in the interval (0, r0 ). Denote by r1 the smallest such zero. Then, we
have that
Suppose that there exists a continuous function ψ : [0, r1 ) → [0, +∞) such
that m−2
1 L
hm (r0 ) + (|1 − α| + ψ(0)) − 1 < 0. (5.5)
2 m
Define function on the interval [0, r1 ) by
m−2
1 1 L
g2 (t) = hm (t) + (|1 − α| + ψ(t))
1 − g0 (t) 2 m
Then, function H2 has zeros in the interval (0, r1 ]. Denote the smallest such zero
by r. Then, we have that
r ≤ r1 < r0 , (5.6)
92 Iterative methods and their dynamics with applications: a contemporary study
and
xn+1 − x∗ ≤ g2 (xn − x∗ )xn − x∗ < xn − x∗ , (5.16)
where the “g” functions are defined previously. Furthermore, for R ∈ [r, r0 ) the
limit point x∗ is the only solution of equation F(x) = 0 in Ū(x∗ , R) ∩ D, where r0
is defined previously.
Proof: We shall show estimates (5.15) and (5.16) using mathematical induction.
By hypothesis x0 ∈ U(x∗ , r) \ {x∗ }. Then, we have in turn the approximation
Local convergence and basins of attraction of a two-step Newton-like method 93
= ···
11 1
= 0 0 · · · 0 F (m) (x∗ + θm−1 θm−2 . . . θ1 (x0 − x∗ )) − F (m) (x∗ )
(5.17)
Then, using (5.6), (5.7), (5.12) and (5.17), we get for x0 = x∗ that
−1
F (m) (x∗ )(x0 − x∗ )m−1 F (x0 ) − F (m) (x∗ )(x0 − x∗ )m−1
! 1 −1 (m) ∗
1 1
≤ x0 − x∗ 1−m 0 0 . . . 0 F (m) (x∗ ) F (x + θm−1 θm−2 . . . θ1
It follows from (5.18) and the Banach lemma on invertible functions [4,5,39,
45] that F (x0 )−1 ∈ Ł(R, R) and
1
F (x0 )−1 F (m) (x∗ ) ≤ m−2 m−2
L0 1 1
x0 − x∗ m−1 1 − ∗
x0 − x + 1 −
2 3 2
(5.19)
Hence, y0 is well defined by the first substep of method (5.2) for n = 0. Using
method (5.2) for n = 0 and (5.10) we obtain the approximation
(5.20)
We shall find some upper bounds on the norms of the integral expression and
F(x0 ) in (5.20). We have in turn the following
1
[F (x∗ + θ1 (x0 − x∗ )) − F (x0 )] (x0 − x∗ )dθ1
0
11
= − 0 0 F (x0 + θ2 (x∗ + θ1 (x0 − x∗ ) − x0 ))(1 − θ1 )(x0 − x∗ )2 dθ1 dθ2
11
= − 0 0 [F (x0 + θ2 (θ1 − 1)(x0 − x∗ )) − F (x∗ )] (1 − θ1 )(x0 − x∗ )2 dθ1 dθ2
111
= 0 0 0 F (x∗ + θ3 (x0 + θ2 (θ1 − 1)(x0 − x∗ ) − x∗ ))(1 − (1 − θ1 )θ2 )(1 − θ1 )
= ···
11 1
= 0 0 . . . 0 F (m) (x∗ + θm θm−2 . . . θ3 (1 − (1 − θ1 )θ2 )(x0 − x∗ )) − F (m) (x∗ )
(5.21)
Local convergence and basins of attraction of a two-step Newton-like method 95
Then, using (5.21) the definition of function hm and (5.12) we get that
1
F (m) (x∗ )−1 F (x∗ + θ1 (x0 − x∗ )) − F (x0 ) (x0 −x∗ )dθ1 ≤ hm (x0 −x∗ )x0 −x∗ m .
0
(5.22)
We also have that
1
F (x0 ) = F (x0 ) − F (x∗ ) = 0 F (x∗ + θ1 (x0 − x∗ ))(x0 − x∗ )dθ1
11 1
= · · · = 0 0 . . . 0 F (m) (x∗ + θm−1 θm−2 . . . θ1 (x0 − x∗ )) (5.23)
×θm−2 . . . θ1 (x0 − x∗ )m−1 dθ1 dθ2 . . . dθm−1
+|1 − α|F (x0 )−1 F (m) (x∗ )F (m) (x∗ )−1 F(x0 )
m−2
∗ |1 − α|L 1
hm (x0 − x ) + x0 − x∗ m
m 2
≤
x0 − x∗ m−1 (1 − g0 (x0 − x∗ ))
= g1 (x0 − x∗ )x0 − x∗ < x0 − x∗ < r,
(5.27)
which shows (5.15) for n = 0 and y0 ∈ U(x∗ , r). We also have that x1 is well
defined by the second substep of method (5.2) for n = 0. Using method (5.2) for
96 Iterative methods and their dynamics with applications: a contemporary study
n = 0, we have that
x1 − x∗ = x0 − x∗ − φ (x0 )F (x0 )−1 F(x0 )
x1 − x∗ ≤ y0 − x∗ + |1 − φ (x0 )|F (x0 )−1 F (m) (x∗ )F (m) (x∗ )−1 F(x0 )
m−1
∗
ψ(x0 − x∗ ) mL 12 x0 − x∗ m−1
≤ g1 (x0 − x ) + x0 − x∗
x0 − x∗ m−1 (1 − g0 (x0 − x∗ ))
= g2 (x0 − x∗ )x0 − x∗ < x0 − x∗ < r,
(5.29)
∗
which shows (5.16) for n = 0 and x1 ∈ U(x , r). By simply replacing x0 , y0 , x1
by xk , yk , xk+1 in the preceding estimates we arrive at estimates (5.15) and (5.16).
Then, from the estimates xk+1 −x∗ ≤ cxk −x∗ < r, x = g2 (x0 −x∗ ) ∈ [0, 1),
we get that lim xk = x∗ and xk+1 ∈ U(x∗ , r). Finally to show the uniqueness part,
k→∞
let y∗ ∈ Ū(x∗ , R) be such that F(y∗ ) = F (y∗ ) = · · · = F (m−1) (y∗ ) = 0, F (m) (y∗ ) =
1
0. Set T = 0 F (y∗ + θ (x∗ − y∗ ))dθ . Then, in view of (5.18) for x0 replaced by
z = y∗ + θ (x∗ − y∗ ) we get in turn that
It follows that T −1 exists. Then, from the identity 0 = F(y∗ ) − F(x∗ ) = T (y∗ −
x∗ ), we conclude that x∗ = y∗ .
of iterative methods, which allow us to distinguish between the good and bad
methods in terms of its numerical properties.
Firstly, some dynamical concepts of complex dynamics that are used in this
work are shown. Given a rational function R : Ĉ → Ĉ, where Ĉ is the Riemann
sphere, the orbit of a point z0 ∈ Ĉ is defined as
The Fatou set of the rational function R, F (R), is the set of points z ∈ Ĉ
whose orbits tend to an attractor (fixed point, periodic orbit or infinity). Its com-
plement in Ĉ is the Julia set, J (R). That means that the basin of attraction of
any fixed point belongs to the Fatou set and the boundaries of these basins of
attraction belong to the Julia set.
Now, we are going to see some of the basin of attractions related to different
choices of α and φ (x). In concrete we have chosen 3 values of α
1
α=
2
α =1
α = −1
and three choices of functions φ
φ (x) = x2
98 Iterative methods and their dynamics with applications: a contemporary study
x
φ (x) =
2
φ (x) = x4 + 1
when method (5.2) is applied to three different polynomials
F(x) = (x − 1)2 (x + 1)
F(x) = (x − 1)3 (x + 1)
F(x) = (x − 1)4 (x + 1).
A point is painted in cyan if the iteration of the method starting in x0 con-
verges to the fixed point 1 (multiplicity greater than one), in magenta if it con-
verges to −1 (simple one) and in yellow if the iteration diverges to ∞. If after
100 iterations the sequence does not converge to any of the roots with a tolerance
of 10−3 or diverges to infinity that point is painted in black.
α = 0.5 α =1 α = −1
Figure 5.1: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)2 (x + 1) with φ (x) = x2 .
In Figure 5.2 we see how the basin change when α change with φ (x) = 2x .
In Figure 5.3 we see how the basin change when α change with φ (x) = x4 +1.
Local convergence and basins of attraction of a two-step Newton-like method 99
α = 0.5 α =1 α = −1
Figure 5.2: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)2 (x + 1) with φ (x) = 2x .
α = 0.5 α =1 α = −1
Figure 5.3: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)2 (x + 1) with φ (x) = x4 + 1.
α = 0.5 α =1 α = −1
Figure 5.4: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)3 (x + 1) with φ (x) = x2 .
α = 0.5 α =1 α = −1
Figure 5.5: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)3 (x + 1) with φ (x) = 2x .
α = 0.5 α =1 α = −1
Figure 5.6: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)3 (x + 1) with φ (x) = x4 + 1.
Local convergence and basins of attraction of a two-step Newton-like method 101
α = 0.5 α =1 α = −1
Figure 5.7: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)4 (x + 1) with φ (x) = x2 .
α = 0.5 α =1 α = −1
Figure 5.8: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)4 (x + 1) with φ (x) = 2x .
In Figure 5.9 we see how the basin change when α change with φ (x) = x4 +1.
α = 0.5 α =1 α = −1
Figure 5.9: Basins of attraction of the method (5.2) associated to the roots of
F(x) = (x − 1)4 (x + 1) with φ (x) = x4 + 1.
1 3
since L = max |(x + 2)ex | = e and
2 −1≤x≤1 2
1 x xn−1 1 1 1
L0 = max |2(1+ +· · ·+ +· · · )+e | =
x
2(1 + + · · · + + · · · ) + e
2 −1≤x≤1 2! x! 2 2 n!
1
= (3e − 4).
2
Now choosing
α = 0.85,
φ (x) = φ = 1 − x
and
ψ(x) = |x|
we obtain that conditions of Theorem hold and method (5.2) with α = 0.85 and
φ (x) = 1 − x converges to x∗ = 0. Moreover,
r0 = 0.962731 . . . ,
r1 = 0.112174 . . .
and
r = 0.051511 . . . .
Example 5.4.2 Let S = R, D = (1, 3) and define function F on D by
L = 743
Local convergence and basins of attraction of a two-step Newton-like method 103
and
L0 = 742.
Now choosing
α = 1,
φ (x) = 3x
and
ψ(x) = t − 3
we obtain that conditions of Theorem hold and method (5.2) with α = 1 and
φ (x) = 3x converges to x∗ = 2. Moreover,
r0 = 0.0136456 . . . ,
r1 = 0.00341139 . . .
and
r = 0.000710411 . . . .
References
[1] S. Amat, S. Busquier and S. Plaza, Dynamics of the King and Jarratt itera-
tions, Aequationes Math., 69 (3), 212–223, 2005.
[2] S. Amat, S. Busquier and S. Plaza, Chaotic dynamics of a third-order
Newton-type method, J. Math. Anal. Appl., 366 (1), 24–32, 2010.
[3] S. Amat, M. A. Hernández and N. Romero, A modified Chebyshev’s itera-
tive method with at least sixth order of convergence, Appl. Math. Comput.,
206 (1), 164–174, 2008.
[4] I. K. Argyros, Convergence and Application of Newton-type Iterations,
Springer, 2008.
[5] I. K. Argyros and S. Hilout, Numerical methods in Nonlinear Analysis,
World Scientific Publ. Comp. New Jersey, 2013.
[6] I. K. Argyros and Á. A. Magreñán, On the convergence of an optimal
fourth-order family of methods and its dynamics, Appl. Math. Comput. 252,
336–346, 2015.
[7] I. K. Argyros and Á. A. Magreñán, Local Convergence and the Dynamics
of a Two-Step Newton-Like Method, Int. J. Bifurcation and Chaos, 26 (5),
1630012 2016.
[8] R. Behl, Development and analysis of some new iterative methods for nu-
merical solutions of nonlinear equations (PhD Thesis), Punjab University,
2013.
[9] D. D. Bruns and J. E. Bailey, Nonlinear feedback control for operating a
nonisothermal CSTR near an unstable steady state, Chem. Eng. Sci., 32,
257–264, 1977.
104
References 105
[10] V. Candela and A. Marquina, Recurrence relations for rational cubic meth-
ods I: The Halley method, Computing, 44, 169–184, 1990.
[11] V. Candela and A. Marquina, Recurrence relations for rational cubic meth-
ods II: The Chebyshev method, Computing, 45 (4), 355–367, 1990.
[12] F. Chicharro, A. Cordero and J. R. Torregrosa, Drawing dynamical and
parameters planes of iterative families and methods, The Scientific World
Journal, Volume 2013, Article ID 780153.
[13] C. Chun, Some improvements of Jarratt’s method with sixth-order conver-
gence, Appl. Math. Comput., 190 (2), 1432–1437, 1990.
[14] A. Cordero, J. Garcı́a-Maimó, J. R. Torregrosa, M. P. Vassileva and P. Vin-
del, Chaos in King’s iterative family, Appl. Math. Let., 26, 842–848, 2013.
[15] A. Cordero, J. M. Gutiérrez, Á. A., Magreñán and J. R. Torregrosa, Stability
analysis of a parametric family of iterative methods for solving nonlinear
models, Appl. Math. Comput., 285, 26–40, 2016.
[16] A. Cordero, L. Feng, L., Á. A., Magreñán and J. R. Torregrosa, A new
fourth-order family for solving nonlinear problems and its dynamics, J.
Math. Chemistry, 53 (3), 893–910, 2015.
[17] J. A. Ezquerro and M. A. Hernández, Recurrence relations for Chebyshev-
type methods, Appl. Math. Optim., 41(2), 227–236, 2000.
[18] J. A. Ezquerro and M. A. Hernández, New iterations of R-order four with
reduced computational cost, BIT Numer. Math., 49, 325–342, 2009.
[19] J. A. Ezquerro and M. A. Hernández, On the R-order of the Halley method,
J. Math. Anal. Appl., 303, 591–601, 2005.
[20] A. Fraile, E. Larrodé, Á. A. Magreñán and J. A. Sicilia, Decision model
for siting transport and logistic facilities in urban environments: A method-
ological approach. J. Comput. Appl. Math., 291, 478–487, 2016.
[21] J. M. Gutiérrez and M. A. Hernández, Recurrence relations for the super-
Halley method, Comput. Math. Appl., 36 (7), 1–8, 1998.
[22] M. Ganesh and M. C. Joshi, Numerical solvability of Hammerstein integral
equations of mixed type, IMA J. Numer. Anal., 11, 21–31, 1991.
[23] M. Garcı́a-Olivo, J. M. Gutiérrez and Á. A. Magreñán, A complex dynam-
ical approach of Chebyshev’s method,SeMA Journal, 71 (1), 57–68, 2015.
[24] Y. H. Geum, Y. I., Kim and Á. A., Magreñán, A biparametric extension
of King’s fourth-order methods and their dynamics, Appl. Math. Comput.,
282, 254–275, 2016.
106 References
Extending the
Kantorovich theory for
solving equations
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
6.2 First convergence improvement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
6.3 Second convergence improvement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
6.1 Introduction
Let X, Y be Banach spaces, D ⊂ X be convex, F : D ⊂ X → Y be a Fréchet
differentiable operator. We shall determine a solution x∗ of the equation
F(x) = 0, (6.1)
108
Extending the Kantorovich theory for solving equations 109
Many problems from Applied Sciences can be solved finding the solutions of
equations in a form like (6.1) [5, 9, 12, 13, 27, 31–33].
The most popular iterative method is undoubtedly Newton’s method defined
by
xn+1 = xn − F (xn )−1 F(xn ), for each n = 0, 1, 2, · · · , (6.2)
where x0 is an initial point. Newton’s method has order of convergence 2 if the
initial point x0 is chosen sufficiently close to the solution x .
The computation of the inverse may be expensive in general. That is why the
modified Newton’s method defined as
and
F (x0 )−1 F(x0 ) ≤ η
respectively.
In a series of papers [1–11, 14–17], we provided weaker sufficient conver-
gence criteria. One of the most recent and weakest criterion is given by
1
h1 = L1 η ≤ , (6.5)
2
where
1 √
L1 = 4L0 + L0 L + L2 + 8L0 L ,
8
where L0 is the center-Lispchitz constant given in
Notice that L0 ≤ L,
1 1
h≤ ⇒ h1 ≤ (6.6)
2 2
110 Iterative methods and their dynamics with applications: a contemporary study
and
h1 L0
→ 0 as → 0. (6.7)
h L
It is well known that in practice the computation of constant L requires that of
constant L0 . Throughout the chapter we denote by U(x, ρ) and Ū(x, ρ) the open
and closed balls with center x ∈ D and radius ρ > 0, respectively.
In this chapter we present two different improvements of criterion (6.5) based
on modified Newton’s method and Newton’s method in Sections 6.2 and 6.3
respectively.
where
L + 3L0 − L2 + 9L02 + 2L0 L
δ = 1−
L
and
η0 = min{η1 , η2 }.
If
η < η1
then
η < η2
then
L0 r0 2 − (L + 2L0 )r0
≤ .
2 2 − Lr0
Due to the choice of b, r and r0 we get
Lbk−1 r0 Lr0
≤ ≤ b for each k = 2, 3, · · · .
2(1 − L0 (1 + b )r0 ) 2(1 − L0 r)
k−1
unique solution x∗ located in the closed ball Ū(x0 , r) of equation F(x) = 0. More-
over the following estimates hold for each n = 0, 1, 2, · · ·
L
xn+1 − x∗ ≤ xn − x∗ 2 . (6.14)
2(1 − L0 xn − x0 )
Proof: It follows from condition (6.12) that
1
h0 = L0 η ≤ (6.15)
2
which implies r0 is well defined and equation F(x) = 0 has a unique solution
x ∈ Ū(x0 , r0 ).
By means of considering Newton’s method we get
and
Lxk −x∗
xk+1 − x∗ ≤ ∗
2(1−L0 xk −x0 ) xk − x
Lbk−1 r0 xk −x∗
≤ 2(1−L0 (xk −x∗ +x∗ −x0 ))
Lbk−1 r0 xk −x∗
≤ 2(1−L0 (bk−1 +1)r0 )
Lbk−1 r0 xk −x∗
≤ 2(1−L0 r0 )
< bxk − x∗
and
≤ bk x∗ − x0 + x∗ − x0
≤ (bk + 1)x∗ − x0
≤ (b + 1)x∗ − x0
1−b2
≤ 1−b r0 ≤ r0 .
As a consequence, xk+1 ∈ U(x0 , r) and lim xk+1 = x∗ .
k→∞
Remark 6.2.2 It follows from (6.16) that Newton’s method converges quadrat-
ically to x , r is given in closed form and (6.12) can be weaker than the Kan-
torovich hypothesis [26]
1
η≤
2L
for sufficiently small L0 .
where
xn+1 − x0
an+1 =
rn+1
and
2
xn+1 − xn
bn+1 = for each n = 0, 1, 2, · · ·
rn+1 − rn
is a tighter majorizing sequence than {tn }. Therefore, it may converge under
weaker convergence criteria than {tn }. As an example, let us consider there exists
a ∈ [0, 1] and b ∈ [0, 1] such that
s0 = 0, s1 = η,
0 (s1 −s0 )
2
s2 = s1 + L̄2(1− L̄0 s1 )
L̄(sn+1 −sn ) 2
sn+2 = sn+1 2(1− L̄0 sn+1 )
for each n = 0, 1, 2, · · ·
where
K = max{aL0 , bL}
and define function
K 2
f (t) = t −t +η
2
and sequence {un } by
Extending the Kantorovich theory for solving equations 115
u0 = 0,
f (un )
un+1 = un − f (un ) for each n = 0, 1, 2, · · ·
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
7.2 Standard results on convex functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
7.3 Semilocal convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
7.4 Special cases and applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
7.1 Introduction
The main task this chapter is to use the iterative methods to find solutions x∗ of
the equation
F(x) = 0, (7.1)
where D : D ⊂ X → Y is a Fréchet-differentiable operator X,Y are Banach spaces
and D ⊂ X. Many problems from Applied Sciences can be solved finding the
solutions of equations in a form like (7.1) [9, 12, 14, 17, 18, 23, 25, 28–34].
119
120 Iterative methods and their dynamics with applications: a contemporary study
The most well-known and used iterative method is Newton’s method due to
its simplicity and goodness
then
h(a)
a− ≤ b,
h (a)
with equality if and only if h is affine in [a, b).
(C1 ) f0 (0) = −1, f (0) > 0, f (0) = −1, f0 (t) ≤ f (t), f0 (t) ≤ f (t) for each
t ∈ [0, R);
122 Iterative methods and their dynamics with applications: a contemporary study
−( f (t) + 2ρ) κρ
κρ := sup (ρ)|(t − ρ) , λρ := sup{t ∈ [ρ, R) : κρ + f (t) < 0}, Θρ := 2 − κρ .
|
ρ<t<R 0 f
(7.6)
Then for any R ∈ [0, Θρ ] and z0 ∈ U(x0 , ρ), the sequence {zn } generated
by the inexact Newton method with initial point z0 and residual relative error
tolerance θ for k = 0, 1, . . . , zn+1 = zn + Sn ,
F (z0 )−1 [F(zn ) + F (zn )Sn ] ≤ θ F (z0 )−1 F(zn ), (7.7)
If, additionally, 0 ≤ θ < κρ/(4 + κρ), then {zn } converges Q-linearly as follows
1 + θ 2θ
x∗ − zn+1 ≤ κ + x∗ − zn , k = 0, 1, . . . .
2 κρ
Remark 7.3.2 (a) If f0 = f , then, Theorem 7.3.1 reduces to the corresponding
one [22, Theorem 2.1]. On the other hand, if f0 < f it constitutes a clear
1 1
improvement, since < . So our theorem is at least as good
| f0 (t)| | f (t)|
as the previous one.
Robust convergence for inexact Newton method 123
(b) If one choose θ = 0 in Theorem 7.3.1 we have the exact Newton method and
its convergence properties. If θ = θn in each iteration and letting θn goes
to zero as n goes to infinity, the penultimate inequality of the Theorem
7.3.1 implies that the generated sequence converges to the solution with
superlinear rate.
Here after we assume that the hypotheses of Theorem 7.3.1 hold. The scalar
function f in the above theorem is called a majorant function for F at point x0 ,
where as the function f0 is the center-majorant function for F at point x0 .
Next, we analyze some basic properties of functions f0 and f . Condition (C2 )
implies strict convexity of f and f0 . It is worth noticing that t∗ is the smallest root
of f (t) = 0 and, since f is convex, if the equation has another one, this second
root is τ̄.
Define
t¯ := sup{t ∈ [0, R) : f (t) < 0}. (7.9)
Proposition 7.3.3 (cf. [22]) The following statements on the majorant function
hold
i) f (t) < 0 for any t ∈ [0, t¯), (and f (t) ≥ 0 for any t ∈ [0, R) [0, t¯));
ii) 0 < t∗ < t¯ ≤ τ̄ ≤ R;
iii) β = − lim f (t), 0 < β < t¯.
t→t¯−
Now, we will prove a special case of Theorem 7.3.1. This case is ρ = 0 and
z0 = x0 . As in [22], in order to simplify the notation in the case ρ = 0, we will
use κ, λ and θ instead of κ0 , λ0 and θ0 respectively:
− f (t) κ
κ := sup , λ := sup{t ∈ [0, R) : κ + f (t) < 0}, Θ := . (7.10)
0<t<R t 2−κ
Firstly, we need a Banach-type result for invertible operators. We will use the
following one that appears in [24].
Proof: We have by (7.9) that f0 (t) < 0, since f0 (t) ≤ f (t). Using (7.3) and (C1 )
we get
F (x0 )−1 F (x) − I = F (x0 )−1 [F (x) − F (x0 )] ≤ f0 (x − x0 ) − f0 (0)
Using Banach Lemma [24] and the preceding inequality above we conclude that
F (x)−1 ∈ L(Y, X ) and
1 1
F (x)−1 F (x0 ) = (F (x0 )−1 F (x))−1 ≤ =− .
1 − ( f0 (t) + 1) f0 (t)
Notice that the corresponding result in [22] uses f instead of f0 to obtain
1
F (x)−1 F (x0 ) ≤ − .
f (t)
But
1 1
− ≤− .
f0 (t) f (t)
Therefore, if f0 (t) < f (t), then our result is more precise and less expensive to
arrive at. Furthermore, this modification influences error bounds and the suffi-
cient convergence criteria.
The linearization errors on F and f are given respectively by
e f (a + b, b) ≤ e f (t + s,t),
1 f (t + s) − f (t) 2
e f (a + b, b) ≤ a , s = 0.
2 s
Next, we bound the linearization error EF using the linearization error on the
majorant function.
Robust convergence for inexact Newton method 125
From the first inequality showed above this corollary proves that F has no zeroes
in the region t∗ < x − x0 < τ̄.
then
1 + θ 2θ
x∗ − y ≤ + x∗ − x, (7.17)
2 κ
1 + θ D− f (λ ) 2 + f0 (λ )
x∗ − y ≤ x ∗ − x + θ x∗ − x. (7.18)
2 | f0 (λ )| | f0 (λ )|
and
−F (x0 )−1 F(x) = F (x0 )−1 [EF (x∗ , x) + F (x)(x∗ − x)] .
126 Iterative methods and their dynamics with applications: a contemporary study
and
F (x0 )−1 F(x) ≤ F (x0 )−1 EF (x∗ , x) + F (x0 )−1 F (x)x∗ − x.
Combining two above inequalities with Proposition 7.3.5, Corollary 7.3.8 with
y = x∗ and s = λ − t and Lemma 7.3.10 we have that
1 1 + θ f (λ ) − f (t)
x∗ − y ≤ x∗ − x + θ [2 + f0 (t)] x∗ − x.
| f0 (t)| 2 λ −t
Since x∗ − x ≤ λ − t, f < −κ < 0 in [0, λ ) and f0 , f are increasing the first
inequality follows from last inequality. Using Proposition 7.2.1 and taking in
account that f0 ≤ f < 0 in [0, λ ) and increasing we obtain the second inequality
from above inequality.
In the following result we present a single inexact Newton iteration with rel-
ative error θ .
If y ∈ X and
Proof: From 7.3.5 and the first inequality in (7.19) we conclude that F (x)−1 ∈
L(Y, X ) and F (x)−1 F (x0 ) ≤ −1/ f0 (t). Therefore, using also triangular in-
equality, (7.20) and the identity
y − x = F (x)−1 F (x0 ) F (x0 )−1 [F(x) + F (x)(y − x)] − F (x0 )−1 F(x) ,
Robust convergence for inexact Newton method 127
we conclude that
−1
y − x ≤ (1 + θ )F (x0 )−1 F(x).
f0 (t)
To end the proof of the first item, use the second inequality on (7.19) and the
above inequality. The second item follows from the first one, the first and the
third inequalities in (7.19) and triangular inequality. Using the definition of the
error (7.13) we have that
F(y) = EF (y, x) + F (x0 ) F (x0 )−1 [F(x) + F (x)(y − x)] .
Hence, using (7.20), the second inequality on (7.19) and the triangle inequality
we have that
F (x0 )−1 F(y) ≤ F (x0 )−1 EF (y, x) + θ F (x0 )−1 F(x)
Using the first item, (7.19) and Lemma 7.3.6 with s = −(1 + θ )( f (t) + ε)/ f0 (t)
we get
−1 f (t) + ε
F (x0 ) EF (y, x) ≤ e f t − (1 + θ ) ,t
f0 (t)
f (t) + ε
= f t − (1 + θ ) + ε + θ ( f (t) + ε).
f0 (t)
Using the two above equation we obtain the latter inequality in the third item.
In view of Lemma 7.3.12, define, for θ = 0, the auxiliary map nθ : [0, t¯) ×
[0, 8) → R × R,
f (t) + ε
nθ (t, ε) := t − (1 + θ ) , ε + 2θ ( f (t) + ε) . (7.21)
f0 (t)
Let
Proof: Since 0 ≤ t < λ , according to (7.10) we have f (t) < −κ < 0. Therefore
t < t+ and ε ≤ ε+ .
As ε ≤ κt, f (t) + ε > 0 and −1 ≤ f (t) < f (t) + κ < 0,
f (t) + ε f (t) + κt
− ≤ −
f0 (t) f0 (t)
(7.23)
f (t) + κt κ f (t) + κt
= − 1+ ≤− (1 − κ)
f0 (t) + κ f0 (t) f0 (t) + κ
The function h(s) := f (s) + κs is differentiable at the point t, h (t) < 0, is strictly
convex and
lim h(t) = 0.
s→λ −
Hence, using Proposition 7.2.2, we have t − h(t)/h (t) < λ , which is equivalent
to
f (t) + κt
− < λ − t. (7.24)
f0 (t) + κ
Using the above inequality with (7.23) and the definition of t+ we conclude that
ε+ ≤ 2θ ( f (t) + ε) + κt
Taking into account the inequalities f (t) + ε > 0 and −1 ≤ f (t) < 0, we have
that
f (t) + ε
f (t) + ε ≤ − .
f0 (t)
Using the two above inequalities with the definition of t+ we obtain ε+ ≤ κt+ .
In order to prove the two last inequalities, note that using the definition (7.14),
we get
Taking into account that f > 0 for every point we conclude that the integral
Robust convergence for inexact Newton method 129
is positive. So, last equality implies that f (t+ ) + ε+ = θ ( f (t) + ε) > 0. Taking
s ∈ [t+ , λ ) and using the convexity of f , we have that
t+ t+ u−t
( f (u) − f (t)) du ≤ ( f (s) − f (u)) du
t t s−t
1 (t+ − t)2
= ( f (s) − f (t)) .
2 s−t
On the other hand, due to f (s) + κ < 0 and −1 ≤ f (t) we can conclude that
f (s) − f (t) f (s) + κ − f (t) − κ
= ≤ 1 − κ.
− f0 (t) − f0 (t)
Using (7.23), (7.24) and the last two above inequalities and taking in account that
(1 + θ )(1 − κ) ≤ 1 − θ we conclude that
1 λ −t
f (t+ ) + ε+ ≤ θ + (1 + θ )2 (1 − κ)2 ( f (t) + ε)
2 s−t
1 λ −t
= θ + (1 − θ )2 ( f (t) + ε),
2 s−t
Definition 7.3.14 (cf. [22]) For 0 ≤ θ , Nθ is the family of maps Nθ : U(x0 , t¯) →
X such that
F (x0 )−1 [F(x) + F (x)(Nθ (x) − x)] ≤ θ F (x0 )−1 F(x), (7.25)
If x ∈ U(x0 , t¯), then F (x)−1 ∈ L(Y, X ). Hence, for θ = 0, N0 has a single ele-
ment, namely the exact Newton iteration map given by
Where nθ and Nθ were defined in (7.21), (7.22) and Definition 3.14 respectively.
Proposition 7.3.16 (cf. [22]) Take 0 ≤ θ ≤ Θ and Nθ ∈ Nθ . Then for any (t, ε) ∈
Ω and x ∈ K(t, ε)
Nθ (K(t, ε)) ⊂ K(nθ (t, ε)) ⊂ K, Nθ (x) − x ≤ t+ − t,
where t+ is the first component of nθ (t, ε). Moreover,
nθ (Ω) ⊂ Ω, Nθ (K) ⊂ K. (7.28)
Proof: Using definitions (7.21), (7.22), Definition 3.14, (7.26), (7.27) with Lem-
mas 7.3.12 and 7.3.13.
Theorem 7.3.17 Take 0 ≤ θ ≤ Ω and Nθ ∈ Nθ . For any (t0 , ε0 ) ∈ Ω and y0 ∈
K(t0 , ε0 ), the sequences
yn+1 = Nθ (yn ), (tn+1 , εn+1 ) = nθ (tn , εn ), k = 0, 1, . . . , (7.29)
are well defined,
yn ∈ K(tn , εn ), (tn , εn ) ∈ Ω, k = 0, 1, . . . , (7.30)
the sequence {tn } is strictly increasing and converges to some (0, λ ], the se-
quence {εn } is non-decreasing and converges to some ε ∈ [0, κλ ],
n
−1 1+θ2
F (x0 ) F(yn ) ≤ f (tn ) + εn ≤ ( f (t0 ) + ε0 ), k = 0, 1, . . . ,
2
(7.31)
the sequence {yn } is contained in U(x0 , λ ) and converges to a point x∗ ∈ Ū(x0 ,t∗ )
which is the unique zero of F in U(x0 , τ̄) and
yn+1 − yn ≤ tn+1 − tn , x∗ − yn ≤
t − tn , k = 0, 1, . . . . (7.32)
Moreover, if the following condition holds
Robust convergence for inexact Newton method 131
(C4 ) λ < R,
then the sequence {yn } satisfies, for k = 0, 1, . . .
1 + θ D− f (λ ) 2 + f0 (λ )
x∗ − yn+1 ≤ + x∗ − yn + θ x∗ − yn . (7.33)
2 | f0 (λ )| | f0 (λ )|
Hence, if 0 ≤ θ < κ/(4 + κ), then {yn } converges Q-linearly as follows
1 + θ 2θ
x∗ − yn+1 ≤ + x∗ − yn , k = 0, 1, . . . . (7.34)
2 κ
Proof: From the assumptions on θ , (t0 , ε0 ), y0 and the two last inclusions on
Proposition 7.3.16 it is clear that the sequences {(tn , εn )} and {yn } as defined in
(7.29) are well defined. Furthermore, since (7.30) holds for k = 0, using the first
inclusion in Proposition 7.3.5 and an inductive argument on k, we conclude that
(7.30) holds for all k. From Proposition 7.3.5, (7.30) and (7.29) while the first
inequality in (7.31) follows from (7.30) and the definition of K(t, ε) in (7.26)
follows the first inequality in (7.32). Using the definition of in (7.22) shows that
Ω ⊂ [0, λ ) × [0, κλ ).
Hence, using (7.30) and the definition of K(t, ε), we have tn ∈ [0, λ ), εn ∈ [0, κλ ),
yn ∈ U(x0 , λ ), k = 0, 1, . . . .
Using Lemma 7.3.13 and (7.22) we conclude that {tn } is strictly increasing,
{εn } is non-decreasing and the second equality in (7.31) holds for all k. Hence,
in view of the first two above inclusions, {tn } and {εn } converge, respectively, to
some ε ∈ [0, κλ ]. Convergence to
t ∈ (0, λ ] and t , together with the first inequal-
ity in (7.32) and the inclusion yn ∈ U(x0 , λ ) implies that yn converges to some
x∗ ∈ Ū(0, λ ) and that the second inequality on (7.32) holds for all k.
Using the inclusion yn ∈ U(x0 , λ ), the first inequality in Corollary 7.3.9 and
(7.31), we have that
n
1+θ2
− f (yn − x0 ) ≤ ( f (t0 ) + ε0 ), k = 0, 1, . . . .
2
According to (7.12), f < −κ in [0, λ ). Therefore, since f (t∗ ) = 0 and t∗ < λ ,
f (t) ≤ −κ(t − t∗ ), t∗ ≤ t < λ .
Hence, if yn − x0 = t∗ , we can combine the two above inequalities, setting
t = yn − x0 in the second, to obtain
n
1+θ2 f (t0 ) + ε0
yn − x0 − t∗ ≤ .
2 κ
Note that the above inequality remain valid even if yn − x0 < t∗ . Thus, tak-
ing the limit k → ∞ in the above inequality we conclude that x∗ − x0 ≤ t∗ .
132 Iterative methods and their dynamics with applications: a contemporary study
Proof: Assumption ρ < β /2 and Proposition 7.3.3-item (iii) proves the first two
inequalities of the proposition. The last inequality follows from the first inequal-
ity and Proposition 7.3.3-item (i).
Now, using all the preceding results we can prove Theorem 7.3.1.
Proof: (Proof of Theorem 7.3.1) We will begin with the case ρ = 0 and z0 = x0 .
Note that, from the definition in (7.10), we have that
κ0 = κ, λ0 = λ , Θ0 = Θ.
Since
(0, 0) ∈ Ω, x0 ∈ K(0, 0),
using Theorem 7.3.17 we conclude that Theorem 7.3.1 holds for ρ = 0. For prov-
ing the general case, take
Using Proposition 7.3.18 and (7.9) we conclude that ρ < t¯/2 and f (ρ) < 0.
Define
−1
g : [0, R − ρ) → R, g(t) = [ f (t + ρ) + 2ρ]. (7.36)
f0 (ρ)
We claim that g is a majorant function for F at point z0 . Trivially, U(z0 , R − ρ) ⊂
D, g (0) = −1, g(0) > 0. Moreover g is also convex and strictly increasing. To
end the proof that g satisfies C1 , C2 and C3 , using Proposition 7.3.3-item (iii) and
second inequality in (7.35), we have that
−1
lim g(t) = (2ρ − β ) < 0.
t→t¯−ρ f0 (ρ)
Robust convergence for inexact Newton method 133
F (z0 )−1 F(z0 ) ≤ F (z0 )−1 F (x0 )F (x0 )−1 F(z0 )
−1
≤ [ f (z0 − x0 ) + 2z0 − x0 ] .
f0 (ρ)
To end the proof that g is a majorant function for F at z0 , take x, y ∈ X such that
F (z0 )−1 F (y) − F (x) ≤ F (z0 )−1 F (x0 ) F (x0 )−1 F (y) − F (x)
−1
≤ f (y − x + x − x0 ) − f (x − x0 ) .
f0 (ρ)
which are the same as (7.5) taking g instead of f . Hence, applying Theorem
7.3.1 for F and the majorant function g at point z0 and ρ = 0, we conclude that
the sequence zn is well defined, remains in U(z0 , λρ ), satisfies (7.8) and converges
to some z∗ ∈ Ū(z0 ,t ∗ , ρ) which is a zero of F, where t∗,ρ is the smallest solution
of g(t) = 0. Using (7.36) we conclude that t∗,ρ is the smallest solution of
f (ρ + t) + 2ρ = 0.
134 Iterative methods and their dynamics with applications: a contemporary study
√ √
1 + α − 1 − 6α + α 2 1 + α + 1 − 6α + α 2
t∗ = , τ̄ = .
4 4
On the other hand, the sequence {xn } satisfies, for k = 0, 1, . . . ,
1 + θ D− f (λ ) f0 (λ ) + 2
x∗ − xk+1 ≤ x ∗ − xn + θ x∗ − xn .
2 | f0 (λ )| | f0 (λ )|
√ √
If, additionally, 0 ≤ θ < (1 − 2 α − α)/(5 − 2 α − α), then {xn } converges
Q-linearly as follows
1+θ 2θ
x∗ − xn+1 ≤ + √ x∗ − xn , k = 0, 1, . . . .
2 1−2 α −α
Proof: The function f : [0, 1/γ) → R
t
f (t) = − 2t + β ,
1 − γt
is a majorant function for F in x0 (cf. [20]). Therefore, all results follow from
Theorem 7.3.1, applied to this particular context by choosing f0 : [0, 1/γ) → R
such that
1
f0 (t) = −2 + .
(1 − γ0t)2
Notice that if F is analytic on U(x0 , 1/γ) we can choose
1
F (x0 )−1 F (k)(x0 ) k−1
γ0 = sup
,
k>1 k!
(cf. [1, 11, 12, 16, 37]). A semilocal convergence result for Newton method is in-
strumental in the complexity analysis of linear and quadratic minimization prob-
lems by means of self-concordant functions (cf. [23, 28]). Also in this setting,
Theorem 7.3.1 provides a semilocal convergence result for Newton method with
a relative error tolerance.
Then, the sequence generated by the inexact Newton method for solving F(x) = 0
with starting point x0 and residual relative error tolerance θ : For k = 0, 1, . . . ,
xn+1 = xn + Sn , F (x0 )−1 [F(xn ) + F (xn )Sn ] ≤ θ F (x0 )−1 F(xn ),
is well defined,
n
−1 1+θ2
F (x0 ) F(xn ) ≤ β , k = 0, 1, . . . .
2
L̄xn+1 − xn 2
xn+1 − xn ≤ for each n = 0, 1, 2, . . . ,
2 (1 − L0 xn+1 − x0 )
where
L0 if n=1
L̄ =
L if k>1
L0 (r1 − r0 )2
r0 = 0, r1 = β , r2 = r1 + ,
2(1 − L0 r1 )
L(rn − rn−1 )2
rn+1 = rn + for each n = 2, 3, . . . .
2(1 − L0 rn )
Then, we have that (cf. [11])
It was shown in [11] that {rn } is a majorizing sequence for {xn } and con-
verges provided that
1
H1 = L1 β ≤ ,
2
where
1 √
L1 = 4L0 + L0 L + L0 L + 8L0 . 2
8
The corresponding results deduced from Theorem 7.3.1 (with f0 = f ) are
given by
Lxn+1 − xn 2
xn+1 − xn ≤ for each n = 0, 1, 2, . . .
2(1 − L0 xn+1 − x0 )
and the corresponding majorizing sequence {vn } is defined by
rn ≤ vn
rn+1 − rn ≤ vn+1 − vn
and
r∗ = lim rn ≤ v∗ = lim vn .
k→∞ k→∞
The inequality in the first two preceding inequalities is strict if L0 < L for
each n = 2, 3, . . .. Finally, notice that
h1 L0
→ 0 as → 0.
h L
The preceding estimate shows by how many times at most the applicability
of Newton’s method is expanded with our approach. Examples where L0 <
L and where the old convergence criteria are not satisfied but the new
convergence criteria are applied can be found for example in [11].
Notice that the Lipschitz conditions in the last two theorems imply the
center-Lipschitz conditions with γ0 ≤ γ and L0 ≤ L. That is the center-
Lipschitz conditions are not additional hypotheses to the Lipschitz condi-
tions.
(c) Another important consideration is that in case of modified inexact Newton
method defined by
x̄n+1 = x̄n + Sn , F (x0 )−1 [F(x̄n ) + F (x0 )Sn ] ≤ θ F (x0 )−1 F(xn )
is used, then f0 can replace f in all preceding results. The resulting con-
vergence criteria are weaker but the rate of convergence is only linear.
However, we can start with the slower method until a certain finite step
N ≥ 1 such that the convergence criteria of Theorem 7.3.1 hold for x0 = x̄N .
Then, we can continue with the faster method. Such an approach has been
given by us in [11–13] for Newton’s method (7.2). Notice that the ap-
proach in [22] shall use f instead of f0 . That means, e.g., in the case of
Newton’s method the sufficient convergence criterion shall be
1
H = Lβ ≤
2
for the modified Newton’s method whereas in our case it will be
1
H0 = L0 β ≤ .
2
Robust convergence for inexact Newton method 139
140
References 141
Inexact
Gauss-Newton-like
method for least square
problems
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
8.2 Auxiliary results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
8.3 Local convergence analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
8.4 Applications and examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
144
Inexact Gauss-Newton-like method for least square problems 145
8.1 Introduction
In this chapter we are interested in locating a solution x∗ of the nonlinear least
squares problem:
1
min G(x) := F(x)T F(x), (8.1)
2
where F is Fréchet-differentiable defined on Rn with values in Rm , m ≥ n.
In Computational Sciences the practice of numerical analysis is essentially
connected to variants of Newton’s method [1]– [15]. In the present chapter we
consider the local convergence of inexact Gauss-Newton method (IGNM) for
solving problem (8.1).
In particular, we define the method as follows [5, 8, 10, 11]:
Let U(x, r) and U(x, r) denote, respectively, for the open and closed ball in Rn
with center x and radius r > 0. Denote by L(Rn , Rm ) the space of bounded linear
operators from Rn into Rm . For k = 0 step 1 until convergence do.
Find the step Δk which satisfies
rk
Bk Δk = −F (xk )T F(xk ) + rk , where ≤ ηk . (8.2)
F (xk )T F(xk )
Set xk+1 = xk + Δk .
Here, Bk is an n × n nonsingular matrix, and {ηk } is a sequence of forcing
terms such that ηk ∈ [0, 1] for each k.
The process is called inexact Gauss-Newton method (IGNM) if Bk =
F (xk )T F(xk ) and it is called inexact Gauss-Newton-like method (IGNLM) if
Bk = B(xk ) approximates F (xk )T F(xk ). There is a plethora of convergence re-
sults for (IGNM) and (IGNLM) under various Lipschitz-type conditions [1]–
[15]. Recently, studies have been focused on the analysis of stopping residual
rk
controls F(x k )
≤ ηk and its effect on convergence properties [3, 5–8, 10, 11] by
considering iterative form where a scaled residual control is performed at each
iteration as follows:
For k = 0 step 1 until convergence do.
Find the step Δk which satisfies
Pk rk
Bk Δk = −F(xk ) + rk , where ≤ θk . (8.3)
Pk F(xk )
Set xk+1 = xk + Δk .
Here, Pk−1 ∈ L(Rm , Rn ) for each k.
In this chapter, we are motivated by different studies [5–7] and optimization
considerations. We show our advantages by considering a combination of weak
Lipschitz and weak center Lipschitz conditions.
The chapter is organized as follows. In Section 8.2, we present some auxil-
iary results on Moore-Penrose inverses. The local convergence of (IGNM) and
(IGNLM) is presented in Section 8.3. Numerical examples are presented in Sec-
tion 8.4.
146 Iterative methods and their dynamics with applications: a contemporary study
A†
B† ≤ . (8.4)
1 − A† E|
Now, we present results involving weak Lipschitz condition ( [15]) in the case
F (x)† = F (x)−1 .
Lemma 8.2.3 Suppose that F is continuously Fréchet-differentiable in
U(x , r), F(x ) = 0, F (x ) has full rank:
(i) If F satisfies the center-Lipschitz condition with L0 average:
θ s(x)
θ
F (x ) − F (x ) ≤ L0 (u)du for each x ∈ U(x , r), 0 ≤ θ ≤ 1,
0
(8.6)
where xθ = x + θ (x − x ), s(x) = x − x and L0 is a positive integrable
function. Then, for β = F (x )† , if
s(x)
β L0 (u)du < 1, (8.7)
0
1 s(x)
F (xθ ) − F (x )s(x)dθ ≤ L0 (u)(s(x) − u)du. (8.8)
0 0
and s(x)
† x − x + β L0 (u)(s(x) − u)du
F (y) F(x) ≤ s(y)
0
(8.9)
1 − β 0 L0 (u)du
for each x and y ∈ U(x , r).
Inexact Gauss-Newton-like method for least square problems 147
Proof: (i) Simply use L0 instead of L in the proof of (ii) in Lemma 2.3 [5, Page
101].
(ii) Using (8.10), we get
1 1 s(x)
F (x) − F (xθ )s(x)dθ ≤ L(u)dus(x)dθ
0 0 θ s(x)
s(x)
= L(u)udu,
0
which shows (8.11). Moreover, since F (x ) has full rank, it follows from Lem-
mas 8.2.1, 8.2.2, 8.2.3 and
s(x)
†
F (x ) F (x) − F (x ) ≤ β L0 (u)du < 1 for each x ∈ U(x , r) (8.13)
0
that F (x) has full rank and
β
[F (x)T F (x)]−1 F (x)T ≤ s(x) for each x ∈ U(x , r). (8.14)
1−β 0
L0 (u)du
Notice that we have the estimates
F (y)† F (xθ ) = I − F (y)† (F (y) − F (xθ )) (8.15)
and
F (y)† F(x) = F (y)† (F(x) − F(x ))
1
= F (y)† F (xθ )dθ (x − x ). (8.16)
0
148 Iterative methods and their dynamics with applications: a contemporary study
Proof are omitted since are analogous to the corresponding ones in [5], where
we replace old estimates
s(x)
β L(u)du < 1, (8.20)
0
1 s(x)
θ
F (x ) − F (x )s(x)dθ ≤ L(u)(s(x) − u)du, (8.21)
0 0
s(x)
† x − x + β L(u)(s(x) − u)du
F (y) F(x) ≤ s(y)
0
(8.22)
1 − β 0 L(u)du
and s(x)
† β L(u)du
F(x) F(x) ≤ s(x0 + s(x)
0
(8.23)
1 − β 0 L(u)du
by the new and more precise (8.7), (8.8), (8.9) and (8.12), respectively.
Next, we present the local convergence results. The first one involving
(IGNM).
Theorem 8.3.1 Suppose x satisfies (8.1), F has a continuous derivative in
U(x , r),
F (x ) has full rank and F (x) satisfies the radius Lipschitz condition with
L average and the center-Lipschitz condition with L0 average where L and
L0 are nondecreasing. Assume Bk = F (xk )T F (xk ), for each k in (8.2), vk =
θk (Pk F (xk )T F (xk ))−1
Pk F (xk )T F (xk ) = θkCond(Pk F (xk )T F (xk )) with vk ≤ v < 1. Let r > 0 satisfy
r √ r
β 0 L(u)udu 2cβ 2 0 L0 (u)du
(1 + v) r + r + v ≤ 1. (8.24)
r(1 − β 0 L0 (u)du) r(1 − β 0 L0 (u)du)
and
s(x0 )
[F (x )T F (x )]−1 F (x )T F (x) − F (x ) ≤ β L0 (u)du
0
r
≤ β L0 (u)du < 1,
0
for each x ∈ U(x , r).
β
[F (x)T F (x)]−1 F (x)T ≤ s(x0 ) ,
1−β 0
L0 (u)du
for each x ∈ U(x , r),
√ s(x)
T −1 T T −1 T 2β 2 0 L0 (u)du
[F (x) F (x)] F (x) − [F (x ) F (x )] F (x ) ≤ s(x) ,
1−β 0 L0 (u)du
for each x ∈ U(x , r).
s(xk )
β L(u)du
xk+1 − x ≤ (1 + vk ) 0
s(xk ) s(xk )2
s(xk )2 (1 − β 0
L0 (u)du)
√ s(x )
2cβ 2 0 k L0 (u)du
+vk s(xk ) + s(x )
1 − β 0 k L0 (u)du
s(x )
β 0 k L(u)du
≤ (1 + v) s(xk ) s(xk )2
s(xk ) (1 − β 0 L0 (u)du)
2
√ s(x )
2cβ 2 0 k L0 (u)du
+ v+ s(x ) s(xk ).
1 − β 0 k L0 (u)du
2
is less than 1.
vk = θk (Pk F (xk )T F (xk ))−1 Pk F (xk )T F (xk ) = θkCond(Pk F (xk )T F (xk ))
with vk ≤ v < 1. Let r > 0 satisfy
r √ 2 r
β ω1 0 L(u)udu 2β ω1 0 L0 (u)du
(1 + v) r + ω2 + ω1 v + r ≤ 1. (8.30)
r(1 − β 0 L0 (u)du) r(1 − β 0 L0 (u)du))
Then (IGNLM) is convergent for all x0 ∈ U(x , r) and
s(x )
β ω1 0 0 L(u)udu
xk+1 − x ≤ (1 + v) s(x0 ) xk − x 2
s(x0 ) (1 − β 0 L0 (u)du)
2
√ 2 s(x0 )
2β ω1 0 L0 (u)du
+ ω2 + ω1 v + s(x ) xk − x
s(x0 )(1 − β 0 0 L0 (u)du)
(8.31)
where c = F(x ), β = [F (x )T F (x )]−1 F (x )T ,
s(x )
β ω1 0 0 L(u)udu
q = (1 + v) s(x )
s(x0 )(1 − β 0 0 L0 (u)du)
√ 2 s(x0 )
2β ω1 0 L0 (u)du
+ω2 + ω1 v + s(x ) (8.32)
s(x0 )(1 − β 0 0 L0 (u)du)
is less than 1.
Proof: Let x0 ∈ U(x , r), where r satisfies (8.30), by the monotonicity of L0 , L,
(8.17) and Lemma (8.2.5), we have
s(x )
β ω1 0 0 L(u)udu
q = (1 + v) s(x ) s(x0 )
s(x0 )2 (1 − β 0 0 L0 (u)du)
√ 2 s(x0 )
2β ω1 0 L0 (u)du
+ω2 + ω1 v + s(x )
s(x0 )(1 − β 0 0 L0 (u)du)
r
β ω1 0 L(u)udu
< (1 + v) r) r
r2 (1 − β 0 L0 (u)du)
√ 2 r
2β ω1 0 L0 (u)du
+ω2 + ω1 v + r ≤ 1,
r(1 − β 0 L0 (u)du)
and
s(x)
[F (x)T F (x)]−1 F (x)T F (x) − F (x ) ≤ β L0 (u)du
0
r
≤ β L0 (u)du < 1,
0
for each x ∈ U(x , r).
154 Iterative methods and their dynamics with applications: a contemporary study
β
[F (x)T F (x)]−1 F (x)T ≤ s(x) , for each x ∈ U(x , r)
1−β 0
L0 (u)du
√ s(x)
−1 T −1 T 2β 2 L0 (u)du
[F (x) F (x)]T
F (x) − [F (x ) F (x )]
T
F (x ) ≤ 0
s(x) ,
1 − β 0 L0 (u)du
for each x ∈ U(x , r).
xk+1 − x = xk − x − B−1 −1
k F (xk ) (F(xk ) − F(x )) + Bk rk
T
1
= xk − x − B−1 T θ
k F (xk ) F (x )(xk − x )dθ
0
+B−1 −1 −1 T −1 T
k Pk Pk rk + Bk F (xk ) F(xk ){[F (x ) F (x )] F (x ) F(x )
T
where β = [F (x )T F (x )]−1 F (x )T . Then (MIGNM) is convergent for all
x0 ∈ U(x , r),
L0 β x0 − x + v L0 β x0 − x (1 + v)
q= + < 1, (8.38)
1 − L0 β x0 − x 2(1 − L0 β x0 − x )
and the inequality (8.23) holds.
where c = F(x ), β = [F (x )T F (x )]−1 F (x )T . Then (IGNLM) is conver-
gent for all x0 ∈ U(x , r),
√
Lβ ω1 (1 + v)x0 − x + 2 2cL0 β 2 ω1
q = ω1 v + ω2 + < 1, (8.40)
2(1 − L0 β x0 − x )
and the inequality (8.31) holds.
Remark 8.4.4 (a) If, L0 = L the results reduce to the corresponding ones
in [5]. On the other hand, if L0 < L the new results constitute an improve-
ment.
(b) The results of Section 5 in [5–7] using only center-Lipschitz condition can
be improved with L0 replacing L.
Notice that we have F(x ) = 0, F (x ) = F (x )−1 = diag {1, 1, 1} and L0 =
e − 1 < L = e.
References
Lavrentiev Regularization
methods for Ill-posed
equations
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
9.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
9.2 Basic assumptions and some preliminary results . . . . . . . . . . . . . . . . . . . . 163
9.3 Error estimates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165
9.3.1 Apriori parameter choice . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166
9.3.2 Aposteriori parameter choice . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166
9.4 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
161
162 Iterative methods and their dynamics with applications: a contemporary study
9.1 Introduction
In this chapter, we consider the problem of approximately solving the nonlin-
ear ill-posed operator equation of the form
F(x) = y, (9.1)
Assumption 10 (1) There exists r > 0 such that U(x0 , r) ⊆ D(F) and F :
U(x0 , r) → X is Fréchet differentiable.
(2) There exists K0 > 0 such that, for all uθ = u + θ (x0 − u) ∈ U(x0 , r), θ ∈
[0, 1] and v ∈ X, there exists an element, say φ (x0 , uθ , v) ∈ X, satisfying
Assumption 11 There exists a constant K > 0 such that, for all x, y ∈ U(x̂, r) and
v ∈ X, there exists an element denoted by P(x, u, v) ∈ X satisfying
Note that condition (9.6) is suitable for both mildly and severely ill-posed prob-
lems [16], [17]. Further note that the source condition (9.6) involves the known
initial approximation x0 whereas the source condition considered in [15] requires
the knowledge of the unknown x̂.
We need some auxiliary results based on Assumption 10.
164 Iterative methods and their dynamics with applications: a contemporary study
Proof: From the fundamental theorem of integration, for any u ∈ U(x0 , r), we
get
1
F(x̂) − F(u) = F (u + t(x̂ − u))(x̂ − u)dt.
0
Hence, by Assumption 10,
F(x̂) − F(u) − F (u)(x̂ − u)
1
= [F (u + t(x̂ − u)) − F (x0 ) + F (x0 ) − F (u)](x̂ − u)dt
0
1
= F (x0 )[φ (x0 , u + t(x̂ − u), u − x̂) − φ (x0 , u, x̂ − u)]dt.
0
Then, by (2), (3) in Assumptions 10 and the inequality (F (u)+αI)−1 F (uθ ) ≤
1, we get
(F (u) + αI)−1 [F(x̂) − F(u) − F (u)(x̂ − u)
1
≤ φ (x0 , u + t(x̂ − u), u − x̂) + φ (x0 , u, x̂ − u)dt.
0
1
≤ [ K0 (u − x0 + x̂ − ut)dt + K0 u − x0 ]x̂ − u
0
1
≤ [ K0 (u − x̂ + x̂ − x0 + x̂ − ut)dt + K0 (u − x̂ + x̂ − x0 )]x̂ − u
0
5K0
≤ x̂ − u2 + 2K0 x̂ − x0 x̂ − u.
2
Theorem 9.2.2 ( [16, Theorem 2.2], also see [22]) Let xα be the solution of (9.4)
with y in place of yδ . Then
(1) xαδ − xα ≤ αδ ,
(2) xα − x̂ ≤ x0 − x̂,
(3) F(xαδ ) − F(xα ) ≤ δ .
Remark 9.2.3 From Theorem 9.2.2 and triangle inequality we have
δ
xαδ − x̂ ≤ + x0 − x̂.
α
Lavrentiev Regularization methods for Ill-posed equations 165
So (9.5) is meaningful if
δ
r> + x0 − x̂.
α
The result now follows from (9.7), (9.8), (9.9) and (9.10).
The following Theorem is a consequence of Theorem 9.2.2 and Theorem
9.3.1.
Theorem 9.3.2 Under the assumptions of Theorem 9.3.1
δ
xαδ − x̂ ≤ C( + ϕ(α))
α
where C is as in Theorem 9.3.1.
Lemma 9.3.5 Let Assumption 10 and assumptions in Proposition 9.3.4 hold and
α := α(δ ) is chosen according to (9.13). Then
(c − 2)δ
α(A0 + αI)−1 (F(xα ) − y) ≥ (9.14)
1 + k1
K0 (2c−1)x0 −x̂
where k1 = c−1 .
Let
aα := α(A0 + αI)−1 (F(xα ) − y).
Then
Lemma 9.3.6 Let the assumptions of Theorem 9.3.1 and Proposition 9.3.4 hold.
Then
168 Iterative methods and their dynamics with applications: a contemporary study
So
1
F(xα ) − F(x̂) = F (x̂ + t(xα − x̂))(xα − x̂)dt
0
1
= A0 (xα − x̂) − F (x̂ + t(xα − x̂))ϕ(x0 , x̂ + t(xα − x̂), xα − x̂)dt.
0
Hence
(c − 2)δ
ψ(ϕ(α)) = αϕ(α) ≥ := ξ δ
1 + k1
(c−2)
where ξ = 1+k1 . Thus α ≥ ϕ −1 (ψ −1 (ξ δ )). This completes the proof.
Theorem 9.3.7 Let Assumption 10 be satisfied and 4K0 x0 − x̂ < 1. Then for
0 < α0 ≤ α,
Proof: Since
F(xα ) − y + α(xα − x0 ) = 0, (9.15)
α − α0
xα −xα0 = (Aα +α0 I)−1 (F(xα )−y)+(Aα +α0 I)−1 [F(xα0 )−F(xα )+Aα (xα −xα0 )].
α
(F(xα ) − y).
Lavrentiev Regularization methods for Ill-posed equations 171
Theorem 9.3.9 Let assumptions of Lemma 9.3.6 hold. If, in addition, 9K0 x0 −
x̂ < 2, then
xαδ − x̂ ≤ ℘ψ −1 (ηδ )
where ℘ = 1
ξ + 1−4K01x0 −x̂ +C and η = max{ξ , β }.
Proof: Firstly, we begin with the case α := α(δ ) ≤ α0 . Then by Theorem 9.3.2
we get
δ
xαδ − x̂ ≤ C( + ϕ(α0 )). (9.21)
α
Now, we follow with the case α := α(δ ) > α0 . In this case by Theorem 9.3.7,
we obtain
δ
xαδ − x̂ ≤ +Cϕ(α0 )
α
α(Aα + α0 I)−1 (F(xα ) − y)
+
α0 (1 − 4K0 x − x̂)
δ βδ
≤ + +Cψ −1 (β δ ).
ϕ −1 ψ −1 (ξ δ ) (1 − 4K0 x − x̂)ϕ −1 ψ −1 (β δ )
λ
Now since ϕ −1 ψ −1 (λ ) = ψ −1 (λ ) we have
ψ −1 (ξ δ ) ψ −1 (β δ )
xαδ − x̂ ≤ + +Cψ −1 (β δ )
ξ 1 − 4K0 x − x̂
1 1
≤ ( + +C)ψ −1 (ηδ ). (9.23)
ξ 1 − 4K0 x0 − x̂
for all u ∈ D(F). Using (9.24), (9.25), Assumptions 10 (2), 11 for x̂ = 0, we get
K0 = 7.5 < K = 15.
where d0 , d1 and d2 are the given parameters. Note that F(x̂) = F(0) = 0. Then it
can easily be seen that, for d2 sufficiently large and d1 sufficiently small, KK0 can
be arbitrarily large.
Lavrentiev Regularization methods for Ill-posed equations 173
We now present two examples where Assumption 11 is not satisfied, but As-
sumption 10 (2) is satisfied.
and so
F (x) − F (x̂) ≤ K0 |x − x̂|.
References
King-Werner-type
√
methods of order 1 + 2
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
10.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176
10.2 Majorizing sequences for King-Werner-type methods . . . . . . . . . . . . . . . 178
10.3 Convergence analysis of King-Werner-type methods . . . . . . . . . . . . . . . . 183
10.4 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188
10.1 Introduction
Iterative methods are used to generate a sequence of approximating a solution x∗
of the nonlinear equation
F(x) = 0, (10.1)
where F is Fréchet-differentiable operator defined on a convex subset D of a
Banach space X with values in a Banach space Y .
176
√
King-Werner-type methods of order 1 + 2 177
xn = xn−1 − F ( xn−1 +y
2
n−1 −1
) F(xn−1 ) f or each n = 1, 2, ...
xn−1 +yn−1 −1 (10.4)
yn = xn − F ( 2 ) F(xn ) f or each n = 1, 2, ....
Lemma 10.2.1
Let L0 > 0, L > 0, s ≥ 0, η > 0 be given parameters. Denote by α the only positive
√
King-Werner-type methods of order 1 + 2 179
t0 = 0,t1 = η, s0 = s
L(s +t −2t )(t −t )
sn = tn + n−1 Ln0 n−1 n n−1
2(1− 2 (tn−1 +sn−1 )) (10.7)
L(sn−1 +tn −2tn−1 )(tn −tn−1 )
tn+1 = tn + L f or each n = 1, 2...
2(1− 20 (tn +sn ))
and
sn < tn+1 (10.10)
for each n = 1, 2, ....
Proof: We have that p(0) = −L < 0 and p(1) = L0 > 0. Then if follows by the
intermediate value theorem that p has roots α ∈ (0, 1). On the other hand, by the
Descartes rule of signs α is the only positive root of polynomial p. Now from
(10.6) and the definition of sequence {tn } we have that
L(sk − tk + tk+1 − tk )
0< ≤α (10.11)
2(1 − L20 (tk + sk ))
and
L(sk − tk + tk+1 − tk )
0< ≤α (10.12)
2(1 − L20 (tk+1 + sk+1 ))
hold for k = 0. Since, (10.11) and (10.12) are satisfied for k = 0 we get from
(10.7) that
0 < sk − tk ≤ α(tk − tk−1 ) ≤ α k η (10.13)
0 < tk+1 − tk ≤ α(tk − tk−1 ) ≤ α k η (10.14)
180 Iterative methods and their dynamics with applications: a contemporary study
1 − α k+1
sk ≤ η < t (10.15)
1−α
1 − α k+1
tk+1 ≤ η < t (10.16)
1−α
and
sk < tk+1
also hold for k = 1. Notice that if we show (10.12) using induction then (10.11)
follows too for all k = 0, 1, 2, .... Evidently, (10.12) holds, if
L k L0 1 − α k+1 1 − α k+2
(α η + α k η) + α( + )η − α ≤ 0. (10.17)
2 2 1−α 1−α
Estimate (10.17) motivates us to introduce recurrent polynomials fk on [0, 1)
defined by
L0
fk (t) = Lt k−1 η + (1 + t + ... + t k + 1 + t + ... + t k+1 )η − 1 (10.18)
2
and show instead of (10.17) that
p1 (t) = L0t 2 + Lt − L.
Then,
−L + L2 + 4L0 L
β=
2L0
is the only positive root of polynomial p1 . Moreover, define recurrent polynomi-
als gk on [0, 1) by
1 − t k+1 L0
gk (t) = L(r2 − r1 )t k−1 + L0 ( )(r2 − r1 ) + (2r1 + (r2 − r1 )t) − 1.
1−t 2
Then, we get
gk+1 (t) = gk (t) + p1 (t)t k−1 (r2 − r1 )
and
gk+1 (β ) = gk (β ) f or each k = 1, 2, ....
On the other hand, define function g∞ on [0, 1) by
Then, we obtain
1
g∞ (β ) = (2L0 r1 + (r2 − r1 )β L0 − 2).
2
Finally, if one defines parameter β0 by
L(q1 + r2 − 2r1 )
β0 = .
2(1 − L20 (r2 + q2 ))
182 Iterative methods and their dynamics with applications: a contemporary study
It is worth noticing that r1 , q1 and r2 depend on the initial data. Then, using the
proof of Lemma 10.2.1, the above definitions and β instead of α we arrive at the
following result.
Lemma 10.2.3 Suppose that
2(1 − L0 r1 )
0 < β0 ≤ β ≤ . (10.26)
(r2 − r1 )L0
r2 ≤ r ≤ r .
and
rn+1 ≤ qn
for each n = 1, 2, ....
From a straight simple induction argument and the hypothesis L0 ≤ L we get
in turn that:
Lemma 10.2.4 Suppose that hypotheses (10.6), (2.22) are satisfied and L0 ≤
L. Then, the following estimates hold
q n ≤ sn ,
rn ≤ tn ,
qn − rn ≤ sn − tn ,
rn+1 − qn ≤ tn+1 − sn ,
rn+1 − rn ≤ tn+1 − tn
and
r ≤ t .
Moreover strict inequality holds in all but the last inequality for each n = 2, 3, ...,
if L0 < L.
√
King-Werner-type methods of order 1 + 2 183
xn − x ≤ t − tn , (10.33)
where, {tn } is given in Lemma 10.2.1. Furthermore, if there exists R > t such
that
U(x0 , R) ⊆ D (10.34)
and
L0 (t + R) ≤ 2, (10.35)
then, the point x is the only solution of equation F(x) = 0 in U(x0 , R).
Proof: We shall show using induction that the following assertions hold:
(Ik ) xk+1 − xk ≤ tk+1 − tk
(IIk ) yk − xk ≤ sk − tk
and
(IIIk ) F ( xk +y k −1
2 ) F (x0 ) ≤ L0
1
.
1− 2 (xk −x0 +yk −x0 )
We have that (Ik ), (IIk ) hold for k = 0 by the initial conditions (10.27)-(10.29).
Using (10.31), (II0 ), (10.6) and (10.16) we get
It follows from (10.36) and the Banach lemma on invertible operators [7] that
184 Iterative methods and their dynamics with applications: a contemporary study
F ( x0 +y0 −1
2 ) ∈ L(Y, X) so that (III0 ) is satisfied. Suppose that (Ik ), (IIk ), (IIIk )
hold for k < n, we get that
n+1
n+1
xn+1 − x0 ≤ xi − xi−1 ≤ (ti − ti−1 ) = tn+1 − t0 = tn+1 < t ,
i=1 i=1
Therefore, F ( xn +yn −1
2 ) ∈ L(Y, X) and (IIIn ) holds. Using (10.4) we get in turn
that
F(xn ) = F(xn ) − F(xn−1 ) − F ( xn−1 +y n−1
)(xn − xn−1 )
1 2
= 0 [F (xn−1 + θ (xn − xn−1 )) − F ( xn−1 +y
2
n−1
)](xn − xn−1 )dθ .
(10.38)
Then, by (10.7), (10.30), (IIIn ) and (10.38) we get
xn+1 − xn ≤
F ( xn +y n −1
2 ) F (x0 )F (x0 )
−1 1
0
[F (xn−1 + θ (xn − xn−1 ))
xn−1 +yn−1
−F ( )]dθ xn − xn−1
2 1
≤ L0
L
2xn−1 + 2θ (xn − xn−1 ) − xn−1 − yn−1 xn − xn−1 dθ
2(1− 2 (tn +sn )) 0
1
≤ L0
L
0
(xn−1 − yn−1 ) + 2θ (xn − xn−1 )xn − xn−1 dθ
2(1− 2(tn +sn ))
≤ L
L
(sn−1 − tn−1 + tn − tn−1 )(tn − tn−1 ) = tn+1 − tn ,
2(1− 20 (tn +sn ))
(10.39)
which shows (In ). Moreover, from (10.4) we get that
yn − xn ≤ F ( xn−1 +y
2
n−1 −1
) F (x0 )F (x0 )−1 F(xn )
L(sn−1 −tn−1 +tn −tn−1 )(tn −tn−1 ) (10.40)
≤ L0 = sn − tn ,
2(1− 2 (tn−1 +sn−1 ))
which shows that (IIn ) holds. The induction is now complete. Thus, {xn } is
a complete sequence in a Banach space X and as such it converges to some
x ∈ U(x0 ,t ) (since U(x0 ,t ) is a closed set). In view of (10.39) we obtain the
estimate
L
F (x0 )−1 F(xn ) ≤ ((sn−1 − tn−1 ) + (tn − tn−1 ))(tn − tn−1 ). (10.41)
2
√
King-Werner-type methods of order 1 + 2 185
and similarly
y1 − x1 ≤ q1 − r1 ,
which justify the definition of q1 , r1 , r2 and consequently the definition of se-
quence {rn }.
Then, we present the following result.
Theorem 10.3.3 Let F : D ⊆ X → Y be a Fréchet-differentiable operator.
Suppose that there exist x0 , y0 ∈ D, L0 > 0, L > 0, η > 0 with L0 ≤ L such that
for each x, y ∈ D
F (x0 )−1 ∈ L(Y, X)
F (x0 )−1 F(x0 ) ≤ η
F (x0 )−1 (F (x) − F (y)) ≤ Lx − y
F (x0 )−1 (F (x) − F (x0 )) ≤ L0 x − x0
U(x0 , r ) ⊆ D
and hypotheses of Lemma 10.2.3 hold with s = 0, where r is given in Lemma
10.2.3. Then, sequence {xn } generated by King-Werner-type method (10.3) with
y0 = x0 is well defined, remains in U(x0 , r ) and converges to a unique solution
x ∈ U(x0 , r ) of equation F(x) = 0. Moreover, the following estimates hold for
each n = 0, 1, 2, ...
xn − x ≤ r − rn ,
where, {rn } is given in Lemma 10.2.3. Furthermore, if there exists R > r such
that
U(x0 , R) ⊆ D
and
L0 (r + R) ≤ 2,
then, the point x is the only solution of equation F(x) = 0 in U(x0 , R).
186 Iterative methods and their dynamics with applications: a contemporary study
It follows from (10.47) and the Banach lemma on invertible operators that
F (x0 )−1 ∈ L(Y, X) and
1
F (x0 )−1 F (x ) ≤ . (10.48)
1 − l0 x − x0
Similarly, we have that
x0 + y0 1 1
− x ≤ (x − x0 + x − y0 ) < (ρ + ρ) = ρ,
2 2 2
x0 +y0
so, 2 ∈ U(x , ρ) and
x0 + y0 x + y0 l
F (x )−1 (F (x )−F ( )) ≤ l0 x − 0 ≤ 0 (x −x0 +x −y0 ) < l0 ρ < 1.
2 2 2
(10.49)
√
King-Werner-type methods of order 1 + 2 187
Using the second substep of (10.4) and (10.42), we get in turn that
x1 − x = x0 − x − F ( x0 +y0 −1 x0 +y0 −1
2 ) F(x0 ) = F ( 2 )
[F ( x0 +y
2 )(x0 − x ) −F(x0 )]
0
x0 +y0 −1 1 (10.51)
= −F ( 2 ) F (x ) 0 F (x )−1 [F (x + θ (x0 − x ))
+y
−F ( 0 2 0 )](x0 − x )dθ .
x
x0 + y0 1
− x ≤ (x0 − x + y0 − x ) < ρ,
2 2
using the first substep of (10.4) and (10.42) we get that
y1 − x = x1 − x − F ( x0 +y 0 −1
2 ) F(x1 )
+y 1
= −F ( 0 2 0 )−1 F (x ) 0 [F (x )−1 (F (x + θ (x1 − x ))
x
(10.53)
−F ( x0 +y
2 ))](x1 − x )dθ .
0
Next, using (10.43), (10.46), (10.50) and (10.53) we obtain in turn that
1 −1
y1 − x ≤ F ( x0 +y 0 −1
2 ) F (x ) 0 F (x ) (F (x + θ (x1 − x ))
x0 +y0
−F ( 2 ))x1 − x dθ
≤ l0
l
( x −x
2
0
+ x −y
2
0
+ x −x
2
1
)x1 − x
1− 2 (x0 −x +y0 −x )
3lρ
≤ 2(1−l0 ρ) x1 − x ≤ x1 − x < ρ,
(10.54)
x1 +y1
which shows y1 ∈ U(x , ρ). We also have that ∈ U(x , ρ). If we replace
2
the role of x0 , y0 by xk , yk in the preceding estimates we obtain as in (10.52) and
(10.54) that
xk+1 − x < xk − x < ρ (10.55)
188 Iterative methods and their dynamics with applications: a contemporary study
and
yk+1 − x < xk+1 − x < ρ, (10.56)
respectively. By letting k → ∞ in (10.55), we obtain limk→∞ xk = x . 2
Remark 10.3.6 Notice that the radius of convergence for Newton’s method
due to Traub [12] or Rheinboldt [11] is given by
2
ξ= . (10.57)
3l
The corresponding radius due to us in [2, 5, 6] is given by
2
ξ1 = . (10.58)
2l0 + l
Comparing (10.46), (10.57) and (10.58), we see that
ρ < ξ < ξ1 . (10.59)
Notice however that King-Werner-type method (10.4) is faster than Newton’s
method. Finally notice that
ρ ρ 1 ξ1 l0
→ 1, → and →3 as → 0. (10.60)
ξ ξ1 3 ξ l
Then, we have for x (t) = 0 (t ∈ [0, 1]) that l = L1 = 3 and l0 = 32 . Then, using
(10.46) we obtain that
1
ρ= .
6
Example 10.4.3 Let also X = Y = C[0, 1] equipped with the max norm and D =
U(0, r) for some r > 1. Define F on D by
1
F(x)(s) = x(s) − y(s) − μ G(s,t)x3 (t)dt, x ∈ C[0, 1], s ∈ [0, 1].
0
y ∈ C[0, 1] is given, μ is a real parameter and the Kernel G is the Green’s function
defined by
(1 − s)t if t ≤ s
G(s,t) =
s(1 − t) if s ≤ t.
Then, the Fréchet derivative of F is defined by
1
(F (x)(w))(s) = w(s) − 3μ G(s,t)x2 (t)w(t)dt, w ∈ C[0, 1], s ∈ [0, 1].
0
Let us choose x0 (s) = y(s) = 1 and |μ| < 83 . Then, we have that
and
6r|μ|
L= .
8 − 3|μ|
Let us simple choose y0 (s) = 1, r = 3 and μ = 12 . Then, we have that
and
L(s + η)
L(s+η)
= 0.057441746, α = 0.711345739,
2(1 − 2 (2 + 2−L0 s )η)
L0
1 − L0 η = 0.928994083.
That is, condition (10.6) is satisfied and Theorem 10.3.1 applies.
References
[10] Y. H. Geum, Y. I., Kim and Á. A., Magreñán, A biparametric extension
of King’s fourth-order methods and their dynamics, Appl. Math. Comput.,
282, 254–275, 2016.
[11] L. V. Kantorovich and G. P. Akilov, Functional Analysis, Pergamon Press,
Oxford, 1982.
[12] R. F. King, Tangent methods for nonlinear equations, Numer. Math., 18,
298–304, 1972.
[13] Á. A. Magreñán, Different anomalies in a Jarratt family of iterative root-
finding methods, Appl. Math. Comput., 233, 29–38, 2014.
[14] Á. A. Magreñán, A new tool to study real dynamics: The convergence
plane, Appl. Math. Comput., 248, 215–224, 2014.
[15] T. J. McDougall and S. J. Wotherspoon, A simple
√ modification of Newton’s
method to achieve convergence of order 1 + 2, Appl. Math. Let., 29, 20–
25, 2014.
[16] F. A. Potra and V. Ptak, Nondiscrete induction and iterative processes [J].
Research Notes in Mathematics, 103, Pitman, Boston, 5, 112–119, 1984.
[17] H. M. Ren, Q. B. Wu and W. H. Bi, On convergence of a new secant like
method for solving nonlinear equations, Appl. Math. Comput., 217, 583–
589, 2010.
[18] W. C. Rheinboldt, An adaptive continuation process for solving systems
of nonlinear equations, Polish Academy of Science, Banach Ctr. Publ., 3,
129–142, 1977.
[19] J. F. Traub, Iterative Methods for the Solution of Equations, Englewood
Cliffs, Prentice Hull, 1984.
√
[20] W. Werner, Uber ein Verfahren der Ordnung 1 + 2 zur Nullstellenbestim-
mung, Numer. Math., 32, 333–342, 1979.
√
[21] W. Werner, Some supplementary results on the 1 + 2 order method for
the solution of nonlinear equations, Numer. Math., 38, 383–392, 1982.
Chapter 11
Generalized equations
and Newton’s method
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
11.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193
11.2 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194
11.3 Semilocal convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
11.1 Introduction
In [18], G. S. Silva considered the problem of approximating the solution of the
generalized equation
F(x) + Q(x) 0, (11.1)
where F : D −→ H is a Fréchet differentiable function, H is a Hilbert space
with inner product ., . and corresponding norm ., D ⊆ H an open set and
T : H ⇒ H is set-valued and maximal monotone. It is well known that the sys-
tem of nonlinear equations and abstract inequality system can be modelled as
equation of the form (11.1) [17]. If ψ : H −→ (−∞, +∞] is a proper lower
193
194 Iterative methods and their dynamics with applications: a contemporary study
F(x) + ∂ ψ(x) 0,
11.2 Preliminaries
Let U(x, ρ) and Ū(x, ρ) stand respectively for open and closed balls in H with
center x ∈ H and radius ρ > 0.
The following items are stated briefly here in order to make the chapter as
selfcontains as possible. More details can be found in [18].
Definition 11.2.1 Let D ⊆ H be an open nonempty subset of H, h : D −→ H be
a Fréchet differentiable function with Fréchet derivative h and Q : H ⇒ H be
a set mapping. The partial linearization of the mapping h + Q at x ∈ H is the
set-valued mapping Lh (x, .) : H ⇒ H given by
Lemma 11.2.4 ( [22]) Let G be a positive operator (i.e.,G(x), x ≥ 0). The fol-
lowing statements about G hold:
G2 = G2 ;
If G−1 exists, then G−1 is a positive operator.
Lemma 11.2.5 ( [21, Lemma 2.2]) Let G be a positive operator. Suppose G−1
exists, then for each x ∈ H we have
x2
G(x), x ≥ .
G−1
F̂ (x0 )−1 F (x) − F (x0 ) ≤ f0 (x − x0 ) − f0 (0). (11.7)
f0 (tn0 )
0
tn+1 = tn0 − , n = 0, 1, 2, . . .
f0 (tn0 )
is strictly increasing, remains in (0,t0∗ ) and converges to t0∗ , where t0∗ is the small-
est zero of function f0 in (0, R). Furthermore, suppose that for each x, y ∈ D1 :=
Ū(x0 , ρ) ∩U(x0 ,t0∗ ) there exists f1 : [0, ρ1 ) −→ R, ρ1 = min{ρ,t0∗ } such that
F̂ (x0 )−1 F (y) − F (x) ≤ f1 (x − y + x − x0 ) − f1 (x − x0 )
(11.8)
and x1 − x0 ≤ f1 (0)
f1 (tn ) f1 (sn )
1
tn+1 = tn1 − (or sn+1 = sn − )
f1 (tn ) f0 (sn )
are well defined, {tn1 } (or sn ) is strictly increasing, remians in (0,t1∗ ) and
converges to t1∗ . Moreover, sequence {xn } generated by generalized Newton’s
method (11.3) is well defined, remains in U(x0 ,t1∗ ) and converges to a point x∗ ∈
Ū(x0 ,t1∗ ), which is the unique solution of generalized equation F(x) + Q(x) 0
in Ū(x0 ,t1∗ ). Furthermore, the following estimates hold:
t1∗ − tn+1
1
xn+1 − x∗ ≤ xn − x∗ 2
(t1∗ − tn1 )2
t1∗ − s1n+1
xn+1 − x∗ ≤ xn − x∗ 2
(t1∗ − s1n )2
sn+1 − sn ≤ tn+1
1
− tn1 ,
Generalized equations and Newton’s method 197
D− f1 (t1∗ )
xn+1 − x∗ ≤ xn − x∗ 2 ,
−2 f1 (t1∗ )
D− f1 (t1∗ )
xn+1 − x∗ ≤ xn − x∗ 2 ,
−2 f0 (t1∗ )
D− f1 (t1∗ ) ∗
t1∗ − tn+1 ≤ (t − tn )2 ,
−2 f1 (t1∗ ) 1
and
D− f1 (t1∗ ) ∗
t1∗ − sn+1 ≤ (t − sn )2 ,
−2 f0 (t1∗ ) 1
where D− stands for the left directional derivative of function f1 .
Remark 11.3.2 (a) Suppose that there exists f : [0, R) −→ R twice continu-
ously differentiable such that for each x, y ∈ U(x0 , ρ)
leading to tn1 ≤ tn , sn ≤ tn ,
1
tn+1 − tn1 ≤ tn+1 − tn (11.13)
f (tn )
t0 = 0, tn+1 = tn − ,
f (tn )
f1 (t) f (s)
− ≤− (11.16)
f1 (t) f (s)
or
f1 (t) f (s)
− ≤− , (11.17)
f0 (t) f (s)
respectively for each t ≤ s. Estimates (11.13) and (11.14) can be strict if
(11.16) and (11.17) hold as strict inequalities.
(b) We have improved the error bounds and the location of the solution x∗
but not necessarily the convergence domain of the generalized Newton’s
method (11.3). We can also show that convergence domain can be im-
proved in some interesting special cases. Let F ≡ {0},
L
f (t) = t 2 − t + η,
2
L0 2
f0 (t) = t −t +η
2
and
L1 2
f1 (t) =
t − t + η,
2
where x1 − x0 ≤ η and L, L0 and L1 are Lipschitz constants satisfying:
L0 ≤ L
and
L1 ≤ L.
The corresponding majorizing sequences are
f (tn ) L(tn − tn−1 )2
t0 = 0, tn+1 = tn −
= tn +
f (tn ) 2(1 − Ltn )
1 1 2
f1 (tn1 ) L1 (tn − tn−1 )
t01 = 0, tn+1
1
= tn1 − = tn1 +
f1 (tn1 ) 2(1 − Ltn1 )
f1 (sn ) − f1 (sn−1 ) − f1 (sn−1 )(sn − sn−1 ) L1 (sn − sn−1 )2
s0 = 0, sn+1 = sn − = s n + .
f0 (sn ) 2(1 − L0 sn )
Then, sequences converge provided, respectively that
1
h = Lη ≤ (11.18)
2
and for the last two
1
h1 = L1 η ≤ ,
2
so
1 1
h≤
=⇒ h1 ≤ .
2 2
It turns out from the proof of Theorem 11.3.1 that sequence {rn } defined
by [6]
L0 (r1 − r0 )2
r0 = 0, r1 = η, r2 = r1 + ,
2(1 − L0 r1 )
L1 (rn+1 − rn )2
rn+2 = rn+1 +
2(1 − L0 rn+1 )
is also a tighter majorizing sequence than the preceding ones for {xn }.
The sufficient convergence condition for {rn } is given by [6]:
1
h2 = Kη ≤ ,
2
where
1 √
K = (4L0 + L1 L0 + 8L02 + L0 L1 ).
8
Then, we have that
1 1
h1 ≤ =⇒ h2 ≤ .
2 2
200 Iterative methods and their dynamics with applications: a contemporary study
Therefore, the old results in [18] have been also improved. Similar im-
provements can follow for the Smale’s alpha theory [2, 6] or Wang’s the-
ory [18,22,24]. Examples where L0 < L or L1 < L or L0 < L1 can be found
in [6]. Notice that (11.18) is the celebrated Newton-Kantorovich hypothe-
sis for solving nonlinear equations using Newton’s method [13] employed
as a sufficient convergence condition in all earlier studies other than ours.
(c) The introduction of (11.8) depends on (11.7) (i.e., f1 depends on f0 ). Such
an introduction was not possible before (i.e., when f was used instead of
f1 ).
Now we can prove the Theorem 11.3.1.
Proof of Theorem 11.3.1 It is clear that the iterates {xn } ∈ D1 which is a more
precise location than Ū(x0 , ρ) used in [18], since D1 ⊆ Ū(x0 , ρ). Then, the defi-
nition of function f1 becomes possible and replaces f in the proof [18], whereas
for the computation on the upper bounds F̂ (x)−1 we use the more precise f0
than f as it is shown in the next perturbation Banach lemma [13].
2
Lemma 11.3.3 Let x0 ∈ D be such that F̄ (x0 ) is a positive operator and
F̄ (x0 )−1 exists. If x−x0 ≤ t < t ∗ , then F̄ (x) is a positive operator and F̄ (x)−1
exists. Furthermore,
F̄ (x0 )−1
F̄ (x)−1 ≤ . (11.19)
f0 (t)
Proof: It is easy to see that
1 1
F̄ (x)− F̄ (x0 ) ≤ F̄ (x)− F̄ (x0 )+ (F̄ (x)− F̄ (x0 ))∗ = F̄ (x)− F̄ (x0 ).
2 2
(11.20)
∗
Let x ∈ Ū(x0 ,t), 0 ≤ t < t . Thus f (t) < 0. Using (h1 ) and (h2 ), we get in turn
1 1
that
F̄ (x0 )−1 F̄ (x) − F̄ (x0 ) ≤ F̄ (x0 )−1 F̄ (x) − F̄ (x0 )
≤ f0 (x − x0 ) − f0 (0)
< f0 (t) + 1 < 1. (11.21)
Now, using Banach’s Lemma on invertible operators, we have F̄ (x)−1 exists.
Moreover, using the above inequality,
F̄ (x0 )−1 F̄ (x0 )−1 F (x0 )−1
F̄ (x)−1 ≤ ≤ = − .
1 − F̄ (x0 )−1 F (x) − F (x0 ) 1 − ( f0 (t) + 1) f0 (t)
Now, using (11.21) we obtain in turn that
1
F̄ (x) − F̄ (x0 ) ≤ . (11.22)
F̄ (x0 )−1
Generalized equations and Newton’s method 201
Consequently, we get
y2
(F̄ (x0 ) − F̄ (x))y, y ≤ F̄ (x0 ) − F̄ (x)y2 ≤ ,
F̄ (x −1
0)
which implies,
y2
F̄ (x0 )y, y − ≤ F̄ (x)y, y. (11.23)
F̄ (x0 )−1
Now since F̄ (x0 ) is a positive operator and F̄ (x0 )−1 exists by assumption, we
get
y2
F̄ (x0 )y, y ≥ . (11.24)
F̄ (x0 )−1
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
12.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
12.2 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
12.3 Local convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 206
12.4 Special cases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 210
12.1 Introduction
In this chapter we are concerned with the study of the generalized equation
204
Newton’s method for generalized equations using restricted domains 205
F(x) + ∂ ψ(x) 0,
for solving (12.1) in an approximate way. We will use the idea of restricted con-
vergence domains to present a convergence analysis of (12.2). In our analysis we
relax the Lipschitz type continuity of the derivative of the operator involved. Our
main goal in this chapter is to find larger convergence domain for the method
(12.2). By means of using the restricted convergence domains, we obtained a
finer convergence analysis, with the advantages (A):
tighter error estimates on the distances involved and
the information on the location of the solution is at least as precise.
These advantages were obtained (under the same computational cost) using
the same or weaker hypotheses as in [36].
The rest of the chapter is organized as follows. In Section 12.2 we present
some preliminaries, then in Section 12.3 we present the local convergence study
of method (12.2). Finally, in Section 12.4 we provide the special cases and nu-
merical examples.
12.2 Preliminaries
In order to make the chapter as self contained as possible we reintroduce some
standard notations and auxiliary results for the monotonicity of set valued oper-
ators presented in [18, 22, 27].
Throughout this chapter we will denote by U(w, ρ),U(w, ρ), the open and
closed balls in B1 , respectively, with center w ∈ H and of radius ρ > 0.
206 Iterative methods and their dynamics with applications: a contemporary study
Lemma 12.2.4 Let G be a positive operator. Suppose that G−1 exists, then for
each x ∈ H we have
x2
G(x), x ≥ .
G−1
Lemma 12.2.5 Let B : H −→ H be a bounded linear operator and I : H −→ H
the identity operator. If B − I < 1 then B is invertible and B−1 ≤ (1−B−I)
1
.
(h0 )
−1
F
(x∗ ) F (x) − F (x∗ ) ≤ f0 (x − x∗ ) − f0 (0), (12.3)
x ∈ D and
−1
F
(x∗ ) F (x) − F (x∗ + θ (x − x∗ )) ≤ f (x − x∗ ) − f (θ x − x∗ )
(12.4)
for each x ∈ D0 = D ∩U(x∗ , R), θ ∈ [0, 1].
(h1 ) f (0) = f0 (0) and f (0) = f0 (0) = −1, f0 (t) ≤ f (t), f0 (t) ≤ f (t) for each
t ∈ [0, R).
(h2 ) f0 , f are convex and strictly increasing.
respectively,
f (tk )
0 ∈ F(xk ) + F (xk )(xk+1 − xk ) + Q(xk+1 ), tk+1 = |tk − |, k = 0, 1, . . . ,
f (tk )
(12.5)
are well defined, {tk } is strictly decreasing, is contained in (0, r) and converges
to 0, {xk } is contained in U(x∗ , r) and converges to the point x∗ which is the
unique solution of the generalized equation F(x) + Q(x) 0 in U(x∗ , σ̄ ), where
σ̄ = min{r, σ } and σ := sup{0 < t < R : f (t) < 0}. Moreover, the sequence { tk+1 tk2
}
is strictly decreasing,
tk+1 ∗ 2 tk+1 f (t0 )
x∗ − xk+1 ≤ [ ]x k − x , ≤ , k = 0, 1, . . . . (12.6)
tk2 tk2 2| f (t0 )|
If, additionally ρ f ρ(ρ)− f (ρ)
f (ρ) = 1 and ρ < R then r = ρ is the optimal convergence
radius. Furthermore, for t ∈ (0, r) and x ∈ Ū(x∗ ,t)
e f (xk − x∗ , 0)
xk+1 − x∗ ≤
| f0 (xk∗ − x∗ )|
e f (xk − x∗ , 0) |η f (t)|
:= αk ≤ ∗ ≤ xk − x∗ 2 , (12.7)
| f (xk − x∗ )| t2
where
e f (s,t) := f (t) − ( f (s) + f (s)(t − s)) for each s,t ∈ [0, R)
and
f (t)
η f (t) := t − for each s,t ∈ [0, ν).
f (t)
Finally, by the second inequality in (12.7) there exists r∗ ≥ r such that
limk−→∞ xk = x∗ , if x0 ∈ U(x∗ , r∗ ) − {x∗ }.
208 Iterative methods and their dynamics with applications: a contemporary study
Before proving the results of previous Theorem we must consider the following
results.
From now on we assume that the hypotheses of Theorem 12.3.1 hold.
Remark 12.3.2 The introduction of the center-Lipschitz-type condition (12.3)
(i.e., function f0 ) lends to the introduction of restricted Lipschitz-type condition
(12.4). The condition used in earlier studies such as [36] is given by
−1
F
(x∗ )F (x) − F (x∗ + θ (x − x∗ )) ≤ f1 (x − x∗ ) − f1 (θ x − x∗ )
(12.8)
for each x ∈ D, θ ∈ [0, 1], where f1 : [0, +∞)R is also twice continuously differ-
entiable. It follows from (12.3), (12.4) and (12.8) that
f0 (t) ≤ f1 (t) (12.9)
f (t) ≤ f (t) (12.10)
for each t ∈ [o, ν) since D0 ⊆ D. If f0 (t) = f1 (t) = f (t) for each t ∈ [0, ν), then
our results reduce to the corresponding ones in [36]. Otherwise, (i.e., if strict
inequality holds in (12.9) or (12.10)) then the new results improve the old ones.
Indeed, let
t f (t) − f1 (t)
r1 := sup{t ∈ (0, ν̄) : − 1 < 1},
t f1 (t)
where ν1 := sup{t ∈ [0, +∞) : f1 (t) < 0}. Then, the error bounds are (corre-
sponding to (12.7)):
e f1 (xk − x∗ , 0)
xk+1 − x∗ ≤
| f1 (xk∗ − x∗ )|
|η f1 (t)|
:= βk ≤ xk − x∗ 2 . (12.11)
t2
In view of the definition of r, r1 and estimates (12.7), (12.9), (12.10) and (12.11),
it is clear that
r1 ≤ r (12.12)
and
αk ≤ βk , k = 1, 2, . . . . (12.13)
Hence, we obtain a larger radius of convergence and tighter error estimates on
the distances involved, leading to a wider choice of initial guesses x∗ and fewer
computations of iterates xk in order to achieve a desired error tolerance. It is
also worth noticing:
The advantages are obtained under the same computational cost due to the
fact in practice the computation of function f1 requires the computation, as spe-
cial cases, of functions f0 and f . The introduction of function f was not possible
before (when only function f1 was used). This inclusion is possible using function
f0 (i.e., f is a function of f0 ).
Newton’s method for generalized equations using restricted domains 209
Next, we present an auxiliary Banach Lemma relating the operator F with the
majorant function f0 .
−1 −1
F
(x∗ ) F
(x) − F
(x∗ ) ≤ F (x∗ ) F (x) − F (x∗ ) ≤ f0 (x − x∗ ) − f0 (0) < 1.
(12.16)
−1
By Banach Lemma, we conclude that F
(x∗ ) exists. Moreover taking into
account the last inequality
−1 −1 −1
−1 F
(x∗ ) F (x∗ ) F (x∗ )
F
(x) ≤ ≤ = .
−1 1 − ( f0 (x − x∗ ) − f0 (0)) | f0 (x − x∗ )|
1 − F
(x∗ ) F (x) − F (x∗ )
The last step follows from the fact that r = min{R, ν}. Moreover, using
(12.16) we have
1
F
(x) − F (x∗ ) ≤
−1
. (12.17)
F
(x∗ )
y2
(F
(x∗ ) − F
(x))y, y ≤ F
(x∗ ) − F (x)y2 ≤
−1
,
F
(x∗ )
which implies
y2
F
(x∗ )y, y −
−1
≤ F
(x)y, y.
F
(x∗ )
210 Iterative methods and their dynamics with applications: a contemporary study
−1
Since F
(x∗ ) is a positive operator and F (x∗ ) exists by assumption, we obtain
by Lemma 12.2.5 that
y2
F
(x∗ )y, y ≥
−1
.
F
(x∗ )
Therefore, combining the two last inequalities we conclude that F (x)y, y ≥ 0,
i.e., F
(x) is a positive operator.
−1
Lemma 12.2.4 shows that F
(x) is a positive operator and F (x) exists, thus
by Lemma 12.2.3 we have that for any y ∈ H
y2
F
(x)y, y ≥
−1
.
F
(x)
2
It is worth noticing that F
(x)y, y = F (x)y, y. Thus by the second part of
0 ∈ F(x) + F (x)(NF+Q (x) − x) + Q(NF+Q (x)), for all x ∈ U(x∗ , r). (12.18)
instead of (12.14). However, we have that (12.14) gives a tighter error estimate
than (12.15), since | f1 (t) ≤ | f0 (t)|. This is really important in the proof of Theo-
rem 12.3.1.
Proof of Theorem 12.3.1 Simply follow the proof of Theorem 4 in [36] but
realize that the iterates xk lie in D0 which is a more precise location than D (used
in [36]) allowing the usage of tighter function f than f1 and also the usage of
tighter function f0 that f1 for the computation of the upper bounds of the inverses
−1
F (x) (i.e., we use (12.14) instead of (12.19)).
L
f (t) = t 2 − t
2
and
L1 2
f1 (t) =
t −t
2
for some positive constants L0 , L and L1 to be determined using a specialized
operator F.
Lxk − x∗ 2
xk+1 − x∗ ≤
2(1 − L0 xk − x∗ )
Lxk − x∗ 2
xk+1 − x∗ ≤
2(1 − Lxk − x∗ )
and
L1 xk − x∗ 2
xk+1 − x∗ ≤ .
2(1 − L1 xk − x∗ )
References
[37] S. Smale, Newton method estimates from data at one point, in: R. Ewing,
K. Gross and C. Martin (Eds.), The Merging of Disciplines: New Directions
in Pure, Applied and Computational Mathematics, pp. 185–196. Springer-
Verlag, New York, 1986.
[38] J. F. Traub and H. Wozniakowski, Convergence and complexity of Newton
iteration for operator equations, J. Assoc. Comput. Mach., 26 (2), 250–258,
1979.
[39] L. U. Uko, Generalized equations and the generalized Newton method,
Math. Programming, 73 (3, Ser. A), 251–268, 1996.
[40] X. Wang, Convergence of Newton’s method and uniqueness of the solution
of equation in Banach space, IMA J. Numer. Anal., 20, 123–134, 2000.
Chapter 13
Secant-like methods
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
13.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216
13.2 Semilocal convergence analysis of the secant method I . . . . . . . . . . . . . . 218
13.3 Semilocal convergence analysis of the secant method II . . . . . . . . . . . . . 226
13.4 Local convergence analysis of the secant method I . . . . . . . . . . . . . . . . . . 227
13.5 Local convergence analysis of the secant method II . . . . . . . . . . . . . . . . . 231
13.6 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 232
13.1 Introduction
In this chapter we study the problem of finding a locally unique solution x of
equation
F(x) = 0, (13.1)
where F is a Fréchet-differentiable operator defined on a convex subset D of a
Banach space X with values in a Banach space Y.
216
Secant-like methods 217
Many problems from Applied Sciences can be solved finding the solutions of
equations in a form like (13.1) [6, 7, 17, 22].
In this chapter, we present a both semilocal/local convergence study for the
secant-like method defined for each n = 0, 1, 2, . . . by
xn+1 = xn − A−1
n F(xn ), An = δ F(yn , zn ), (13.2)
y − x0 ≤ h1 (u − x0 , v − x0 )
and
z − x0 ≤ h2 (u − x0 , v − x0 ).
A−1
0 F(x0 ) ≤ η
and
F (x0 )−1 (δ F(x, y) − F (z)) ≤ f1 (x − z, y − z).
and
1
ψ(t) = f1 (t + θ η + h1 (t + c,t),t + θ η + h2 (t + c,t))dθ .
0
Now using the conditions (C) we present the main semilocal convergence
result.
Theorem 13.2.1 Suppose that conditions (C) hold. Then, the sequence {xn }
generated by method (13.2) starting at x−1 and x0 is well defined, remains in
U(x0 , r) for all n ≥ 0 and converges to a solution x∗ ∈ Ū(x0 , r) of equation
F(x) = 0, which is unique in Ū(x0 , r1 ) ∩ D.
Proof: We shall first show using mathematical induction that sequence {xn } is
well defined and belongs in U(x0 , r). By (C2 )-(C4 ), y0 = g1 (x−1 , x0 ) ∈ D, z0 =
g2 (x−1 , x0 ) ∈ D and A−1
0 ∈ L(Y, X ). It follows that x1 is well defined and x1 −
x0 = A−1 0 F(x0 ) < η < r, since r is a solution of equation (13.8). Then, we
have that x1 ∈ U(x0 , r). Moreover, y1 = g1 (x0 , x1 ) and z1 = g2 (x0 , x1 ) are well
defined by the last two hypotheses in (C5 ). By the second condition in (C4 ), we
get that
It follows from (13.9) and the Banach lemma on invertible operators [7, 18,
22] that A−1
1 ∈ L(Y, X ) and
1 1
A−1
1 F (x0 ) ≤ ≤ . (13.10)
1 − f0 (h1 (0, η), h2 (0, η)) 1 − ϕ(r)
Therefore, x2 is well defined.
We can write using method (13.2) for n = 0 and (C1 ) that
Then, using the first and second condition in (C4 ), (C3 ) and (13.11) we get in
Secant-like methods 221
turn that
1
F (x0 )−1 F(x1 ) = 0
F (x0 )−1 [F (x0 + θ (x1 − x0 )) − A0 ] dθ (x1 − x0 )
1
≤ 0
f1 (x0 − x0 + θ x1 − x0 + y0 − x0 ,
≤ ψ(r)x1 − x0 .
(13.12)
By (13.2), (13.10), (C5 ) and (13.12), we obtain
x2 − x1 = A−1 −1 −1
1 F(x1 ) ≤ A1 F (x0 )F (x0 ) F(x1 )
ψ(r)x1 − x0 (13.13)
≤ ≤ qx1 − x0 < η
1 − ϕ(r)
and
1 − q2 η
x2 − x0 = x2 − x1 + x1 − x0 ≤ x1 − x0 < = r.
1−q 1−q
(13.14)
That is x2 ∈ U(x0 , r). On the other hand, y2 = g1 (x1 , x2 ) and z2 = g2 (x1 , x2 ) are
well defined by the last two hypotheses in (C5 ). Then, as in (13.9), we have that
F (x0 )−1 (A2 − F (x0 )) ≤ f0 (h1 (η, r), h2 (η, r)) ≤ ϕ(r) < 1.
Then, from the first, second hypotheses in (C4 ), (C3 ) and (13.11) we have as
in (13.12) that
1
F (x0 )−1 F(x2 ) ≤ 0 f1 (x1 − x0 + θ x2 − x1 + y1 − x0 ,
≤ ψ(r)x2 − x1 .
(13.17)
We get
x3 − x2 = A−1 −1
2 F(x0 )F (x0 ) F(x2 )
ψ(r)
≤ x2 − x1 ≤ qx2 − x1 < q2 x1 − x0 < η
1 − ϕ(r)
and
1 − q3 η
x3 − x0 = x3 − x2 + x2 − x1 + x1 − x0 ≤ η< = r.
1−q 1−q
≤ ψ(r)x3 − x2 .
(13.19)
In view of (13.18) and (13.19) we get that
x4 − x3 = A−1 −1
3 F (x0 )F (x0 ) F(x3 )
ψ(r)
≤ x3 − x2 ≤ qx3 − x2 < q3 x1 − x0 < η
1 − ϕ(r)
and
x4 − x0 = x4 − x3 + x3 − x2 + x2 − x1 + x1 − x0
1 − q4 η
≤ η< = r.
1−q 1−q
Then, we get, as in (13.18) that y4 , z4 are well defined and
1
A−1
4 F (x0 ) ≤ (13.20)
1 − ϕ(r)
and
1
F (x0 )−1 F(x4 ) ≤ 0
f1 (r + θ η + h1 (r, r), r + θ η + h2 (r, r))dθ x4 − x3
≤ ψ(r)x3 − x2 .
(13.21)
Then, we have from (13.21) that
x5 − x4 = A−1 −1
4 F (x0 )F (x0 ) F(x4 )
ψ(r)
≤ x4 − x3 ≤ qx4 − x3 < q4 x1 − x0 < η
1 − ϕ(r)
and
x5 − x0 = x5 − x4 + x4 − x3 + x3 − x2 + x2 − x1 + x1 − x0
1 − q5 η
≤ η< = r.
1−q 1−q
224 Iterative methods and their dynamics with applications: a contemporary study
1
A−1
k F (x0 ) ≤ ,
1 − ϕ(r)
1 − qj qk
= xk+1 − xk < x1 − x0 .
1−q 1−q
(13.23)
for j = 1, 2, . . . and q < 1. It follows from (13.23) that sequence {xk } is com-
plete in a Banach space X and as such it converges to some x∗ ∈ Ū(x0 , r) (since
Ū(x0 , r) is a closed set). From the estimate
and letting k → ∞ we deduce that F(x∗ ) = 0. Next, we will study the uniqueness
part. Let y∗ ∈ Ū(x0 , r) be such that F(y∗ ) = 0. Then, using the last condition in
1
(C4 ) and (C7 ), we get in turn that for Q = 0 F (y∗ + θ (x∗ − y∗ ))dθ
Secant-like methods 225
1
F (x0 )−1 (F (x0 ) − Q) ≤ 0
f0 (y∗ + θ (x∗ − y∗ ) − x0 ,
y∗ + θ (x∗ − y∗ ) − x0 )dθ
1
≤ 0
f0 (θ x∗ − x0 + (1 − θ )y∗ − x0 ,
θ x∗ − x0 + (1 − θ )y∗ − x0 )dθ
1
f0 (θ r + (1 − θ )r1 , θ r + (1 − θ )r1 )dθ < 1.
≤ 0
(13.24)
If follows from (13.24) and the Banach’s Lemma on invertible operators that
Q−1 ∈ L(Y, X ). Then, using the identity 0 = F(x∗ ) − F(y∗ ) = Q(y∗ − x∗ ), we
conclude that y∗ = x∗ . .
Remark 13.2.2 (a) The condition A−10 ∈ L(Y, X ) can be dropped as follows.
Suppose that
f0 (h1 (c, 0), h2 (c, 0)) < 1. (13.25)
Then, we have by the third condition in (C4 ) that
1
A0 F (x0 )−1 ≤ . (13.26)
1 − f0 (h1 (c, 0), h2 (c, 0))
Then, due to the estimate
F (x0 )−1 F(x0 )
x1 − x0 ≤ A−1 −1
0 F (x0 )F (x0 ) F(x0 ) ≤ ,
1 − f0 (h1 (c, 0), h2 (c, 0))
we can define
F (x0 )−1 F(x0 )
η= . (13.27)
1 − f0 (h1 (c, 0), h2 (c, 0))
Then, the conclusions of Theorem 2.1 hold with (13.25) replacing condi-
tion A−1
0 ∈ L(Y, X ) and with η given by (13.27).
(b) In view of the second and third condition in (C4 ) we have that
x−1 − x0 ≤ c.
y − x0 ≤ h1 (u − x0 , v − x0 )
and
z − x0 ≤ h2 (u − x0 , v − x0 )
A−1
0 F(x0 ) ≤ η
A−1
0 (δ F(x, y) − A0 ) ≤ f 0 (x − y0 , y − z0 )
and
A−1
0 (δ F(x, y) − δ F(u, v)) ≤ f 1 (x − u, y − v)
has at least one positive zero and the smallest positive zero, denoted by R,
satisfies
ϕ(R) < 1,
h1 (c, 0) ≤ R̄, h1 (R, R) ≤ R̄
and
h2 (c, 0) ≤ R̄, h2 (R, R) ≤ R̄, for some R̄ ≥ R.
Set q = q(R).
(H6 ) Ū(x0 , R̄) ⊆ D.
(H7 ) There exists R1 ≥ R such that
Using the (H) conditions we present the following semilocal convergence result.
Theorem 13.3.1 Suppose that the conditions (H) hold. Then, the sequence {xn }
generated by method (13.2) starting at x−1 and x0 is well defined, remains in
U(x0 , R) for all n ≥ 0 and converges to a solution x∗ ∈ Ū(x0 , R) of equation
F(x) = 0, which is unique in Ū(x0 , R1 ) ∩ D.
It is worth noticing that the second and third conditions in (H4 ) do not neces-
sarily imply the Fréchet-differentiability of F.
y − x∗ ≤ h1 (u − x∗ , v − x∗ )
and
z − x∗ ≤ h2 (u − x∗ , v − x∗ )
228 Iterative methods and their dynamics with applications: a contemporary study
and
F (x∗ )−1 (δ F(x, y) − F (z)) ≤ f1 (x − z, y − z)
and 1
ψ(t) = f1 (θt + h1 (t,t), θt + h2 (t,t))dθ .
0
Moreover, define functions q and p by
ψ(t)
q(t) = .
1 − ϕ(t)
and
1
p(t) = f1 (θt + h1 (t,t), θt + h2 (t,t))dθ + f0 (h1 (t,t), h2 (t,t)) − 1.
0
Theorem 13.4.1 Suppose that the conditions (C∗ ) hold. Then, the sequence {xn }
generated for x−1 , x0 ∈ U(x∗ , r) \ {x∗ } by method (13.2) is well defined, remains
in U(x∗ , r) and converges to the solution x∗ of equation F(x) = 0, which is unique
in Ū(x0 , r1 ) ∩ D.
Proof: The proof is similar to the proof of Theorem 13.2.1. We shall first show
using mathematical induction that sequence {xn } is well defined and belongs in
U(x∗ , r). We have by the first condition in (C4∗ ) that
1 1
A−1
0 F (x0 ) ≤ ≤ .
1 − f0 (h1 (x−1 − x∗ , x0 − x∗ ), h2 (x−1 − x∗ , x0 − x∗ )) 1 − ϕ(r)
Thus, x1 is well defined. Then, we have by the second substep of condition (C4∗ )
that
1
F (x∗ )−1 ( 0 F (x∗ ) + θ (x0 − x∗ )) − A0 )dθ (x0 − x∗ )
1
≤ 0
f1 (x∗ + θ (x0 − x∗ ) − y0 , x∗ + θ (x0 − x∗ ) − z0 )dθ x0 − x∗
1
≤ 0
f1 (θ x0 − x∗ + y0 − x∗ , θ x0 − x∗ + z0 − x∗ )dθ x0 − x∗
1
≤ 0
f1 (θ x0 − x∗ + h1 (x−1 − x∗ , x0 − x∗ ), θ x0 − x∗
= −A−1 ∗ ∗
0 (F(x0 ) − F(x ) − A0 (x0 − x ))
= −A−1
1 ∗
0 0
F (x + θ (x0 − x∗ ) − A0 ) dθ (x0 − x∗ )
230 Iterative methods and their dynamics with applications: a contemporary study
we obtain
1
x1 − x∗ ≤ A−1 ∗
0 F(x ) 0
F (x∗ )−1 (F (x∗ + θ (x0 − x∗ )) − A0 )
dθ x0 − x∗
= ϕ(r) < 1,
so
1
A−1 ∗
1 F (x ) ≤ .
1 − ϕ(r)
That is x2 is well defined. Then, we obtain
1
F (x∗ )−1 0 F (x∗ + θ (x1 − x∗ )) − A1 dθ x1 − x∗
which shows that lim xk = x∗ and xk+1 ∈ U(x∗ , r). The part related to the unique-
k→∞
ness has been shown in Theorem 13.2.1. .
Secant-like methods 231
h1 (R, R) ≤ R̄
and
h2 (R, R) ≤ R̄, for some R̄ ≥ R.
Set q = q(R).
(H6∗ ) Ū(x0 , R̄) ⊆ D.
(H7∗ ) There exists R1 ≥ R such that
f0 (0, R1 ) < 1.
Theorem 13.5.1 Suppose that the conditions (H ∗ ) hold. Then, the sequence
{xn } generated for x−1 , x0 ∈ Ū(x∗ , R) \ x∗ by method (13.2) is well defined, re-
mains in U(x∗ , R) and converges to the solution x∗ of equation F(x) = 0, which
is unique in Ū(x0 , R1 ) ∩ D.
The results in Sections 13.4 and 13.5 hold for Newton-like methods, if we set
x−1 = x0 and zn = yn = xn , for each n = 0, 1, 2 . . ..
where x ∈ C[0, 1] and the kernel G is the Green’s function in [0, 1] × [0, 1].
We use a discretization process and transform equation (13.29) into a finite
dimensional problem. For this, we follow the process described in Section 13.2
in [12] with m = 8 and we obtain the following system of nonlinear equations:
F(x) ≡ x − 1 − Avx = 0, F : R 8 → R8 , (13.30)
where
a = (x1 , x2 , . . . , x8 )T , 1 = (1, 1, . . . , 1)T , A = (ai j )8i, j=1 , vx = (x12 , x22 , . . . , x82 )T .
We use the divided difference of first order of F as [u, v; F] = I − B, where
B = (bi j )8i, j=1 with bi j = ai j (u j + v j ).
If we choose the starting points x−1 = ( 10 , 10 , . . . , 10
7 7 7 T
) and x0 =
( 10 , 10 , . . . , 10 ) , method (13.5) with λ = 2 and the max-norm, we obtain c = 11
18 18 18 T 1
10 ,
y0 = ( 4 , 4 , . . . , 4 ) , β = 1.555774 . . ., η = 0.7839875 . . ., m = 0.257405 . . . and
5 5 5 T
the polynomial
0.257406
p(t) = −0.789051 + t 1 −
1 − 0.19223(0.55 + 2t)
Secant-like methods 233
has no real roots so the conditions in [12] are not satisfied so we cannot ensure
the convergence of method (13.5).
Next, we are going to use our (C) conditions in order to ensure the conver-
gence of the methods. We consider g1 (x, y) = y+x 2 and g2 (x, y) = y, and we define
the functions
s t
h1 (s,t) = + ,
2 2
h2 (s,t) = t,
f0 (s,t) = 0.0103507 . . . (s + t)
and
f1 (s,t) = 0.047827 . . . (s + t),
which satisfy the (C3 ) and (C4 ) conditions. Moreover, we define
ϕ(t) = 0.0207015t,
So, all the conditions of Theorem 13.2.1 are satisfied and a consequence we can
ensure the convergence of method (13.5).
234 Iterative methods and their dynamics with applications: a contemporary study
x0 = ( 14
10 , 10 , . . . , 10 ) , method (13.4) with a = 0.8, b = 0.75 and the max-norm,
14 14 T
we obtain c = 0.2,
Mδ γ 2 = 0.507089 . . . ≤ 0.5.
and
h2 (c, 0) = 0.745613 . . . ≤ r̄, h2 (r, r) = r̄ . . . ≤ r̄,
Finally to show that condition (C7 ) is verified we choose as r1 = 0.6 and we
obtain that
1
f0 (θ r + (1 − θ )r1 , θ r + (1 − θ )r1 )dθ = 0.0695705 . . . < 1.
0
So, all the conditions of Theorem 13.2.2 are satisfied and a consequence we can
ensure the convergence of method (13.4).
Example 13.6.3 Let X = Y = R3 , D = U(0, 1), x∗ = (0, 0, 0)T and define func-
tion F on D by
F(x, y, z) = (ex − 1, y2 + y, z)T . (13.31)
We have that for u = (x, y, z)T
⎛ ⎞
ex 0 0
F (u) = ⎝ 0
2y + 1 0 ⎠ , (13.32)
0 0 1
Using the norm of the maximum of the rows and (13.31)–(13.32) we see that since
F (x∗ ) = diag{1, 1, 1}, In this case, we take the domain Ω = B(x0 , R), where
x0 = 0.5 and R = 0.3. Choosing x−1 = 0.4, it is easy to see that now, we are
going to use our (C∗ ) conditions in order to ensure the convergence of the method
(13.7). We define the functions
h1 (s,t) = t
h2 (s,t) = s,
e
f0 (s,t) = (s + t)
2
and
f1 (s,t) = e(s + t)
which satisfy the (H3∗ ) and (H4∗ ) conditions. Moreover, we define
ϕ(t) = et,
ψ(t) = 3et,
φ (t) 3et
q(t) = =
1 − ϕ(t) 1 − et
and
1
p(t) = f1 (θt +h1 (t,t), θt +h2 (t,t))dθ + f0 (h1 (t,t), h2 (t,t))−1 = 10.8731t −1.
0
236 Iterative methods and their dynamics with applications: a contemporary study
So, all the conditions of Theorem 13.2.4 are satisfied and as a consequence we
can ensure the convergence of method (13.7).
Example 13.6.4 In order to show the applicability of our theory in a real prob-
lem we are going to consider the following quartic equation that describes the
fraction of the nitrogen-hydrogen feed that gets converted to ammonia, called the
fractional conversion showed in [13,25]. In Figure 13.1 they process of ammonia
is shown.
For 250 atm and 500C, this equation takes the form:
In this case, we take the domain Ω = B(x0 , R), where x0 = 0.5 and R = 0.3.
Choose x−1 = 0.4. We are going to use our (C∗ ) conditions in order to ensure
the convergence of the method (13.6). Define the functions
h1 (s,t) = s + 2t
h2 (s,t) = s,
f0 (s,t) = 0.111304(32.2642s + 20.5458t)
and
f1 (s,t) = 0.111304(40.9686s + 30.7779t),
which satisfy the (C3∗ ) and (C4∗ ) conditions. Moreover, we define
ϕ(t) = 13.0603t,
ψ(t) = 21.0985t,
Secant-like methods 237
φ (t) −1 + 34.1587t
q(t) = =
1 − ϕ(t) 1 − 13.0603t
and
1
p(t) = f1 (θt +h1 (t,t), θt +h2 (t,t))dθ + f0 (h1 (t,t), h2 (t,t))−1 = 34.1587t −1.
0
So, all the conditions of Theorem 13.2.3 are satisfied and as a consequence we
can ensure the convergence of method (13.6).
References
King-Werner-like
methods free of
derivatives
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
14.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 241
14.2 Semilocal convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 242
14.3 Local convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245
14.4 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246
14.1 Introduction
Recently, Argyros and Ren in [6] studied King-Werner-like methods for approx-
imating a locally unique solution x of equation
F(x) = 0, (14.1)
241
242 Iterative methods and their dynamics with applications: a contemporary study
xn+1 = xn − A−1
n F(xn )
(14.2)
yn+1 = xn+1 − A−1
n F(xn+1 ),
where x0 , y0 are initial points, An = [xn , yn ; F] and [x, y; F] denotes a divided dif-
ference of order one for operator F at points x, y ∈ Ω [2, 4, 7] satisfying
Suppose that
L(t1 + s0 ) 2L0t1
0< ≤ α ≤ 1− , (14.5)
1 − L0 (t1 + s1 + s0 ) 1 − L0 s0
where
s1 = t1 + L(t1 + s0 )t1 . (14.6)
Then, scalar sequence {tn } defined for each n = 1, 2, . . . by
−tn +sn −tn )(tn+1 −tn )
t0 = 0, sn+1 = tn+1 + L(tn+1
1−L0 (tn −t0 +sn +s0 ) , f or each n = 1, 2, . . . ,
L(tn+1 −tn +sn −tn )(tn+1 −tn ) (14.7)
tn+2 = tn+1 + 1−L0 (tn+1 −t0 +sn+1 +s0 ) , f or each n = 0, 1, 2, . . .
King-Werner-like methods free of derivatives 243
t1 ≤ t ≤ t . (14.9)
Throughout this chapter, we denote by U(w, ρ), U(w, ρ), the open and closed
balls in B1 , respectively, with center w ∈ B1 and of radius ρ > 0.
Now, using {tn } as a majorizing sequence we present our main semilocal
result.
A−1
0 ([x, y; F] − [z, v; F]) ≤ L(x − z + y − v), (14.17)
1
for each x, y, z ∈ Ω1 = Ω ∩U(x0 , )
2L0
U(x0 ,t ) ⊆ Ω (14.18)
and hypotheses of Lemma 14.2.1 hold, where A0 = [x0 ; y0 ; F] and t is given in
Lemma 14.2.1. Then, sequence {xn } generated by method (14.2) is well defined,
remains in U(x0 ,t ) and converges to a unique solution x ∈ U(x0 ,t ) of equation
F(x) = 0. Moreover, the following estimates hold for each n = 0, 1, 2, ...
xn − x ≤ t − tn . (14.19)
244 Iterative methods and their dynamics with applications: a contemporary study
Remark 14.3.2 (a) For the special case B1 = B2 = R, the radius of convergence
ball for method (14.2) is given in [10] by
s
ρ = , (14.36)
M
where s ≈ 0.55279 is a constant and M > 0 is the upper bound for
|F(x )−1 F (x)| in the given domain Ω. Using (14.31) we have
That is, we can choose l =Simply set l0 = l, we have from (14.33) that
M
2.
√
2 2(3 − 2) 0.63432 s
ρ= √ = ≈ > = ρ . (14.38)
(3 + 2)M 5M M M
Therefore, even in this special case, a bigger radius of convergence ball for
method (14.2) has been given in Theorem 14.3.1.
(b) Notice that we have
l0 ≤ l1 and l ≤ l1 (14.39)
A−1
0 ([x, y; F] − [z, u; F]) ≤ l1 (x − z + y − u) for each x, y, z, u ∈ Ω.
The radius given in [6]:
1
ρ0 = √ ≤ ρ. (14.40)
(1 + 2)l1 + l0
Moreover, if l < l1 , then ρ0 < ρ and the new error bounds (14.34) and (14.35)
are tighter than the old ones in [6] using ρ0 instead of ρ.
That is to say, the Lipschitz condition (14.31) and the center-Lipschitz condition
1
n using ρ using ρ0
new (14.35) old (14.35)
3 0.0498 0.0828
4 0.0031 0.0142
5 2.9778e-06 1.7305e-04
6 1.7108e-13 4.4047e-09
6 5.4309e-31 3.4761e-20
Hence the new results are more precise than the old ones in [6].
Example 14.4.2 Let B1 = B2 = C[0, 1], the space of continuous functions de-
fined on [0, 1], equipped with the max norm and Ω = U(0, 1). Define function F
on Ω, given by
1
F(x)(s) = x(s) − 5 stx3 (t)dt, (14.46)
0
and the divided difference of F is defined by
1
[x, y; F] = F (tx + (1 − t)y)dt. (14.47)
0
Then, we have
1
[F (x)y](s) = y(s) − 15 stx2 (t)y(t)dt, f or all y ∈ Ω. (14.48)
0
We have x (s) = 0 for all s ∈ [0, 1], l0 = 3.75 and l = l1 = 7.5. Using Theorem
248 Iterative methods and their dynamics with applications: a contemporary study
14.3.1, we can deduce that the radius of convergence ball for method (14.2) is
given by
1
ρ0 = ρ = √ ≈ 0.039052429. (14.49)
(1 + 2)l + 2l0
Example 14.4.3 Let B1 = B2 = C[0, 1] be equipped with the max norm and Ω =
U(0, r) for some r > 1. Define F on Ω by
1
F(x)(s) = x(s) − y(s) − μ G(s,t)x3 (t)dt, x ∈ C[0, 1], s ∈ [0, 1].
0
y ∈ C[0, 1] is given, μ is a real parameter and the Kernel G is the Green’s function
defined by
(1 − s)t if t ≤ s
G(s,t) =
s(1 − t) if s ≤ t.
Then, the Fréchet derivative of F is defined by
1
(F (x)(w))(s) = w(s) − 3μ G(s,t)x2 (t)w(t)dt, w ∈ C[0, 1], s ∈ [0, 1].
0
Let us choose x0 (s) = y0 (s) = y(s) = 1 and |μ| < 83 . Then, we have that
I − A0 ≤ 38 μ, A−1
0 ∈ L(B2 , B1 ),
−1 |μ| 3(1+r)|μ|
A0 ≤ 8−3|μ| , s0 = 0, t1 = 8−3|μ|
8
, L0 = 2(8−3|μ|) ,
and
3r|μ|
L= .
8 − 3|μ|
Let us choose r = 3 and μ = 12 . Then, we have that
and
L(t1 + s0 )
≈ 0.057441746, α ≈ 0.711345739,
1 − L0 (t1 + s1 + s0 )
2L0t1
1− ≈ 0.928994083.
1 − L0 s0
That is, condition (14.5) is satisfied and Theorem 14.2.2 applies.
References
Müller’s method
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
15.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
15.2 Convergence ball for method (15.1.2) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 252
15.3 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
15.1 Introduction
In this chapter we are concerned with approximating a solution of the equation
f (x) = 0, (15.1)
where f is defined on an open domain or closed domain D on a real space R.
The widely used [4] Müller’s method is defined in [2] by
2C
xn+1 = xn − n , n = 0, 1, 2, ..., x−2 , x−1 , x0 ∈ D, (15.2)
Bn ± B2n − 4AnCn
where,
An = f [xn , xn−1 , xn−2 ], Bn = f [xn , xn−1 ]+An (xn − xn−1 ), Cn = f (xn ), (15.3)
251
252 Iterative methods and their dynamics with applications: a contemporary study
and f [., .], f [., ., .] are divided differences of order one and two, respectively (see
[3]). The sign in the denominator of (15.2) is chosen so as to give the larger value.
It is a free-derivative method and has a convergence order 1.839 . . . under rea-
sonable conditions [2]. Xie in [5] established a semilocal convergence Theorem
of the method by means of bounding third and fourth derivatives. Bi et al. [6] pre-
sented a new semilocal convergence Theorem of the method using γ−condition.
Wu et al. [7] gave the convergence ball and error analysis of the method under
the hypotheses that the second-order and third-order derivative of function f are
bounded.
In this chapter, we will use a weaker condition to provide a new estimate
on the radius of convergence ball of Müller’s method. We will assume that f is
differentiable in D, f (x ) = 0, and the following Lipschitz conditions are true:
and
where,
2K
C= √ ,
4K 2 + 2KK
and {Fn } is Fibonacci sequence, and is defined by F−1 = 1, F0 = 1 and Fn+1 =
Fn + Fn−1 for any n = 0, 1, ....
Proof: We will use induction. First of all, denote en = x − xn (n = −2, −1, ...).
Let x−2 , x−1 , x0 ∈ U(x , R ) be distinct points. By (15.3) and (15.4), we have
| f (x )−1 A0 | = | f (x )−1 f [x0 , x−1 , x−2 ]| = | f (x )−1 f [x0 ,x−2x0]− f [x−1 ,x−2 ]
−x−1 | ≤ K.
(15.9)
Now, using x−1 , x0 ∈ U(x , R ), (15.3), (15.5) and (15.9), we get
|1 − f (x )−1 B0 | = |1 − f (x )−1 ( f [x0 , x−1 ] + A0 (e−1 − e0 ))|
= | f (x )−1 ( f [x , x ] − f [x , x0 ] + f [x0 , x ] − f [x0 , x−1 ] + A0 (e0 − e−1 ))|
≤ K |e0 | + K|e−1 | + K(|e0 | + |e−1 |) = (K + K )|e0 | + 2K|e−1 |
< (3K + K )R = √
3K+K
< 1,
5K+K +2 4K +2KK
2
(15.10)
which means
√
4K 2 +2KK
| f (x )−1 B0 | > 1 − (3K + K )R =
2K+2
√ >0 (15.11)
5K+K +2 4K 2 +2KK
and √
5K+K +2
√ 4K +2KK .
2
|B−1
0 f (x )| < 1
1−(3K+K )R = (15.12)
2K+2 4K 2 +2KK
Using x0 ∈ U(x , R ), f (x ) = 0, (15.3) and (15.5), we have
| f (x )−1C0 | = | f (x )−1 ( f (x0 ) − f (x ))| = | f (x )−1 ( f [x , x0 ]
− f [x , x ] + f [x , x ])e0 | (15.13)
≤ K |e0 |2 + |e0 | < K R + R .
2
(15.14)
Therefore, x1 can be defined. Denoting sign(s) as sign(s) = 1 if s ≥ 0 and
sign(s) = −1 if s < 0. Now, considering (15.2), we obtain
√
B −sign(B0 ) B20 −4A0C0
e1 = e0 + 2C√0 2 = e0 + 0 2A0
B0 +sign(B0 ) B0 −4A0C0
√
2A0 e0 +B0 −sign(B0 ) B20 −4A0C0 (2A0 e0 +B0 )2 −B20 +4A0C0
= 2A0 = √ 2 (15.15)
2A0 (2A0 e0 +B0 +sign(B0 ) B0 −4A0C0 )
2(A0 e20 +B0 e0 +C0 )
= √ ,
2A0 e0 +B0 +sign(B0 ) B20 −4A0C0
254 Iterative methods and their dynamics with applications: a contemporary study
so we have
2|A0 e20 +B0 e0 +C0 |
|e1 | = √ . (15.16)
|2A0 e0 +B0 +sign(B0 ) B20 −4A0C0 |
Next we shall show 2A0 e0 + B0 has the same sign as B0 . In fact, using (15.9),
(15.6) and (15.12), we get
2A0 e0 +B0
= 1 + 2 f (x )−1 A0 f (x )B−1 −1 −1
0 e0 ≥ 1 − 2| f (x ) A0 || f (x )B0 ||e0 |
B0 √
2K(5K+K +2 4K 2 +2KK )
> 1− √ R
√ 2K+2 4K 2 +2KK
4K 2 +2KK
= √ > 0.
K+ 2 4K +2KK
(15.17)
Hence, by (15.16), (15.3), (15.9), (15.4), (15.14), (15.11) and (15.17), we obtain
2|A0 e20 +B0 e0 +C0 |
|e1 | = √ = 2|(A0 e0 +B0 )e√0 − f [x ,x0 ]e0 |
|2A0 e0 +B0 |+ B20 −4A0C0 |2A0 e0 +B0 |+ B20 −4A0C0
2|A0 e−1 +A0 (e0 −e−1 )+ f [x0 ,x−1 ]−A0 (e0 −e−1 )− f [x0 ,x ]|
= √ |e0 |
|2A0 e0 +B0 |+ B20 −4A0C0
2| f (x )−1 (A0 e−1 + f [x0 ,x−1 ]− f [x0 ,x ])|
= 2A e +B
√ |e0 |
| f (x )−1 B0 | 0 B0 0 + ( f (x )−1 B0 )2 −4 f (x )−1 A0 f (x )−1C0
0
(15.18)
4K|e−1 |
≤ √
2
2K+2 4K +2KK
√
4K 2 +2KK
|e0 |
√ √
5K+K +2 √
4K 2 +2KK K+ 4K 2 +2KK
2K(5K+K +2 4K 2 +2KK )
= √ |e0 ||e−1 |.
4K 2 +2KK
Denote
ρn = C |eRn| , n = 0, 1, 2, ...., (15.21)
we obtain
ρn+1 ≤ ρn ρn−1 , n = 0, 1, 2, .... (15.22)
Clearly the following relation holds:
F
ρn+1 ≤ ρ0Fn ρ−1
n−1
, n = 0, 1, 2, .... (15.23)
Remark 15.2.2
(a) Condition (15.4) used in Theorem 15.2.1 is weaker than bounded conditions
of the second-order and third-order derivative of function f used in [7]. In
fact, suppose that f is twice differentiable on D, and
K ≤ K (15.26)
holds in general and KK can be arbitrarily large [1]. Therefore, the condition K ≤
K in Theorem 15.2.1 involves no loss of generality. If we use only condition
(15.4) in Theorem 15.2.1, we can establish a similar Theorem by replacing (15.6)
and (15.8) by
1
R = √ (15.27)
2(3 + 6)K
and
√
|x − xn+1 | ≤ ((4 + 2 6)K)Fn+1 −1 |x − x0 |Fn |x − x−1 |Fn−1 , f or any n = 0, 1, 2, ...,
(15.28)
respectively. On the other hand, if (15.26) holds strictly, it is clear that the fol-
lowing inequalities are true:
R < R (15.29)
and
√ C 2K(5K + K )
(4 + 2 6)K < = √ 2 + 4K, (15.30)
R 4K + 2KK
which means that we have bigger radius of convergence ball of Müller’s method
and tighter errors by using conditions (15.4) and (15.5) simultaneously instead
of using only condition (15.4).
256 Iterative methods and their dynamics with applications: a contemporary study
Then, we have x = 0,
x + 3, 0 ≤ x ≤ 1,
f (x) = |x| + 3 = (15.40)
−x + 3, −1 ≤ x < 0
and ⎧
⎨ 1, 0 < x ≤ 1,
f (x) = does not exist, x = 0, (15.41)
⎩
−1, −1 ≤ x < 0.
Notice that, for any x, y, u, v ∈ D, we have
1
| f (x )−1 ( f [x, y] − f [u, v])| = 13 | 0 ( f (tx + (1 − t)y) − f (tu + (1 − t)v))dt|
1
= 13 | 0 (|tx + (1 − t)y| − |tu + (1 − t)v|)dt|
1
≤ 13 0 ||tx + (1 − t)y| − |tu + (1 − t)v||dt
1
≤ 13 0 |tx + (1 − t)y − (tu + (1 − t)v)|dt
1
≤ 13 0 (t|x − u| + (1 − t)|y − v|)dt
= 16 (|x − u| + |y − v|).
(15.42)
That is, condition (15.4) holds for K = 6 but f (x) is not differentiable at 0 ∈ D.
1
References
258
Chapter 16
Generalized Newton
method with applications
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
16.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259
16.2 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 260
16.3 Semilocal convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 261
16.1 Introduction
In this chapter we are interested in the approximately solving the generalized
equation: Find x∗ ∈ H such that
0 ∈ F(x∗ ) + T (x∗ ). (16.1)
where F : H −→ H is a Fréchet differentiable function, H is a Hilbert space and
T : H ⇒ H is a set valued maximal monotone operator.
Importance of studying problem (16.1) and its applications in the physical
and engineering sciences and many other areas can be found in [1]– [33].
259
260 Iterative methods and their dynamics with applications: a contemporary study
for n = 1, 2, 3, . . ..
Using classical Lipschitz condition Uko [26, 27] established the convergence
of (16.2). The convergence (16.2) has been studied by many other authors under
various conditions (see Robinson [17], Josephy [12] (also see [18, 28])).
The convergence of generalized Newton method (16.2) was shown in [29,33]
using Lipschitz continuity conditions on F . However, there are problems where
the Lipschitz continuity of F does not hold. Motivated by this constrains, we
present a convergence analysis of the generalized Newton method (16.2) using
generalized continuity on F . Our results are weaker even, if we specialize the
conditions on F to the condition given in [23].
The rest of the chapter is organized as follows: In Section 16.2 we present
some needed preliminaries. Then, in Section 16.3 the convergence of generalized
Newton method (16.2) is presented.
16.2 Preliminaries
Let x ∈ H and r > 0. Throughout this chapter we will use U(x, r) and U(x, r) to
denote the open and closed metric ball, respectively, at x with radius r
(a) monotone if
u−v, x−y ≥ 0 for each y ∈ domT and v ∈ T (y) ⇒ x ∈ domT and u ∈ T (x).
(16.5)
Let G : H −→ H be a bounded linear operator. Then G ( := 1 (G + G∗ ) where
2
G∗ is the adjoint of G. Hereafter, we assume that T : H ⇒ H is a set valued
maximal monotone operator and F : H −→ H is a Fréchet differentiable operator.
at x̄ ∈ U(x̄, R), if
−1
x−x̄
F
(x̄)
F (x) − F (x̄) ≤ L0 (u)du,
0
−1
x−x̄+y−x
F
(x̄)
F (y) − F (x) ≤
L(u)du,
x−x̄
−1
x−x̄+y−x
F
(x̄)
F (y) − F (x) ≤
L1 (u)du
x−x̄
for each x, y ∈ U(x̄, R), with x − x̄ + y − x < R and L1 is a non-negative non-
decreasing integrable function on the interval [0, R].
262 Iterative methods and their dynamics with applications: a contemporary study
sn ≤ tn ≤ r1 (16.14)
Proof: The proof of part (a) and that sequence {tn } non-decreasingly converges
to r1 can be found in [29, 33]. Using simple induction (16.6), (16.7), (16.9),
(16.10) estimates (16.14) and (16.15) are obtained. Therefore, sequence {sn } is
non-decreasing and bounded above by r1 and as such it converges to s∗ which
satisfies (16.16) 2
We also need the following lemma.
Lemma 16.3.5 Let r < R0 . Let also x̄ ∈ H be such that F (x̄) is a positive op-
−1
erator and F (x̄) exists. Suppose that F
(x̄)F satisfies the center-Lipschitz
condition with L0 −average at x̄ ∈ U(x̄, r). Then, for each x ∈ U(x̄, r), F (x) is a
−1
positive operator, F (x̄) exists and
−1
−1 F (x̄)
F
(x) ≤ x−x̄ . (16.17)
1− 0 L0 (u)du
In view of (16.6), (16.17) and (16.18), estimate (16.17) is more precise than
(16.18), if L0 (u) < L1 (u).
It follows from (16.28) that sequence {xk } is complete in a Hilbert space H and
as such it converges to some x∗ ∈ U(x0 , s∗ ) (since U(x∗ , s∗ ) is a closed set).
Moreover, since T is a maximal monotone and F ∈ C1 , we deduce that x∗ solves
(16.1). Furthermore, (16.20) follows from (16.21) by using standard majorization
techniques [1, 4, 5]. 2
Next, we present the following result related to the uniqueness.
Proposition 16.3.8 Suppose that the hypotheses of Theorem 16.3.7 hold except
the center-Lipschitz (L0 , L) condition for x1 ∈ H such that
Remark 16.3.9 (a) Notice that function L1 is not needed in the proof of The-
orem 4.1 in [33], since it can be replaced by L0 which is more precise
than L1 (see (16.6)) in all stages of the proof. On the other hand, in view
of (16.16), we obtain a better information on the uniqueness of x∗ .
(b) The results obtained in this chapter can be improved even further, if in
Definition 16.3.2 we consider instead the center Lipschitz condition in
266 Iterative methods and their dynamics with applications: a contemporary study
−1
x−x̄+y−x
F
(x̄)
F (y) − F (x) ≤ K(u)du,
x−x̄
αn ≤ sn , (16.31)
L(tn − tn−1 )2
t0 = 0, t1 = β , tn+1 = tn + , (16.34)
2(1 − Ltn )
L(sn − sn−1 )2
s0 = 0, s1 = β , sn+1 = sn + , (16.35)
2(1 − L0 sn )
K(αn − αn−1 )2
α0 = 0, α1 = β , αn+1 = αn + . (16.36)
2(1 − L0 αn )
Generalized Newton method with applications 267
1 1 1
hk ≤ ⇒ h1 ≤ ⇒ h2 ≤ (16.40)
2 2 2
but not necessarily vice versa, unless, if L0 = L = K.
Examples, where L0 < K < L < L1 can be found in [6].
Hence, we have also extended the convergence domain in this case. Similar
advantages are obtained in the case of the Smale’s α−theory [23, 24] of Wang’s
γ−condition [30], [33].
References
Newton-secant methods
with values in a cone
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
17.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 271
17.2 Convergence of the Newton-secant method . . . . . . . . . . . . . . . . . . . . . . . . . 273
17.1 Introduction
We study the variational inclusion
271
272 Iterative methods and their dynamics with applications: a contemporary study
(b) Compute e : x − x1 ; x0 := x1 ; x1 := x.
3. Return x.
Remark 17.1.1 Notice that the continuity of the linear operator F (xm ) and G
being closed and convex, imply that the feasible set of (17.3) is a closed convex
set for all m. Hence, the existence of a feasible point x̄ implies that each solution
of (17.3) lies in the intersection of the feasible set of (17.3) with the closed ball of
center xm and radius x̄ − xm . Then, by [6–10] a solution of (17.3) exists, since
X is reflexive and the function x − xm is weakly lower semicontinuous.
The convergence of Newton-secant method (17.3) was shown in [19] using
Lipschitz continuity conditions on F and divided differences of order one and
two for G. However, there are problems where the Lipschitz continuity of F does
not hold or the divided differences of order two of G do not exist. Motivated
by these constrains, we present a convergence analysis of the Newton-secant
method (17.3) using generalized continuity on F and hypotheses only on the
divided difference of order one for G. Our results are weaker even, if we special-
ize the conditions on F to the condition given in [19]. This way we expand the
applicability of Newton-secant method (17.3).
Newton-secant methods with values in a cone 273
The rest of the chapter is organized as follows: In Section 17.2 we present the
convergence of Newton-secant method (17.3).
and
p(t) = w(t − a) + w2 (t,t − a,t − a). (17.8)
∗
Denote by t the smallest such zero. Then, scalar sequence {tn } defined by
t0 = 0, t1 = a,t2 ≥ b,
1
c
tn+1 = tn + [ w(θ (tn − tn−1 ))dθ
1 − cpn 0
+w1 (tn − tn−1 ,tn − tn−2 ,tn−1 − tn−2 )](tn − tn−1 ) (17.9)
b ≤ t ∗ ≤ t ∗∗ , (17.11)
where
pn = w(tn − a) + w2 (tn−1 ,tn−1 − a,tn ,tn−1 − t1 ),
c
q=
1 − cp
and
p = p(t ∗ ).
274 Iterative methods and their dynamics with applications: a contemporary study
0 ≤ t2 − t1 ≤ t1 − t0 (17.12)
t0 ≤ t1 ≤ t2 ≤ . . . ≤ tk ≤ tk+1
and
Q−1 (x0 , x1 ) ≤ c
and
x1 − x0 ≤ a.
Newton-secant methods with values in a cone 275
where sequence {tm } is defined by (17.9). By the second hypothesis in (ii), (iv)
and (17.9), we obtain the existence of x2 which solves (17.1) for m = 1, x1 −
x0 ≤ a = t1 − t0 , x2 − x1 ≤ b − a = t2 − t1 , which shows (17.17) for m = 0, 1.
Suppose that (17.17) holds for i = 3, . . . m, where x1 , x2 , . . . xm are defined by
(17.3). Then, we get that xi − x0 ≤ ti − t0 < t ∗ . That is xi ∈ U(x0 ,t ∗ ).
Now, it is easy that
that
Q(xi−1 , xi )x = (Q(x0 , x1 ) − Bi )x. (17.19)
By (ii), (iii), (17.19), Lemma 17.2.1 and the induction hypotheses we get
It follows from (17.21) and the Banach perturbation Lemma [13] that
and
Q−1 (x0 , x1 )
Q−1 (xi−1 , xi ) ≤
1 − A−1 (x0 , x1 )Bi
c
≤ . (17.22)
1 − cpi
The existence of xi solving (17.3) with i = k follows from the fact that Q(xi−1 , xi ) :
X −→ Y. Next, we must find a solution of
The right hand side of (17.23) is contained in the cone C, since xm solves (ii).
That is any x satisfying (17.23) is feasible for (17.3). Using (17.23), we can get
x as the solution of
The right hand side of (17.24) contains an element of least norm, so there exists
x̄ satisfying (17.22) and (17.23) such that
In view of (17.9), (iii), (17.22), (17.25) and the induction hypotheses we obtain
1
c
x̄ − xm ≤ [ w(θ xm − xm−1 )dθ xm − xm−1
1 − cpm 0
+[xm , xm−1 ; G] − [xm−2 , xm−1 ; G]xm − xm−1 ]
1
c
≤ [ w(θ xm − xm−1 )dθ
1 − cpm 0
+w1 (xm − xm−1 , xm − xm−2 , xm−1 − xm−2 )]xm − xm−1
1
c
≤ [ w(θ (tm − tm−1 ))dθ
1 − cpm 0
+w1 (tm − tm−1 ,tm − tm−2 ,tm−1 − tm−2 )]|tm − tm−1 |
= tm+1 − tm . (17.26)
By Lemma 17.2.1, sequence {tm } is complete. In view of (17.26), sequence {xm }
is complete in a Banach space X and
xm+1 − xm ≤ x̄ − xm ≤ tm+1 − tm .
2
Remark 17.2.3 (a) Our results can specialize to the corresponding ones in
[19]. Choose w(t) = Lt, w1 (t) = L1 and w2 (t,t,t,t) = L2 . But even in this
case, our results are weaker, since we do not use the hypothesis on the
divided difference of order two
[x, y, z; G] ≤ K
but use instead the second hypothesis in (iii) which involves only the di-
vided difference of order one. Moreover, the results in [19] cannot be used
to solve the numerical examples at the end of the chapter, since F is not
Lipschitz continuous. However, our results can apply to solve inclusion
problems. Hence, we have expanded the applicability of Newton-secant
method (17.3).
(b) Our results can be improved even further as follows: Let r0 be defined as
the smallest positive zero of equation
w2 (t,t,t,t) = 1.
Define D0 = D ∩U(x0 , r0 ). Suppose that the first and second hypotheses
in (iii) are replaced by
F (x) − F (y) ≤ w̄(x − y)
and
[x, y; G] − [z, y; G] ≤ w̄1 (x − z, x − y, z − y)
278 Iterative methods and their dynamics with applications: a contemporary study
[1] I. K. Argyros, S. George and Á. A. Magreñán, Local convergence for multi-
point-parametric Chebyshev-Halley-type methods of high convergence or-
der, J. Comput. Appl. Math., 282, 215–224, 2015.
[2] I. K. Argyros and S. George, Ball comparison between two optimal eight-
order methods under weak conditions, SeMA, 73 (1), 1–1, 2015.
[3] J. P. Aubin, Lipschitz behaviour of solutions to convex minimization prob-
lems, Math. Oper. Res., 9, 87–111, 1984.
[4] J. P. Aubin and H. Frankowska, Set-Valued Analysis, Systems Control
Found. Appl. Boston, MA, Birkhauser Boston, Inc.,1990.
[5] E. Catinas, On some iterative methods for solving nonlinear equations, Rev.
Anal. Nwner. Thdor. Approx., 23, 1, 47–53, 1994.
[6] A. L. Dontchev and W. W. Hager, An inverse mapping theorem for set
valued maps, Proc. Amer. Math. Soc., 121, 2, 481–489, 1994.
[7] A. L. Dontchev, Local convergence of the Newton method for generalized
equation, G. R. Acad. Sci., Paris, Ser. I, 322, 4, 327–331, 1996.
[8] A. L. Dontchev, Uniform convergence of the Newton method for Aubin
continuous maps, Serdica Math. J., 22, 3, 385–398, 1996.
[9] A. L. Dontchev, M. Quincampoix and N. Zlateva, Aubin Criterion for Met-
ric Regularity, J. Convex Anal., 13, 2, 281–297 (2006).
[10] A. L. Dontchev and R. T. Rockafellar, Implicit functions and solution map-
pings. Springer Monographs in Mathematics. New York, Springer (2009).
[11] M. H. Geoffroy and A. Pietrus, Local convergence of some iterative meth-
ods for generalized equations, J. Math. Anal Appl., 290, 2, 497–505, 2004.
279
280 References
Gauss-Newton method
with applications to
convex optimization
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
18.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 282
18.2 Gauss-Newton Algorithm and Quasi-Regularity condition . . . . . . . . . . . 283
18.2.1 Gauss-Newton Algorithm GNA . . . . . . . . . . . . . . . . . . . . . . . . . . . . 283
18.2.2 Quasi Regularity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 284
18.3 Semilocal convergence for GNA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 286
18.4 Specializations and numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291
18.1 Introduction
In this chapter we will study the convex composite optimizations problem.
We present a convergence analysis based on the majorant function presented
282
Gauss-Newton method with applications to convex optimization 283
Now, let ρ ∈ [1, ∞[, Δ ∈]0, ∞] and for each x ∈ IRn , define DΔ (x) by
Where d(x,W ) denotes the distance from x to W in the finite dimensional Banach
space containing W . It is worth noticing that the set DΔ (x) (x ∈ IRn ) is nonempty
and is the solution of the following convex optimization problem
min
n
h(F(x) + F (x) d), (18.5)
d∈IR , d≤Δ
which can be solved by well known methods such as the subgradient or cutting
plane or bundle methods (see, e.g., [12, 19, 25–27]). Throughout the chapter we
will denote U(x, r) as the open ball in Rn (or Rm ) centered at x and of radius
r > 0 and by U(x, r) its closure. Let W be a closed convex subset of Rn (or Rm ).
The negative polar of W denoted by W is defined as
In order for us to make the chapter as self contained as possible, the notion of
quasi-regularity is also re-introduced (see, e.g., [12, 17, 21]).
Proposition 18.2.2 (see, e.g., [12,17,21,25]) Let x0 be a regular point of (18.8).
Then, there are constants R0 > 0 and β > 0 such that (18.10) holds for R0 and
β (·) = β . Therefore, x0 is a quasi-regular point with the quasi-regular radius
Rx0 ≥ R0 and the quasi-regular bound function βx0 ≤ β on [0, R0 ].
Remark 18.2.3 (a) D(x) can be considered as the solution set of the lin-
earized problem associated to (18.8)
F(x) + F (x) d ∈ C. (18.12)
(b) If C defined in (18.8) is the set of all minimum points of h and if there
exists d0 ∈ D(x) with d0 ≤ Δ, then d0 ∈ DΔ (x) and for each d ∈ Rn , we
have the following equivalence
d ∈ DΔ (x) ⇐⇒ d ∈ D(x) ⇐⇒ d ∈ D∞ (x). (18.13)
(c) Let Rx0 denote the supremum of R0 such that (18.10) holds for some func-
tion β defined in Definition 18.2.1. Let R0 ∈ [0, Rx0 ] and BR (x0 ) denotes
the set of function β defined on [0, R0 ) such that (18.10) holds. Define
βx0 (t) = inf{β (t) : β ∈ BRx0 (x0 )} for each t ∈ [0, Rx0 ). (18.14)
All function β ∈ BR (x0 ) with lim β (t) < +∞ can be extended to an ele-
t→R−
ment of BRx0 (x0 ) and we have that
βx0 (t) = inf{β (t) : β ∈ BR (x0 )} for each t ∈ [0, R0 ). (18.15)
Rx0 and βx0 are called the quasi-regular radius and the quasi-regular func-
tion of the quasi-regular point x0 , respectively.
286 Iterative methods and their dynamics with applications: a contemporary study
Definition 18.3.1 Let R > 0, x0 ∈ IRn and F : IRn → IRm be continuously Fréchet-
differentiable. A twice-differentiable function f0 : [0, R) → R is called a center-
majorant function for F on U(x0 , R), if for each x ∈ U(x0 , R),
(h00 ) F (x) − F (x0 ) ≤ f0 (x − x0 ) − f0 (0);
(h01 ) f0 (0) = 0, f0 (0) = −1;
and
(h02 ) f0 is convex and strictly increasing.
Definition 18.3.2 [5,8,17] Let x0 ∈ IRn and F : IRn → IRm be continuously differ-
entiable. Define R0 = sup{t ∈ [0, R) : f0 (t) < 0}. A twice-differentiable function
f : [0, R0 ) → R is called a majorant function for F on U(x0 , R0 ), if for each
x, y ∈ U(x0 , R0 ), x − x0 + y − x < R0 ,
(h0 ) F (y) − F (x) ≤ f (y − x + x − x0 ) − f (x − x0 );
(h1 ) f (0) = 0, f (0) = −1;
and
(h2 ) f is convex and strictly increasing.
Let ρ > 0 and α > 0 be fixed and define auxiliary function ϕ : [0, R0 ) → R
by
ϕ(t) = ρ + (α − 1)t + α f (t). (18.16)
We shall use the following hypotheses
Gauss-Newton method with applications to convex optimization 287
(h4 ) there exists s∗ ∈ (0, R) such that for each t ∈ (0, s∗ ), ϕ(t) > 0 and ϕ(s∗ ) = 0;
(h5 ) ϕ(s∗ ) < 0.
From now on we assume the hypotheses (h0 ) − (h4 ) and (h00 ) − (h02 ) which
will be called the hypotheses (H).
Now, we provide the main convergence result of the Gauss-Newton method
generated by the Algorithm GNA for solving (18.1).
Theorem 18.3.4 Suppose that the (H) conditions are satisfied. Then,
(i) sequence {sk } generated by the Gauss-Newton method for s0 = 0, sk+1 = sk −
ϕ(sk )
ϕ (sk ) for solving equation ψ(t) = 0 is: well defined; strictly increasing;
remains in [0, s∗ ) and converges Q-linearly to s∗ .
Let η ∈ [1, ∞], Δ ∈ (0, ∞] and h : Rm → R be real-valued convex with
minimizer set C such that C = ∅.
(ii) Suppose that x0 ∈ Rn is a quasi-regular point of the inclusion
F(x) ∈ C,
with the quasi-regular radius rx0 and the quasi-regular bound function
βx0 defined by (18.2.14) and (18.2.15), respectively. If d(F(x0 ),C) > 0,
s∗ ≤ rx0 , Δ ≥ ρ ≥ ηβx0 (0)d(F(x0 ),C),
- .
ηβx0 (t) ∗
α ≥ sup :ρ ≤t <s
ηβx0 (t)(1 + f (t)) + 1
(b) Let us specialize conditions (18.8)–(18.10) even further in the case when
L0 , K and L are constant functions and α = 1. Then, function correspond-
ing to (18.16) reduce to
L
ψ(t) = t 2 − t + ρ (18.26)
2
[17, 23] and
K 2
ϕ(t) = t − t + ρ, (18.27)
2
respectively. In this case the convergence criteria become, respectively
1
h = Lρ ≤ (18.28)
2
and
1
h1 = Kρ ≤ . (18.29)
2
Notice that
1 1
h≤ =⇒ h1 ≤ (18.30)
2 2
but not necessarily vice versa. Unless, if K = L. Criterion (18.28) is the
celebrated Kantorovich hypotheses for the semilocal convergence of New-
ton’s method [20]. In the case of Wang’s condition [29] we have
γt 2
ϕ(t) = − t + ρ,
1 − γt
βt 2
ψ(t) = − t + ρ,
1 − βt
2γ 1
L(u) = , γ > 0, 0 ≤ t ≤
(1 − γu)3 γ
and
2β 1
K(u) = , β > 0, 0 ≤ t ≤
(1 − β u)3 β
with convergence criteria, given respectively by
√
H = γρ ≤ 3 − 2 2 (18.31)
√
H1 = β ρ ≤ 3 − 2 2. (18.32)
Then, we also have
√ √
H ≤ 3 − 2 2 =⇒ H1 ≤ 3 − 2 2
(c) Concerning the error bounds and the limit of majorizing sequence, let us
define majorizing sequence {rα,k } by
ϕ(rα,k )
rα,0 = 0; rα,k+1 = rα,k − (r
ϕα,0 α,k )
Suppose that
ϕ(r) ϕ(s)
− (r) ≤ − ϕ (s)
ϕα,0
for each r, s ∈ [0, R0 ] with r ≤ s. According to the proof of Theorem 18.3.1
sequence {rα,k } is also a majorizing sequence for GNA. On the other hand,
a simple inductive argument shows that
rk ≤ sk ,
rk+1 − rk ≤ sk+1 − sk
and
r∗ = lim rk ≤ s∗ .
k−→∞
ϕ(s) ψ(t)
−
≤− (18.33)
ϕ (s) ψ (t)
sα,k ≤ tα,k
Directional Newton
methods and restricted
domains
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
19.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 295
19.2 Semilocal convergence analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 298
19.1 Introduction
Let F : D ⊆ Rn → R be a differentiable function. In computer graphics, we often
need to compute and display the intersection C = A ∩ B of two surfaces A and
B in R3 [5], [6]. If the two surfaces are explicitly given by
295
296 Iterative methods and their dynamics with applications: a contemporary study
and
B = {(u, v, w)T : w = F2 (u, v)},
then the solution x = (u , v , w )T ∈ C must satisfy the nonlinear equation
F1 (u , v ) = F2 (u , v )
and
w = F1 (u , v ).
Therefore, we must solve a nonlinear equation in two variables x = (u, v)T of
the form
F(x) = F1 (x) − F2 (x) = 0.
The marching method can be used to compute the intersection C. In this
method, we first need to compute a starting point x0 = (u0 , v0 , w0 )T ∈ C, and
then compute the succeeding intersection points by succesive updating.
G(x) 2 = 0,
if the zero of function G is isolated or locally isolated and if the rounding error
is neglected [3], [7], [10], [11], [12].
F(xk )
xk+1 = xk − dk (k ≥ 0)
∇F(xk ) · dk
Now, we will explain DNM. First of all, we begin by choosing an initial guess
x0 ∈ U0 , where F is differentiable and a direction vector d0 .
Set θ0 = 0 to obtain the Newton iteration for f , that is the next point:
f (0)
v1 = − .
f (0)
The iteration for F is
F(x0 )
x1 = x0 − d0 .
∇F(x0 ) · d0
In [5], Levin and Ben–Israel provide a semilocal convergence analysis for the
DNM.
The quadratic convergence of the method was established for directions
dk sufficiently close to the gradients ∇F(xk ), and under standard Newton–
Kantorovich–type hypotheses [1]– [3], [7].
1. Weaker hypotheses;
2. Larger convergence domain for DNM;
3. Finer error bounds on the distances xk+1 − xk , xk − x (k ≥ 0);
4. An at least as precise information on the location of the zero x .
A numerical example where our results apply, but the earlier ones in [5] are
violated is also provided at the end of this chapter.
Here after, we use the Euclidean inner product, the corresponding norm
x , and the corresponding matrix norm A , except in Section 19.3 where
the ∞–norm is used for vectors and matrices, denoted by x ∞ , and A ∞ ,
respectively.
298 Iterative methods and their dynamics with applications: a contemporary study
Lemma 19.2.1 Suppose that there exist constants L0 > 0, L > 0 and η ≥ 0 such
that:
1
q0 := L η ≤ (19.1)
2
where,
1 √
L= 4 L0 + L0 L + 8 L0 + L0 L .
2 (19.2)
8
Then, sequence {tk } (k ≥ 0) given by
F(x0 ) = 0, ∇F(x0 ) = 0.
F(x0 )
h0 = − d0 ,
∇F(x0 ) · d0
x1 = x0 + h0 .
Directional Newton methods and restricted domains 299
(ii) For F ∈ C 2 [D], there exist constants M0 > 0 and M > 0 such that
1
p0 = |F(x0 )| M |∇F(x0 ) · d0 |−2 ≤ , (19.8)
2
and
B0 := B(x0 ,t ) ⊆ D (19.9)
where
1 √
M= 4 M0 + M0 M + 8 M0 + M0 M .
2 (19.10)
8
(iii) Sequence {xk } (k ≥ 0) given by
xk+1 = xk + hk , (19.11)
where,
F(xk )
hk = − dk , (19.12)
∇F(xk ) · dk
satisfies
∠(dk , ∇F(xk )) ≤ ∠(d0 , ∇F(x0 )) k ≥ 0, (19.13)
where, each dk ∈ Rn is such that dk = 1.
and
xk − x ≤ t − tk , (19.15)
where, iteration {tk } is given by (19.3), for
Proof: The iterates {xn } are shown to be in D0 which is more precise location
than D used in [1, 5], since D0 ⊆ D leading to advantages. We shall show using
mathematical induction on k ≥ 0:
and
B(xk+1 ,t − tk+1 ) ⊆ B(xk ,t − tk ). (19.18)
For each z ∈ B(x1 ,t − t1 ), we have
z − x0 ≤ z − x1 + x1 − x0
≤ t − t1 + t1 − t0 = t − t0 ,
and
xk + t (xk+1 − xk ) − x0 ≤ tk + t (tk+1 − tk ) ≤ t , t ∈ [0, 1].
Using condition (19.6) for x = xk , we obtain:
F(xk−1 )
= (dk−1 · ∇F(xk−1 )) + F(xk )
(F(xk−1 ) · dk−1 )
−F(xk−1 )
= F(xk ).
(19.21)
Directional Newton methods and restricted domains 301
M hk−1 2
≤
2 |∇F(xk ) · dk |
M hk−1 2 ∇F(x0 )
≤
2 ∇F(xk ) |∇F(x0 ) · d0 |
M hk−1 2 ∇F(x0 )
≤
2 ( ∇F(x0 ) −M0 tk ) |∇F(x0 ) · d0 |
M hk−1 2
≤
2 (1− ∇F(x0 ) −1 M0 tk ) |∇F(x0 ) · d0 |
M (tk − tk−1 )2
≤
2 (1 − |∇F(x0 ) · d0 |−1 M0 tk ) |∇F(x0 ) · d0 |
= tk+1 − tk ,
Lemma 19.2.1 implies that {tn } is a Cauchy sequence. It then follows from
(19.17), and (19.18) that {xn } is a Cauchy sequence too. Since B0 is a closed set,
it follows that {xn } converges to some x ∈ B0 .
Notice that
M0 ≤ M1 (19.25)
and
M ≤ M1 , ( since D0 ⊆ D) (19.26)
M1
hold in general, and can be arbitrarily large [1]– [3].
M0
Directional Newton methods and restricted domains 303
But, we have
M < N. (19.29)
That is,
1 1
p≤ =⇒ p0 ≤ , (19.30)
2 2
but not necessarily vice verca (unless if M0 = M = M1 ).
Moreover corresponding error bounds as well as the information on the location
of we improved if M0 < M or M < M1 .
Examples were M0 < M < M < M1 can be found in [1–4]. The rest of the
results in [1] (and consequently the results in [5] can be improved along the
same lines) It is worth noting that in the proof of Theorem 2.2 in [3], we used M1
instead of M (i.e., D instead of D0 ). However, the iterates xn remain in D0 which
is more precise domain than D, since D0 ⊆ D.
References
Expanding the
applicability of the
Gauss-Newton method for
convex optimization
under restricted
convergence domains and
majorant conditions
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
20.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 307
306
Gauss-Newton method for convex optimization 307
20.1 Introduction
In this chapter we are concerned with the convex composite optimizations prob-
lem. This work is mainly motivated by the work in [17,23]. We present a conver-
gence analysis of Gauss–Newton method (defined by Algorithm (GNA) in Sec.
20.2). The convergence of GNA is based on the majorant function in [17] (to
be precised in Sec. 20.2). They follow the same formulation using the majorant
function provided in [23] (see [21,23,28,29]). In [3,5,8], a convergence analysis
in a Banach space setting was given for (GNM) defined by
−1
xk+1 = xk − F (xk )+ F (xk ) F (xk )+ F(xk ) for each k = 0, 1, 2, . . . ,
where x0 is an initial point and F (x)+ in the Moore-Penrose inverse [11–13,
19, 26] of operator F (x) with F : Rn → Rm being continuously differentiable.
In [23], a semilocal convergence analysis using a combination of a majorant
and a center majorant function was given with the advantages (A): tighter error
estimates on the distances involved and the information on the location of the
solution is at least as precise. These advantages were obtained (under the same
computational cost) using same or weaker sufficient convergence hypotheses.
Here, we extend the same advantages (A) but to hold for GNA.
The chapter is organized as follows: Section 20.2 contains the definition of
GNA. In order for us to make the chapter as self contained as possible, the notion
of quasi-regularity is also re-introduced (see, e.g., [12, 17, 21]). The semilocal
convergence analysis of GNA is presented in Section 20.3. Numerical examples
and applications of our theoretical results and favorable comparisons to earlier
studies (see, e.g., [12, 17, 18, 21, 22]) are presented in Section 20.4.
25, 27]. We assume that the minimum hmin of the function h is attained. Problem
(20.1) is related to
F(x) ∈ C, (20.2)
where
C = argmin h (20.3)
is the set of all minimum points of h.
Let ξ ∈ [1, ∞[, Δ ∈]0, ∞] and for each x ∈ IRn , define DΔ (x) by
DΔ (x) = {d ∈ IRn : d ≤ Δ, h(F(x) + F (x) d) ≤ h(F(x) + F (x) d )
for all d ∈ IRn with d ≤ Δ}.
(20.4)
Let x0 ∈ IRn be an initial point. The Gauss-Newton algorithm GNA associated
with (ξ , Δ, x0 ) as defined in [12] (see also [17]) is as follows:
Algorithm GNA : (ξ , Δ, x0 )
Here, d(x,W ) denotes the distance from x to W in the finite dimensional Banach
space containing W . Note that the set DΔ (x) (x ∈ IRn ) is nonempty and is the
solution of the following convex optimization problem
min
n
h(F(x) + F (x) d), (20.5)
d∈IR , d≤Δ
which can be solved by well known methods such as the subgradient or cutting
plane or bundle methods (see, e.g., [12, 19, 25–27]).
Let U(x, r) denote the open ball in Rn (or Rm ) centered at x and of radius
r > 0. By U(x, r) we denote its closure. Let W be a closed convex subset of Rn
(or Rm ). The negative polar of W denoted by W is defined as
W = {z : < z, w >≤ 0 for each w ∈ W }. (20.6)
T −1 y = {x ∈ Rn : y ∈ T x} and A = inf a .
a∈A
Consider the inclusion
F(x) ∈ C, (20.8)
where C is a closed convex set in R . Let x ∈ R and
m n
(b) If C defined in (20.8) is the set of all minimum points of h and if there
exists d0 ∈ D(x) with d0 ≤ Δ, then d0 ∈ DΔ (x) and for each d ∈ Rn , we
have the following equivalence
d ∈ DΔ (x) ⇐⇒ d ∈ D(x) ⇐⇒ d ∈ D∞ (x). (20.13)
(c) Let Rx0 denote the supremum of R0 such that (20.10) holds for some func-
tion β defined in Definition 20.2.1. Let R0 ∈ [0, Rx0 ] and BR (x0 ) denotes
the set of function β defined on [0, R0 ) such that (20.10) holds. Define
βx0 (t) = inf{β (t) : β ∈ BRx0 (x0 )} for each t ∈ [0, Rx0 ). (20.14)
All function β ∈ BR (x0 ) with lim− β (t) < +∞ can be extended to an ele-
t→R
ment of BRx0 (x0 ) and we have that
βx0 (t) = inf{β (t) : β ∈ BR (x0 )} for each t ∈ [0, R0 ). (20.15)
Rx0 and βx0 are called the quasi-regular radius and the quasi-regular func-
tion of the quasi-regular point x0 , respectively.
310 Iterative methods and their dynamics with applications: a contemporary study
Definition 20.3.1 Let R > 0, x0 ∈ IRn and F : IRn → IRm be continuously Fréchet-
differentiable. A twice-differentiable function f0 : [0, R) → R is called a center-
majorant function for F on U(x0 , R), if for each x ∈ U(x0 , R),
(h00 ) F (x) − F (x0 ) ≤ f0 (x − x0 ) − f0 (0);
(h01 ) f0 (0) = 0, f0 (0) = −1;
and
(h02 ) f0 is convex and strictly increasing.
Definition 20.3.2 [5,8,17] Let x0 ∈ IRn and F : IRn → IRm be continuously differ-
entiable. Define R0 = sup{t ∈ [0, R) : f0 (t) < 0}. A twice-differentiable function
f : [0, R0 ) → R is called a majorant function for F on U(x0 , R0 ), if for each
x, y ∈ U(x0 , R0 ), x − x0 + y − x < R0 ,
(h0 ) F (y) − F (x) ≤ f (y − x + x − x0 ) − f (x − x0 );
(h1 ) f (0) = 0, f (0) = −1;
and
(h2 ) f is convex and strictly increasing.
(h4 ) there exists s∗ ∈ (0, R) such that for each t ∈ (0, s∗ ), ϕ(t) > 0 and ϕ(s∗ ) = 0;
(h5 ) ϕ(s∗ ) < 0.
From now on we assume the hypotheses (h0 ) − (h4 ) and (h00 ) − (h02 ) which
will be called the hypotheses (H).
Next, we present the main semi-local convergence result of the Gauss-
Newton method generated by the Algorithm GNA for solving (20.1).
Theorem 20.3.4 Suppose that the (H) conditions are satisfied. Then,
(i) sequence {sk } generated by the Gauss-Newton method for s0 = 0, sk+1 = sk −
ϕ(sk )
ϕ (sk ) for solving equation ψ(t) = 0 is: well defined; strictly increasing;
remains in [0, s∗ ) and converges Q-linearly to s∗ .
Let η ∈ [1, ∞], Δ ∈ (0, ∞] and h : Rm → R be real-valued convex with
minimizer set C such that C = ∅.
(ii) Suppose that x0 ∈ Rn is a quasi-regular point of the inclusion
F(x) ∈ C,
with the quasi-regular radius rx0 and the quasi-regular bound function βx0
defined by (20.14) and (20.15), respectively. If d(F(x0 ),C) > 0, s∗ ≤ rx0 ,
Δ ≥ ξ ≥ ηβx0 (0)d(F(x0 ),C),
- .
ηβx0 (t) ∗
α ≥ sup :ξ ≤t <s
ηβx0 (t)(1 + f (t)) + 1
Proof: Simply replace function g in [23] (see also [17]) by function f in the
proof, where g is a majorant function for F on U(x0 , R). That is we have instead
of (h0 ):
(h0 )
and
f (t) ≤ g (t) for each t ∈ [0, R0 ). (20.23)
Therefore, if
then the following advantages denoted by (A) are obtained: weaker suf-
ficient convergence criteria, tighter error bounds on the distances xn −
x∗ , xn+1 − xn and an at least as precise information on the location of
the solution x∗ . These advantages are obtained using less computational
cost, since in practice the computation of function g requires the compu-
tation of functions f0 and f as special cases. It is also worth noticing that
under (h00 ) function f0 is defined and therefore R0 which is at least as small
as R. Therefore the majorant function to satisfy (i.e., f ) is at least as small
as the majorant function satisfying (h0 ) (i.e., g ) leading to the advantages
Gauss-Newton method for convex optimization 313
of the new approach over the approach in [17] or [23]. Indeed, we have
that if function ψ has a solution t ∗ , then, since ϕ(t ∗ ) ≤ ψ(t ∗ ) = 0 and
ϕ(0) = ψ(0) = ξ > 0, we get that function ϕ has a solution r∗ such that
r∗ ≤ t ∗ . (20.25)
but not necessarily vice versa. If also follows from (20.25) that the new
information about the location of the solution x∗ is at least as precise as
the one given in [18].
(b) Let us specialize conditions (20.8)–(20.10) even further in the case when
L0 , K and L are constant functions and α = 1. Then, function correspond-
ing to (20.16) reduce to
L
ψ(t) = t 2 − t + ξ (20.26)
2
[17, 23] and
K 2
ϕ(t) = t −t +ξ, (20.27)
2
respectively. In this case the convergence criteria become, respectively
1
h = Lξ ≤ (20.28)
2
and
1
h1 = Kξ ≤ . (20.29)
2
Notice that
1 1
h≤ =⇒ h1 ≤ (20.30)
2 2
but not necessarily vice versa. Unless, if K = L. Criterion (20.28) is
the famous for its simplicity and clarity Kantorovich hypotheses for the
semilocal convergence of Newton’s method to a solution x∗ of equation
F(x) = 0 [20]. In the case of Wang’s condition [29] we have
γt 2
ϕ(t) = −t +ξ,
1 − γt
βt 2
ψ(t) = −t +ξ,
1 − βt
2γ 1
L(u) = , γ > 0, 0 ≤ t ≤
(1 − γu)3 γ
and
2β 1
K(u) = , β > 0, 0 ≤ t ≤
(1 − β u)3 β
314 Iterative methods and their dynamics with applications: a contemporary study
Suppose that
ϕ(r) ϕ(s)
− ≤−
ϕα,0 (r) ϕ (s)
for each r, s ∈ [0, R0 ] with r ≤ s. According to the proof of Theorem 20.3.1
sequence {rα,k } is also a majorizing sequence for GNA.
Moreover, a simple inductive argument shows that
rk ≤ sk ,
rk+1 − rk ≤ sk+1 − sk
and
r∗ = lim rk ≤ s∗ .
k−→∞
sα,k ≤ tα,k
Gauss-Newton method for convex optimization 315
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
21.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 319
21.2 Local convergence analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320
21.3 Numerical examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 326
21.1 Introduction
Consider the problem of approximating a locally unique solution x∗ of the non-
linear equation
F(x) = 0, (21.1)
where F is a Fréchet-differentiable operator defined on a convex subset D of
a Banach space X with values in a Banach space Y. The equation (21.1) cov-
ers wide range of problems in classical analysis and applications [1–30]. Closed
form solutions of these nonlinear equations exist only for few special cases which
319
320 Iterative methods and their dynamics with applications: a contemporary study
may not be of much practical value. Therefore solutions of these nonlinear equa-
tions (21.1) are approximated by iterative methods.
Newton-like methods are undoubtedly the most popular method for approxi-
mating a locally unique solution x∗ provided that the initial point is close enough
to the solution [1–30]. The number of function evaluations per step increases
with the order of convergence. In the scalar case the efficiency index [3, 6, 21]
1
EI = p m provides a measure of balance where p is the order of the method and
m is the number of function evaluations.
It is well known that according to the Kung-Traub conjuncture the conver-
gence of any multi-point method without memory cannot exceed the upper bound
2m−1 [21] (called the optimal order). Hence the optimal order for a method
with three function evaluations per step is 4. The corresponding efficiency in-
1
dex is EI = 4 3 = 1.58740... which is better than Newtons method which is
1
EI = 2 2 = 1.414.... Therefore, the study of new optimal methods of order four
is important.
In the present paper, we consider the following method:
1
yn = xn − F (xn )−1 F(xn )
2
1
zn = (4yn − xn )
3
un = yn + (F (xn ) − 3F (zn ))−1 F(xn ) (21.2)
vn = un + 2(F (xn ) − 3F (zn ))−1 F(un )
xn+1 = vn + 2(F (xn ) − 3F (zn ))−1 F(vn ).
Using hypotheses up to the first derivative of function F and contractions on a
Banach space setting we proved the convergence of the method. Moreover we
avoid Taylor expansions and use instead Lipschitz parameters. We do not have
to use higher order derivatives to show the convergence of method (21.2). This
way we expand the applicability of method (21.2).
The rest of the paper is organized as follows. In Section 21.2 we present the
local convergence analysis. We also provide a radius of convergence, computable
error bounds and uniqueness result not given in the earlier studies using Taylor
expansions [13, 18, 20]. Special cases and numerical examples are presented in
the concluding Section 21.3.
(b) There exists function w0 : [0, +∞) → [0, +∞) continuous, non-
decreasing with w0 (0) = 0 such that for each x ∈ D,
(c) Let R = max{r1 , r2 , r}. For r0 defined by (21.3), set D0 = D ∩ U(x∗ , r̄0 )
where r̄0 = max{r1 , r2 , r0 }. There exists functions w, v : [0, +∞) →
[0, +∞) continuous, non-decreasing with w(0) = 0 such that for each
x, y ∈ D0 ,
F (x∗ )−1 (F (x) − F (y) ≤ w(x − y)
and
F (x∗ )−1 F (x) ≤ v(x − x∗ ).
(d) U(x∗ , R) ⊆ D.
Then, the sequence {xn } generated for x0 ∈ U(x∗ , r) − {x∗ } is well de-
fined for each n = 0, 1, 2, . . . and converges to x∗ . Moreover, the following
estimates hold for each n = 0, 1, · · ·
It follows from (21.12) and the Banach Lemma on invertible operators [4, 5, 28]
that F (x)−1 ∈ L(Y, X) and
1
F (x)−1 F (x∗ ) ≤ . (21.13)
1 − w0 (x − x∗ )
Notice that x∗ + θ (x0 − x∗ ) − x∗ = θ x0 − x∗ < r, for each θ ∈ [0, 1], so x∗ +
θ (x0 − x∗ ) ∈ U(x∗ , r).
Then, using (21.4), (21.5) (for i = 3), (c), (21.13), (21.14) and (21.15), we get
in turn that
1
y0 − x∗ ≤ F (x0 )−1 F (x∗ ) 0 F (x∗ )−1 [F (x∗ + θ (x0 − x∗ ))
−F (x0 )]dθ x0 − x∗ + 12 F (x0 )−1 F(x∗ )F (x∗ )−1 F(x0 )
1
w((1−θ )x0 −x∗ )dθ + 12 01 v(θ x0 −x∗ )dθ
≤ 0
1−w0 (x0 −x∗ ) ,
≤ g1 (r)r = r1 ,
(21.16)
which shows (21.6) for n = 0 and y0 ∈ U(x∗ , r1 ).
Then, from the second sub-step of method (21.2) and (21.16) we also get that
z0 − x∗ = ∗ ∗
3 4(y0 − x ) + (x0 − x )
1
∗ ∗
≤ 3 [4(y0 − x ) + (x0 − x )]
1
(21.17)
∗ ∗
≤ 3 [4g1 (x0 − x ) + 1](x0 − x )
1
≤ g2 (r)r = r2 ,
which shows (21.17) for n = 0 and z0 ∈ U(x∗ , r2 ). Next, we must show that
A0 = (F (x0 ) − 3F (z0 ))−1 ∈ L(Y, X). Using (b), (21.4) and (21.17), we obtain in
324 Iterative methods and their dynamics with applications: a contemporary study
turn that
(−2F (x∗ ))−1 [A0 − F (x∗ )]
≤ 12 [F (x∗ )−1 [F (x0 ) − F (x∗ )] + 3F (x∗ )−1 (F (z0 ) − F (x∗ ))]
≤ 12 [w0 (x∗ − x0 ) + 3w0 (g2 (x∗ − x0 )x∗ − x0 )] = p(x∗ − x0 ) ≤ p(r) < 1,
(21.18)
so,
1
A−1 ∗
0 F (x ) ≤ . (21.19)
2(1 − p(x0 − x∗ ))
Hence, u0 , v0 and x1 are well defined. Then, using (21.4), (21.5), (21.13), (21.16),
(21.17), (c) and (21.19), we obtain in turn that
1
w((1−θ )(x0 −x∗ ))dθ x0 −x∗
≤ 0
1−w0 (x0 −x∗ )
[w0 (x0 −x∗ )+w0 (g2 (x0 −x∗ )x0 −x∗ ) 01 v(θ x0 −x∗ )dθ ]x0 −x∗
+ 34 (1−w0 (x0 −x ))(1−p(x0 −x∗ ))
∗
1
v(θ x0 −x∗ )dθ x0 −x∗
v0 − x∗ ≤ u0 − x∗ + 2 0
1−p(x0 −x∗ )
and
∗
2v0 −x
x1 − x∗ ≤ v0 − x∗ + 1−p(x ∗
0 −x )
so, Q−1 ∈ L(Y, X). Then, from the identity 0 = F(y∗ ) − F(x∗ ) = Q(y∗ − x∗ ), we
deduce that x∗ = y∗ . 2
Remark 21.2.2 (a) The radius r1 was obtained by Argyros in [4] as the con-
vergence radius for Newton’s method under condition (21.9)-(21.11). No-
tice that the convergence radius for Newton’s method given independently
by Rheinboldt [19] and Traub [21] is given by
2
ρ= < r1 .
3L
As an example, let us consider the function f (x) = ex − 1. Then x∗ = 0. Set
Ω = U(0, 1). Then, we have that L0 = e − 1 < l = e, so ρ = 0.24252961 <
r1 = 0.324947231.
Moreover, the new error bounds [2, 14] are:
L
xn+1 − x∗ ≤ xn − x∗ 2 ,
1 − L0 xn − x∗
whereas the old ones [26, 28]
L
xn+1 − x∗ ≤ xn − x∗ 2 .
1 − Lxn − x∗
Clearly, the new error bounds are more precise, if L0 < L. Clearly, we do
not expect the radius of convergence of method (21.2) given by r to be
larger than r1 (see (21.5)).
(b) The local results can be used for projection methods such as Arnoldi’s
method, the generalized minimum residual method (GMREM), the gen-
eralized conjugate method (GCM) for combined Newton/finite projection
methods and in connection to the mesh independence principle in order to
develop the cheapest and most efficient mesh refinement strategy [4–7].
326 Iterative methods and their dynamics with applications: a contemporary study
(c) The results can be also be used to solve equations where the operator F 0
satisfies the autonomous differential equation [5, 7, 15, 17]:
F 0 (x) = P(F(x)),
where P : Y →Y is a known continuous operator. Since F 0 (x∗ ) = P(F(x∗ ))
= P(0), we can apply the results without actually knowing the solution x∗ .
Let as an example F(x) = ex − 1. Then, we can choose P(x) = x + 1 and
x∗ = 0.
(d) It is worth noticing that method (21.2) are not changing if we use the
new instead of the old conditions [23]. Moreover, for the error bounds in
practice we can use the computational order of convergence (COC)
ln kxkxn+2 −xn+1 k
n+1 −xn k
ξ= kxn+1 −xn k
, for each n = 1, 2, . . .
ln kx n −xn−1 k
Notice that using the (21.11) conditions, we get w0 (t) = L0t, w(t) = Lt, v(t) = L,
1
L0 = e − 1, L = e L0 . The parameters are
Example 21.3.2 Let X = Y = C[0, 1], the space of continuous functions defined
on [0, 1] and be equipped with the max norm. Let D = U(0, 1). Define function F
on D by
1
F(ϕ)(x) = ϕ(x) − 5 xθ ϕ(θ )3 dθ . (21.25)
0
We have that
1
F (ϕ(ξ ))(x) = ξ (x) − 15 xθ ϕ(θ )2 ξ (θ )dθ , for each ξ ∈ D.
0
Then, we get that x∗ = 0, w0 (t) = L0t, w(t) = Lt, v(t) = 2, L0 = 7.5, L = 15. The
parameters for method are
[24] F. A. Potra and V. Pták, Nondiscrete Induction and Iterative Processes, in:
Research Notes in Mathematics, Vol. 103, Pitman, Boston, 1984.
[25] H. Ren and Q. Wu, Convergence ball and error analysis of a family of
iterative methods with cubic convergence, Appl. Math. Comput., 209, 369–
378, 2009.
[26] W. C. Rheinboldt, An adaptive continuation process for solving systems of
nonlinear equations, In: Mathematical models and numerical methods (A.
N. Tikhonov et al. eds.) pub. 3, 129–142 Banach Center, Warsaw Poland,
1977.
[27] J. R. Sharma, P. K. Guha and R. Sharma, An efficient fourth order weighted-
Newton method for systems of nonlinear equations, Numer. Algorithms, 62
(2), 307–323, 2013.
[28] J. F. Traub, Iterative methods for the solution of equations, AMS Chelsea
Publishing, 1982.
[29] S. Weerakoon and T. G. I. Fernando, A variant of Newton’s method with
accelerated third-order convergence, Appl. Math. Let., 13, 87–93, 2000.
[30] T. Zhanlar, O. Chulunbaatar and G. Ankhbayar, On Newton-type methods
with fourth and fifth-order convergence, Bulletin of PFUR Series Mathe-
matics. Information Sciences, Physics, 2 (2), 29–34, 2010.
Chapter 22
Expanding Kantorovich’s
theorem for solving
generalized equations
Ioannis K. Argyros
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505,
USA, Email: iargyros@cameron.edu
CONTENTS
22.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 331
22.2 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 332
22.3 Semilocal convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333
22.1 Introduction
In [18], G. S. Silva considered the problem of approximating the solution of the
generalized equation
F(x) + Q(x) 0, (22.1)
where F : D −→ H is a Fréchet differentiable function, H is a Hilbert space
with inner product ., . and corresponding norm ., D ⊆ H an open set and T :
331
332 Iterative methods and their dynamics with applications: a contemporary study
F(x) + ∂ ψ(x) 0,
has been studied by several authors [1]– [24]. In [13], Kantorovich obtained a
convergence result for Newton’s method for solving the equation F(x) = 0 under
some assumptions on the derivative F (x0 ) and F (x0 )−1 F(x0 ). Kantorovich,
used the majorization principle to prove his results. Later in [16], Robinson con-
sidered generalization of the Kantorovich theorem of the type F(x) ∈ K, where
K is a nonempty closed and convex cone, and obtained convergence results and
error bounds for this method. Josephy [12], considered a semilocal Newton’s
method of the kind (22.3) in order to solving (22.1) with F = NC the normal
cone mapping of a convex set C ⊂ R2 .
The rest of this paper is organized as follows. Preliminaries are given in Sec-
tion 22.2 and the main results are presented in the concluding Section 22.3.
22.2 Preliminaries
Let U(x, ρ) and Ū(x, ρ) stand respectively for open and closed balls in H with
center x ∈ H and radius ρ > 0. The following Definitions and Lemmas are used
for proving our results. These items are stated briefly here in order to make the
paper as self contains as possible. More details can be found in [18].
Definition 22.2.1 Let D ⊆ H be an open nonempty subset of H, h : D −→ H be
a Fréchet differentiable function with Fréchet derivative h and Q : H ⇒ H be
a set mapping. The partial linearization of the mapping h + Q at x ∈ H is the
set-valued mapping Lh (x, .) : H ⇒ H given by
Lemma 22.2.4 ( [22]) Let G be a positive operator (i.e.,G(x), x ≥ 0). The fol-
lowing statements about G hold:
G2 = G2 ;
If G−1 exists, then G−1 is a positive operator.
Lemma 22.2.5 ( [21, Lemma 2.2]) Let G be a positive operator. Suppose G−1
exists, then for each x ∈ H we have
x2
G(x), x ≥ .
G−1
F̂ (x0 )−1 F (x) − F (x0 ) ≤ f0 (x − x0 ) − f0 (0). (22.7)
f0 (tn0 )
0
tn+1 = tn0 − , n = 0, 1, 2, . . .
f0 (tn0 )
is strictly increasing, remains in (0,t0∗ ) and converges to t0∗ , where t0∗ is the small-
est zero of function f0 in (0, R). Furthermore, suppose that for each x, y ∈ D1 :=
Ū(x0 , ρ) ∩U(x0 ,t0∗ ) there exists f1 : [0, ρ1 ) −→ R, ρ1 = min{ρ,t0∗ } such that
F̂ (x0 )−1 F (y) − F (x) ≤ f1 (x − y + x − x0 ) − f1 (x − x0 )
(22.8)
and x1 − x0 ≤ f1 (0)
f1 (tn1 ) f1 (sn )
1
tn+1 = tn1 − (or sn+1 = sn − )
f1 (tn )
1 f0 (sn )
are well defined, {tn1 } (or sn ) is strictly increasing, remains in (0,t1∗ ) and
converges to t1∗ . Moreover, sequence {xn } generated by generalized Newton’s
method (22.3) is well defined, remains in U(x0 ,t1∗ ) and converges to a point x∗ ∈
Ū(x0 ,t1∗ ), which is the unique solution of generalized equation F(x) + Q(x) 0
in Ū(x0 ,t1∗ ). Furthermore, the following estimates hold:
t1∗ − tn+1
1
xn+1 − x∗ ≤ xn − x∗ 2
(t1∗ − tn1 )2
t1∗ − s1n+1
xn+1 − x∗ ≤ xn − x∗ 2
(t1∗ − s1n )2
sn+1 − sn ≤ tn+1
1
− tn1 ,
Expanding Kantorovich’s theorem for solving generalized equations 335
D− f1 (t1∗ )
xn+1 − x∗ ≤ xn − x∗ 2 ,
−2 f1 (t1∗ )
D− f1 (t1∗ )
xn+1 − x∗ ≤ xn − x∗ 2 ,
−2 f0 (t1∗ )
D− f1 (t1∗ ) ∗
t1∗ − tn+1 ≤ (t − tn )2 ,
−2 f1 (t1∗ ) 1
and
D− f1 (t1∗ ) ∗
t1∗ − sn+1 ≤ (t − sn )2 ,
−2 f0 (t1∗ ) 1
where D− stands for the left directional derivative of function f1 .
Remark 22.3.2 (a) Suppose that there exists f : [0, R) −→ R twice continu-
ously differentiable such that for each x, y ∈ U(x0 , ρ)
leading to tn1 ≤ tn , sn ≤ tn ,
1
tn+1 − tn1 ≤ tn+1 − tn (22.13)
f (tn )
t0 = 0, tn+1 = tn − ,
f (tn )
f1 (t) f (s)
− ≤− (22.16)
f1 (t) f (s)
or
f1 (t) f (s)
− ≤− , (22.17)
f0 (t) f (s)
respectively for each t ≤ s. Estimates (22.13) and (22.14) can be strict if
(22.16) and (22.17) hold as strict inequalities.
(b) So far we have improved the error bounds and the location of the solution
x∗ but not necessarily the convergence domain of the generalized New-
ton’s method (22.3). We can also show that convergence domain can be
improved in some interesting special cases. Let F ≡ {0},
L
f (t) = t 2 − t + η,
2
L0 2
f0 (t) = t −t +η
2
and
L1 2
f1 (t) =
t − t + η,
2
where x1 − x0 ≤ η and L, L0 and L1 are Lipschitz constants satisfying:
L0 ≤ L
and
L1 ≤ L.
The corresponding majorizing sequences are
Proof of Theorem 22.3.1 Simply notice that the iterates {xn } ∈ D1 which is a
more precise location than Ū(x0 , ρ) used in [18], since D1 ⊆ Ū(x0 , ρ). Then,
the definition of function f1 becomes possible and replaces f in the proof [18],
whereas for the computation on the upper bounds F̂ (x)−1 we use the more
precise f0 than f as it is shown in the next perturbation Banach lemma [13].
2
Lemma 22.3.3 Let x0 ∈ D be such that F̄ (x0 ) is a positive operator and
F̄ (x0 )−1 exists. If x−x0 ≤ t < t ∗ , then F̄ (x) is a positive operator and F̄ (x)−1
exists. Moreover,
F̄ (x0 )−1
F̄ (x)−1 ≤ . (22.19)
f0 (t)
Proof: Observe that
1 1
F̄ (x)− F̄ (x0 ) ≤ F̄ (x)− F̄ (x0 )+ (F̄ (x)− F̄ (x0 ))∗ = F̄ (x)− F̄ (x0 ).
2 2
(22.20)
Let x ∈ Ū(x0 ,t), 0 ≤ t < t ∗ . Thus f (t) < 0. Using, (h11 ) and (h12 ), we obtain that
F̄ (x0 )−1 F̄ (x) − F̄ (x0 ) ≤ F̄ (x0 )−1 F̄ (x) − F̄ (x0 )
≤ f0 (x − x0 ) − f0 (0)
< f0 (t) + 1 < 1. (22.21)
Hence, there is no guarantee under the results in [18] that Newton’s method con-
√
verges to the solution x∗ = 3 p. However, our convergence criteria are satisfied:
1 1
q1 = L1 η ≤ for each p ∈ (0.4619831630, ). (22.27)
2 2
and
1 1
q2 = Kη ≤ for each p ∈ (0.2756943786, ). (22.28)
2 2
In order for us to compare the majorizing sequences {tn }, {sn }, {rn }, let p = 0.6.
Then, old condition (22.26) as well as new conditions (22.27) and (22.28) are
satisfied. Then, we present the following table showing that the new error bounds
are tighter. It is worth noticing that these advantages are obtained under the same
computational cost since the evaluation of the old Lipschitz constants L requires
the computation of the new constants L0 , L1 and K as special cases.
303, 307, 308, 311–315, 320–322, 326, 327, Euclidean inner product 298
332–334, 337–341 Euler-Chebyshev method 90
convergence analysis 2, 3, 25, 27, 46, 50, 91, 93, exact Newton iteration map 130
121, 145, 149, 177–179, 184, 187, 206, 217, Ezquerro et al. method 218
219, 227, 228, 232, 243, 261, 263, 273, 283,
284, 296, 298, 299, 308, 320–322, 334 F
convergence ball 246–249, 252, 253, 256
Fatou set 72, 98
convergence hypotheses 284, 308
Fibonacci sequence 246, 254
convergence improvement 109, 111, 114
¿QLWHGLPHQVLRQDO%DQDFKVSDFH
convex composite optimization 283, 284, 308
¿[HGSRLQWSUREOHPV
convex composite optimization problems 284,
¿[HGSRLQWWKHRU\
308
fractional conversion 237
convex optimization 283, 285, 307, 309
Fréchet-derivative 121, 173, 178, 189, 212, 328
convex subset 1, 5, 11, 12, 177, 217, 219, 227,
Fréchet-differentiable operator 1, 25, 26, 31, 109,
228, 232, 243, 285, 309, 320
120, 177, 184, 186, 187, 207, 217, 219, 228,
convex type 195, 333
243, 244, 246, 262, 284, 308, 320, 322
critical point 63, 72, 74–79, 98
full rank 147, 148, 150, 151, 153, 155, 157, 158
cutting method 285, 309
fundamental theorem of integration 165, 170
D
G
Derivative free 179, 242, 253
Gauss-Legendre quadrature formula 41
different multiplicity 13, 15
Gauss-Newton Algorithm 283, 284, 308, 309
differentiable operator 1, 25, 26, 31, 109, 120,
Gauss-Newton method 146, 153, 283, 284, 288,
177, 184, 186, 187, 207, 217, 219, 228, 243,
307, 308, 312
244, 246, 262, 284, 308, 320, 322
generalized conjugate method 326
directional derivative 198, 298, 336
Generalized equations 194, 205, 332
Directional Newton methods 296
generalized minimum residual methods 326
discrepancy principal 167
Generalized Newton method 260, 261
discrepancy principle 48, 167
geometric construction 23, 140, 292, 316
discretization process 41, 233
Green’s kernel 11
divided difference 218, 227, 232, 233, 243, 248,
Green’s function 190, 249
253, 272, 273, 278
GLYLGHGGLIIHUHQFHRI¿UVWRUGHU
H
Dynamical study 75
dynamics 63, 71, 72, 90, 91, 98 Halley’s method 1–3, 5, 8, 11, 13–22, 179
Hammerstein integral equation of the second kind
E 11, 39
Hilbert space 47, 163, 194, 206, 260, 266, 332
HI¿FLHQF\LQGH[
Holder type 167
HI¿FLHQWPHVKUH¿QHPHQWVWUDWHJ\
Eighth order method 320
I
element of least norm 277
equality-constraint optimization problem 297 ill-posed equations 46, 162
error 3, 39, 40, 42, 46, 48, 50, 55, 56, 66–68, 70, ill-posed problems 46, 164
116, 121, 123, 125, 127, 128, 135–137, 140, implicit functions 202, 213, 268, 279, 341
162–164, 166, 167, 195, 199, 206, 209, 211, inclusion 132, 209, 272, 273, 275, 285, 288, 310,
212, 219, 247, 248, 253, 256, 284, 286, 289, 312
291, 297, 298, 304, 308, 310, 313, 315, 321, Inexact Gauss-Newton method 146, 153
326, 327, 333, 337, 341 Inexact Gauss-Newton-like method 145, 146, 153
error bounds 3, 39, 40, 46, 50, 55, 116, 125, 137, Inexact Newton’s method 120, 123, 126, 131,
140, 166, 195, 199, 209, 212, 219, 247, 286, 135, 137, 139
289, 291, 298, 304, 310, 313, 315, 321, 326, initial points 2, 10, 90, 110, 123, 179, 197, 218,
327, 333, 337, 341 243, 265, 273, 284, 285, 308, 309, 321, 335
error estimates 42, 121, 162, 163, 166, 206, 209, inner product 47, 163, 194, 298, 332
219, 248, 284, 308 integral expression 95
Index 345
interior point method 142, 203, 269, 342 majorizing sequence 26, 27, 29, 32, 34, 38, 41,
intermediate value theorem 92, 180, 322 42, 50, 54, 114, 115, 138, 177, 179, 200,
invariant Fatou component 74 219, 243–245, 263, 274, 291, 299, 311, 315,
invertible operators 5, 12, 28, 29, 33, 113, 124, 338, 340, 341
184, 187, 201, 221, 226, 324, 339 mathematical induction 8, 29, 93, 152, 156, 221,
iterative method 47, 63, 65, 66, 69–71, 75, 97, 230, 265, 301, 323
98, 110, 120, 121, 177, 178, 321 mathematical modelling 2
iterative regularization method 47 mathematical programming 297
iteratively regularized Lavrentiev method 47, 49 matrix form 41
matrix norm 298
J maximal monotone 194, 196, 206, 207, 260, 262,
265, 266, 333, 334
Julia set 72, 98
maximal monotone operator 207, 260, 262, 265
max-norm 11, 39, 58, 233, 235
K
mean value theorem 28
Kantorovich hypothesis 39, 110, 114, 138, 201, mesh independence principle 326
339 minimization algorithms 141, 293, 317
Kantorovich theorem 26, 110, 121, 195, 292, 316, monotone process of convex and concave type
332, 333 281
Kantorovich theory 40, 109, 111, 113, 115, 135, monotone processes 281
292, 316 monotonically decreasing 263
Kantorovich-type convergence criterion 116 monotonically increasing 48, 55, 149, 164, 263
k-Fréchet-differentiable 1, 26, 120, 136, 146, Moore-Penrose inverses 146
147, 177, 184, 186, 187, 217–219, 228, 243, Müller’s method 252, 253, 256
244, 246, 284, 308, 320
King-Werner-like methods 242 N
King-Werner-type method 177, 179, 182, 184,
necessary conditions 235
186, 187, 189
negative polar 285, 309
Kung and Traub method 65
Newton iteration mapping 211
Kung-Traub conjuncture 321
Newton-Kantorovich hypotheses 290, 314
Kurchatov method 218
Newton-Kantorovich theorem 26, 292, 316
Newton-secant cone 273
L
Newton-Secant method 272–274, 278
Laguerre’s method 90 Newton’s method 25–27, 29, 31, 39–41, 90,
Lavrentiev regularization 46, 162, 163, 167 110–115, 121, 137, 139, 146, 178, 179, 189,
Lavrentiev regularization method 162 194, 195, 197, 199, 201, 205–207, 219, 263,
Least-change Secant update methods 305 265, 267, 290, 292, 314, 316, 326, 333–335,
library ARPREC 71 337, 339, 341
linearized problem 286, 310 nitrogen-hydrogen feed 237
Lipschitz parameters 321 noisy data 47, 163
Lipschitz-type conditions 146, 219 nonlinear Hammerstein integral equation of the
local analysis 213, 328 second kind 11, 39
local convergence 89–91, 93, 145, 146, 149, 150, Nonlinear Ill-posed equations 46
178, 179, 187, 205–207, 217–219, 228, 232, nonlinear ill-posed operator equation 163
242, 243, 246, 308, 311, 312, 320–322 nonlinear least square problems 146
local convergence analysis 91, 93, 145, 149, 178, nonlinear monotone operators 46
187, 217, 219, 228, 232, 243, 320–322 norm of the maximum of the rows 236
local convergence study 206, 218, 219, 243 normal cone mapping 195, 333
local study 2 normed convex process 273
logarithmic type 167 numerical examples 1, 3, 10, 25, 27, 37, 39, 46,
50, 57, 90, 91, 102, 146, 162–164, 173, 177,
M 179, 189, 206, 217, 219, 233, 242, 243, 247,
252, 253, 257, 278, 283, 284, 292, 298, 308,
majorant function 121, 124, 125, 133–137, 196, 316, 320, 321, 327
210, 283, 284, 287, 289, 308, 311, 313, 134
346 Iterative Methods and Their Dynamics with Applications: A Contemporary Study
U W
uniqueness ball 11, 13 :DQJ¶VȖWKHRU\
uniqueness of solution 25, 32 weak center Lipschitz condition 146
weak Lipschitz condition 147
V weaker conditions 48, 253
valued operator 196, 206, 207, 261, 272, 273, 334 Werner method 177–179, 181, 184, 186, 187,
variational inequalities 202, 203, 268, 270, 341, 189, 242
342
variational inequality problem 195, 206, 333