• Tidak ada hasil yang ditemukan

On Newton-Kantorovich Method for Solving the Nonlinear Operator Equation

N/A
N/A
Protected

Academic year: 2024

Membagikan "On Newton-Kantorovich Method for Solving the Nonlinear Operator Equation"

Copied!
13
0
0

Teks penuh

(1)

Research Article

On Newton-Kantorovich Method for Solving the Nonlinear Operator Equation

Hameed Husam Hameed,

1,2

Z. K. Eshkuvatov,

3,4

Anvarjon Ahmedov,

4,5

and N. M. A. Nik Long

1,4

1Department of Mathematics, Faculty of Science, Universiti Putra Malaysia (UPM), Selangor, Malaysia

2Technical Institute of Alsuwerah, The Middle Technical University, Baghdad, Iraq

3Faculty of Science and Technology, Universiti Sains Islam Malaysia (USIM), Negeri Sembilan, Malaysia

4Institute for Mathematical Research, Universiti Putra Malaysia (UPM), Selangor, Malaysia

5Department of Process and Food Engineering, Faculty of Engineering, Universiti Putra Malaysia (UPM), Selangor, Malaysia

Correspondence should be addressed to Z. K. Eshkuvatov; [email protected] Received 19 July 2014; Accepted 6 October 2014

Academic Editor: Gaohang Yu

Copyright © 2015 Hameed Husam Hameed et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

We develop the Newton-Kantorovich method to solve the system of2×2nonlinear Volterra integral equations where the unknown function is in logarithmic form. A new majorant function is introduced which leads to the increment of the convergence interval.

The existence and uniqueness of approximate solution are proved and a numerical example is provided to show the validation of the method.

1. Introduction

Nonlinear phenomenon appears in many scientific areas such as physics, fluid mechanics, population models, chem- ical kinetics, economic systems, and medicine and can be modeled by system of nonlinear integral equations. The difficulty lies in finding the exact solution for such system.

Alternatively, the approximate or numerical solutions can be sought. One of the well known approximate method is Newton-Kantorovich method which reduces the nonlinear into sequence of linear integral equations. The the approxi- mate solution is then obtained by processing the convergent sequence. In 1939, Kantorovich [1] presented an iterative method for functional equation in Banach space and derived the convergence theorem for Newton method. In 1948, Kantorovich [2] proved a semilocal convergence theorem for Newton method in Banach space, later known as the Newton-Kantorovich method. Uko and Argyros [3] proved a weak Kantorovich-type theorem which gives the same conclusion under the weaker conditions. Shen and Li [4] have

established the Kantorovich-type convergence criterion for inexact Newton methods, assuming that the first derivative of an operator satisfies the Lipschitz condition. Argyros [5]

provided a sufficient condition for the semilocal convergence of Newton’s method to a locally unique solution of a nonlinear operator equation. Saberi-Nadjafi and Heidari [6] introduced a combination of the Newton-Kantorovich and quadrature methods to solve the nonlinear integral equation of Urysohn type in the systematic procedure. Ezquerro et al. [7] studied the nonlinear integral equation of mixed Hammerstein type using Newton-Kantorovich method with majorant principle.

Ezquerro et al. [8] provided the semilocal convergence of Newton method in Banach space under a modification of the classic conditions of Kantorovich. There are many methods of solving the system of nonlinear integral equations, for example, product integration method [9], Adomian method [10], RBF network method [11], biorthogonal system method [12], Chebyshev wavelets method [13], analytical method [14], reproducing kernel method [15], step method [16], and single term Wlash series [17]. In 2003, Boikov and Tynda [18]

Volume 2015, Article ID 219616, 12 pages http://dx.doi.org/10.1155/2015/219616

(2)

implemented the Newton-Kantorovich method to the follow- ing system:

𝑥 (𝑡) − ∫𝑡

𝑦(𝑡)ℎ (𝑡, 𝜏) 𝑔 (𝜏) 𝑥 (𝜏) 𝑑𝜏 = 0,

𝑡

𝑦(𝑡)𝑘 (𝑡, 𝜏) [1 − 𝑔 (𝜏)] 𝑥 (𝜏) 𝑑𝜏 = 𝑓 (𝑡) ,

(1)

where0 < 𝑡0 ≤ 𝑡 ≤ 𝑇,𝑦(𝑡) < 𝑡, and the functionsℎ(𝑡, 𝜏), 𝑘(𝑡, 𝜏) ∈ 𝐶[𝑡0,𝑇]×[𝑡0,𝑇],𝑓(𝑡),𝑔(𝑡) ∈ 𝐶[𝑡0,𝑇], and(0 < 𝑔(𝑡) < 1).

In 2010, Eshkuvatov et al. [19] used the Newton-Kantorovich hypothesis to solve the system of nonlinear Volterra integral equation of the form

𝑥 (𝑡) − ∫𝑡

𝑦(𝑡)ℎ (𝑡, 𝜏) 𝑥2(𝜏) 𝑑𝜏 = 0,

𝑡

𝑦(𝑡)𝑘 (𝑡, 𝜏) 𝑥2(𝜏) 𝑑𝜏 = 𝑓 (𝑡) ,

(2)

where 𝑥(𝑡) and 𝑦(𝑡) are unknown functions defined on [𝑡0, ∞), 𝑡0 > 0, andℎ(𝑡, 𝜏), 𝑘(𝑡, 𝜏) ∈ 𝐶[𝑡0,∞]×[𝑡0,∞],𝑓(𝑡) ∈ 𝐶[𝑡0,∞]. In 2010, Eshkuvatov et al. [20] developed the modified Newton-Kantorovich to obtain an approximate solution of system with the form

𝑥 (𝑡) − ∫𝑡

𝑦(𝑡)𝐻 (𝑡, 𝜏) 𝑥𝑛(𝜏) 𝑑𝜏 = 0,

𝑡

𝑦(𝑡)𝐾 (𝑡, 𝜏) 𝑥𝑛(𝜏) 𝑑𝜏 = 𝑓 (𝑡) ,

(3)

where0 < 𝑡0 ≤ 𝑡 ≤ 𝑇, 𝑦(𝑡) < 𝑡, and the functions𝐻(𝑡, 𝜏), 𝐾(𝑡, 𝜏) ∈ 𝐶[𝑡0,∞]×[𝑡0,∞],𝑓(𝑡) ∈ 𝐶[𝑡0,∞], and the unknown functions𝑥(𝑡) ∈ 𝐶[𝑡0,∞],𝑦(𝑡) ∈ 𝐶1[𝑡0,∞],𝑦(𝑡) < 𝑡.

In this paper, we consider the systems of nonlinear integral equation of the form

𝑥 (𝑡) − ∫𝑡

𝑦(𝑡)ℎ (𝑡, 𝜏)log|𝑥 (𝜏)| 𝑑𝜏 = 𝑔 (𝑡) ,

𝑡

𝑦(𝑡)𝑘 (𝑡, 𝜏)log|𝑥 (𝜏)| 𝑑𝜏 = 𝑓 (𝑡) ,

(4)

where0 < 𝑡0 ≤ 𝑡 ≤ 𝑇,𝑦(𝑡) < 𝑡,𝑥(𝑡) ̸= 0,ℎ(𝑡, 𝜏), ℎ𝜏(𝑡, 𝜏), 𝑘(𝑡, 𝜏), 𝑘𝜏(𝑡, 𝜏) ∈ 𝐶(𝐷) and the unknown functions𝑥(𝑡) ∈ 𝐶[𝑡0, 𝑇],𝑦(𝑡) ∈ 𝐶1[𝑡0, 𝑇]to be determined, and𝐷 = [𝑡0, 𝑇] × [𝑡0, 𝑇].

The paper is organized as follows, inSection 2, Newton- Kantorovich method for the system of integral equations(4) is presented. Section 3deals with mixed method followed by discretizations. In Section 4, the rate of convergence of the method is investigated. Lastly,Section 5demonstrates the numerical example to verify the validity and accuracy of the proposed method, followed by the conclusion inSection 6.

2. Newton-Kantorovich Method for the System

Let us rewrite the system of nonlinear Volterra integral equa- tion(4)in the operator form

𝑃 (𝑋) = (𝑃1(𝑋) , 𝑃2(𝑋)) = 0, (5)

where𝑋 = (𝑥(𝑡), 𝑦(𝑡))and

𝑃1(𝑋) = 𝑥 (𝑡) − ∫𝑡

𝑦(𝑡)ℎ (𝑡, 𝜏)log|𝑥 (𝜏)| 𝑑𝜏 − 𝑔 (𝑡) , 𝑃2(𝑋) = ∫𝑡

𝑦(𝑡)𝑘 (𝑡, 𝜏)log|𝑥 (𝜏)| 𝑑𝜏 − 𝑓 (𝑡) .

(6)

To solve(5)we use initial iteration of Newton-Kantorovich method which is of the form

𝑃󸀠(𝑋0) (𝑋 − 𝑋0) + 𝑃 (𝑋0) = 0, (7)

where𝑋0 = (𝑥0(𝑡), 𝑦0(𝑡))is the initial guess and𝑥0(𝑡) and 𝑦0(𝑡) can be any continuous functions provided that 𝑡0 <

𝑦(𝑡) < 𝑡and𝑥(𝑡) ̸= 0.

The Frechet derivative of𝑃(𝑋)at the point𝑋0is defined as

𝑃󸀠(𝑋0) 𝑋

= (lim

𝑠 → 0

1

𝑠[𝑃1(𝑋0+ 𝑠𝑋) − 𝑃1(𝑋)] ,

𝑠 → 0lim 1

𝑠[𝑃2(𝑋0+ 𝑠𝑋) − 𝑃2(𝑋)])

= (lim

𝑠 → 0

1

𝑠[𝑃1(𝑥0+ 𝑠𝑥, 𝑦0+ 𝑠𝑦) − 𝑃1(𝑥0, 𝑦0)] ,

𝑠 → 0lim 1

𝑠[𝑃2(𝑥0+ 𝑠𝑥, 𝑦0+ 𝑠𝑦) − 𝑃2(𝑥0, 𝑦0)])

= (lim

𝑠 → 0[𝜕𝑃1(𝑥0, 𝑦0)

𝜕𝑥 𝑠𝑥 +𝜕𝑃1(𝑥0, 𝑦0)

𝜕𝑦 𝑠𝑦

+1 2(𝜕2𝑃1

𝜕𝑥2 (𝑥0+ 𝜃𝑠𝑥, 𝑦0+ 𝛿𝑠𝑦) 𝑠2𝑥2 + 2𝜕2𝑃1

𝜕𝑥𝜕𝑦(𝑥0+ 𝜃𝑠𝑥, 𝑦0+ 𝛿𝑠𝑦) 𝑠2𝑥𝑦 + 𝜕2𝑃1

𝜕𝑦2 (𝑥0+ 𝜃𝑠𝑥, 𝑦0+ 𝛿𝑠𝑦) 𝑠𝑦2)] ,

(3)

𝑠 → 0lim 1 𝑠[𝜕𝑃2

𝜕𝑥 (𝑥0, 𝑦0) 𝑠𝑥 +𝜕𝑃2

𝜕𝑦 (𝑥0, 𝑦0) 𝑠𝑦 +1

2(𝜕2𝑃2

𝜕𝑥2 (𝑥0+ 𝜃𝑠𝑥, 𝑦0+ 𝛿𝑠𝑦) 𝑠2𝑥2 + 2𝜕2𝑃2

𝜕𝑥𝜕𝑦(𝑥0+ 𝜃𝑠𝑥, 𝑦0+ 𝛿𝑠𝑦) 𝑠2𝑥𝑦 +𝜕2𝑃2

𝜕𝑦2 (𝑥0+ 𝜃𝑠𝑥, 𝑦0+ 𝛿𝑠𝑦) 𝑠𝑦2)])

= (𝜕𝑃1(𝑥0, 𝑦0)

𝜕𝑥 𝑥 + 𝜕𝑃1(𝑥0, 𝑦0)

𝜕𝑦 𝑦,

𝜕𝑃2(𝑥0, 𝑦0)

𝜕𝑥 𝑥 + 𝜕𝑃2(𝑥0, 𝑦0)

𝜕𝑦 𝑦) .

(8) Hence,

𝑃󸀠(𝑋0) 𝑋 = (

𝜕𝑃1

𝜕𝑥󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)

𝜕𝑃1

𝜕𝑦󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)

𝜕𝑃2

𝜕𝑥󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)

𝜕𝑃2

𝜕𝑦󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)

) (𝑥𝑦). (9)

From(7)and(9)it follows that

𝜕𝑃1

𝜕𝑥󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)(Δ𝑥 (𝑡)) + 𝜕𝑃1

𝜕𝑦󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)(Δ𝑦 (𝑡))

= −𝑃1(𝑥0(𝑡) , 𝑦0(𝑡)) ,

𝜕𝑃2

𝜕𝑥󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)(Δ𝑥 (𝑡)) + 𝜕𝑃2

𝜕𝑦󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)(Δ𝑦 (𝑡))

= −𝑃2(𝑥0(𝑡) , 𝑦0(𝑡)) ,

(10)

whereΔ𝑥(𝑡) = 𝑥1(𝑡) − 𝑥0(𝑡),Δ𝑦(𝑡) = 𝑦1(𝑡) − 𝑦0(𝑡), and(𝑥0(𝑡), 𝑦0(𝑡))is the initial given functions. To solve(10)with respect toΔ𝑥andΔ𝑦we need to compute all partial derivatives:

𝜕𝑃1

𝜕𝑥󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)=lim

𝑠 → 0

1

𝑠(𝑃1(𝑥0+ 𝑠𝑥, 𝑦0) − 𝑃1(𝑥0, 𝑦0))

=lim

𝑠 → 0

1 𝑠[𝑠𝑥 (𝑡)

− ∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏) (log󵄨󵄨󵄨󵄨𝑥0(𝜏) + 𝑠𝑥 (𝜏)󵄨󵄨󵄨󵄨

− log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨) 𝑑𝜏]

= 𝑥 (𝑡) − ∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏) 𝑥 (𝜏) 𝑥0(𝜏)𝑑𝜏,

𝜕𝑃1

𝜕𝑦󵄨󵄨󵄨󵄨

󵄨󵄨󵄨󵄨(𝑥0,𝑦0)=lim

𝑠 → 0

1

𝑠(𝑃1(𝑥0, 𝑦0+ 𝑠𝑦) − 𝑃1(𝑥0, 𝑦0))

=lim

𝑠 → 0

1

𝑠[∫𝑦0(𝑡)+𝑠𝑦(𝑡)

𝑦0(𝑡) ℎ (𝑡, 𝜏)log󵄨󵄨󵄨󵄨(𝑥0(𝜏))󵄨󵄨󵄨󵄨 𝑑𝜏]

= ℎ (𝑡, 𝑦0(𝑡))log󵄨󵄨󵄨󵄨𝑥0(𝑦0(𝑡))󵄨󵄨󵄨󵄨 𝑦 (𝑡) ,

(11) and in the same manner we obtain

𝜕𝑃2

𝜕𝑥󵄨󵄨󵄨󵄨

󵄨󵄨󵄨󵄨(𝑥0,𝑦0)= ∫𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏) 𝑥 (𝜏) 𝑥0(𝜏)𝑑𝜏,

𝜕𝑃2

𝜕𝑦󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨(𝑥0,𝑦0)= −𝑘 (𝑡, 𝑦0(𝑡))log󵄨󵄨󵄨󵄨𝑥0(𝑦0(𝑡))󵄨󵄨󵄨󵄨 𝑦 (𝑡) .

(12)

So that from(10)–(12)it follows that Δ𝑥 (𝑡) − ∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏)Δ𝑥 (𝜏) 𝑥0(𝜏)𝑑𝜏

+ ℎ (𝑡, 𝑦0(𝑡))log󵄨󵄨󵄨󵄨𝑥0(𝑦0(𝑡))󵄨󵄨󵄨󵄨 Δ𝑦 (𝑡)

= ∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 − 𝑥0(𝑡) + 𝑔 (𝑡) ,

𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏)Δ𝑥 (𝜏) 𝑥0(𝜏)𝑑𝜏

− 𝑘 (𝑡, 𝑦0(𝑡))log󵄨󵄨󵄨󵄨𝑥0(𝑦0(𝑡))󵄨󵄨󵄨󵄨 Δ𝑦 (𝑡)

= − ∫𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 + 𝑓 (𝑡) .

(13)

Equation(13)is a linear, and, by solving it forΔ𝑥andΔ𝑦, we obtain(𝑥1(𝑡), 𝑦1(𝑡)). By continuing this process, a sequence of approximate solution(𝑥𝑚(𝑡), 𝑦𝑚(𝑡))can be evaluated from 𝑃󸀠(𝑋0) Δ𝑋𝑚+ 𝑃 (𝑋𝑚) = 0, (14) which is equivalent to the system

Δ𝑥𝑚(𝑡) − ∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏 + ℎ (𝑡, 𝑦0(𝑡))log󵄨󵄨󵄨󵄨𝑥0(𝑦0(𝑡))󵄨󵄨󵄨󵄨 Δ𝑦𝑚(𝑡)

= ∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 − 𝑥0(𝑡) + 𝑔 (𝑡) ,

𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏

− 𝑘 (𝑡, 𝑦0(𝑡))log󵄨󵄨󵄨󵄨𝑥0(𝑦0(𝑡))󵄨󵄨󵄨󵄨 Δ𝑦𝑚(𝑡)

= − ∫𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 + 𝑓 (𝑡) ,

(15)

whereΔ𝑥𝑚(𝑡) = 𝑥𝑚(𝑡)−𝑥𝑚−1(𝑡)andΔ𝑦𝑚(𝑡) = 𝑦𝑚(𝑡)−𝑦𝑚−1(𝑡), 𝑚 = 1, 2, 3, . . ..

(4)

Thus, one should solve a system of two linear Volterra integral equations to find each successive approximation.

Let us eliminateΔ𝑦(𝑡)from the system(13) by finding the expression ofΔ𝑦(𝑡)from the first equation of this system and substitute it in the second equation to yield

Δ𝑦 (𝑡) = 1 𝐻 (𝑡)[∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏) [Δ𝑥 (𝜏)

𝑥0(𝜏) +log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨] 𝑑𝜏

− [Δ𝑥 (𝑡) + 𝑥0(𝑡) − 𝑔 (𝑡)]] ,

𝐺 (𝑡) [∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏) [Δ𝑥 (𝜏)

𝑥0(𝜏) +log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨] 𝑑𝜏

− [Δ𝑥 (𝑡) + 𝑥0(𝑡) − 𝑔 (𝑡)]]

= ∫𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏)Δ𝑥 (𝜏) 𝑥0(𝜏)𝑑𝜏

− ∫𝑡

𝑦0(𝑡)𝑘 (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 + 𝑓 (𝑡) ,

(16) where 𝐺(𝑡) = 𝑘(𝑡, 𝑦0(𝑡))/ℎ(𝑡, 𝑦0(𝑡)) and 𝐻(𝑡) = 1/[ℎ(𝑡, 𝑦0(𝑡))log|𝑥0(𝑦0(𝑡))|], and the second equation of(16)yields

Δ𝑥 (𝑡) − ∫𝑡

𝑦0(𝑡)𝑘1(𝑡, 𝜏)Δ𝑥 (𝜏)

𝑥0(𝜏)𝜏 = 𝐹0(𝑡) , (17) where

𝑘1(𝑡, 𝜏) = ℎ (𝑡, 𝜏) −𝑘 (𝑡, 𝜏) 𝐺 (𝑡) , 𝐺 (𝑡) = 𝑘 (𝑡, 𝑦0(𝑡))

ℎ (𝑡, 𝑦0(𝑡)), 𝑘 (𝑡, 𝑦0(𝑡)) ̸= 0 ∀𝑡 ∈ [𝑡0, 𝑇] , 𝐹0(𝑡) = ∫𝑡

𝑦0(𝑡)𝑘1(𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥0(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 − 𝑥0(𝑡) + 𝑔 (𝑡) + 𝑓 (𝑡) 𝐺 (𝑡).

(18) In an analogous way,Δ𝑦𝑚(𝑡)andΔ𝑥𝑚(𝑡)can be written in the form

Δ𝑦𝑚(𝑡)

= 1

𝐻 (𝑡)[∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏 + ∫𝑡

𝑦𝑚−1(𝑡)ℎ (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏

− Δ𝑥𝑚(𝑡) − 𝑥𝑚−1(𝑡) + 𝑔 (𝑡)] ,

(19)

Δ𝑥𝑚(𝑡) − ∫𝑡

𝑦0(𝑡)𝑘1(𝑡, 𝜏)Δ𝑥𝑚(𝜏)

𝑥0(𝜏) 𝑑𝜏 = 𝐹𝑚−1(𝑡) , (20)

where

𝐹𝑚−1(𝑡) = ∫𝑡

𝑦𝑚−1(𝑡)𝑘1(𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 − 𝑥𝑚−1(𝑡) + 𝑔 (𝑡) +𝑓 (𝑡)

𝐺 (𝑡).

(21)

3. The Mixed Method (Simpson

and Trapezoidal) for Approximate Solution

At each step of the iterative process we have to find the solution of(18)and(20)on the closed interval[𝑡0, 𝑇]. To do this the grid (𝜔) of points𝑡𝑖 = 𝑡0+ 𝑖ℎ,𝑖 = 1, 2, 3, . . . , 2𝑁, ℎ = (𝑇−𝑡0)/2𝑁is introduced, and by the collocation method with mixed rule we require that the approximate solution satisfies(18)and(20). Hence

Δ𝑥𝑚(𝑡0) = −𝑥𝑚−1(𝑡0) + 𝑔 (𝑡0) +𝑓 (𝑡0)

𝐺 (𝑡0), (22) Δ𝑥𝑚(𝑡2𝑖) − ∫𝑡2𝑖

𝑦0(𝑡2𝑖)𝑘1(𝑡2𝑖, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏

= 𝐹𝑚−1(𝑡2𝑖) , 𝑖 = 1, 2, . . . , 𝑁.

(23)

On the grid (𝜔) we setV2𝑖= 𝑦0(𝑡2𝑖), suct that

𝑡V2𝑖 = {𝑡V2𝑖, 𝑡0≤ 𝑦0(𝑡2𝑖) < 𝑡2𝑖−2,

𝑡2𝑖, 𝑡2𝑖−2≤ 𝑦0(𝑡2𝑖) < 𝑡2𝑖. (24)

Consequently, the system(23)can be written in the form

Δ𝑥𝑚(𝑡2𝑖) − ∫𝑡V2𝑖

𝑦0(𝑡2𝑖)𝑘1(𝑡2𝑖, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏

− ∑𝑖−1

𝑗=V2𝑖𝑡2𝑗+2

𝑡2𝑗

𝑘1(𝑡2𝑖, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏

= 𝐹𝑚−1(𝑡2𝑖) , 𝑖 = 1, 2, . . . , 𝑁.

(25)

By computing the integral in(26)using tapezoidal formula on the first integrals and Simpson formula on the second integral, we consider two cases.

Case 1.WhenV2𝑖 ̸= 2𝑖,𝑖 = 1, 2, . . . , 𝑁, then

Δ𝑥𝑚(𝑡2𝑖) = 𝐹𝑚−1(𝑡2𝑖) + 𝐴 (𝑖) + 𝐵 (𝑖) + 𝐶 (𝑖) 1 − ((𝑡2𝑖− 𝑡2𝑖−2) /6 ⋅ 𝑥0(𝑡2𝑖)) 𝑘1(𝑡2𝑖, 𝑡2𝑖),

(26)

(5)

where

𝐴 (𝑖) = 0.5 (𝑡V2𝑖− 𝑦0(𝑡2𝑖))

× [𝑘1(𝑡2𝑖, 𝑡V2𝑖)Δ𝑥𝑚(𝑡V2𝑖)

𝑥0(𝑡V2𝑖) + 𝑘1(𝑡2𝑖, 𝑦0(𝑡2𝑖))

× Δ𝑥𝑚(𝑡V2𝑖) (𝑡V2𝑖− 𝑦0(𝑡2𝑖)) (𝑡V2𝑖− 𝑡V2𝑖−2) (𝑥0(𝑦0(𝑡2𝑖))) + 𝑘1(𝑡2𝑖, 𝑦0(𝑡2𝑖))

×Δ𝑥𝑚(𝑡V2𝑖−2) (𝑦0(𝑡2𝑖) − 𝑡V2𝑖−2) (𝑡V2𝑖− 𝑡V2𝑖−2) (𝑥0(𝑦0(𝑡2𝑖))) ] , 𝐵 (𝑖) = 𝑖−2

𝑗=V2𝑖

(𝑡2𝑗+2− 𝑡2𝑗) 6

× [𝑘1(𝑡2𝑖, 𝑡2𝑗)Δ𝑥𝑚(𝑡2𝑗) 𝑥0(𝑡2𝑗)

+ 4𝑘1(𝑡2𝑖, 𝑡2𝑗+1)Δ𝑥𝑚(𝑡2𝑗+1) 𝑥0(𝑡2𝑗+1) + 𝑘1(𝑡2𝑖, 𝑡2𝑗+2)Δ𝑥𝑚(𝑡2𝑗+2)

𝑥0(𝑡2𝑗+2) ] , 𝐶 (𝑖) =(𝑡2𝑖− 𝑡2𝑖−2)

6 [𝑘1(𝑡2𝑖, 𝑡2𝑖−2)Δ𝑥𝑚(𝑡2𝑖−2) 𝑥0(𝑡2𝑖−2) + 4𝑘1(𝑡2𝑖, 𝑡2𝑖−1)Δ𝑥𝑚(𝑡2𝑖−1)

𝑥0(𝑡2𝑖−1) ] . (27)

Case 2.WhenV2𝑖= 2𝑖,𝑖 = 1, 2, . . . , 𝑁, then

Δ𝑥𝑚(𝑡2𝑖) =𝐷1(𝑖)

𝐷2(𝑖), (28)

where

𝐷1(𝑖) = 𝐹𝑚−1(𝑡2𝑖) + 0.5𝑘1(𝑡2𝑖, 𝑦0(𝑡2𝑖))

× [Δ𝑥𝑚(𝑡2𝑖−2) 𝑥0(𝑦0(𝑡2𝑖))

(𝑡2𝑖− 𝑦0(𝑡2𝑖)) (𝑦0(𝑡2𝑖) − 𝑡2𝑖−2) 𝑡2𝑖− 𝑡2𝑖−2 ] , 𝐷2(𝑖) = [1 − 0.5 (𝑡2𝑖− 𝑦0(𝑡2))𝑘1(𝑡2𝑖, 𝑡2𝑖)

𝑥0(𝑡2𝑖)

− 0.5𝑘1(𝑡2𝑖, 𝑦0(𝑡2𝑖)) (𝑡2𝑖− 𝑦0(𝑡2𝑖))2 𝑥0(𝑦0(𝑡2𝑖)) (𝑡2𝑖− 𝑡2𝑖−2)] .

(29)

Also, to computeΔ𝑦𝑚(𝑡) on the grid (𝜔), (18) can be re- presented in the form

Δ𝑦𝑚(𝑡2𝑖) = 1 𝐻 (𝑡2𝑖)

× [∫𝑡2𝑖

𝑦0(𝑡2𝑖)ℎ (𝑡2𝑖, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏 + ∫𝑡2𝑖

𝑦𝑚−1(𝑡2𝑖)ℎ (𝑡2𝑖, 𝜏)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏

− Δ𝑥𝑚(𝑡2𝑖) − 𝑥𝑚−1(𝑡2𝑖) + 𝑔 (𝑡2𝑖) ] . (30)

Let us setV2𝑖= 𝑦0(𝑡2𝑖)and𝑢2𝑖= 𝑦𝑚−1(𝑡2𝑖)and

𝑡V2𝑖 ={ {{

𝑡2𝑖, 𝑡2𝑖−2 ≤ 𝑦0(𝑡2𝑖) < 𝑡2𝑖, 𝑡V2𝑖, 𝑡0≤ 𝑦0(𝑡2𝑖) < 𝑡2𝑖−2,

𝑡𝑢2𝑖={ {{

𝑡2𝑖, 𝑡2𝑖−2≤ 𝑦𝑚−1(𝑡2𝑖) < 𝑡2𝑖, 𝑡𝑢2𝑖, 𝑡0≤ 𝑦𝑚−1(𝑡2𝑖) < 𝑡2𝑖−2.

(31)

Then(30)can be written as

Δ𝑦𝑚(𝑡2𝑖) = 1 𝐻 (𝑡2𝑖)

× [∫𝑡V2𝑖

𝑦0(𝑡2𝑖)ℎ (𝑡2𝑖, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏 + ∑𝑖−1

𝑗=V2𝑖

𝑡2𝑗+2

𝑡2𝑗

ℎ (𝑡2𝑖, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏 + ∫𝑡𝑢2𝑖

𝑦𝑚−1(𝑡2𝑖)ℎ (𝑡2𝑖, 𝜏)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏 + 𝑖−1

𝑗=𝑢2𝑖𝑡2𝑗+2

𝑡2𝑗

ℎ (𝑡2𝑖, 𝜏)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏

− Δ𝑥𝑚(𝑡2𝑖) − 𝑥𝑚−1(𝑡2𝑖) + 𝑔 (𝑡2𝑖) ] , (32)

and by applying mixed formula for(32)we obtain the follow- ing four cases.

(6)

Case 1.WhenV2𝑖 ̸= 2𝑖and𝑢2𝑖 ̸= 2𝑖, we have Δ𝑦𝑚(𝑡2𝑖)

= 1

𝐻 (𝑡2𝑖)

× [ [

0.5 (𝑡V2𝑖− 𝑦0(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡V2𝑖)Δ𝑥𝑚(𝑡V2𝑖) 𝑥0(𝑡V2𝑖)

+ ℎ (𝑡2𝑖, 𝑦0(𝑡2𝑖))Δ𝑥𝑚(𝑦0(𝑡2𝑖)) 𝑥0(𝑦0(𝑡2𝑖)) )

+ 𝑖−1

𝑗=V2𝑖

(𝑡2𝑗+2− 𝑡2𝑗) 6

× (ℎ (𝑡2𝑖, 𝑡2𝑗)Δ𝑥𝑚(𝑡2𝑗) 𝑥0(𝑡2𝑗)

+ 4ℎ (𝑡2𝑖, 𝑡2𝑗+1)Δ𝑥𝑚(𝑡2𝑗+1) 𝑥0(𝑡2𝑗+1)

+ ℎ (𝑡2𝑖, 𝑡2𝑗+2)Δ𝑥𝑚(𝑡2𝑗+2) 𝑥0(𝑡2𝑗+2) ) + 0.5 (𝑡𝑢2𝑖− 𝑦𝑚−1(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡𝑢2𝑖)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡𝑢2𝑖)󵄨󵄨󵄨󵄨󵄨

+ ℎ (𝑡2𝑖, 𝑦𝑚−1(𝑡2𝑖))log󵄨󵄨󵄨󵄨(𝑥𝑚−1(𝑦𝑚−1(𝑡2𝑖)))󵄨󵄨󵄨󵄨 ) + 𝑖−1

𝑗=𝑢2𝑖

(𝑡2𝑗+2− 𝑡2𝑗) 6

× (ℎ (𝑡2𝑖, 𝑡2𝑗)log(𝑥𝑚−1(𝑡2𝑗)) + 4ℎ (𝑡2𝑖, 𝑡2𝑗+1)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑗+1)󵄨󵄨󵄨󵄨󵄨

+ ℎ (𝑡2𝑖, 𝑡2𝑗+2)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑗+2)󵄨󵄨󵄨󵄨󵄨)

− Δ𝑥𝑚(𝑡2𝑖) − 𝑥𝑚−1(𝑡2𝑖) + 𝑔 (𝑡2𝑖) ] ] .

(33)

Case 2.IfV2𝑖= 2𝑖and𝑢2𝑖 ̸= 2𝑖, then Δ𝑦𝑚(𝑡2𝑖)

= 1

𝐻 (𝑡2𝑖)

× [ [

0.5 (𝑡2𝑖− 𝑦0(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡2𝑖)Δ𝑥𝑚(𝑡V2𝑖) 𝑥0(𝑡2𝑖)

+ ℎ (𝑡2𝑖, 𝑦0(𝑡2𝑖))Δ𝑥𝑚(𝑦0(𝑡2𝑖)) 𝑥0(𝑦0(𝑡2𝑖)) ) + 0.5 (𝑡𝑢2𝑖− 𝑦𝑚−1(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡𝑢2𝑖)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡𝑢2𝑖)󵄨󵄨󵄨󵄨󵄨

+ ℎ (𝑡2𝑖, 𝑦𝑚−1(𝑡2𝑖))log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑦𝑚−1(𝑡2𝑖))󵄨󵄨󵄨󵄨 ) + ∑𝑖−1

𝑗=𝑢2𝑖

(𝑡2𝑗+2− 𝑡2𝑗) 6

× (ℎ (𝑡2𝑖, 𝑡2𝑗)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑗)󵄨󵄨󵄨󵄨󵄨

+ 4ℎ (𝑡2𝑖, 𝑡2𝑗+1)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑗+1)󵄨󵄨󵄨󵄨󵄨

+ ℎ (𝑡2𝑖, 𝑡2𝑗+2)log󵄨󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑗+2)󵄨󵄨󵄨󵄨󵄨)

− Δ𝑥𝑚(𝑡2𝑖) − 𝑥𝑚−1(𝑡2𝑖) + 𝑔 (𝑡2𝑖) ] ] .

(34)

Case 3.WhenV2𝑖 ̸= 2𝑖and𝑢2𝑖= 2𝑖, we get Δ𝑦𝑚(𝑡2𝑖)

= 1

𝐻 (𝑡2𝑖)

× [ [

0.5 (𝑡V2𝑖− 𝑦0(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡V2𝑖)Δ𝑥𝑚(𝑡V2𝑖) 𝑥0(𝑡V2𝑖)

+ ℎ (𝑡2𝑖, 𝑦0(𝑡2𝑖))Δ𝑥𝑚(𝑦0(𝑡2𝑖)) 𝑥0(𝑦0(𝑡2𝑖)) )

(7)

+ 𝑖−1

𝑗=V2𝑖

(𝑡2𝑗+2− 𝑡2𝑗) 6

× (ℎ (𝑡2𝑖, 𝑡2𝑗)Δ𝑥𝑚(𝑡2𝑗) 𝑥0(𝑡2𝑗) + 4ℎ (𝑡2𝑖, 𝑡2𝑗+1)Δ𝑥𝑚(𝑡2𝑗+1)

𝑥0(𝑡2𝑗+1) + ℎ (𝑡2𝑖, 𝑡2𝑗+2)Δ𝑥𝑚(𝑡2𝑗+2) 𝑥0(𝑡2𝑗+2) ) + 0.5 (𝑡2𝑖− 𝑦𝑚−1(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡2𝑖)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑖)󵄨󵄨󵄨󵄨

+ ℎ (𝑡2𝑖, 𝑦𝑚−1(𝑡2𝑖))log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑦𝑚−1(𝑡2𝑖))󵄨󵄨󵄨󵄨 )

− Δ𝑥𝑚(𝑡2𝑖) − 𝑥𝑚−1(𝑡2𝑖) + 𝑔 (𝑡2𝑖) ] ] .

(35) Case 4.IfV2𝑖= 2𝑖and𝑢2𝑖= 2𝑖, then

Δ𝑦𝑚(𝑡2𝑖)

= 1

𝐻 (𝑡2𝑖)

× [ [

0.5 (𝑡2𝑖− 𝑦0(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡2𝑖)Δ𝑥𝑚(𝑡2𝑖) 𝑥0(𝑡2𝑖)

+ ℎ (𝑡2𝑖, 𝑦0(𝑡2𝑖))Δ𝑥𝑚(𝑦0(𝑡2𝑖)) 𝑥0(𝑦0(𝑡2𝑖)) ) + 0.5 (𝑡2𝑖− 𝑦𝑚−1(𝑡2𝑖))

× (ℎ (𝑡2𝑖, 𝑡2𝑖)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑡2𝑖)󵄨󵄨󵄨󵄨

+ ℎ (𝑡2𝑖, 𝑦𝑚−1(𝑡2𝑖))log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝑦𝑚−1(𝑡2𝑖))󵄨󵄨󵄨󵄨 )

− Δ𝑥𝑚(𝑡2𝑖) − 𝑥𝑚−1(𝑡2𝑖) + 𝑔 (𝑡2𝑖) ] ] .

(36)

Thus,(32)can be computed by one of(33)–(36)according to the cases.

4. The Convergence Analysis of the Method

On the basis of general theorems of Newton-Kantorovich method [21, Chapter XVIII] for the convergence, we state the following theorem regarding the successive approximations described by(18)–(20).

First, consider the following classes of functions:

(i)𝐶[𝑡0,𝑇]the set of all continuous functions𝑓(𝑡)defined on the interval[𝑡0, 𝑇],

(ii)𝐶[𝑡0,𝑇]×[𝑡0,𝑇]the set of all continuous functions𝜓(𝑡, 𝜏) defined on the region[𝑡0, 𝑇] × [𝑡0, 𝑇],

(iii)𝐶 = {𝑋 : 𝑋 = (𝑥(𝑡), 𝑦(𝑡)) : 𝑥(𝑡), 𝑦(𝑡) ∈ 𝐶[𝑡0,𝑇]}, (iv)𝐶<[𝑡0,𝑇]= {𝑦(𝑡) ∈ 𝐶1[𝑡0,𝑇]: 𝑦(𝑡) < 𝑡}.

And define the following norms

‖𝑥‖ = max

𝑡∈[𝑡0,𝑇]|𝑥 (𝑡)| ,

‖Δ𝑋‖𝐶=max{‖Δ𝑥‖𝐶[𝑡0,𝑇], 󵄩󵄩󵄩󵄩Δ𝑦󵄩󵄩󵄩󵄩𝐶[𝑡0,𝑇]} ,

‖𝑋‖𝐶1 =max{‖𝑥‖𝐶[𝑡0,𝑇], 󵄩󵄩󵄩󵄩󵄩𝑥󸀠󵄩󵄩󵄩󵄩󵄩𝐶[𝑡0,𝑇]} ,

󵄩󵄩󵄩󵄩󵄩𝑋󵄩󵄩󵄩󵄩󵄩𝐶=max{‖𝑥‖𝐶[𝑡0,𝑇], 󵄩󵄩󵄩󵄩𝑦󵄩󵄩󵄩󵄩𝐶[𝑡0,𝑇]}

‖ℎ (𝑡, 𝜏)‖ = 𝐻1, 󵄩󵄩󵄩󵄩󵄩ℎ󸀠𝜏(𝑡, 𝜏)󵄩󵄩󵄩󵄩󵄩 = 𝐻1󸀠,

‖𝑘 (𝑡, 𝜏)‖ = 𝐻2, 󵄩󵄩󵄩󵄩󵄩𝑘󸀠𝜏(𝑡, 𝜏)󵄩󵄩󵄩󵄩󵄩 = 𝐻2󸀠,

󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩1

𝑥0󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 1 𝑥0(𝑡)󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 = 𝑐1,

󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩

󵄩 1

𝑥20󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩= max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 1 𝑥20(𝑡)󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨= 𝑐2,

󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩 1

𝐺 (𝑡)󵄩󵄩󵄩󵄩󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 1 𝐺 (𝑡)󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 = 𝑐3,

󵄩󵄩󵄩󵄩𝑥0󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨𝑥0(𝑡)󵄨󵄨󵄨󵄨 = 𝐻3,

󵄩󵄩󵄩󵄩󵄩𝑥󸀠0󵄩󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨󵄨𝑥󸀠0(𝑡)󵄨󵄨󵄨󵄨󵄨 = 𝐻3󸀠,

𝑡∈[𝑡min0,𝑇]󵄨󵄨󵄨󵄨𝑦0(𝑡)󵄨󵄨󵄨󵄨 = 𝐻4,

󵄩󵄩󵄩󵄩log󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨log(𝑥 (𝑡))󵄨󵄨󵄨󵄨 = 𝐻5,

󵄩󵄩󵄩󵄩𝑔󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨𝑔(𝑡)󵄨󵄨󵄨󵄨 = 𝐻6,

󵄩󵄩󵄩󵄩𝑓󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]󵄨󵄨󵄨󵄨𝑓(𝑡)󵄨󵄨󵄨󵄨 = 𝐻7.

(37)

Let

𝜂1=max{𝐻1𝑐2(𝑇 − 𝐻4) , 𝐻1𝑐1, 𝐻󸀠1𝐻5 + 𝐻1𝐻3󸀠𝑐1,

𝐻2𝑐2(𝑇 − 𝐻4) , 𝐻2𝑐1, 𝐻2󸀠𝐻5+ 𝐻2𝐻3󸀠𝑐1} . (38)

(8)

Let us consider real valued function

𝜓 (𝑡) = 𝐾 (𝑡 − 𝑡0)2− (1 + 𝐾𝜂) (𝑡 − 𝑡0) + 𝜂, (39) where𝐾 > 0and𝜂are nonnegative real coefficients.

Theorem 1. Assume that the operator𝑃(𝑋) = 0in(5)is de- fined inΩ = {𝑋 ∈ 𝐶([𝑡0, 𝑇]) : ‖𝑋 − 𝑋0‖ ≤ 𝑅}and has continuous second derivative in closed ballΩ0 = {𝑋 ∈ 𝐶([𝑡0, 𝑇]) : ‖𝑋 − 𝑋0‖ ≤ 𝑟}where𝑇 = 𝑡0+ 𝑟 ≤ 𝑡0+ 𝑅. Suppose the following conditions are satisfied:

(1)‖Γ0𝑃(𝑋0)‖ ≤ 𝜂/(1 + 𝐾𝜂),

(2)‖Γ0𝑃󸀠󸀠(𝑋)‖ ≤ 2𝐾/(1+𝐾𝜂), when‖𝑋−𝑋0‖ ≤ 𝑡−𝑡0≤ 𝑟, where𝐾and𝜂as in(39). Then the function𝜓(𝑡)defined by (39)majorizes the operator𝑃(𝑋).

Proof. Let us rewrite(5)and(39)in the form

𝑡 = 𝜙 (𝑡) , 𝜙 (𝑡) = 𝑡 + 𝑐0𝜓 (𝑡) , (40) 𝑋 = 𝑆 (𝑋) , 𝑆 (𝑋) = 𝑋 − Γ0𝑃 (𝑋) , (41) where𝑐0= −1/𝜓󸀠(𝑡0) = 1/(1 + 𝐾𝜂)andΓ0= [𝑃󸀠(𝑋0)]−1.

Let us show that (40) and (41) satisfy the majorizing conditions [21, Theorem 1, page 525]. In fact

󵄩󵄩󵄩󵄩𝑆(𝑋0) − 𝑋0󵄩󵄩󵄩󵄩 = 󵄩󵄩󵄩󵄩−Γ0𝑃 (𝑋0)󵄩󵄩󵄩󵄩 ≤ 𝜂

1 + 𝐾𝜂 = 𝜙 (𝑡0) − 𝑡0, (42) and for the‖𝑋 − 𝑋0‖ ≤ 𝑡 − 𝑡0with the Remark in [21, Remark 1, page 504] we have

󵄩󵄩󵄩󵄩󵄩𝑆󸀠(𝑋)󵄩󵄩󵄩󵄩󵄩 =󵄩󵄩󵄩󵄩󵄩𝑆󸀠(𝑋) − 𝑆󸀠(𝑋0)󵄩󵄩󵄩󵄩󵄩

≤ ∫𝑋

𝑋0󵄩󵄩󵄩󵄩󵄩𝑆󸀠󸀠(𝑋)󵄩󵄩󵄩󵄩󵄩 𝑑𝑋 = ∫

𝑋

𝑋0󵄩󵄩󵄩󵄩󵄩Γ0𝑃󸀠󸀠(𝑋)󵄩󵄩󵄩󵄩󵄩 𝑑𝑋

≤ ∫𝑡

𝑡0

𝑐0𝜓󸀠󸀠(𝜏) 𝑑𝜏 = ∫𝑡

𝑡0

2𝐾 1 + 𝐾𝜂𝑑𝜏

= 2𝐾

1 + 𝐾𝜂(𝑡 − 𝑡0) = 𝜙󸀠(𝑡) .

(43)

Hence𝜓(𝑡) = 0is a majorant function of𝑃(𝑋) = 0.

Theorem 2. Let the functions𝑓(𝑡), 𝑔(𝑡) ∈ 𝐶[𝑡0,𝑇],𝑥0(𝑡) ∈ 𝐶1[𝑡0, 𝑇],𝑥0(𝑦0(𝑡)) ̸= 0,𝑥20(𝑡) ̸= 0, and the kernelsℎ(𝑡, 𝜏), 𝑘(𝑡, 𝜏) ∈ 𝐶1[𝑡0,𝑇]×[𝑡0,𝑇]and(𝑥0(𝑡), 𝑦0(𝑡)) ∈ Ω0; then

(1)the system(7)has unique solution in the interval[𝑡0, 𝑇]; that is, there exists Γ0, and ‖Γ0‖ ≤ ∑𝑗=1(𝑐1𝐻1 + 𝑐1𝑐3𝐻2)𝑗((𝑇 − 𝐻4)𝑗−1/(𝑗 − 1)!) = 𝜂2,

(2)‖Δ𝑋‖ ≤ 𝜂/(1 + 𝐾𝜂), (3)‖𝑃󸀠󸀠(𝑋)‖ ≤ 𝜂1,

(4)𝜂 > 1/𝐾and𝑟 < 𝜂 + 𝑡0,

where𝐾 and 𝜂 as in (39). Then the system(4) has unique solution𝑋 in the closed ballΩ0 and the sequence𝑋𝑚(𝑡) = (𝑥𝑚(𝑡), 𝑦𝑚(𝑡)),𝑚 ≥ 0of successive approximations

Δ𝑦𝑚(𝑡) = 1 𝐻 (𝑡)[∫𝑡

𝑦0(𝑡)ℎ (𝑡, 𝜏)Δ𝑥𝑚(𝜏) 𝑥0(𝜏) 𝑑𝜏 + ∫𝑡

𝑦𝑚−1(𝑡)ℎ (𝑡, 𝜏)log󵄨󵄨󵄨󵄨𝑥𝑚−1(𝜏)󵄨󵄨󵄨󵄨 𝑑𝜏

− Δ𝑥𝑚(𝑡) − 𝑥𝑚−1(𝑡) + 𝑔 (𝑡) ] ,

Δ𝑥𝑚(𝑡) − ∫𝑡

𝑦0(𝑡)𝑘1(𝑡, 𝜏)Δ𝑥𝑚(𝜏)

𝑥0(𝜏) 𝑑𝜏 = 𝐹𝑚−1(𝑡) ,

(44)

whereΔ𝑥𝑚(𝑡) = 𝑥𝑚(𝑡)−𝑥𝑚−1(𝑡)andΔ𝑦𝑚(𝑡) = 𝑦𝑚(𝑡)−𝑦𝑚−1(𝑡), 𝑚 = 2, 3, . . ., and𝑋𝑚converge to the solution𝑋. The rate of convergence is given by

󵄩󵄩󵄩󵄩𝑋− 𝑋𝑚󵄩󵄩󵄩󵄩 ≤ ( 21 + 𝐾𝜂)𝑚(1

𝐾) . (45)

Proof. It is shown that(7)is reduced to(17). Since(17)is a linear Volterra integral equation of 2nd kind with respect to Δ𝑥(𝑡)and since𝑘(𝑡, 𝑦0(𝑡)) ̸= 0, ∀𝑡 ∈ [𝑡0, 𝑇]which implies that the kernel𝑘1(𝑡, 𝜏)defined by(18)is continues it follows that (17) has a unique solution which can be obtained by the method of successive approximations. Then the function Δ𝑦(𝑡)is uniquely determined from(16). Hence the existence ofΓ0is archived.

To verify thatΓ0is bounded we need to establish the resol- vent kernelΓ0(𝑡, 𝜏)of(17), so we assume the integral operator 𝑈from𝐶[𝑡0, 𝑇] → 𝐶[𝑡0, 𝑇]is given by

𝑍 = 𝑈 (Δ𝑥) , 𝑍 (𝑡) = ∫𝑡

𝑦0(𝑡)𝑘2(𝑡, 𝜏) Δ𝑥 (𝜏) 𝑑𝜏, (46) where𝑘2(𝑡, 𝜏) = 𝑘1(𝑡, 𝜏)/𝑥0(𝜏), and𝑘1(𝑡, 𝜏)is defined in(18).

Due to(46),(17)can be written as

Δ𝑥 − 𝑈 (Δ𝑥) = 𝐹0. (47)

The solutionΔ𝑥of(47)is expressed in terms of𝐹0by means of the formula

Δ𝑥 = 𝐹0+ 𝐵 (𝐹0) , (48) where 𝐵is an integral operator and can be expanded as a series in powers of𝑈[21, Theorem 1, page 378]:

𝐵 (𝐹0) = 𝑈 (𝐹0) + 𝑈2(𝐹0) + ⋅ ⋅ ⋅ + 𝑈𝑛(𝐹0) + ⋅ ⋅ ⋅ , (49) and it is known that the powers of𝑈are also integral opera- tors. In fact

𝑍𝑛 = 𝑈𝑛, 𝑍𝑛(𝑡) = ∫𝑡

𝑦0(𝑡)𝑘(𝑛)2 (𝑡, 𝜏) Δ𝑥 (𝜏) 𝑑𝜏, (𝑛 = 1, 2, . . .) ,

(50)

where𝑘(𝑛)2 is the iterated kernel.

(9)

Substituting(50)into(48)we obtain an expression for the solution of(47):

Δ𝑥= 𝐹0(𝑡) +∑

𝑗=1

𝑡

𝑦0(𝑡)𝑘(𝑗)2 (𝑡, 𝜏) 𝐹0(𝜏) 𝑑𝜏. (51) Next, we show that the series in(51)is convergent uniformly for all𝑡 ∈ [𝑡0, 𝑇]. Since

󵄨󵄨󵄨󵄨𝑘2(𝑡, 𝜏)󵄨󵄨󵄨󵄨 =󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨𝑘1(𝑡, 𝜏) 𝑥0(𝜏) 󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨

≤󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨ℎ (𝑡, 𝜏)

𝑥0(𝜏)󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 +󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 𝑘 (𝑡, 𝜏)

𝑥0(𝜏) 𝐺 (𝑡)󵄨󵄨󵄨󵄨󵄨󵄨󵄨󵄨 ≤ 𝑐1𝐻1+ 𝑐1𝑐3𝐻2. (52)

Let𝑀 = 𝑐1𝐻1+ 𝑐1𝑐3𝐻2; then by mathematical induction we get

󵄨󵄨󵄨󵄨󵄨𝑘(2)2 (𝑡, 𝜏)󵄨󵄨󵄨󵄨󵄨 ≤ ∫

𝑡

𝑦0(𝑡)󵄨󵄨󵄨󵄨𝑘2(𝑡, 𝑢) 𝑘2(𝑢, 𝜏)󵄨󵄨󵄨󵄨 𝑑𝑢 ≤ 𝑀2(𝑡 − 𝐻4) (1)! ,

󵄨󵄨󵄨󵄨󵄨𝑘(3)2 (𝑡, 𝜏)󵄨󵄨󵄨󵄨󵄨 ≤ ∫

𝑡

𝑦0(𝑡)󵄨󵄨󵄨󵄨󵄨𝑘2(𝑡, 𝑢) 𝑘(2)2 (𝑢, 𝜏)󵄨󵄨󵄨󵄨󵄨 𝑑𝑢 ≤ 𝑀3(𝑡 − 𝐻4)2 (2)! , ...

󵄨󵄨󵄨󵄨󵄨𝑘(𝑛)2 (𝑡, 𝜏)󵄨󵄨󵄨󵄨󵄨 ≤ ∫

𝑡

𝑦0(𝑡)󵄨󵄨󵄨󵄨󵄨𝑘2(𝑡, 𝑢) 𝑘(𝑛−1)2 (𝑢, 𝜏)󵄨󵄨󵄨󵄨󵄨 𝑑𝑢

≤ 𝑀𝑛(𝑡 − 𝐻4)𝑛−1 (𝑛 − 1)! ,

(𝑛 = 1, 2, . . .) ; (53) then

󵄩󵄩󵄩󵄩𝑈𝑛󵄩󵄩󵄩󵄩 = max

𝑡∈[𝑡0,𝑇]𝑡

𝑦0(𝑡)󵄨󵄨󵄨󵄨󵄨𝑘(𝑛)2 (𝑡, 𝜏)󵄨󵄨󵄨󵄨󵄨 𝑑𝜏 ≤𝑀𝑛(𝑇 − 𝐻4)(𝑛−1) (𝑛 − 1)! .

(54) Therefore the𝑛th root test of the sequence yields

√‖𝑈𝑛 𝑛‖ ≤ 𝑀 (𝑇 − 𝐻4)1−1/𝑛

√(𝑛 − 1)!𝑛 󳨀󳨀󳨀→𝑛 → ∞0. (55) Hence𝜌 = 1/lim𝑛 → ∞√‖𝑈𝑛 𝑛‖ = ∞and a Volterra integral equations(17)has no characteristic values. Since the series in(51)converges uniformly(48)can be written in terms of resolvent kernel of(17):

Δ𝑥 = 𝐹0+ ∫𝑡

𝑦0(𝑡)Γ0(𝑡, 𝜏) 𝐹0(𝜏) 𝑑𝜏, (56) where

Γ0(𝑡, 𝜏) =∑

𝑗=1𝑘(𝑗)2 (𝑡, 𝜏) . (57)

Since the series in(57)is convergent we obtain

󵄩󵄩󵄩󵄩Γ0󵄩󵄩󵄩󵄩 = 󵄩󵄩󵄩󵄩𝐵(𝐹0)󵄩󵄩󵄩󵄩 ≤∑

𝑗=1󵄩󵄩󵄩󵄩󵄩𝑈𝑗󵄩󵄩󵄩󵄩󵄩 ≤

𝑗=1𝑀𝑗(𝑇 − 𝐻)𝑗−1

(𝑗 − 1)! ≤ 𝜂2. (58)

To establish the validity of second condition, let us represent operator equation

𝑃 (𝑋) = 0, (59)

as in(41)and its the successive approximations is

𝑋𝑛+1= 𝑆 (𝑋𝑛) , (𝑛 = 0, 1, 2, . . .) . (60)

For initial guess𝑋0we have

𝑆 (𝑋0) = 𝑋0− Γ0𝑃 (𝑋0) . (61)

From second condition of (Theorem 1) we have

󵄩󵄩󵄩󵄩Γ0𝑃 (𝑋0)󵄩󵄩󵄩󵄩 =󵄩󵄩󵄩󵄩𝑆(𝑋0) − 𝑋0󵄩󵄩󵄩󵄩

= 󵄩󵄩󵄩󵄩𝑋1− 𝑋0󵄩󵄩󵄩󵄩 = ‖Δ𝑋‖ ≤ 𝜂1 + 𝐾𝜂. (62) In addition, we need to show that‖𝑃󸀠󸀠(𝑋)‖ ≤ 𝜂1for all 𝑋 ∈ Ω0 where 𝜂1 is defined in (38). It is known that the second derivative 𝑃󸀠󸀠(𝑋0)(𝑋, 𝑋) of the nonlinear operator 𝑃(𝑋) is described by 3-dimensional array 𝑃󸀠󸀠(𝑋0)𝑋𝑋 = (𝐷1, 𝐷2)(𝑋, 𝑋), which is called bilinear operator; that is, 𝑃󸀠󸀠(𝑋0)(𝑋𝑋) = 𝐵(𝑋0, 𝑋, 𝑋)where

𝑃󸀠󸀠(𝑋0) (𝑋, 𝑋)

=lim

𝑠 → 0

1

𝑠[𝑃󸀠(𝑥0+ 𝑠𝑋) − 𝑃󸀠(𝑋0)]

= {lim

𝑠 → 0

1 𝑠[(𝜕𝑃1

𝜕𝑥 (𝑥0+ 𝑠𝑥, 𝑦0+ 𝑠𝑦) − 𝜕𝑃1

𝜕𝑥 (𝑥0, 𝑦0)) 𝑥 + (𝜕𝑃1

𝜕𝑦 (𝑥0+ 𝑠𝑥, 𝑦0+ 𝑠𝑦) − 𝜕𝑃1

𝜕𝑦 (𝑥0, 𝑦0)) 𝑦] ,

𝑠 → 0lim 1 𝑠[(𝜕𝑃2

𝜕𝑥 (𝑥0+ 𝑠𝑥, 𝑦0+ 𝑠𝑦) − 𝜕𝑃2

𝜕𝑥 (𝑥0, 𝑦0)) 𝑥 + (𝜕𝑃2

𝜕𝑦 (𝑥0+ 𝑠𝑥, 𝑦0+ 𝑠𝑦) −𝜕𝑃2

𝜕𝑦 (𝑥0, 𝑦0)) 𝑦]}

Referensi

Dokumen terkait

Submit your manuscripts at http://www.hindawi.com Computer Games Technology International Journal of Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi

Submit your manuscripts at http://www.hindawi.com Stem Cells International Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation

Submit your manuscripts at http://www.hindawi.com Stem Cells International Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation

Submit your manuscripts at http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Inorganic Chemistry International Journal of Hindawi Publishing

Submit your manuscripts at http://www.hindawi.com Stem Cells International Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation

Submit your manuscripts at http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Anatomy Research International Peptides Hindawi Publishing

Submit your manuscripts at http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 MathematicsJournal of Hindawi Publishing Corporation

Submit your manuscripts at http://www.hindawi.com Stem Cells International Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation