• Tidak ada hasil yang ditemukan

Directory UMM :Data Elmu:jurnal:J-a:Journal of Computational And Applied Mathematics:Vol103.Issue2:

N/A
N/A
Protected

Academic year: 2017

Membagikan "Directory UMM :Data Elmu:jurnal:J-a:Journal of Computational And Applied Mathematics:Vol103.Issue2:"

Copied!
15
0
0

Teks penuh

(1)

Convergence of partially asynchronous block

quasi-Newton methods for nonlinear systems of equations

Jian-Jun Xu∗

Department of Mathematics, Temple University, Philadelphia, PA 19122, United States Received 4 May 1998; received in revised form 7 October 1998

Abstract

In this paper, a partially asynchronous block Broyden method is presented for solving nonlinear systems of equations of the formF(x) = 0. Sucient conditions that guarantee its local convergence are given. In particular, local convergence is shown when the JacobianF′(x) is an H-matrix, wherexis the zero point ofF. The results are extended to Schubert’s method. A connection with discrete Schwarz alternating procedure is also shown. c1999 Elsevier Science B.V. All rights reserved.

Keywords:Parallel iterative methods; Asynchronous iterations; Quasi-Newton methods; Schwarz alternating procedure

1. Introduction

Chazan and Miranker rst introduced an asynchronous method for the solution of nonsingular linear system of equations of the form Ax=b [9]. They proved that the asynchronous point method converges to the solution if and only if A is an H-matrix [29]. This idea was extended to nonlinear systems of equations, and there is now a considerable understanding on the convergence properties of asynchronous iterative methods for nonlinear problems. Many authors have concentrated on xed point problems of the form x=G(x). Convergence properties have been shown if G satises some contraction properties; see, e.g., [3–5, 13]. The xed point mapping G does not always arise directly from the formulation of physical problems, however, and when it does, the contracting properties are not always easily veriable.

In this paper, we consider the problem of nding x∗ which satises the following nonlinear system

of equations:

F(x) = 0; (1.1)

where F= (f1; : : : ; fn)T is a nonlinear operator from Rn into itself.

E-mail: [email protected].

(2)

In the literature, one can nd two general approaches for the asynchronous solutions of (1.1). First, for an asynchronous implementation of Newton’s method, one can use two processes, one updating the iterates, and the other evaluating the Jacobian matrix; see [3, 7, 20]. Also see [31] for the analogue of Broyden’s method. More recently, asynchronous block methods, or more general multisplitting methods were proposed. In [2, 6] the nonlinear term was treated explicitly because of the weak nonlinearity. In [14, 22], convergence properties were shown where F is an M-function, see also [12, 15] for point methods. In [12, 14, 15, 22], exact solutions for subproblems were assumed. The use of Newton-type methods for approximate solutions of the subproblems and local convergence theorems for the corresponding asynchronous methods were given in [1, 27, 32].

In this paper, we consider partially asynchronous block quasi-Newton methods. Suppose F and x in (1.1) are conformally partitioned as follows:

F= (FT

1; : : : ; FLT)T; x= (xT1; : : : ; xLT)T;

where Fi:Rn→Rni; Fi= (fi1; : : : ; fini)

T; x

i∈Rni; xi= (xi1; : : : ; xini)

T; i= 1; : : : ; L: Let S

i={i1; : : : ; ini},

the partition chosen is such that SL

i=1Si={1; : : : ; n}; Si∩Sj=∅; i6=j; i; j= 1; : : : ; L. This partition

may correspond to a decomposition of the domain of a nonlinear partial dierential equations, see Section 5.

The system (1.1) can be rewritten as

Fl(x1; : : : ; xl; : : : ; xL) = 0; l= 1; : : : ; L: (1.1)′

We consider the following nonlinear block method. Given initial values of x= (xT

1; : : : ; xTL)T, repeat

the following procedure until convergence: For l= 1; : : : ; L

Solve for yl in Fl(x1; : : : ; xl−1; yl; xl+1; : : : ; xL) = 0;

for all l∈S⊆{1; : : : ; L};

Set xl=yl for all l∈S:

(1.2)

In (1.2) the subset S can be chosen in a dynamic fashion. The classical nonlinear block-Jacobi method and nonlinear block-Gauss–Seidel method [4, 23] are two examples of such selections. For the purpose of parallel processing, the nonlinear block Jacobi method is nearly ideal, since up to L processors can each perform one of the iterations in (1.2) with the others simultaneously. It is synchronous in the sense that to begin the computation of the next iterate, each processor has to wait until all processors have completed their current iteration. In practice we need to add a synchronization mechanism to ensure that the algorithm is carried out correctly. The time necessary to carry out the synchronization mechanism, as well as the time a processor must wait until its data is ready to continue the calculation, adds an overhead to the computation. By removing this synchronization and letting the processors continue their calculations according to the information currently available, we obtain the so-called asynchronous parallel methods; see, e.g., [3, 4, 9].

(3)

The advantages of quasi-Newton methods for (1.1) are that they involve evaluating only O(n) scalar functions in each iteration step, which is much cheaper than Newton’s method, they do not involve computation of the derivative of F, and they still have superlinear convergence; see, e.g. [10, 11] and the references given therein. In this paper, we consider the use of quasi-Newton methods instead of Newton’s method. We will consider two quasi-Newton methods: Broyden’s method [8] and Schubert’s method [24].

The main purpose of this paper is to give sucient conditions which guarantee local convergence of partially asynchronous block quasi-Newton methods. To this end, we assume in the rest of the paper that F and the partition satisfy the following conditions:

F is continuously dierentiable on an open convex set ⊂Rn: (1.3)

There exists a constant ¿0 such that

kF′(x)−F′(x∗)k6kx−x∗k; for x∈: (1.4)

All the matrices @Fi(x∗)=@xi; i= 1; : : : ; L; are nonsingular, moreover

(|D(x∗)−1(F′(x∗)−D(x∗))|)¡1; (1.5)

where D(x∗) = diag(@F

1(x∗)=@x1; : : : ; @FL(x∗)=@xL) is a block diagonal matrix, and we use the

nota-tion

@Fi(x)

@xj

= 

     

@fi1(x)

@xj1

· · · @fi1(x)

@xjnj

· · · ·

@fini(x)

@xj1

· · · @fini(x)

@xjnj

     

i; j= 1; : : : ; L;

and

F′(x) = 

    

@F1(x)

@x1

· · · @F1(x)

@xL

· · · ·

@FL(x)

@x1

· · · @FL(x)

@xL

    

:

Remark 1. Conditions (1.3) and (1.4) are standard. Condition (1.5) is natural for the convergence. Consider the linear case, i.e., F(x) =Ax−b, and let F and x be partitioned as before. Suppose that

A−1 exists. Since F(x) =A, condition (1.5) is necessary and sucient for the convergence of the

asynchronous block method for the linear equations F(x) = 0; see, e.g., [4, 9].

Remark 2. Condition (1.5) holds when the Jacobian matrix F′(x) is an H-matrix, since F(x) =

D(x∗)(D(x)F(x)) is an H-splitting of F(x); see e.g., [5, 15]. On the other hand, if each

(4)

H-matrix. The class of H-matrices includes M-matrices, and also the class of strictly or irreducibly diagonally dominant matrices [29].

In Section 2, we present both the computational and mathematical models of partially asynchronous block Broyden’s method. The local convergence theorem is proved in Section 3 under conditions (1.3)–(1.5). In Section 4, we point out that local convergence is still guaranteed if Schubert’s method [24] is used instead of Broyden’s method. The connection of the block method with the Schwarz alternating procedure (see, e.g., [18, 19, 25, 30]) is discussed in Section 5.

The weighted maximum norms is used as an important tool in proofs of asynchronous iterative methods; see, e.g., [3, 4, 6, 9, 13, 17, 26, 27]. Given an vectorw∈Rn; w ¿0. Letw= (wT

1; : : : ; wTL)T

be partitioned conformally with x, then we dene

kxkw= max{kxikwi;16i6L}= max

(

|xij|

wij

;16j6ni;16i6L

)

:

Given a matrix A∈Rn×n. Let A= (A

ij)L×L be partitioned conformally with F′(x), then the norms

k · kw;k · kwi induce the following matrix norm:

kAkw= max

kAxk

w

kxkw

: x6= 0; x∈Rn

:

We extend the weighted maximum norms for rectangular matrices as follows:

k(Ai1; : : : ; AiL)kwi= max

k(A

i1; : : : ; AiL)xkwi kxkw

: x6= 0; x∈Rn

and

kAijkwi= max

(

kAijxjkwi kxjkwj

: xj6= 0; xj∈Rnj

)

:

We immediately have

kAii(Ai1; : : : ; AiL)kwi6kAiikwi· k(Ai1; : : : ; AiL)kwi; (1.6)

k|A|kw=kAkw; k|(Ai1; : : : ; AiL)|kwi=k(Ai1; : : : ; AiL)kwi; (1.7)

kAkw= max{k(Ai1; : : : ; AiL)kwi;16i6L}; (1.8)

and

kAijkwi6k(Ai1; : : : ; AiL)kwi; 16i; j6L: (1.9)

(5)

(1:5)′ All matrices @F

i(x∗)=@xi; i= 1; : : : ; L; are nonsingular, moreover, there exists 0¡1, and a

vector w ¿0 such that kD(x∗)−1(F(x)D(x))k

w¡ 0:

2. Partially asynchronous block Broyden method

Broyden’s method [8] for (1.1) is as follows:

  

  

x(k+ 1) =x(k)−B(k)−1F(x(k));

B(k+ 1) =B(k) + (1=z(k)Tz(k))·(y(k)−B(k)z(k))z(k)T; k= 1;2; : : : ;

where y(k) =F(x(k+ 1))−F(x(k)); z(k) =x(k+ 1)−x(k);

where x(0) and B(0) are initial approximation to the solution x∗ and initial approximation the Jacobian F′(x), respectively.

We will rst describe a computational model for partially asynchronous block Broyden’s method, i.e., how the computer is actually programmed to execute its instructions, then we give a mathematics model, i.e., how the execution is described mathematically in order to analyze convergence [26].

Let us as before assume that there are L processors. We design L processes. Each processor perform computation in one process. In each process, say the lth process, we use one step of Broyden’s method to update the lth block xl in the lth subproblem in (1.2), l= 1; : : : ; L. More

precisely, the computational model can be written as the following pseudo-code: Given an initial guess x, and initial matrices Bl; l= 1; : : : ; L.

Process l(l= 1; : : : ; L) Initialization for x; Bl

Step 1: Compute Fl:=Fl(x);

Step 2: Set x:=xl; Fl:=Fl;

Step 3: Compute xl:=xl−B−l 1Fl;

Step 4: Send xl to the other processes;

Step 5: Check the termination criterion, if satised then stop;

Step 6: Receive the values of x (except xl) from the other processes;

Step 7: Compute Fl:=Fl(x);

Step 8: Compute z:=xl−xl; y:=Fl−Fl;

Step 9: Compute Bl:=Bl+ (1=zTz)·(y−Blz)zT;

Step 10: Goto step 2;

where Bl;Fl;x; y; z are local variables, only x is a global variable. In each cycle of a process, say

the lth process, one step of Broyden’s update is performed for a subproblem of nl dimension. The

L processes can be done in parallel. There is no explicit synchronization among the processes.

Remark 3. Since Broyden update is a rank-one update, the Sherman–Morrison formula can be used to update the inverse of Bl. Let Hl=Bl−1, then initialization for Bl is changed forHl, Steps 3 and 9

are rewritten as follows while the other steps remain unchanged: Step 3′: Compute x

l:=xl−Hl·Fl;

Step 9′: H

(6)

One can also uses QR factorization technique to solve the linear system in step 3; see, e.g., [11]. We give the following mathematical model in order to analyze the convergence.

Let the initial approximations x(0) = (x1(0)T; : : : ; xL(0)T)T, and Bi(0); i= 1; : : : ; Lbe given, then for

k= 0;1; : : : ; i= 1; : : : ; L;

                     

                     

xi(k+ 1) =xi(si(k))−Bi(k)−1Fi(u(k)) ifi∈I(k)

xi(k+ 1) =xi(k) else

Bi(k+ 1) =Bi(k) + (1=zi(k)Tzi(k))·(yi(k)−Bi(k)zi(k))zi(k)T ifi∈I(k)

Bi(k+ 1) =Bi(k) else;

whereu(k) = (x1(s1(k))T; : : : ; xL(sL(k))T)T;

yi(k) =Fi(x(k + 1))−Fi(v(k; i));

v(k; i) = (x1(k+ 1)T; : : : ; xi−1(k+ 1)T;

xi(si(k))T; xi+1(k+ 1)T; : : : ; xL(k+ 1)T)T;

zi(k) =xi(k+ 1)−xi(si(k));

where I(k) and sj(k) satisfy the following conditions

(2.1) for some xed integer p ¿0; k−p ¡ sj(k)6k; j= 1; : : : ; L; k= 0;1; : : :

(2.2) each I(k); k= 0;1; : : : are subset of {1; : : : ; L}, and the set {k: j∈I(k)} is unbounded for all j= 1; : : : ; L.

Conditions (2.1) and (2.2) may be interpreted as follows: At each instant of time k, all the ith block component xi; i∈I(k), are updated while the remaining are unchanged. The updating uses the

rst block component of x(s1(k)), the second block component of x(s2(k)), etc. Then the matrices

Bi; i∈I(k), are updated by using the new results. Each pair of block component and matrix is

updated innitely many often, and no update uses a value of a component which was produced by an update p or more steps previously.

The assumption of the existence of uniform bound p for delays appeared in the rst asynchronous model in [9]. It is not an additional constraint in most practical implementations [3]. This type of asynchronous model was called partially asynchronous method in [4].

We will refer the above mathematical model as partially asynchronous block Broyden’s (PABB) method.

If we take si(k) =k; I(k) ={1; : : : ; L} for all k and i, then PABB method describes block-Jacobi

Broyden method.

If we take si(k) =k; I(k) ={k+ 1 (modL)} for all k, then PABB method describes block-Gauss–

Seidel Broyden method.

3. Local convergence theorem

(7)

Theorem 3.1. Suppose that (1:3)–(1:5) are satised; then there exist two constants ¿0; ¿0;

such that if kx(0)−x∗kw6 for w ¿0 dened in (1:5)′; kBi(0)−(@Fi(x∗))=@xik26; i= 1; : : : ; L;

then the sequence {x(k)} generated by PABB method converges to x∗.

Before we give the proof of Theorem 3.1, we need several Lemmas. Because of the norm equiv-alence, we may assume that the norm in (1.4) is k · kw, and there exist two constants 1; 2¿0

such that

kAk261kAkwˆ; kAkwˆ62kAk2; (3.1)

where ˆw=w if A∈Rn×n, and ˆw=w

i if A∈Rni×n(orRni×nj).

Lemma 3.2 (Ortega and Rheinboldt [23]). IfF:RnRm is Gateaux-dierentiable in an open

(8)

exists S(x∗; 1) ={x:kx−x∗kw61} ⊂; such that

kD(x∗)−1(F′(x)−D(x))kw¡ 0; forx∈S(x∗; 1)

thus by (1.8), (3.3) holds.

(9)

while by (3.1), (3.2)

thus we have proved Lemma 3.6.

Proof of Theorem 3.1. Suppose that for some xed ¿0; we have

We use mathematical induction to show the following inequalities:

kxi(k+ 1)−x∗i kwi6rkx(s(k))−x∗kw if i∈I(k);

and as a byproduct we get

(10)
(11)

Thus if i∈I(m), the uniform bound p for the delays, each block component is updated at least once for every 2p

(12)

If i =∈I(m);

Bi(m+ 1)−

@Fi(x∗)

@xi

2

=

Bi(m)−

@Fi(x∗)

@xi

2

62:

Therefore (3.8) holds for all k; and then (3.9) holds for all k; thus {x(k)} convergences to x∗. By Remark 2, we have the following corollary.

Corollary 3.7. Suppose that (1.3)–(1.4) are satised; and F′(x) is an H-matrix; then for any

partitioning of F and x; there exist two constants ¿0; ¿0; such that if kx(0)−x∗kw6 for

w ¿0 dened in (1.5)′; kB

i(0)−(@Fi(x∗))=@xik26; i= 1; : : : ; L; the sequence {x(k)} generated by

PABB method converges to x∗.

4. Extension to Schubert’s method

Schubert’s method [24] is a variation of Broyden’s method which is of interest in the case that the Jacobian is sparse. In this variation, the approximate Jacobian is forced to have the same sparsity pattern as the Jacobian.

Given a vector function F= (f1; : : : ; fm)T∈Rm → Rm; we dene its sparsity pattern as a matrix

PF= (pij)m×m; where

pij=

0 if f

i(x) is independent of xj

1 else i; j= 1; : : : ; m:

Let PFl= (plij) be the sparsity pattern of Fl; l= 1; : : : ; L. Using Schubert’s method instead of

Broy-den’s method for each subproblem in (1.2) leads to the following computational model: Given initial solution x= (xT

1; : : : ; xTL)T; and matrices Bl.

Process l(l= 1; : : : L) Initialization for x= (xT

1; : : : ; xTl; : : : ; xLT)T; Bl

Steps 1–8 and 10 are the same as the computational model of PABB method; step 9: if plij= 1 then uij←zj else uij←0; i; j= 1; : : : ; nl;

Bl←Bl+ nl

X

i=1

ui6=0

eieTi

(y−Blz)

uT

iui

uT

i;

where ui= (ui1; : : : ; uinl)

T; i= 1; : : : ; n

l; and ei∈Rnl is the unit vector of which the ith component is

1, i= 1; : : : ; nl.

Its mathematical model can be described similarly as PABB method.

(13)

5. Connection with the discrete Schwarz alternating procedure

In this section, suppose that there are overlaps among the blocks x1; : : : ; xL; and correspondingly

among F1; : : : ; FL. This may happen when discrete Schwarz alternating procedure (see, e.g., [18, 19,

25, 30]) is used to solve (1.1), where each block xl corresponds to the unknowns in a subdomain.

In this case, the computational models need not change, but we need to modify the mathematical model so that the previous analysis can be applied.

Suppose that PL

l= 1nl=m ¿ n; we dene a vector x= ( xT1; : : : ;xTL)T∈Rm from x∈Rn by setting

xl=xl; l= 1; : : : ; L: Denote this mapping from Rn to Rm by x=(x), similarly we obtain F=

( FT

1; : : : ;FTL)T=(F). We can thus rewrite (1.1) by repeating the overlapping unknowns and the

corresponding equations:

Fl( xl;Zl) = 0; l= 1; : : : ; L; (5.1)

where the vector Zl is a subset of

S

i6=l{xl}; since Fl depends only on n variables. The vector Zl

may not be unique, since the overlapping unknowns may have dierent choices of sub-indices. As before, we can update the approximate values of the lth block of unknowns by exploiting the lth block equations. Note that x∗=(x∗) is the solution of (5.1). On the other hand, we can recover the solution x∗ from x∗ by simply discarding the redundant overlapping parts in x∗. The previous analysis now can be applied to (5.1).

If F(x) is dierentiable and satises the Lipschitz condition, then F( x) will also be dierentiable and satisfy the Lipschitz condition with the same Lipschitz constant. Condition (1.5) for F does not mean a similar condition for F. When F′(x) is an H-matrix, however, we can show that F( x)

is also an H-matrix.

Theorem 5.1. If F′(x) is an M-matrix; then F( x) is also an M-matrix; where x=(x);

F=(F∗).

Proof. For simplicity, let A=F′(x); A= F( x). The diagonal entries of A are composed of the

diagonal entries of (@Fl( x∗))=@xl; l= 1; : : : ; L. Since each (@Fl( x∗))=@xl is a diagonal block of A;the

diagonal entries of A are diagonal entries of A; thus positive. The nonzero o-diagonal entries of A

are also o-diagonal entries of A; thus nonpositive. Let A= D−B; A =D−B; where D; D are the diagonal parts of A and A; respectively. To show A is an M-matrix, we need only to show that A−1

is nonnegative. It is enough to show that ( D−1B)¡1; see, e.g., [23, 28]. Since A is an M-matrix,

(D−1B)¡1; and D−1B is nonnegative, there exists a positive vector vRn such that D−1Bv ¡ v.

Let v=(v); then we have D−1Bv ¡ v. Thus ( D−1B)¡1; see, e.g., [4].

Corollary 5.2. If F′(x) is an H-matrix; then F( x) is also an H-matrix; where x=(x);

F=(F∗).

Proof. Since F′(x) is an H-matrix, the comparison matrix of F(x) is an M-matrix; see, e.g. [6,

16]. Thus by Theorem 5.1, the comparison matrix of F′( x) is also an M-matrix, i.e. F( x) is an

(14)

Theorem 5.1 is more general than those similar theorems in [14, 22], since the above augmentation process allows components of x to belong to more than two subdomains.

Acknowledgements

The author would like to thank Professor Daniel B. Szyld for his careful reading of the manuscript which improved the presentation, and also thank the referees for the valuable comments.

References

[1] Z.-Z. Bai, On the convergence of parallel chaotic nonlinear multisplitting Newton-type methods, Comput. Appl. Math. 80 (1997) 317– 334.

[2] Z.-Z. Bai, V. Migallon, J. Penades, D.B. Szyld, Block and asynchronous two stage methods for mildly nonlinear systems, Research Report 97-31, Dept. Math., Temple Univ., 1997, in Numer. Math., to appear.

[3] G. Baudet, Asynchronous iterative methods for multiprocessors, J. ACM 25 (1978) 226 –344.

[4] D.P. Bertsekas, J.N. Tistisklis, Parallel and Distributed Computation, Prentice-Hall, Englewood Clis, NJ, 1989. [5] J. Bhai, J.C. Miellou, K. Rhor, Asynchronous multisplitting methods for nonlinear xed point problems, Numer.

Algorithms 15 (1997) 315 – 345.

[6] A. Bhaya, E. Kaszkurewicz, F. Mota, Asynchronous block-iterative method for almost linear equations, Linear Algebra Appl. 154 –156 (1991) 487– 508.

[7] A. Bojanczyk, Optimal asynchronous Newton method for the solution of nonlinear equations, J. ACM 31 (1984) 792 – 803.

[8] C.G. Broyden, A class of methods for solving nonlinear simultaneous equations, Math. Comp. 19 (1965) 577– 593. [9] D. Chazan, W. Miranker, Chaotic relaxation, Linear Algebra Appl. 2 (1969) 199 –222.

[10] J.E. Dennis, J.J. More, Quasi-Newton methods, motivation and theory, SIAM Rev. 19 (1977) 46 – 89.

[11] J.E. Dennis, B.S. Robert, Numerical methods for unconstrained optimization and nonlinear equations, Prentice-Hall, Englewood Clis, NJ, 1983.

[12] D. El Baz, M-functions and parallel asynchronous algorithms, SIAM J. Numer. Anal. 27 (1990) 136 –140. [13] M.N. El. Tarazi, Some Convergence results for asynchronous algorithms, Numer. Math. 39 (1982) 325 –340. [14] D.J. Evans, D.-R. Wang, An asynchronous algorithm for solving a class of a nonlinear simultaneous equations,

Parallel Comput. 17 (1991) 165 –180.

[15] A. Frommer, D.B. Szyld, Asynchronous two-stage iterative methods, Numer. Math. 69 (1994) 141–153.

[16] A. Frommer, Generalized nonlinear diagonal dominance and applications to asynchronous iterative methods, Comp. Appl. Math. 38 (1991) 105 –124.

[17] A. Frommer, H. Schwandt, D.B. Szyld, Asynchronous weighted additive Schwarz method, Electronic Trans. Numer. Anal. 5 (1997) 48 – 61.

[18] L.-S. Kang, L.L. Sun, Y.-P. Chen, Asynchronous Parallel Methods for Problems of Mathematical Physics (Chinese), Academic Press of China, 1985.

[19] S.-B. Kim, A. Hadjidimos, E.N. Houstis, J.R. Rice, Multi-parametrized Schwarz alterating methods for elliptic boundary value problems, Math. Comput. Simulation 42 (1996) 47–76.

[20] H.T. Kung, Synchronized and asynchronous parallel algorithms for multiprocessors. in: J.F. Traub, (Ed.), New Directions and Recent Results in Algorithms and Complexity, Academic Press, New York, 1976.

[21] E.S. Marwil, Convergence results for Schubert’s method for solving sparse nonlinear equations, SIAM J. Numer. Anal. 16 (1979) 588 – 604.

[22] J.C. Miellou, D. El Baz, P. Spiteri, A new class of asynchronous iterative algorithms with order intervals, Math. Comput. 67 (1998) 237– 225.

(15)

[24] L.M. Schubert, Modication of a quasi-Newton method for nonlinear equations with sparse Jacobian, Math. Comput. 24 (1970) 27– 30.

[25] B. Smith, P. Bjrstad, W. Gropp, Domain decomposition: Parallel multilevel methods for elliptic partial dierential equations, Cambridge University Press, Cambridge, 1996.

[26] D.B. Szyld, Dierent models of parallel asynchronous iterations with overlapping blocks, Comput. Appl. Math. 17 (1998) 100 –115.

[27] D.B. Szyld, J.-J. Xu, Asynchronous nonlinear multisplitting methods, in prepartion. [28] R.S. Varga, Matrix iterative analysis, Prentice-Hall, Englewood Clis, NJ, 1962.

[29] R.S. Varga, On recurring theorems on diagonal dominance, Linear Algebra Appl. 13 (1976) 1– 9. [30] W.P. Tang, Generalized Schwarz splittings, SIAM J. Sci. Statist. Comput. 13 (1992) 573 – 595.

[31] J.-J. Xu, An asynchronous parallel quasi-Newton methods for nonlinear equations (Chinese), in: L.-S. Kang et al. (Eds.), Proc. Nat. Conf. on Parallel Algorithms, Zhuo-Zhou, China, 1989, pp. 113 –117.

Referensi

Dokumen terkait

Subjek penelitian penderita kelainan refraksi miopia yang dikoreksi dengan menggunakan pemeriksaan subjektif dibanding dengan hasil koreksi dengan pemeriksaan

Analisis Kandungan Sikap Pada Buku Teks Ipa Kurikulum 2013 Dan Implementasinya Dalam Pembelajaran Di Smp. Universitas Pendidikan Indonesia | repository.upi.edu

Building on input from across the scientiic community, this report identiies key priorities for Antarctic and Southern Ocean research over the coming decade, along with

Barito Hilir Pelabuhan I Trisakti Banjarmasin

Menindaklanjuti Berita Acara Pembuktian Dokumen Prakualifikasi Pekerjaan Jasa Konsultansi Pengawasan Lanjutan Pembangunan Fasilitas Pelabuhan Laut Panasahan

Through these theological criteria of issues, the development of Islamic philanthropy for social justice in Indonesia will become more effective towards social change on the grounds

Further test conducted on the components of intellectual capital also shows that only human capital has a significant impact toward firm’s pr ofitability and only human capital

Struktur histologi magnum itik Magelang, itik Tegal dan itik Pengging menunjukkan sel-sel epithel yang berdifferensiasi menjadi sel-sel epithel bersilia dan sel