a derivation of extrapolation algorithms based on error ... · and which gave rise to...

22
JOURNA, OF COMPUTATIONAL AND APPUED MATHEMATICS ELSEVIER Journal of Computational and Applied Mathematics 66 (1996) 5-26 A derivation of extrapolation algorithms based on error estimates Claude Brezinski*, Ana C. Matos Laboratoire d'Analyse Numbrique et d'Optimisation, UFR IEEA-M3, Universitk des Sciences et Technologies de Lille, 59655 Villeneuve d'Ascq Cedex, France Received 20 April 1994; revised 13 November 1994 Abstract In this paper, we shall emphasize the role played by error estimates and annihilation difference operators in the construction of extrapolation processes. It is shown that this approach leads to a unified derivation of many extrapola- tion algorithms and related devices, to general results about their kernels and that it opens the way to many new algorithms. Their convergence and acceleration properties could also be studied within this framework. Keywords." Extrapolation methods; Convergence acceleration A M S classification: 65B05; 65B10 1. Introduction The first methods used for accelerating the convergence of sequences were the linear summation processes which go back to Euler, Cesaro, Hausdorff and others; see [42], for example. Among them is also the Richardson extrapolation process based on polynomial extrapolation at zero [43] and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear convergence acceleration method was Aitken's A2 process [1]. It was generalized by Shanks in 1955 [47] and, one year later, Wynn produced his e-algorithm [56] for its implementation. Since then, many other extrapolation algorithms have been proposed and studied; see [20, 54] for a review and 1-16] for some history. Quite a general framework has been constructed along the years for the theory of extrapolation methods which, nowadays, lies on a firm basis; see [28, 54] and the first chapter of 1-20]. The situation is quite different for the practical construction of extrapolation methods, that is for the algorithms and, up to now, there was no systematic way for deriving them. Thus, in survey * Corresponding author. E-mail: [email protected]. 0377-0427/96/$15.00 © 1996 Elsevier Science B.V. All rights reserved SSDI 0377-0427(95)001 57-3

Upload: others

Post on 17-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

JOURNA, OF COMPUTATIONAL AND APPUED MATHEMATICS

ELSEVIER Journal of Computational and Applied Mathematics 66 (1996) 5-26

A derivation of extrapolat ion algorithms based on error estimates

Claude Brezinski*, Ana C. Matos

Laboratoire d'Analyse Numbrique et d'Optimisation, UFR IEEA-M3, Universitk des Sciences et Technologies de Lille, 59655 Villeneuve d'Ascq Cedex, France

Received 20 April 1994; revised 13 November 1994

Abstract

In this paper, we shall emphasize the role played by error estimates and annihilation difference operators in the construction of extrapolation processes. It is shown that this approach leads to a unified derivation of many extrapola- tion algorithms and related devices, to general results about their kernels and that it opens the way to many new algorithms. Their convergence and acceleration properties could also be studied within this framework.

Keywords." Extrapolation methods; Convergence acceleration

AMS classification: 65B05; 65B10

1. Introduction

The first methods used for accelerating the convergence of sequences were the linear summation processes which go back to Euler, Cesaro, Hausdorff and others; see [42], for example. Among them is also the Richardson extrapolation process based on polynomial extrapolation at zero [43] and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear convergence acceleration method was Aitken's A 2 process [1]. It was generalized by Shanks in 1955 [47] and, one year later, Wynn produced his e-algorithm [56] for its implementation. Since then, many other extrapolation algorithms have been proposed and studied; see [20, 54] for a review and 1-16] for some history.

Quite a general framework has been constructed along the years for the theory of extrapolation methods which, nowadays, lies on a firm basis; see [28, 54] and the first chapter of 1-20].

The situation is quite different for the practical construction of extrapolation methods, that is for the algorithms and, up to now, there was no systematic way for deriving them. Thus, in survey

* Corresponding author. E-mail: [email protected].

0377-0427/96/$15.00 © 1996 Elsevier Science B.V. All rights reserved SSDI 0 3 7 7 - 0 4 2 7 ( 9 5 ) 0 0 1 5 7 - 3

Page 2: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

6 C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26

papers or books on the subject (see the bibliography at the end of the paper), each algorithm was usually presented separately from the others with no logical link between them.

It is our purpose, in this paper, to propose a systematic approach to extrapolation algorithms and their construction. In fact, such an approach was already used in [49, 50] for Levin's transforms, but its formalism is due to Weniger [51]. It is based on remainder (or error) estimates and annihilation operators. We think that this approach is a very interesting and powerful one and that it has not been sufficiently exploited. In this paper, we shall develop this point of view and show its impact on the subject. We shall not enter here into all the details (they will be treated in subsequent publications) but we only intend to open the way. As we shall see below, this approach leads to a better understanding of the mechanism of extrapolation algorithms, it gives us a frame- work where all the processes actually known can be included find it also provides us with new algorithms and new theoretical results.

2. The scenery

Let us begin by some definitions. Sequences will be denoted by letters without any index and their terms by the same letter (unless

indicated) with a subscript. If u = (u,) and v = (v,), we shall make use of the notation u/v = (u,,/v,,) and a similar notation for the product. A sequence will always start with the index O.

Definition 1 (Wimp [55, p. 196]). Let 5 p be the set of complex sequences. A difference operator L is a linear mapping of 5 ¢ into itself:

L : u = (u.) L (u ) =

Such an operator L is represented by an infinite matrix or, in other terms, by the sequence (L,) of linear forms mapping u into the nth term of the sequence L(u). This is why, for briefness, the nth number (L(u)), of the sequence L(u) will be denoted by L,(u) or, sometimes, by L(u,). The notation L = l (for example L = A) means that Vn, L,(u) = l(u.) (for example, Vn, L,(u) = Au,).

It is well known [3] that the most general difference operator is defined by

q,

L , ( u ) = Z G,(n)u,+i, i = - p ,

where p. and q, are nonnegative integers which can eventually depend on n, ui -- 0 for i < 0, and the Gfs are given functions of n which can also depend on auxiliary fixed sequences. (In theory, with some supplementary assumptions on the Gi(n)'s, q, can be infinite. However, in practical situations, we shall only consider the finite case.) It must be clearly understood that if the auxiliary sequences on which the Gfs could depend also depend on some terms of the sequence (u.) itself, then these terms arefixed in the Gfs and thus the operator L is still a linear one. In other terms, for defining L, we must first choose the fixed sequence u which is used in the Gi's and then keep the same Gfs for all the sequences to which L is applied. Thus we have L(u + v) = L(u) + L(v) since, after choosing the sequence u which enters into the definition of the auxiliary sequences G~, these G~'s are kept fixed. Such an assumption is usual for extrapolation algorithms; see [17].

Page 3: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26 7

This remark is very important since, in fact, it allows us to use the Toeplitz theorem for proving the convergence of the transformed sequence to the same limit as the initial sequence. However, when the Gi(n) depend on the sequence to be transformed, the convergence is ensured only for the sequence under consideration and not for all converging sequences as is the case for linear summation processes when the Gi(n)'s do not depend on the initial sequence. This point will be discussed in more detail in the sequel.

Definition 2 (Weniger [51, p. 212]). L is called an annihilation difference operator for the sequence a = (an) if 3N such that Vn ~ N, Ln(a) = O.

We can assume without any loss of generality, that N = 0.

Definition 3 (Weniger [51, p. 212]). The sequence (Dn) is called a remainder (or error) estimate of the sequence (Sn) if Vn, S~ - S, = anDn where (an) is an unknown sequence and S~ a (usually unknown) number. If (S,) converges to So, then So is called its limit and, otherwise, its antilimit.

As explained in [20], the first step in the construction of an extrapolation method is to assume that the sequence under consideration has a certain behavior. In other terms, one should construct an algorithm able to find the exact limit (or antilimit) So of certain sequences. This will be achieved by using an annihilation difference operator, as in [51], but with a slightly generalized approach.

We assume that the sequence S = (Sn) satisfies V n,

Sac - Sn = anDn,

where a = (an) is an unknown sequence and D = (Dn) a known one. We want to construct a sequence transformation T:(Sn)~--~(Tn) such that 3N, Vn/> N, Tn = Sac. We assume also that 3b = (bn) such that L(b) is known and L is an annihilation difference operator for a - b. Thus, we have

S. a n

D.

S~,D

Dn

and

S~

D. Sn

b. = a. - b.. D.

Applying L to both sides of this relation and using its linearity property, we obtain Vn/> N

So~L.(1/D) -- L.(S/D) - L.(b) = L.(a -- b) = 0

and it follows that

L.(S/D) + L.(b) Soc =

L.(1/D)

Thus, if we define the transformation T:(S.)~--~(Tn) by

L.(S/D) + Ln(b) T n - -

L.(1/D)

Page 4: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

8 C. Brezinski, A,C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26

then, by construction, Vn >/N, T, = S~ if and only if Vn >~ N, S~ - S, = a , D , . We recall the

Definition 4 (Brezinski and Redivo Zaglia [20, p. 4]). Let T : (S,) ~ (T.) be a sequence transforma- tion. The kerne l o,~- of T is the set of sequences such that 3S~, 3N, Vn >~ N, T, = S~.

Remark. I f a sequence e = (~,) and an operator L are known such that L ( e a - b) = 0, then we can write

- s . =

and thus

L . ( a S / D ) + L . ( b ) S ~ =

L.(o~/O )

This remark will be useful later. The simplest form of the transformation T is obtained when b = (0). Unless specified (see Sections

3.1, 3.9, 3.14 and 3.15), we shall only consider this case. Obviously, if the sequence (D,) is invariant under translation of the sequence (S,) (that is, if it

remains the same when a constant is added to all the terms of (S,)), then the transformation T is translative (that is, the same constant is added to all the terms of (T,)). As proved in [14], a necessary and sufficient condition for that property to hold is that T, could be written as

T , = F , ( S o , ... , Sk(,))

with

L ( X o , . . .

r , ( x o . . . . , Xk(,)) = D~,(Xo, ~.. , Xg(,))

and DZf , identically zero, where D f , denotes the sum of the partial derivatives off , with respect to Xo, . . . , Xk(,). The remark that all the translative sequence transformations can be written under this form was originally made by Benchiboun [4]. The transformation T, as obtained above, is also homogeneous which means that if, when all the terms of(S,) are multiplied by a nonzero constant c, the D,'s become c D , , then the T,'s are also multiplied by the same constant c. A transformation which is translative and homogeneous is called quasi-linear.

The vector, matrix and confluent cases can also be included into this framework. If S~, S, and a, are vectors and D. numbers, then we have

S ~ L , ( 1 / D ) - L , ( S / D ) = L , (a ) ,

where L, is applied componentwise. If, now, S~, S,, a, and D, are p x p complex matrices and 5 ~ is the set of sequences of complex p x p matrices, then two cases have to be considered, the right case and the left one as for Pad6 approximants for matrix series [2, Vol. 2, pp. 50if].

We have to assume that L(c~u.) = ~ L ( u , ) in the right case and that L ( u , ~ ) = L ( u , ) ~ in the left case where ~ is an arbitrary matrix.

(i) R i g h t case: We write

S ~ - S , = a , D , , i.e., S ~ D ~ 1 _ S , D ~ 1 = a, .

Page 5: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26

Thus

S ~ L ( D ; x) _ L ( S , D ; x) = L (a , )

and the transformation T is defined by

T , = L ( S . D ~ 1) [L(D~- 1)] - 1

(ii) Left case: We write

So - S , = D,a , , i.e., D~-1So~ - D,-1S, = a,.

Thus

L ( D ; 1 ) S ~ - L ( D [ I S , ) = L (a , )

and the transformation T is defined by

T , = [ L ( D ; 1)]- 1 L ( D ; 1Sn).

These questions are discussed in detail in [-24]. Let us now consider the confluent base where S, a and D are functions of a variable t. We have

So - S( t ) = a(t)O(t).

If L is a linear operator on the set of functions, then

S ~ L ( 1 / D ( t ) ) - L (S ( t ) /D( t ) ) = L(a( t ) )

and the transformation T is given by

T( t ) = L(S( t ) /O( t ) ) /L (1 /O( t ) ) .

Obviously, it is also possible to consider the vector and matrix confluent cases.

3. Some extrapolation algorithms

Let us now consider different choices of the linear operator L and some particular choices of the error estimates (D,). We shall see that they lead to different well-known transformations and we shall also propose some natural generalizations.

Sections 3.1-3.8 concern a linear operator which is independent of the sequence (S,). In Sections 3.9-3.11, L depends on (S,). The case where L is a linear combination of several other operators is considered in Sections 3.12 and 3.13. Finally, generalizations with a correction factor are treated in Sections 3.14 and 3.15.

3.1. A simple transformation

The simplest difference operator is the identity. In this case, the most general form of the transformation T leads to

T , = S, + b,D, .

Page 6: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

10 C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5 26

The kernel of this transformation is the set of sequences of the form S~ - S, = b,D,. Such a transformation was considered in [21] where some theoretical results about it can be found.

3.2. The O-procedure

If Vn, a, = a, b, = 0, and if L = A, the forward difference operator, then we have

D,+ I S, - D,S,+ I AS, T , = = S, - - - D , .

D,+ a - D, AD,

This transformation was considered in [10]. It is called the O-procedure. When D, = x, - x, where (x,) is a known sequence converging to a known limit x, then the

second standard process of Germain-Bonne [30] is recovered. For some particular choices of the sequence (x,), new acceleration processes based on convergence tests for sequences and error estimates can be constructed as introduced in [15] and developed in [34].

3.3. Summation processes

If L,(u) = Y~"=o Gi(n)ui where the Gi(n)'s are given numbers, if L(b) = 0, and if Vn, D, = 1, then T is a linear summat ion process as defined, for example, in [54]. The convergence of the sequence (T,) to S~ for all sequences (S,) converging to S~ is given by the Toeplitz theorem. Their acceleration properties were studied in [53].

3.4. Column and diagonal transformations

If b, = 0, D, = AS, and L = A k, then for k = 1 we obtain Aitken's A 2 process, while k = 2 corresponds to the second column of the 0-algorithm [5]. The case of an arbitrary value of k was considered by D r u m m o n d [29].

Because of this example, let us discuss the case where the linear forms defining L also depend on a second integer k. We shall denote them by L,,a and consider, as before, the difference operator L defined by the sequence (L, = L,,k), h,: a fixed value of k or the operator obtained by reversing the roles of n and k, that is the operator L' defined by the sequence (L~, = L,,k)k for a fixed value of n. We shall see, in the sequel, many other examples of such linear forms.

Thus, for the above operator, we obtain two alternatives: (i) k f i xed and n varying: In this case we shall consider the sequence L = A k as above (that is

L.(u) = Aau,). We obtain, for n = 0, 1, . . . ,

T R

and we (ii) n

=

and we

= dk(S./ZIS.)/a (1/ZIS.)

shall speak about a column transformation. f ixed and k varying: In that case we shall consider the sequence L~ = zl g (that is ,4au,) which gives, for k = O, 1, ... ,

= Ak(S. /AS.)Mk(1MS.)

shall speak about a diagonal transformation.

Page 7: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26 11

The reasons for these names are clearly understood if we set

T~n)= Ak(Sn/ASn) /Ak(1 /ASn)

and if we display these quantities into a two-dimensional table as follows:

T,o o, T(o 1) T~ °) 7"'o2, °, T(o 3) T~ 2) T(21) T(3 °)

So, n indicates the minimal index of the sequence S which is used in the computat ion of T~ n) while the index k is related to the number of terms of S needed and, thus, it is, in some sense, a measure of the complexity of the sequence transformation.

Obviously, it is also possible to define other difference operators by letting n and k vary arbitrarily•

3.5. Lev in ' s t ransforms

Ifbn = 0 and ifa. is a polynomial of degree k - 1 in n + 1, then it is well known that L = d k is an annihilation difference operator for (an). This remark is the basis used by Levin [33] for construct- ing his various sequence transformations based on different choices of the sequence (Dn) and by Weniger [51] for extending them. Levin's transforms can be generalized by taking, for an, a polynomal of degree k - 1 in xn. In that case, the annihilation difference operator is the divided difference operator 6k that will be defined in the next subsection and Levin's transformations appear as a generalization of Richardson's. The generalization of the Richardson process introduc- ed by Sidi [48] can also be put into this framework and Drummond 's method [29] as well. The connections between these processes are discussed in more detail in [51, pp. 236-238, 20, pp. 116-119]. Levin's transforms will be discussed again in Sections 4.1 and 4.2.

3.6. R ichardson extrapolat ion

Let us now consider the well-known Richardson extrapolation procedure [43]. It consists in assuming that

S . = S o - c lxn ckx~,

where (x,) is a given auxiliary sequence. Thus

S o - S , = (cl + ... + c , xk , -~ )x , ,

which corresponds to D, = x, and a, = cl + ... + CkXkn -1. The divided difference operator •k of order k at the points xi is an annihilation operator for the sequence (a.). Let us recall that this

Page 8: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

1 2 C. Brezinsla', A.C. Matos /Journal o f Computational and Applied Mathematics 66 (1996) 5 -26

operator is recursively defined by

1) - 6 k + l ( u . ) -

Xn+k+ 1 - - X n

with 6o(U,) = u,. Thus, we are in the case, described in Section 3.4, of an operator depending on a second index k. Using the same notations as above, we can construct a two-dimensional array by

T~,) _ 6k(S,/x,) 6k(1/X.)

with Tt0 ") = S,. By construction, if S~ - S, has the form above, then Vn, Tk t") = S~. We shall now prove that these numbers are identical to the numbers Utk ") constructed by the

Richardson extrapolation scheme

X lr r ( n ) n + k U k - 1 ) = - x . V _+l 1)

Xn+ k - - X n

with U~0 ") = S,. By using the recurrence relation between divided differences, it is easy to prove by induction that

6a(1/x,) = ( - 1)R/(x, "" X.+k).

We can also prove by induction that

6k(S,/x,)-- (--1)k U~ "), X n . . . Xn+ k

where Utk ") is the quantity computed by the Richardson scheme. The property is true for k = 0. Assuming that it is true for k, we have

• T(n + 1) , T(n) X n ' " X n + k [ . , I k - - X n + 1 " " X n + k + l ( - ) k

6k+,(S,/x,) = (-- 1) k -...-ZW-----.-2Y- . . . . . . . . . . . . • XnXn+ l " " Xn+kXn+k+ l (Xn+k + l - - Xn)

Thus, using the Richardson scheme, we obtain 6k+~(S,/x,)= 6k+ ~(1/X,) tr(")~k+ 1 which shows that Tk ~") = Utk "). A similar derivation can be found in [51, pp. 246-247]. The expression for the divided differences of 1/x, is also given [39, p. 8].

Thus, the Richardson extrapolation method can also be implemented via the recurrence relations for divided differences as already mentioned in [48]. However, the Richardson scheme is simpler because it is a relation between the Tk~")'s themselves instead of two separate recurrence relations for their numerators and their denominators. If the Richardson extrapolation method has to be applied simultaneously to several sequences with the same auxiliary sequences (x,) then, since the denominators are the same, it could be preferable to use the preceding scheme based on divided differences.

Using the usual determinantal formula for divided differences and the definitions of Tk t") given above, we recover the expression of these quantities as a ratio of two determinants, see [20, p. 72]. It will probably be possible to derive new extrapolation schemes by using the recurrence relationship for generalized divided differences obtained by M/ihlbach [40].

Page 9: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinski, A.C. Matos /Journal of Computational and Applied Mathematics 66 (1996) 5-26 13

3.7. Overholt's process

The Overholt process [41] is defined by

A ( v ( n ) / ( A S n + k ) k+ 1) v(n) k + l A (1/(AS.+k) k+ ')

with V(o ") = S.. For this algorithm, it is known that the quantities Vk (") can be expressed as a ratio of determinants but these determinants have not yet been found [26].

F rom the above expression, we immediately see that Vn, --k+l~"(") = So if and only if there exists a constant Ck such that

S o - V~ ") = c k ( A S . + k ) k+ ~.

Thus, assuming that the constant ck is replaced by a polynomial of degree mk -- 1 in n, we obtain a generalization of Overholt 's process:

v ( n ) A m ~ ( V ( k n ) / ( A S n + k ) k + 1) k + l = Amk(1/(ASn+k)k+l)

If Ck is replaced by a polynomial in x., then divided differences have to be used instead of the operator A.

3.8. The E-algorithm

Let us now consider some more difficult examples where, in particular, the sequences Gi in the definition (given above) of the most general difference operator can depend on some terms of the sequence (S.).

Let us take

L.(u) = L'k(U) =

I Un Uk + x . . . Un + k [

g~(n) gl(n + 1) ... gx(n + k)

I • . . ,

gk(n) gk(n + 1) .-. gk(n + k)

where the (gi(n)) are given auxiliary sequences which can depend on (S,) itself and define T as above. If we set

M.(u) =

Un Uk+ 1 "'" U n + k

gl(n)D, gl(n + l)D.+l ... gl(n + k)D.+k

gk(n)D, gk(n + 1)D.+x "'" gk(n + k)D.+k

Page 10: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

14 C. Brezinsla', A.C. M a t o s / J o u r n a l o f Computa t ional and App l i ed Mathemat ics 66 (1996) 5 - 2 6

then we also have

T. - m L . ( S / D ) M . ( S )

L . ( 1 / O ) M.(1)

and a similar proper ty for L~,. For more properties of this type, see [17]. If Vn, D, = 1 and S, = x ", then the formal b ior thogonal polynomials defined in [19] are

recovered. If D, = 1, if b, = y k= 1 cig~(n), where the c~'s are arbitrary numbers , and if the operator L is defined by the sequence (L,) of linear forms given above, then the t ransformat ion T corres- ponds to the kth co lumn of the E-algori thm [9, 31] with the (g~(n))'s as auxiliary sequences. If L' is defined by the sequence of linear forms (L~,), then T corresponds to the nth diagonal of the E-algori thm. Another approach to the E-algor i thm and some related algori thms can be found in [23]. It is also related to annihi lat ion difference operators, but does not make use of determinants in their definition•

The E-algori thm is the most general existing extrapolat ion algorithm. Almost all the sequence t ransformations actually known are part icular cases of the E-algori thm and they can be recovered by different choices of the auxilary sequences g~. We shall now study one of them in more details.

Let us take

L. (u ) = L i (u ) =

Un Un+ l . . . Un+ k

/ I S n A S h + 1 "'" ASn+k

A S n + k - 1 A S n + k . . . A S n + 2 k - 1

If D. = 1, if b, = y k=lc iAS,+i_ 1, where the ci's are arbitrary numbers , and if the operator L is defined by the sequence (L,), then T is the kth co lumn of the Shanks t ransformat ion [47], or, in other words, the 2kth co lumn of the e-algorithm of Wynn [56]. If L' is defined by the sequence (L~,), then T corresponds to the nth diagonal of the Shanks t ransformat ion or the e-algorithm. The first and second generalizations of the e-algorithm [6] can also be put into this f ramework since they correspond to replacing the opera tor A by a more general one [7, 45].

3.9. Pad~ and Padk- type approximants

Let us consider, in the e-algorithm, the part icular case S = (S,) where S, = Co + c~z + ... + c ,z ~

(that is the nth partial sum of the series S~ = f ( z ) = Co + c l z + c2z2+ ...). Since we have 1 Zn + 1 S~ - S, = c,+ + . . . , then it corresponds to D, = 1 and we shall take

Lp(u) = L'q(u) =

ZqUp _ q zq - 1 lip _ q + 1 • • • IAp

C p - q + l C p - q + 2 . . . Cp+ 1

ep Cp+ l . . . Cp+q

with the convent ion that c~ = 0 for i < 0.

Page 11: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C Brezinski, A.C. Matos/Journal o f Computational and Applied Mathematics 66 (1996) 5-26 15

If b, = 52~= l C p + i : +i, and if the operator L is defined by the sequence (Lp), then Tp is the [p /q] Pad+ approximant of the series f at the point z [25]. We obtain

f ( z ) L ( 1 ) - L ( S ) = L((Cp+lZ "+1 + ... )p).

Using the definition of L, we easily see that

L ( ( c p + 1 Z p + I 4- " '" )p) =

zq(Cp_q+ lZp -q+ .~_ . . . ) "'" Cp+ l Zp+ I -~- . . .

Cp-q+ 1 "'" Cp+ 1

Cp """ Cp+q

= ( zqO(z ,+ 1))

since the terms in z p÷ 1, ... ,zp+q in the first row disappear by linear combination with the other rows. This is the usual approximation-through-order property of the Padb approximants and we have, for a fixed value of q,

Lp(S ) /Lp (1 ) = [p /q]s ( z ) .

Of course, the same property holds if p is fixed and q changes. If we set

L , ( u ) = a(ok)zku. + a ~ ) z k - l u . + , + . " + a~k)u.+k,

then, for b. = 0 and D. = 1, it is easy to see that

f ( z ) L . ( 1 ) - L . ( S ) = O ( : +k+ 1),

which shows that L . ( S ) / L . ( 1 ) is the ( n + k / k ) Pad6-type approximant of f with Vk(Z) = a~ ) + "" + a~k)z k as its generating polynomial [8] since L.(1) is a polynomial of degree k and L . ( S ) is a polynomial of degree n + k.

3.10. O-algorithm

The 0-algorithm [5] can be written as

A gD(n + I)/F~(n) I, V2k /t'~2k + 1 )

O~k)+ Z = A(1/D~)k+I)

with

o(n) / ] (n+ 1) r~(n) 2 k + l : t " 2 k - 1 -[- L"2k,

D~n) = 1/(0~ n+ l) - - O~ n))

and 0~)1 = 0, 0(o ") = S.. Thus, the 0-algorithm can also be put into our framework since it corresponds to an iterative use

of our basic procedure with S = (0~k + 1)), D = (D~k)+ 1) and b = (0). From theoretical reasons, it is known [26] that the quantities n(.) V2k can be expressed as a ratio of two determinants but these determinants have not yet been found.

Page 12: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

16 C. Brezinski, A.C. Matos /Journal o f Computational and Applied Mathematics 66 (1996) 5-26

From the above expression, we immediately see that Vn, t,,2k+2 t)(n) ~--- So0 if and only if there exists a constant Ck such that

Soo 0 (2n ; 1) ~ t3(n) (n) - - = t . k V 2 k + l = Ck/AOzk+l.

Thus, assuming that the constant CR is replaced by a polynomial of degree mk -- 1 in n, we obtain a generalization of the 0-algorithm:

/U2k + 1) O~)k + 2 : mk (n)

A (1/Dzk+I)

If CR is replaced by a polynomial in x,, then divided differences have to be used instead of the operator A.

3.11. p-algorithm

Extrapolation at infinity by a rational function consists in assuming that

S~xk + clxk -1 + ... + ck S. = k b l x k - 1 x . + + "" + bk

It leads to a slight generalization of the p-algorithm of Wynn [57] whose quantities are defined by a ratio of two determinants as follows (we only indicate their first rows):

k - ,(,) I1 S, x . x , S , ... x k-1 x . iS . xkS, I 1 k - ~'2k= l1 S. x . x , S , ... x k- x , 1S, xkl

It corresponds to D, = 1 and L,(u) = I1 S, x , x . S . . . , x , x , S , These ratios of determinants can be recursively computed by the p-algorithm whose rules are

p(") p~-+,') + (x,+k+ l - x,)l(p~ "+1) p~")) k+l~---

with P~)x = 0 and p(o ") -- S,. These quantities are the usual reciprocal differences which play, in rational interpolation and in continued fractions, the role of divided differences in polynomial interpolation.

3.12. Composite transformations

Let L (1), L (2), . . . ,L (k) be difference operators and (bl ")) given sequences of numbers. We shall define the difference operator L by

k L , = ~ v,h!")l. (i)-.

i = 1

and the transformation T as usual by T, = L, (S /D) /L , (1 /D) . Defining the transformation T(i) ' (S,) ~ (T~ ")) by

T[")= L(.I)(S/D)/L~i)(1/D).

Page 13: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinsla', A.C. Matos /Journal o f Computational and Applied Mathematics 66 (1996) 5-26 17

we have k

c (n) Y (") T.=y , i = l

with

c(.) h!.) L(.i)(1/D) i = v , L.(1/D)"

Thus k

_(n) 1 Z Ci

i=1

and we recover the composite sequence transformations introduced in [13]. We want to choose the bl")'s so that the kernel of the transformation T contains the kernels of the

t ransformat ionsT ~). So let us assume that 3i, L")(a) = 0. In order that, Vn, T, = So, we must have -t") 0. Since, in that case, T} ") So~ and v~h !") :~ 0 and Vj -¢ i, vj/,~.") = 0. Thus, c-~-t"J = 1 and Vj :~ i, c~ = =

A T} "~ = 0, a possible choice for L, when V n, D, = 1, is given by

L~" "" L~ k'

AT~ ") ... AT~ ") E l l ~ o

AT~ n'+k-2) ... AT~ n'+k-2)

This transformation, given in [13], generalizes Shanks' which is recovered for the choice T~ ") = S, +i_ 1.

A possible generalization consists in taking

k

L,(u) = • bl")L.(u/D~i)), i=1

where the D")'s are known sequences.

3.13. Least-squares extrapolation

T,, as defined in Section 2, is the value min imiz ingf ( t ) = (tL,(1/D) - L,(S/D)) 2. Let us now define T, as the value minimizing

k

f ( t ) = Z [tL~,°(1/D) -L~,')(S/D)] 2, i = 0

where L tl), L t2 ) , . . . , L (k) are difference operators. We obtain

El:ok L~i)(1/D)L~i)(S/D ) T n = k

o 2

If the difference operators are taken as L~,i)(u) = L, + i(u), where L is some difference operator, then we recover an extrapolation procedure, called extrapolation in the least-squares sense, which was studied by Cordellier [27] (see also [11]).

Page 14: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

18 C. Brezinski, A. C MaWs~Journal of Computational and Applied Mathematics 66 (1996) 5-26

If we define the operator L' by

k

= L . (1 /D)L . (u), L'~(u) ~. ,) (i) i = 0

then T, = L'~(S/D)/L' ,(1/D) which shows that such an extrapolation process can also be considered as the composi te sequence transformation corresponding to the choice bl ") = L~°(1/D).

3.14. Cauchy- type approximants

If we define ft, by

ft, = -- L , ( b ) / L , ( S / D ) ,

then the most general form of the transformation T, as given in Section 2 with an auxiliary sequence b, can be written:

L , ( S / D ) T, = (1 - ft,) L , ( 1 / D ) '

which shows that we have, in fact, introduced a correction factor in its simplest form (correspond- ing to b = (0)).

Let us see now how, using this expression, the Cauchy-type approximants can be put into our framework. Let f ( z )=~ ,F=oCi zi be a formal power series and f , ( z ) its partial sums. Let g(z) = ~F=odiz i be a known auxiliary series with partial sums g,(z). We set r ,(z) = g(z) - g,(z). The Cauchy-type approximants [18] are defined by

_ _ -- L,i=oa i ' C.(z ) - h . (z) ~'" z i g(z) g(z)

where h.(z) = ZT=o a f is the nth partial sum of the se r i e s f ( z )g ( z ) . Thus

ai = doci + dlCi-1 + "'" + dico

and C.(z ) can also be written as

. . . . B(n)f Iz~ C.(z ) B{o"' fo(z ) + B~"'f,(z) + + . j . , ,

with BI "} -- d ,_ i z " - i / g ( z ) for i -- 0, ... ,n. Taking D, -- 1 and defining the operator L by

Ln(g ) = ~ dn- i zn- iu i , i = o

we obtain

C.(z) = (1 -- ft.) - - L.(S/D) L . ( 1 / D )

with ft. -- r.(z) /g(z) . It must be noticed that (ft.) tends to 0 when n goes to infinity. Acceleration properties of the Cauchy-type approximants were studied in [35-37].

Page 15: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinsla', A.C. Matos /Journal of Computational and Applied Mathematics 66 (1996) 5-26 19

3.15. Error control

Procedures for controlling the error in convergence acceleration methods were introduced in [-12-1. They consist in setting

Ln(S /D) Ln(a) Tn - S - - - S - - - enD',

L,(1/D) Ln(1 /D)

with en = Ln(a) and D'. = 1/Ln(1/D). Then, a second sequence transformation T * is defined by

Ln(S /D) T* = (1 - fin) Ln(1 /D)

= T n - 7nO'n = Tn(~n)

with ~'n = f lnLn(S/D). It can be proved that, under certain assumptions, the limit So of the sequence (Sn) belongs to the interval with bounds Tn(Tn) and T n ( - 7n).

In our case, if

A Tn 1

AD'. 1 + D'n+ 1/D'n

and if

then

lim Ln(1 /D) _ lim Ln+l(a______) 4: +_ 1, ._.~ L n + I ( 1 / D ) n--o~ Ln(a)

T* -- S lim - - - 0. n - ~ T n - - S

Moreover, if we set

T*(e) = Tn -- (1 + e)ynD'n,

then, under the preceding assumptions, Ve :~ 0, 3N, Vn ~> N, So belongs to the interval whose bounds are T * ( e ) and T*( - e).

These results can be applied to the Cauchy-type approximants introduced in the preceding example. These approximants are recovered by the choice

r.(z) - - - h n ( z )

o(z) and we have

lim C . ( z ) - f ( z ) _ 0 n-*~ fn(Z) -- f ( z )

if

lim rn(Z) = 1. ,-~ o~ On(Z) -- hn ( z ) / f ( z )

Page 16: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

20 C Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26

4. Composition of operators

In this section it will be easier to denote by L(Un) the nth term of the sequence L(u). We shall use again several difference operators denoted by L ").

4.1. A par t i cu lar case

As in Section 2, let us assume that

S ~ - Sn = anDn.

In partical situations, most of the time, an annihilation operator for the sequence (an) is not known. Thus we begin by choosing an operator L (1). Applying it to the previous relation leads to a first sequence transformation defined by

T~ n) = L(1)(Sn/Dn)/L(1)(1/Dn)

and we have

Soo - T~ n) = L(1)(an)/L(1)(1/Dn).

Setting a(m n) = L(1)(an) and D~ n) = 1/L(1)(1/Dn), we have

S~ -- T~ n' = a~')O~ n'.

Let us now choose another difference operator L (2) (which can be the same as the first one) and apply it to the previous relation. It leads to a second sequence transformation defined by

T(2n) = L(Z)( T~n)/D(ln))/L(2)(1/D(1 n))

and we have

S~o -- T (n) = L(2)(a(ln))/L(2)(1/m(1 n))

It is easy to see that

T~n) L(Z)L(1)(Sn/Dn)

and so on. Thus we have the following iterative use of the procedure. We set

T(o n) = Sn and D(o n) = Dn.

Then, for k = O, 1, . . . , we set

T(n) L(k+ 1)(T2n)/o~ n))

k + l = L(g+I)(I/D(k,))

and

D(n) 1/L(k + i)(1/D~n)). k + l "~-

Page 17: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 H996) 5-26

We have

S~ Ten) ~(") 13(") - - jt k + 1 = ~ k + l U k + l

with

a(.) L(i+ 1)(a~")). k + l

It can be easily checked that

T~ ") = L (k) ''" L(1)(S. /D.) L(k). . . L(x)(1/D.) '

a~ ") = L(k). . . L(1)(a.)

and

Dtk ") = 1/Ltk). . . L(1)(1/D,).

For example, if V k, L (k) is the opera tor A, we have

T'"' T~"' - A T~ ") D~. ) k + a = AD~.-----~

where A operates on the superscripts n, and

O(n) r~(n) r'~(n + 1) / A 1~(n) k + l ~ - - " t ' k L " k / z ~ L " k •

21

This is a part icular case of the a lgor i thm (30) given by Homeier [32] when his quantit ies A rtk ") are all taken equal to 1.

As another applicat ion of this iterative procedure, let us consider the family of sequence t ransformat ions defined by

T~k ") = L(k)(Sn/Dn) L(k)(1/D,) '

where the difference operators L (k) satisfy the following recurrence relation:

L (k+ 1)(u.) = ~k)L(k)(u. + 1 ) + fl~.k)L(k)(u.)

with L~°)(u.) = u., (a~*)) and (fl~k)) being given auxiliary sequences of numbers. Then, we can write

L (k+ 1)(S./D.) C~.k)L(k)(s.+ 1/D.+ 1 ) + fl~.k)L(k)(S./D.) T ( n ) = =

--k+l L(k+I)(1/D.) a~k)L(k)(1/D.+l ) + fl~k)L(k)(1/D.)

o~(k) l ( k ) l l / r ~ ~,-F,(n+ 1 ) . ~ ~'/ '- ' .+lJ--k + fl~.k)L(k)(1/D.)T~ ")

a~'L(*)(1/Dn+ 1) + flt.k)L(k)(1/Dn)

So we get

T(n) k + l - -

~ ( k + 1,( Ttkn)/D~kn) )

~ ( k + i )(1/D~k)) '

Page 18: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

22 C Brezinski, A. C Matos /Journal of Computational and Applied Mathematics 66 (1996) 5-26

with ,~(k)

£f(k+ 1)(Un) = 1 + p,, Un,

D(n k) = 1/L(k)(1/Dn) = 1/~f(kLLf(k- 1)... ~f(1)(1/D,)"

We can easily see that

D(n) 1/¢.~(k + 1)(1/D~n)), k+l

which shows that T(n) Xk+ 1 can be obtained by composing difference operators. As an example, let us consider Levin's transforms introduced in Section 3.5. The algorithm for

their implementation given in [50] enables us to write

T(k n) = Ak(( n + 1) k- ISn/Dn) __ L(k)(sn/nn) Ak((n + 1) k- 1/Dn) L(k)(1/On)

with

L(k+l)(Un) = L(k)(Un+I) (n + 1)(n + k + 1) k-1 - (n + k + 2 ) k L(k)(Un), L(°)(Un) = Un

and Dn chosen according to the Levin's transform under consideration. We immediately see that this algorithm fits into the previous iterative procedure by choosing

0~(k) 1 and fl~) (n + 1)(n + k + 1) k - 1 n = = - (n + k + 2) k

4.2. The general case

It is also possible to iterate as before, but with arbitrary D(kn)'s no more related to the D~ )- :'s by the above recurrence relation. This is the most general case and it includes, for example, the E-algorithm since we have

E(n) = A ( E ~ ) 1/g(kn)1 ,k)

(1/g~)- ,.k)

with E(o ") = Sn and the quantities Uk,i "(") computed by

1,,/gk- 1,k) ~k,i A (n) (1/gk- 1,k)

with .(n) o,i = g i ( n ) .

Overholrs process (Section 3.7) and the 0-algorithm (Section 3.10) also fit into this category. Let us set

b~n) = 1

1) '

a~n) ~(n) r(k) l_(n)

Page 19: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinsla', A.C Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26 23

We have

T i n) = L(k)(b~) a L(k- 1)(... L(Z)(b(ln)L(1)(Sn/Dn) ... )

L(k)(b~ )- 1Ltk- 1)(... L(2)(b]n)L(1)(1/Dn) ... )

Let us denote, in this general case, by T , the transformation (Sn )~ (Tk (n)) for a fixed value of k. The kernel of such a composed transformation was studied in [22] where the following results were proved.

Theorem 5. The kernel o f Tk+ 1 contains the kernel o f Tk.

Theorem 6. The kernel o f Tk is the set o f sequences such that Vn,

S~ - Sn = anDn with L(k)(b~& i L(g- 1)(... L(2)(b~n)LtX)(an)... ) = O.

When the operators L (i) are all identical to A then, by the theory developed in [26], we know that the quantities T~ ") can be expressed as a ratio of two determinants. However, these determinants are only known in a few particular cases and for many transformations they still have to be found.

Let us now give three examples. (i) We consider the sequence of transformations given by

T(o n) = Sn,

T(n) = T~n, ~n(AT~n)) 2 k + l

an+ IAT~ "+ x) -- anAT~ ")"

These transformations were studied by Matos [34] in the cases an = n and an = n log n. It is easy to see that they can be obtained by composition of operators in two different ways. First of all, these transformations can be written as

T(o n) = Sn,

,r(n) _ L(T~n)/D~ n)) " l k + 1

L(1/D~ n))

with L = A and D(k ") = ariA T(k "). The same sequence of transformations can also be obtained as in Section 4.1 with

D~ ~+1) D~.) L(k+l)(Un) = A,rtn+ l) Un+ a anAT(k m Un an+ 1 z3 a k

and

D(n) 1/L(k + 1)(1/D~n)). k + l

(ii) The transformations defined by

T(o n) = Sn,

A T(k n) T(n) = Ttk n) k + l

P k + l - - 1

Page 20: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

24 C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26

are recovered by the choice L = A and Dtk")= p~,+ i. Such t ransformations were in t roduced by Matos [34]. They were proved to accelerate the convergence of m o n o t o n e sequences such that, Vn,

A S . = a lp] + a2Pn2 + "'"

with 1 > pl > P2 > "'" > 0 and Vi , ai # O, that is, we have, Vk,

"-(") S)/(Ttk ") S) O. l i m { l k + l - - - - ~--- n--~ 09

(iii) Let us consider again Levin's t ransforms as defined in the previous section. If we set 7~ k) = n . . . (n + k) = (n)k+ 1, the P o c h h a m m e r symbol, then we can write

L (k+ 1)(u.) = A(Ltk)(u.)7~k)(n + k ) k - i ) , l'~k- 1 ,(k) (n + k + . j Y.+l

which gives

T(k. ) _ Ltk)(S./D.) _ A ((L (k- 1)(S. /D.) /L(k- 1)(1/D.))7(. k- 1)(n + k - 1) k- 2Ltk- i)(1/D.))

L(k)(1/D.) A (?~- 1)(n + k - 1) k- 2L(k- 1)(1/D,))

Setting

D~) = 1 1 ~)(k- 1)(n + k - 1) k - 2L(k- 1)(1/D.)'

we get

Tkt, ) = A (Tkt~ 1/D(kn)-1 ) A(1/D~)-i)

In this case, the Dtk")'s are not related by D~ ") = 1/A(1/D~)-I) but it is easy to see that they satisfy

Dtk. ) = 1 n(n + k)A(1/D~)- i ) '

which shows that Levin's t ransforms also fit into the general case with

b~k ") = n(n + k).

I terat ion of sequence t ransformations is considered in [52] and their kernels are studied in [22].

5. Conclusions

As can be seen from what precedes, the f ramework developed in this paper is quite powerful and interesting and it could certainly be extended furthermore. In particular, an impor tan t question will be to study the convergence and acceleration properties of extrapolat ion algori thms within this formalism in order to obtain more general results than those actually known. Some results in this direction have already been obtained by Matos [38]. The mechanism int roduced in [23] has been extended to the vector and matr ix cases in [24]. Including in this mechanism the recent interpreta- t ion of the vector e-algorithm of Wynn [58] obtained by Salam [46] is under consideration.

Page 21: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

C. Brezinski, A.C. Matos/Journal of Computational and Applied Mathematics 66 (1996) 5-26 25

Acknowledgements

W e would like to t h a n k the two referees for their careful r ead ing of the mancusc r ip t and for suggest ing several improvemen t s .

This w o r k was s u p p o r t e d by the H u m a n Capi ta l and Mobi l i t y P r o g r a m m e of the E u r o p e a n C o m m u n i t y , P ro jec t R O L L S , C o n t r a c t C H R X - C T 9 3 - 0 4 1 6 ( D G 12 CO MA ) .

References

[1] A.C. Aitken• •n Bern•u••i• s numeri•a• s••uti•n •f a•gebraic equati•ns• Proc. Roy. Soc. Edinburgh 46 ( •926) 289-3•5. [2] G.A. Baker Jr. and P.R. Graves-Morris, Pad6 Approximants, 2 Vols. (Addison-Wesley, Reading, MA, 1981). [3] S. Banach, Throrie des Operations Linraires, Monografie Matematyczne, Vol. 1 (Warszawa, 1932). [4] M.D. Benchiboun, Etude de certaines grn~ralisations du A 2 d'Aitken et comparaison de procadrs d'accalrration de

la convergence, Thrse 3~me cycle, Universit~ des Sciences et Technologies de Lile, 1987. [5] C. Brezinski, Acehl~ration de suites ~ convergence logarithmique, C.R. Acad. Sci. Paris 273A (1971) 727-730. [6] C. Brezinski, Etudes sur les e et p-algorithmes, Numer. Math. 17 (1971) 153-162. [7] C. Brezinski, Conditions d'application et de convergence de pro~d~s d'extrapolation, Numer. Math. 20 (1972)

64 -79. [8] C. Brezinski, Padr-Type Approximation and General Orthogonal Polynomials, Internat. Ser. Numer. Math., Vol. 50

(Birkhfiuser, Basel, 1980). [9] C. Brezinski, A general extrapolation algorithm, Numer. Math. 35 (1980) 175-187.

[10] C. Brezinski, Some new convergence accelerating methods, Math. Comp. 39 (1982) 133-145. [11] C. Brezinski, Algorithm 585: a subroutine for the general interpolation and extrapolation problems, ACM Trans.

Math. Software 8 (1982) 290-301. [12] C. Brezinski, Error control in convergence acceleration processes, IMA J. Number. Anal. 3 (1983) 65-80. [13] C. Brezinski, Composite sequence transformations, Numer. Math. 46 (1985) 311-321. [14] C. Brezinski, Quasi-linear extrapolation processes, in: R.P. Agarwal et al., Eds., Numerical Mathematics, Singapore,

1988, Internat. Ser. Numer. Math., Voi. 86 (Birkh~iuser, Basel, 1988) 61-78. [15] C. Brezinski, A new approach to convergence acceleration methods, in: A. Cuyt, Ed., Nonlinear Numerical Methods

and Rational Approximation (Reidel, Dordrecht, 1988) 373-405. [16] C. Brezinski, A survey ofiterative extrapolation by the E-algorithm, Det Kon#. Norske Vid. Selsk. skr. 2 (1989) 1-26. [17] C. Brezinski, Algebraic properties of the E-transformation, in: Numerical Analysis and Mathematical Modelling,

Banach Center Publications, Vol. 24 (PWN, Warsaw, 1990) 85-90. [18] C. Brezinski, The asymptotic behavior of sequences and new series transformations based on the Cauchy product,

Rocky Mountain J. Math. 21 (1991) 71-84. [19] C. Brezinski, Biorthogonality and its Applications to Numerical Analysis (Marcel Dekker, New York, 1992). [20] C. Brezinski and M. Redivo Zaglia, Extrapolation Methods. Theory and Practice (North-Holland, Amsterdam,

1991). [21] C. Brezinski and M. Redivo Zaglia, Construction of extrapolation processes, Appl. Numer. Math. 8 (1991) 11-23. [22] C. Brezinski and M. Redivo-Zaglia, On the kernel of sequence transformations, Appl. Numer. Math. 16 (1-2) (1994)

239-244. [23] C. Brezinski and M. Redivo-Zaglia, A general extrapolation procedure revisited, Adv. Comput. Math. 2 (1994)

461-477. [24] C. Brezinski and A. Salam, Matrix and vector sequence transformations revisited, Proc. Edinburgh Math. Soc. 38

(1995) 495-510. [25] C. Brezinski and J. Van Iseghem, Pad6 approximations, in: P.G. Ciarlet and J.L. Lions, Eds., Handbook of

Numerical Analysis, Vol. III (North-Holland, Amsterdam, 1994) 47-222. [26] C. Brezinski and G. Walz, Sequences of transformations and triangular recursion schemes with applications in

numerical analysis, J. Comput. Appl. Math. 34 (1991) 361-383.

Page 22: A derivation of extrapolation algorithms based on error ... · and which gave rise to Romberg's method for accelerating the convergence of the trapezoidal rule [44]. The first nonlinear

26 C. Brezinski, A.C. Matos /Journal of Computational and Applied Mathematics 66 (1996) 5-26

[27] F. Cordellier, Interpolation rationnelle et autres questions: Aspects algorithmiques et num6riques, Th~se d'Etat, Universit~ des Sciences et Technologies de Lille, 1989.

[28] J.P. Delahaye, Sequence Transformations (Springer, Berlin, 1988). [29] J.E. Drummond, A formula for accelerating the convergence of a general series, Bull. Austral. Math. Soc. 6 (1972)

69 -74. [30] B. Germain-Bonne, Estimation de la limite de suites et formalisation de proc6d6s d'acc~16ration de la convergence,

Th6se d'Etat, Universit6 des Sciences et Technologies de Lille, 1978. [31] T. H~tvie, Generalized Neville type extrapolation schemes, BIT 19 (1979) 204-213. [32] H.H.H. Homeier, Some applications of nonlinear convergence accelerators, Internat. J. Quantum Chem. 45 (1993)

545-562. [33] D. Levin, Development of non-linear transformations for improving convergence of sequences, lnternat. J. Comput.

Math. B3 (1973) 371-388. [34] A.C. Matos, Acceleration methods based on convergence tests, Numer. Math. 58 (1990) 329-340. [35] A.C. Matos, Construction of new transformations for lacunary power series based on the Cauchy-type approxi-

mants, Appl. Numer. Math. 7 (1991) 493-507. [36] A.C. Matos, On the choice of the denominator in Pad6 and Cauchy-type approximation, in: C. Brezinski et al., Eds.,

Orthogonal Polynomials and their Applications (Baltzer, Basel, 1991) 347-352. [37] A.C. Matos, Some new acceleration methods for periodic-linearly convergent power series, BIT 31 (1991) 686-696. [38] A.C. Matos, Linear difference operators and acceleration methods, to appear. [39] L.M. Milne Thomson, The Calculus of Finite Differences (Chelsea, New York, 1954). [40] G. Mfihlbach, A recurrence formula for generalized divided differences and some applications, J. Approx. Theory

9 (1973) 165-172. [41] K.J. Overholt, Extended Aitken acceleration, BIT 5 (1965) 122-132. [42] R. Powell and S.M. Shah, Summability Theory and its Applications (Van Nostrand Reinhold, London, 1972;

Prentice-Hall of India, New Delhi, 2nd revised ed., 1988). [43] L.F. Richardson, The approximate arithmetic solution by finite differences of physical problems involving

differential equations, with an application to the stress in a masonery dam, Philos. Trans. Roy. Soc. London Ser. A210 (1910) 307-357.

[44] W. Romberg, Vereinfachte numerische Integration, Kgl. Norske Vid. Selsk. Forsk. 28 (1955) 30-36. [45] A. Salam, On a generalization of the e-algorithm, J. Comput. Appl. Math. 46 (1993) 455-464. [46] A. Salam, Non-commutative extrapolation algorithms, Numer. Algorithms 7 (1994) 225-251. [47] D. Shanks, Non linear transformations of divergent and slowly convergent sequences, J. Math. Phys. 34 (1955)

1-42. [48] A. Sidi, An algorithm for a special case of a generalization of the Richardson extrapolation process, Numer. Math.

38 (1982) 299-307. [49] A.C. Smith, A one-parameter method for accelerating the convergence of sequences and series, Comput. Math. Appl.

4 (1978) 157-171. [50] D.A. Smith and W.F. Ford, Acceleration of linear and logarithmic convergence, SIAM J. Numer. Anal. 16 (1979)

223-240. [51-] E.J. Weniger, Nonlinear sequence transformations for the acceleration of convergence and the summation of

divergent series, Comput. Phys. Reports 10 (1989) 189-371. [52] E.J. Weniger, On the derivation of iterated sequence transformations for the acceleration of convergence and the

summation of divergent series, Comput. Phys. Commun. 64 (1991) 19-45. [53] J. Wimp, Toeplitz arrays, linear sequence transformations and orthogonal polynomials, Numer. Math. 23 (1974)

1-17. [54] J. Wimp, Sequence Transformations and their Applications (Academic Press, New York, 1981). [55] J. Wimp, Computation with Recurrence Relations (Pitman, Boston, 1984). [-56] P. Wynn, On a device for computing the em(Sn) transformation, MTAC 10 (1956) 91-96. [57] P. Wynn, On a procrustean technique for the numerical transformation of slowly convergent sequences and series,

Proc. Cambridge Philos. Soc. 52 (1956) 663-671. [58] P. Wynn, Acceleration techniques for iterated vector and matrix problems, Math. Comp. 16 (1962) 301-322.