fbpx
Wikipedia

Rate of convergence

In numerical analysis, the order of convergence and the rate of convergence of a convergent sequence are quantities that represent how quickly the sequence approaches its limit. A sequence that converges to is said to have order of convergence and rate of convergence if

[1]

The rate of convergence is also called the asymptotic error constant. Note that this terminology is not standardized and some authors will use rate where this article uses order (e.g., [2]).

In practice, the rate and order of convergence provide useful insights when using iterative methods for calculating numerical approximations. If the order of convergence is higher, then typically fewer iterations are necessary to yield a useful approximation. Strictly speaking, however, the asymptotic behavior of a sequence does not give conclusive information about any finite part of the sequence.

Similar concepts are used for discretization methods. The solution of the discretized problem converges to the solution of the continuous problem as the grid size goes to zero, and the speed of convergence is one of the factors of the efficiency of the method. However, the terminology, in this case, is different from the terminology for iterative methods.

Series acceleration is a collection of techniques for improving the rate of convergence of a series discretization. Such acceleration is commonly accomplished with sequence transformations.

Convergence speed for iterative methods

Q-convergence definitions

Suppose that the sequence   converges to the number  . The sequence is said to converge Q-linearly to   if there exists a number   such that

 

The number   is called the rate of convergence.[3]

The sequence is said to converge Q-superlinearly to   (i.e. faster than linearly) in all the cases where   and also the case   if

 .[4]

The sequence is said to converge Q-sublinearly to   (i.e. slower than linearly) if

 

The sequence   converges logarithmically to   if the sequence converges sublinearly and additionally if

 .[5]

Note that unlike previous definitions, logarithmic convergence is not called "Q-logarithmic."

In order to further classify convergence, the order of convergence is defined as follows. The sequence is said to converge with order   to   for   if

 

for some positive constant   if  , and   if  .[6] In particular, convergence with order

  •   is called linear convergence (if  ),
  •   is called quadratic convergence,
  •   is called cubic convergence,
  • etc.

Some sources require that   is strictly greater than   since the   case requires   so is best treated separately.[7] It is not necessary, however, that   be an integer. For example, the secant method, when converging to a regular, simple root, has an order of φ ≈ 1.618.[citation needed]

In the definitions above, the "Q-" stands for "quotient" because the terms are defined using the quotient between two successive terms.[8]: 619  Often, however, the "Q-" is dropped and a sequence is simply said to have linear convergence, quadratic convergence, etc.

Order estimation

A practical method to calculate the order of convergence for a sequence is to calculate the following sequence, which converges to  

 [9]

R-convergence definition

The Q-convergence definitions have a shortcoming in that they do not include some sequences, such as the sequence   below, which converge reasonably fast, but whose rate is variable. Therefore, the definition of rate of convergence is extended as follows.

Suppose that   converges to  . The sequence is said to converge R-linearly to   if there exists a sequence   such that

 

and   converges Q-linearly to zero.[3] The "R-" prefix stands for "root". [8]: 620 

Examples

Consider the sequence

 

It can be shown that this sequence converges to  . To determine the type of convergence, we plug the sequence into the definition of Q-linear convergence,

 

Thus, we find that   converges Q-linearly and has a convergence rate of  . More generally, for any  , the sequence   converges linearly with rate  .

The sequence

 

also converges linearly to 0 with rate 1/2 under the R-convergence definition, but not under the Q-convergence definition. (Note that   is the floor function, which gives the largest integer that is less than or equal to  .)

The sequence

 

converges superlinearly. In fact, it is quadratically convergent.

Finally, the sequence

 

converges sublinearly and logarithmically.

 
Linear, linear, superlinear (quadratic), and sublinear rates of convergence

Convergence speed for discretization methods

A similar situation exists for discretization methods designed to approximate a function  , which might be an integral being approximated by numerical quadrature, or the solution of an ordinary differential equation (see example below). The discretization method generates a sequence  , where each successive   is a function of   along with the grid spacing   between successive values of the independent variable  . The important parameter here for the convergence speed to   is the grid spacing  , inversely proportional to the number of grid points, i.e. the number of points in the sequence required to reach a given value of  .

In this case, the sequence   is said to converge to the sequence   with order q if there exists a constant C such that

 

This is written as   using big O notation.

This is the relevant definition when discussing methods for numerical quadrature or the solution of ordinary differential equations.[example needed]

A practical method to estimate the order of convergence for a discretization method is pick step sizes   and   and calculate the resulting errors   and  . The order of convergence is then approximated by the following formula:

 [citation needed]

which comes from writing the truncation error, at the old and new grid spacings, as

 

The error   is, more specifically, a global truncation error (GTE), in that it represents a sum of errors accumulated over all   iterations, as opposed to a local truncation error (LTE) over just one iteration.

Example of discretization methods

Consider the ordinary differential equation

 

with initial condition  . We can solve this equation using the Forward Euler scheme for numerical discretization:

 

which generates the sequence

 

In terms of  , this sequence is as follows, from the Binomial theorem:

 

The exact solution to this ODE is  , corresponding to the following Taylor expansion in   for  :

 

In this case, the truncation error is

 

so   converges to   with a convergence rate  .

Examples (continued)

The sequence   with   was introduced above. This sequence converges with order 1 according to the convention for discretization methods.[why?]

The sequence   with  , which was also introduced above, converges with order q for every number q. It is said to converge exponentially using the convention for discretization methods. However, it only converges linearly (that is, with order 1) using the convention for iterative methods.[why?]

Recurrent sequences and fixed points

The case of recurrent sequences   which occurs in dynamical systems and in the context of various fixed point theorems is of particular interest. Assuming that the relevant derivatives of f are continuous, one can (easily) show that for a fixed point   such that  , one has at linear convergence for any starting value   sufficiently close to p. If   and  , then one has at least quadratic convergence, and so on. If  , then one has a repulsive fixed point and no starting value will produce a sequence converging to p (unless one directly jumps to the point p itself).

Acceleration of convergence

Many methods exist to increase the rate of convergence of a given sequence, i.e. to transform a given sequence into one converging faster to the same limit. Such techniques are in general known as "series acceleration". The goal of the transformed sequence is to reduce the computational cost of the calculation. One example of series acceleration is Aitken's delta-squared process. (It should be noted, though, that these methods in general (and in particular Aitken's method) do not increase the order of convergence, and are useful only if initially the convergence is not faster than linear: If   convergences linearly, one gets a sequence   that still converges linearly (except for pathologically designed special cases), but faster in the sense that  . On the other hand, if the convergence is already of order ≥ 2, Aitken's method will bring no improvement.)

References

  1. ^ Ruye, Wang (2015-02-12). "Order and rate of convergence". hmc.edu. Retrieved 2020-07-31.
  2. ^ Senning, Jonathan R. "Computing and Estimating the Rate of Convergence" (PDF). gordon.edu. Retrieved 2020-08-07.
  3. ^ a b Bockelman, Brian (2005). "Rates of Convergence". math.unl.edu. Retrieved 2020-07-31.
  4. ^ Arnold, Mark. "Order of Convergence" (PDF). University of Arkansas. Retrieved 2022-12-13.{{cite web}}: CS1 maint: url-status (link)
  5. ^ Van Tuyl, Andrew H. (1994). "Acceleration of convergence of a family of logarithmically convergent sequences" (PDF). Mathematics of Computation. 63 (207): 229–246. doi:10.2307/2153571. JSTOR 2153571. Retrieved 2020-08-02.
  6. ^ Hundley, Douglas. "Rate of Convergence" (PDF). Whitman College. Retrieved 2020-12-13.{{cite web}}: CS1 maint: url-status (link)
  7. ^ Porta, F. A. (1989). "On Q-Order and R-Order of Convergence" (PDF). Journal of Optimization Theory and Applications. 63 (3): 415–431. doi:10.1007/BF00939805. S2CID 116192710. Retrieved 2020-07-31.
  8. ^ a b Nocedal, Jorge; Wright, Stephen J. (2006). Numerical Optimization (2nd ed.). Berlin, New York: Springer-Verlag. ISBN 978-0-387-30303-1.
  9. ^ Senning, Jonathan R. "Computing and Estimating the Rate of Convergence" (PDF). gordon.edu. Retrieved 2020-08-07.

Literature

The simple definition is used in

The extended definition is used in

  • Walter Gautschi (1997), Numerical analysis: an introduction, Birkhäuser, Boston. ISBN 0-8176-3895-4.
  • Endre Süli and David Mayers (2003), An introduction to numerical analysis, Cambridge University Press. ISBN 0-521-00794-1.

The Big O definition is used in

  • Richard L. Burden and J. Douglas Faires (2001), Numerical Analysis (7th ed.), Brooks/Cole. ISBN 0-534-38216-9

The terms Q-linear and R-linear are used in; The Big O definition when using Taylor series is used in

rate, convergence, numerical, analysis, order, convergence, rate, convergence, convergent, sequence, quantities, that, represent, quickly, sequence, approaches, limit, sequence, displaystyle, that, converges, displaystyle, said, have, order, convergence, displ. In numerical analysis the order of convergence and the rate of convergence of a convergent sequence are quantities that represent how quickly the sequence approaches its limit A sequence x n displaystyle x n that converges to x displaystyle x is said to have order of convergence q 1 displaystyle q geq 1 and rate of convergence m displaystyle mu if lim n x n 1 x x n x q m displaystyle lim n rightarrow infty frac left x n 1 x right left x n x right q mu 1 The rate of convergence m displaystyle mu is also called the asymptotic error constant Note that this terminology is not standardized and some authors will use rate where this article uses order e g 2 In practice the rate and order of convergence provide useful insights when using iterative methods for calculating numerical approximations If the order of convergence is higher then typically fewer iterations are necessary to yield a useful approximation Strictly speaking however the asymptotic behavior of a sequence does not give conclusive information about any finite part of the sequence Similar concepts are used for discretization methods The solution of the discretized problem converges to the solution of the continuous problem as the grid size goes to zero and the speed of convergence is one of the factors of the efficiency of the method However the terminology in this case is different from the terminology for iterative methods Series acceleration is a collection of techniques for improving the rate of convergence of a series discretization Such acceleration is commonly accomplished with sequence transformations Contents 1 Convergence speed for iterative methods 1 1 Q convergence definitions 1 1 1 Order estimation 1 2 R convergence definition 1 3 Examples 2 Convergence speed for discretization methods 2 1 Example of discretization methods 2 2 Examples continued 3 Recurrent sequences and fixed points 4 Acceleration of convergence 5 References 6 LiteratureConvergence speed for iterative methods EditQ convergence definitions Edit Suppose that the sequence x k displaystyle x k converges to the number L displaystyle L The sequence is said to converge Q linearly to L displaystyle L if there exists a number m 0 1 displaystyle mu in 0 1 such that lim k x k 1 L x k L m displaystyle lim k to infty frac x k 1 L x k L mu The number m displaystyle mu is called the rate of convergence 3 The sequence is said to converge Q superlinearly to L displaystyle L i e faster than linearly in all the cases where q gt 1 displaystyle q gt 1 and also the case q 1 m 0 displaystyle q 1 mu 0 if lim k x k 1 L x k L q m displaystyle lim k to infty frac x k 1 L x k L q mu 4 The sequence is said to converge Q sublinearly to L displaystyle L i e slower than linearly if lim k x k 1 L x k L 1 displaystyle lim k to infty frac x k 1 L x k L 1 The sequence x k displaystyle x k converges logarithmically to L displaystyle L if the sequence converges sublinearly and additionally if lim k x k 2 x k 1 x k 1 x k 1 displaystyle lim k to infty frac x k 2 x k 1 x k 1 x k 1 5 Note that unlike previous definitions logarithmic convergence is not called Q logarithmic In order to further classify convergence the order of convergence is defined as follows The sequence is said to converge with order q displaystyle q to L displaystyle L for q 1 displaystyle q geq 1 if lim k x k 1 L x k L q M displaystyle lim k to infty frac x k 1 L x k L q M for some positive constant 0 lt M lt displaystyle 0 lt M lt infty if q gt 1 displaystyle q gt 1 and 0 lt M lt 1 displaystyle 0 lt M lt 1 if q 1 displaystyle q 1 6 In particular convergence with order q 1 displaystyle q 1 is called linear convergence if M lt 1 displaystyle M lt 1 q 2 displaystyle q 2 is called quadratic convergence q 3 displaystyle q 3 is called cubic convergence etc Some sources require that q displaystyle q is strictly greater than 1 displaystyle 1 since the q 1 displaystyle q 1 case requires M lt 1 displaystyle M lt 1 so is best treated separately 7 It is not necessary however that q displaystyle q be an integer For example the secant method when converging to a regular simple root has an order of f 1 618 citation needed In the definitions above the Q stands for quotient because the terms are defined using the quotient between two successive terms 8 619 Often however the Q is dropped and a sequence is simply said to have linear convergence quadratic convergence etc Order estimation Edit A practical method to calculate the order of convergence for a sequence is to calculate the following sequence which converges to q displaystyle q q log x k 1 x k x k x k 1 log x k x k 1 x k 1 x k 2 displaystyle q approx frac log left frac x k 1 x k x k x k 1 right log left frac x k x k 1 x k 1 x k 2 right 9 R convergence definition Edit The Q convergence definitions have a shortcoming in that they do not include some sequences such as the sequence b k displaystyle b k below which converge reasonably fast but whose rate is variable Therefore the definition of rate of convergence is extended as follows Suppose that x k displaystyle x k converges to L displaystyle L The sequence is said to converge R linearly to L displaystyle L if there exists a sequence e k displaystyle varepsilon k such that x k L e k for all k displaystyle x k L leq varepsilon k quad text for all k and e k displaystyle varepsilon k converges Q linearly to zero 3 The R prefix stands for root 8 620 Examples Edit Consider the sequence a k 1 1 2 1 4 1 8 1 16 1 32 1 2 k displaystyle a k left 1 frac 1 2 frac 1 4 frac 1 8 frac 1 16 frac 1 32 ldots frac 1 2 k right It can be shown that this sequence converges to L 0 displaystyle L 0 To determine the type of convergence we plug the sequence into the definition of Q linear convergence lim k 1 2 k 1 0 1 2 k 0 lim k 2 k 2 k 1 1 2 displaystyle lim k to infty frac left 1 2 k 1 0 right left 1 2 k 0 right lim k to infty frac 2 k 2 k 1 frac 1 2 Thus we find that a k displaystyle a k converges Q linearly and has a convergence rate of m 1 2 displaystyle mu 1 2 More generally for any c R m 1 1 displaystyle c in mathbb R mu in 1 1 the sequence c m k displaystyle c mu k converges linearly with rate m displaystyle mu The sequence b k 1 1 1 4 1 4 1 16 1 16 1 4 k 2 displaystyle b k left 1 1 frac 1 4 frac 1 4 frac 1 16 frac 1 16 ldots frac 1 4 left lfloor frac k 2 right rfloor ldots right also converges linearly to 0 with rate 1 2 under the R convergence definition but not under the Q convergence definition Note that x displaystyle lfloor x rfloor is the floor function which gives the largest integer that is less than or equal to x displaystyle x The sequence c k 1 2 1 4 1 16 1 256 1 65 536 1 2 2 k displaystyle c k left frac 1 2 frac 1 4 frac 1 16 frac 1 256 frac 1 65 536 ldots frac 1 2 2 k ldots right converges superlinearly In fact it is quadratically convergent Finally the sequence d k 1 1 2 1 3 1 4 1 5 1 6 1 k 1 displaystyle d k left 1 frac 1 2 frac 1 3 frac 1 4 frac 1 5 frac 1 6 ldots frac 1 k 1 ldots right converges sublinearly and logarithmically Linear linear superlinear quadratic and sublinear rates of convergenceConvergence speed for discretization methods EditThis section needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed August 2020 Learn how and when to remove this template message This section may require cleanup to meet Wikipedia s quality standards The specific problem is There appears to be a mixture of defining convergence with regards to grid points n displaystyle n and with step size h displaystyle h Section should be modified for consistency and include an explanation of alternative equivalent definitions Please help improve this section if you can August 2020 Learn how and when to remove this template message A similar situation exists for discretization methods designed to approximate a function y f x displaystyle y f x which might be an integral being approximated by numerical quadrature or the solution of an ordinary differential equation see example below The discretization method generates a sequence y 0 y 1 y 2 y 3 displaystyle y 0 y 1 y 2 y 3 where each successive y j displaystyle y j is a function of y j 1 y j 2 displaystyle y j 1 y j 2 along with the grid spacing h displaystyle h between successive values of the independent variable x displaystyle x The important parameter here for the convergence speed to y f x displaystyle y f x is the grid spacing h displaystyle h inversely proportional to the number of grid points i e the number of points in the sequence required to reach a given value of x displaystyle x In this case the sequence y n displaystyle y n is said to converge to the sequence f x n displaystyle f x n with order q if there exists a constant C such that y n f x n lt C h q for all n displaystyle y n f x n lt Ch q text for all n This is written as y n f x n O h q displaystyle y n f x n mathcal O h q using big O notation This is the relevant definition when discussing methods for numerical quadrature or the solution of ordinary differential equations example needed A practical method to estimate the order of convergence for a discretization method is pick step sizes h new displaystyle h text new and h old displaystyle h text old and calculate the resulting errors e new displaystyle e text new and e old displaystyle e text old The order of convergence is then approximated by the following formula q log e new e old log h new h old displaystyle q approx frac log e text new e text old log h text new h text old citation needed which comes from writing the truncation error at the old and new grid spacings as e y n f x n O h q displaystyle e y n f x n mathcal O h q The error e displaystyle e is more specifically a global truncation error GTE in that it represents a sum of errors accumulated over all n displaystyle n iterations as opposed to a local truncation error LTE over just one iteration Example of discretization methods Edit Consider the ordinary differential equation d y d x k y displaystyle frac dy dx kappa y with initial condition y 0 y 0 displaystyle y 0 y 0 We can solve this equation using the Forward Euler scheme for numerical discretization y n 1 y n h k y n displaystyle frac y n 1 y n h kappa y n which generates the sequence y n 1 y n 1 h k displaystyle y n 1 y n 1 h kappa In terms of y 0 y 0 displaystyle y 0 y 0 this sequence is as follows from the Binomial theorem y n y 0 1 h k n y 0 1 n h k n n 1 h 2 k 2 2 displaystyle y n y 0 1 h kappa n y 0 left 1 nh kappa n n 1 frac h 2 kappa 2 2 right The exact solution to this ODE is y f x y 0 exp k x displaystyle y f x y 0 exp kappa x corresponding to the following Taylor expansion in h k displaystyle h kappa for h k 1 displaystyle h kappa ll 1 f x n f n h y 0 exp k n h y 0 exp k h n y 0 1 h k h 2 k 2 2 n y 0 1 n h k n 2 h 2 k 2 2 displaystyle f x n f nh y 0 exp kappa nh y 0 left exp kappa h right n y 0 left 1 h kappa frac h 2 kappa 2 2 right n y 0 left 1 nh kappa frac n 2 h 2 kappa 2 2 right In this case the truncation error is e y n f x n n h 2 k 2 2 O h 2 displaystyle e y n f x n frac nh 2 kappa 2 2 mathcal O h 2 so y n displaystyle y n converges to f x n displaystyle f x n with a convergence rate q 2 displaystyle q 2 Examples continued Edit The sequence d k displaystyle d k with d k 1 k 1 displaystyle d k 1 k 1 was introduced above This sequence converges with order 1 according to the convention for discretization methods why The sequence a k displaystyle a k with a k 2 k displaystyle a k 2 k which was also introduced above converges with order q for every number q It is said to converge exponentially using the convention for discretization methods However it only converges linearly that is with order 1 using the convention for iterative methods why Recurrent sequences and fixed points EditThe case of recurrent sequences x n 1 f x n displaystyle x n 1 f x n which occurs in dynamical systems and in the context of various fixed point theorems is of particular interest Assuming that the relevant derivatives of f are continuous one can easily show that for a fixed point f p p displaystyle f p p such that f p lt 1 displaystyle f p lt 1 one has at linear convergence for any starting value x 0 displaystyle x 0 sufficiently close to p If f p 0 displaystyle f p 0 and f p lt 1 displaystyle f p lt 1 then one has at least quadratic convergence and so on If f p gt 1 displaystyle f p gt 1 then one has a repulsive fixed point and no starting value will produce a sequence converging to p unless one directly jumps to the point p itself Acceleration of convergence EditMany methods exist to increase the rate of convergence of a given sequence i e to transform a given sequence into one converging faster to the same limit Such techniques are in general known as series acceleration The goal of the transformed sequence is to reduce the computational cost of the calculation One example of series acceleration is Aitken s delta squared process It should be noted though that these methods in general and in particular Aitken s method do not increase the order of convergence and are useful only if initially the convergence is not faster than linear If x n displaystyle x n convergences linearly one gets a sequence a n displaystyle a n that still converges linearly except for pathologically designed special cases but faster in the sense that lim a n L x n L 0 displaystyle lim a n L x n L 0 On the other hand if the convergence is already of order 2 Aitken s method will bring no improvement References Edit Ruye Wang 2015 02 12 Order and rate of convergence hmc edu Retrieved 2020 07 31 Senning Jonathan R Computing and Estimating the Rate of Convergence PDF gordon edu Retrieved 2020 08 07 a b Bockelman Brian 2005 Rates of Convergence math unl edu Retrieved 2020 07 31 Arnold Mark Order of Convergence PDF University of Arkansas Retrieved 2022 12 13 a href Template Cite web html title Template Cite web cite web a CS1 maint url status link Van Tuyl Andrew H 1994 Acceleration of convergence of a family of logarithmically convergent sequences PDF Mathematics of Computation 63 207 229 246 doi 10 2307 2153571 JSTOR 2153571 Retrieved 2020 08 02 Hundley Douglas Rate of Convergence PDF Whitman College Retrieved 2020 12 13 a href Template Cite web html title Template Cite web cite web a CS1 maint url status link Porta F A 1989 On Q Order and R Order of Convergence PDF Journal of Optimization Theory and Applications 63 3 415 431 doi 10 1007 BF00939805 S2CID 116192710 Retrieved 2020 07 31 a b Nocedal Jorge Wright Stephen J 2006 Numerical Optimization 2nd ed Berlin New York Springer Verlag ISBN 978 0 387 30303 1 Senning Jonathan R Computing and Estimating the Rate of Convergence PDF gordon edu Retrieved 2020 08 07 Literature EditThe simple definition is used in Michelle Schatzman 2002 Numerical analysis a mathematical introduction Clarendon Press Oxford ISBN 0 19 850279 6 The extended definition is used in Walter Gautschi 1997 Numerical analysis an introduction Birkhauser Boston ISBN 0 8176 3895 4 Endre Suli and David Mayers 2003 An introduction to numerical analysis Cambridge University Press ISBN 0 521 00794 1 The Big O definition is used in Richard L Burden and J Douglas Faires 2001 Numerical Analysis 7th ed Brooks Cole ISBN 0 534 38216 9The terms Q linear and R linear are used in The Big O definition when using Taylor series is used in Nocedal Jorge Wright Stephen J 2006 Numerical Optimization 2nd ed Berlin New York Springer Verlag pp 619 620 ISBN 978 0 387 30303 1 Retrieved from https en wikipedia org w index php title Rate of convergence amp oldid 1128285608, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.