Jump to content

Rate of convergence

From Wikipedia, the free encyclopedia

In mathematical analysis, particularly numerical analysis, the rate of convergence and order of convergence of a sequence that converges to a limit are any of several characterizations of how quickly that sequence approaches its limit. These are broadly divided into asymptotic rates and orders of convergence, describing how quickly a sequence further approaches its limit once it is already close to it, and non-asymptotic rates and orders of convergence, which describe how quickly sequences approach their limits from starting points that are not necessarily close to their limits.

Asymptotic rates and orders of convergence have particular importance both in practical numerics and in formal proof, and they are the primary focus of this article. In practical numerics, asymptotic rates and orders of convergence follow two common conventions for two types of sequences: the first for sequences of iterations of an iterative numerical method and the second for sequences of successively more accurate numerical discretizations of a target. In formal mathematics, rates of convergence and orders of convergence are often described using asymptotic notation commonly called "big O notation," which can be used to encompass both of the prior conventions.

For iterative methods, a sequence that converges to is said to have order of convergence and rate of convergence if

[1][2]

Where greater methodological precision is required, these rates and orders of convergence are known specifically as the rates and orders of Q-convergence, short for quotient-convergence, since the limit in question is a quotient of error terms.[2] The rate of convergence may also be called the asymptotic error constant. Furthermore some authors will use rate where this article uses order (e.g., [3]).

Similar concepts are used for sequences of discretizations. For instance, ideally the solution of a differential equation discretized via a regular grid will converge to the solution of the continuous differential equation as the grid spacing goes to zero and grid points go to infinity, and if so the rate and order of that convergence are an important characterization of the efficiency of the grid discretization method. A sequence of approximate grid solutions of some problem that converges to a true solution with a corresponding sequence of regular grid spacings that converge to 0 is said to have order of convergence and rate of convergence if

where the absolute value symbols stand for a function metric for the space of solutions such as the uniform norm. Similar considerations also apply for non-grid discretization schemes such as the polygon meshes of a finite element method or the basis sets in computational chemistry: in general, the appropriate definition of will involve the asymptotic limit of the ratio of some approximation error term above to a -power of a discretization scale parameter below.

In practice, the rate and order of convergence of a sequence of iterates or approximations provide useful insights when using iterative methods and discretization methods for calculating numerical approximations. Strictly speaking, however, the asymptotic behavior of a sequence does not give conclusive information about any finite part of the sequence.

Series acceleration is a collection of techniques for improving the rate of convergence of the sequence of partial sums of a series and possibly its order of convergence, also. These accelerations are commonly accomplished with sequence transformations.

Rates of convergence for iterative methods

[edit]

Convergence rate definitions

[edit]

Suppose that the sequence converges to the number . The sequence is said to converge with order to , and with a rate of convergence , if

for some positive constant if and if .[2][4][5] It is not necessary that be an integer. For example, the secant method, when converging to a regular, simple root, has an order of φ ≈ 1.618.[citation needed] This is technically called Q-convergence, short for quotient-convergence, and the rates and orders are called rates and orders of Q-convergence in certain technical settings where alternative rate definitions are more appropriate; see § R-convergence below.

Convergence with order

  • is called linear convergence and the sequence is said to converge linearly to .
  • is called quadratic convergence.
  • is called cubic convergence.

In addition, when for a sequence or for any sequence such that

that sequence is said to converge superlinearly to (i.e., faster than linearly).[2][6] A sequence is said to converge sublinearly to (i.e., slower than linearly) if it converges and

A sequence converges logarithmically to if the sequence converges sublinearly and [7]

R-convergence

[edit]

The definitions of Q-convergence rates have a shortcoming in that they do not naturally capture the convergence behavior of sequences that do not converge with an asymptotically constant rate with every step, such as the staggered geometric progression below that gets closer to its limit only every other step.

In such cases, a closely related but more technical definition of rate of convergence called R-convergence is more appropriate; The "R-" prefix stands for "root."[2][8]: 620  A sequence that converges to is said to converge at least R-linearly if there exists an error-bounding sequence such that and converges Q-linearly to zero; analogous definitions hold for R-superlinear convergence, R-sublinear convergence, R-quadratic convergence, and so on.[2][9]

In order to define the rates and orders of R-convergence, one uses the rate and order of Q-convergence of am error-bounding sequence chosen such that no other error-bounding sequence could have been chosen that would converge with a faster rate and order; any provides a lower bound on the rate and order of R-convergence and the greatest lower bound gives the exact rate and order of R-convergence.

Order estimation

[edit]

A practical method to calculate the order of convergence for a sequence generated by a fixed point iteration is to calculate the following sequence, which converges to the order :[10]

For numerical approximation of an exact value through a numerical method of order q see.[11]

Examples

[edit]

The geometric progression converges to . Plugging the sequence into the definition of Q-linear convergence (i.e., order of convergence 1) shows that

Thus converges Q-linearly with a convergence rate of ; see the first plot of the figure below.

More generally, for any , a geometric progression converges linearly with rate and the sequence of partial sums of a geometric series also converges linearly with rate . The same holds also for geometric progressions and geometric series parameterized by any complex numbers

The staggered geometric progression using the floor function that gives the largest integer that is less than or equal to converges R-linearly to 0 with rate 1/2, but it does not converge Q-linearly; see the second plot of the figure below. The defining Q-linear convergence limits do not exist for this sequence because one subsequence of error quotients (the sequence of quotients taken from odd steps) has a different limit than another subsequence (the sequence of quotients taken from even steps). Generally, for any staggered geometric progression , the sequence will not converge Q-linearly but will converge R-linearly with rate this example highlights why the "R" in R-linear convergence is short for "root."

The sequence converges to zero Q-superlinearly. In fact, it is quadratically convergent with a quadratic convergence rate of 1. It is shown in the third plot of the figure below.

Finally, the sequence converges to zero Q-sublinearly and logarithmically and its convergence is shown as the fourth plot of the figure below.

Plot showing the different rates of convergence for the sequences ak, bk, ck and dk.
Log-linear plots of the example sequences ak, bk, ck, and dk that exemplify linear, linear, superlinear (quadratic), and sublinear rates of convergence, respectively.

Rates of convergence for discretization methods

[edit]

A similar situation exists for discretization methods designed to approximate a function , which might be an integral being approximated by numerical quadrature, or the solution of an ordinary differential equation (see example below). The discretization method generates a sequence , where each successive is a function of along with the grid spacing between successive values of the independent variable . The important parameter here for the rate of convergence to is the grid spacing , inversely proportional to the number of grid points, i.e. the number of points in the sequence required to reach a given value of .

In this case, the sequence is said to converge to the sequence with order q if there exists a constant C such that

This is written as using big O notation.

This is the relevant definition when discussing methods for numerical quadrature or the solution of ordinary differential equations (ODEs).[example needed]

A practical method to estimate the order of convergence for a discretization method is pick step sizes and and calculate the resulting errors and . The order of convergence is then approximated by the following formula:

[citation needed]

which comes from writing the truncation error, at the old and new grid spacings, as

The error is, more specifically, a global truncation error (GTE), in that it represents a sum of errors accumulated over all iterations, as opposed to a local truncation error (LTE) over just one iteration.

Example of discretization methods

[edit]

Consider the ordinary differential equation

with initial condition . We can approximate a solution to this equation using the forward Euler method for numerical discretization:

which implies the first-order linear recurrence with constant coefficients

Given , the sequence satisfying that recurrence is the geometric progression

The exact analytical solution to the differential equation is , corresponding to the following Taylor expansion in for :

In this case, the truncation error is

so converges to with a convergence rate .

Examples (continued)

[edit]

The sequence with was introduced above. This sequence converges with order 1 according to the convention for discretization methods.[why?]

The sequence with , which was also introduced above, converges with order q for every number q. It is said to converge exponentially using the convention for discretization methods. However, it only converges linearly (that is, with order 1) using the convention for iterative methods.[why?]

Recurrent sequences and fixed points

[edit]

Recurrent sequences define discrete dynamical systems and have important general applications in mathematics through various fixed-point theorems about their convergence behavior. When f is continuously differentiable, given a fixed point p, such that , the fixed point is an attractive fixed point and the recurrent sequence will converge at least linearly to p for any starting value sufficiently close to p. If and , then the recurrent sequence will converge at least quadratically, and so on. If , then the fixed point is a repulsive fixed point and sequences cannot converge to p from its immediate neighborhoods, though they may still jump to p directly from outside of its local neighborhoods.

Acceleration of convergence rates

[edit]

Many methods exist to increase the rate of convergence of a given sequence, i.e., to transform one given sequence into a second one that converges more quickly to the same limit. Such techniques are in general known as "series acceleration" methods. These may reduce the computational costs of approximating the limits of the transformed sequences. One example of series acceleration is Aitken's delta-squared process. These methods in general (and in particular Aitken's method) do not increase the order of convergence, and are useful only if initially the convergence is not faster than linear: if converges linearly, one gets a sequence that still converges linearly (except for pathologically designed special cases), but faster in the sense that . On the other hand, if the convergence is already of order ≥ 2, Aitken's method will bring no improvement.

References

[edit]
  1. ^ Ruye, Wang (2015-02-12). "Order and rate of convergence". hmc.edu. Retrieved 2020-07-31.
  2. ^ a b c d e f Nocedal, Jorge; Wright, Stephen J. (1999). Numerical Optimization (1st ed.). New York, NY: Springer. pp. 28–29. ISBN 978-0-387-98793-4.
  3. ^ Senning, Jonathan R. "Computing and Estimating the Rate of Convergence" (PDF). gordon.edu. Retrieved 2020-08-07.
  4. ^ Hundley, Douglas. "Rate of Convergence" (PDF). Whitman College. Retrieved 2020-12-13.
  5. ^ Porta, F. A. (1989). "On Q-Order and R-Order of Convergence" (PDF). Journal of Optimization Theory and Applications. 63 (3): 415–431. doi:10.1007/BF00939805. S2CID 116192710. Retrieved 2020-07-31.
  6. ^ Arnold, Mark. "Order of Convergence" (PDF). University of Arkansas. Retrieved 2022-12-13.
  7. ^ Van Tuyl, Andrew H. (1994). "Acceleration of convergence of a family of logarithmically convergent sequences" (PDF). Mathematics of Computation. 63 (207): 229–246. doi:10.2307/2153571. JSTOR 2153571. Retrieved 2020-08-02.
  8. ^ Nocedal, Jorge; Wright, Stephen J. (2006). Numerical Optimization (2nd ed.). Berlin, New York: Springer-Verlag. ISBN 978-0-387-30303-1.
  9. ^ Bockelman, Brian (2005). "Rates of Convergence". math.unl.edu. Retrieved 2020-07-31.
  10. ^ Senning, Jonathan R. "Computing and Estimating the Rate of Convergence" (PDF). gordon.edu. Retrieved 2020-08-07.
  11. ^ Senning, Jonathan R. "Verifying Numerical Convergence Rates" (PDF). Retrieved 2024-02-09.

Literature

[edit]

The simple definition is used in

The extended definition is used in

  • Walter Gautschi (1997), Numerical analysis: an introduction, Birkhäuser, Boston. ISBN 0-8176-3895-4.
  • Endre Süli and David Mayers (2003), An introduction to numerical analysis, Cambridge University Press. ISBN 0-521-00794-1.

The Big O definition is used in

  • Richard L. Burden and J. Douglas Faires (2001), Numerical Analysis (7th ed.), Brooks/Cole. ISBN 0-534-38216-9

The terms Q-linear and R-linear are used in