fbpx
Wikipedia

Iterative method

In computational mathematics, an iterative method is a mathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems, in which the n-th approximation is derived from the previous ones.

A specific implementation with termination criteria for a given iterative method like gradient descent, hill climbing, Newton's method, or quasi-Newton methods like BFGS, is an algorithm of the iterative method. An iterative method is called convergent if the corresponding sequence converges for given initial approximations. A mathematically rigorous convergence analysis of an iterative method is usually performed; however, heuristic-based iterative methods are also common.

In contrast, direct methods attempt to solve the problem by a finite sequence of operations. In the absence of rounding errors, direct methods would deliver an exact solution (for example, solving a linear system of equations by Gaussian elimination). Iterative methods are often the only choice for nonlinear equations. However, iterative methods are often useful even for linear problems involving many variables (sometimes on the order of millions), where direct methods would be prohibitively expensive (and in some cases impossible) even with the best available computing power.[1]

Attractive fixed points edit

If an equation can be put into the form f(x) = x, and a solution x is an attractive fixed point of the function f, then one may begin with a point x1 in the basin of attraction of x, and let xn+1 = f(xn) for n ≥ 1, and the sequence {xn}n ≥ 1 will converge to the solution x. Here xn is the nth approximation or iteration of x and xn+1 is the next or n + 1 iteration of x. Alternately, superscripts in parentheses are often used in numerical methods, so as not to interfere with subscripts with other meanings. (For example, x(n+1) = f(x(n)).) If the function f is continuously differentiable, a sufficient condition for convergence is that the spectral radius of the derivative is strictly bounded by one in a neighborhood of the fixed point. If this condition holds at the fixed point, then a sufficiently small neighborhood (basin of attraction) must exist.

Linear systems edit

In the case of a system of linear equations, the two main classes of iterative methods are the stationary iterative methods, and the more general Krylov subspace methods.

Stationary iterative methods edit

Introduction edit

Stationary iterative methods solve a linear system with an operator approximating the original one; and based on a measurement of the error in the result (the residual), form a "correction equation" for which this process is repeated. While these methods are simple to derive, implement, and analyze, convergence is only guaranteed for a limited class of matrices.

Definition edit

An iterative method is defined by

 

and for a given linear system   with exact solution   the error by

 

An iterative method is called linear if there exists a matrix   such that

 

and this matrix is called the iteration matrix. An iterative method with a given iteration matrix   is called convergent if the following holds

 

An important theorem states that for a given iterative method and its iteration matrix   it is convergent if and only if its spectral radius   is smaller than unity, that is,

 

The basic iterative methods work by splitting the matrix   into

 

and here the matrix   should be easily invertible. The iterative methods are now defined as

 

From this follows that the iteration matrix is given by

 

Examples edit

Basic examples of stationary iterative methods use a splitting of the matrix   such as

 

where   is only the diagonal part of  , and   is the strict lower triangular part of  . Respectively,   is the strict upper triangular part of  .

  • Richardson method:  
  • Jacobi method:  
  • Damped Jacobi method:  
  • Gauss–Seidel method:  
  • Successive over-relaxation method (SOR):  
  • Symmetric successive over-relaxation (SSOR):  

Linear stationary iterative methods are also called relaxation methods.

Krylov subspace methods edit

Krylov subspace methods work by forming a basis of the sequence of successive matrix powers times the initial residual (the Krylov sequence). The approximations to the solution are then formed by minimizing the residual over the subspace formed. The prototypical method in this class is the conjugate gradient method (CG) which assumes that the system matrix   is symmetric positive-definite. For symmetric (and possibly indefinite)   one works with the minimal residual method (MINRES). In the case of non-symmetric matrices, methods such as the generalized minimal residual method (GMRES) and the biconjugate gradient method (BiCG) have been derived.

Convergence of Krylov subspace methods edit

Since these methods form a basis, it is evident that the method converges in N iterations, where N is the system size. However, in the presence of rounding errors this statement does not hold; moreover, in practice N can be very large, and the iterative process reaches sufficient accuracy already far earlier. The analysis of these methods is hard, depending on a complicated function of the spectrum of the operator.

Preconditioners edit

The approximating operator that appears in stationary iterative methods can also be incorporated in Krylov subspace methods such as GMRES (alternatively, preconditioned Krylov methods can be considered as accelerations of stationary iterative methods), where they become transformations of the original operator to a presumably better conditioned one. The construction of preconditioners is a large research area.

History edit

Jamshīd al-Kāshī used iterative methods to calculate the sine of 1° and π in The Treatise of Chord and Sine to high precision. An early iterative method for solving a linear system appeared in a letter of Gauss to a student of his. He proposed solving a 4-by-4 system of equations by repeatedly solving the component in which the residual was the largest[citation needed].

The theory of stationary iterative methods was solidly established with the work of D.M. Young starting in the 1950s. The conjugate gradient method was also invented in the 1950s, with independent developments by Cornelius Lanczos, Magnus Hestenes and Eduard Stiefel, but its nature and applicability were misunderstood at the time. Only in the 1970s was it realized that conjugacy based methods work very well for partial differential equations, especially the elliptic type.

See also edit

References edit

  1. ^ Amritkar, Amit; de Sturler, Eric; Świrydowicz, Katarzyna; Tafti, Danesh; Ahuja, Kapil (2015). "Recycling Krylov subspaces for CFD applications and a new hybrid recycling solver". Journal of Computational Physics. 303: 222. arXiv:1501.03358. Bibcode:2015JCoPh.303..222A. doi:10.1016/j.jcp.2015.09.040.

External links edit

  • Templates for the Solution of Linear Systems
  • Y. Saad: Iterative Methods for Sparse Linear Systems, 1st edition, PWS 1996

iterative, method, computational, mathematics, iterative, method, mathematical, procedure, that, uses, initial, value, generate, sequence, improving, approximate, solutions, class, problems, which, approximation, derived, from, previous, ones, specific, implem. In computational mathematics an iterative method is a mathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems in which the n th approximation is derived from the previous ones A specific implementation with termination criteria for a given iterative method like gradient descent hill climbing Newton s method or quasi Newton methods like BFGS is an algorithm of the iterative method An iterative method is called convergent if the corresponding sequence converges for given initial approximations A mathematically rigorous convergence analysis of an iterative method is usually performed however heuristic based iterative methods are also common In contrast direct methods attempt to solve the problem by a finite sequence of operations In the absence of rounding errors direct methods would deliver an exact solution for example solving a linear system of equations A x b displaystyle A mathbf x mathbf b by Gaussian elimination Iterative methods are often the only choice for nonlinear equations However iterative methods are often useful even for linear problems involving many variables sometimes on the order of millions where direct methods would be prohibitively expensive and in some cases impossible even with the best available computing power 1 Contents 1 Attractive fixed points 2 Linear systems 2 1 Stationary iterative methods 2 1 1 Introduction 2 1 2 Definition 2 1 3 Examples 2 2 Krylov subspace methods 2 2 1 Convergence of Krylov subspace methods 2 3 Preconditioners 2 4 History 3 See also 4 References 5 External linksAttractive fixed points editIf an equation can be put into the form f x x and a solution x is an attractive fixed point of the function f then one may begin with a point x1 in the basin of attraction of x and let xn 1 f xn for n 1 and the sequence xn n 1 will converge to the solution x Here xn is the nth approximation or iteration of x and xn 1 is the next or n 1 iteration of x Alternately superscripts in parentheses are often used in numerical methods so as not to interfere with subscripts with other meanings For example x n 1 f x n If the function f is continuously differentiable a sufficient condition for convergence is that the spectral radius of the derivative is strictly bounded by one in a neighborhood of the fixed point If this condition holds at the fixed point then a sufficiently small neighborhood basin of attraction must exist Linear systems editIn the case of a system of linear equations the two main classes of iterative methods are the stationary iterative methods and the more general Krylov subspace methods Stationary iterative methods edit Introduction edit Stationary iterative methods solve a linear system with an operator approximating the original one and based on a measurement of the error in the result the residual form a correction equation for which this process is repeated While these methods are simple to derive implement and analyze convergence is only guaranteed for a limited class of matrices Definition edit An iterative method is defined by x k 1 PS x k k 0 displaystyle mathbf x k 1 Psi mathbf x k quad k geq 0 nbsp and for a given linear system A x b displaystyle A mathbf x mathbf b nbsp with exact solution x displaystyle mathbf x nbsp the error by e k x k x k 0 displaystyle mathbf e k mathbf x k mathbf x quad k geq 0 nbsp An iterative method is called linear if there exists a matrix C R n n displaystyle C in mathbb R n times n nbsp such that e k 1 C e k k 0 displaystyle mathbf e k 1 C mathbf e k quad forall k geq 0 nbsp and this matrix is called the iteration matrix An iterative method with a given iteration matrix C displaystyle C nbsp is called convergent if the following holds lim k C k 0 displaystyle lim k rightarrow infty C k 0 nbsp An important theorem states that for a given iterative method and its iteration matrix C displaystyle C nbsp it is convergent if and only if its spectral radius r C displaystyle rho C nbsp is smaller than unity that is r C lt 1 displaystyle rho C lt 1 nbsp The basic iterative methods work by splitting the matrix A displaystyle A nbsp into A M N displaystyle A M N nbsp and here the matrix M displaystyle M nbsp should be easily invertible The iterative methods are now defined as M x k 1 N x k b k 0 displaystyle M mathbf x k 1 N mathbf x k b quad k geq 0 nbsp From this follows that the iteration matrix is given by C I M 1 A M 1 N displaystyle C I M 1 A M 1 N nbsp Examples edit Basic examples of stationary iterative methods use a splitting of the matrix A displaystyle A nbsp such as A D L U D diag a i i i displaystyle A D L U quad D text diag a ii i nbsp where D displaystyle D nbsp is only the diagonal part of A displaystyle A nbsp and L displaystyle L nbsp is the strict lower triangular part of A displaystyle A nbsp Respectively U displaystyle U nbsp is the strict upper triangular part of A displaystyle A nbsp Richardson method M 1 w I w 0 displaystyle M frac 1 omega I quad omega neq 0 nbsp Jacobi method M D displaystyle M D nbsp Damped Jacobi method M 1 w D w 0 displaystyle M frac 1 omega D quad omega neq 0 nbsp Gauss Seidel method M D L displaystyle M D L nbsp Successive over relaxation method SOR M 1 w D L w 0 displaystyle M frac 1 omega D L quad omega neq 0 nbsp Symmetric successive over relaxation SSOR M 1 w 2 w D w L D 1 D w U w 0 2 displaystyle M frac 1 omega 2 omega D omega L D 1 D omega U quad omega not in 0 2 nbsp Linear stationary iterative methods are also called relaxation methods Krylov subspace methods edit Krylov subspace methods work by forming a basis of the sequence of successive matrix powers times the initial residual the Krylov sequence The approximations to the solution are then formed by minimizing the residual over the subspace formed The prototypical method in this class is the conjugate gradient method CG which assumes that the system matrix A displaystyle A nbsp is symmetric positive definite For symmetric and possibly indefinite A displaystyle A nbsp one works with the minimal residual method MINRES In the case of non symmetric matrices methods such as the generalized minimal residual method GMRES and the biconjugate gradient method BiCG have been derived Convergence of Krylov subspace methods edit Since these methods form a basis it is evident that the method converges in N iterations where N is the system size However in the presence of rounding errors this statement does not hold moreover in practice N can be very large and the iterative process reaches sufficient accuracy already far earlier The analysis of these methods is hard depending on a complicated function of the spectrum of the operator Preconditioners edit The approximating operator that appears in stationary iterative methods can also be incorporated in Krylov subspace methods such as GMRES alternatively preconditioned Krylov methods can be considered as accelerations of stationary iterative methods where they become transformations of the original operator to a presumably better conditioned one The construction of preconditioners is a large research area History edit Jamshid al Kashi used iterative methods to calculate the sine of 1 and p in The Treatise of Chord and Sine to high precision An early iterative method for solving a linear system appeared in a letter of Gauss to a student of his He proposed solving a 4 by 4 system of equations by repeatedly solving the component in which the residual was the largest citation needed The theory of stationary iterative methods was solidly established with the work of D M Young starting in the 1950s The conjugate gradient method was also invented in the 1950s with independent developments by Cornelius Lanczos Magnus Hestenes and Eduard Stiefel but its nature and applicability were misunderstood at the time Only in the 1970s was it realized that conjugacy based methods work very well for partial differential equations especially the elliptic type See also edit nbsp Mathematics portalClosed form expression Iterative refinement Kaczmarz method Non linear least squares Numerical analysis Root finding algorithmReferences edit Amritkar Amit de Sturler Eric Swirydowicz Katarzyna Tafti Danesh Ahuja Kapil 2015 Recycling Krylov subspaces for CFD applications and a new hybrid recycling solver Journal of Computational Physics 303 222 arXiv 1501 03358 Bibcode 2015JCoPh 303 222A doi 10 1016 j jcp 2015 09 040 External links edit nbsp Wikimedia Commons has media related to Iterative methods Templates for the Solution of Linear Systems Y Saad Iterative Methods for Sparse Linear Systems 1st edition PWS 1996 Retrieved from https en wikipedia org w index php title Iterative method amp oldid 1158740832, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.