fbpx
Wikipedia

Fixed-point iteration

In numerical analysis, fixed-point iteration is a method of computing fixed points of a function.

More specifically, given a function defined on the real numbers with real values and given a point in the domain of , the fixed-point iteration is

which gives rise to the sequence of iterated function applications which is hoped to converge to a point . If is continuous, then one can prove that the obtained is a fixed point of , i.e.,

More generally, the function can be defined on any metric space with values in that same space.

Examples edit

 
The fixed-point iteration xn+1 = sin xn with initial value x0 = 2 converges to 0. This example does not satisfy the assumptions of the Banach fixed-point theorem and so its speed of convergence is very slow.
  • A first simple and useful example is the Babylonian method for computing the square root of a > 0, which consists in taking  , i.e. the mean value of x and a/x, to approach the limit   (from whatever starting point  ). This is a special case of Newton's method quoted below.
  • The fixed-point iteration   converges to the unique fixed point of the function   for any starting point   This example does satisfy (at the latest after the first iteration step) the assumptions of the Banach fixed-point theorem. Hence, the error after n steps satisfies   (where we can take  , if we start from  .) When the error is less than a multiple of   for some constant q, we say that we have linear convergence. The Banach fixed-point theorem allows one to obtain fixed-point iterations with linear convergence.
  • The requirement that f is continuous is important, as the following example shows. The iteration
     
    converges to 0 for all values of  . However, 0 is not a fixed point of the function
     
    as this function is not continuous at  , and in fact has no fixed points.

Attracting fixed points edit

 
The fixed point iteration xn+1 = cos xn with initial value x1 = −1.

An attracting fixed point of a function f is a fixed point xfix of f with a neighborhood U of "close enough" points around xfix such that for any value of x in U, the fixed-point iteration sequence

 
is contained in U and converges to xfix. The basin of attraction of xfix is the largest such neighborhood U.[1]

The natural cosine function ("natural" means in radians, not degrees or other units) has exactly one fixed point, and that fixed point is attracting. In this case, "close enough" is not a stringent criterion at all—to demonstrate this, start with any real number and repeatedly press the cos key on a calculator (checking first that the calculator is in "radians" mode). It eventually converges to the Dottie number (about 0.739085133), which is a fixed point. That is where the graph of the cosine function intersects the line  .[2]

Not all fixed points are attracting. For example, 0 is a fixed point of the function f(x) = 2x, but iteration of this function for any value other than zero rapidly diverges. We say that the fixed point of   is repelling.

An attracting fixed point is said to be a stable fixed point if it is also Lyapunov stable.

A fixed point is said to be a neutrally stable fixed point if it is Lyapunov stable but not attracting. The center of a linear homogeneous differential equation of the second order is an example of a neutrally stable fixed point.

Multiple attracting points can be collected in an attracting fixed set.

Banach fixed-point theorem edit

The Banach fixed-point theorem gives a sufficient condition for the existence of attracting fixed points. A contraction mapping function   defined on a complete metric space has precisely one fixed point, and the fixed-point iteration is attracted towards that fixed point for any initial guess   in the domain of the function. Common special cases are that (1)   is defined on the real line with real values and is Lipschitz continuous with Lipschitz constant  , and (2) the function f is continuously differentiable in an open neighbourhood of a fixed point xfix, and  .

Although there are other fixed-point theorems, this one in particular is very useful because not all fixed-points are attractive. When constructing a fixed-point iteration, it is very important to make sure it converges to the fixed point. We can usually use the Banach fixed-point theorem to show that the fixed point is attractive.

Attractors edit

Attracting fixed points are a special case of a wider mathematical concept of attractors. Fixed-point iterations are a discrete dynamical system on one variable. Bifurcation theory studies dynamical systems and classifies various behaviors such as attracting fixed points, periodic orbits, or strange attractors. An example system is the logistic map.

Iterative methods edit

In computational mathematics, an iterative method is a mathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems, in which the n-th approximation is derived from the previous ones. Convergent fixed-point iterations are mathematically rigorous formalizations of iterative methods.

Iterative method examples edit

  • Newton's method is a root-finding algorithm for finding roots of a given differentiable function  . The iteration is  

    If we write  , we may rewrite the Newton iteration as the fixed-point iteration  .

    If this iteration converges to a fixed point   of g, then  , so  

    therefore  , that is,   is a root of  . Under the assumptions of the Banach fixed-point theorem, the Newton iteration, framed as the fixed-point method, demonstrates linear convergence. However, a more detailed analysis shows quadratic convergence, i.e.,  , under certain circumstances.
  • Halley's method is similar to Newton's method when it works correctly, but its error is   (cubic convergence). In general, it is possible to design methods that converge with speed   for any  . As a general rule, the higher the k, the less stable it is, and the more computationally expensive it gets. For these reasons, higher order methods are typically not used.
  • Runge–Kutta methods and numerical ordinary differential equation solvers in general can be viewed as fixed-point iterations. Indeed, the core idea when analyzing the A-stability of ODE solvers is to start with the special case  , where   is a complex number, and to check whether the ODE solver converges to the fixed point   whenever the real part of   is negative.[a]
  • The Picard–Lindelöf theorem, which shows that ordinary differential equations have solutions, is essentially an application of the Banach fixed-point theorem to a special sequence of functions which forms a fixed-point iteration, constructing the solution to the equation. Solving an ODE in this way is called Picard iteration, Picard's method, or the Picard iterative process.
  • The iteration capability in Excel can be used to find solutions to the Colebrook equation to an accuracy of 15 significant figures.[3][4]
  • Some of the "successive approximation" schemes used in dynamic programming to solve Bellman's functional equation are based on fixed-point iterations in the space of the return function.[5][6]
  • The cobweb model of price theory corresponds to the fixed-point iteration of the composition of the supply function and the demand function.[7]

Convergence acceleration edit

The speed of convergence of the iteration sequence can be increased by using a convergence acceleration method such as Anderson acceleration and Aitken's delta-squared process. The application of Aitken's method to fixed-point iteration is known as Steffensen's method, and it can be shown that Steffensen's method yields a rate of convergence that is at least quadratic.

Chaos game edit

 
Sierpinski triangle created using IFS, selecting all members at each iteration

The term chaos game refers to a method of generating the fixed point of any iterated function system (IFS). Starting with any point x0, successive iterations are formed as xk+1 = fr(xk), where fr is a member of the given IFS randomly selected for each iteration. Hence the chaos game is a randomized fixed-point iteration. The chaos game allows plotting the general shape of a fractal such as the Sierpinski triangle by repeating the iterative process a large number of times. More mathematically, the iterations converge to the fixed point of the IFS. Whenever x0 belongs to the attractor of the IFS, all iterations xk stay inside the attractor and, with probability 1, form a dense set in the latter.

See also edit

References edit

  1. ^ One may also consider certain iterations A-stable if the iterates stay bounded for a long time, which is beyond the scope of this article.
  1. ^ Rassias, Themistocles M.; Pardalos, Panos M. (17 September 2014). Mathematics Without Boundaries: Surveys in Pure Mathematics. Springer. ISBN 978-1-4939-1106-6.
  2. ^ Weisstein, Eric W. "Dottie Number". Wolfram MathWorld. Wolfram Research, Inc. Retrieved 23 July 2016.
  3. ^ M A Kumar (2010), Solve Implicit Equations (Colebrook) Within Worksheet, Createspace, ISBN 1-4528-1619-0
  4. ^ Brkic, Dejan (2017) Solution of the Implicit Colebrook Equation for Flow Friction Using Excel, Spreadsheets in Education (eJSiE): Vol. 10: Iss. 2, Article 2. Available at: https://sie.scholasticahq.com/article/4663-solution-of-the-implicit-colebrook-equation-for-flow-friction-using-excel
  5. ^ Bellman, R. (1957). Dynamic programming, Princeton University Press.
  6. ^ Sniedovich, M. (2010). Dynamic Programming: Foundations and Principles, Taylor & Francis.
  7. ^ Onozaki, Tamotsu (2018). "Chapter 2. One-Dimensional Nonlinear Cobweb Model". Nonlinearity, Bounded Rationality, and Heterogeneity: Some Aspects of Market Economies as Complex Systems. Springer. ISBN 978-4-431-54971-0.

Further reading edit

  • Burden, Richard L.; Faires, J. Douglas (1985). "Fixed-Point Iteration". Numerical Analysis (Third ed.). PWS Publishers. ISBN 0-87150-857-5.
  • Hoffman, Joe D.; Frankel, Steven (2001). "Fixed-Point Iteration". Numerical Methods for Engineers and Scientists (Second ed.). New York: CRC Press. pp. 141–145. ISBN 0-8247-0443-6.
  • Judd, Kenneth L. (1998). "Fixed-Point Iteration". Numerical Methods in Economics. Cambridge: MIT Press. pp. 165–167. ISBN 0-262-10071-1.
  • Sternberg, Shlomo (2010). "Iteration and fixed points". Dynamical Systems (First ed.). Dover Publications. ISBN 978-0486477053.
  • Shashkin, Yuri A. (1991). "9. The Iteration Method". Fixed Points (First ed.). American Mathematical Society. ISBN 0-8218-9000-X.
  • Rosa, Alessandro (2021). "An episodic history of the staircased iteration diagram". Antiquitates Mathematicae. 15: 3–90. doi:10.14708/am.v15i1.7056. S2CID 247259939.

External links edit

  • Fixed-point algorithms online
  • Fixed-point iteration online calculator (Mathematical Assistant on Web)

fixed, point, iteration, this, article, needs, additional, citations, verification, please, help, improve, this, article, adding, citations, reliable, sources, unsourced, material, challenged, removed, find, sources, news, newspapers, books, scholar, jstor, 20. This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Fixed point iteration news newspapers books scholar JSTOR May 2010 Learn how and when to remove this template message In numerical analysis fixed point iteration is a method of computing fixed points of a function More specifically given a function f displaystyle f defined on the real numbers with real values and given a point x 0 displaystyle x 0 in the domain of f displaystyle f the fixed point iteration isx n 1 f x n n 0 1 2 displaystyle x n 1 f x n n 0 1 2 dots which gives rise to the sequence x 0 x 1 x 2 displaystyle x 0 x 1 x 2 dots of iterated function applications x 0 f x 0 f f x 0 displaystyle x 0 f x 0 f f x 0 dots which is hoped to converge to a point x fix displaystyle x text fix If f displaystyle f is continuous then one can prove that the obtained x fix displaystyle x text fix is a fixed point of f displaystyle f i e f x fix x fix displaystyle f x text fix x text fix More generally the function f displaystyle f can be defined on any metric space with values in that same space Contents 1 Examples 2 Attracting fixed points 2 1 Banach fixed point theorem 2 2 Attractors 3 Iterative methods 3 1 Iterative method examples 3 2 Convergence acceleration 4 Chaos game 5 See also 6 References 7 Further reading 8 External linksExamples edit nbsp The fixed point iteration xn 1 sin xn with initial value x0 2 converges to 0 This example does not satisfy the assumptions of the Banach fixed point theorem and so its speed of convergence is very slow A first simple and useful example is the Babylonian method for computing the square root of a gt 0 which consists in taking f x 1 2 a x x displaystyle f x frac 1 2 left frac a x x right nbsp i e the mean value of x and a x to approach the limit x a displaystyle x sqrt a nbsp from whatever starting point x 0 0 displaystyle x 0 gg 0 nbsp This is a special case of Newton s method quoted below The fixed point iteration x n 1 cos x n displaystyle x n 1 cos x n nbsp converges to the unique fixed point of the function f x cos x displaystyle f x cos x nbsp for any starting point x 0 displaystyle x 0 nbsp This example does satisfy at the latest after the first iteration step the assumptions of the Banach fixed point theorem Hence the error after n steps satisfies x n x q n 1 q x 1 x 0 C q n displaystyle x n x leq q n over 1 q x 1 x 0 Cq n nbsp where we can take q 0 85 displaystyle q 0 85 nbsp if we start from x 0 1 displaystyle x 0 1 nbsp When the error is less than a multiple of q n displaystyle q n nbsp for some constant q we say that we have linear convergence The Banach fixed point theorem allows one to obtain fixed point iterations with linear convergence The requirement that f is continuous is important as the following example shows The iteration x n 1 x n 2 x n 0 1 x n 0 displaystyle x n 1 begin cases frac x n 2 amp x n neq 0 1 amp x n 0 end cases nbsp converges to 0 for all values of x 0 displaystyle x 0 nbsp However 0 is not a fixed point of the function f x x 2 x 0 1 x 0 displaystyle f x begin cases frac x 2 amp x neq 0 1 amp x 0 end cases nbsp as this function is not continuous at x 0 displaystyle x 0 nbsp and in fact has no fixed points Attracting fixed points edit nbsp The fixed point iteration xn 1 cos xn with initial value x1 1 An attracting fixed point of a function f is a fixed point xfix of f with a neighborhood U of close enough points around xfix such that for any value of x in U the fixed point iteration sequencex f x f f x f f f x displaystyle x f x f f x f f f x dots nbsp is contained in U and converges to xfix The basin of attraction of xfix is the largest such neighborhood U 1 The natural cosine function natural means in radians not degrees or other units has exactly one fixed point and that fixed point is attracting In this case close enough is not a stringent criterion at all to demonstrate this start with any real number and repeatedly press the cos key on a calculator checking first that the calculator is in radians mode It eventually converges to the Dottie number about 0 739085133 which is a fixed point That is where the graph of the cosine function intersects the line y x displaystyle y x nbsp 2 Not all fixed points are attracting For example 0 is a fixed point of the function f x 2x but iteration of this function for any value other than zero rapidly diverges We say that the fixed point of f x 2 x displaystyle f x 2x nbsp is repelling An attracting fixed point is said to be a stable fixed point if it is also Lyapunov stable A fixed point is said to be a neutrally stable fixed point if it is Lyapunov stable but not attracting The center of a linear homogeneous differential equation of the second order is an example of a neutrally stable fixed point Multiple attracting points can be collected in an attracting fixed set Banach fixed point theorem edit The Banach fixed point theorem gives a sufficient condition for the existence of attracting fixed points A contraction mapping function f displaystyle f nbsp defined on a complete metric space has precisely one fixed point and the fixed point iteration is attracted towards that fixed point for any initial guess x 0 displaystyle x 0 nbsp in the domain of the function Common special cases are that 1 f displaystyle f nbsp is defined on the real line with real values and is Lipschitz continuous with Lipschitz constant L lt 1 displaystyle L lt 1 nbsp and 2 the function f is continuously differentiable in an open neighbourhood of a fixed point xfix and f x fix lt 1 displaystyle f x text fix lt 1 nbsp Although there are other fixed point theorems this one in particular is very useful because not all fixed points are attractive When constructing a fixed point iteration it is very important to make sure it converges to the fixed point We can usually use the Banach fixed point theorem to show that the fixed point is attractive Attractors edit Attracting fixed points are a special case of a wider mathematical concept of attractors Fixed point iterations are a discrete dynamical system on one variable Bifurcation theory studies dynamical systems and classifies various behaviors such as attracting fixed points periodic orbits or strange attractors An example system is the logistic map Iterative methods editMain article Iterative method In computational mathematics an iterative method is a mathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems in which the n th approximation is derived from the previous ones Convergent fixed point iterations are mathematically rigorous formalizations of iterative methods Iterative method examples edit Newton s method is a root finding algorithm for finding roots of a given differentiable function f x displaystyle f x nbsp The iteration is x n 1 x n f x n f x n textstyle x n 1 x n frac f x n f x n nbsp If we write g x x f x f x textstyle g x x frac f x f x nbsp we may rewrite the Newton iteration as the fixed point iteration x n 1 g x n textstyle x n 1 g x n nbsp If this iteration converges to a fixed point x fix displaystyle x text fix nbsp of g then x fix g x fix x fix f x fix f x fix textstyle x text fix g x text fix x text fix frac f x text fix f x text fix nbsp so f x fix f x fix 0 textstyle f x text fix f x text fix 0 nbsp therefore f x fix 0 displaystyle f x text fix 0 nbsp that is x fix displaystyle x text fix nbsp is a root of f displaystyle f nbsp Under the assumptions of the Banach fixed point theorem the Newton iteration framed as the fixed point method demonstrates linear convergence However a more detailed analysis shows quadratic convergence i e x n x fix lt C q 2 n textstyle x n x text fix lt Cq 2 n nbsp under certain circumstances Halley s method is similar to Newton s method when it works correctly but its error is x n x fix lt C q 3 n displaystyle x n x text fix lt Cq 3 n nbsp cubic convergence In general it is possible to design methods that converge with speed C q k n displaystyle Cq k n nbsp for any k N displaystyle k in mathbb N nbsp As a general rule the higher the k the less stable it is and the more computationally expensive it gets For these reasons higher order methods are typically not used Runge Kutta methods and numerical ordinary differential equation solvers in general can be viewed as fixed point iterations Indeed the core idea when analyzing the A stability of ODE solvers is to start with the special case y a y displaystyle y ay nbsp where a displaystyle a nbsp is a complex number and to check whether the ODE solver converges to the fixed point y fix 0 displaystyle y text fix 0 nbsp whenever the real part of a displaystyle a nbsp is negative a The Picard Lindelof theorem which shows that ordinary differential equations have solutions is essentially an application of the Banach fixed point theorem to a special sequence of functions which forms a fixed point iteration constructing the solution to the equation Solving an ODE in this way is called Picard iteration Picard s method or the Picard iterative process The iteration capability in Excel can be used to find solutions to the Colebrook equation to an accuracy of 15 significant figures 3 4 Some of the successive approximation schemes used in dynamic programming to solve Bellman s functional equation are based on fixed point iterations in the space of the return function 5 6 The cobweb model of price theory corresponds to the fixed point iteration of the composition of the supply function and the demand function 7 Convergence acceleration edit The speed of convergence of the iteration sequence can be increased by using a convergence acceleration method such as Anderson acceleration and Aitken s delta squared process The application of Aitken s method to fixed point iteration is known as Steffensen s method and it can be shown that Steffensen s method yields a rate of convergence that is at least quadratic Chaos game edit nbsp Sierpinski triangle created using IFS selecting all members at each iteration Main article Chaos game The term chaos game refers to a method of generating the fixed point of any iterated function system IFS Starting with any point x0 successive iterations are formed as xk 1 fr xk where fr is a member of the given IFS randomly selected for each iteration Hence the chaos game is a randomized fixed point iteration The chaos game allows plotting the general shape of a fractal such as the Sierpinski triangle by repeating the iterative process a large number of times More mathematically the iterations converge to the fixed point of the IFS Whenever x0 belongs to the attractor of the IFS all iterations xk stay inside the attractor and with probability 1 form a dense set in the latter See also editFixed point combinator Cobweb plot Markov chain Infinite compositions of analytic functions Rate of convergenceReferences edit One may also consider certain iterations A stable if the iterates stay bounded for a long time which is beyond the scope of this article Rassias Themistocles M Pardalos Panos M 17 September 2014 Mathematics Without Boundaries Surveys in Pure Mathematics Springer ISBN 978 1 4939 1106 6 Weisstein Eric W Dottie Number Wolfram MathWorld Wolfram Research Inc Retrieved 23 July 2016 M A Kumar 2010 Solve Implicit Equations Colebrook Within Worksheet Createspace ISBN 1 4528 1619 0 Brkic Dejan 2017 Solution of the Implicit Colebrook Equation for Flow Friction Using Excel Spreadsheets in Education eJSiE Vol 10 Iss 2 Article 2 Available at https sie scholasticahq com article 4663 solution of the implicit colebrook equation for flow friction using excel Bellman R 1957 Dynamic programming Princeton University Press Sniedovich M 2010 Dynamic Programming Foundations and Principles Taylor amp Francis Onozaki Tamotsu 2018 Chapter 2 One Dimensional Nonlinear Cobweb Model Nonlinearity Bounded Rationality and Heterogeneity Some Aspects of Market Economies as Complex Systems Springer ISBN 978 4 431 54971 0 Further reading editBurden Richard L Faires J Douglas 1985 Fixed Point Iteration Numerical Analysis Third ed PWS Publishers ISBN 0 87150 857 5 Hoffman Joe D Frankel Steven 2001 Fixed Point Iteration Numerical Methods for Engineers and Scientists Second ed New York CRC Press pp 141 145 ISBN 0 8247 0443 6 Judd Kenneth L 1998 Fixed Point Iteration Numerical Methods in Economics Cambridge MIT Press pp 165 167 ISBN 0 262 10071 1 Sternberg Shlomo 2010 Iteration and fixed points Dynamical Systems First ed Dover Publications ISBN 978 0486477053 Shashkin Yuri A 1991 9 The Iteration Method Fixed Points First ed American Mathematical Society ISBN 0 8218 9000 X Rosa Alessandro 2021 An episodic history of the staircased iteration diagram Antiquitates Mathematicae 15 3 90 doi 10 14708 am v15i1 7056 S2CID 247259939 External links editFixed point algorithms online Fixed point iteration online calculator Mathematical Assistant on Web Retrieved from https en wikipedia org w index php title Fixed point iteration amp oldid 1189645416, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.