fbpx
Wikipedia

Finite difference method

In numerical analysis, finite-difference methods (FDM) are a class of numerical techniques for solving differential equations by approximating derivatives with finite differences. Both the spatial domain and time interval (if applicable) are discretized, or broken into a finite number of steps, and the value of the solution at these discrete points is approximated by solving algebraic equations containing finite differences and values from nearby points.

Finite difference methods convert ordinary differential equations (ODE) or partial differential equations (PDE), which may be nonlinear, into a system of linear equations that can be solved by matrix algebra techniques. Modern computers can perform these linear algebra computations efficiently which, along with their relative ease of implementation, has led to the widespread use of FDM in modern numerical analysis.[1] Today, FDM are one of the most common approaches to the numerical solution of PDE, along with finite element methods.[1]

Derivation from Taylor's polynomial

First, assuming the function whose derivatives are to be approximated is properly behaved, by Taylor's theorem, we can create a Taylor series expansion

 

where n! denotes the factorial of n, and Rn(x) is a remainder term, denoting the difference between the Taylor polynomial of degree n and the original function. We will derive an approximation for the first derivative of the function "f" by first truncating the Taylor polynomial:

 

Setting, x0=a we have,

 

Dividing across by h gives:

 

Solving for f'(a):

 

Assuming that   is sufficiently small, the approximation of the first derivative of "f" is:

 

This is, not coincidentally, similar to the definition of derivative, which is given as:

 

except for the limit towards zero (the method is named after this).

Accuracy and order

The error in a method's solution is defined as the difference between the approximation and the exact analytical solution. The two sources of error in finite difference methods are round-off error, the loss of precision due to computer rounding of decimal quantities, and truncation error or discretization error, the difference between the exact solution of the original differential equation and the exact quantity assuming perfect arithmetic (that is, assuming no round-off).

 
The finite difference method relies on discretizing a function on a grid.

To use a finite difference method to approximate the solution to a problem, one must first discretize the problem's domain. This is usually done by dividing the domain into a uniform grid (see image to the right). This means that finite-difference methods produce sets of discrete numerical approximations to the derivative, often in a "time-stepping" manner.

An expression of general interest is the local truncation error of a method. Typically expressed using Big-O notation, local truncation error refers to the error from a single application of a method. That is, it is the quantity   if   refers to the exact value and   to the numerical approximation. The remainder term of a Taylor polynomial is convenient for analyzing the local truncation error. Using the Lagrange form of the remainder from the Taylor polynomial for  , which is

 , where  ,

the dominant term of the local truncation error can be discovered. For example, again using the forward-difference formula for the first derivative, knowing that  ,

 

and with some algebraic manipulation, this leads to

 

and further noting that the quantity on the left is the approximation from the finite difference method and that the quantity on the right is the exact quantity of interest plus a remainder, clearly that remainder is the local truncation error. A final expression of this example and its order is:

 

This means that, in this case, the local truncation error is proportional to the step sizes. The quality and duration of simulated FDM solution depends on the discretization equation selection and the step sizes (time and space steps). The data quality and simulation duration increase significantly with smaller step size.[2] Therefore, a reasonable balance between data quality and simulation duration is necessary for practical usage. Large time steps are useful for increasing simulation speed in practice. However, time steps which are too large may create instabilities and affect the data quality.[3][4]

The von Neumann and Courant-Friedrichs-Lewy criteria are often evaluated to determine the numerical model stability.[3][4][5][6]

Example: ordinary differential equation

For example, consider the ordinary differential equation

 

The Euler method for solving this equation uses the finite difference quotient

 

to approximate the differential equation by first substituting it for u'(x) then applying a little algebra (multiplying both sides by h, and then adding u(x) to both sides) to get

 

The last equation is a finite-difference equation, and solving this equation gives an approximate solution to the differential equation.

Example: The heat equation

Consider the normalized heat equation in one dimension, with homogeneous Dirichlet boundary conditions

 
  (boundary condition)
  (initial condition)

One way to numerically solve this equation is to approximate all the derivatives by finite differences. We partition the domain in space using a mesh   and in time using a mesh  . We assume a uniform partition both in space and in time, so the difference between two consecutive space points will be h and between two consecutive time points will be k. The points

 

will represent the numerical approximation of  

Explicit method

 
The stencil for the most common explicit method for the heat equation.

Using a forward difference at time   and a second-order central difference for the space derivative at position   (FTCS) we get the recurrence equation:

 

This is an explicit method for solving the one-dimensional heat equation.

We can obtain   from the other values this way:

 

where  

So, with this recurrence relation, and knowing the values at time n, one can obtain the corresponding values at time n+1.   and   must be replaced by the boundary conditions, in this example they are both 0.

This explicit method is known to be numerically stable and convergent whenever  .[7] The numerical errors are proportional to the time step and the square of the space step:

 

Implicit method

 
The implicit method stencil.

If we use the backward difference at time   and a second-order central difference for the space derivative at position   (The Backward Time, Centered Space Method "BTCS") we get the recurrence equation:

 

This is an implicit method for solving the one-dimensional heat equation.

We can obtain   from solving a system of linear equations:

 

The scheme is always numerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step:

 

Crank–Nicolson method

Finally if we use the central difference at time   and a second-order central difference for the space derivative at position   ("CTCS") we get the recurrence equation:

 

This formula is known as the Crank–Nicolson method.

 
The Crank–Nicolson stencil.

We can obtain   from solving a system of linear equations:

 

The scheme is always numerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over both the time step and the space step:

 

Comparison

To summarize, usually the Crank–Nicolson scheme is the most accurate scheme for small time steps. For larger time steps, the implicit scheme works better since it is less computationally demanding. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive.

Here is an example. The figures below present the solutions given by the above methods to approximate the heat equation

 

with the boundary condition

 

The exact solution is

 
Comparison of Finite Difference Methods
 
Explicit method (not stable)
 
Implicit method (stable)
 
Crank-Nicolson method (stable)

Example: The Laplace operator

The (continuous) Laplace operator in  -dimensions is given by  . The discrete Laplace operator   depends on the dimension  .

In 1D the Laplace operator is approximated as

 

This approximation is usually expressed via the following stencil

 

and which represents a symmetric, tridiagonal matrix. For an equidistant grid one gets a Toeplitz matrix.

The 2D case shows all the characteristics of the more general nD case. Each second partial derivative needs to be approximated similar to the 1D case

 

which is usually given by the following stencil

 

Consistency

Consistency of the above-mentioned approximation can be shown for highly regular functions, such as  . The statement is

 

To prove this, one needs to substitute Taylor Series expansions up to order 3 into the discrete Laplace operator.

Properties

Subharmonic

Similar to continuous subharmonic functions one can define subharmonic functions for finite-difference approximations  

 

Mean value

One can define a general stencil of positive type via

 

If   is (discrete) subharmonic then the following mean value property holds

 

where the approximation is evaluated on points of the grid, and the stencil is assumed to be of positive type.

A similar mean value property also holds for the continuous case.

Maximum principle

For a (discrete) subharmonic function   the following holds

 

where   are discretizations of the continuous domain  , respectively the boundary  .

A similar maximum principle also holds for the continuous case.

The SBP-SAT method

The SBP-SAT (summation by parts - simultaneous approximation term) method is a stable and accurate technique for discretizing and imposing boundary conditions of a well-posed partial differential equation using high order finite differences.[8][9]

The method is based on finite differences where the differentiation operators exhibit summation-by-parts properties. Typically, these operators consist of differentiation matrices with central difference stencils in the interior with carefully chosen one-sided boundary stencils designed to mimic integration-by-parts in the discrete setting. Using the SAT technique, the boundary conditions of the PDE are imposed weakly, where the boundary values are "pulled" towards the desired conditions rather than exactly fulfilled. If the tuning parameters (inherent to the SAT technique) are chosen properly, the resulting system of ODE's will exhibit similar energy behavior as the continuous PDE, i.e. the system has no non-physical energy growth. This guarantees stability if an integration scheme with a stability region that includes parts of the imaginary axis, such as the fourth order Runge-Kutta method, is used. This makes the SAT technique an attractive method of imposing boundary conditions for higher order finite difference methods, in contrast to for example the injection method, which typically will not be stable if high order differentiation operators are used.

See also

References

  1. ^ a b Christian Grossmann; Hans-G. Roos; Martin Stynes (2007). Numerical Treatment of Partial Differential Equations. Springer Science & Business Media. p. 23. ISBN 978-3-540-71584-9.
  2. ^ Arieh Iserles (2008). A first course in the numerical analysis of differential equations. Cambridge University Press. p. 23. ISBN 9780521734905.
  3. ^ a b Hoffman JD; Frankel S (2001). Numerical methods for engineers and scientists. CRC Press, Boca Raton.
  4. ^ a b Jaluria Y; Atluri S (1994). "Computational heat transfer". Computational Mechanics. 14 (5): 385–386. Bibcode:1994CompM..14..385J. doi:10.1007/BF00377593. S2CID 119502676.
  5. ^ Majumdar P (2005). Computational methods for heat and mass transfer (1st ed.). Taylor and Francis, New York.
  6. ^ Smith GD (1985). Numerical solution of partial differential equations: finite difference methods (3rd ed.). Oxford University Press.
  7. ^ Crank, J. The Mathematics of Diffusion. 2nd Edition, Oxford, 1975, p. 143.
  8. ^ Bo Strand (1994). "Summation by Parts for Finite Difference Approximations for d/dx". Journal of Computational Physics. 110 (1): 47–67. Bibcode:1994JCoPh.110...47S. doi:10.1006/jcph.1994.1005.
  9. ^ Mark H. Carpenter; David I. Gottlieb; Saul S. Abarbanel (1994). "Time-stable boundary conditions for finite-difference schemes solving hyperbolic systems: Methodology and application to high-order compact schemes". Journal of Computational Physics. 111 (2): 220–236. Bibcode:1994JCoPh.111..220C. doi:10.1006/jcph.1994.1057. hdl:2060/19930013937.

Further reading

  • K.W. Morton and D.F. Mayers, Numerical Solution of Partial Differential Equations, An Introduction. Cambridge University Press, 2005.
  • Autar Kaw and E. Eric Kalu, Numerical Methods with Applications, (2008) [1]. Contains a brief, engineering-oriented introduction to FDM (for ODEs) in Chapter 08.07.
  • John Strikwerda (2004). Finite Difference Schemes and Partial Differential Equations (2nd ed.). SIAM. ISBN 978-0-89871-639-9.
  • Smith, G. D. (1985), Numerical Solution of Partial Differential Equations: Finite Difference Methods, 3rd ed., Oxford University Press
  • Peter Olver (2013). Introduction to Partial Differential Equations. Springer. Chapter 5: Finite differences. ISBN 978-3-319-02099-0..
  • Randall J. LeVeque, Finite Difference Methods for Ordinary and Partial Differential Equations, SIAM, 2007.
  • Sergey Lemeshevsky, Piotr Matus, Dmitriy Poliakov(Eds): "Exact Finite-Difference Schemes", De Gruyter (2016). DOI: https://doi.org/10.1515/9783110491326 .

finite, difference, method, this, article, multiple, issues, please, help, improve, discuss, these, issues, talk, page, learn, when, remove, these, template, messages, this, article, lead, section, short, adequately, summarize, points, please, consider, expand. This article has multiple issues Please help improve it or discuss these issues on the talk page Learn how and when to remove these template messages This article s lead section may be too short to adequately summarize the key points Please consider expanding the lead to provide an accessible overview of all important aspects of the article April 2015 This article may be too technical for most readers to understand Please help improve it to make it understandable to non experts without removing the technical details April 2015 Learn how and when to remove this template message This article s tone or style may not reflect the encyclopedic tone used on Wikipedia See Wikipedia s guide to writing better articles for suggestions June 2020 Learn how and when to remove this template message Learn how and when to remove this template message In numerical analysis finite difference methods FDM are a class of numerical techniques for solving differential equations by approximating derivatives with finite differences Both the spatial domain and time interval if applicable are discretized or broken into a finite number of steps and the value of the solution at these discrete points is approximated by solving algebraic equations containing finite differences and values from nearby points Finite difference methods convert ordinary differential equations ODE or partial differential equations PDE which may be nonlinear into a system of linear equations that can be solved by matrix algebra techniques Modern computers can perform these linear algebra computations efficiently which along with their relative ease of implementation has led to the widespread use of FDM in modern numerical analysis 1 Today FDM are one of the most common approaches to the numerical solution of PDE along with finite element methods 1 Contents 1 Derivation from Taylor s polynomial 2 Accuracy and order 3 Example ordinary differential equation 4 Example The heat equation 4 1 Explicit method 4 2 Implicit method 4 3 Crank Nicolson method 4 4 Comparison 5 Example The Laplace operator 5 1 Consistency 5 2 Properties 5 2 1 Subharmonic 5 2 2 Mean value 5 2 3 Maximum principle 6 The SBP SAT method 7 See also 8 References 9 Further readingDerivation from Taylor s polynomial EditFirst assuming the function whose derivatives are to be approximated is properly behaved by Taylor s theorem we can create a Taylor series expansion f x 0 h f x 0 f x 0 1 h f 2 x 0 2 h 2 f n x 0 n h n R n x displaystyle f x 0 h f x 0 frac f x 0 1 h frac f 2 x 0 2 h 2 cdots frac f n x 0 n h n R n x where n denotes the factorial of n and Rn x is a remainder term denoting the difference between the Taylor polynomial of degree n and the original function We will derive an approximation for the first derivative of the function f by first truncating the Taylor polynomial f x 0 h f x 0 f x 0 h R 1 x displaystyle f x 0 h f x 0 f x 0 h R 1 x Setting x0 a we have f a h f a f a h R 1 x displaystyle f a h f a f a h R 1 x Dividing across by h gives f a h h f a h f a R 1 x h displaystyle f a h over h f a over h f a R 1 x over h Solving for f a f a f a h f a h R 1 x h displaystyle f a f a h f a over h R 1 x over h Assuming that R 1 x displaystyle R 1 x is sufficiently small the approximation of the first derivative of f is f a f a h f a h displaystyle f a approx f a h f a over h This is not coincidentally similar to the definition of derivative which is given as f a lim h 0 f a h f a h displaystyle f a lim h to 0 frac f a h f a h except for the limit towards zero the method is named after this Accuracy and order EditSee also Finite difference coefficient The error in a method s solution is defined as the difference between the approximation and the exact analytical solution The two sources of error in finite difference methods are round off error the loss of precision due to computer rounding of decimal quantities and truncation error or discretization error the difference between the exact solution of the original differential equation and the exact quantity assuming perfect arithmetic that is assuming no round off The finite difference method relies on discretizing a function on a grid To use a finite difference method to approximate the solution to a problem one must first discretize the problem s domain This is usually done by dividing the domain into a uniform grid see image to the right This means that finite difference methods produce sets of discrete numerical approximations to the derivative often in a time stepping manner An expression of general interest is the local truncation error of a method Typically expressed using Big O notation local truncation error refers to the error from a single application of a method That is it is the quantity f x i f i displaystyle f x i f i if f x i displaystyle f x i refers to the exact value and f i displaystyle f i to the numerical approximation The remainder term of a Taylor polynomial is convenient for analyzing the local truncation error Using the Lagrange form of the remainder from the Taylor polynomial for f x 0 h displaystyle f x 0 h which isR n x 0 h f n 1 3 n 1 h n 1 displaystyle R n x 0 h frac f n 1 xi n 1 h n 1 where x 0 lt 3 lt x 0 h displaystyle x 0 lt xi lt x 0 h the dominant term of the local truncation error can be discovered For example again using the forward difference formula for the first derivative knowing that f x i f x 0 i h displaystyle f x i f x 0 ih f x 0 i h f x 0 f x 0 i h f 3 2 i h 2 displaystyle f x 0 ih f x 0 f x 0 ih frac f xi 2 ih 2 and with some algebraic manipulation this leads to f x 0 i h f x 0 i h f x 0 f 3 2 i h displaystyle frac f x 0 ih f x 0 ih f x 0 frac f xi 2 ih and further noting that the quantity on the left is the approximation from the finite difference method and that the quantity on the right is the exact quantity of interest plus a remainder clearly that remainder is the local truncation error A final expression of this example and its order is f x 0 i h f x 0 i h f x 0 O h displaystyle frac f x 0 ih f x 0 ih f x 0 O h This means that in this case the local truncation error is proportional to the step sizes The quality and duration of simulated FDM solution depends on the discretization equation selection and the step sizes time and space steps The data quality and simulation duration increase significantly with smaller step size 2 Therefore a reasonable balance between data quality and simulation duration is necessary for practical usage Large time steps are useful for increasing simulation speed in practice However time steps which are too large may create instabilities and affect the data quality 3 4 The von Neumann and Courant Friedrichs Lewy criteria are often evaluated to determine the numerical model stability 3 4 5 6 Example ordinary differential equation EditFor example consider the ordinary differential equation u x 3 u x 2 displaystyle u x 3u x 2 The Euler method for solving this equation uses the finite difference quotient u x h u x h u x displaystyle frac u x h u x h approx u x to approximate the differential equation by first substituting it for u x then applying a little algebra multiplying both sides by h and then adding u x to both sides to get u x h u x h 3 u x 2 displaystyle u x h u x h 3u x 2 The last equation is a finite difference equation and solving this equation gives an approximate solution to the differential equation Example The heat equation EditConsider the normalized heat equation in one dimension with homogeneous Dirichlet boundary conditions U t U x x displaystyle U t U xx U 0 t U 1 t 0 displaystyle U 0 t U 1 t 0 boundary condition U x 0 U 0 x displaystyle U x 0 U 0 x initial condition One way to numerically solve this equation is to approximate all the derivatives by finite differences We partition the domain in space using a mesh x 0 x J displaystyle x 0 x J and in time using a mesh t 0 t N displaystyle t 0 t N We assume a uniform partition both in space and in time so the difference between two consecutive space points will be h and between two consecutive time points will be k The points u x j t n u j n displaystyle u x j t n u j n will represent the numerical approximation of u x j t n displaystyle u x j t n Explicit method Edit The stencil for the most common explicit method for the heat equation Using a forward difference at time t n displaystyle t n and a second order central difference for the space derivative at position x j displaystyle x j FTCS we get the recurrence equation u j n 1 u j n k u j 1 n 2 u j n u j 1 n h 2 displaystyle frac u j n 1 u j n k frac u j 1 n 2u j n u j 1 n h 2 This is an explicit method for solving the one dimensional heat equation We can obtain u j n 1 displaystyle u j n 1 from the other values this way u j n 1 1 2 r u j n r u j 1 n r u j 1 n displaystyle u j n 1 1 2r u j n ru j 1 n ru j 1 n where r a k h 2 displaystyle r alpha k h 2 So with this recurrence relation and knowing the values at time n one can obtain the corresponding values at time n 1 u 0 n displaystyle u 0 n and u J n displaystyle u J n must be replaced by the boundary conditions in this example they are both 0 This explicit method is known to be numerically stable and convergent whenever r 1 2 displaystyle r leq 1 2 7 The numerical errors are proportional to the time step and the square of the space step D u O k O h 2 displaystyle Delta u O k O h 2 Implicit method Edit The implicit method stencil If we use the backward difference at time t n 1 displaystyle t n 1 and a second order central difference for the space derivative at position x j displaystyle x j The Backward Time Centered Space Method BTCS we get the recurrence equation u j n 1 u j n k u j 1 n 1 2 u j n 1 u j 1 n 1 h 2 displaystyle frac u j n 1 u j n k frac u j 1 n 1 2u j n 1 u j 1 n 1 h 2 This is an implicit method for solving the one dimensional heat equation We can obtain u j n 1 displaystyle u j n 1 from solving a system of linear equations 1 2 r u j n 1 r u j 1 n 1 r u j 1 n 1 u j n displaystyle 1 2r u j n 1 ru j 1 n 1 ru j 1 n 1 u j n The scheme is always numerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step The errors are linear over the time step and quadratic over the space step D u O k O h 2 displaystyle Delta u O k O h 2 Crank Nicolson method Edit Finally if we use the central difference at time t n 1 2 displaystyle t n 1 2 and a second order central difference for the space derivative at position x j displaystyle x j CTCS we get the recurrence equation u j n 1 u j n k 1 2 u j 1 n 1 2 u j n 1 u j 1 n 1 h 2 u j 1 n 2 u j n u j 1 n h 2 displaystyle frac u j n 1 u j n k frac 1 2 left frac u j 1 n 1 2u j n 1 u j 1 n 1 h 2 frac u j 1 n 2u j n u j 1 n h 2 right This formula is known as the Crank Nicolson method The Crank Nicolson stencil We can obtain u j n 1 displaystyle u j n 1 from solving a system of linear equations 2 2 r u j n 1 r u j 1 n 1 r u j 1 n 1 2 2 r u j n r u j 1 n r u j 1 n displaystyle 2 2r u j n 1 ru j 1 n 1 ru j 1 n 1 2 2r u j n ru j 1 n ru j 1 n The scheme is always numerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step The errors are quadratic over both the time step and the space step D u O k 2 O h 2 displaystyle Delta u O k 2 O h 2 Comparison Edit To summarize usually the Crank Nicolson scheme is the most accurate scheme for small time steps For larger time steps the implicit scheme works better since it is less computationally demanding The explicit scheme is the least accurate and can be unstable but is also the easiest to implement and the least numerically intensive Here is an example The figures below present the solutions given by the above methods to approximate the heat equation U t a U x x a 1 p 2 displaystyle U t alpha U xx quad alpha frac 1 pi 2 with the boundary condition U 0 t U 1 t 0 displaystyle U 0 t U 1 t 0 The exact solution is U x t 1 p 2 e t sin p x displaystyle U x t frac 1 pi 2 e t sin pi x Comparison of Finite Difference Methods Explicit method not stable Implicit method stable Crank Nicolson method stable Example The Laplace operator EditThe continuous Laplace operator in n displaystyle n dimensions is given by D u x i 1 n i 2 u x displaystyle Delta u x sum i 1 n partial i 2 u x The discrete Laplace operator D h u displaystyle Delta h u depends on the dimension n displaystyle n In 1D the Laplace operator is approximated as D u x u x u x h 2 u x u x h h 2 D h u x displaystyle Delta u x u x approx frac u x h 2u x u x h h 2 Delta h u x This approximation is usually expressed via the following stencil D h 1 h 2 1 2 1 displaystyle Delta h frac 1 h 2 begin bmatrix 1 amp 2 amp 1 end bmatrix and which represents a symmetric tridiagonal matrix For an equidistant grid one gets a Toeplitz matrix The 2D case shows all the characteristics of the more general nD case Each second partial derivative needs to be approximated similar to the 1D case D u x y u x x x y u y y x y u x h y 2 u x y u x h y h 2 u x y h 2 u x y u x y h h 2 u x h y u x h y 4 u x y u x y h u x y h h 2 D h u x y displaystyle begin aligned Delta u x y amp u xx x y u yy x y amp approx frac u x h y 2u x y u x h y h 2 frac u x y h 2u x y u x y h h 2 amp frac u x h y u x h y 4u x y u x y h u x y h h 2 amp Delta h u x y end aligned which is usually given by the following stencil D h 1 h 2 1 1 4 1 1 displaystyle Delta h frac 1 h 2 begin bmatrix amp 1 1 amp 4 amp 1 amp 1 end bmatrix Consistency Edit Consistency of the above mentioned approximation can be shown for highly regular functions such as u C 4 W displaystyle u in C 4 Omega The statement is D u D h u O h 2 displaystyle Delta u Delta h u mathcal O h 2 To prove this one needs to substitute Taylor Series expansions up to order 3 into the discrete Laplace operator Properties Edit Subharmonic Edit Similar to continuous subharmonic functions one can define subharmonic functions for finite difference approximations u h displaystyle u h D h u h 0 displaystyle Delta h u h leq 0 Mean value Edit One can define a general stencil of positive type via a N a W a C a E a S a i gt 0 a C i N E S W a i displaystyle begin bmatrix amp alpha N alpha W amp alpha C amp alpha E amp alpha S end bmatrix quad alpha i gt 0 quad alpha C sum i in N E S W alpha i If u h displaystyle u h is discrete subharmonic then the followingmean value property holds u h x C i N E S W a i u h x i i N E S W a i displaystyle u h x C leq frac sum i in N E S W alpha i u h x i sum i in N E S W alpha i where the approximation is evaluated on points of the grid and the stencil is assumed to be of positive type A similar mean value property also holds for the continuous case Maximum principle Edit For a discrete subharmonic function u h displaystyle u h the following holds max W h u h max W h u h displaystyle max Omega h u h leq max partial Omega h u h where W h W h displaystyle Omega h partial Omega h are discretizations of the continuous domain W displaystyle Omega respectively the boundary W displaystyle partial Omega A similar maximum principle also holds for the continuous case The SBP SAT method EditThe SBP SAT summation by parts simultaneous approximation term method is a stable and accurate technique for discretizing and imposing boundary conditions of a well posed partial differential equation using high order finite differences 8 9 The method is based on finite differences where the differentiation operators exhibit summation by parts properties Typically these operators consist of differentiation matrices with central difference stencils in the interior with carefully chosen one sided boundary stencils designed to mimic integration by parts in the discrete setting Using the SAT technique the boundary conditions of the PDE are imposed weakly where the boundary values are pulled towards the desired conditions rather than exactly fulfilled If the tuning parameters inherent to the SAT technique are chosen properly the resulting system of ODE s will exhibit similar energy behavior as the continuous PDE i e the system has no non physical energy growth This guarantees stability if an integration scheme with a stability region that includes parts of the imaginary axis such as the fourth order Runge Kutta method is used This makes the SAT technique an attractive method of imposing boundary conditions for higher order finite difference methods in contrast to for example the injection method which typically will not be stable if high order differentiation operators are used See also EditFinite element method Finite difference Finite difference time domain Infinite difference method Stencil numerical analysis Finite difference coefficients Five point stencil Lax Richtmyer theorem Finite difference methods for option pricing Upwind differencing scheme for convection Central differencing scheme Discrete Poisson equation Discrete Laplace operatorReferences Edit a b Christian Grossmann Hans G Roos Martin Stynes 2007 Numerical Treatment of Partial Differential Equations Springer Science amp Business Media p 23 ISBN 978 3 540 71584 9 Arieh Iserles 2008 A first course in the numerical analysis of differential equations Cambridge University Press p 23 ISBN 9780521734905 a b Hoffman JD Frankel S 2001 Numerical methods for engineers and scientists CRC Press Boca Raton a b Jaluria Y Atluri S 1994 Computational heat transfer Computational Mechanics 14 5 385 386 Bibcode 1994CompM 14 385J doi 10 1007 BF00377593 S2CID 119502676 Majumdar P 2005 Computational methods for heat and mass transfer 1st ed Taylor and Francis New York Smith GD 1985 Numerical solution of partial differential equations finite difference methods 3rd ed Oxford University Press Crank J The Mathematics of Diffusion 2nd Edition Oxford 1975 p 143 Bo Strand 1994 Summation by Parts for Finite Difference Approximations for d dx Journal of Computational Physics 110 1 47 67 Bibcode 1994JCoPh 110 47S doi 10 1006 jcph 1994 1005 Mark H Carpenter David I Gottlieb Saul S Abarbanel 1994 Time stable boundary conditions for finite difference schemes solving hyperbolic systems Methodology and application to high order compact schemes Journal of Computational Physics 111 2 220 236 Bibcode 1994JCoPh 111 220C doi 10 1006 jcph 1994 1057 hdl 2060 19930013937 Further reading EditK W Morton and D F Mayers Numerical Solution of Partial Differential Equations An Introduction Cambridge University Press 2005 Autar Kaw and E Eric Kalu Numerical Methods with Applications 2008 1 Contains a brief engineering oriented introduction to FDM for ODEs in Chapter 08 07 John Strikwerda 2004 Finite Difference Schemes and Partial Differential Equations 2nd ed SIAM ISBN 978 0 89871 639 9 Smith G D 1985 Numerical Solution of Partial Differential Equations Finite Difference Methods 3rd ed Oxford University Press Peter Olver 2013 Introduction to Partial Differential Equations Springer Chapter 5 Finite differences ISBN 978 3 319 02099 0 Randall J LeVeque Finite Difference Methods for Ordinary and Partial Differential Equations SIAM 2007 Sergey Lemeshevsky Piotr Matus Dmitriy Poliakov Eds Exact Finite Difference Schemes De Gruyter 2016 DOI https doi org 10 1515 9783110491326 Retrieved from https en wikipedia org w index php title Finite difference method amp oldid 1126400243, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.