fbpx
Wikipedia

Newton polynomial

In the mathematical field of numerical analysis, a Newton polynomial, named after its inventor Isaac Newton,[1] is an interpolation polynomial for a given set of data points. The Newton polynomial is sometimes called Newton's divided differences interpolation polynomial because the coefficients of the polynomial are calculated using Newton's divided differences method.

Definition edit

Given a set of k + 1 data points

 

where no two xj are the same, the Newton interpolation polynomial is a linear combination of Newton basis polynomials

 

with the Newton basis polynomials defined as

 

for j > 0 and  .

The coefficients are defined as

 

where

 

is the notation for divided differences.

Thus the Newton polynomial can be written as

 

Newton forward divided difference formula edit

The Newton polynomial can be expressed in a simplified form when   are arranged consecutively with equal spacing.

If   are consecutively arranged and equally spaced with   for i = 0, 1, ..., k and some variable x is expressed as  , then the difference   can be written as  . So the Newton polynomial becomes

 

This is called the Newton forward divided difference formula.[citation needed]

Newton backward divided difference formula edit

If the nodes are reordered as  , the Newton polynomial becomes

 

If   are equally spaced with   for i = 0, 1, ..., k and  , then,

 

This is called the Newton backward divided difference formula.[citation needed]

Significance edit

Newton's formula is of interest because it is the straightforward and natural differences-version of Taylor's polynomial. Taylor's polynomial tells where a function will go, based on its y value, and its derivatives (its rate of change, and the rate of change of its rate of change, etc.) at one particular x value. Newton's formula is Taylor's polynomial based on finite differences instead of instantaneous rates of change.

Polynomial interpolation edit

For a polynomial   of degree less than or equal to n, that interpolates   at the nodes   where  . Let   be the polynomial of degree less than or equal to n+1 that interpolates   at the nodes   where  . Then   is given by:

 

where   and  .

Proof:

This can be shown for the case where  :

 
and when  :
 

By the uniqueness of interpolated polynomials of degree less than  ,   is the required polynomial interpolation. The function can thus be expressed as:

 

where the factors   are divided differences. Thus, Newton polynomials are used to provide polynomial interpolation formula of n points.[2]


Taking   for some unknown function in Newton divided difference formulas, if the representation of x in the previous sections was instead taken to be  , in terms of forward differences, the Newton forward interpolation formula is expressed as:

 
whereas for the same in terms of backward differences, the Newton backward interpolation formula is expressed as:
 
This follows since relationship between divided differences and forward differences is given as:[3]
 
whereas for backward differences, it is given as:[citation needed]
 

Addition of new points edit

As with other difference formulas, the degree of a Newton interpolating polynomial can be increased by adding more terms and points without discarding existing ones. Newton's form has the simplicity that the new points are always added at one end: Newton's forward formula can add new points to the right, and Newton's backward formula can add new points to the left.

The accuracy of polynomial interpolation depends on how close the interpolated point is to the middle of the x values of the set of points used. Obviously, as new points are added at one end, that middle becomes farther and farther from the first data point. Therefore, if it isn't known how many points will be needed for the desired accuracy, the middle of the x-values might be far from where the interpolation is done.

Gauss, Stirling, and Bessel all developed formulae to remedy that problem.[4]

Gauss's formula alternately adds new points at the left and right ends, thereby keeping the set of points centered near the same place (near the evaluated point). When so doing, it uses terms from Newton's formula, with data points and x values renamed in keeping with one's choice of what data point is designated as the x0 data point.

Stirling's formula remains centered about a particular data point, for use when the evaluated point is nearer to a data point than to a middle of two data points.

Bessel's formula remains centered about a particular middle between two data points, for use when the evaluated point is nearer to a middle than to a data point.

Bessel and Stirling achieve that by sometimes using the average of two differences, and sometimes using the average of two products of binomials in x, where Newton's or Gauss's would use just one difference or product. Stirling's uses an average difference in odd-degree terms (whose difference uses an even number of data points); Bessel's uses an average difference in even-degree terms (whose difference uses an odd number of data points).

Strengths and weaknesses of various formulae edit

For any given finite set of data points, there is only one polynomial of least possible degree that passes through all of them. Thus, it is appropriate to speak of the "Newton form", or Lagrange form, etc., of the interpolation polynomial. However, different methods of computing this polynomial can have differing computational efficiency. There are several similar methods, such as those of Gauss, Bessel and Stirling. They can be derived from Newton's by renaming the x-values of the data points, but in practice they are important.

Bessel vs. Stirling edit

The choice between Bessel and Stirling depends on whether the interpolated point is closer to a data point, or closer to a middle between two data points.

A polynomial interpolation's error approaches zero, as the interpolation point approaches a data-point. Therefore, Stirling's formula brings its accuracy improvement where it is least needed and Bessel brings its accuracy improvement where it is most needed.

So, Bessel's formula could be said to be the most consistently accurate difference formula, and, in general, the most consistently accurate of the familiar polynomial interpolation formulas.

Divided-Difference Methods vs. Lagrange edit

Lagrange is sometimes said to require less work, and is sometimes recommended for problems in which it is known, in advance, from previous experience, how many terms are needed for sufficient accuracy.

The divided difference methods have the advantage that more data points can be added, for improved accuracy. The terms based on the previous data points can continue to be used. With the ordinary Lagrange formula, to do the problem with more data points would require re-doing the whole problem.

There is a "barycentric" version of Lagrange that avoids the need to re-do the entire calculation when adding a new data point. But it requires that the values of each term be recorded.

But the ability, of Gauss, Bessel and Stirling, to keep the data points centered close to the interpolated point gives them an advantage over Lagrange, when it isn't known, in advance, how many data points will be needed.

Additionally, suppose that one wants to find out if, for some particular type of problem, linear interpolation is sufficiently accurate. That can be determined by evaluating the quadratic term of a divided difference formula. If the quadratic term is negligible—meaning that the linear term is sufficiently accurate without adding the quadratic term—then linear interpolation is sufficiently accurate. If the problem is sufficiently important, or if the quadratic term is nearly big enough to matter, then one might want to determine whether the sum of the quadratic and cubic terms is large enough to matter in the problem.

Of course, only a divided-difference method can be used for such a determination.

For that purpose, the divided-difference formula and/or its x0 point should be chosen so that the formula will use, for its linear term, the two data points between which the linear interpolation of interest would be done.

The divided difference formulas are more versatile, useful in more kinds of problems.

The Lagrange formula is at its best when all the interpolation will be done at one x value, with only the data points' y values varying from one problem to another, and when it is known, from past experience, how many terms are needed for sufficient accuracy.

With the Newton form of the interpolating polynomial a compact and effective algorithm exists for combining the terms to find the coefficients of the polynomial.[5]

Accuracy edit

When, with Stirling's or Bessel's, the last term used includes the average of two differences, then one more point is being used than Newton's or other polynomial interpolations would use for the same polynomial degree. So, in that instance, Stirling's or Bessel's is not putting an N−1 degree polynomial through N points, but is, instead, trading equivalence with Newton's for better centering and accuracy, giving those methods sometimes potentially greater accuracy, for a given polynomial degree, than other polynomial interpolations.

General case edit

For the special case of xi = i, there is a closely related set of polynomials, also called the Newton polynomials, that are simply the binomial coefficients for general argument. That is, one also has the Newton polynomials   given by

 

In this form, the Newton polynomials generate the Newton series. These are in turn a special case of the general difference polynomials which allow the representation of analytic functions through generalized difference equations.

Main idea edit

Solving an interpolation problem leads to a problem in linear algebra where we have to solve a system of linear equations. Using a standard monomial basis for our interpolation polynomial we get the very complicated Vandermonde matrix. By choosing another basis, the Newton basis, we get a system of linear equations with a much simpler lower triangular matrix which can be solved faster.

For k + 1 data points we construct the Newton basis as

 

Using these polynomials as a basis for   we have to solve

 

to solve the polynomial interpolation problem.

This system of equations can be solved iteratively by solving

 

Derivation edit

While the interpolation formula can be found by solving a linear system of equations, there is a loss of intuition in what the formula is showing and why Newton's interpolation formula works is not readily apparent. To begin, we will need to establish two facts first:

Fact 1. Reversing the terms of a divided difference leaves it unchanged:  

The proof of this is an easy induction: for   we compute

 

Induction step: Suppose the result holds for any divided difference involving at most   terms. Then using the induction hypothesis in the following 2nd equality we see that for a divided difference involving   terms we have

 

We formulate next Fact 2 which for purposes of induction and clarity we also call Statement   ( ) :

Fact 2. ( ) : If   are any   points with distinct  -coordinates and   is the unique polynomial of degree (at most)   whose graph passes through these   points then there holds the relation

 

Proof. (It will be helpful for fluent reading of the proof to have the precise statement and its subtlety in mind:   is defined by passing through   but the formula also speaks at both sides of an additional arbitrary point   with  -coordinate distinct from the other  .)

We again prove these statements by induction. To show   let   be any one point and let   be the unique polynomial of degree 0 passing through  . Then evidently   and we can write

 
as wanted.

Proof of   assuming   already established: Let   be the polynomial of degree (at most)   passing through  

With   being the unique polynomial of degree (at most)   passing through the points  , we can write the following chain of equalities, where we use in the penultimate equality that Stm  applies to  :


 

The induction hypothesis for   also applies to the second equality in the following computation, where   is added to the points defining   :

 

Now look at   By the definition of   this polynomial passes through   and, as we have just shown, it also passes through   Thus it is the unique polynomial of degree   which passes through these points. Therefore this polynomial is   i.e.:  

Thus we can write the last line in the first chain of equalities as ` ' and have thus established that

 
So we established  , and hence completed the proof of Fact 2.

Now look at Fact 2: It can be formulated this way: If   is the unique polynomial of degree at most   whose graph passes through the points   then   is the unique polynomial of degree at most   passing through points   So we see Newton interpolation permits indeed to add new interpolation points without destroying what has already been computed.

Taylor polynomial edit

The limit of the Newton polynomial if all nodes coincide is a Taylor polynomial, because the divided differences become derivatives.

 

Application edit

As can be seen from the definition of the divided differences new data points can be added to the data set to create a new interpolation polynomial without recalculating the old coefficients. And when a data point changes we usually do not have to recalculate all coefficients. Furthermore, if the xi are distributed equidistantly the calculation of the divided differences becomes significantly easier. Therefore, the divided-difference formulas are usually preferred over the Lagrange form for practical purposes.

Examples edit

The divided differences can be written in the form of a table. For example, for a function f is to be interpolated on points  . Write

 

Then the interpolating polynomial is formed as above using the topmost entries in each column as coefficients.

For example, suppose we are to construct the interpolating polynomial to f(x) = tan(x) using divided differences, at the points

     
     
     
     
     
     

Using six digits of accuracy, we construct the table

 

Thus, the interpolating polynomial is

 

Given more digits of accuracy in the table, the first and third coefficients will be found to be zero.

Another example:

The sequence   such that   and  , i.e., they are   from   to  .

You obtain the slope of order   in the following way:

  •  
  •  
  •  

As we have the slopes of order  , it is possible to obtain the next order:

  •  
  •  

Finally, we define the slope of order  :

  •  

Once we have the slope, we can define the consequent polynomials:

  •  .
  •  
  •  .
  •  

See also edit

References edit

  1. ^ Dunham, William (1990). "7". Journey Through Genius: The Great Theorems of Mathematics. Kanak Agrawal, Inc. pp. 155–183. ISBN 9780140147391. Retrieved 24 October 2019.
  2. ^ Epperson, James F. (2013). An introduction to numerical methods and analysis (2nd ed.). Hoboken, NJ: Wiley. ISBN 978-1-118-36759-9.
  3. ^ Burden, Richard L.; Faires, J. Douglas (2011). Numerical Analysis (9th ed.). Cengage Learning. p. 129. ISBN 9780538733519.
  4. ^ Hamming, Richard W. (1986). Numerical methods for scientists and engineers (Unabridged republ. of the 2. ed. (1973) ed.). New York: Dover. ISBN 978-0-486-65241-2.
  5. ^ Stetekluh, Jeff. "Algorithm for the Newton Form of the Interpolating Polynomial".

External links edit

    newton, polynomial, mathematical, field, numerical, analysis, named, after, inventor, isaac, newton, interpolation, polynomial, given, data, points, sometimes, called, newton, divided, differences, interpolation, polynomial, because, coefficients, polynomial, . In the mathematical field of numerical analysis a Newton polynomial named after its inventor Isaac Newton 1 is an interpolation polynomial for a given set of data points The Newton polynomial is sometimes called Newton s divided differences interpolation polynomial because the coefficients of the polynomial are calculated using Newton s divided differences method Contents 1 Definition 1 1 Newton forward divided difference formula 1 2 Newton backward divided difference formula 2 Significance 2 1 Polynomial interpolation 3 Addition of new points 4 Strengths and weaknesses of various formulae 4 1 Bessel vs Stirling 4 2 Divided Difference Methods vs Lagrange 4 3 Accuracy 5 General case 6 Main idea 7 Derivation 8 Taylor polynomial 9 Application 9 1 Examples 10 See also 11 References 12 External linksDefinition editGiven a set of k 1 data points x 0 y 0 x j y j x k y k displaystyle x 0 y 0 ldots x j y j ldots x k y k nbsp where no two xj are the same the Newton interpolation polynomial is a linear combination of Newton basis polynomials N x j 0 k a j n j x displaystyle N x sum j 0 k a j n j x nbsp with the Newton basis polynomials defined as n j x i 0 j 1 x x i displaystyle n j x prod i 0 j 1 x x i nbsp for j gt 0 and n 0 x 1 displaystyle n 0 x equiv 1 nbsp The coefficients are defined as a j y 0 y j displaystyle a j y 0 ldots y j nbsp where y 0 y j displaystyle y 0 ldots y j nbsp is the notation for divided differences Thus the Newton polynomial can be written as N x y 0 y 0 y 1 x x 0 y 0 y k x x 0 x x 1 x x k 1 displaystyle N x y 0 y 0 y 1 x x 0 cdots y 0 ldots y k x x 0 x x 1 cdots x x k 1 nbsp Newton forward divided difference formula edit The Newton polynomial can be expressed in a simplified form when x 0 x 1 x k displaystyle x 0 x 1 dots x k nbsp are arranged consecutively with equal spacing If x 0 x 1 x k displaystyle x 0 x 1 dots x k nbsp are consecutively arranged and equally spaced with x i x 0 i h displaystyle x i x 0 ih nbsp for i 0 1 k and some variable x is expressed as x x 0 s h displaystyle x x 0 sh nbsp then the difference x x i displaystyle x x i nbsp can be written as s i h displaystyle s i h nbsp So the Newton polynomial becomes N x y 0 y 0 y 1 s h y 0 y k s s 1 s k 1 h k i 0 k s s 1 s i 1 h i y 0 y i i 0 k s i i h i y 0 y i displaystyle begin aligned N x amp y 0 y 0 y 1 sh cdots y 0 ldots y k s s 1 cdots s k 1 h k amp sum i 0 k s s 1 cdots s i 1 h i y 0 ldots y i amp sum i 0 k s choose i i h i y 0 ldots y i end aligned nbsp This is called the Newton forward divided difference formula citation needed Newton backward divided difference formula edit If the nodes are reordered as x k x k 1 x 0 displaystyle x k x k 1 dots x 0 nbsp the Newton polynomial becomes N x y k y k y k 1 x x k y k y 0 x x k x x k 1 x x 1 displaystyle N x y k y k y k 1 x x k cdots y k ldots y 0 x x k x x k 1 cdots x x 1 nbsp If x k x k 1 x 0 displaystyle x k x k 1 dots x 0 nbsp are equally spaced with x i x k k i h displaystyle x i x k k i h nbsp for i 0 1 k and x x k s h displaystyle x x k sh nbsp then N x y k y k y k 1 s h y k y 0 s s 1 s k 1 h k i 0 k 1 i s i i h i y k y k i displaystyle begin aligned N x amp y k y k y k 1 sh cdots y k ldots y 0 s s 1 cdots s k 1 h k amp sum i 0 k 1 i s choose i i h i y k ldots y k i end aligned nbsp This is called the Newton backward divided difference formula citation needed Significance editFurther information Finite differences Newton s series Newton s formula is of interest because it is the straightforward and natural differences version of Taylor s polynomial Taylor s polynomial tells where a function will go based on its y value and its derivatives its rate of change and the rate of change of its rate of change etc at one particular x value Newton s formula is Taylor s polynomial based on finite differences instead of instantaneous rates of change Polynomial interpolation edit Main article Polynomial interpolation For a polynomial p n displaystyle p n nbsp of degree less than or equal to n that interpolates f displaystyle f nbsp at the nodes x i displaystyle x i nbsp where i 0 1 2 3 n displaystyle i 0 1 2 3 cdots n nbsp Let p n 1 displaystyle p n 1 nbsp be the polynomial of degree less than or equal to n 1 that interpolates f displaystyle f nbsp at the nodes x i displaystyle x i nbsp where i 0 1 2 3 n n 1 displaystyle i 0 1 2 3 cdots n n 1 nbsp Then p n 1 displaystyle p n 1 nbsp is given by p n 1 x p n x a n 1 w n x displaystyle p n 1 x p n x a n 1 w n x nbsp where w n x i 0 n x x i textstyle w n x prod i 0 n x x i nbsp and a n 1 f x n 1 p n x n 1 w n x n 1 textstyle a n 1 f x n 1 p n x n 1 over w n x n 1 nbsp Proof This can be shown for the case where i 0 1 2 3 n displaystyle i 0 1 2 3 cdots n nbsp p n 1 x i p n x i a n 1 j 0 n x i x j p n x i displaystyle p n 1 x i p n x i a n 1 prod j 0 n x i x j p n x i nbsp and when i n 1 displaystyle i n 1 nbsp p n 1 x n 1 p n x n 1 f x n 1 p n x n 1 w n x n 1 w n x n 1 f x n 1 displaystyle p n 1 x n 1 p n x n 1 f x n 1 p n x n 1 over w n x n 1 w n x n 1 f x n 1 nbsp By the uniqueness of interpolated polynomials of degree less than n 1 displaystyle n 1 nbsp p n 1 x p n x a n 1 w n x textstyle p n 1 x p n x a n 1 w n x nbsp is the required polynomial interpolation The function can thus be expressed as p n x a 0 a 1 x x 0 a 2 x x 0 x x 1 a n x x 0 x x n 1 textstyle p n x a 0 a 1 x x 0 a 2 x x 0 x x 1 cdots a n x x 0 cdots x x n 1 nbsp where the factors a i displaystyle a i nbsp are divided differences Thus Newton polynomials are used to provide polynomial interpolation formula of n points 2 Taking y i f x i displaystyle y i f x i nbsp for some unknown function in Newton divided difference formulas if the representation of x in the previous sections was instead taken to be x x j s h displaystyle x x j sh nbsp in terms of forward differences the Newton forward interpolation formula is expressed as f x N x N x j s h i 0 k s i D i f x j displaystyle f x approx N x N x j sh sum i 0 k s choose i Delta i f x j nbsp whereas for the same in terms of backward differences the Newton backward interpolation formula is expressed as f x N x N x j s h i 0 k 1 i s i i f x j displaystyle f x approx N x N x j sh sum i 0 k 1 i s choose i nabla i f x j nbsp This follows since relationship between divided differences and forward differences is given as 3 y j y j 1 y j n 1 n h n D n y j displaystyle y j y j 1 ldots y j n frac 1 n h n Delta n y j nbsp whereas for backward differences it is given as citation needed y j y j 1 y j n 1 n h n n y j displaystyle y j y j 1 ldots y j n frac 1 n h n nabla n y j nbsp Addition of new points editAs with other difference formulas the degree of a Newton interpolating polynomial can be increased by adding more terms and points without discarding existing ones Newton s form has the simplicity that the new points are always added at one end Newton s forward formula can add new points to the right and Newton s backward formula can add new points to the left The accuracy of polynomial interpolation depends on how close the interpolated point is to the middle of the x values of the set of points used Obviously as new points are added at one end that middle becomes farther and farther from the first data point Therefore if it isn t known how many points will be needed for the desired accuracy the middle of the x values might be far from where the interpolation is done Gauss Stirling and Bessel all developed formulae to remedy that problem 4 Gauss s formula alternately adds new points at the left and right ends thereby keeping the set of points centered near the same place near the evaluated point When so doing it uses terms from Newton s formula with data points and x values renamed in keeping with one s choice of what data point is designated as the x0 data point Stirling s formula remains centered about a particular data point for use when the evaluated point is nearer to a data point than to a middle of two data points Bessel s formula remains centered about a particular middle between two data points for use when the evaluated point is nearer to a middle than to a data point Bessel and Stirling achieve that by sometimes using the average of two differences and sometimes using the average of two products of binomials in x where Newton s or Gauss s would use just one difference or product Stirling s uses an average difference in odd degree terms whose difference uses an even number of data points Bessel s uses an average difference in even degree terms whose difference uses an odd number of data points Strengths and weaknesses of various formulae editFor any given finite set of data points there is only one polynomial of least possible degree that passes through all of them Thus it is appropriate to speak of the Newton form or Lagrange form etc of the interpolation polynomial However different methods of computing this polynomial can have differing computational efficiency There are several similar methods such as those of Gauss Bessel and Stirling They can be derived from Newton s by renaming the x values of the data points but in practice they are important Bessel vs Stirling edit The choice between Bessel and Stirling depends on whether the interpolated point is closer to a data point or closer to a middle between two data points A polynomial interpolation s error approaches zero as the interpolation point approaches a data point Therefore Stirling s formula brings its accuracy improvement where it is least needed and Bessel brings its accuracy improvement where it is most needed So Bessel s formula could be said to be the most consistently accurate difference formula and in general the most consistently accurate of the familiar polynomial interpolation formulas Divided Difference Methods vs Lagrange edit Lagrange is sometimes said to require less work and is sometimes recommended for problems in which it is known in advance from previous experience how many terms are needed for sufficient accuracy The divided difference methods have the advantage that more data points can be added for improved accuracy The terms based on the previous data points can continue to be used With the ordinary Lagrange formula to do the problem with more data points would require re doing the whole problem There is a barycentric version of Lagrange that avoids the need to re do the entire calculation when adding a new data point But it requires that the values of each term be recorded But the ability of Gauss Bessel and Stirling to keep the data points centered close to the interpolated point gives them an advantage over Lagrange when it isn t known in advance how many data points will be needed Additionally suppose that one wants to find out if for some particular type of problem linear interpolation is sufficiently accurate That can be determined by evaluating the quadratic term of a divided difference formula If the quadratic term is negligible meaning that the linear term is sufficiently accurate without adding the quadratic term then linear interpolation is sufficiently accurate If the problem is sufficiently important or if the quadratic term is nearly big enough to matter then one might want to determine whether the sum of the quadratic and cubic terms is large enough to matter in the problem Of course only a divided difference method can be used for such a determination For that purpose the divided difference formula and or its x0 point should be chosen so that the formula will use for its linear term the two data points between which the linear interpolation of interest would be done The divided difference formulas are more versatile useful in more kinds of problems The Lagrange formula is at its best when all the interpolation will be done at one x value with only the data points y values varying from one problem to another and when it is known from past experience how many terms are needed for sufficient accuracy With the Newton form of the interpolating polynomial a compact and effective algorithm exists for combining the terms to find the coefficients of the polynomial 5 Accuracy edit When with Stirling s or Bessel s the last term used includes the average of two differences then one more point is being used than Newton s or other polynomial interpolations would use for the same polynomial degree So in that instance Stirling s or Bessel s is not putting an N 1 degree polynomial through N points but is instead trading equivalence with Newton s for better centering and accuracy giving those methods sometimes potentially greater accuracy for a given polynomial degree than other polynomial interpolations General case editFor the special case of xi i there is a closely related set of polynomials also called the Newton polynomials that are simply the binomial coefficients for general argument That is one also has the Newton polynomials p n z displaystyle p n z nbsp given by p n z z n z z 1 z n 1 n displaystyle p n z z choose n frac z z 1 cdots z n 1 n nbsp In this form the Newton polynomials generate the Newton series These are in turn a special case of the general difference polynomials which allow the representation of analytic functions through generalized difference equations Main idea editSolving an interpolation problem leads to a problem in linear algebra where we have to solve a system of linear equations Using a standard monomial basis for our interpolation polynomial we get the very complicated Vandermonde matrix By choosing another basis the Newton basis we get a system of linear equations with a much simpler lower triangular matrix which can be solved faster For k 1 data points we construct the Newton basis as n 0 x 1 n j x i 0 j 1 x x i j 1 k displaystyle n 0 x 1 qquad n j x prod i 0 j 1 x x i qquad j 1 ldots k nbsp Using these polynomials as a basis for P k displaystyle Pi k nbsp we have to solve 1 0 1 x 1 x 0 1 x 2 x 0 x 2 x 0 x 2 x 1 1 x k x 0 j 0 k 1 x k x j a 0 a k y 0 y k displaystyle begin bmatrix 1 amp amp ldots amp amp 0 1 amp x 1 x 0 amp amp amp 1 amp x 2 x 0 amp x 2 x 0 x 2 x 1 amp amp vdots vdots amp vdots amp amp ddots amp 1 amp x k x 0 amp ldots amp ldots amp prod j 0 k 1 x k x j end bmatrix begin bmatrix a 0 vdots a k end bmatrix begin bmatrix y 0 vdots y k end bmatrix nbsp to solve the polynomial interpolation problem This system of equations can be solved iteratively by solving i 0 j a i n i x j y j j 0 k displaystyle sum i 0 j a i n i x j y j qquad j 0 dots k nbsp Derivation editWhile the interpolation formula can be found by solving a linear system of equations there is a loss of intuition in what the formula is showing and why Newton s interpolation formula works is not readily apparent To begin we will need to establish two facts first Fact 1 Reversing the terms of a divided difference leaves it unchanged y 0 y n y n y 0 displaystyle y 0 ldots y n y n ldots y 0 nbsp The proof of this is an easy induction for n 1 displaystyle n 1 nbsp we compute y 0 y 1 y 1 y 0 x 1 x 0 y 0 y 1 x 0 x 1 y 1 y 0 displaystyle y 0 y 1 frac y 1 y 0 x 1 x 0 frac y 0 y 1 x 0 x 1 y 1 y 0 nbsp Induction step Suppose the result holds for any divided difference involving at most n 1 displaystyle n 1 nbsp terms Then using the induction hypothesis in the following 2nd equality we see that for a divided difference involving n 2 displaystyle n 2 nbsp terms we have y 0 y n 1 y 1 y n 1 y 0 y n x n 1 x 0 y n y 0 y n 1 y 1 x 0 x n 1 y n 1 y 0 displaystyle y 0 ldots y n 1 frac y 1 ldots y n 1 y 0 ldots y n x n 1 x 0 frac y n ldots y 0 y n 1 ldots y 1 x 0 x n 1 y n 1 ldots y 0 nbsp We formulate next Fact 2 which for purposes of induction and clarity we also call Statement n displaystyle n nbsp Stm n displaystyle text Stm n nbsp Fact 2 Stm n displaystyle text Stm n nbsp If x 0 y 0 x n 1 y n 1 displaystyle x 0 y 0 ldots x n 1 y n 1 nbsp are any n displaystyle n nbsp points with distinct x displaystyle x nbsp coordinates and P P x displaystyle P P x nbsp is the unique polynomial of degree at most n 1 displaystyle n 1 nbsp whose graph passes through these n displaystyle n nbsp points then there holds the relation y 0 y n x n x 0 x n x n 1 y n P x n displaystyle y 0 ldots y n x n x 0 cdot ldots cdot x n x n 1 y n P x n nbsp Proof It will be helpful for fluent reading of the proof to have the precise statement and its subtlety in mind P displaystyle P nbsp is defined by passing through x 0 y 0 x n 1 y n 1 displaystyle x 0 y 0 x n 1 y n 1 nbsp but the formula also speaks at both sides of an additional arbitrary point x n y n displaystyle x n y n nbsp with x displaystyle x nbsp coordinate distinct from the other x i displaystyle x i nbsp We again prove these statements by induction To show Stm 1 displaystyle text Stm 1 nbsp let x 0 y 0 displaystyle x 0 y 0 nbsp be any one point and let P x displaystyle P x nbsp be the unique polynomial of degree 0 passing through x 0 y 0 displaystyle x 0 y 0 nbsp Then evidently P x y 0 displaystyle P x y 0 nbsp and we can write y 0 y 1 x 1 x 0 y 1 y 0 x 1 x 0 x 1 x 0 y 1 y 0 y 1 P x 1 displaystyle y 0 y 1 x 1 x 0 frac y 1 y 0 x 1 x 0 x 1 x 0 y 1 y 0 y 1 P x 1 nbsp as wanted Proof of Stm n 1 displaystyle text Stm n 1 nbsp assuming Stm n displaystyle text Stm n nbsp already established Let P x displaystyle P x nbsp be the polynomial of degree at most n displaystyle n nbsp passing through x 0 y 0 x n y n displaystyle x 0 y 0 ldots x n y n nbsp With Q x displaystyle Q x nbsp being the unique polynomial of degree at most n 1 displaystyle n 1 nbsp passing through the points x 1 y 1 x n y n displaystyle x 1 y 1 ldots x n y n nbsp we can write the following chain of equalities where we use in the penultimate equality that Stmn displaystyle n nbsp applies to Q displaystyle Q nbsp y 0 y n 1 x n 1 x 0 x n 1 x n y 1 y n 1 y 0 y n x n 1 x 0 x n 1 x 0 x n 1 x n y 1 y n 1 y 0 y n x n 1 x 1 x n 1 x n y 1 y n 1 x n 1 x 1 x n 1 x n y 0 y n x n 1 x 1 x n 1 x n y n 1 Q x n 1 y 0 y n x n 1 x 1 x n 1 x n y n 1 Q x n 1 y 0 y n x n 1 x 1 x n 1 x n displaystyle begin aligned amp y 0 ldots y n 1 x n 1 x 0 cdot ldots cdot x n 1 x n amp frac y 1 ldots y n 1 y 0 ldots y n x n 1 x 0 x n 1 x 0 cdot ldots cdot x n 1 x n amp left y 1 ldots y n 1 y 0 ldots y n right x n 1 x 1 cdot ldots cdot x n 1 x n amp y 1 ldots y n 1 x n 1 x 1 cdot ldots cdot x n 1 x n y 0 ldots y n x n 1 x 1 cdot ldots cdot x n 1 x n amp y n 1 Q x n 1 y 0 ldots y n x n 1 x 1 cdot ldots cdot x n 1 x n amp y n 1 Q x n 1 y 0 ldots y n x n 1 x 1 cdot ldots cdot x n 1 x n end aligned nbsp The induction hypothesis for Q displaystyle Q nbsp also applies to the second equality in the following computation where x 0 y 0 displaystyle x 0 y 0 nbsp is added to the points defining Q displaystyle Q nbsp Q x 0 y 0 y n x 0 x 1 x 0 x n Q x 0 y n y 0 x 0 x n x 0 x 1 Q x 0 y 0 Q x 0 y 0 P x 0 displaystyle begin aligned amp Q x 0 y 0 ldots y n x 0 x 1 cdot ldots cdot x 0 x n amp Q x 0 y n ldots y 0 x 0 x n cdot ldots cdot x 0 x 1 amp Q x 0 y 0 Q x 0 amp y 0 amp P x 0 end aligned nbsp Now look at Q x y 0 y n x x 1 x x n displaystyle Q x y 0 ldots y n x x 1 cdot ldots cdot x x n nbsp By the definition of Q displaystyle Q nbsp this polynomial passes through x 1 y 1 x n y n displaystyle x 1 y 1 x n y n nbsp and as we have just shown it also passes through x 0 y 0 displaystyle x 0 y 0 nbsp Thus it is the unique polynomial of degree n displaystyle leq n nbsp which passes through these points Therefore this polynomial is P x displaystyle P x nbsp i e P x Q x y 0 y n x x 1 x x n displaystyle P x Q x y 0 ldots y n x x 1 cdot ldots cdot x x n nbsp Thus we can write the last line in the first chain of equalities as y n 1 P x n 1 displaystyle y n 1 P x n 1 nbsp and have thus established that y 0 y n 1 x n 1 x 0 x n 1 x n y n 1 P x n 1 displaystyle y 0 ldots y n 1 x n 1 x 0 cdot ldots cdot x n 1 x n y n 1 P x n 1 nbsp So we established Stm n 1 displaystyle text Stm n 1 nbsp and hence completed the proof of Fact 2 Now look at Fact 2 It can be formulated this way If P displaystyle P nbsp is the unique polynomial of degree at most n 1 displaystyle n 1 nbsp whose graph passes through the points x 0 y 0 x n 1 y n 1 displaystyle x 0 y 0 x n 1 y n 1 nbsp then P x y 0 y n x x 0 x x n 1 displaystyle P x y 0 ldots y n x x 0 cdot ldots cdot x x n 1 nbsp is the unique polynomial of degree at most n displaystyle n nbsp passing through points x 0 y 0 x n 1 y n 1 x n y n displaystyle x 0 y 0 x n 1 y n 1 x n y n nbsp So we see Newton interpolation permits indeed to add new interpolation points without destroying what has already been computed Taylor polynomial editThe limit of the Newton polynomial if all nodes coincide is a Taylor polynomial because the divided differences become derivatives lim x 0 x n z z f x 0 f x 0 x 1 3 x 0 f x 0 x n 3 x 0 3 x n 1 f z f z 3 z f n z n 3 z n displaystyle begin aligned amp lim x 0 dots x n to z dots z f x 0 f x 0 x 1 cdot xi x 0 dots f x 0 dots x n cdot xi x 0 cdot dots cdot xi x n 1 amp f z f z cdot xi z dots frac f n z n cdot xi z n end aligned nbsp Application editAs can be seen from the definition of the divided differences new data points can be added to the data set to create a new interpolation polynomial without recalculating the old coefficients And when a data point changes we usually do not have to recalculate all coefficients Furthermore if the xi are distributed equidistantly the calculation of the divided differences becomes significantly easier Therefore the divided difference formulas are usually preferred over the Lagrange form for practical purposes Examples edit The divided differences can be written in the form of a table For example for a function f is to be interpolated on points x 0 x n displaystyle x 0 ldots x n nbsp Write x 0 f x 0 f x 1 f x 0 x 1 x 0 x 1 f x 1 f x 2 f x 1 x 2 x 1 f x 1 f x 0 x 1 x 0 x 2 x 0 f x 2 f x 1 x 2 x 1 x 2 f x 2 x n f x n displaystyle begin matrix x 0 amp f x 0 amp amp amp amp f x 1 f x 0 over x 1 x 0 amp x 1 amp f x 1 amp amp f x 2 f x 1 over x 2 x 1 f x 1 f x 0 over x 1 x 0 over x 2 x 0 amp amp f x 2 f x 1 over x 2 x 1 amp x 2 amp f x 2 amp amp vdots amp amp vdots amp vdots amp amp amp vdots amp amp vdots amp x n amp f x n amp amp end matrix nbsp Then the interpolating polynomial is formed as above using the topmost entries in each column as coefficients For example suppose we are to construct the interpolating polynomial to f x tan x using divided differences at the points n displaystyle n nbsp x n displaystyle x n nbsp f x n displaystyle f x n nbsp 0 displaystyle 0 nbsp 3 2 displaystyle tfrac 3 2 nbsp 14 1014 displaystyle 14 1014 nbsp 1 displaystyle 1 nbsp 3 4 displaystyle tfrac 3 4 nbsp 0 931596 displaystyle 0 931596 nbsp 2 displaystyle 2 nbsp 0 displaystyle 0 nbsp 0 displaystyle 0 nbsp 3 displaystyle 3 nbsp 3 4 displaystyle tfrac 3 4 nbsp 0 931596 displaystyle 0 931596 nbsp 4 displaystyle 4 nbsp 3 2 displaystyle tfrac 3 2 nbsp 14 1014 displaystyle 14 1014 nbsp Using six digits of accuracy we construct the table 3 2 14 1014 17 5597 3 4 0 931596 10 8784 1 24213 4 83484 0 0 0 0 1 24213 4 83484 3 4 0 931596 10 8784 17 5597 3 2 14 1014 displaystyle begin matrix tfrac 3 2 amp 14 1014 amp amp amp amp amp amp 17 5597 amp amp amp tfrac 3 4 amp 0 931596 amp amp 10 8784 amp amp amp amp 1 24213 amp amp 4 83484 amp 0 amp 0 amp amp 0 amp amp 0 amp amp 1 24213 amp amp 4 83484 amp tfrac 3 4 amp 0 931596 amp amp 10 8784 amp amp amp amp 17 5597 amp amp amp tfrac 3 2 amp 14 1014 amp amp amp amp end matrix nbsp Thus the interpolating polynomial is 14 1014 17 5597 x 3 2 10 8784 x 3 2 x 3 4 4 83484 x 3 2 x 3 4 x 0 x 3 2 x 3 4 x x 3 4 0 00005 1 4775 x 0 00001 x 2 4 83484 x 3 displaystyle begin aligned amp 14 1014 17 5597 x tfrac 3 2 10 8784 x tfrac 3 2 x tfrac 3 4 4 83484 x tfrac 3 2 x tfrac 3 4 x 0 x tfrac 3 2 x tfrac 3 4 x x tfrac 3 4 amp 0 00005 1 4775x 0 00001x 2 4 83484x 3 end aligned nbsp Given more digits of accuracy in the table the first and third coefficients will be found to be zero Another example The sequence f 0 displaystyle f 0 nbsp such that f 0 1 6 f 0 2 9 f 0 3 2 displaystyle f 0 1 6 f 0 2 9 f 0 3 2 nbsp and f 0 4 5 displaystyle f 0 4 5 nbsp i e they are 6 9 2 5 displaystyle 6 9 2 5 nbsp from x 0 1 displaystyle x 0 1 nbsp to x 3 4 displaystyle x 3 4 nbsp You obtain the slope of order 1 displaystyle 1 nbsp in the following way f 1 x 0 x 1 f 0 x 1 f 0 x 0 x 1 x 0 9 6 2 1 3 displaystyle f 1 x 0 x 1 frac f 0 x 1 f 0 x 0 x 1 x 0 frac 9 6 2 1 3 nbsp f 1 x 1 x 2 f 0 x 2 f 0 x 1 x 2 x 1 2 9 3 2 7 displaystyle f 1 x 1 x 2 frac f 0 x 2 f 0 x 1 x 2 x 1 frac 2 9 3 2 7 nbsp f 1 x 2 x 3 f 0 x 3 f 0 x 2 x 3 x 2 5 2 4 3 3 displaystyle f 1 x 2 x 3 frac f 0 x 3 f 0 x 2 x 3 x 2 frac 5 2 4 3 3 nbsp As we have the slopes of order 1 displaystyle 1 nbsp it is possible to obtain the next order f 2 x 0 x 1 x 2 f 1 x 1 x 2 f 1 x 0 x 1 x 2 x 0 7 3 3 1 5 displaystyle f 2 x 0 x 1 x 2 frac f 1 x 1 x 2 f 1 x 0 x 1 x 2 x 0 frac 7 3 3 1 5 nbsp f 2 x 1 x 2 x 3 f 1 x 2 x 3 f 1 x 1 x 2 x 3 x 1 3 7 4 2 5 displaystyle f 2 x 1 x 2 x 3 frac f 1 x 2 x 3 f 1 x 1 x 2 x 3 x 1 frac 3 7 4 2 5 nbsp Finally we define the slope of order 3 displaystyle 3 nbsp f 3 x 0 x 1 x 2 x 3 f 2 x 1 x 2 x 3 f 2 x 0 x 1 x 2 x 3 x 0 5 5 4 1 10 3 displaystyle f 3 x 0 x 1 x 2 x 3 frac f 2 x 1 x 2 x 3 f 2 x 0 x 1 x 2 x 3 x 0 frac 5 5 4 1 frac 10 3 nbsp Once we have the slope we can define the consequent polynomials p 0 x 6 displaystyle p 0 x 6 nbsp p 1 x 6 3 x 1 displaystyle p 1 x 6 3 x 1 nbsp p 2 x 6 3 x 1 5 x 1 x 2 displaystyle p 2 x 6 3 x 1 5 x 1 x 2 nbsp p 3 x 6 3 x 1 5 x 1 x 2 10 3 x 1 x 2 x 3 displaystyle p 3 x 6 3 x 1 5 x 1 x 2 frac 10 3 x 1 x 2 x 3 nbsp See also editDe numeris triangularibus et inde de progressionibus arithmeticis Magisteria magna a work by Thomas Harriot describing similar methods for interpolation written 50 years earlier than Newton s work but not published until 2009 Newton series Neville s schema Polynomial interpolation Lagrange form of the interpolation polynomial Bernstein form of the interpolation polynomial Hermite interpolation Carlson s theorem Table of Newtonian seriesReferences edit Dunham William 1990 7 Journey Through Genius The Great Theorems of Mathematics Kanak Agrawal Inc pp 155 183 ISBN 9780140147391 Retrieved 24 October 2019 Epperson James F 2013 An introduction to numerical methods and analysis 2nd ed Hoboken NJ Wiley ISBN 978 1 118 36759 9 Burden Richard L Faires J Douglas 2011 Numerical Analysis 9th ed Cengage Learning p 129 ISBN 9780538733519 Hamming Richard W 1986 Numerical methods for scientists and engineers Unabridged republ of the 2 ed 1973 ed New York Dover ISBN 978 0 486 65241 2 Stetekluh Jeff Algorithm for the Newton Form of the Interpolating Polynomial External links editModule for the Newton Polynomial by John H Mathews Retrieved from https en wikipedia org w index php title Newton polynomial amp oldid 1189644821, wikipedia, wiki, book, books, library,

    article

    , read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.