fbpx
Wikipedia

Woodbury matrix identity

In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury,[1][2] says that the inverse of a rank-k correction of some matrix can be computed by doing a rank-k correction to the inverse of the original matrix. Alternative names for this formula are the matrix inversion lemma, Sherman–Morrison–Woodbury formula or just Woodbury formula. However, the identity appeared in several papers before the Woodbury report.[3][4]

The Woodbury matrix identity is[5]

where A, U, C and V are conformable matrices: A is n×n, C is k×k, U is n×k, and V is k×n. This can be derived using blockwise matrix inversion.

While the identity is primarily used on matrices, it holds in a general ring or in an Ab-category.

The Woodbury matrix identity allows cheap computation of inverses and solutions to linear equations. However, little is known about the numerical stability of the formula. There are no published results concerning its error bounds. Anecdotal evidence [6] suggests that it may diverge even for seemingly benign examples (when both the original and modified matrices are well-conditioned).

Discussion edit

To prove this result, we will start by proving a simpler one. Replacing A and C with the identity matrix I, we obtain another identity which is a bit simpler:

 
To recover the original equation from this reduced identity, set   and  .

This identity itself can be viewed as the combination of two simpler identities. We obtain the first identity from

 
thus,
 
and similarly
 
The second identity is the so-called push-through identity[7]
 
that we obtain from
 
after multiplying by   on the right and by   on the left.

Putting all together,

 
where the first and second equality come from the first and second identity, respectively.

Special cases edit

When   are vectors, the identity reduces to the Sherman–Morrison formula.

In the scalar case, the reduced version is simply

 

Inverse of a sum edit

If n = k and U = V = In is the identity matrix, then

 

Continuing with the merging of the terms of the far right-hand side of the above equation results in Hua's identity

 

Another useful form of the same identity is

 

which, unlike those above, is valid even if   is singular, and has a recursive structure that yields

 
if the spectral radius of   is less than one. That is, if the above sum converges then it is equal to  .

This form can be used in perturbative expansions where B is a perturbation of A.

Variations edit

Binomial inverse theorem edit

If A, B, U, V are matrices of sizes n×n, k×k, n×k, k×n, respectively, then

 

provided A and B + BVA−1UB are nonsingular. Nonsingularity of the latter requires that B−1 exist since it equals B(I + VA−1UB) and the rank of the latter cannot exceed the rank of B.[7]

Since B is invertible, the two B terms flanking the parenthetical quantity inverse in the right-hand side can be replaced with (B−1)−1, which results in the original Woodbury identity.

A variation for when B is singular and possibly even non-square:[7]

 

Formulas also exist for certain cases in which A is singular.[8]

Pseudoinverse with positive semidefinite matrices edit

In general Woodbury's identity is not valid if one or more inverses are replaced by (Moore–Penrose) pseudoinverses. However, if   and   are positive semidefinite, and   (implying that   is itself positive semidefinite), then the following formula provides a generalization:[9][10]

 

where   can be written as   because any positive semidefinite matrix is equal to   for some  .

Derivations edit

Direct proof edit

The formula can be proven by checking that   times its alleged inverse on the right side of the Woodbury identity gives the identity matrix:

 

Alternative proofs edit

Algebraic proof

First consider these useful identities,

 

Now,

 
Derivation via blockwise elimination

Deriving the Woodbury matrix identity is easily done by solving the following block matrix inversion problem

 

Expanding, we can see that the above reduces to

 
which is equivalent to  . Eliminating the first equation, we find that  , which can be substituted into the second to find  . Expanding and rearranging, we have  , or  . Finally, we substitute into our  , and we have  . Thus,
 

We have derived the Woodbury matrix identity.

Derivation from LDU decomposition

We start by the matrix

 
By eliminating the entry under the A (given that A is invertible) we get
 

Likewise, eliminating the entry above C gives

 

Now combining the above two, we get

 

Moving to the right side gives

 
which is the LDU decomposition of the block matrix into an upper triangular, diagonal, and lower triangular matrices.

Now inverting both sides gives

 

We could equally well have done it the other way (provided that C is invertible) i.e.

 

Now again inverting both sides,

 

Now comparing elements (1, 1) of the RHS of (1) and (2) above gives the Woodbury formula

 

Applications edit

This identity is useful in certain numerical computations where A−1 has already been computed and it is desired to compute (A + UCV)−1. With the inverse of A available, it is only necessary to find the inverse of C−1 + VA−1U in order to obtain the result using the right-hand side of the identity. If C has a much smaller dimension than A, this is more efficient than inverting A + UCV directly. A common case is finding the inverse of a low-rank update A + UCV of A (where U only has a few columns and V only a few rows), or finding an approximation of the inverse of the matrix A + B where the matrix B can be approximated by a low-rank matrix UCV, for example using the singular value decomposition.

This is applied, e.g., in the Kalman filter and recursive least squares methods, to replace the parametric solution, requiring inversion of a state vector sized matrix, with a condition equations based solution. In case of the Kalman filter this matrix has the dimensions of the vector of observations, i.e., as small as 1 in case only one new observation is processed at a time. This significantly speeds up the often real time calculations of the filter.

In the case when C is the identity matrix I, the matrix   is known in numerical linear algebra and numerical partial differential equations as the capacitance matrix.[4]

See also edit

Notes edit

  1. ^ Max A. Woodbury, Inverting modified matrices, Memorandum Rept. 42, Statistical Research Group, Princeton University, Princeton, NJ, 1950, 4pp MR38136
  2. ^ Max A. Woodbury, The Stability of Out-Input Matrices. Chicago, Ill., 1949. 5 pp. MR32564
  3. ^ Guttmann, Louis (1946). "Enlargement methods for computing the inverse matrix". Ann. Math. Statist. 17 (3): 336–343. doi:10.1214/aoms/1177730946.
  4. ^ a b Hager, William W. (1989). "Updating the inverse of a matrix". SIAM Review. 31 (2): 221–239. doi:10.1137/1031049. JSTOR 2030425. MR 0997457.
  5. ^ Higham, Nicholas (2002). Accuracy and Stability of Numerical Algorithms (2nd ed.). SIAM. p. 258. ISBN 978-0-89871-521-7. MR 1927606.
  6. ^ "MathOverflow discussion". MathOverflow.
  7. ^ a b c Henderson, H. V.; Searle, S. R. (1981). "On deriving the inverse of a sum of matrices" (PDF). SIAM Review. 23 (1): 53–60. doi:10.1137/1023004. hdl:1813/32749. JSTOR 2029838.
  8. ^ Kurt S. Riedel, "A Sherman–Morrison–Woodbury Identity for Rank Augmenting Matrices with Application to Centering", SIAM Journal on Matrix Analysis and Applications, 13 (1992)659-662, doi:10.1137/0613040 preprint MR1152773
  9. ^ Bernstein, Dennis S. (2018). Scalar, Vector, and Matrix Mathematics: Theory, Facts, and Formulas (Revised and expanded ed.). Princeton: Princeton University Press. p. 638. ISBN 9780691151205.
  10. ^ Schott, James R. (2017). Matrix analysis for statistics (Third ed.). Hoboken, New Jersey: John Wiley & Sons, Inc. p. 219. ISBN 9781119092483.
  • Press, WH; Teukolsky, SA; Vetterling, WT; Flannery, BP (2007), "Section 2.7.3. Woodbury Formula", Numerical Recipes: The Art of Scientific Computing (3rd ed.), New York: Cambridge University Press, ISBN 978-0-521-88068-8

External links edit

woodbury, matrix, identity, mathematics, specifically, linear, algebra, named, after, woodbury, says, that, inverse, rank, correction, some, matrix, computed, doing, rank, correction, inverse, original, matrix, alternative, names, this, formula, matrix, invers. In mathematics specifically linear algebra the Woodbury matrix identity named after Max A Woodbury 1 2 says that the inverse of a rank k correction of some matrix can be computed by doing a rank k correction to the inverse of the original matrix Alternative names for this formula are the matrix inversion lemma Sherman Morrison Woodbury formula or just Woodbury formula However the identity appeared in several papers before the Woodbury report 3 4 The Woodbury matrix identity is 5 A U C V 1 A 1 A 1 U C 1 V A 1 U 1 V A 1 displaystyle left A UCV right 1 A 1 A 1 U left C 1 VA 1 U right 1 VA 1 where A U C and V are conformable matrices A is n n C is k k U is n k and V is k n This can be derived using blockwise matrix inversion While the identity is primarily used on matrices it holds in a general ring or in an Ab category The Woodbury matrix identity allows cheap computation of inverses and solutions to linear equations However little is known about the numerical stability of the formula There are no published results concerning its error bounds Anecdotal evidence 6 suggests that it may diverge even for seemingly benign examples when both the original and modified matrices are well conditioned Contents 1 Discussion 1 1 Special cases 1 1 1 Inverse of a sum 1 2 Variations 1 2 1 Binomial inverse theorem 1 2 2 Pseudoinverse with positive semidefinite matrices 2 Derivations 2 1 Direct proof 2 2 Alternative proofs 3 Applications 4 See also 5 Notes 6 External linksDiscussion editTo prove this result we will start by proving a simpler one Replacing A and C with the identity matrix I we obtain another identity which is a bit simpler I U V 1 I U I V U 1 V displaystyle left I UV right 1 I U left I VU right 1 V nbsp To recover the original equation from this reduced identity set U A 1 X displaystyle U A 1 X nbsp and V C Y displaystyle V CY nbsp This identity itself can be viewed as the combination of two simpler identities We obtain the first identity fromI I P 1 I P I P 1 I P 1 P displaystyle I I P 1 I P I P 1 I P 1 P nbsp thus I P 1 I I P 1 P displaystyle I P 1 I I P 1 P nbsp and similarly I P 1 I P I P 1 displaystyle I P 1 I P I P 1 nbsp The second identity is the so called push through identity 7 I U V 1 U U I V U 1 displaystyle I UV 1 U U I VU 1 nbsp that we obtain from U I V U I U V U displaystyle U I VU I UV U nbsp after multiplying by I V U 1 displaystyle I VU 1 nbsp on the right and by I U V 1 displaystyle I UV 1 nbsp on the left Putting all together I U V 1 I U V I U V 1 I U I V U 1 V displaystyle left I UV right 1 I UV left I UV right 1 I U left I VU right 1 V nbsp where the first and second equality come from the first and second identity respectively Special cases edit When V U displaystyle V U nbsp are vectors the identity reduces to the Sherman Morrison formula In the scalar case the reduced version is simply1 1 u v 1 u v 1 u v displaystyle frac 1 1 uv 1 frac uv 1 uv nbsp Inverse of a sum edit If n k and U V In is the identity matrix then A B 1 A 1 A 1 B 1 A 1 1 A 1 A 1 A 1 A B 1 I 1 displaystyle begin aligned left A B right 1 amp A 1 A 1 left B 1 A 1 right 1 A 1 1ex amp A 1 A 1 left AB 1 I right 1 end aligned nbsp Continuing with the merging of the terms of the far right hand side of the above equation results in Hua s identity A B 1 A 1 A A B 1 A 1 displaystyle left A B right 1 A 1 left A A B 1 A right 1 nbsp Another useful form of the same identity is A B 1 A 1 A 1 B A B 1 displaystyle left A B right 1 A 1 A 1 B left A B right 1 nbsp which unlike those above is valid even if B displaystyle B nbsp is singular and has a recursive structure that yields A B 1 k 0 A 1 B k A 1 displaystyle left A B right 1 sum k 0 infty left A 1 B right k A 1 nbsp if the spectral radius of A 1 B displaystyle A 1 B nbsp is less than one That is if the above sum converges then it is equal to A B 1 displaystyle A B 1 nbsp This form can be used in perturbative expansions where B is a perturbation of A Variations edit Binomial inverse theorem edit If A B U V are matrices of sizes n n k k n k k n respectively then A U B V 1 A 1 A 1 U B B B V A 1 U B 1 B V A 1 displaystyle left A UBV right 1 A 1 A 1 UB left B BVA 1 UB right 1 BVA 1 nbsp provided A and B BVA 1UB are nonsingular Nonsingularity of the latter requires that B 1 exist since it equals B I VA 1UB and the rank of the latter cannot exceed the rank of B 7 Since B is invertible the two B terms flanking the parenthetical quantity inverse in the right hand side can be replaced with B 1 1 which results in the original Woodbury identity A variation for when B is singular and possibly even non square 7 A U B V 1 A 1 A 1 U I B V A 1 U 1 B V A 1 displaystyle A UBV 1 A 1 A 1 U I BVA 1 U 1 BVA 1 nbsp Formulas also exist for certain cases in which A is singular 8 Pseudoinverse with positive semidefinite matrices edit In general Woodbury s identity is not valid if one or more inverses are replaced by Moore Penrose pseudoinverses However if A displaystyle A nbsp and C displaystyle C nbsp are positive semidefinite and V U H displaystyle V U mathrm H nbsp implying that A U C V displaystyle A UCV nbsp is itself positive semidefinite then the following formula provides a generalization 9 10 X X H Y Y H Z Z H I Y Z H X H E X I Y Z Z I X X Y E I X Y I Z Z F 1 X Y H F I I Z Z Y H X X H Y I Z Z displaystyle begin aligned left XX mathrm H YY mathrm H right amp left ZZ mathrm H right left I YZ right mathrm H X mathrm H EX left I YZ right Z amp left I XX right Y E amp I X Y left I Z Z right F 1 left X Y right mathrm H F amp I left I Z Z right Y mathrm H left XX mathrm H right Y left I Z Z right end aligned nbsp where A U C U H displaystyle A UCU mathrm H nbsp can be written as X X H Y Y H displaystyle XX mathrm H YY mathrm H nbsp because any positive semidefinite matrix is equal to M M H displaystyle MM mathrm H nbsp for some M displaystyle M nbsp Derivations editDirect proof edit The formula can be proven by checking that A U C V displaystyle A UCV nbsp times its alleged inverse on the right side of the Woodbury identity gives the identity matrix A U C V A 1 A 1 U C 1 V A 1 U 1 V A 1 I U C 1 V A 1 U 1 V A 1 U C V A 1 U C V A 1 U C 1 V A 1 U 1 V A 1 I U C V A 1 U C 1 V A 1 U 1 V A 1 U C V A 1 U C 1 V A 1 U 1 V A 1 I U C V A 1 U U C V A 1 U C 1 V A 1 U 1 V A 1 I U C V A 1 U C C 1 V A 1 U C 1 V A 1 U 1 V A 1 I U C V A 1 U C V A 1 I displaystyle begin aligned amp left A UCV right left A 1 A 1 U left C 1 VA 1 U right 1 VA 1 right amp left I U left C 1 VA 1 U right 1 VA 1 right left UCVA 1 UCVA 1 U left C 1 VA 1 U right 1 VA 1 right amp left I UCVA 1 right left U left C 1 VA 1 U right 1 VA 1 UCVA 1 U left C 1 VA 1 U right 1 VA 1 right amp I UCVA 1 left U UCVA 1 U right left C 1 VA 1 U right 1 VA 1 amp I UCVA 1 UC left C 1 VA 1 U right left C 1 VA 1 U right 1 VA 1 amp I UCVA 1 UCVA 1 amp I end aligned nbsp Alternative proofs edit Algebraic proof First consider these useful identities U U C V A 1 U U C C 1 V A 1 U A U C V A 1 U A U C V 1 U C A 1 U C 1 V A 1 U 1 displaystyle begin aligned U UCVA 1 U amp UC left C 1 VA 1 U right left A UCV right A 1 U left A UCV right 1 UC amp A 1 U left C 1 VA 1 U right 1 end aligned nbsp Now A 1 A U C V 1 A U C V A 1 A U C V 1 I U C V A 1 A U C V 1 A U C V 1 U C V A 1 A U C V 1 A 1 U C 1 V A 1 U 1 V A 1 displaystyle begin aligned A 1 amp left A UCV right 1 left A UCV right A 1 amp left A UCV right 1 left I UCVA 1 right amp left A UCV right 1 left A UCV right 1 UCVA 1 amp left A UCV right 1 A 1 U left C 1 VA 1 U right 1 VA 1 end aligned nbsp Derivation via blockwise elimination Deriving the Woodbury matrix identity is easily done by solving the following block matrix inversion problem A U V C 1 X Y I 0 displaystyle begin bmatrix A amp U V amp C 1 end bmatrix begin bmatrix X Y end bmatrix begin bmatrix I 0 end bmatrix nbsp Expanding we can see that the above reduces to A X U Y I V X C 1 Y 0 displaystyle begin cases AX UY I VX C 1 Y 0 end cases nbsp which is equivalent to A U C V X I displaystyle A UCV X I nbsp Eliminating the first equation we find that X A 1 I U Y displaystyle X A 1 I UY nbsp which can be substituted into the second to find V A 1 I U Y C 1 Y displaystyle VA 1 I UY C 1 Y nbsp Expanding and rearranging we have V A 1 C 1 V A 1 U Y displaystyle VA 1 left C 1 VA 1 U right Y nbsp or C 1 V A 1 U 1 V A 1 Y displaystyle left C 1 VA 1 U right 1 VA 1 Y nbsp Finally we substitute into our A X U Y I displaystyle AX UY I nbsp and we have A X U C 1 V A 1 U 1 V A 1 I displaystyle AX U left C 1 VA 1 U right 1 VA 1 I nbsp Thus A U C V 1 X A 1 A 1 U C 1 V A 1 U 1 V A 1 displaystyle A UCV 1 X A 1 A 1 U left C 1 VA 1 U right 1 VA 1 nbsp We have derived the Woodbury matrix identity Derivation from LDU decomposition We start by the matrix A U V C displaystyle begin bmatrix A amp U V amp C end bmatrix nbsp By eliminating the entry under the A given that A is invertible we get I 0 V A 1 I A U V C A U 0 C V A 1 U displaystyle begin bmatrix I amp 0 VA 1 amp I end bmatrix begin bmatrix A amp U V amp C end bmatrix begin bmatrix A amp U 0 amp C VA 1 U end bmatrix nbsp Likewise eliminating the entry above C gives A U V C I A 1 U 0 I A 0 V C V A 1 U displaystyle begin bmatrix A amp U V amp C end bmatrix begin bmatrix I amp A 1 U 0 amp I end bmatrix begin bmatrix A amp 0 V amp C VA 1 U end bmatrix nbsp Now combining the above two we get I 0 V A 1 I A U V C I A 1 U 0 I A 0 0 C V A 1 U displaystyle begin bmatrix I amp 0 VA 1 amp I end bmatrix begin bmatrix A amp U V amp C end bmatrix begin bmatrix I amp A 1 U 0 amp I end bmatrix begin bmatrix A amp 0 0 amp C VA 1 U end bmatrix nbsp Moving to the right side gives A U V C I 0 V A 1 I A 0 0 C V A 1 U I A 1 U 0 I displaystyle begin bmatrix A amp U V amp C end bmatrix begin bmatrix I amp 0 VA 1 amp I end bmatrix begin bmatrix A amp 0 0 amp C VA 1 U end bmatrix begin bmatrix I amp A 1 U 0 amp I end bmatrix nbsp which is the LDU decomposition of the block matrix into an upper triangular diagonal and lower triangular matrices Now inverting both sides gives A U V C 1 I A 1 U 0 I 1 A 0 0 C V A 1 U 1 I 0 V A 1 I 1 I A 1 U 0 I A 1 0 0 C V A 1 U 1 I 0 V A 1 I A 1 A 1 U C V A 1 U 1 V A 1 A 1 U C V A 1 U 1 C V A 1 U 1 V A 1 C V A 1 U 1 1 displaystyle begin aligned begin bmatrix A amp U V amp C end bmatrix 1 amp begin bmatrix I amp A 1 U 0 amp I end bmatrix 1 begin bmatrix A amp 0 0 amp C VA 1 U end bmatrix 1 begin bmatrix I amp 0 VA 1 amp I end bmatrix 1 8pt amp begin bmatrix I amp A 1 U 0 amp I end bmatrix begin bmatrix A 1 amp 0 0 amp left C VA 1 U right 1 end bmatrix begin bmatrix I amp 0 VA 1 amp I end bmatrix 8pt amp begin bmatrix A 1 A 1 U left C VA 1 U right 1 VA 1 amp A 1 U left C VA 1 U right 1 left C VA 1 U right 1 VA 1 amp left C VA 1 U right 1 end bmatrix qquad mathrm 1 end aligned nbsp We could equally well have done it the other way provided that C is invertible i e A U V C I U C 1 0 I A U C 1 V 0 0 C I 0 C 1 V I displaystyle begin bmatrix A amp U V amp C end bmatrix begin bmatrix I amp UC 1 0 amp I end bmatrix begin bmatrix A UC 1 V amp 0 0 amp C end bmatrix begin bmatrix I amp 0 C 1 V amp I end bmatrix nbsp Now again inverting both sides A U V C 1 I 0 C 1 V I 1 A U C 1 V 0 0 C 1 I U C 1 0 I 1 I 0 C 1 V I A U C 1 V 1 0 0 C 1 I U C 1 0 I A U C 1 V 1 A U C 1 V 1 U C 1 C 1 V A U C 1 V 1 C 1 C 1 V A U C 1 V 1 U C 1 2 displaystyle begin aligned begin bmatrix A amp U V amp C end bmatrix 1 amp begin bmatrix I amp 0 C 1 V amp I end bmatrix 1 begin bmatrix A UC 1 V amp 0 0 amp C end bmatrix 1 begin bmatrix I amp UC 1 0 amp I end bmatrix 1 8pt amp begin bmatrix I amp 0 C 1 V amp I end bmatrix begin bmatrix left A UC 1 V right 1 amp 0 0 amp C 1 end bmatrix begin bmatrix I amp UC 1 0 amp I end bmatrix 8pt amp begin bmatrix left A UC 1 V right 1 amp left A UC 1 V right 1 UC 1 C 1 V left A UC 1 V right 1 amp C 1 C 1 V left A UC 1 V right 1 UC 1 end bmatrix qquad mathrm 2 end aligned nbsp Now comparing elements 1 1 of the RHS of 1 and 2 above gives the Woodbury formula A U C 1 V 1 A 1 A 1 U C V A 1 U 1 V A 1 displaystyle left A UC 1 V right 1 A 1 A 1 U left C VA 1 U right 1 VA 1 nbsp Applications editThis identity is useful in certain numerical computations where A 1 has already been computed and it is desired to compute A UCV 1 With the inverse of A available it is only necessary to find the inverse of C 1 VA 1U in order to obtain the result using the right hand side of the identity If C has a much smaller dimension than A this is more efficient than inverting A UCV directly A common case is finding the inverse of a low rank update A UCV of A where U only has a few columns and V only a few rows or finding an approximation of the inverse of the matrix A B where the matrix B can be approximated by a low rank matrix UCV for example using the singular value decomposition This is applied e g in the Kalman filter and recursive least squares methods to replace the parametric solution requiring inversion of a state vector sized matrix with a condition equations based solution In case of the Kalman filter this matrix has the dimensions of the vector of observations i e as small as 1 in case only one new observation is processed at a time This significantly speeds up the often real time calculations of the filter In the case when C is the identity matrix I the matrix I V A 1 U displaystyle I VA 1 U nbsp is known in numerical linear algebra and numerical partial differential equations as the capacitance matrix 4 See also editSherman Morrison formula Schur complement Matrix determinant lemma formula for a rank k update to a determinant Invertible matrix Moore Penrose pseudoinverse Updating the pseudoinverseNotes edit Max A Woodbury Inverting modified matrices Memorandum Rept 42 Statistical Research Group Princeton University Princeton NJ 1950 4pp MR38136 Max A Woodbury The Stability of Out Input Matrices Chicago Ill 1949 5 pp MR32564 Guttmann Louis 1946 Enlargement methods for computing the inverse matrix Ann Math Statist 17 3 336 343 doi 10 1214 aoms 1177730946 a b Hager William W 1989 Updating the inverse of a matrix SIAM Review 31 2 221 239 doi 10 1137 1031049 JSTOR 2030425 MR 0997457 Higham Nicholas 2002 Accuracy and Stability of Numerical Algorithms 2nd ed SIAM p 258 ISBN 978 0 89871 521 7 MR 1927606 MathOverflow discussion MathOverflow a b c Henderson H V Searle S R 1981 On deriving the inverse of a sum of matrices PDF SIAM Review 23 1 53 60 doi 10 1137 1023004 hdl 1813 32749 JSTOR 2029838 Kurt S Riedel A Sherman Morrison Woodbury Identity for Rank Augmenting Matrices with Application to Centering SIAM Journal on Matrix Analysis and Applications 13 1992 659 662 doi 10 1137 0613040 preprint MR1152773 Bernstein Dennis S 2018 Scalar Vector and Matrix Mathematics Theory Facts and Formulas Revised and expanded ed Princeton Princeton University Press p 638 ISBN 9780691151205 Schott James R 2017 Matrix analysis for statistics Third ed Hoboken New Jersey John Wiley amp Sons Inc p 219 ISBN 9781119092483 Press WH Teukolsky SA Vetterling WT Flannery BP 2007 Section 2 7 3 Woodbury Formula Numerical Recipes The Art of Scientific Computing 3rd ed New York Cambridge University Press ISBN 978 0 521 88068 8External links editSome matrix identities Weisstein Eric W Woodbury formula MathWorld Retrieved from https en wikipedia org w index php title Woodbury matrix identity amp oldid 1218579540 Binomial inverse theorem, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.