fbpx
Wikipedia

Moore–Penrose inverse

In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix is the most widely known generalization of the inverse matrix.[1][2][3][4] It was independently described by E. H. Moore[5] in 1920, Arne Bjerhammar[6] in 1951, and Roger Penrose[7] in 1955. Earlier, Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse.

A common use of the pseudoinverse is to compute a "best fit" (least squares) solution to a system of linear equations that lacks a solution (see below under § Applications). Another use is to find the minimum (Euclidean) norm solution to a system of linear equations with multiple solutions. The pseudoinverse facilitates the statement and proof of results in linear algebra.

The pseudoinverse is defined and unique for all matrices whose entries are real or complex numbers. It can be computed using the singular value decomposition. In the special case where is a normal matrix (for example, a Hermitian matrix), the pseudoinverse annihilates the kernel of and acts as a traditional inverse of on the subspace orthogonal to the kernel.

Notation

In the following discussion, the following conventions are adopted.

  •   will denote one of the fields of real or complex numbers, denoted  ,  , respectively. The vector space of   matrices over   is denoted by  .
  • For  ,   and   denote the transpose and Hermitian transpose (also called conjugate transpose) respectively. If  , then  .
  • For  ,   (standing for "range") denotes the column space (image) of   (the space spanned by the column vectors of  ) and   denotes the kernel (null space) of  .
  • Finally, for any positive integer  ,   denotes the   identity matrix.

Definition

For  , a pseudoinverse of A is defined as a matrix   satisfying all of the following four criteria, known as the Moore–Penrose conditions:[7][8]

  1.   need not be the general identity matrix, but it maps all column vectors of A to themselves:
     
  2.   acts like a weak inverse:
     
  3.   is Hermitian:
     
  4.   is also Hermitian:
     

  exists for any matrix A , but, when the latter has full rank (that is, the rank of A is  ), then   can be expressed as a simple algebraic formula.

In particular, when   has linearly independent columns (and thus matrix   is invertible),   can be computed as

 

This particular pseudoinverse constitutes a left inverse, since, in this case,  .

When A has linearly independent rows (matrix   is invertible),   can be computed as

 

This is a right inverse, as  .

Properties

Existence and uniqueness

The pseudoinverse exists and is unique: for any matrix  , there is precisely one matrix  , that satisfies the four properties of the definition.[8]

A matrix satisfying the first condition of the definition is known as a generalized inverse. If the matrix also satisfies the second definition, it is called a generalized reflexive inverse. Generalized inverses always exist but are not in general unique. Uniqueness is a consequence of the last two conditions.

Basic properties

Proofs for the properties below can be found at b:Topics in Abstract Algebra/Linear algebra.

  • If   has real entries, then so does  .
  • If   is invertible, its pseudoinverse is its inverse. That is,  .[9]: 243 
  • For any   and any     where   (We use   for "reciprocal".)
  • More generally, for any   and any   the rectangular diagonal matrix
     
    where  
  • It follows that the pseudoinverse of a zero matrix is its transpose.
  • The pseudoinverse of the pseudoinverse is the original matrix:  .[9]: 245 
  • Pseudoinversion commutes with transposition, complex conjugation, and taking the conjugate transpose:[9]: 245 
     
  • The pseudoinverse of a scalar multiple of   is the reciprocal multiple of  :`
     
    for  .

Identities

The following identity formula can be used to cancel or expand certain subexpressions involving pseudoinverses:

 
Equivalently, substituting   for   gives
 
while substituting   for   gives
 

Reduction to Hermitian case

The computation of the pseudoinverse is reducible to its construction in the Hermitian case. This is possible through the equivalences:

 
 

as   and   are Hermitian.

Products

Suppose  . Then the following are equivalent:[10]

  1.  
  2.  
  3.  
  4.  
  5.  

The following are sufficient conditions for  :

  1.   has orthonormal columns (then  ),   or
  2.   has orthonormal rows (then  ),   or
  3.   has linearly independent columns (then   ) and   has linearly independent rows (then  ),   or
  4.  , or
  5.  .

The following is a necessary condition for  :

  1.  

The last sufficient condition yields the equalities

 

NB: The equality   does not hold in general. See the counterexample:

 

Projectors

  and   are orthogonal projection operators, that is, they are Hermitian ( ,  ) and idempotent (  and  ). The following hold:

  •   and  
  •   is the orthogonal projector onto the range of   (which equals the orthogonal complement of the kernel of  ).
  •   is the orthogonal projector onto the range of   (which equals the orthogonal complement of the kernel of  ).
  •   is the orthogonal projector onto the kernel of  .
  •   is the orthogonal projector onto the kernel of  .[8]

The last two properties imply the following identities:

  •  
  •  

Another property is the following: if   is Hermitian and idempotent (true if and only if it represents an orthogonal projection), then, for any matrix   the following equation holds:[11]

 

This can be proven by defining matrices  ,  , and checking that   is indeed a pseudoinverse for   by verifying that the defining properties of the pseudoinverse hold, when   is Hermitian and idempotent.

From the last property it follows that, if   is Hermitian and idempotent, for any matrix  

 

Finally, if   is an orthogonal projection matrix, then its pseudoinverse trivially coincides with the matrix itself, that is,  .

Geometric construction

If we view the matrix as a linear map   over the field   then   can be decomposed as follows. We write   for the direct sum,   for the orthogonal complement,   for the kernel of a map, and   for the image of a map. Notice that   and  . The restriction   is then an isomorphism. This implies that   on   is the inverse of this isomorphism, and is zero on  

In other words: To find   for given   in  , first project   orthogonally onto the range of  , finding a point   in the range. Then form  , that is, find those vectors in   that   sends to  . This will be an affine subspace of   parallel to the kernel of  . The element of this subspace that has the smallest length (that is, is closest to the origin) is the answer   we are looking for. It can be found by taking an arbitrary member of   and projecting it orthogonally onto the orthogonal complement of the kernel of  .

This description is closely related to the minimum-norm solution to a linear system.

Subspaces

 

Limit relations

The pseudoinverse are limits:

 
(see Tikhonov regularization). These limits exist even if   or   do not exist.[8]: 263 

Continuity

In contrast to ordinary matrix inversion, the process of taking pseudoinverses is not continuous: if the sequence   converges to the matrix   (in the maximum norm or Frobenius norm, say), then   need not converge to  . However, if all the matrices   have the same rank as  ,   will converge to  .[12]

Derivative

The derivative of a real valued pseudoinverse matrix which has constant rank at a point   may be calculated in terms of the derivative of the original matrix:[13]

 

Examples

Since for invertible matrices the pseudoinverse equals the usual inverse, only examples of non-invertible matrices are considered below.

  • For   the pseudoinverse is   (Generally, the pseudoinverse of a zero matrix is its transpose.) The uniqueness of this pseudoinverse can be seen from the requirement  , since multiplication by a zero matrix would always produce a zero matrix.
  • For   the pseudoinverse is  

    Indeed,   and thus  

    Similarly,   and thus  
  • For    
  • For     (The denominators are  .)
  • For    
  • For   the pseudoinverse is   For this matrix, the left inverse exists and thus equals  , indeed,  

Special cases

Scalars

It is also possible to define a pseudoinverse for scalars and vectors. This amounts to treating these as matrices. The pseudoinverse of a scalar   is zero if   is zero and the reciprocal of   otherwise:

 

Vectors

The pseudoinverse of the null (all zero) vector is the transposed null vector. The pseudoinverse of a non-null vector is the conjugate transposed vector divided by its squared magnitude:

 

Linearly independent columns

If the rank of   is identical to its column rank,  , (for  ,) there are   linearly independent columns, and   is invertible. In this case, an explicit formula is:[14]

 

It follows that   is then a left inverse of  :    .

Linearly independent rows

If the rank of   is identical to its row rank,  , (for  ,) there are   linearly independent rows, and   is invertible. In this case, an explicit formula is:

 

It follows that   is a right inverse of  :    .

Orthonormal columns or rows

This is a special case of either full column rank or full row rank (treated above). If   has orthonormal columns ( ) or orthonormal rows ( ), then:

 

Normal matrices

If   is normal; that is, it commutes with its conjugate transpose, then its pseudoinverse can be computed by diagonalizing it, mapping all nonzero eigenvalues to their inverses, and mapping zero eigenvalues to zero. A corollary is that   commuting with its transpose implies that it commutes with its pseudoinverse.

Orthogonal projection matrices

This is a special case of a normal matrix with eigenvalues 0 and 1. If   is an orthogonal projection matrix, that is,   and  , then the pseudoinverse trivially coincides with the matrix itself:

 

Circulant matrices

For a circulant matrix  , the singular value decomposition is given by the Fourier transform, that is, the singular values are the Fourier coefficients. Let   be the Discrete Fourier Transform (DFT) matrix; then[15]

 

Construction

Rank decomposition

Let   denote the rank of  . Then   can be (rank) decomposed as   where   and   are of rank  . Then  .

The QR method

For   computing the product   or   and their inverses explicitly is often a source of numerical rounding errors and computational cost in practice. An alternative approach using the QR decomposition of   may be used instead.

Consider the case when   is of full column rank, so that  . Then the Cholesky decomposition  , where   is an upper triangular matrix, may be used. Multiplication by the inverse is then done easily by solving a system with multiple right-hand sides,

 

which may be solved by forward substitution followed by back substitution.

The Cholesky decomposition may be computed without forming   explicitly, by alternatively using the QR decomposition of  , where   has orthonormal columns,  , and   is upper triangular. Then

 

so   is the Cholesky factor of  .

The case of full row rank is treated similarly by using the formula   and using a similar argument, swapping the roles of   and  .

Singular value decomposition (SVD)

A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition.[14][8][16] If   is the singular value decomposition of  , then  . For a rectangular diagonal matrix such as  , we get the pseudoinverse by taking the reciprocal of each non-zero element on the diagonal, leaving the zeros in place, and then transposing the matrix. In numerical computation, only elements larger than some small tolerance are taken to be nonzero, and the others are replaced by zeros. For example, in the MATLAB or GNU Octave function pinv, the tolerance is taken to be t = ε⋅max(m, n)⋅max(Σ), where ε is the machine epsilon.

The computational cost of this method is dominated by the cost of computing the SVD, which is several times higher than matrix–matrix multiplication, even if a state-of-the art implementation (such as that of LAPACK) is used.

The above procedure shows why taking the pseudoinverse is not a continuous operation: if the original matrix   has a singular value 0 (a diagonal entry of the matrix   above), then modifying   slightly may turn this zero into a tiny positive number, thereby affecting the pseudoinverse dramatically as we now have to take the reciprocal of a tiny number.

Block matrices

Optimized approaches exist for calculating the pseudoinverse of block structured matrices.

The iterative method of Ben-Israel and Cohen

Another method for computing the pseudoinverse (cf. Drazin inverse) uses the recursion

 

which is sometimes referred to as hyper-power sequence. This recursion produces a sequence converging quadratically to the pseudoinverse of   if it is started with an appropriate   satisfying  . The choice   (where  , with   denoting the largest singular value of  )[17] has been argued not to be competitive to the method using the SVD mentioned above, because even for moderately ill-conditioned matrices it takes a long time before   enters the region of quadratic convergence.[18] However, if started with   already close to the Moore–Penrose inverse and  , for example  , convergence is fast (quadratic).

Updating the pseudoinverse

For the cases where   has full row or column rank, and the inverse of the correlation matrix (  for   with full row rank or   for full column rank) is already known, the pseudoinverse for matrices related to   can be computed by applying the Sherman–Morrison–Woodbury formula to update the inverse of the correlation matrix, which may need less work. In particular, if the related matrix differs from the original one by only a changed, added or deleted row or column, incremental algorithms exist that exploit the relationship.[19][20]

Similarly, it is possible to update the Cholesky factor when a row or column is added, without creating the inverse of the correlation matrix explicitly. However, updating the pseudoinverse in the general rank-deficient case is much more complicated.[21][22]

Software libraries

High-quality implementations of SVD, QR, and back substitution are available in standard libraries, such as LAPACK. Writing one's own implementation of SVD is a major programming project that requires a significant numerical expertise. In special circumstances, such as parallel computing or embedded computing, however, alternative implementations by QR or even the use of an explicit inverse might be preferable, and custom implementations may be unavoidable.

The Python package NumPy provides a pseudoinverse calculation through its functions matrix.I and linalg.pinv; its pinv uses the SVD-based algorithm. SciPy adds a function scipy.linalg.pinv that uses a least-squares solver.

The MASS package for R provides a calculation of the Moore–Penrose inverse through the ginv function.[23] The ginv function calculates a pseudoinverse using the singular value decomposition provided by the svd function in the base R package. An alternative is to employ the pinv function available in the pracma package.

The Octave programming language provides a pseudoinverse through the standard package function pinv and the pseudo_inverse() method.

In Julia (programming language), the LinearAlgebra package of the standard library provides an implementation of the Moore–Penrose inverse pinv() implemented via singular-value decomposition.[24]

Applications

Linear least-squares

The pseudoinverse provides a least squares solution to a system of linear equations.[25] For  , given a system of linear equations

 

in general, a vector   that solves the system may not exist, or if one does exist, it may not be unique. The pseudoinverse solves the "least-squares" problem as follows:

  •  , we have   where   and   denotes the Euclidean norm. This weak inequality holds with equality if and only if   for any vector  ; this provides an infinitude of minimizing solutions unless   has full column rank, in which case   is a zero matrix.[26] The solution with minimum Euclidean norm is  [26]

This result is easily extended to systems with multiple right-hand sides, when the Euclidean norm is replaced by the Frobenius norm. Let  .

  •  , we have   where   and   denotes the Frobenius norm.

Obtaining all solutions of a linear system

If the linear system

 

has any solutions, they are all given by[27]

 

for arbitrary vector  . Solution(s) exist if and only if  .[27] If the latter holds, then the solution is unique if and only if   has full column rank, in which case   is a zero matrix. If solutions exist but   does not have full column rank, then we have an indeterminate system, all of whose infinitude of solutions are given by this last equation.

Minimum norm solution to a linear system

For linear systems   with non-unique solutions (such as under-determined systems), the pseudoinverse may be used to construct the solution of minimum Euclidean norm   among all solutions.

  • If   is satisfiable, the vector   is a solution, and satisfies   for all solutions.

This result is easily extended to systems with multiple right-hand sides, when the Euclidean norm is replaced by the Frobenius norm. Let  .

  • If   is satisfiable, the matrix   is a solution, and satisfies   for all solutions.

Condition number

Using the pseudoinverse and a matrix norm, one can define a condition number for any matrix:

 

A large condition number implies that the problem of finding least-squares solutions to the corresponding system of linear equations is ill-conditioned in the sense that small errors in the entries of   can lead to huge errors in the entries of the solution.[28]

Generalizations

Besides for matrices over real and complex numbers, the conditions hold for matrices over biquaternions, also called "complex quaternions".[29]

In order to solve more general least-squares problems, one can define Moore–Penrose inverses for all continuous linear operators   between two Hilbert spaces   and  , using the same four conditions as in our definition above. It turns out that not every continuous linear operator has a continuous linear pseudoinverse in this sense.[28] Those that do are precisely the ones whose range is closed in  .

A notion of pseudoinverse exists for matrices over an arbitrary field equipped with an arbitrary involutive automorphism. In this more general setting, a given matrix doesn't always have a pseudoinverse. The necessary and sufficient condition for a pseudoinverse to exist is that  , where   denotes the result of applying the involution operation to the transpose of  . When it does exist, it is unique.[30] Example: Consider the field of complex numbers equipped with the identity involution (as opposed to the involution considered elsewhere in the article); do there exist matrices that fail to have pseudoinverses in this sense? Consider the matrix  . Observe that   while  . So this matrix doesn't have a pseudoinverse in this sense.

In abstract algebra, a Moore–Penrose inverse may be defined on a *-regular semigroup. This abstract definition coincides with the one in linear algebra.

See also

Notes

  1. ^ Ben-Israel & Greville 2003, p. 7.
  2. ^ Campbell & Meyer 1991, p. 10.
  3. ^ Nakamura 1991, p. 42.
  4. ^ Rao & Mitra 1971, p. 50–51.
  5. ^ Moore, E. H. (1920). "On the reciprocal of the general algebraic matrix". Bulletin of the American Mathematical Society. 26 (9): 394–95. doi:10.1090/S0002-9904-1920-03322-7.
  6. ^ Bjerhammar, Arne (1951). "Application of calculus of matrices to method of least squares; with special references to geodetic calculations". Trans. Roy. Inst. Tech. Stockholm. 49.
  7. ^ a b Penrose, Roger (1955). "A generalized inverse for matrices". Proceedings of the Cambridge Philosophical Society. 51 (3): 406–13. Bibcode:1955PCPS...51..406P. doi:10.1017/S0305004100030401.
  8. ^ a b c d e Golub, Gene H.; Charles F. Van Loan (1996). Matrix computations (3rd ed.). Baltimore: Johns Hopkins. pp. 257–258. ISBN 978-0-8018-5414-9.
  9. ^ a b c Stoer, Josef; Bulirsch, Roland (2002). Introduction to Numerical Analysis (3rd ed.). Berlin, New York: Springer-Verlag. ISBN 978-0-387-95452-3..
  10. ^ Greville, T. N. E. (1966-10-01). "Note on the Generalized Inverse of a Matrix Product". SIAM Review. 8 (4): 518–521. Bibcode:1966SIAMR...8..518G. doi:10.1137/1008107. ISSN 0036-1445.
  11. ^ Maciejewski, Anthony A.; Klein, Charles A. (1985). "Obstacle Avoidance for Kinematically Redundant Manipulators in Dynamically Varying Environments". International Journal of Robotics Research. 4 (3): 109–117. doi:10.1177/027836498500400308. hdl:10217/536. S2CID 17660144.
  12. ^ Rakočević, Vladimir (1997). "On continuity of the Moore–Penrose and Drazin inverses" (PDF). Matematički Vesnik. 49: 163–72.
  13. ^ Golub, G. H.; Pereyra, V. (April 1973). "The Differentiation of Pseudo-Inverses and Nonlinear Least Squares Problems Whose Variables Separate". SIAM Journal on Numerical Analysis. 10 (2): 413–32. Bibcode:1973SJNA...10..413G. doi:10.1137/0710036. JSTOR 2156365.
  14. ^ a b Ben-Israel & Greville 2003.
  15. ^ Stallings, W. T.; Boullion, T. L. (1972). "The Pseudoinverse of an r-Circulant Matrix". Proceedings of the American Mathematical Society. 34 (2): 385–88. doi:10.2307/2038377. JSTOR 2038377.
  16. ^ Linear Systems & Pseudo-Inverse
  17. ^ Ben-Israel, Adi; Cohen, Dan (1966). "On Iterative Computation of Generalized Inverses and Associated Projections". SIAM Journal on Numerical Analysis. 3 (3): 410–19. Bibcode:1966SJNA....3..410B. doi:10.1137/0703035. JSTOR 2949637.pdf
  18. ^ Söderström, Torsten; Stewart, G. W. (1974). "On the Numerical Properties of an Iterative Method for Computing the Moore–Penrose Generalized Inverse". SIAM Journal on Numerical Analysis. 11 (1): 61–74. Bibcode:1974SJNA...11...61S. doi:10.1137/0711008. JSTOR 2156431.
  19. ^ Gramß, Tino (1992). Worterkennung mit einem künstlichen neuronalen Netzwerk (PhD dissertation). Georg-August-Universität zu Göttingen. OCLC 841706164.
  20. ^ Emtiyaz, Mohammad (February 27, 2008). "Updating Inverse of a Matrix When a Column is Added/Removed" (PDF).
moore, penrose, inverse, mathematics, particular, linear, algebra, displaystyle, matrix, displaystyle, most, widely, known, generalization, inverse, matrix, independently, described, moore, 1920, arne, bjerhammar, 1951, roger, penrose, 1955, earlier, erik, iva. In mathematics and in particular linear algebra the Moore Penrose inverse A displaystyle A of a matrix A displaystyle A is the most widely known generalization of the inverse matrix 1 2 3 4 It was independently described by E H Moore 5 in 1920 Arne Bjerhammar 6 in 1951 and Roger Penrose 7 in 1955 Earlier Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903 When referring to a matrix the term pseudoinverse without further specification is often used to indicate the Moore Penrose inverse The term generalized inverse is sometimes used as a synonym for pseudoinverse A common use of the pseudoinverse is to compute a best fit least squares solution to a system of linear equations that lacks a solution see below under Applications Another use is to find the minimum Euclidean norm solution to a system of linear equations with multiple solutions The pseudoinverse facilitates the statement and proof of results in linear algebra The pseudoinverse is defined and unique for all matrices whose entries are real or complex numbers It can be computed using the singular value decomposition In the special case where A displaystyle A is a normal matrix for example a Hermitian matrix the pseudoinverse A displaystyle A annihilates the kernel of A displaystyle A and acts as a traditional inverse of A displaystyle A on the subspace orthogonal to the kernel Contents 1 Notation 2 Definition 3 Properties 3 1 Existence and uniqueness 3 2 Basic properties 3 2 1 Identities 3 3 Reduction to Hermitian case 3 4 Products 3 5 Projectors 3 6 Geometric construction 3 7 Subspaces 3 8 Limit relations 3 9 Continuity 3 10 Derivative 4 Examples 5 Special cases 5 1 Scalars 5 2 Vectors 5 3 Linearly independent columns 5 4 Linearly independent rows 5 5 Orthonormal columns or rows 5 6 Normal matrices 5 7 Orthogonal projection matrices 5 8 Circulant matrices 6 Construction 6 1 Rank decomposition 6 2 The QR method 6 3 Singular value decomposition SVD 6 4 Block matrices 6 5 The iterative method of Ben Israel and Cohen 6 6 Updating the pseudoinverse 6 7 Software libraries 7 Applications 7 1 Linear least squares 7 2 Obtaining all solutions of a linear system 7 3 Minimum norm solution to a linear system 7 4 Condition number 8 Generalizations 9 See also 10 Notes 11 References 12 External linksNotation EditIn the following discussion the following conventions are adopted k displaystyle mathbb k will denote one of the fields of real or complex numbers denoted R displaystyle mathbb R C displaystyle mathbb C respectively The vector space of m n displaystyle m times n matrices over k displaystyle mathbb k is denoted by k m n displaystyle mathbb k m times n For A k m n displaystyle A in mathbb k m times n A T displaystyle A textsf T and A displaystyle A denote the transpose and Hermitian transpose also called conjugate transpose respectively If k R displaystyle mathbb k mathbb R then A A T displaystyle A A textsf T For A k m n displaystyle A in mathbb k m times n ran A displaystyle operatorname ran A standing for range denotes the column space image of A displaystyle A the space spanned by the column vectors of A displaystyle A and ker A displaystyle ker A denotes the kernel null space of A displaystyle A Finally for any positive integer n displaystyle n I n k n n displaystyle I n in mathbb k n times n denotes the n n displaystyle n times n identity matrix Definition EditFor A k m n displaystyle A in mathbb k m times n a pseudoinverse of A is defined as a matrix A k n m displaystyle A in mathbb k n times m satisfying all of the following four criteria known as the Moore Penrose conditions 7 8 A A displaystyle AA need not be the general identity matrix but it maps all column vectors of A to themselves A A A A displaystyle AA A A A displaystyle A acts like a weak inverse A A A A displaystyle A AA A A A displaystyle AA is Hermitian A A A A displaystyle left AA right AA A A displaystyle A A is also Hermitian A A A A displaystyle left A A right A A A displaystyle A exists for any matrix A but when the latter has full rank that is the rank of A is min m n displaystyle min m n then A displaystyle A can be expressed as a simple algebraic formula In particular when A displaystyle A has linearly independent columns and thus matrix A A displaystyle A A is invertible A displaystyle A can be computed asA A A 1 A displaystyle A left A A right 1 A This particular pseudoinverse constitutes a left inverse since in this case A A I displaystyle A A I When A has linearly independent rows matrix A A displaystyle AA is invertible A displaystyle A can be computed asA A A A 1 displaystyle A A left AA right 1 This is a right inverse as A A I displaystyle AA I Properties EditExistence and uniqueness Edit The pseudoinverse exists and is unique for any matrix A displaystyle A there is precisely one matrix A displaystyle A that satisfies the four properties of the definition 8 A matrix satisfying the first condition of the definition is known as a generalized inverse If the matrix also satisfies the second definition it is called a generalized reflexive inverse Generalized inverses always exist but are not in general unique Uniqueness is a consequence of the last two conditions Basic properties Edit Proofs for the properties below can be found at b Topics in Abstract Algebra Linear algebra If A displaystyle A has real entries then so does A displaystyle A If A displaystyle A is invertible its pseudoinverse is its inverse That is A A 1 displaystyle A A 1 9 243 For any n Z 0 displaystyle n in mathbb Z geq 0 and any v k n displaystyle v in mathbb k n diag v diag r v displaystyle operatorname diag v operatorname diag r circ v where r x 1 x k 0 0 0 displaystyle r x 1 x in mathbb k setminus 0 cup 0 0 We use r displaystyle r for reciprocal More generally for any m n Z 0 displaystyle m n in mathbb Z geq 0 and any v k min m n displaystyle v in mathbb k operatorname min m n the rectangular diagonal matrix m n a a a min m n 0 a a v a a min m n n m a a a min m n 0 a a r v a a min m n displaystyle begin aligned amp m times n setminus a a a in min m n times 0 cup a a v a a in min m n amp n times m setminus a a a in min m n times 0 cup a a r v a a in min m n end aligned where r x 1 x k 0 0 0 displaystyle r x 1 x in mathbb k setminus 0 cup 0 0 It follows that the pseudoinverse of a zero matrix is its transpose The pseudoinverse of the pseudoinverse is the original matrix A A displaystyle left A right A 9 245 Pseudoinversion commutes with transposition complex conjugation and taking the conjugate transpose 9 245 A T A T A A A A displaystyle left A textsf T right left A right textsf T quad left overline A right overline A quad left A right left A right The pseudoinverse of a scalar multiple of A displaystyle A is the reciprocal multiple of A displaystyle A a A a 1 A displaystyle left alpha A right alpha 1 A for a 0 displaystyle alpha neq 0 Identities Edit The following identity formula can be used to cancel or expand certain subexpressions involving pseudoinverses A A A A A A A displaystyle A A A A A A A Equivalently substituting A displaystyle A for A displaystyle A gives A A A A A A A displaystyle A A A A A A A while substituting A displaystyle A for A displaystyle A gives A A A A A A A displaystyle A A A A A A A Reduction to Hermitian case Edit The computation of the pseudoinverse is reducible to its construction in the Hermitian case This is possible through the equivalences A A A A displaystyle A left A A right A A A A A displaystyle A A left AA right as A A displaystyle A A and A A displaystyle AA are Hermitian Products Edit Suppose A k m n B k n p displaystyle A in mathbb k m times n B in mathbb k n times p Then the following are equivalent 10 A B B A displaystyle AB B A A A B B A B B A B B A A B A A B textstyle begin aligned A ABB A amp BB A BB A AB amp A AB end aligned A A B B A A B B A A B B A A B B displaystyle begin aligned left A ABB right amp A ABB left A ABB right amp A ABB end aligned A A B B A A B B B B A A displaystyle A ABB A ABB BB A A A A B B A B A B B B A A A B A B displaystyle begin aligned A AB amp B AB AB BB A amp A AB AB end aligned The following are sufficient conditions for A B B A displaystyle AB B A A displaystyle A has orthonormal columns then A A A A I n displaystyle A A A A I n or B displaystyle B has orthonormal rows then B B B B I n displaystyle BB BB I n or A displaystyle A has linearly independent columns then A A I displaystyle A A I and B displaystyle B has linearly independent rows then B B I displaystyle BB I or B A displaystyle B A or B A displaystyle B A The following is a necessary condition for A B B A displaystyle AB B A A A B B B B A A displaystyle A A BB BB A A The last sufficient condition yields the equalities A A A A A A A A displaystyle begin aligned left AA right amp A A left A A right amp A A end aligned NB The equality A B B A displaystyle AB B A does not hold in general See the counterexample 1 1 0 0 0 0 1 1 1 1 0 0 1 2 0 1 2 0 1 4 0 1 4 0 0 1 2 0 1 2 1 2 0 1 2 0 0 0 1 1 1 1 0 0 displaystyle Biggl begin pmatrix 1 amp 1 0 amp 0 end pmatrix begin pmatrix 0 amp 0 1 amp 1 end pmatrix Biggr begin pmatrix 1 amp 1 0 amp 0 end pmatrix begin pmatrix tfrac 1 2 amp 0 tfrac 1 2 amp 0 end pmatrix quad neq quad begin pmatrix tfrac 1 4 amp 0 tfrac 1 4 amp 0 end pmatrix begin pmatrix 0 amp tfrac 1 2 0 amp tfrac 1 2 end pmatrix begin pmatrix tfrac 1 2 amp 0 tfrac 1 2 amp 0 end pmatrix begin pmatrix 0 amp 0 1 amp 1 end pmatrix begin pmatrix 1 amp 1 0 amp 0 end pmatrix Projectors Edit P A A displaystyle P AA and Q A A displaystyle Q A A are orthogonal projection operators that is they are Hermitian P P displaystyle P P Q Q displaystyle Q Q and idempotent P 2 P displaystyle P 2 P and Q 2 Q displaystyle Q 2 Q The following hold P A A Q A displaystyle PA AQ A and A P Q A A displaystyle A P QA A P displaystyle P is the orthogonal projector onto the range of A displaystyle A which equals the orthogonal complement of the kernel of A displaystyle A Q displaystyle Q is the orthogonal projector onto the range of A displaystyle A which equals the orthogonal complement of the kernel of A displaystyle A I Q I A A displaystyle I Q I A A is the orthogonal projector onto the kernel of A displaystyle A I P I A A displaystyle I P I AA is the orthogonal projector onto the kernel of A displaystyle A 8 The last two properties imply the following identities A I A A I A A A 0 displaystyle A left I A A right left I AA right A 0 A I A A I A A A 0 displaystyle A left I AA right left I A A right A 0 Another property is the following if A k n n displaystyle A in mathbb k n times n is Hermitian and idempotent true if and only if it represents an orthogonal projection then for any matrix B k m n displaystyle B in mathbb k m times n the following equation holds 11 A B A B A displaystyle A BA BA This can be proven by defining matrices C B A displaystyle C BA D A B A displaystyle D A BA and checking that D displaystyle D is indeed a pseudoinverse for C displaystyle C by verifying that the defining properties of the pseudoinverse hold when A displaystyle A is Hermitian and idempotent From the last property it follows that if A k n n displaystyle A in mathbb k n times n is Hermitian and idempotent for any matrix B k n m displaystyle B in mathbb k n times m A B A A B displaystyle AB A AB Finally if A displaystyle A is an orthogonal projection matrix then its pseudoinverse trivially coincides with the matrix itself that is A A displaystyle A A Geometric construction Edit If we view the matrix as a linear map A k n k m displaystyle A mathbb k n to mathbb k m over the field k displaystyle mathbb k then A k m k n displaystyle A mathbb k m to mathbb k n can be decomposed as follows We write displaystyle oplus for the direct sum displaystyle perp for the orthogonal complement ker displaystyle ker for the kernel of a map and ran displaystyle operatorname ran for the image of a map Notice that k n ker A ker A displaystyle mathbb k n left ker A right perp oplus ker A and k m ran A ran A displaystyle mathbb k m operatorname ran A oplus left operatorname ran A right perp The restriction A ker A ran A displaystyle A left ker A right perp to operatorname ran A is then an isomorphism This implies that A displaystyle A on ran A displaystyle operatorname ran A is the inverse of this isomorphism and is zero on ran A displaystyle left operatorname ran A right perp In other words To find A b displaystyle A b for given b displaystyle b in k m displaystyle mathbb k m first project b displaystyle b orthogonally onto the range of A displaystyle A finding a point p b displaystyle p b in the range Then form A 1 p b displaystyle A 1 p b that is find those vectors in k n displaystyle mathbb k n that A displaystyle A sends to p b displaystyle p b This will be an affine subspace of k n displaystyle mathbb k n parallel to the kernel of A displaystyle A The element of this subspace that has the smallest length that is is closest to the origin is the answer A b displaystyle A b we are looking for It can be found by taking an arbitrary member of A 1 p b displaystyle A 1 p b and projecting it orthogonally onto the orthogonal complement of the kernel of A displaystyle A This description is closely related to the minimum norm solution to a linear system Subspaces Edit ker A ker A ran A ran A displaystyle begin aligned ker left A right amp ker left A right operatorname ran left A right amp operatorname ran left A right end aligned Limit relations Edit The pseudoinverse are limits A lim d 0 A A d I 1 A lim d 0 A A A d I 1 displaystyle A lim delta searrow 0 left A A delta I right 1 A lim delta searrow 0 A left AA delta I right 1 see Tikhonov regularization These limits exist even if A A 1 displaystyle left AA right 1 or A A 1 displaystyle left A A right 1 do not exist 8 263 Continuity Edit In contrast to ordinary matrix inversion the process of taking pseudoinverses is not continuous if the sequence A n displaystyle left A n right converges to the matrix A displaystyle A in the maximum norm or Frobenius norm say then A n displaystyle A n need not converge to A displaystyle A However if all the matrices A n displaystyle A n have the same rank as A displaystyle A A n displaystyle A n will converge to A displaystyle A 12 Derivative Edit The derivative of a real valued pseudoinverse matrix which has constant rank at a point x displaystyle x may be calculated in terms of the derivative of the original matrix 13 d d x A x A d d x A A A A T d d x A T I A A I A A d d x A T A T A displaystyle frac mathrm d mathrm d x A x A left frac mathrm d mathrm d x A right A A A textsf T left frac mathrm d mathrm d x A textsf T right left I AA right left I A A right left frac text d text d x A textsf T right A textsf T A Examples EditSince for invertible matrices the pseudoinverse equals the usual inverse only examples of non invertible matrices are considered below For A 0 0 0 0 displaystyle A begin pmatrix 0 amp 0 0 amp 0 end pmatrix the pseudoinverse is A 0 0 0 0 displaystyle A begin pmatrix 0 amp 0 0 amp 0 end pmatrix Generally the pseudoinverse of a zero matrix is its transpose The uniqueness of this pseudoinverse can be seen from the requirement A A A A displaystyle A A AA since multiplication by a zero matrix would always produce a zero matrix For A 1 0 1 0 displaystyle A begin pmatrix 1 amp 0 1 amp 0 end pmatrix the pseudoinverse is A 1 2 1 2 0 0 displaystyle A begin pmatrix frac 1 2 amp frac 1 2 0 amp 0 end pmatrix Indeed A A 1 2 1 2 1 2 1 2 displaystyle A A begin pmatrix frac 1 2 amp frac 1 2 frac 1 2 amp frac 1 2 end pmatrix and thus A A A 1 0 1 0 A displaystyle A A A begin pmatrix 1 amp 0 1 amp 0 end pmatrix A Similarly A A 1 0 0 0 displaystyle A A begin pmatrix 1 amp 0 0 amp 0 end pmatrix and thus A A A 1 2 1 2 0 0 A displaystyle A A A begin pmatrix frac 1 2 amp frac 1 2 0 amp 0 end pmatrix A For A 1 0 1 0 displaystyle A begin pmatrix 1 amp 0 1 amp 0 end pmatrix A 1 2 1 2 0 0 displaystyle A begin pmatrix frac 1 2 amp frac 1 2 0 amp 0 end pmatrix For A 1 0 2 0 displaystyle A begin pmatrix 1 amp 0 2 amp 0 end pmatrix A 1 5 2 5 0 0 displaystyle A begin pmatrix frac 1 5 amp frac 2 5 0 amp 0 end pmatrix The denominators are 5 1 2 2 2 displaystyle 5 1 2 2 2 For A 1 1 1 1 displaystyle A begin pmatrix 1 amp 1 1 amp 1 end pmatrix A 1 4 1 4 1 4 1 4 displaystyle A begin pmatrix frac 1 4 amp frac 1 4 frac 1 4 amp frac 1 4 end pmatrix For A 1 0 0 1 0 1 displaystyle A begin pmatrix 1 amp 0 0 amp 1 0 amp 1 end pmatrix the pseudoinverse is A 1 0 0 0 1 2 1 2 displaystyle A begin pmatrix 1 amp 0 amp 0 0 amp frac 1 2 amp frac 1 2 end pmatrix For this matrix the left inverse exists and thus equals A displaystyle A indeed A A 1 0 0 1 displaystyle A A begin pmatrix 1 amp 0 0 amp 1 end pmatrix Special cases EditScalars Edit It is also possible to define a pseudoinverse for scalars and vectors This amounts to treating these as matrices The pseudoinverse of a scalar x displaystyle x is zero if x displaystyle x is zero and the reciprocal of x displaystyle x otherwise x 0 if x 0 x 1 otherwise displaystyle x begin cases 0 amp mbox if x 0 x 1 amp mbox otherwise end cases Vectors Edit The pseudoinverse of the null all zero vector is the transposed null vector The pseudoinverse of a non null vector is the conjugate transposed vector divided by its squared magnitude x 0 T if x 0 x x x otherwise displaystyle vec x begin cases vec 0 textsf T amp text if vec x vec 0 dfrac vec x vec x vec x amp text otherwise end cases Linearly independent columns Edit If the rank of A displaystyle A is identical to its column rank n displaystyle n for n m displaystyle n leq m there are n displaystyle n linearly independent columns and A A displaystyle A A is invertible In this case an explicit formula is 14 A A A 1 A displaystyle A left A A right 1 A It follows that A displaystyle A is then a left inverse of A displaystyle A A A I n displaystyle A A I n Linearly independent rows Edit If the rank of A displaystyle A is identical to its row rank m displaystyle m for m n displaystyle m leq n there are m displaystyle m linearly independent rows and A A displaystyle AA is invertible In this case an explicit formula is A A A A 1 displaystyle A A left AA right 1 It follows that A displaystyle A is a right inverse of A displaystyle A A A I m displaystyle AA I m Orthonormal columns or rows Edit This is a special case of either full column rank or full row rank treated above If A displaystyle A has orthonormal columns A A I n displaystyle A A I n or orthonormal rows A A I m displaystyle AA I m then A A displaystyle A A Normal matrices Edit If A displaystyle A is normal that is it commutes with its conjugate transpose then its pseudoinverse can be computed by diagonalizing it mapping all nonzero eigenvalues to their inverses and mapping zero eigenvalues to zero A corollary is that A displaystyle A commuting with its transpose implies that it commutes with its pseudoinverse Orthogonal projection matrices Edit This is a special case of a normal matrix with eigenvalues 0 and 1 If A displaystyle A is an orthogonal projection matrix that is A A displaystyle A A and A 2 A displaystyle A 2 A then the pseudoinverse trivially coincides with the matrix itself A A displaystyle A A Circulant matrices Edit For a circulant matrix C displaystyle C the singular value decomposition is given by the Fourier transform that is the singular values are the Fourier coefficients Let F displaystyle mathcal F be the Discrete Fourier Transform DFT matrix then 15 C F S F C F S F displaystyle begin aligned C amp mathcal F cdot Sigma cdot mathcal F C amp mathcal F cdot Sigma cdot mathcal F end aligned Construction EditRank decomposition Edit Let r min m n displaystyle r leq min m n denote the rank of A k m n displaystyle A in mathbb k m times n Then A displaystyle A can be rank decomposed as A B C displaystyle A BC where B k m r displaystyle B in mathbb k m times r and C k r n displaystyle C in mathbb k r times n are of rank r displaystyle r Then A C B C C C 1 B B 1 B displaystyle A C B C left CC right 1 left B B right 1 B The QR method Edit For k R C displaystyle mathbb k in mathbb R mathbb C computing the product A A displaystyle AA or A A displaystyle A A and their inverses explicitly is often a source of numerical rounding errors and computational cost in practice An alternative approach using the QR decomposition of A displaystyle A may be used instead Consider the case when A displaystyle A is of full column rank so that A A A 1 A displaystyle A left A A right 1 A Then the Cholesky decomposition A A R R displaystyle A A R R where R displaystyle R is an upper triangular matrix may be used Multiplication by the inverse is then done easily by solving a system with multiple right hand sides A A A 1 A A A A A R R A A displaystyle A left A A right 1 A quad Leftrightarrow quad left A A right A A quad Leftrightarrow quad R RA A which may be solved by forward substitution followed by back substitution The Cholesky decomposition may be computed without forming A A displaystyle A A explicitly by alternatively using the QR decomposition of A Q R displaystyle A QR where Q displaystyle Q has orthonormal columns Q Q I displaystyle Q Q I and R displaystyle R is upper triangular ThenA A Q R Q R R Q Q R R R displaystyle A A QR QR R Q QR R R so R displaystyle R is the Cholesky factor of A A displaystyle A A The case of full row rank is treated similarly by using the formula A A A A 1 displaystyle A A left AA right 1 and using a similar argument swapping the roles of A displaystyle A and A displaystyle A Singular value decomposition SVD Edit A computationally simple and accurate way to compute the pseudoinverse is by using the singular value decomposition 14 8 16 If A U S V displaystyle A U Sigma V is the singular value decomposition of A displaystyle A then A V S U displaystyle A V Sigma U For a rectangular diagonal matrix such as S displaystyle Sigma we get the pseudoinverse by taking the reciprocal of each non zero element on the diagonal leaving the zeros in place and then transposing the matrix In numerical computation only elements larger than some small tolerance are taken to be nonzero and the others are replaced by zeros For example in the MATLAB or GNU Octave function pinv the tolerance is taken to be t e max m n max S where e is the machine epsilon The computational cost of this method is dominated by the cost of computing the SVD which is several times higher than matrix matrix multiplication even if a state of the art implementation such as that of LAPACK is used The above procedure shows why taking the pseudoinverse is not a continuous operation if the original matrix A displaystyle A has a singular value 0 a diagonal entry of the matrix S displaystyle Sigma above then modifying A displaystyle A slightly may turn this zero into a tiny positive number thereby affecting the pseudoinverse dramatically as we now have to take the reciprocal of a tiny number Block matrices Edit Optimized approaches exist for calculating the pseudoinverse of block structured matrices The iterative method of Ben Israel and Cohen Edit Another method for computing the pseudoinverse cf Drazin inverse uses the recursionA i 1 2 A i A i A A i displaystyle A i 1 2A i A i AA i which is sometimes referred to as hyper power sequence This recursion produces a sequence converging quadratically to the pseudoinverse of A displaystyle A if it is started with an appropriate A 0 displaystyle A 0 satisfying A 0 A A 0 A displaystyle A 0 A left A 0 A right The choice A 0 a A displaystyle A 0 alpha A where 0 lt a lt 2 s 1 2 A displaystyle 0 lt alpha lt 2 sigma 1 2 A with s 1 A displaystyle sigma 1 A denoting the largest singular value of A displaystyle A 17 has been argued not to be competitive to the method using the SVD mentioned above because even for moderately ill conditioned matrices it takes a long time before A i displaystyle A i enters the region of quadratic convergence 18 However if started with A 0 displaystyle A 0 already close to the Moore Penrose inverse and A 0 A A 0 A displaystyle A 0 A left A 0 A right for example A 0 A A d I 1 A displaystyle A 0 left A A delta I right 1 A convergence is fast quadratic Updating the pseudoinverse Edit For the cases where A displaystyle A has full row or column rank and the inverse of the correlation matrix A A displaystyle AA for A displaystyle A with full row rank or A A displaystyle A A for full column rank is already known the pseudoinverse for matrices related to A displaystyle A can be computed by applying the Sherman Morrison Woodbury formula to update the inverse of the correlation matrix which may need less work In particular if the related matrix differs from the original one by only a changed added or deleted row or column incremental algorithms exist that exploit the relationship 19 20 Similarly it is possible to update the Cholesky factor when a row or column is added without creating the inverse of the correlation matrix explicitly However updating the pseudoinverse in the general rank deficient case is much more complicated 21 22 Software libraries Edit High quality implementations of SVD QR and back substitution are available in standard libraries such as LAPACK Writing one s own implementation of SVD is a major programming project that requires a significant numerical expertise In special circumstances such as parallel computing or embedded computing however alternative implementations by QR or even the use of an explicit inverse might be preferable and custom implementations may be unavoidable The Python package NumPy provides a pseudoinverse calculation through its functions matrix I and linalg pinv its pinv uses the SVD based algorithm SciPy adds a function scipy linalg pinv that uses a least squares solver The MASS package for R provides a calculation of the Moore Penrose inverse through the ginv function 23 The ginv function calculates a pseudoinverse using the singular value decomposition provided by the svd function in the base R package An alternative is to employ the pinv function available in the pracma package The Octave programming language provides a pseudoinverse through the standard package function pinv and the pseudo inverse method In Julia programming language the LinearAlgebra package of the standard library provides an implementation of the Moore Penrose inverse pinv implemented via singular value decomposition 24 Applications EditLinear least squares Edit See also Linear least squares mathematics The pseudoinverse provides a least squares solution to a system of linear equations 25 For A k m n displaystyle A in mathbb k m times n given a system of linear equationsA x b displaystyle Ax b in general a vector x displaystyle x that solves the system may not exist or if one does exist it may not be unique The pseudoinverse solves the least squares problem as follows x k n displaystyle forall x in mathbb k n we have A x b 2 A z b 2 displaystyle left Ax b right 2 geq left Az b right 2 where z A b displaystyle z A b and 2 displaystyle cdot 2 denotes the Euclidean norm This weak inequality holds with equality if and only if x A b I A A w displaystyle x A b left I A A right w for any vector w displaystyle w this provides an infinitude of minimizing solutions unless A displaystyle A has full column rank in which case I A A displaystyle left I A A right is a zero matrix 26 The solution with minimum Euclidean norm is z displaystyle z 26 This result is easily extended to systems with multiple right hand sides when the Euclidean norm is replaced by the Frobenius norm Let B k m p displaystyle B in mathbb k m times p X k n p displaystyle forall X in mathbb k n times p we have A X B F A Z B F displaystyle AX B mathrm F geq AZ B mathrm F where Z A B displaystyle Z A B and F displaystyle cdot mathrm F denotes the Frobenius norm Obtaining all solutions of a linear system Edit If the linear systemA x b displaystyle Ax b has any solutions they are all given by 27 x A b I A A w displaystyle x A b left I A A right w for arbitrary vector w displaystyle w Solution s exist if and only if A A b b displaystyle AA b b 27 If the latter holds then the solution is unique if and only if A displaystyle A has full column rank in which case I A A displaystyle I A A is a zero matrix If solutions exist but A displaystyle A does not have full column rank then we have an indeterminate system all of whose infinitude of solutions are given by this last equation Minimum norm solution to a linear system Edit For linear systems A x b displaystyle Ax b with non unique solutions such as under determined systems the pseudoinverse may be used to construct the solution of minimum Euclidean norm x 2 displaystyle x 2 among all solutions If A x b displaystyle Ax b is satisfiable the vector z A b displaystyle z A b is a solution and satisfies z 2 x 2 displaystyle z 2 leq x 2 for all solutions This result is easily extended to systems with multiple right hand sides when the Euclidean norm is replaced by the Frobenius norm Let B k m p displaystyle B in mathbb k m times p If A X B displaystyle AX B is satisfiable the matrix Z A B displaystyle Z A B is a solution and satisfies Z F X F displaystyle Z mathrm F leq X mathrm F for all solutions Condition number Edit Using the pseudoinverse and a matrix norm one can define a condition number for any matrix cond A A A displaystyle mbox cond A A left A right A large condition number implies that the problem of finding least squares solutions to the corresponding system of linear equations is ill conditioned in the sense that small errors in the entries of A displaystyle A can lead to huge errors in the entries of the solution 28 Generalizations EditBesides for matrices over real and complex numbers the conditions hold for matrices over biquaternions also called complex quaternions 29 In order to solve more general least squares problems one can define Moore Penrose inverses for all continuous linear operators A H 1 H 2 displaystyle A H 1 rightarrow H 2 between two Hilbert spaces H 1 displaystyle H 1 and H 2 displaystyle H 2 using the same four conditions as in our definition above It turns out that not every continuous linear operator has a continuous linear pseudoinverse in this sense 28 Those that do are precisely the ones whose range is closed in H 2 displaystyle H 2 A notion of pseudoinverse exists for matrices over an arbitrary field equipped with an arbitrary involutive automorphism In this more general setting a given matrix doesn t always have a pseudoinverse The necessary and sufficient condition for a pseudoinverse to exist is that rank A rank A A rank A A displaystyle operatorname rank A operatorname rank left A A right operatorname rank left AA right where A displaystyle A denotes the result of applying the involution operation to the transpose of A displaystyle A When it does exist it is unique 30 Example Consider the field of complex numbers equipped with the identity involution as opposed to the involution considered elsewhere in the article do there exist matrices that fail to have pseudoinverses in this sense Consider the matrix A 1 i T displaystyle A begin bmatrix 1 amp i end bmatrix textsf T Observe that rank A A T 1 displaystyle operatorname rank left AA textsf T right 1 while rank A T A 0 displaystyle operatorname rank left A textsf T A right 0 So this matrix doesn t have a pseudoinverse in this sense In abstract algebra a Moore Penrose inverse may be defined on a regular semigroup This abstract definition coincides with the one in linear algebra See also EditDrazin inverse Hat matrix Inverse element Linear least squares mathematics Pseudo determinant Von Neumann regular ringNotes Edit Ben Israel amp Greville 2003 p 7 Campbell amp Meyer 1991 p 10 Nakamura 1991 p 42 Rao amp Mitra 1971 p 50 51 Moore E H 1920 On the reciprocal of the general algebraic matrix Bulletin of the American Mathematical Society 26 9 394 95 doi 10 1090 S0002 9904 1920 03322 7 Bjerhammar Arne 1951 Application of calculus of matrices to method of least squares with special references to geodetic calculations Trans Roy Inst Tech Stockholm 49 a b Penrose Roger 1955 A generalized inverse for matrices Proceedings of the Cambridge Philosophical Society 51 3 406 13 Bibcode 1955PCPS 51 406P doi 10 1017 S0305004100030401 a b c d e Golub Gene H Charles F Van Loan 1996 Matrix computations 3rd ed Baltimore Johns Hopkins pp 257 258 ISBN 978 0 8018 5414 9 a b c Stoer Josef Bulirsch Roland 2002 Introduction to Numerical Analysis 3rd ed Berlin New York Springer Verlag ISBN 978 0 387 95452 3 Greville T N E 1966 10 01 Note on the Generalized Inverse of a Matrix Product SIAM Review 8 4 518 521 Bibcode 1966SIAMR 8 518G doi 10 1137 1008107 ISSN 0036 1445 Maciejewski Anthony A Klein Charles A 1985 Obstacle Avoidance for Kinematically Redundant Manipulators in Dynamically Varying Environments International Journal of Robotics Research 4 3 109 117 doi 10 1177 027836498500400308 hdl 10217 536 S2CID 17660144 Rakocevic Vladimir 1997 On continuity of the Moore Penrose and Drazin inverses PDF Matematicki Vesnik 49 163 72 Golub G H Pereyra V April 1973 The Differentiation of Pseudo Inverses and Nonlinear Least Squares Problems Whose Variables Separate SIAM Journal on Numerical Analysis 10 2 413 32 Bibcode 1973SJNA 10 413G doi 10 1137 0710036 JSTOR 2156365 a b Ben Israel amp Greville 2003 Stallings W T Boullion T L 1972 The Pseudoinverse of an r Circulant Matrix Proceedings of the American Mathematical Society 34 2 385 88 doi 10 2307 2038377 JSTOR 2038377 Linear Systems amp Pseudo Inverse Ben Israel Adi Cohen Dan 1966 On Iterative Computation of Generalized Inverses and Associated Projections SIAM Journal on Numerical Analysis 3 3 410 19 Bibcode 1966SJNA 3 410B doi 10 1137 0703035 JSTOR 2949637 pdf Soderstrom Torsten Stewart G W 1974 On the Numerical Properties of an Iterative Method for Computing the Moore Penrose Generalized Inverse SIAM Journal on Numerical Analysis 11 1 61 74 Bibcode 1974SJNA 11 61S doi 10 1137 0711008 JSTOR 2156431 Gramss Tino 1992 Worterkennung mit einem kunstlichen neuronalen Netzwerk PhD dissertation Georg August Universitat zu Gottingen OCLC 841706164 Emtiyaz Mohammad February 27 2008 Updating Inverse of a Matrix When a Column is Added Removed PDF span, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.