fbpx
Wikipedia

Rank (linear algebra)

In linear algebra, the rank of a matrix A is the dimension of the vector space generated (or spanned) by its columns.[1][2][3] This corresponds to the maximal number of linearly independent columns of A. This, in turn, is identical to the dimension of the vector space spanned by its rows.[4] Rank is thus a measure of the "nondegenerateness" of the system of linear equations and linear transformation encoded by A. There are multiple equivalent definitions of rank. A matrix's rank is one of its most fundamental characteristics.

The rank is commonly denoted by rank(A) or rk(A);[2] sometimes the parentheses are not written, as in rank A.[i]

Main definitions edit

In this section, we give some definitions of the rank of a matrix. Many definitions are possible; see Alternative definitions for several of these.

The column rank of A is the dimension of the column space of A, while the row rank of A is the dimension of the row space of A.

A fundamental result in linear algebra is that the column rank and the row rank are always equal. (Three proofs of this result are given in § Proofs that column rank = row rank, below.) This number (i.e., the number of linearly independent rows or columns) is simply called the rank of A.

A matrix is said to have full rank if its rank equals the largest possible for a matrix of the same dimensions, which is the lesser of the number of rows and columns. A matrix is said to be rank-deficient if it does not have full rank. The rank deficiency of a matrix is the difference between the lesser of the number of rows and columns, and the rank.

The rank of a linear map or operator   is defined as the dimension of its image:[5][6][7][8]

 
where   is the dimension of a vector space, and   is the image of a map.

Examples edit

The matrix

 
has rank 2: the first two columns are linearly independent, so the rank is at least 2, but since the third is a linear combination of the first two (the first column plus the second), the three columns are linearly dependent so the rank must be less than 3.

The matrix

 
has rank 1: there are nonzero columns, so the rank is positive, but any pair of columns is linearly dependent. Similarly, the transpose
 
of A has rank 1. Indeed, since the column vectors of A are the row vectors of the transpose of A, the statement that the column rank of a matrix equals its row rank is equivalent to the statement that the rank of a matrix is equal to the rank of its transpose, i.e., rank(A) = rank(AT).

Computing the rank of a matrix edit

Rank from row echelon forms edit

A common approach to finding the rank of a matrix is to reduce it to a simpler form, generally row echelon form, by elementary row operations. Row operations do not change the row space (hence do not change the row rank), and, being invertible, map the column space to an isomorphic space (hence do not change the column rank). Once in row echelon form, the rank is clearly the same for both row rank and column rank, and equals the number of pivots (or basic columns) and also the number of non-zero rows.

For example, the matrix A given by

 
can be put in reduced row-echelon form by using the following elementary row operations:
 
The final matrix (in reduced row echelon form) has two non-zero rows and thus the rank of matrix A is 2.

Computation edit

When applied to floating point computations on computers, basic Gaussian elimination (LU decomposition) can be unreliable, and a rank-revealing decomposition should be used instead. An effective alternative is the singular value decomposition (SVD), but there are other less computationally expensive choices, such as QR decomposition with pivoting (so-called rank-revealing QR factorization), which are still more numerically robust than Gaussian elimination. Numerical determination of rank requires a criterion for deciding when a value, such as a singular value from the SVD, should be treated as zero, a practical choice which depends on both the matrix and the application.

Proofs that column rank = row rank edit

Proof using row reduction edit

The fact that the column and row ranks of any matrix are equal forms is fundamental in linear algebra. Many proofs have been given. One of the most elementary ones has been sketched in § Rank from row echelon forms. Here is a variant of this proof:

It is straightforward to show that neither the row rank nor the column rank are changed by an elementary row operation. As Gaussian elimination proceeds by elementary row operations, the reduced row echelon form of a matrix has the same row rank and the same column rank as the original matrix. Further elementary column operations allow putting the matrix in the form of an identity matrix possibly bordered by rows and columns of zeros. Again, this changes neither the row rank nor the column rank. It is immediate that both the row and column ranks of this resulting matrix is the number of its nonzero entries.

We present two other proofs of this result. The first uses only basic properties of linear combinations of vectors, and is valid over any field. The proof is based upon Wardlaw (2005).[9] The second uses orthogonality and is valid for matrices over the real numbers; it is based upon Mackiw (1995).[4] Both proofs can be found in the book by Banerjee and Roy (2014).[10]

Proof using linear combinations edit

Let A be an m × n matrix. Let the column rank of A be r, and let c1, ..., cr be any basis for the column space of A. Place these as the columns of an m × r matrix C. Every column of A can be expressed as a linear combination of the r columns in C. This means that there is an r × n matrix R such that A = CR. R is the matrix whose ith column is formed from the coefficients giving the ith column of A as a linear combination of the r columns of C. In other words, R is the matrix which contains the multiples for the bases of the column space of A (which is C), which are then used to form A as a whole. Now, each row of A is given by a linear combination of the r rows of R. Therefore, the rows of R form a spanning set of the row space of A and, by the Steinitz exchange lemma, the row rank of A cannot exceed r. This proves that the row rank of A is less than or equal to the column rank of A. This result can be applied to any matrix, so apply the result to the transpose of A. Since the row rank of the transpose of A is the column rank of A and the column rank of the transpose of A is the row rank of A, this establishes the reverse inequality and we obtain the equality of the row rank and the column rank of A. (Also see Rank factorization.)

Proof using orthogonality edit

Let A be an m × n matrix with entries in the real numbers whose row rank is r. Therefore, the dimension of the row space of A is r. Let x1, x2, …, xr be a basis of the row space of A. We claim that the vectors Ax1, Ax2, …, Axr are linearly independent. To see why, consider a linear homogeneous relation involving these vectors with scalar coefficients c1, c2, …, cr:

 
where v = c1x1 + c2x2 + ⋯ + crxr. We make two observations: (a) v is a linear combination of vectors in the row space of A, which implies that v belongs to the row space of A, and (b) since Av = 0, the vector v is orthogonal to every row vector of A and, hence, is orthogonal to every vector in the row space of A. The facts (a) and (b) together imply that v is orthogonal to itself, which proves that v = 0 or, by the definition of v,
 
But recall that the xi were chosen as a basis of the row space of A and so are linearly independent. This implies that c1 = c2 = ⋯ = cr = 0. It follows that Ax1, Ax2, …, Axr are linearly independent.

Now, each Axi is obviously a vector in the column space of A. So, Ax1, Ax2, …, Axr is a set of r linearly independent vectors in the column space of A and, hence, the dimension of the column space of A (i.e., the column rank of A) must be at least as big as r. This proves that row rank of A is no larger than the column rank of A. Now apply this result to the transpose of A to get the reverse inequality and conclude as in the previous proof.

Alternative definitions edit

In all the definitions in this section, the matrix A is taken to be an m × n matrix over an arbitrary field F.

Dimension of image edit

Given the matrix  , there is an associated linear mapping

 
defined by
 
The rank of   is the dimension of the image of  . This definition has the advantage that it can be applied to any linear map without need for a specific matrix.

Rank in terms of nullity edit

Given the same linear mapping f as above, the rank is n minus the dimension of the kernel of f. The rank–nullity theorem states that this definition is equivalent to the preceding one.

Column rank – dimension of column space edit

The rank of A is the maximal number of linearly independent columns   of A; this is the dimension of the column space of A (the column space being the subspace of Fm generated by the columns of A, which is in fact just the image of the linear map f associated to A).

Row rank – dimension of row space edit

The rank of A is the maximal number of linearly independent rows of A; this is the dimension of the row space of A.

Decomposition rank edit

The rank of A is the smallest integer k such that A can be factored as  , where C is an m × k matrix and R is a k × n matrix. In fact, for all integers k, the following are equivalent:

  1. the column rank of A is less than or equal to k,
  2. there exist k columns   of size m such that every column of A is a linear combination of  ,
  3. there exist an   matrix C and a   matrix R such that   (when k is the rank, this is a rank factorization of A),
  4. there exist k rows   of size n such that every row of A is a linear combination of  ,
  5. the row rank of A is less than or equal to k.

Indeed, the following equivalences are obvious:  . For example, to prove (3) from (2), take C to be the matrix whose columns are   from (2). To prove (2) from (3), take   to be the columns of C.

It follows from the equivalence   that the row rank is equal to the column rank.

As in the case of the "dimension of image" characterization, this can be generalized to a definition of the rank of any linear map: the rank of a linear map f : VW is the minimal dimension k of an intermediate space X such that f can be written as the composition of a map VX and a map XW. Unfortunately, this definition does not suggest an efficient manner to compute the rank (for which it is better to use one of the alternative definitions). See rank factorization for details.

Rank in terms of singular values edit

The rank of A equals the number of non-zero singular values, which is the same as the number of non-zero diagonal elements in Σ in the singular value decomposition  .

Determinantal rank – size of largest non-vanishing minor edit

The rank of A is the largest order of any non-zero minor in A. (The order of a minor is the side-length of the square sub-matrix of which it is the determinant.) Like the decomposition rank characterization, this does not give an efficient way of computing the rank, but it is useful theoretically: a single non-zero minor witnesses a lower bound (namely its order) for the rank of the matrix, which can be useful (for example) to prove that certain operations do not lower the rank of a matrix.

A non-vanishing p-minor (p × p submatrix with non-zero determinant) shows that the rows and columns of that submatrix are linearly independent, and thus those rows and columns of the full matrix are linearly independent (in the full matrix), so the row and column rank are at least as large as the determinantal rank; however, the converse is less straightforward. The equivalence of determinantal rank and column rank is a strengthening of the statement that if the span of n vectors has dimension p, then p of those vectors span the space (equivalently, that one can choose a spanning set that is a subset of the vectors): the equivalence implies that a subset of the rows and a subset of the columns simultaneously define an invertible submatrix (equivalently, if the span of n vectors has dimension p, then p of these vectors span the space and there is a set of p coordinates on which they are linearly independent).

Tensor rank – minimum number of simple tensors edit

The rank of A is the smallest number k such that A can be written as a sum of k rank 1 matrices, where a matrix is defined to have rank 1 if and only if it can be written as a nonzero product   of a column vector c and a row vector r. This notion of rank is called tensor rank; it can be generalized in the separable models interpretation of the singular value decomposition.

Properties edit

We assume that A is an m × n matrix, and we define the linear map f by f(x) = Ax as above.

  • The rank of an m × n matrix is a nonnegative integer and cannot be greater than either m or n. That is,
     
    A matrix that has rank min(m, n) is said to have full rank; otherwise, the matrix is rank deficient.
  • Only a zero matrix has rank zero.
  • f is injective (or "one-to-one") if and only if A has rank n (in this case, we say that A has full column rank).
  • f is surjective (or "onto") if and only if A has rank m (in this case, we say that A has full row rank).
  • If A is a square matrix (i.e., m = n), then A is invertible if and only if A has rank n (that is, A has full rank).
  • If B is any n × k matrix, then
     
  • If B is an n × k matrix of rank n, then
     
  • If C is an l × m matrix of rank m, then
     
  • The rank of A is equal to r if and only if there exists an invertible m × m matrix X and an invertible n × n matrix Y such that
     
    where Ir denotes the r × r identity matrix.
  • Sylvester’s rank inequality: if A is an m × n matrix and B is n × k, then[ii]
     
    This is a special case of the next inequality.
  • The inequality due to Frobenius: if AB, ABC and BC are defined, then[iii]
     
  • Subadditivity:
     
    when A and B are of the same dimension. As a consequence, a rank-k matrix can be written as the sum of k rank-1 matrices, but not fewer.
  • The rank of a matrix plus the nullity of the matrix equals the number of columns of the matrix. (This is the rank–nullity theorem.)
  • If A is a matrix over the real numbers then the rank of A and the rank of its corresponding Gram matrix are equal. Thus, for real matrices
     
    This can be shown by proving equality of their null spaces. The null space of the Gram matrix is given by vectors x for which   If this condition is fulfilled, we also have  [11]
  • If A is a matrix over the complex numbers and   denotes the complex conjugate of A and A the conjugate transpose of A (i.e., the adjoint of A), then
     

Applications edit

One useful application of calculating the rank of a matrix is the computation of the number of solutions of a system of linear equations. According to the Rouché–Capelli theorem, the system is inconsistent if the rank of the augmented matrix is greater than the rank of the coefficient matrix. If on the other hand, the ranks of these two matrices are equal, then the system must have at least one solution. The solution is unique if and only if the rank equals the number of variables. Otherwise the general solution has k free parameters where k is the difference between the number of variables and the rank. In this case (and assuming the system of equations is in the real or complex numbers) the system of equations has infinitely many solutions.

In control theory, the rank of a matrix can be used to determine whether a linear system is controllable, or observable.

In the field of communication complexity, the rank of the communication matrix of a function gives bounds on the amount of communication needed for two parties to compute the function.

Generalization edit

There are different generalizations of the concept of rank to matrices over arbitrary rings, where column rank, row rank, dimension of column space, and dimension of row space of a matrix may be different from the others or may not exist.

Thinking of matrices as tensors, the tensor rank generalizes to arbitrary tensors; for tensors of order greater than 2 (matrices are order 2 tensors), rank is very hard to compute, unlike for matrices.

There is a notion of rank for smooth maps between smooth manifolds. It is equal to the linear rank of the derivative.

Matrices as tensors edit

Matrix rank should not be confused with tensor order, which is called tensor rank. Tensor order is the number of indices required to write a tensor, and thus matrices all have tensor order 2. More precisely, matrices are tensors of type (1,1), having one row index and one column index, also called covariant order 1 and contravariant order 1; see Tensor (intrinsic definition) for details.

The tensor rank of a matrix can also mean the minimum number of simple tensors necessary to express the matrix as a linear combination, and that this definition does agree with matrix rank as here discussed.

See also edit

Notes edit

  1. ^ Alternative notation includes   from Katznelson & Katznelson (2008, p. 52, §2.5.1) and Halmos (1974, p. 90, § 50).
  2. ^ Proof: Apply the rank–nullity theorem to the inequality
     
  3. ^ Proof. The map
     
    is well-defined and injective. We thus obtain the inequality in terms of dimensions of kernel, which can then be converted to the inequality in terms of ranks by the rank–nullity theorem. Alternatively, if   is a linear subspace then  ; apply this inequality to the subspace defined by the orthogonal complement of the image of   in the image of  , whose dimension is  ; its image under   has dimension  .

References edit

  1. ^ Axler (2015) pp. 111-112, §§ 3.115, 3.119
  2. ^ a b Roman (2005) p. 48, § 1.16
  3. ^ Bourbaki, Algebra, ch. II, §10.12, p. 359
  4. ^ a b Mackiw, G. (1995), "A Note on the Equality of the Column and Row Rank of a Matrix", Mathematics Magazine, 68 (4): 285–286, doi:10.1080/0025570X.1995.11996337
  5. ^ Hefferon (2020) p. 200, ch. 3, Definition 2.1
  6. ^ Katznelson & Katznelson (2008) p. 52, § 2.5.1
  7. ^ Valenza (1993) p. 71, § 4.3
  8. ^ Halmos (1974) p. 90, § 50
  9. ^ Wardlaw, William P. (2005), "Row Rank Equals Column Rank", Mathematics Magazine, 78 (4): 316–318, doi:10.1080/0025570X.2005.11953349, S2CID 218542661
  10. ^ Banerjee, Sudipto; Roy, Anindya (2014), Linear Algebra and Matrix Analysis for Statistics, Texts in Statistical Science (1st ed.), Chapman and Hall/CRC, ISBN 978-1420095388
  11. ^ Mirsky, Leonid (1955). An introduction to linear algebra. Dover Publications. ISBN 978-0-486-66434-7.

Sources edit

Further reading edit

  • Roger A. Horn and Charles R. Johnson (1985). Matrix Analysis. Cambridge University Press. ISBN 978-0-521-38632-6.
  • Kaw, Autar K. Two Chapters from the book Introduction to Matrix Algebra: 1. Vectors [1] and System of Equations [2]
  • Mike Brookes: Matrix Reference Manual. [3]

rank, linear, algebra, linear, algebra, rank, matrix, dimension, vector, space, generated, spanned, columns, this, corresponds, maximal, number, linearly, independent, columns, this, turn, identical, dimension, vector, space, spanned, rows, rank, thus, measure. In linear algebra the rank of a matrix A is the dimension of the vector space generated or spanned by its columns 1 2 3 This corresponds to the maximal number of linearly independent columns of A This in turn is identical to the dimension of the vector space spanned by its rows 4 Rank is thus a measure of the nondegenerateness of the system of linear equations and linear transformation encoded by A There are multiple equivalent definitions of rank A matrix s rank is one of its most fundamental characteristics The rank is commonly denoted by rank A or rk A 2 sometimes the parentheses are not written as in rank A i Contents 1 Main definitions 2 Examples 3 Computing the rank of a matrix 3 1 Rank from row echelon forms 3 2 Computation 4 Proofs that column rank row rank 4 1 Proof using row reduction 4 2 Proof using linear combinations 4 3 Proof using orthogonality 5 Alternative definitions 5 1 Dimension of image 5 2 Rank in terms of nullity 5 3 Column rank dimension of column space 5 4 Row rank dimension of row space 5 5 Decomposition rank 5 6 Rank in terms of singular values 5 7 Determinantal rank size of largest non vanishing minor 5 8 Tensor rank minimum number of simple tensors 6 Properties 7 Applications 8 Generalization 9 Matrices as tensors 10 See also 11 Notes 12 References 13 Sources 14 Further readingMain definitions editIn this section we give some definitions of the rank of a matrix Many definitions are possible see Alternative definitions for several of these The column rank of A is the dimension of the column space of A while the row rank of A is the dimension of the row space of A A fundamental result in linear algebra is that the column rank and the row rank are always equal Three proofs of this result are given in Proofs that column rank row rank below This number i e the number of linearly independent rows or columns is simply called the rank of A A matrix is said to have full rank if its rank equals the largest possible for a matrix of the same dimensions which is the lesser of the number of rows and columns A matrix is said to be rank deficient if it does not have full rank The rank deficiency of a matrix is the difference between the lesser of the number of rows and columns and the rank The rank of a linear map or operator F displaystyle Phi nbsp is defined as the dimension of its image 5 6 7 8 rank F dim img F displaystyle operatorname rank Phi dim operatorname img Phi nbsp where dim displaystyle dim nbsp is the dimension of a vector space and img displaystyle operatorname img nbsp is the image of a map Examples editThe matrix 1 0 1 0 1 1 0 1 1 displaystyle begin bmatrix 1 amp 0 amp 1 0 amp 1 amp 1 0 amp 1 amp 1 end bmatrix nbsp has rank 2 the first two columns are linearly independent so the rank is at least 2 but since the third is a linear combination of the first two the first column plus the second the three columns are linearly dependent so the rank must be less than 3 The matrixA 1 1 0 2 1 1 0 2 displaystyle A begin bmatrix 1 amp 1 amp 0 amp 2 1 amp 1 amp 0 amp 2 end bmatrix nbsp has rank 1 there are nonzero columns so the rank is positive but any pair of columns is linearly dependent Similarly the transpose A T 1 1 1 1 0 0 2 2 displaystyle A mathrm T begin bmatrix 1 amp 1 1 amp 1 0 amp 0 2 amp 2 end bmatrix nbsp of A has rank 1 Indeed since the column vectors of A are the row vectors of the transpose of A the statement that the column rank of a matrix equals its row rank is equivalent to the statement that the rank of a matrix is equal to the rank of its transpose i e rank A rank AT Computing the rank of a matrix editRank from row echelon forms edit Main article Gaussian elimination A common approach to finding the rank of a matrix is to reduce it to a simpler form generally row echelon form by elementary row operations Row operations do not change the row space hence do not change the row rank and being invertible map the column space to an isomorphic space hence do not change the column rank Once in row echelon form the rank is clearly the same for both row rank and column rank and equals the number of pivots or basic columns and also the number of non zero rows For example the matrix A given byA 1 2 1 2 3 1 3 5 0 displaystyle A begin bmatrix 1 amp 2 amp 1 2 amp 3 amp 1 3 amp 5 amp 0 end bmatrix nbsp can be put in reduced row echelon form by using the following elementary row operations 1 2 1 2 3 1 3 5 0 2 R 1 R 2 R 2 1 2 1 0 1 3 3 5 0 3 R 1 R 3 R 3 1 2 1 0 1 3 0 1 3 R 2 R 3 R 3 1 2 1 0 1 3 0 0 0 2 R 2 R 1 R 1 1 0 5 0 1 3 0 0 0 displaystyle begin aligned begin bmatrix 1 amp 2 amp 1 2 amp 3 amp 1 3 amp 5 amp 0 end bmatrix amp xrightarrow 2R 1 R 2 to R 2 begin bmatrix 1 amp 2 amp 1 0 amp 1 amp 3 3 amp 5 amp 0 end bmatrix xrightarrow 3R 1 R 3 to R 3 begin bmatrix 1 amp 2 amp 1 0 amp 1 amp 3 0 amp 1 amp 3 end bmatrix amp xrightarrow R 2 R 3 to R 3 begin bmatrix 1 amp 2 amp 1 0 amp 1 amp 3 0 amp 0 amp 0 end bmatrix xrightarrow 2R 2 R 1 to R 1 begin bmatrix 1 amp 0 amp 5 0 amp 1 amp 3 0 amp 0 amp 0 end bmatrix end aligned nbsp The final matrix in reduced row echelon form has two non zero rows and thus the rank of matrix A is 2 Computation edit When applied to floating point computations on computers basic Gaussian elimination LU decomposition can be unreliable and a rank revealing decomposition should be used instead An effective alternative is the singular value decomposition SVD but there are other less computationally expensive choices such as QR decomposition with pivoting so called rank revealing QR factorization which are still more numerically robust than Gaussian elimination Numerical determination of rank requires a criterion for deciding when a value such as a singular value from the SVD should be treated as zero a practical choice which depends on both the matrix and the application Proofs that column rank row rank editProof using row reduction edit The fact that the column and row ranks of any matrix are equal forms is fundamental in linear algebra Many proofs have been given One of the most elementary ones has been sketched in Rank from row echelon forms Here is a variant of this proof It is straightforward to show that neither the row rank nor the column rank are changed by an elementary row operation As Gaussian elimination proceeds by elementary row operations the reduced row echelon form of a matrix has the same row rank and the same column rank as the original matrix Further elementary column operations allow putting the matrix in the form of an identity matrix possibly bordered by rows and columns of zeros Again this changes neither the row rank nor the column rank It is immediate that both the row and column ranks of this resulting matrix is the number of its nonzero entries We present two other proofs of this result The first uses only basic properties of linear combinations of vectors and is valid over any field The proof is based upon Wardlaw 2005 9 The second uses orthogonality and is valid for matrices over the real numbers it is based upon Mackiw 1995 4 Both proofs can be found in the book by Banerjee and Roy 2014 10 Proof using linear combinations edit Let A be an m n matrix Let the column rank of A be r and let c1 cr be any basis for the column space of A Place these as the columns of an m r matrix C Every column of A can be expressed as a linear combination of the r columns in C This means that there is an r n matrix R such that A CR R is the matrix whose i th column is formed from the coefficients giving the i th column of A as a linear combination of the r columns of C In other words R is the matrix which contains the multiples for the bases of the column space of A which is C which are then used to form A as a whole Now each row of A is given by a linear combination of the r rows of R Therefore the rows of R form a spanning set of the row space of A and by the Steinitz exchange lemma the row rank of A cannot exceed r This proves that the row rank of A is less than or equal to the column rank of A This result can be applied to any matrix so apply the result to the transpose of A Since the row rank of the transpose of A is the column rank of A and the column rank of the transpose of A is the row rank of A this establishes the reverse inequality and we obtain the equality of the row rank and the column rank of A Also see Rank factorization Proof using orthogonality edit Let A be an m n matrix with entries in the real numbers whose row rank is r Therefore the dimension of the row space of A is r Let x1 x2 xr be a basis of the row space of A We claim that the vectors Ax1 Ax2 Axr are linearly independent To see why consider a linear homogeneous relation involving these vectors with scalar coefficients c1 c2 cr 0 c 1 A x 1 c 2 A x 2 c r A x r A c 1 x 1 c 2 x 2 c r x r A v displaystyle 0 c 1 A mathbf x 1 c 2 A mathbf x 2 cdots c r A mathbf x r A c 1 mathbf x 1 c 2 mathbf x 2 cdots c r mathbf x r A mathbf v nbsp where v c1x1 c2x2 crxr We make two observations a v is a linear combination of vectors in the row space of A which implies that v belongs to the row space of A and b since Av 0 the vector v is orthogonal to every row vector of A and hence is orthogonal to every vector in the row space of A The facts a and b together imply that v is orthogonal to itself which proves that v 0 or by the definition of v c 1 x 1 c 2 x 2 c r x r 0 displaystyle c 1 mathbf x 1 c 2 mathbf x 2 cdots c r mathbf x r 0 nbsp But recall that the xi were chosen as a basis of the row space of A and so are linearly independent This implies that c1 c2 cr 0 It follows that Ax1 Ax2 Axr are linearly independent Now each Axi is obviously a vector in the column space of A So Ax1 Ax2 Axr is a set of r linearly independent vectors in the column space of A and hence the dimension of the column space of A i e the column rank of A must be at least as big as r This proves that row rank of A is no larger than the column rank of A Now apply this result to the transpose of A to get the reverse inequality and conclude as in the previous proof Alternative definitions editIn all the definitions in this section the matrix A is taken to be an m n matrix over an arbitrary field F Dimension of image edit Given the matrix A displaystyle A nbsp there is an associated linear mappingf F n F m displaystyle f F n to F m nbsp defined by f x A x displaystyle f x Ax nbsp The rank of A displaystyle A nbsp is the dimension of the image of f displaystyle f nbsp This definition has the advantage that it can be applied to any linear map without need for a specific matrix Rank in terms of nullity edit Given the same linear mapping f as above the rank is n minus the dimension of the kernel of f The rank nullity theorem states that this definition is equivalent to the preceding one Column rank dimension of column space edit The rank of A is the maximal number of linearly independent columns c 1 c 2 c k displaystyle mathbf c 1 mathbf c 2 dots mathbf c k nbsp of A this is the dimension of the column space of A the column space being the subspace of Fm generated by the columns of A which is in fact just the image of the linear map f associated to A Row rank dimension of row space edit The rank of A is the maximal number of linearly independent rows of A this is the dimension of the row space of A Decomposition rank edit The rank of A is the smallest integer k such that A can be factored as A C R displaystyle A CR nbsp where C is an m k matrix and R is a k n matrix In fact for all integers k the following are equivalent the column rank of A is less than or equal to k there exist k columns c 1 c k displaystyle mathbf c 1 ldots mathbf c k nbsp of size m such that every column of A is a linear combination of c 1 c k displaystyle mathbf c 1 ldots mathbf c k nbsp there exist an m k displaystyle m times k nbsp matrix C and a k n displaystyle k times n nbsp matrix R such that A C R displaystyle A CR nbsp when k is the rank this is a rank factorization of A there exist k rows r 1 r k displaystyle mathbf r 1 ldots mathbf r k nbsp of size n such that every row of A is a linear combination of r 1 r k displaystyle mathbf r 1 ldots mathbf r k nbsp the row rank of A is less than or equal to k Indeed the following equivalences are obvious 1 2 3 4 5 displaystyle 1 Leftrightarrow 2 Leftrightarrow 3 Leftrightarrow 4 Leftrightarrow 5 nbsp For example to prove 3 from 2 take C to be the matrix whose columns are c 1 c k displaystyle mathbf c 1 ldots mathbf c k nbsp from 2 To prove 2 from 3 take c 1 c k displaystyle mathbf c 1 ldots mathbf c k nbsp to be the columns of C It follows from the equivalence 1 5 displaystyle 1 Leftrightarrow 5 nbsp that the row rank is equal to the column rank As in the case of the dimension of image characterization this can be generalized to a definition of the rank of any linear map the rank of a linear map f V W is the minimal dimension k of an intermediate space X such that f can be written as the composition of a map V X and a map X W Unfortunately this definition does not suggest an efficient manner to compute the rank for which it is better to use one of the alternative definitions See rank factorization for details Rank in terms of singular values edit The rank of A equals the number of non zero singular values which is the same as the number of non zero diagonal elements in S in the singular value decomposition A U S V displaystyle A U Sigma V nbsp Determinantal rank size of largest non vanishing minor edit The rank of A is the largest order of any non zero minor in A The order of a minor is the side length of the square sub matrix of which it is the determinant Like the decomposition rank characterization this does not give an efficient way of computing the rank but it is useful theoretically a single non zero minor witnesses a lower bound namely its order for the rank of the matrix which can be useful for example to prove that certain operations do not lower the rank of a matrix A non vanishing p minor p p submatrix with non zero determinant shows that the rows and columns of that submatrix are linearly independent and thus those rows and columns of the full matrix are linearly independent in the full matrix so the row and column rank are at least as large as the determinantal rank however the converse is less straightforward The equivalence of determinantal rank and column rank is a strengthening of the statement that if the span of n vectors has dimension p then p of those vectors span the space equivalently that one can choose a spanning set that is a subset of the vectors the equivalence implies that a subset of the rows and a subset of the columns simultaneously define an invertible submatrix equivalently if the span of n vectors has dimension p then p of these vectors span the space and there is a set of p coordinates on which they are linearly independent Tensor rank minimum number of simple tensors edit Main articles Tensor rank decomposition and Tensor rank The rank of A is the smallest number k such that A can be written as a sum of k rank 1 matrices where a matrix is defined to have rank 1 if and only if it can be written as a nonzero product c r displaystyle c cdot r nbsp of a column vector c and a row vector r This notion of rank is called tensor rank it can be generalized in the separable models interpretation of the singular value decomposition Properties editWe assume that A is an m n matrix and we define the linear map f by f x Ax as above The rank of an m n matrix is a nonnegative integer and cannot be greater than either m or n That is rank A min m n displaystyle operatorname rank A leq min m n nbsp A matrix that has rank min m n is said to have full rank otherwise the matrix is rank deficient Only a zero matrix has rank zero f is injective or one to one if and only if A has rank n in this case we say that A has full column rank f is surjective or onto if and only if A has rank m in this case we say that A has full row rank If A is a square matrix i e m n then A is invertible if and only if A has rank n that is A has full rank If B is any n k matrix then rank A B min rank A rank B displaystyle operatorname rank AB leq min operatorname rank A operatorname rank B nbsp If B is an n k matrix of rank n then rank A B rank A displaystyle operatorname rank AB operatorname rank A nbsp If C is an l m matrix of rank m then rank C A rank A displaystyle operatorname rank CA operatorname rank A nbsp The rank of A is equal to r if and only if there exists an invertible m m matrix X and an invertible n n matrix Y such that X A Y I r 0 0 0 displaystyle XAY begin bmatrix I r amp 0 0 amp 0 end bmatrix nbsp where Ir denotes the r r identity matrix Sylvester s rank inequality if A is an m n matrix and B is n k then ii rank A rank B n rank A B displaystyle operatorname rank A operatorname rank B n leq operatorname rank AB nbsp This is a special case of the next inequality The inequality due to Frobenius if AB ABC and BC are defined then iii rank A B rank B C rank B rank A B C displaystyle operatorname rank AB operatorname rank BC leq operatorname rank B operatorname rank ABC nbsp Subadditivity rank A B rank A rank B displaystyle operatorname rank A B leq operatorname rank A operatorname rank B nbsp when A and B are of the same dimension As a consequence a rank k matrix can be written as the sum of k rank 1 matrices but not fewer The rank of a matrix plus the nullity of the matrix equals the number of columns of the matrix This is the rank nullity theorem If A is a matrix over the real numbers then the rank of A and the rank of its corresponding Gram matrix are equal Thus for real matrices rank A T A rank A A T rank A rank A T displaystyle operatorname rank A mathrm T A operatorname rank AA mathrm T operatorname rank A operatorname rank A mathrm T nbsp This can be shown by proving equality of their null spaces The null space of the Gram matrix is given by vectors x for which A T A x 0 displaystyle A mathrm T A mathbf x 0 nbsp If this condition is fulfilled we also have 0 x T A T A x A x 2 displaystyle 0 mathbf x mathrm T A mathrm T A mathbf x left A mathbf x right 2 nbsp 11 If A is a matrix over the complex numbers and A displaystyle overline A nbsp denotes the complex conjugate of A and A the conjugate transpose of A i e the adjoint of A then rank A rank A rank A T rank A rank A A rank A A displaystyle operatorname rank A operatorname rank overline A operatorname rank A mathrm T operatorname rank A operatorname rank A A operatorname rank AA nbsp Applications editOne useful application of calculating the rank of a matrix is the computation of the number of solutions of a system of linear equations According to the Rouche Capelli theorem the system is inconsistent if the rank of the augmented matrix is greater than the rank of the coefficient matrix If on the other hand the ranks of these two matrices are equal then the system must have at least one solution The solution is unique if and only if the rank equals the number of variables Otherwise the general solution has k free parameters where k is the difference between the number of variables and the rank In this case and assuming the system of equations is in the real or complex numbers the system of equations has infinitely many solutions In control theory the rank of a matrix can be used to determine whether a linear system is controllable or observable In the field of communication complexity the rank of the communication matrix of a function gives bounds on the amount of communication needed for two parties to compute the function Generalization editThere are different generalizations of the concept of rank to matrices over arbitrary rings where column rank row rank dimension of column space and dimension of row space of a matrix may be different from the others or may not exist Thinking of matrices as tensors the tensor rank generalizes to arbitrary tensors for tensors of order greater than 2 matrices are order 2 tensors rank is very hard to compute unlike for matrices There is a notion of rank for smooth maps between smooth manifolds It is equal to the linear rank of the derivative Matrices as tensors editMatrix rank should not be confused with tensor order which is called tensor rank Tensor order is the number of indices required to write a tensor and thus matrices all have tensor order 2 More precisely matrices are tensors of type 1 1 having one row index and one column index also called covariant order 1 and contravariant order 1 see Tensor intrinsic definition for details The tensor rank of a matrix can also mean the minimum number of simple tensors necessary to express the matrix as a linear combination and that this definition does agree with matrix rank as here discussed See also editMatroid rank Nonnegative rank linear algebra Rank differential topology Multicollinearity Linear dependenceNotes edit Alternative notation includes r F displaystyle rho Phi nbsp from Katznelson amp Katznelson 2008 p 52 2 5 1 and Halmos 1974 p 90 50 Proof Apply the rank nullity theorem to the inequality dim ker A B dim ker A dim ker B displaystyle dim ker AB leq dim ker A dim ker B nbsp Proof The mapC ker A B C ker B C ker A B ker B displaystyle C ker ABC ker BC to ker AB ker B nbsp is well defined and injective We thus obtain the inequality in terms of dimensions of kernel which can then be converted to the inequality in terms of ranks by the rank nullity theorem Alternatively if M displaystyle M nbsp is a linear subspace then dim A M dim M displaystyle dim AM leq dim M nbsp apply this inequality to the subspace defined by the orthogonal complement of the image of B C displaystyle BC nbsp in the image of B displaystyle B nbsp whose dimension is rank B rank B C displaystyle operatorname rank B operatorname rank BC nbsp its image under A displaystyle A nbsp has dimension rank A B rank A B C displaystyle operatorname rank AB operatorname rank ABC nbsp References edit Axler 2015 pp 111 112 3 115 3 119 a b Roman 2005 p 48 1 16 Bourbaki Algebra ch II 10 12 p 359 a b Mackiw G 1995 A Note on the Equality of the Column and Row Rank of a Matrix Mathematics Magazine 68 4 285 286 doi 10 1080 0025570X 1995 11996337 Hefferon 2020 p 200 ch 3 Definition 2 1 Katznelson amp Katznelson 2008 p 52 2 5 1 Valenza 1993 p 71 4 3 Halmos 1974 p 90 50 Wardlaw William P 2005 Row Rank Equals Column Rank Mathematics Magazine 78 4 316 318 doi 10 1080 0025570X 2005 11953349 S2CID 218542661 Banerjee Sudipto Roy Anindya 2014 Linear Algebra and Matrix Analysis for Statistics Texts in Statistical Science 1st ed Chapman and Hall CRC ISBN 978 1420095388 Mirsky Leonid 1955 An introduction to linear algebra Dover Publications ISBN 978 0 486 66434 7 Sources editAxler Sheldon 2015 Linear Algebra Done Right Undergraduate Texts in Mathematics 3rd ed Springer ISBN 978 3 319 11079 0 Halmos Paul Richard 1974 1958 Finite Dimensional Vector Spaces Undergraduate Texts in Mathematics 2nd ed Springer ISBN 0 387 90093 4 Hefferon Jim 2020 Linear Algebra 4th ed Orthogonal Publishing L3C ISBN 978 1 944325 11 4 Katznelson Yitzhak Katznelson Yonatan R 2008 A Terse Introduction to Linear Algebra American Mathematical Society ISBN 978 0 8218 4419 9 Roman Steven 2005 Advanced Linear Algebra Undergraduate Texts in Mathematics 2nd ed Springer ISBN 0 387 24766 1 Valenza Robert J 1993 1951 Linear Algebra An Introduction to Abstract Mathematics Undergraduate Texts in Mathematics 3rd ed Springer ISBN 3 540 94099 5 Further reading editRoger A Horn and Charles R Johnson 1985 Matrix Analysis Cambridge University Press ISBN 978 0 521 38632 6 Kaw Autar K Two Chapters from the book Introduction to Matrix Algebra 1 Vectors 1 and System of Equations 2 Mike Brookes Matrix Reference Manual 3 Retrieved from https en wikipedia org w index php title Rank linear algebra amp oldid 1214965544 Main definitions, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.