fbpx
Wikipedia

Diagonalizable matrix

In linear algebra, a square matrix  is called diagonalizable or non-defective if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix  and a diagonal matrix such that , or equivalently . (Such , are not unique.) For a finite-dimensional vector space , a linear map  is called diagonalizable if there exists an ordered basis of  consisting of eigenvectors of . These definitions are equivalent: if  has a matrix representation as above, then the column vectors of  form a basis consisting of eigenvectors of , and the diagonal entries of  are the corresponding eigenvalues of ; with respect to this eigenvector basis,  is represented by . Diagonalization is the process of finding the above  and .

Diagonalizing a matrix makes many subsequent computations easier. One can raise a diagonal matrix  to a power by simply raising the diagonal entries to that power. The determinant of a diagonal matrix is simply the product of all diagonal entries. Such computations generalize easily to .

The geometric transformation represented by a diagonalizable matrix is an inhomogeneous dilation (or anisotropic scaling), meaning that it scales the space by a different amount in different directions. In particular, the direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue. A inhomogeneous dilation is in contrast to a homogeneous dilation, which scales by the same amount in every direction.

A square matrix that is not diagonalizable is called defective. It can happen that a matrix with real entries is defective over the real numbers, meaning that is impossible for any invertible and diagonal with real entries, but it is possible with complex entries, so that is diagonalizable over the complex numbers. For example, this is the case for a generic rotation matrix.

Many results for diagonalizable matrices hold only over an algebraically closed field (such as the complex numbers). In this case, diagonalizable matrices are dense in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a small perturbation; and the Jordan normal form theorem states that any matrix is uniquely the sum of a diagonalizable matrix and a nilpotent matrix. Over an algebraically closed field, diagonalizable matrices are equivalent to semi-simple matrices.

Definition edit

A square   matrix,  , with entries in a field   is called diagonalizable or nondefective if there exists an   invertible matrix (i.e. an element of the general linear group GLn(F)),  , such that   is a diagonal matrix. Formally,

 

Characterization edit

The fundamental fact about diagonalizable maps and matrices is expressed by the following:

  • An   matrix   over a field   is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to  , which is the case if and only if there exists a basis of   consisting of eigenvectors of  . If such a basis has been found, one can form the matrix   having these basis vectors as columns, and   will be a diagonal matrix whose diagonal entries are the eigenvalues of  . The matrix   is known as a modal matrix for  .
  • A linear map   is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to  , which is the case if and only if there exists a basis of   consisting of eigenvectors of  . With respect to such a basis,   will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of  .

The following sufficient (but not necessary) condition is often useful.

  • An   matrix   is diagonalizable over the field   if it has   distinct eigenvalues in  , i.e. if its characteristic polynomial has   distinct roots in  ; however, the converse may be false. Consider
     
    which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form (similar to  )
     
    and change of basis matrix  :
     
    The converse fails when   has an eigenspace of dimension higher than 1. In this example, the eigenspace of   associated with the eigenvalue 2 has dimension 2.
  • A linear map   with   is diagonalizable if it has   distinct eigenvalues, i.e. if its characteristic polynomial has   distinct roots in  .

Let   be a matrix over  . If   is diagonalizable, then so is any power of it. Conversely, if   is invertible,   is algebraically closed, and   is diagonalizable for some   that is not an integer multiple of the characteristic of  , then   is diagonalizable. Proof: If   is diagonalizable, then   is annihilated by some polynomial  , which has no multiple root (since  ) and is divided by the minimal polynomial of  .

Over the complex numbers  , almost every matrix is diagonalizable. More precisely: the set of complex   matrices that are not diagonalizable over  , considered as a subset of  , has Lebesgue measure zero. One can also say that the diagonalizable matrices form a dense subset with respect to the Zariski topology: the non-diagonalizable matrices lie inside the vanishing set of the discriminant of the characteristic polynomial, which is a hypersurface. From that follows also density in the usual (strong) topology given by a norm. The same is not true over  .

The Jordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its nilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its Jordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.

Diagonalization edit

 
The diagonalization of a symmetric matrix can be interpreted as a rotation of the axes to align them with the eigenvectors.

If a matrix   can be diagonalized, that is,

 

then:

 

Writing   as a block matrix of its column vectors  

 

the above equation can be rewritten as

 

So the column vectors of   are right eigenvectors of  , and the corresponding diagonal entry is the corresponding eigenvalue. The invertibility of   also suggests that the eigenvectors are linearly independent and form a basis of  . This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The row vectors of   are the left eigenvectors of  .

When a complex matrix   is a Hermitian matrix (or more generally a normal matrix), eigenvectors of   can be chosen to form an orthonormal basis of  , and   can be chosen to be a unitary matrix. If in addition,   is a real symmetric matrix, then its eigenvectors can be chosen to be an orthonormal basis of   and   can be chosen to be an orthogonal matrix.

For most practical work matrices are diagonalized numerically using computer software. Many algorithms exist to accomplish this.

Simultaneous diagonalization edit

A set of matrices is said to be simultaneously diagonalizable if there exists a single invertible matrix   such that   is a diagonal matrix for every   in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizable matrices commutes if and only if the set is simultaneously diagonalizable.[1]: p. 64 

The set of all   diagonalizable matrices (over  ) with   is not simultaneously diagonalizable. For instance, the matrices

 

are diagonalizable but not simultaneously diagonalizable because they do not commute.

A set consists of commuting normal matrices if and only if it is simultaneously diagonalizable by a unitary matrix; that is, there exists a unitary matrix   such that   is diagonal for every   in the set.

In the language of Lie theory, a set of simultaneously diagonalizable matrices generates a toral Lie algebra.

Examples edit

Diagonalizable matrices edit

  • Involutions are diagonalizable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal.
  • Finite order endomorphisms are diagonalizable over   (or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with roots of unity on the diagonal. This follows since the minimal polynomial is separable, because the roots of unity are distinct.
  • Projections are diagonalizable, with 0s and 1s on the diagonal.
  • Real symmetric matrices are diagonalizable by orthogonal matrices; i.e., given a real symmetric matrix  ,   is diagonal for some orthogonal matrix  . More generally, matrices are diagonalizable by unitary matrices if and only if they are normal. In the case of the real symmetric matrix, we see that  , so clearly   holds. Examples of normal matrices are real symmetric (or skew-symmetric) matrices (e.g. covariance matrices) and Hermitian matrices (or skew-Hermitian matrices). See spectral theorems for generalizations to infinite-dimensional vector spaces.

Matrices that are not diagonalizable edit

In general, a rotation matrix is not diagonalizable over the reals, but all rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as Jordan normal form.

Some matrices are not diagonalizable over any field, most notably nonzero nilpotent matrices. This happens more generally if the algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider

 

This matrix is not diagonalizable: there is no matrix   such that   is a diagonal matrix. Indeed,   has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.

Some real matrices are not diagonalizable over the reals. Consider for instance the matrix

 

The matrix   does not have any real eigenvalues, so there is no real matrix   such that   is a diagonal matrix. However, we can diagonalize   if we allow complex numbers. Indeed, if we take

 

then   is diagonal. It is easy to find that   is the rotation matrix which rotates counterclockwise by angle  

Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.

How to diagonalize a matrix edit

Diagonalizing a matrix is the same process as finding its eigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix

 

The roots of the characteristic polynomial   are the eigenvalues  . Solving the linear system   gives the eigenvectors   and  , while   gives  ; that is,   for  . These vectors form a basis of  , so we can assemble them as the column vectors of a change-of-basis matrix   to get:

 
We may see this equation in terms of transformations:   takes the standard basis to the eigenbasis,  , so we have:
 
so that   has the standard basis as its eigenvectors, which is the defining property of  .

Note that there is no preferred order of the eigenvectors in  ; changing the order of the eigenvectors in   just changes the order of the eigenvalues in the diagonalized form of  .[2]

Application to matrix functions edit

Diagonalization can be used to efficiently compute the powers of a matrix  :

 

and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix   with eigenvalues   in the example above we compute:

 

This approach can be generalized to matrix exponential and other matrix functions that can be defined as power series. For example, defining  , we have:

 

This is particularly useful in finding closed form expressions for terms of linear recursive sequences, such as the Fibonacci numbers.

Particular application edit

For example, consider the following matrix:

 

Calculating the various powers of   reveals a surprising pattern:

 

The above phenomenon can be explained by diagonalizing  . To accomplish this, we need a basis of   consisting of eigenvectors of  . One such eigenvector basis is given by

 

where ei denotes the standard basis of Rn. The reverse change of basis is given by

 

Straightforward calculations show that

 

Thus, a and b are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that

 

Switching back to the standard basis, we have

 

The preceding relations, expressed in matrix form, are

 

thereby explaining the above phenomenon.

Quantum mechanical application edit

In quantum mechanical and quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent Schrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional Hilbert space.

A very common approximation is to truncate Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on the variational principle, valid for Hamiltonians that are bounded from below.

First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.

See also edit

Notes edit

References edit

  1. ^ Horn, Roger A.; Johnson, Charles R. (2013). Matrix Analysis, second edition. Cambridge University Press. ISBN 9780521839402.
  2. ^ Anton, H.; Rorres, C. (22 Feb 2000). Elementary Linear Algebra (Applications Version) (8th ed.). John Wiley & Sons. ISBN 978-0-471-17052-5.

diagonalizable, matrix, this, article, about, matrix, diagonalization, linear, algebra, other, uses, diagonalization, linear, algebra, square, matrix, displaystyle, called, diagonalizable, defective, similar, diagonal, matrix, there, exists, invertible, matrix. This article is about matrix diagonalization in linear algebra For other uses see Diagonalization In linear algebra a square matrix A displaystyle A is called diagonalizable or non defective if it is similar to a diagonal matrix i e if there exists an invertible matrix P displaystyle P and a diagonal matrix D displaystyle D such that P 1 A P D displaystyle P 1 AP D or equivalently A P D P 1 displaystyle A PDP 1 Such P displaystyle P D displaystyle D are not unique For a finite dimensional vector space V displaystyle V a linear map T V V displaystyle T V to V is called diagonalizable if there exists an ordered basis of V displaystyle V consisting of eigenvectors of T displaystyle T These definitions are equivalent if T displaystyle T has a matrix representation T P D P 1 displaystyle T PDP 1 as above then the column vectors of P displaystyle P form a basis consisting of eigenvectors of T displaystyle T and the diagonal entries of D displaystyle D are the corresponding eigenvalues of T displaystyle T with respect to this eigenvector basis A displaystyle A is represented by D displaystyle D Diagonalization is the process of finding the above P displaystyle P and D displaystyle D Diagonalizing a matrix makes many subsequent computations easier One can raise a diagonal matrix D displaystyle D to a power by simply raising the diagonal entries to that power The determinant of a diagonal matrix is simply the product of all diagonal entries Such computations generalize easily to A P D P 1 displaystyle A PDP 1 The geometric transformation represented by a diagonalizable matrix is an inhomogeneous dilation or anisotropic scaling meaning that it scales the space by a different amount in different directions In particular the direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue A inhomogeneous dilation is in contrast to a homogeneous dilation which scales by the same amount in every direction A square matrix that is not diagonalizable is called defective It can happen that a matrix A displaystyle A with real entries is defective over the real numbers meaning that A P D P 1 displaystyle A PDP 1 is impossible for any invertible P displaystyle P and diagonal D displaystyle D with real entries but it is possible with complex entries so that A displaystyle A is diagonalizable over the complex numbers For example this is the case for a generic rotation matrix Many results for diagonalizable matrices hold only over an algebraically closed field such as the complex numbers In this case diagonalizable matrices are dense in the space of all matrices which means any defective matrix can be deformed into a diagonalizable matrix by a small perturbation and the Jordan normal form theorem states that any matrix is uniquely the sum of a diagonalizable matrix and a nilpotent matrix Over an algebraically closed field diagonalizable matrices are equivalent to semi simple matrices Contents 1 Definition 2 Characterization 3 Diagonalization 4 Simultaneous diagonalization 5 Examples 5 1 Diagonalizable matrices 5 2 Matrices that are not diagonalizable 5 3 How to diagonalize a matrix 6 Application to matrix functions 6 1 Particular application 7 Quantum mechanical application 8 See also 9 Notes 10 ReferencesDefinition editA square n n displaystyle n times n nbsp matrix A displaystyle A nbsp with entries in a field F displaystyle F nbsp is called diagonalizable or nondefective if there exists an n n displaystyle n times n nbsp invertible matrix i e an element of the general linear group GLn F P displaystyle P nbsp such that P 1 A P displaystyle P 1 AP nbsp is a diagonal matrix Formally A F n n diagonalizable P GL n F P 1 A P diagonal displaystyle A in F n times n text diagonalizable iff exists P in operatorname GL n F P 1 AP text diagonal nbsp Characterization editThe fundamental fact about diagonalizable maps and matrices is expressed by the following An n n displaystyle n times n nbsp matrix A displaystyle A nbsp over a field F displaystyle F nbsp is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to n displaystyle n nbsp which is the case if and only if there exists a basis of F n displaystyle F n nbsp consisting of eigenvectors of A displaystyle A nbsp If such a basis has been found one can form the matrix P displaystyle P nbsp having these basis vectors as columns and P 1 A P displaystyle P 1 AP nbsp will be a diagonal matrix whose diagonal entries are the eigenvalues of A displaystyle A nbsp The matrix P displaystyle P nbsp is known as a modal matrix for A displaystyle A nbsp A linear map T V V displaystyle T V to V nbsp is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to dim V displaystyle dim V nbsp which is the case if and only if there exists a basis of V displaystyle V nbsp consisting of eigenvectors of T displaystyle T nbsp With respect to such a basis T displaystyle T nbsp will be represented by a diagonal matrix The diagonal entries of this matrix are the eigenvalues of T displaystyle T nbsp The following sufficient but not necessary condition is often useful An n n displaystyle n times n nbsp matrix A displaystyle A nbsp is diagonalizable over the field F displaystyle F nbsp if it has n displaystyle n nbsp distinct eigenvalues in F displaystyle F nbsp i e if its characteristic polynomial has n displaystyle n nbsp distinct roots in F displaystyle F nbsp however the converse may be false Consider 1 3 1 3 5 1 3 3 1 displaystyle begin bmatrix 1 amp 3 amp 1 3 amp 5 amp 1 3 amp 3 amp 1 end bmatrix nbsp which has eigenvalues 1 2 2 not all distinct and is diagonalizable with diagonal form similar to A displaystyle A nbsp 1 0 0 0 2 0 0 0 2 displaystyle begin bmatrix 1 amp 0 amp 0 0 amp 2 amp 0 0 amp 0 amp 2 end bmatrix nbsp and change of basis matrix P displaystyle P nbsp 1 1 1 1 1 0 1 0 3 displaystyle begin bmatrix 1 amp 1 amp 1 1 amp 1 amp 0 1 amp 0 amp 3 end bmatrix nbsp The converse fails when A displaystyle A nbsp has an eigenspace of dimension higher than 1 In this example the eigenspace of A displaystyle A nbsp associated with the eigenvalue 2 has dimension 2 A linear map T V V displaystyle T V to V nbsp with n dim V displaystyle n dim V nbsp is diagonalizable if it has n displaystyle n nbsp distinct eigenvalues i e if its characteristic polynomial has n displaystyle n nbsp distinct roots in F displaystyle F nbsp Let A displaystyle A nbsp be a matrix over F displaystyle F nbsp If A displaystyle A nbsp is diagonalizable then so is any power of it Conversely if A displaystyle A nbsp is invertible F displaystyle F nbsp is algebraically closed and A n displaystyle A n nbsp is diagonalizable for some n displaystyle n nbsp that is not an integer multiple of the characteristic of F displaystyle F nbsp then A displaystyle A nbsp is diagonalizable Proof If A n displaystyle A n nbsp is diagonalizable then A displaystyle A nbsp is annihilated by some polynomial x n l 1 x n l k displaystyle left x n lambda 1 right cdots left x n lambda k right nbsp which has no multiple root since l j 0 displaystyle lambda j neq 0 nbsp and is divided by the minimal polynomial of A displaystyle A nbsp Over the complex numbers C displaystyle mathbb C nbsp almost every matrix is diagonalizable More precisely the set of complex n n displaystyle n times n nbsp matrices that are not diagonalizable over C displaystyle mathbb C nbsp considered as a subset of C n n displaystyle mathbb C n times n nbsp has Lebesgue measure zero One can also say that the diagonalizable matrices form a dense subset with respect to the Zariski topology the non diagonalizable matrices lie inside the vanishing set of the discriminant of the characteristic polynomial which is a hypersurface From that follows also density in the usual strong topology given by a norm The same is not true over R displaystyle mathbb R nbsp The Jordan Chevalley decomposition expresses an operator as the sum of its semisimple i e diagonalizable part and its nilpotent part Hence a matrix is diagonalizable if and only if its nilpotent part is zero Put in another way a matrix is diagonalizable if each block in its Jordan form has no nilpotent part i e each block is a one by one matrix Diagonalization editSee also Eigendecomposition of a matrix nbsp The diagonalization of a symmetric matrix can be interpreted as a rotation of the axes to align them with the eigenvectors If a matrix A displaystyle A nbsp can be diagonalized that is P 1 A P l 1 0 0 0 l 2 0 0 0 l n displaystyle P 1 AP begin bmatrix lambda 1 amp 0 amp cdots amp 0 0 amp lambda 2 amp cdots amp 0 vdots amp vdots amp ddots amp vdots 0 amp 0 amp cdots amp lambda n end bmatrix nbsp then A P P l 1 0 0 0 l 2 0 0 0 l n displaystyle AP P begin bmatrix lambda 1 amp 0 amp cdots amp 0 0 amp lambda 2 amp cdots amp 0 vdots amp vdots amp ddots amp vdots 0 amp 0 amp cdots amp lambda n end bmatrix nbsp Writing P displaystyle P nbsp as a block matrix of its column vectors a i displaystyle boldsymbol alpha i nbsp P a 1 a 2 a n displaystyle P begin bmatrix boldsymbol alpha 1 amp boldsymbol alpha 2 amp cdots amp boldsymbol alpha n end bmatrix nbsp the above equation can be rewritten as A a i l i a i i 1 2 n displaystyle A boldsymbol alpha i lambda i boldsymbol alpha i qquad i 1 2 dots n nbsp So the column vectors of P displaystyle P nbsp are right eigenvectors of A displaystyle A nbsp and the corresponding diagonal entry is the corresponding eigenvalue The invertibility of P displaystyle P nbsp also suggests that the eigenvectors are linearly independent and form a basis of F n displaystyle F n nbsp This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization The row vectors of P 1 displaystyle P 1 nbsp are the left eigenvectors of A displaystyle A nbsp When a complex matrix A C n n displaystyle A in mathbb C n times n nbsp is a Hermitian matrix or more generally a normal matrix eigenvectors of A displaystyle A nbsp can be chosen to form an orthonormal basis of C n displaystyle mathbb C n nbsp and P displaystyle P nbsp can be chosen to be a unitary matrix If in addition A R n n displaystyle A in mathbb R n times n nbsp is a real symmetric matrix then its eigenvectors can be chosen to be an orthonormal basis of R n displaystyle mathbb R n nbsp and P displaystyle P nbsp can be chosen to be an orthogonal matrix For most practical work matrices are diagonalized numerically using computer software Many algorithms exist to accomplish this Simultaneous diagonalization editSee also Simultaneous triangularisability Weight representation theory and Positive definite matrix A set of matrices is said to be simultaneously diagonalizable if there exists a single invertible matrix P displaystyle P nbsp such that P 1 A P displaystyle P 1 AP nbsp is a diagonal matrix for every A displaystyle A nbsp in the set The following theorem characterizes simultaneously diagonalizable matrices A set of diagonalizable matrices commutes if and only if the set is simultaneously diagonalizable 1 p 64 The set of all n n displaystyle n times n nbsp diagonalizable matrices over C displaystyle mathbb C nbsp with n gt 1 displaystyle n gt 1 nbsp is not simultaneously diagonalizable For instance the matrices 1 0 0 0 and 1 1 0 0 displaystyle begin bmatrix 1 amp 0 0 amp 0 end bmatrix quad text and quad begin bmatrix 1 amp 1 0 amp 0 end bmatrix nbsp are diagonalizable but not simultaneously diagonalizable because they do not commute A set consists of commuting normal matrices if and only if it is simultaneously diagonalizable by a unitary matrix that is there exists a unitary matrix U displaystyle U nbsp such that U A U displaystyle U AU nbsp is diagonal for every A displaystyle A nbsp in the set In the language of Lie theory a set of simultaneously diagonalizable matrices generates a toral Lie algebra Examples editDiagonalizable matrices edit Involutions are diagonalizable over the reals and indeed any field of characteristic not 2 with 1 on the diagonal Finite order endomorphisms are diagonalizable over C displaystyle mathbb C nbsp or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism with roots of unity on the diagonal This follows since the minimal polynomial is separable because the roots of unity are distinct Projections are diagonalizable with 0s and 1s on the diagonal Real symmetric matrices are diagonalizable by orthogonal matrices i e given a real symmetric matrix A displaystyle A nbsp Q T A Q displaystyle Q mathrm T AQ nbsp is diagonal for some orthogonal matrix Q displaystyle Q nbsp More generally matrices are diagonalizable by unitary matrices if and only if they are normal In the case of the real symmetric matrix we see that A A T displaystyle A A mathrm T nbsp so clearly A A T A T A displaystyle AA mathrm T A mathrm T A nbsp holds Examples of normal matrices are real symmetric or skew symmetric matrices e g covariance matrices and Hermitian matrices or skew Hermitian matrices See spectral theorems for generalizations to infinite dimensional vector spaces Matrices that are not diagonalizable edit In general a rotation matrix is not diagonalizable over the reals but all rotation matrices are diagonalizable over the complex field Even if a matrix is not diagonalizable it is always possible to do the best one can and find a matrix with the same properties consisting of eigenvalues on the leading diagonal and either ones or zeroes on the superdiagonal known as Jordan normal form Some matrices are not diagonalizable over any field most notably nonzero nilpotent matrices This happens more generally if the algebraic and geometric multiplicities of an eigenvalue do not coincide For instance consider C 0 1 0 0 displaystyle C begin bmatrix 0 amp 1 0 amp 0 end bmatrix nbsp This matrix is not diagonalizable there is no matrix U displaystyle U nbsp such that U 1 C U displaystyle U 1 CU nbsp is a diagonal matrix Indeed C displaystyle C nbsp has one eigenvalue namely zero and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1 Some real matrices are not diagonalizable over the reals Consider for instance the matrix B 0 1 1 0 displaystyle B left begin array rr 0 amp 1 1 amp 0 end array right nbsp The matrix B displaystyle B nbsp does not have any real eigenvalues so there is no real matrix Q displaystyle Q nbsp such that Q 1 B Q displaystyle Q 1 BQ nbsp is a diagonal matrix However we can diagonalize B displaystyle B nbsp if we allow complex numbers Indeed if we take Q 1 i i 1 displaystyle Q begin bmatrix 1 amp i i amp 1 end bmatrix nbsp then Q 1 B Q displaystyle Q 1 BQ nbsp is diagonal It is easy to find that B displaystyle B nbsp is the rotation matrix which rotates counterclockwise by angle 8 p 2 textstyle theta frac pi 2 nbsp Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable How to diagonalize a matrix edit Diagonalizing a matrix is the same process as finding its eigenvalues and eigenvectors in the case that the eigenvectors form a basis For example consider the matrix A 0 1 2 0 1 0 1 1 3 displaystyle A left begin array rrr 0 amp 1 amp 2 0 amp 1 amp 0 1 amp 1 amp 3 end array right nbsp The roots of the characteristic polynomial p l det l I A displaystyle p lambda det lambda I A nbsp are the eigenvalues l 1 1 l 2 1 l 3 2 displaystyle lambda 1 1 lambda 2 1 lambda 3 2 nbsp Solving the linear system I A v 0 displaystyle left I A right mathbf v mathbf 0 nbsp gives the eigenvectors v 1 1 1 0 displaystyle mathbf v 1 1 1 0 nbsp and v 2 0 2 1 displaystyle mathbf v 2 0 2 1 nbsp while 2 I A v 0 displaystyle left 2I A right mathbf v mathbf 0 nbsp gives v 3 1 0 1 displaystyle mathbf v 3 1 0 1 nbsp that is A v i l i v i displaystyle A mathbf v i lambda i mathbf v i nbsp for i 1 2 3 displaystyle i 1 2 3 nbsp These vectors form a basis of V R 3 displaystyle V mathbb R 3 nbsp so we can assemble them as the column vectors of a change of basis matrix P displaystyle P nbsp to get P 1 A P 1 0 1 1 2 0 0 1 1 1 0 1 2 0 1 0 1 1 3 1 0 1 1 2 0 0 1 1 1 0 0 0 1 0 0 0 2 D displaystyle P 1 AP left begin array rrr 1 amp 0 amp 1 1 amp 2 amp 0 0 amp 1 amp 1 end array right 1 left begin array rrr 0 amp 1 amp 2 0 amp 1 amp 0 1 amp 1 amp 3 end array right left begin array rrr 1 amp 0 amp 1 1 amp 2 amp 0 0 amp 1 amp 1 end array right begin bmatrix 1 amp 0 amp 0 0 amp 1 amp 0 0 amp 0 amp 2 end bmatrix D nbsp We may see this equation in terms of transformations P displaystyle P nbsp takes the standard basis to the eigenbasis P e i v i displaystyle P mathbf e i mathbf v i nbsp so we have P 1 A P e i P 1 A v i P 1 l i v i l i e i displaystyle P 1 AP mathbf e i P 1 A mathbf v i P 1 lambda i mathbf v i lambda i mathbf e i nbsp so that P 1 A P displaystyle P 1 AP nbsp has the standard basis as its eigenvectors which is the defining property of D displaystyle D nbsp Note that there is no preferred order of the eigenvectors in P displaystyle P nbsp changing the order of the eigenvectors in P displaystyle P nbsp just changes the order of the eigenvalues in the diagonalized form of A displaystyle A nbsp 2 Application to matrix functions editDiagonalization can be used to efficiently compute the powers of a matrix A P D P 1 displaystyle A PDP 1 nbsp A k P D P 1 k P D P 1 P D P 1 P D P 1 P D P 1 P D P 1 P P 1 P D P 1 P D k P 1 displaystyle begin aligned A k amp left PDP 1 right k left PDP 1 right left PDP 1 right cdots left PDP 1 right amp PD left P 1 P right D left P 1 P right cdots left P 1 P right DP 1 PD k P 1 end aligned nbsp and the latter is easy to calculate since it only involves the powers of a diagonal matrix For example for the matrix A displaystyle A nbsp with eigenvalues l 1 1 2 displaystyle lambda 1 1 2 nbsp in the example above we compute A k P D k P 1 1 0 1 1 2 0 0 1 1 1 k 0 0 0 1 k 0 0 0 2 k 1 0 1 1 2 0 0 1 1 1 2 2 k 1 2 k 2 2 k 1 0 1 0 1 2 k 1 2 k 1 2 k 1 displaystyle begin aligned A k PD k P 1 amp left begin array rrr 1 amp 0 amp 1 1 amp 2 amp 0 0 amp 1 amp 1 end array right begin bmatrix 1 k amp 0 amp 0 0 amp 1 k amp 0 0 amp 0 amp 2 k end bmatrix left begin array rrr 1 amp 0 amp 1 1 amp 2 amp 0 0 amp 1 amp 1 end array right 1 1em amp begin bmatrix 2 2 k amp 1 2 k amp 2 2 k 1 0 amp 1 amp 0 1 2 k amp 1 2 k amp 1 2 k 1 end bmatrix end aligned nbsp This approach can be generalized to matrix exponential and other matrix functions that can be defined as power series For example defining exp A I A 1 2 A 2 1 3 A 3 textstyle exp A I A frac 1 2 A 2 frac 1 3 A 3 cdots nbsp we have exp A P exp D P 1 1 0 1 1 2 0 0 1 1 e 1 0 0 0 e 1 0 0 0 e 2 1 0 1 1 2 0 0 1 1 1 2 e e 2 e e 2 2 e 2 e 2 0 e 0 e e 2 e e 2 e 2 e 2 displaystyle begin aligned exp A P exp D P 1 amp left begin array rrr 1 amp 0 amp 1 1 amp 2 amp 0 0 amp 1 amp 1 end array right begin bmatrix e 1 amp 0 amp 0 0 amp e 1 amp 0 0 amp 0 amp e 2 end bmatrix left begin array rrr 1 amp 0 amp 1 1 amp 2 amp 0 0 amp 1 amp 1 end array right 1 1em amp begin bmatrix 2e e 2 amp e e 2 amp 2e 2e 2 0 amp e amp 0 e e 2 amp e e 2 amp e 2e 2 end bmatrix end aligned nbsp This is particularly useful in finding closed form expressions for terms of linear recursive sequences such as the Fibonacci numbers Particular application edit For example consider the following matrix M a b a 0 b displaystyle M begin bmatrix a amp b a 0 amp b end bmatrix nbsp Calculating the various powers of M displaystyle M nbsp reveals a surprising pattern M 2 a 2 b 2 a 2 0 b 2 M 3 a 3 b 3 a 3 0 b 3 M 4 a 4 b 4 a 4 0 b 4 displaystyle M 2 begin bmatrix a 2 amp b 2 a 2 0 amp b 2 end bmatrix quad M 3 begin bmatrix a 3 amp b 3 a 3 0 amp b 3 end bmatrix quad M 4 begin bmatrix a 4 amp b 4 a 4 0 amp b 4 end bmatrix quad ldots nbsp The above phenomenon can be explained by diagonalizing M displaystyle M nbsp To accomplish this we need a basis of R 2 displaystyle mathbb R 2 nbsp consisting of eigenvectors of M displaystyle M nbsp One such eigenvector basis is given by u 1 0 e 1 v 1 1 e 1 e 2 displaystyle mathbf u begin bmatrix 1 0 end bmatrix mathbf e 1 quad mathbf v begin bmatrix 1 1 end bmatrix mathbf e 1 mathbf e 2 nbsp where ei denotes the standard basis of Rn The reverse change of basis is given by e 1 u e 2 v u displaystyle mathbf e 1 mathbf u qquad mathbf e 2 mathbf v mathbf u nbsp Straightforward calculations show that M u a u M v b v displaystyle M mathbf u a mathbf u qquad M mathbf v b mathbf v nbsp Thus a and b are the eigenvalues corresponding to u and v respectively By linearity of matrix multiplication we have that M n u a n u M n v b n v displaystyle M n mathbf u a n mathbf u qquad M n mathbf v b n mathbf v nbsp Switching back to the standard basis we have M n e 1 M n u a n e 1 M n e 2 M n v u b n v a n u b n a n e 1 b n e 2 displaystyle begin aligned M n mathbf e 1 amp M n mathbf u a n mathbf e 1 M n mathbf e 2 amp M n left mathbf v mathbf u right b n mathbf v a n mathbf u left b n a n right mathbf e 1 b n mathbf e 2 end aligned nbsp The preceding relations expressed in matrix form are M n a n b n a n 0 b n displaystyle M n begin bmatrix a n amp b n a n 0 amp b n end bmatrix nbsp thereby explaining the above phenomenon Quantum mechanical application editIn quantum mechanical and quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes The basic reason is that the time independent Schrodinger equation is an eigenvalue equation albeit in most of the physical situations on an infinite dimensional Hilbert space A very common approximation is to truncate Hilbert space to finite dimension after which the Schrodinger equation can be formulated as an eigenvalue problem of a real symmetric or complex Hermitian matrix Formally this approximation is founded on the variational principle valid for Hamiltonians that are bounded from below First order perturbation theory also leads to matrix eigenvalue problem for degenerate states See also editDefective matrix Scaling geometry Triangular matrix Semisimple operator Diagonalizable group Jordan normal form Weight module associative algebra generalization Orthogonal diagonalizationNotes editReferences edit Horn Roger A Johnson Charles R 2013 Matrix Analysis second edition Cambridge University Press ISBN 9780521839402 Anton H Rorres C 22 Feb 2000 Elementary Linear Algebra Applications Version 8th ed John Wiley amp Sons ISBN 978 0 471 17052 5 Retrieved from https en wikipedia org w index php title Diagonalizable matrix amp oldid 1182135695 Simultaneous diagonalization, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.