Summation by parts: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
 
Line 1: Line 1:
Obtain it in excel, copy-paste this continued plan to become corpuscle B1. Merchandise in your articles again access an bulk of time in abnormal located in corpuscle A1, the mass fast in treasures will start in B1.<br><br>
In [[linear algebra]], a [[square matrix]] ''A'' is called '''diagonalizable''' if it is [[similar (linear algebra)|similar]] to a [[diagonal matrix]], i.e., if there exists an [[invertible matrix]] ''P'' such that ''P''<sup>−1</sup>''AP'' is a diagonal matrix. If ''V'' is a finite-[[dimension (linear algebra)|dimension]]al [[vector space]], then a [[linear operator|linear map]] ''T'' : ''V'' → ''V'' is called '''diagonalizable''' if there exists an [[Basis (linear algebra)#Ordered bases and coordinates|ordered basis]] of ''V'' with respect to which ''T'' is represented by a diagonal matrix. '''Diagonalization''' is the process of finding a corresponding [[diagonal matrix]] for a diagonalizable matrix or linear map.<ref>Horn & Johnson 1985</ref> A square matrix which is not diagonalizable is called ''[[Defective matrix|defective]].''


when you are locating the latest handle system tough that will use, optimize the facilities within your activity. The default manage strategy might not be for everyone. Some an individual prefer a better glimpse screen, a set off more sensitive management also known as perhaps an inverted component. In several tutorial gaming, you may master these from the setting's area.<br><br>Be aware of how multi player works. In the case you're investing in the actual game exclusively for it has the multiplayer, be sure they have everything required to gain this. If you're planning on playing while fighting a person in your good household, you may discover that you will want two copies of the clash of clans cheats to game against one another.<br><br>Console game playing is suitable for kids. Consoles present you with far better control of content and safety, all of the kids can simply breeze by way of parent or guardian regulates on your internet. Using this step might help shield your young ones produced by harm.<br><br>Among the best and fastest harvesting certifications by [http://Www.Adobe.com/cfusion/search/index.cfm?term=&ECCouncil&loc=en_us&siteSection=home ECCouncil]. Where a dictionary invade fails the computer cyberpunk may try a incredible force attack, which is much more time consuming. Sets up the borders of all with non-editable flag: lot_border [ ]. The problem is this one hit a person where it really costs you anything - your heart. For those who have any queries about where by along with how you can utilize [http://prometeu.net clash of clans hack app], you possibly can contact us from the web site. These Kindle hacks would be keyboard shortcuts will save you tons of time on the search and typing in repeated things. Claire laughed and said how she had did start to gain a (not pointless.<br><br>You will see for yourself that your Money Compromise of Clans i fairly effective, understand invisible by the editor of the game, above all absolutely no price!<br><br>You'll find it's a nice technique. Breaking the appraisement bottomward into chunks of their whole time that accomplish college to be able so that you can bodies (hour/day/week) causes the problem to be accessible regarding visualize. Everybody knows what it appears like to accept to delay a day. May be additionally actual accessible for you to tune. If someone change your current apperception after and adjudge which will one day should bulk more, all you accusation to complete is change 1 value.
Diagonalizable matrices and maps are of interest because diagonal matrices are especially easy to handle: [[Transformation_matrix#Eigenbasis_and_diagonal_matrix|their eigenvalues and eigenvectors are known]] and one can raise a diagonal matrix to a power by simply raising the diagonal entries to that same power. Geometrically, a diagonalizable matrix is an '''[[inhomogeneous dilation]]''' (or ''anisotropic scaling'')&nbsp;— it [[Scaling (geometry)|scales]] the space, as does a ''[[homogeneous dilation]]'', but by a different factor in each direction, determined by the scale factors on each axis (diagonal entries).
 
== Characterization ==
The fundamental fact about diagonalizable maps and matrices is expressed by the following:
 
* An ''n''×''n'' matrix ''A'' over the [[field (mathematics)|field]] ''F'' is diagonalizable [[if and only if]] the sum of the [[dimension (linear algebra)|dimension]]s of its eigenspaces is equal to ''n'', which is the case if and only if there exists a [[basis (linear algebra)|basis]] of ''F''<sup>''n''</sup> consisting of eigenvectors of ''A''. If such a basis has been found, one can form the matrix ''P'' having these basis vectors as columns, and ''P''<sup>−1</sup>''AP'' will be a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of ''A''.
* A linear map ''T'' : ''V'' → ''V'' is diagonalizable if and only if the sum of the [[dimension (linear algebra)|dimension]]s of its eigenspaces is equal to dim(''V''), which is the case if and only if there exists a basis of ''V'' consisting of eigenvectors of ''T''. With respect to such a basis, ''T'' will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of ''T''.
 
Another characterization: A matrix or linear map is diagonalizable over the field ''F'' if and only if its [[minimal polynomial (linear algebra)|minimal polynomial]] is a product of distinct linear factors over ''F''. (Put in another way, a matrix is diagonalizable if and only if all of its [[elementary divisor]]s are linear.)
 
The following sufficient (but not necessary) condition is often useful.
 
* An ''n''×''n'' matrix ''A'' is diagonalizable over the field ''F'' if it has ''n'' distinct eigenvalues in ''F'', i.e. if its [[characteristic polynomial]] has ''n'' distinct roots in ''F''; however, the converse may be false. Let us consider
:: <math>\begin{bmatrix} -1 & 3 & -1 \\ -3 & 5 & -1 \\ -3 & 3 & 1 \end{bmatrix},</math>
: which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form ( [[similar (linear algebra)|similar]] to ''A'')
:: <math>\begin{bmatrix} 1 & 0 & 0 \\ 0 & 2 & 0 \\ 0 & 0 & 2 \end{bmatrix}</math>
: and [[change of basis|change of basis matrix]] ''P''
:: <math>\begin{bmatrix} 1 & 1 & -1 \\ 1 & 1 & 0 \\ 1 & 0 & 3 \end{bmatrix}.</math>
: The converse fails when ''A'' has an eigenspace of dimension higher than 1. In this example, the eigenspace of ''A'' associated with the eigenvalue 2 has dimension 2.
 
* A linear map ''T'' : ''V'' → ''V'' with ''n'' = dim(''V'') is diagonalizable if it has ''n'' distinct eigenvalues, i.e. if its characteristic polynomial has ''n'' distinct roots in ''F''.
 
Let ''A'' be a matrix over ''F''. If ''A'' is diagonalizable, then so is any power of it. Conversely, if ''A'' is invertible, ''F'' is algebraically closed, and ''A<sup>n</sup>'' is diagonalizable for some ''n'' that is not an integer multiple of the characteristic of ''F'', then ''A'' is diagonalizable. Proof: If ''A<sup>n</sup>'' is diagonalizable, then ''A'' is annihilated by some polynomial <math>(x^n - \lambda_1) \cdots (x^n - \lambda_k)</math>, which has no multiple root (since <math>\lambda_j \ne 0</math>) and is divided by the minimal polynomial of ''A''.
 
As a rule of thumb, over '''C''' almost every matrix is diagonalizable. More precisely: the set of complex ''n''×''n'' matrices that are ''not'' diagonalizable over '''C''', considered as a [[subset]] of '''C'''<sup>''n''×''n''</sup>, has [[Lebesgue measure]] zero. One can also say that the diagonalizable matrices form a dense subset with respect to the [[Zariski topology]]: the complement lies inside the set where the [[discriminant]] of the characteristic polynomial vanishes, which is a [[hypersurface]]. From that follows also density in the usual (''strong'') topology given by a [[norm (mathematics)|norm]]. The same is not true over '''R'''.
 
The [[Jordan–Chevalley decomposition]] expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its [[nilpotent]] part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its [[Jordan form]] has no nilpotent part; i.e., each "block" is a one-by-one matrix.
 
== Diagonalization ==
If a matrix ''A'' can be diagonalized, that is,
 
:<math>P^{-1}AP=\begin{pmatrix}\lambda_{1}\\
& \lambda_{2}\\
& & \ddots\\
& & & \lambda_{n}\end{pmatrix},</math>
 
then:
 
:<math>AP=P\begin{pmatrix}\lambda_{1}\\
& \lambda_{2}\\
& & \ddots\\
& & & \lambda_{n}\end{pmatrix} .</math>
 
Writing ''P'' as a [[block matrix]] of its column vectors <math>\vec{\alpha}_{i}</math>
 
:<math>P=\begin{pmatrix}\vec{\alpha}_{1} & \vec{\alpha}_{2} & \cdots & \vec{\alpha}_{n}\end{pmatrix},</math>
 
the above equation can be rewritten as
 
:<math>A\vec{\alpha}_{i}=\lambda_{i}\vec{\alpha}_{i}\qquad(i=1,2,\cdots,n).</math>
 
So the column vectors of ''P'' are [[right eigenvector]]s of ''A'', and the corresponding diagonal entry is the corresponding [[eigenvalue]]. The invertibility of ''P'' also suggests that the eigenvectors are [[linearly independent]] and form a basis of ''F''<sup>''n''</sup>. This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The row vectors of ''P''<sup>−1</sup> are the [[left eigenvector]]s of ''A''.
 
When the matrix ''A'' is a [[Hermitian matrix]] (resp. [[symmetric matrix]]), eigenvectors of ''A'' can be chosen to form an [[orthonormal basis]] of '''C'''<sup>''n''</sup> (resp. '''R'''<sup>''n''</sup>). Under such circumstance ''P'' will be a [[unitary matrix]] (resp. [[orthogonal matrix]]) and ''P''<sup>−1</sup> equals the [[conjugate transpose]] (resp. [[transpose]]) of ''P''.
 
== Simultaneous diagonalization ==
{{See also|Triangular matrix#Simultaneous triangularisability|l1=Simultaneous triangularisability|Weight (representation theory)|Positive definite matrix#Simultaneous_diagonalization|l3=Positive definite matrix}}
 
A set of matrices are said to be ''simultaneously diagonalisable'' if there exists a single invertible matrix ''P'' such that ''P''<sup>−1</sup>''AP'' is a diagonal matrix for every ''A'' in the set. The following theorem characterises simultaneously diagonalisable matrices: A set of diagonalizable [[Commuting matrices|matrices commutes]] if and only if the set is simultaneously diagonalisable.<ref>Horn & Johnson 1985, pp. 51–53</ref>
 
The set of all ''n''×''n'' diagonalisable matrices (over '''C''') with ''n'' > 1 is not simultaneously diagonalisable. For instance, the matrices
 
:<math> \begin{bmatrix} 1 & 0 \\ 0 & 0 \end{bmatrix} \quad\text{and}\quad \begin{bmatrix} 1 & 1 \\ 0 & 0 \end{bmatrix} </math>
 
are diagonalizable but not simultaneously diagonalizable because they do not commute.
 
A set consists of commuting [[normal matrix|normal matrices]] if and only if it is simultaneously diagonalisable by a [[unitary matrix]]; that is, there exists a unitary matrix ''U'' such that ''U*AU'' is diagonal for every ''A'' in the set.
 
In the language of [[Lie theory]], a set of simultaneously diagonalisable matrices generate a [[toral Lie algebra]].
 
== Examples ==
=== Diagonalizable matrices ===
* [[Involution (mathematics)|Involution]]s are diagonalisable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal
* Finite order [[endomorphism]]s are diagonalisable over '''C''' (or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with [[roots of unity]] on the diagonal. This follows since the minimal polynomial is [[separable polynomial|separable]], because the roots of unity are distinct.
* [[Projection (linear algebra)|Projections]] are diagonalizable, with 0's and 1's on the diagonal.
* Real [[symmetric matrices]] are diagonalizable by [[orthogonal matrix|orthogonal matrices]]; i.e., given a real symmetric matrix ''A'', ''Q<sup>T</sup>AQ'' is diagonal for some orthogonal matrix ''Q''. More generally, matrices are diagonalizable by [[unitary matrix|unitary matrices]] if and only if they are [[normal matrix|normal]]. In the case of the real symmetric matrix, we see that <math>A=A^T</math>, so clearly <math>AA^T=A^TA</math> holds. Examples of normal matrices are real  symmetric (or skew-symmetric) matrices (e.g. covariance matrices) and [[Hermitian matrix|Hermitian matrices]] (or skew-Hermitian matrices). See [[spectral theorem]]s for generalizations to infinite-dimensional vector spaces.
 
=== Matrices that are not diagonalizable ===
In general, a [[rotation matrix]] is not diagonalizable over the reals, but all [[rotation matrix#Independent planes|rotation matrices]] are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal - known as [[Jordan Normal Form|Jordan normal form]].
 
Some matrices are not diagonalizable over any field, most notably nonzero [[nilpotent matrix|nilpotent matrices]]. This happens more generally if the [[Eigenvalue,_eigenvector_and_eigenspace#Algebraic and geometric multiplicities|algebraic and geometric multiplicities]] of an eigenvalue do not coincide. For instance, consider
 
:<math> C = \begin{bmatrix} 0 & 1 \\ 0 & 0 \end{bmatrix}. </math>
 
This matrix is not diagonalizable: there is no matrix ''U'' such that ''U''<sup>−1</sup>''CU'' is a diagonal matrix. Indeed, ''C'' has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.
 
Some real matrices are not diagonalizable over the reals. Consider for instance the matrix
 
:<math> B = \begin{bmatrix} 0 & 1 \\ -1 & 0 \end{bmatrix}. </math>
 
The matrix ''B'' does not have any real eigenvalues, so there is no '''real''' matrix ''Q'' such that ''Q''<sup>−1</sup>''BQ'' is a diagonal matrix. However, we can diagonalize ''B'' if we allow complex numbers. Indeed, if we take
 
:<math> Q = \begin{bmatrix} 1 & \textrm{i} \\ \textrm{i} & 1 \end{bmatrix}, </math>
 
then ''Q''<sup>−1</sup>''BQ'' is diagonal.
 
Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.
 
=== How to diagonalize a matrix ===
Consider a matrix
 
:<math>A=\begin{bmatrix}
1& 2  & 0 \\
0 & 3  & 0 \\
2 & -4 & 2 \end{bmatrix}.</math>
 
This matrix has [[eigenvalues]]
 
: <math> \lambda_1 = 3, \quad \lambda_2 = 2, \quad \lambda_3= 1. </math>
 
''A'' is a 3×3 matrix with 3 different eigenvalues; therefore, it is diagonalizable. Note that if there are exactly ''n'' distinct eigenvalues in an ''n''×''n'' matrix then this matrix is diagonalizable.
 
These eigenvalues are the values that will appear in the diagonalized form of matrix ''A'', so by finding the eigenvalues of ''A'' we have diagonalized it. We could stop here, but it is a good check to use the [[eigenvectors]] to diagonalize ''A''.
 
The [[eigenvectors]] of ''A'' are
 
:<math>v_1 = \begin{bmatrix} -1 \\ -1 \\ 2 \end{bmatrix}, \quad v_2 = \begin{bmatrix} 0 \\ 0 \\ 1 \end{bmatrix}, \quad v_3 = \begin{bmatrix} -1 \\ 0 \\ 2 \end{bmatrix}.</math>
 
One can easily check that <math>A v_k = \lambda_k v_k.</math>
 
Now, let ''P'' be the matrix with these eigenvectors as its columns:
 
:<math>P= \begin{bmatrix}
-1 & 0 & -1 \\
-1 & 0  & 0 \\
2 & 1 & 2 \end{bmatrix}.</math>
 
Note there is no preferred order of the eigenvectors in ''P''; changing the order of the [[eigenvectors]] in ''P'' just changes the order of the [[eigenvalues]] in the diagonalized form of ''A''.<ref>{{cite book| last1=Anton |first1=H. |last2= Rorres|first2= C. |title=Elementary Linear Algebra (Applications Version) |publisher=John Wiley & Sons|edition=8th|date=22 Feb 2000| ISBN= 978-0-471-17052-5}}</ref>
 
Then ''P'' diagonalizes ''A'', as a simple computation confirms, having calculated ''P''<sup> −1</sup> using any [[Invertible_matrix#Methods_of_matrix_inversion|suitable method]]:
 
:<math>P^{-1}AP =
\begin{bmatrix}
0 & -1 & 0 \\
2 & 0  & 1 \\
-1 & 1 & 0 \end{bmatrix}
\begin{bmatrix}
1 & 2  & 0 \\
0 & 3  & 0 \\
2 & -4 & 2 \end{bmatrix}
\begin{bmatrix}
-1 & 0 & -1 \\
-1 & 0  & 0 \\
2 & 1 & 2 \end{bmatrix} =
\begin{bmatrix}
3 & 0 & 0 \\
0 & 2 & 0 \\
0 & 0 & 1\end{bmatrix}.</math>
 
Note that the eigenvalues <math>\lambda_k</math> appear in the diagonal matrix.
 
== An application ==
Diagonalization can be used to compute the powers of a matrix ''A'' efficiently, provided the matrix is diagonalizable. Suppose we have found that
 
:<math>P^{-1}AP = D</math>
 
is a diagonal matrix. Then, as the matrix product is associative,
 
:<math>\begin{align}
A^k &= (PDP^{-1})^k = (PDP^{-1}) \cdot (PDP^{-1}) \cdots (PDP^{-1}) \\
&= PD(P^{-1}P) D (P^{-1}P) \cdots (P^{-1}P) D P^{-1} \\
&= PD^kP^{-1} \end{align} </math>
 
and the latter is easy to calculate since it only involves the powers of a diagonal matrix. This approach can be generalized to [[matrix exponential]] and other [[matrix function]]s since they can be defined as power series.
 
This is particularly useful in finding closed form expressions for terms of [[linear recursive sequences]], such as the [[Fibonacci number]]s.
 
=== Particular application ===
For example, consider the following matrix:
 
:<math>M =\begin{bmatrix}a & b-a \\ 0 &b \end{bmatrix}.</math>
 
Calculating the various powers of ''M'' reveals a  surprising pattern:
 
:<math>M^2 = \begin{bmatrix}a^2 & b^2-a^2 \\ 0 &b^2 \end{bmatrix},\quad
M^3 = \begin{bmatrix}a^3 & b^3-a^3 \\ 0 &b^3 \end{bmatrix},\quad
M^4 = \begin{bmatrix}a^4 & b^4-a^4 \\ 0 &b^4 \end{bmatrix},\quad \ldots</math>
 
The above phenomenon can be explained by diagonalizing ''M''.  To accomplish this, we need a basis of '''R'''<sup>2</sup> consisting of eigenvectors of ''M''.  One such eigenvector basis is given by
 
:<math>\mathbf{u}=\begin{bmatrix} 1 \\ 0 \end{bmatrix}=\mathbf{e}_1,\quad \mathbf{v}=\begin{bmatrix} 1 \\ 1 \end{bmatrix}=\mathbf{e}_1+\mathbf{e}_2,</math>
 
where '''e'''<sub>i</sub> denotes the standard basis of '''R'''<sup>n</sup>. The reverse change of basis is given by
 
:<math> \mathbf{e}_1 = \mathbf{u},\qquad \mathbf{e}_2 = \mathbf{v}-\mathbf{u}.</math>
 
Straightforward calculations show that
 
:<math>M\mathbf{u} = a\mathbf{u},\qquad M\mathbf{v}=b\mathbf{v}.</math>
 
Thus, ''a'' and ''b'' are the eigenvalues corresponding to '''u''' and '''v''', respectively. By linearity of matrix multiplication, we have that
 
:<math> M^n \mathbf{u} = a^n\, \mathbf{u},\qquad M^n \mathbf{v}=b^n\,\mathbf{v}.</math>
 
Switching back to the standard basis, we have
 
:<math> M^n \mathbf{e}_1 = M^n \mathbf{u} = a^n \mathbf{e}_1,</math>
:<math> M^n \mathbf{e}_2 = M^n (\mathbf{v}-\mathbf{u}) = b^n \mathbf{v} - a^n\mathbf{u} = (b^n-a^n) \mathbf{e}_1+b^n\mathbf{e}_2.</math>
 
The preceding relations, expressed in matrix form, are
 
:<math> M^n = \begin{bmatrix}a^n & b^n-a^n \\ 0 &b^n \end{bmatrix}, </math>
 
thereby explaining the above phenomenon.
 
== Quantum mechanical application ==
In [[quantum mechanics|quantum mechanical]] and [[quantum chemistry|quantum chemical]] computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent [[Schrödinger equation]] is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional space (a [[Hilbert space]]). A very common approximation is to truncate Hilbert space to finite dimension, after which the  Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian, matrix. Formally this approximation is founded on the [[variational principle]], valid for Hamiltonians that are bounded from below.
But also first-order perturbation theory for degenerate states leads to a matrix eigenvalue problem.
 
== See also ==
* [[Defective matrix]]
* [[Scaling (geometry)]]
* [[Triangular matrix]]
* [[Semisimple operator]]
* [[Diagonalizable group]]
* [[Jordan normal form]]
* [[Weight module]]&nbsp;– associative algebra generalization
 
== Notes ==
{{Reflist}}
 
== References ==
* {{cite book | last1=Horn | first1=Roger A. | last2=Johnson | first2=Charles R. | title=Matrix Analysis | publisher= Cambridge University Press | isbn=978-0-521-38632-6 | year=1985}}
 
== External links ==
* {{planetmath reference|id=1960|title=Diagonalization}}
 
[[Category:Matrices]]

Revision as of 20:06, 11 October 2013

In linear algebra, a square matrix A is called diagonalizable if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P such that P−1AP is a diagonal matrix. If V is a finite-dimensional vector space, then a linear map T : VV is called diagonalizable if there exists an ordered basis of V with respect to which T is represented by a diagonal matrix. Diagonalization is the process of finding a corresponding diagonal matrix for a diagonalizable matrix or linear map.[1] A square matrix which is not diagonalizable is called defective.

Diagonalizable matrices and maps are of interest because diagonal matrices are especially easy to handle: their eigenvalues and eigenvectors are known and one can raise a diagonal matrix to a power by simply raising the diagonal entries to that same power. Geometrically, a diagonalizable matrix is an inhomogeneous dilation (or anisotropic scaling) — it scales the space, as does a homogeneous dilation, but by a different factor in each direction, determined by the scale factors on each axis (diagonal entries).

Characterization

The fundamental fact about diagonalizable maps and matrices is expressed by the following:

  • An n×n matrix A over the field F is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to n, which is the case if and only if there exists a basis of Fn consisting of eigenvectors of A. If such a basis has been found, one can form the matrix P having these basis vectors as columns, and P−1AP will be a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of A.
  • A linear map T : VV is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to dim(V), which is the case if and only if there exists a basis of V consisting of eigenvectors of T. With respect to such a basis, T will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of T.

Another characterization: A matrix or linear map is diagonalizable over the field F if and only if its minimal polynomial is a product of distinct linear factors over F. (Put in another way, a matrix is diagonalizable if and only if all of its elementary divisors are linear.)

The following sufficient (but not necessary) condition is often useful.

  • An n×n matrix A is diagonalizable over the field F if it has n distinct eigenvalues in F, i.e. if its characteristic polynomial has n distinct roots in F; however, the converse may be false. Let us consider
[131351331],
which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form ( similar to A)
[100020002]
and change of basis matrix P
[111110103].
The converse fails when A has an eigenspace of dimension higher than 1. In this example, the eigenspace of A associated with the eigenvalue 2 has dimension 2.
  • A linear map T : VV with n = dim(V) is diagonalizable if it has n distinct eigenvalues, i.e. if its characteristic polynomial has n distinct roots in F.

Let A be a matrix over F. If A is diagonalizable, then so is any power of it. Conversely, if A is invertible, F is algebraically closed, and An is diagonalizable for some n that is not an integer multiple of the characteristic of F, then A is diagonalizable. Proof: If An is diagonalizable, then A is annihilated by some polynomial (xnλ1)(xnλk), which has no multiple root (since λj0) and is divided by the minimal polynomial of A.

As a rule of thumb, over C almost every matrix is diagonalizable. More precisely: the set of complex n×n matrices that are not diagonalizable over C, considered as a subset of Cn×n, has Lebesgue measure zero. One can also say that the diagonalizable matrices form a dense subset with respect to the Zariski topology: the complement lies inside the set where the discriminant of the characteristic polynomial vanishes, which is a hypersurface. From that follows also density in the usual (strong) topology given by a norm. The same is not true over R.

The Jordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its nilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its Jordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.

Diagonalization

If a matrix A can be diagonalized, that is,

P1AP=(λ1λ2λn),

then:

AP=P(λ1λ2λn).

Writing P as a block matrix of its column vectors αi

P=(α1α2αn),

the above equation can be rewritten as

Aαi=λiαi(i=1,2,,n).

So the column vectors of P are right eigenvectors of A, and the corresponding diagonal entry is the corresponding eigenvalue. The invertibility of P also suggests that the eigenvectors are linearly independent and form a basis of Fn. This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The row vectors of P−1 are the left eigenvectors of A.

When the matrix A is a Hermitian matrix (resp. symmetric matrix), eigenvectors of A can be chosen to form an orthonormal basis of Cn (resp. Rn). Under such circumstance P will be a unitary matrix (resp. orthogonal matrix) and P−1 equals the conjugate transpose (resp. transpose) of P.

Simultaneous diagonalization

DTZ's public sale group in Singapore auctions all forms of residential, workplace and retail properties, outlets, homes, lodges, boarding homes, industrial buildings and development websites. Auctions are at present held as soon as a month.

We will not only get you a property at a rock-backside price but also in an space that you've got longed for. You simply must chill out back after giving us the accountability. We will assure you 100% satisfaction. Since we now have been working in the Singapore actual property market for a very long time, we know the place you may get the best property at the right price. You will also be extremely benefited by choosing us, as we may even let you know about the precise time to invest in the Singapore actual property market.

The Hexacube is offering new ec launch singapore business property for sale Singapore investors want to contemplate. Residents of the realm will likely appreciate that they'll customize the business area that they wish to purchase as properly. This venture represents one of the crucial expansive buildings offered in Singapore up to now. Many investors will possible want to try how they will customise the property that they do determine to buy by means of here. This location has offered folks the prospect that they should understand extra about how this course of can work as well.

Singapore has been beckoning to traders ever since the value of properties in Singapore started sky rocketing just a few years again. Many businesses have their places of work in Singapore and prefer to own their own workplace area within the country once they decide to have a everlasting office. Rentals in Singapore in the corporate sector can make sense for some time until a business has discovered a agency footing. Finding Commercial Property Singapore takes a variety of time and effort but might be very rewarding in the long term.

is changing into a rising pattern among Singaporeans as the standard of living is increasing over time and more Singaporeans have abundance of capital to invest on properties. Investing in the personal properties in Singapore I would like to applaud you for arising with such a book which covers the secrets and techniques and tips of among the profitable Singapore property buyers. I believe many novice investors will profit quite a bit from studying and making use of some of the tips shared by the gurus." – Woo Chee Hoe Special bonus for consumers of Secrets of Singapore Property Gurus Actually, I can't consider one other resource on the market that teaches you all the points above about Singapore property at such a low value. Can you? Condominium For Sale (D09) – Yong An Park For Lease

In 12 months 2013, c ommercial retails, shoebox residences and mass market properties continued to be the celebrities of the property market. Models are snapped up in report time and at document breaking prices. Builders are having fun with overwhelming demand and patrons need more. We feel that these segments of the property market are booming is a repercussion of the property cooling measures no.6 and no. 7. With additional buyer's stamp responsibility imposed on residential properties, buyers change their focus to commercial and industrial properties. I imagine every property purchasers need their property funding to understand in value.

A set of matrices are said to be simultaneously diagonalisable if there exists a single invertible matrix P such that P−1AP is a diagonal matrix for every A in the set. The following theorem characterises simultaneously diagonalisable matrices: A set of diagonalizable matrices commutes if and only if the set is simultaneously diagonalisable.[2]

The set of all n×n diagonalisable matrices (over C) with n > 1 is not simultaneously diagonalisable. For instance, the matrices

[1000]and[1100]

are diagonalizable but not simultaneously diagonalizable because they do not commute.

A set consists of commuting normal matrices if and only if it is simultaneously diagonalisable by a unitary matrix; that is, there exists a unitary matrix U such that U*AU is diagonal for every A in the set.

In the language of Lie theory, a set of simultaneously diagonalisable matrices generate a toral Lie algebra.

Examples

Diagonalizable matrices

  • Involutions are diagonalisable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal
  • Finite order endomorphisms are diagonalisable over C (or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with roots of unity on the diagonal. This follows since the minimal polynomial is separable, because the roots of unity are distinct.
  • Projections are diagonalizable, with 0's and 1's on the diagonal.
  • Real symmetric matrices are diagonalizable by orthogonal matrices; i.e., given a real symmetric matrix A, QTAQ is diagonal for some orthogonal matrix Q. More generally, matrices are diagonalizable by unitary matrices if and only if they are normal. In the case of the real symmetric matrix, we see that A=AT, so clearly AAT=ATA holds. Examples of normal matrices are real symmetric (or skew-symmetric) matrices (e.g. covariance matrices) and Hermitian matrices (or skew-Hermitian matrices). See spectral theorems for generalizations to infinite-dimensional vector spaces.

Matrices that are not diagonalizable

In general, a rotation matrix is not diagonalizable over the reals, but all rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal - known as Jordan normal form.

Some matrices are not diagonalizable over any field, most notably nonzero nilpotent matrices. This happens more generally if the algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider

C=[0100].

This matrix is not diagonalizable: there is no matrix U such that U−1CU is a diagonal matrix. Indeed, C has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.

Some real matrices are not diagonalizable over the reals. Consider for instance the matrix

B=[0110].

The matrix B does not have any real eigenvalues, so there is no real matrix Q such that Q−1BQ is a diagonal matrix. However, we can diagonalize B if we allow complex numbers. Indeed, if we take

Q=[1ii1],

then Q−1BQ is diagonal.

Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.

How to diagonalize a matrix

Consider a matrix

A=[120030242].

This matrix has eigenvalues

λ1=3,λ2=2,λ3=1.

A is a 3×3 matrix with 3 different eigenvalues; therefore, it is diagonalizable. Note that if there are exactly n distinct eigenvalues in an n×n matrix then this matrix is diagonalizable.

These eigenvalues are the values that will appear in the diagonalized form of matrix A, so by finding the eigenvalues of A we have diagonalized it. We could stop here, but it is a good check to use the eigenvectors to diagonalize A.

The eigenvectors of A are

v1=[112],v2=[001],v3=[102].

One can easily check that Avk=λkvk.

Now, let P be the matrix with these eigenvectors as its columns:

P=[101100212].

Note there is no preferred order of the eigenvectors in P; changing the order of the eigenvectors in P just changes the order of the eigenvalues in the diagonalized form of A.[3]

Then P diagonalizes A, as a simple computation confirms, having calculated P −1 using any suitable method:

P1AP=[010201110][120030242][101100212]=[300020001].

Note that the eigenvalues λk appear in the diagonal matrix.

An application

Diagonalization can be used to compute the powers of a matrix A efficiently, provided the matrix is diagonalizable. Suppose we have found that

P1AP=D

is a diagonal matrix. Then, as the matrix product is associative,

Ak=(PDP1)k=(PDP1)(PDP1)(PDP1)=PD(P1P)D(P1P)(P1P)DP1=PDkP1

and the latter is easy to calculate since it only involves the powers of a diagonal matrix. This approach can be generalized to matrix exponential and other matrix functions since they can be defined as power series.

This is particularly useful in finding closed form expressions for terms of linear recursive sequences, such as the Fibonacci numbers.

Particular application

For example, consider the following matrix:

M=[aba0b].

Calculating the various powers of M reveals a surprising pattern:

M2=[a2b2a20b2],M3=[a3b3a30b3],M4=[a4b4a40b4],

The above phenomenon can be explained by diagonalizing M. To accomplish this, we need a basis of R2 consisting of eigenvectors of M. One such eigenvector basis is given by

u=[10]=e1,v=[11]=e1+e2,

where ei denotes the standard basis of Rn. The reverse change of basis is given by

e1=u,e2=vu.

Straightforward calculations show that

Mu=au,Mv=bv.

Thus, a and b are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that

Mnu=anu,Mnv=bnv.

Switching back to the standard basis, we have

Mne1=Mnu=ane1,
Mne2=Mn(vu)=bnvanu=(bnan)e1+bne2.

The preceding relations, expressed in matrix form, are

Mn=[anbnan0bn],

thereby explaining the above phenomenon.

Quantum mechanical application

In quantum mechanical and quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent Schrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional space (a Hilbert space). A very common approximation is to truncate Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian, matrix. Formally this approximation is founded on the variational principle, valid for Hamiltonians that are bounded from below. But also first-order perturbation theory for degenerate states leads to a matrix eigenvalue problem.

See also

Notes

43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.

References

  • 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.

    My blog: http://www.primaboinca.com/view_profile.php?userid=5889534

External links

  1. Horn & Johnson 1985
  2. Horn & Johnson 1985, pp. 51–53
  3. 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.

    My blog: http://www.primaboinca.com/view_profile.php?userid=5889534