A Different Perspective of Diagonalization

An (atttempt at an) intuitive approach to similar matrix decomposition.

Diagonalization

While this works for any dimension matrix, this will be of the example where the matrix is a 2×2.

Suppose A is a 2×2 matrix with eigenvectors v1,v2 associated with eigenvalues λ1,λ2 respectively. By definition of an eigenvector, this would imply that for any scalars c1,c2.

(1)A(c1v1+c2v2)=λ1c1v1+λ2c2v2

Say we apply A a second time…

A(A(c1v1+c2v2))=A((λ1c1)v1+(λ2c2)v2)A2(c1v1+c2v2)=λ1(λ1c1)v1+λ2(λ2c2)v2A2(c1v1+c2v2)=λ12c1v1+λ22c2v2

We could easily repeat this as many times as we wanted. So it is also the case that

(2)An(c1v1+c2v2)=λ1nc1v1+λ2nc2v2

Since for those vectors, the transformation A is just simple scaling, in the situation where we have to repeatedly apply A to some vector w, it would clearly be preferable to get that vector in terms of v1 and v2. This seems like a job for change of basis…

So let us define the eigenbasis (3)B={v1,v2}

The change of basis matrix PεB (where ε is the standard basis ε={e1,e2}), then is

(4)PεB=(v1v2)

By extension,

(5)PBε=(PεB)1

So now suppose that

w=c1v1+c2v2

This tells us that the coordinate vector of w with respect to the eigenbasis B, (w)B, is

(6)w=c1v1+c2v2(w)B=(c1c2)

We know from (2) then that

Anw=An(c1v1+c2v2)=λ1nc1v1+λ2nc2v2

So

(Anw)B=(λ1nc1λ2nc2)

But that’s also the transformation

(Anw)B=(λ1nc1λ2nc2)=(λ1n00λ2n)(c1c2)

If we call the diagonal matrix D=(λ100λ2), then we get

(7)(Anw)B=Dn(w)B

We’re so close now. Next, we change back to the standard basis by applying PεB to both sides,

PεB(Anw)B=PεBDn(w)B

(8)Anw=PεBDn(w)B

To get things entirely in terms of the standard basis, we start by rewriting

(w)B=PBε(w)ε

Using (5) and the fact that (w)ε=w (by definition of the standard basis),

(9)(w)B=(PεB)1w

we substitute into (8)

Anw=PεBDn(PεB)1w

If we just denote PεB as P, then we finally get

(10)Anw=PDnP1w

General Decomposition

In general, if we know how a matrix A acts on a basis, then we may either construct or decompose it via a PDP1 decomposition.

Complex Eigendecomposition of Real Matrices

Let’s take the example of a real 2×2 with complex eigenvalues. Suppose A is such a matrix with a complex eigenvalue a+bi, where b0, associated with a complex eigenvector v.

(11)Av=(a+bi)v

Let’s get a bit more specific by decomposing v into its real and imaginary parts.

A(Re(v)+iIm(v))=(a+bi)(Re(v)+iIm(v))ARe(v)+i(AIm(v))=(aRe(v)bIm(v))+i(bRe(v)+aIm(v))

We can equate the real and imaginary parts on both sides,

(12)ARe(v)=aRe(v)bIm(v),AIm(v)=bRe(v)+aIm(v)

Now we know that {Re(v),Im(v)} is a basis of R2 because if they were linearly dependent, then Im(v)=kRe(v). That would change (11) into

A(1+ki)Re(v)=(a+bi)(1+ki)Re(v)

Dividing by the common 1+ki,

ARe(v)=(a+bi)Re(v)

This implies a contradiction since the imaginary part of the left side is zero due to A being real, but the right is not if b0. Therefore,

(13)B={Re(v),Im(v)} is a basis for R2

It then follows that

(14)P=(Re(v)Im(v))

is invertible.

Well, we know how A acts on the basis B (13):

(ARe(v))B=(ab),(AIm(v))B=(ba)

So our matrix D is then,

(15)D=(abba)

And our eigendecomposition of A is

(16)A=(Re(v)Im(v))(abba)(Re(v)Im(v))1

It can be verified that the eigenvalues of this particular D are λ=a±bi.

Footnotes