6.4. Complex Eigenvalues (and Eigenvectors)#

In the previous sections we hinted at the possibility to allow eigenvalues to be complex numbers. For an \(n\times n\) matrix \(A\) the eigenvalues are the zeros of the characteristic polynomial \(p_A(\lambda)\) of \(A\). Even if the matrix is real, these zeros may be complex. We start with an example to explore this until now unknown territory.

Example 6.4.1

Consider the matrix \(A = \left[\begin{array}{cc} 1 & -2 \\ 1 & 3 \end{array}\right]\).

We first compute the characteristic polynomial

\[\begin{split} p_A(\lambda) = \left|\begin{array}{cc} 1-\lambda & -2 \\ 1 & 3-\lambda \end{array}\right|= (1-\lambda)(3-\lambda) +2 = \lambda^2 - 4\lambda +5. \end{split}\]

Completing the square we see that

\[ p_A(\lambda) = 0 \,\, \iff \,\, \lambda^2 - 4\lambda +5 = (\lambda -2)^2 + 1 = 0 \,\, \iff \,\, \lambda = 2 \pm i. \]

So the eigenvalues are complex numbers. Can we still find eigenvectors? We sure can, but they will not be real vectors. To find an eigenvector for \(\lambda_1 = 2+i\), as before we have

\[ A\vect{v} = \lambda_1\vect{v} \quad\iff\quad (A - \lambda_1I)\vect{v} = \vect{0}. \]

So we have to solve a homogeneous system of linear equations, where now the coefficient matrix contains complex numbers. This slightly complicates the computation, but for this \(2 \times 2\) matrix things don’t get too bad.

\[\begin{split} \left[\begin{array}{cc|c} 1 - (2+i) & -2& 0 \\ 1 & 3- (2+i) & 0 \end{array}\right] = \left[\begin{array}{cc|c} -1 - i & -2& 0 \\ 1 & 1-i & 0 \end{array}\right] \sim \left[\begin{array}{cc|c} 0 & \class{blue}0 & 0 \\ 1 & 1-i & 0 \end{array}\right] , \end{split}\]

where the row operation we invoke is: add the second row \((1+i)\) times to the first row. The blue \class{blue}{0} is the result of the evaluation of

\[ -2 + (1+i)(1-i). \]

We can read off a solution (i.e., complex eigenvector):   \(\vect{v} = \left[\begin{array}{c} -1+i \\1 \end{array}\right]\).

To check that we have indeed an eigenvector is also slightly more involved than in the real case.

\[\begin{split} \left[\begin{array}{cc} 1 & -2 \\ 1 & 3 \end{array}\right] \left[\begin{array}{c} -1+i \\1 \end{array}\right] = \left[\begin{array}{c} -1+i-2 \\-1+i +3 \end{array}\right] = \left[\begin{array}{c} -3+i \\ 2+i \end{array}\right] \end{split}\]

and that agrees with

\[\begin{split} (2+i)\left[\begin{array}{c} -1+i \\1 \end{array}\right] = \left[\begin{array}{c} -3+i \\2+i \end{array}\right]. \end{split}\]

Alternatively, to simplify the augmented matrix

\[\begin{split} \left[\begin{array}{cc|c} -1 - i & -2& 0 \\ 1 & 1-i & 0 \end{array}\right] , \end{split}\]

we could have used the entry \(-2\) in the first row of the augmented matrix to create a zero in the second row. More specifically, if we add the first row \(\dfrac{1-i}{2}\) times to the second row, we get

\[\begin{split} \left[\begin{array}{cc|c} -1 - i & -2& 0 \\ 1 & 1-i & 0 \end{array}\right] \sim \left[\begin{array}{cc|c} -1-i & -2& 0 \\ 0 & 0 & 0 \end{array}\right] \sim \left[\begin{array}{cc|c} 1+i & 2& 0 \\ 0 & 0 & 0 \end{array}\right] . \end{split}\]

Then the vector \(\vect{u} = \left[\begin{array}{c} -2 \\ 1+i \end{array}\right] \) is a natural candidate for an eigenvector.

At first sight it seems that we have found two linearly independent eigenvectors for the eigenvalue \(\lambda_1 = 2+i\). However, closer inspection shows that

\[\begin{split} \vect{u} = \left[\begin{array}{c} -2 \\ 1+i \end{array}\right] = (1+i) \left[\begin{array}{c} -1+i \\1 \end{array}\right] = (1+i)\vect{v}, \end{split}\]

so the two vectors are complex multiples of each other, and hence are not linearly independent.

For the other eigenvalue we can proceed in the same manner and find (for instance) the eigenvector   \(\vect{v}_2 =\left[\begin{array}{c} -2 \\ 1-i \end{array}\right] \).

Note that the two eigenvalues are each others complex conjugate, and that the same holds for the corresponding eigenvectors, that is, if we define the complex conjugate of a vector component wise.

6.4.1. Vectors and matrices with complex entries#

We need a few definitions to settle matters a bit more formally. In the remainder of this section matrices (so in particular vectors) are allowed to have complex numbers as entries. If the entries are supposed to be real numbers we explicitly state this by speaking of a real matrix (or a real vector).

Definition 6.4.1

The complex conjugate \(\overline{A}\) of a matrix \(A\) is defined component wise:

if \(\quad A = \left[\begin{array}{ccc} a_{11} & \ldots & a_{1n} \\ \vdots & &\vdots \\ a_{m1} & \ldots & a_{mn} \end{array}\right] \quad\) then \(\quad\overline{A} = \left[\begin{array}{ccc} \overline{a_{11}} & \ldots & \overline{a_{1n}} \\ \vdots & &\vdots \\ \overline{a_{m1}} & \ldots & \overline{a_{mn}} \end{array}\right] \).

Proposition 6.4.1

  1. If \(A\) and \(B\) are two \(m\times n\) matrices, then \(\overline{A+B} = \overline{A}+\overline{B}\).

  2. If \(A\) and \(C\) are two matrices for which the product \(AC\) exists,
    then \(\overline{AC} = \overline{A}\) \(\overline{C}\).

Proof of Proposition 6.4.1

The statements follow immediately from the definitions of the sum and product of two matrices, and of the corresponding rules of complex arithmetic that say

\[ \overline{z+w} = \overline{z} + \overline{w} \quad\text{and} \quad \overline{z\cdot w} = \overline{z}\cdot\overline{w}, \quad \text{for all } z,w \in \C. \]

With this we can put the outcomes in Example 6.4.1 in a broader perspective.

Proposition 6.4.2

Suppose \(A\) is a real matrix, and \(\lambda = \alpha + \beta i\), with \(\beta \neq 0\), is an eigenvalue of \(A\). Then the following properties hold:

  1. \(\overline{\lambda} = \alpha - \beta i\) is an eigenvalue too.

  2. If \(\vect{v} = \vect{u}+i\vect{w}\), where \(\vect{u}\) and \(\vect{w}\) are real vectors, is an eigenvector for \(\lambda\), then \(\overline{\vect{v}} = \vect{u}-i\vect{w}\) is an eigenvector for \(\overline{\lambda}\).

Proof of Proposition 6.4.2

Suppose \(A(\vect{u}+i\vect{w}) = (\alpha + \beta i)(\vect{u}+i\vect{w})\).

Complex conjugation yields that

\[ \overline{A(\vect{u}+i\vect{w})} = \overline{A}\,\overline{(\vect{u}+i\vect{w})} = A(\vect{u}-i\vect{w}), \]

since \(A\) is a real matrix, and \(\vect{u}\) and \(\vect{v}\) are supposed to be real. At the same time

\[ \overline{A(\vect{u}+i\vect{w})} = \overline{(\alpha + \beta i)(\vect{u}+i\vect{w})} = (\alpha - \beta i)(\vect{u}-i\vect{w}). \]

Equating the last expressions of both derivations yields the desired result.

\[ A(\vect{u}-i\vect{w}) = (\alpha - \beta i)(\vect{u}-i\vect{w}). \]

This states exactly that \(A\) has the eigenvalue \(\overline{\lambda} = \alpha - \beta i\) with the corresponding eigenvector \(\vect{u}-i\vect{w}\).

Things look especially simple in the next example.

Example 6.4.2

Let \(A\) be the matrix \(A= \left[\begin{array}{cc} a & -b \\ b & a \end{array}\right]\), with \(b \neq 0\).

The characteristic polynomial of \(A\) is then given by \(p_A(\lambda) = (a-\lambda)^2 + b^2\).

So \(A\) has the complex eigenvalues

\[ \lambda_{1,2} = a \pm bi. \]

From row reduction of the augmented matrix

\[\begin{eqnarray*} [A - (a+bi)I | \vect{0}] &=& \left[\begin{array}{cc|c} a-(a+bi) & -b &0\\ b & a-(a+bi)&0 \end{array}\right]\\ &=& \left[\begin{array}{cc|c} -bi & -b&0 \\ b & -bi&0 \end{array}\right] \\ &\sim& \left[\begin{array}{cc|c} bi & b &0\\ 0 & 0&0 \end{array}\right], \end{eqnarray*}\]

we see that \(\vect{v} = \left[\begin{array}{c} 1 \\ -i \end{array}\right]\) is an eigenvector for \(\lambda_1 = a+bi\). By taking conjugates (according to Proposition 6.4.2) we can conclude that \(\vect{w} = \left[\begin{array}{c} 1 \\ i \end{array}\right]\), i.e. the conjugate of \(\vect{v}\), will be an eigenvector for \(\lambda_2= a-bi\).

There is a nice geometric interpretation for matrices of the form \(\left[\begin{array}{cc} a & -b \\ b & a \end{array}\right]\).

Or rather, for the corresponding linear transformation.

Proposition 6.4.3

The linear transformation \(T:\R^2\to\R^2\) given by

\[\begin{split} T(\vect{x}) = \left[\begin{array}{cc} a & -b \\ b & a \end{array}\right] \vect{x}, \quad a,b \in \R, \quad b \neq 0, \end{split}\]

can be described as a rotation followed by a “scaling”.

In fact, it holds that

\(A = \left[\begin{array}{cc} a & -b \\ b & a \end{array}\right] = r \left[\begin{array}{cc} \cos({\varphi}) & -\sin({\varphi}) \\ \sin({\varphi}) & \cos({\varphi}) \end{array}\right] \), for some \(r > 0\) and angle \(\varphi\).

Proof of Proposition 6.4.3

Both columns of \(A\) have length \(r = \sqrt{a^2 + (\pm b)^2} = \sqrt{a^2 + b^2}\).

If we take out this factor, we get

\[\begin{split} \left[\begin{array}{cc} a & -b \\ b & a \end{array}\right] = r \left[\begin{array}{cc} a/r & -b/r \\ b/r & a/r \end{array}\right] . \end{split}\]

The first column \(\left[\begin{array}{c} a/r \\ b/r \end{array}\right]\) has length one, so it lies on the unit circle, and hence must be of the form

\[\begin{split} \left[\begin{array}{c} a/r \\ b/r \end{array}\right] = \left[\begin{array}{c} \cos(\varphi) \\ \sin(\varphi) \end{array}\right] , \quad \text{for some angle } \varphi, \end{split}\]

and this makes that

\[\begin{split} \left[\begin{array}{cc} a & -b \\ b & a \end{array}\right] = r \left[\begin{array}{cc} a/r & -b/r \\ b/r & a/r \end{array}\right] = r \left[\begin{array}{cc} \cos(\varphi) & - \sin(\varphi)\\ \sin(\varphi) & \cos(\varphi) \end{array}\right] . \end{split}\]

So the ‘action’ of the matrix \(A = \left[\begin{array}{cc} a & -b \\ b & a \end{array}\right]\) on a vector \(\vect{x}\) is a (counterclockwise) rotation over the angle \(\varphi\) followed by a scaling (stretching/dilatation) with a factor \(r\).

The scaling factor \(r\) and the angle \(\varphi\) are exactly the polar coordinates of the eigenvalue \(\lambda = a + bi\). That is,

\[ a + bi = r(a/r + b/r i) = r (\cos(\varphi) + i \sin(\varphi)), \]

where

\[ r = |\lambda|, \quad \varphi = \text{arg}\,\lambda. \]

Proposition 6.4.3 can be generalized as follows. If a real \(n \times n\) matrix \(A\) has a non-real eigenvalue, there is always a rotation ‘hidden’ in the transformation \(T: \vect{x} \mapsto A\vect{x}\).

Proposition 6.4.4

Suppose the real \(n \times n\) matrix \(A\) has a complex eigenvalue \(\lambda = \alpha - \beta i\), with \(\beta \neq 0\). Then there exist two linearly independent real vectors \(\vect{u}\) and \(\vect{w}\) for which

\[\begin{split} \begin{cases} A\vect{u} = \alpha\vect{u} + \beta\vect{w} \\ A\vect{w} = -\beta\vect{u} + \alpha\vect{w}. \end{cases} \end{split}\]

That means that the two-dimensional subspace \(S = \Span{\vect{u},\vect{w}}\) is invariant under the linear transformation \(T\) that has \(A\) as a standard matrix.

Proof of Proposition 6.4.4

Let \(\vect{v}\) be an eigenvector for \(\lambda=\alpha - \beta i\). So, \(\vect{v} = \vect{u}+i\vect{w}\), for real vectors \(\vect{u}\) and \(\vect{w}\). Note that \(\vect{v}\) cannot be a real vector.

\(\vect{u}\) and \(\vect{w}\) must be linearly independent, because if \(\vect{w}= \vect{0}\), then \(\vect{v}\) would be a real eigenvector, and if \(\vect{u}= k\vect{w}\) for some \(k\) in \(\R\), then \(\vect{v} = (k+i)\vect{w}\), and then \(\vect{w}\) would be a real eigenvector for the complex eigenvalue \(\lambda\) of the real matrix \(A\). And that is unheard of.

So we conclude that \(\vect{u}\) and \(\vect{w}\) must be linearly independent.

We now have

\[ A(\vect{u}+i\vect{w}) = (\alpha - \beta i)(\vect{u}+i\vect{w}) = \alpha\vect{u}+\beta \vect{w} -i\beta\vect{u} + i\alpha\vect{w}. \]

Comparing real and imaginary parts we conclude that indeed

(6.4.1)#\[\begin{split}\begin{cases} A\vect{u} = \alpha\vect{u} + \beta\vect{w} \\ A\vect{w} = -\beta\vect{u} + \alpha\vect{w}. \end{cases}\end{split}\]

If we apply the above to the case \(n = 2\), we can rewrite equation (6.4.1) as

\[\begin{split} A [\,\vect{u}\,\, \vect{w}\,] = [\,\vect{u}\,\, \vect{w}\,] \left[\begin{array}{cc} \alpha & -\beta \\ \beta & \alpha \end{array}\right]. \end{split}\]

So if we define \(P\) to be the matrix \([\,\vect{u}\,\, \vect{w}\,]\) then we have \(AP = PC\), where

\[\begin{split} C = \left[\begin{array}{cc} \alpha & -\beta \\ \beta & \alpha \end{array}\right]. \end{split}\]

This more or less settles the following proposition.

Proposition 6.4.5

Suppose \(A\) is a \(2 \times 2\) matrix with eigenvalues \(\alpha \pm \beta i\), with \(\beta \neq 0\).
Then \(A\) can be written as

\[\begin{split} A = PCP^{-1} = P\left[\begin{array}{cc} \alpha & -\beta \\ \beta & \alpha \end{array}\right] P^{-1} \end{split}\]

for some invertible matrix \(P\).

Proposition 6.4.3 states that \(C\) can be written as

\[\begin{split} C = r\left[\begin{array}{cc} \cos(\varphi) & -\sin(\varphi) \\ \sin(\varphi) & \cos(\varphi) \end{array}\right] . \end{split}\]

We can formulate this as there being a hidden rotation in \(A\).

To be specific,

\[\begin{split} P = [ \,\vect{u}\,\, \vect{w}\, ] =\left[\begin{array}{cc}u_1 & w_1 \\ u_2&w_2 \end{array}\right], \end{split}\]

where \(\vect{u} + i \vect{w}\) is an eigenvector for \(\lambda = \alpha - \beta i\).
Let us illustrate matters with the following example.

Example 6.4.3

The matrix \(A = \left[\begin{array}{cc} 1 & -2 \\ 1 & 3 \end{array}\right]\) of Example 6.4.1 has the eigenvalues \(\lambda_{1,2} = \alpha \pm \beta i = 2 \pm i\). An eigenvector for \(2-i\) is given by

\[\begin{split} \vect{v} = \left[\begin{array}{c} -2 \\ 1-i \end{array}\right] = \left[\begin{array}{c} -2 \\ 1 \end{array}\right] + i\left[\begin{array}{c} 0 \\ -1 \end{array}\right]. \end{split}\]

Let us establish the identity

\[\begin{split} P^{-1}AP = \left[\begin{array}{cc} \alpha & -\beta \\ \beta & \alpha \end{array}\right] = \left[\begin{array}{cc} 2 & -1 \\ 1 & 2 \end{array}\right]. \end{split}\]

According to Proposition 6.4.5 we can take

\[\begin{split} P = \left[\begin{array}{cc} -2 & 0 \\ 1 & -1 \end{array}\right], \end{split}\]

with

\[\begin{split} P^{-1} = \dfrac12\left[\begin{array}{cc} -1 & 0 \\ -1 & -2 \end{array}\right]. \end{split}\]

We then get

\[\begin{split} \begin{array}{rcl} P^{-1}AP &=& \dfrac12\left[\begin{array}{cc} -1 & 0 \\ -1 & -2 \end{array}\right] \left[\begin{array}{cc} 1 & -2 \\ 1 & 3 \end{array}\right] \left[\begin{array}{cc} -2 & 0 \\ 1 & -1 \end{array}\right] \\ &=& \dfrac12\left[\begin{array}{cc} -1 & 2 \\ -3 & -4 \end{array}\right] \left[\begin{array}{cc} -2 & 0 \\ 1 & -1 \end{array}\right] \\ &=& \dfrac12 \left[\begin{array}{cc} 4 & -2 \\ 2 & 4 \end{array}\right] = \left[\begin{array}{cc} 2 & -1 \\ 1 & 2 \end{array}\right] = \left[\begin{array}{cc} \alpha & -\beta \\ \beta & \alpha \end{array}\right] . \end{array} \end{split}\]

We conclude this section by reconsidering diagonalizability of real matrices if we allow complex eigenvalues.

6.4.2. Complex diagonalizability#

Let us first generalize the definition:

Definition 6.4.2

A matrix is \(A\) is called complex diagonalizable if it can be written in the form

\[ A = PDP^{-1}. \]

where \(D\) is a diagonal matrix, and \(P\) and \(D\) may contain complex entries. We then say that \(PDP^{-1}\) is a diagonalization of \(A\).

Just like in the real case diagonalizability has all to do with the existence of enough (possibly complex) eigenvectors. The derivation is the same as in Section 6.3, we only repeat the conclusion.

Proposition 6.4.6

An \(n\times n\) matrix is (complex) diagonalizable if and only if there exists a basis of eigenvectors for \(\mathbb{C}^n\).

In that case, if \(\vect{v}_1, \ldots, \vect{v}_n\) are \(n\) linearly independent eigenvectors for the eigenvalues \(\lambda_1, \ldots, \lambda_n\), then

\[\begin{split} A = PDP^{-1} \quad \text{for} \quad D = \left[\begin{array}{cccc} \lambda_1 & 0 & \ldots & 0 \\ 0 & \lambda_2 & \ldots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \ldots & \lambda_n\end{array}\right], \quad P = [ \,\vect{v}_1\,\, \vect{v}_2 \, \ldots \, \vect{v}_n\,]. \end{split}\]

Example 6.4.4

The matrix \(A = \left[\begin{array}{cc} 1 & -2 \\ 1 & 3 \end{array}\right]\) of Example 6.4.1 has the eigenvalues \(\lambda_1 = 2 + i\), \(\lambda_2 = 2 - i\), with corresponding eigenvectors \(\vect{v}_1 = \left[\begin{array}{c} -1+i \\1 \end{array}\right]\), \(\vect{v}_2 = \left[\begin{array}{c} -1-i \\1 \end{array}\right]\).

It follows that

\[\begin{split} \left[\begin{array}{cc} 1 & -2 \\ 1 & 3 \end{array}\right] = \left[\begin{array}{cc} -1+i & -1-i \\1 &1 \end{array}\right] {\left[\begin{array}{cc} 2+i & 0 \\0 &2-i \end{array}\right] \left[\begin{array}{cc} -1+i & -1-i \\1 &1 \end{array}\right] }^{-1}. \end{split}\]

Which you are challenged to check by a careful calculation.

Remark 6.4.1

  • When the context requires it, we will specify whether we mean real diagonalizable or complex diagonalizable.

  • Since the real numbers are contained in the complex numbers, a matrix that is real diagonalizable is automatically complex diagonalizable.

  • The definition also makes sense for matrices with complex numbers as entries. However, we will not pursue that track.

For a matrix to be real diagonalizable Theorem 6.3.1 states that two conditions must be satisfied. One is that the characteristic polynomial of \(A\) must have \(n\) real roots, counting multiplicities. The second is that for each eigenvalue the geometric multiplicity must be equal to the algebraic multiplicity. The Fundamental Theorem of Algebra guarantees that each polynomial of degree \(n\) has \(n\) roots. So if we allow complex eigenvalues, the first condition is automatically satisfied. We thus find the following criterion for complex diagonalizability of a (possibly complex) matrix \(A\).

Proposition 6.4.7

A matrix \(A\) is complex diagonalizable if and only if for each eigenvalue the geometric multiplicity is equal to the algebraic multiplicity.

6.4.3. Grasple Exercises#

Grasple Exercise 6.4.1

https://embed.grasple.com/exercises/bf1aef1c-7948-4b32-951e-d79940282bfb?id=91545

Given a complex eigenvector of a \(2\times 2\) matrix, find the corresponding eigenvalue.

Grasple Exercise 6.4.2

https://embed.grasple.com/exercises/4e7c813d-ce37-42a2-be36-9dffb42d5f0b?id=91546

Given a complex eigenvector of a \(2\times 2\) matrix, find the corresponding eigenvalue.

Grasple Exercise 6.4.3

https://embed.grasple.com/exercises/66dcb059-1c37-477c-9f23-8bd1bb26fb44?id=92368

To find the (complex) eigenvalues of a \(2 \times 2\) matrix \(A\).

Grasple Exercise 6.4.4

https://embed.grasple.com/exercises/507921c2-d568-44d9-a254-536440ca613e?id=91553

To find the (complex) eigenvalues of a \(2 \times 2\) matrix \(A\).

Grasple Exercise 6.4.5

https://embed.grasple.com/exercises/593e9ddd-a7b8-4617-958e-95f328e28e80?id=91547

To find the (complex) eigenvalues and bases of their eigenspaces of a \(2 \times 2\) matrix \(A\).

Grasple Exercise 6.4.6

https://embed.grasple.com/exercises/f60a57c2-b0b3-45de-b18c-e0da1bbef601?id=91555

Given a complex eigenvector of a real 2x2 matrix, point out other eigenvectors.

Grasple Exercise 6.4.7

https://embed.grasple.com/exercises/440c03f9-9a5a-40a6-aa1a-77660a588b20?id=91556

To find an invertible matrix \(P\) and a scaling-rotation matrix \(C\) for which \(A = PCP^{-1}\).

Grasple Exercise 6.4.8

https://embed.grasple.com/exercises/f3a6972e-f17b-4984-b0ee-38012ec542b3?id=91563

To find the complex eigenvalues of a \(2 \times 2\) matrix (of a special form).

Grasple Exercise 6.4.9

https://embed.grasple.com/exercises/79c7c433-fbcc-41b8-85d2-b1938a26d40e?id=91566

T/F? Every real \(5 \times 5\) matrix has at least one real eigenvalue.

Grasple Exercise 6.4.10

https://embed.grasple.com/exercises/1720d806-7602-41b3-9beb-569967e74c84?id=92543

To find a complex diagonalization of a real \(2\times 2\) matrix (eigenvalues given).

Grasple Exercise 6.4.11

https://embed.grasple.com/exercises/510f78e6-717e-4550-8564-7bbf2c2cf673?id=55388

To find a complex diagonalization of a real \(3\times 3\) matrix (one eigenvalue given).