Eigenvalues as the magnitude of the matrix

| 3 min read

Eigenvalues are a notoriously hard topic to understand. Their definition can be unintuitive, making it difficult to understand their usefulness.

Recently, I've started explaining eigenvalues as the magnitude of a matrix, that generalizes the notion of "big" and "small" numbers.

Magnitude of real numbers

Take an arbitrary real number x. What is xn if n is large? If x=0 or 1 the answer is 0 or 1. If x>1 then xn is a large number. If 1<x<1 then xn will be close to 0.

As shwon above, we can actually say quite a lot about xn, without knowing what x actually is. We only had to know its sign and magnitude.

How do we generalize this to matrices? Is it even useful, talking about the high power of a matrix?

Dynamical systems

Let's model the following very simple system: a car moving at a constant speed of v, starting at position x0. After one second, the position is: x1=x0+v After two seconds: x2=x1+v

This can be written in the following matrix form:

[xnv]=[1101][xn1v]

If we call sn=[xnv], the expression can be further simplified as: sn=Asn1=AAsn2==Ans0, where A=[1101].

As we can see, the behaviour of the car can be described by exponentiating the matrix A. Of course, this is a rather simple example. However, many physical simulations can be written in this form, called a linear dynamical system. The linear part refers to the fact the update from sn1 to sn is a linear function (a matrix multiplication). Dynamical system means that sn only depends on the previous state sn1, and not on the step count n.

It turns out, that many physical models can be represented as a dynamical system (e.g. wheather predictions, planetary mechanics). Even if the update step is not linear, it can often be linearized or approximated. Therefore, the ability to efficiently analyze power of matrices is crucial. It turns out, eigenvalues will be a great help for us.

Eigenvalues for powers

Assume that the matrix A has an eigendecomposition, that is it can be converted to the form A=QDQ1, where D is a diagonal matrix holding the eigenvalues, and Q is a matrix whose columns hold the eigenvectors. Now we can write An=(QDQ1)(QDQ1)(QDQ1)=QDnQ1 The power Dn can be calculated quickly, by raising each eigenvalue to the power of n: Dn=[λ1λ2]n=[λ1nλ2n]

What does this form tell us? If all eigenvalues are between -1 and 1, then Dn will approach zero as n gets larger. Consequently, An will also approach zero. If A has an eigenvalue larger than 1, one column of Dn will go to infinity, thus An will have a direction where it blows up. Finally, if there is an eigenvalue less than -1, the matrix An will diverge. These statements are analogous to the statements we had on scalar numbers.

Sometimes you have your eigenvalues from different magnitudes. Depending on your problem, it could be fine that the matrix blows up in one direction (the corresponding eigenvalue is larger than 1) while converges in other directions (the corresponding eigenvalues are between -1 and 1).

Note, that in the above derivation we assumed the matrix can be decomposed nicely. However, it turns out, even if the matrix does not have an eigendecomposition, we can still make these statements about its eigenvalues. If all eigenvalues of the matrix have an absolute value less than 1, An converges to 0.

In summary, we can draw conclusions about the convergence of matrix powers, allowing us to better analyze and solve problems. These conclusions are analogous to those we can make about scalar numbers.