Eigenvalues as the magnitude of the matrix
Eigenvalues are a notoriously hard topic to understand. Their definition can be unintuitive, making it difficult to understand their usefulness.
Recently, I've started explaining eigenvalues as the magnitude of a matrix, that generalizes the notion of "big" and "small" numbers.
Magnitude of real numbers
Take an arbitrary real number
As shwon above, we can actually say quite a lot about
How do we generalize this to matrices? Is it even useful, talking about the high power of a matrix?
Dynamical systems
Let's model the following very simple system: a car moving at a constant speed of
This can be written in the following matrix form:
If we call
As we can see, the behaviour of the car can be described by exponentiating the matrix
It turns out, that many physical models can be represented as a dynamical system (e.g. wheather predictions, planetary mechanics). Even if the update step is not linear, it can often be linearized or approximated. Therefore, the ability to efficiently analyze power of matrices is crucial. It turns out, eigenvalues will be a great help for us.
Eigenvalues for powers
Assume that the matrix
What does this form tell us? If all eigenvalues are between -1 and 1, then
Sometimes you have your eigenvalues from different magnitudes. Depending on your problem, it could be fine that the matrix blows up in one direction (the corresponding eigenvalue is larger than 1) while converges in other directions (the corresponding eigenvalues are between -1 and 1).
Note, that in the above derivation we assumed the matrix can be decomposed nicely. However, it turns out, even if the matrix does not have an eigendecomposition, we can still make these statements about its eigenvalues. If all eigenvalues of the matrix have an absolute value less than 1,
In summary, we can draw conclusions about the convergence of matrix powers, allowing us to better analyze and solve problems. These conclusions are analogous to those we can make about scalar numbers.