Why does the sum of eigenvalues equal to trace in terms of linear transformations

eigenvalues-eigenvectorslinear algebralinear-transformationstrace

While studying eigenvectors, I was confronted with two statements:

  • The product of the eigenvalues of some matrix $A$ is equal to the determinant of $A$

  • The trace of $A$ is equal to the sum of its eigenvalues

The thing is I am trying to understand every topic I learn in terms of linear transformations. For example, the first statement made sense to me because since the determinant is how much area scaled after a linear transformation and we are stretching $2$ vectors by their eigenvalues therefore determinant equals the product of eigenvalues.

But when I try to understand the second statement I can not relate the trace of a matrix and its eigenvalues because I can't also understand what the trace of a matrix tells us about a linear transformation.

I would love it if you can give me an intuitive explanation or let me know if there are topics that I haven't studied that prevent me from understanding this.

Best Answer

You can think of the trace geometrically as the derivative of the determinant at the identity. What this means is that for a fixed matrix $A$ and for small $\varepsilon > 0$ we have

$$\det(I + \varepsilon A) = 1 + \varepsilon \text{ tr}(A) + O(\varepsilon^2).$$

This isn't hard to prove directly from the Leibniz formula for the determinant. On the other hand, if $A$ has eigenvalues $\lambda_i$, then $I + \varepsilon A$ has eigenvalues $1 + \varepsilon \lambda_i$, and the product of these is

$$\det(I + \varepsilon A) = \prod_{i=1}^n (1 + \varepsilon \lambda_i) = 1 + \varepsilon \sum_{i=1}^n \lambda_i + O(\varepsilon^2).$$

So we see that $\text{tr}(A) = \sum_{i=1}^n \lambda_i$ is the sum of the eigenvalues as desired. In other words, the fact that the trace is the sum of the eigenvalues is the derivative of the fact that the determinant is the product of the eigenvalues.