[Math] Matrix Determinant Identity

determinantlinear algebramatrices

I have come across an observation about the determinant of a matrix, but I don't know how to prove it in general. Let me demonstrate it through an example.

$$
\begin{align}
\left|
\begin{matrix}
1 & 2 & 3 & 4 \\
5 & 6 & 7 & 8 \\
9 & 10 & 11 & 12 \\
13 & 14 & 15 & 16 \\
\end{matrix}
\right|
\cdot
\left|
\begin{matrix}
6 & 7 \\
10 & 11 \\
\end{matrix}
\right|
&=
\left|
\begin{matrix}
1 & 2 & 3 \\
5 & 6 & 7 \\
9 & 10 & 11 \\
\end{matrix}
\right|
\cdot\left|
\begin{matrix}
6 & 7 & 8 \\
10 & 11 & 12 \\
14 & 15 & 16 \\
\end{matrix}
\right|\\
&-
\left|
\begin{matrix}
2 & 3 & 4 \\
6 & 7 & 8 \\
10 & 11 & 12 \\
\end{matrix}
\right|
\cdot
\left|
\begin{matrix}
5 & 6 & 7 \\
9 & 10 & 11 \\
13 & 14 & 15 \\
\end{matrix}
\right|
\end{align}
$$

What this essentially says is that you can find the determinant of a larger matrix by breaking it down into this "ad-bc"-style of smaller determinants, however I can't work out why this should be true. I've expanded out the determinant for a 3×3 example in full generality and confirmed that it works, but I'd like to know if there's a more straightforward reason for this observation.

Best Answer

Let's step back and ask what the determinant is and why it is useful. Then we'll get to this particular property.

Permit me to introduce a new product of vectors, called the wedge product. If $a, b$ are vectors, then their wedge product is $a \wedge b$. The result is not a vector, but instead, we interpret it geometrically as an oriented plane--exactly the plane perpendicular to $a \times b$, as a matter of fact, but we can continue wedging--for instance, forming $a \wedge b \wedge c$, which represents a volume.

Of course, in 3d space, there is only one unit volume (you could call it $\hat x \wedge \hat y \wedge \hat z$, but let me call it $i$ for short. All other volumes are scalar multiples of this volume, at least in terms of magnitude and orientation (you may be thinking "orientation?" I submit that a coordinate system following the right-hand rule is oppositedly oriented to one following a left-hand rule, and the unit volumes formed by wedging their unit vectors are oppositely oriented). It is for this reason that the volume elements are often called pseudoscalars, because they are only different from scalars by how they have orientations where scalars do not.

How does this relate to matrices? Well, matrices are used to represent linear operators on vectors, but these operators can act on wedge products of vectors or pseudoscalars too. The "matrices" you use to represent these extensions of the original operator are different from the original matrix. We define the relationships by a simple rule. If $\underline T$ is a linear operator on a vector, then we define $\underline T(a \wedge b) = \underline T(a) \wedge \underline T(b)$, and so on (note that the cross product does not have this nice property except under rotations).

But again, we said that there is only one unit pseudoscalar ($i$), so the action of a linear operator on $i$ must be some scalar multiple of $i$. That is, if $\alpha$ is a scalar,

$$\underline T(i) = \alpha i$$

We define this number $\alpha$ to be the determinant, telling us geometrically how the unit volume is shrunk or dilated (or changes orientation) under the action of a linear operator.

You can find the determinant of a linear operator by writing down the matrix representation and wedging the vectors that appear there. This is perfectly well-founded. You just need to know that $a \wedge b = - b \wedge a$--vectors anticommute under the wedge--and that the wedge is associative. Knowing these properties makes it possible to do computations with the wedge.

(If you're puzzled why wedging vectors gives the determinant, feel free to ask and I'll clarify this point.)

So let's take three vectors $f, g, h$ that appear in the matrix representation of a linear operator and wedge them to find the determinant. Let $f = f^x e_x + f^y e_y + f^z e_z$ and so on. We can then write out the following:

$$\begin{align*} f \wedge g \wedge h = f^x e_x \wedge (g \wedge h) + f^y e_y \wedge (g \wedge h) + f^z e_z \wedge (g \wedge h)\end{align*}$$

I've expanded the wedge product through linearity (the distributive property, which I failed to mention earlier but is also valid for the wedge). This is the foundation for the technique you'e come across, which is called Laplace expansion, or cofactor expansion, or expansion by minors. The antisymmetry of the wedge means that we can write $f^x e_x \wedge (g \wedge h)$ as

$$f^x e_x \wedge (g \wedge h) = f^x e_x \wedge (g^y e_y + g^z e_z) \wedge (h^y e_y + h^z e_z)$$

Why? Because $e_x \wedge e_x = 0$ always, so if any $e_x$ appeared in $g$ or $h$, they would be irrelevant. We can ignore them, and instead, we find the "determinant" of a linear operator on the $yz$-plane. Thus, the method of expansion by minors follows by recursively expanding a single vector through linearity to reduce the subsequent wedge products to a wedge product (a "determinant") you might already know. The method has its roots in the relationship between the determinant and volumes, how all the terms that make up the determinant must contain 1 and only 1 component of some vector from each of the coordinate directions--no more, no less.

Related Question