Linear Algebra Help – Identity of a Matrix and finding a basis of the Kernel of that Matrix

linear algebra

Edit: Oops I had the wrong link, I've updated it

Edit: I mistakenly entered some info about the matrix $A$ (it is actually $c$ x $c$), that might have thrown others off. Also, it may not be that all columns add to zero. If one adds and obtains a negative value, I believe this would be like a chemical "sink", and thus allowable. I don't believe there is mention of it being a "stochastic" matrix, although this would fit the trend of their papers, and why I wrote it in the first place. Especially because the examples they provide have columns summing to $0$, I put it.

I've been reading this paper Flows, scaling, and the control of moment hierarchies for stochastic chemical reaction networks, and have come to a linear algebra issue I'm not quite understanding.

Say you have a matrix $A$ which is an $c \times c$ matrix, and another matrix $Y$ which is $p \times c$. Matrix $A$ has the property that if you sum over each columnn, the sum is $0$ (This seems to hold, but may not be true, see note above).

Now following this paper (section IVA, pages of interest are right-side of page 10),

Let $\left\{e_{\alpha}\right\}_{\alpha=1}^s$ be a basis for ker$(YA)^{\perp} \subseteq \textbf{R}^c$.

Let $\left\{\tilde{e}_{\beta} \right\}_{\beta=1}^{\delta}$ be a basis for ker$(YA)/$ker$(A)$.

It follows that jointly $\left\{\left\{e_{\alpha}\right\}_{\alpha=1}^s,\left\{\tilde{e}_{\beta} \right\}_{\beta=1}^{\delta}\right\}$ form a basis for ker$(A^{\perp}) \subseteq \textbf{R}^c$.

How is this collection a basis for ker$(A^{\perp})$?

Furthermore, they argue that

$$
A
= A
\left\{
\sum_{\alpha=1}^{s} e_{\alpha} e_{\alpha}^T
+ \sum_{\beta=1}^{\delta} \tilde{e}_{\beta} \tilde{e}_{\beta}^T
\right\}
$$

in the following equation $(40)$ on beginning of page 11. I'm not sure I see exactly why this is the case. Why is this so?

Best Answer

(Answering own question)

Here we are given that

$A: W \to V$ and $Y: V \to U$, and the composition is then $A \circ Y : W \to U$.

From an earlier part (of the paper, the section on the left side of page 10), we see that

$$\textrm{dim(Im}(A)) = \textrm{dim(Im}(YA)) + \textrm{dim(Ker(}Y) \cap \textrm{Im}(A))$$

Look at the last term; it tells us that vectors that are in Ker$(Y)$ that are also in $Im(A)$ are of interest. Now look at middle term; it tells us that vectors that are in $Im(A)$ and then in $Im(Y)$ are of interest. This gives us all of $Im(A)$. Hence the equality of dimensions.

Now taking a similar approach for the basis question, we can first see that because row space and image are interchangeable (correct me if I'm wrong), that Ker($YA)^{\perp} = $RowSpace$(YA) =$ Im$(YA)$. So now we just need to show that Ker$(YA)/$Ker$(A) =$ Ker$(Y)$ $\cap$ Im$(A)$. Observe that Ker($YA)$ will have two kinds of vectors: they will be (1) those from Ker($A$), because these map to the zero vector in $V$ space, so are taken to zero vector in $U$ space via linear map $Y$, and (2) those from Ker($Y$), because they map to the zero vector in $U$ by definition. The quotient vector space Ker($YA)$/Ker($A$) has coset elements, treating elements in Ker($A)$ together as the identity coset.

Ker($Y) \cap$ Im$(A)$ is pretty much Ker($Y$). Rank-nullity theorem tells us there are dim$(V)$ - dim(Im($A$)) vectors in Ker$(A$). So then we know that there are dim(Ker($YA$))/(dim$(V)$-dim(Im($A$))) elements in the quotient space, by Lagrange's theorem. This tells us that, for example, if dim(Ker($A)) = 1$ (zero vector only), then each coset has one element (because Im$(A$) spans whole vector space $V$, everything in Ker($Y$) is in Im$(A$)). If there is more than one vector in the kernel of $A$, say $n$, then the image doesn't span whole vector space, so the cosets have $n$ elements each (equality of coset sizes is another group theory theorem). Regardless, it tells how many vectors of Ker$(Y)$ are in the span of Im$(A)$, and gives these as representatives (to us as basis vectors). So Ker$(YA)/$Ker$(A)$ tells us the vectors of Ker$(Y)$ that are in the span of Im$(A)$. So Ker$(YA)/$Ker$(A) \subseteq $Ker$(Y) \cap $Im$(A)$. The intersection Ker$(Y)$ $\cap$ Im$(A)$ basically tells us the same thing, backwards: what is the region of vector space that satisfies the intersection? Then when we find an orthonormal basis for that region, we will have found what would be coset representatives of Ker$(YA)/$Ker$(A)$ (I know that was very hand-wavy, but it works for me at the moment). So we can see how Ker$(Y) \cap $Im$(A) \subseteq $Ker$(YA)/$Im$(A)$, and thus Ker$(YA) \cap $Im$(A) = $Ker$(YA)/$Im$(A)$. So we have that the bases of Ker$(YA)^{\perp}$ and Ker$(YA)$/Ker($A$) together give a basis for ker($A)^{\perp} = $Im$(A)$.

For part two:

Note that given matrix $M$, and a complete orthonormal basis $e_i$ ($e_i$ will be a column vector, and $e_i^T$ a row vector), we can put $M$ into a new basis by multiplying it by the sum of the dyads of $e_i$:

$$M = (M\cdot e_1)e_1^T + (M\cdot e_2)e_2^T + ... + (M\cdot e_n)e_n^T = \sum_{i=1}^n(M\cdot e_i)e_i^T,$$

where $n$ is the dimension of the vector space of interest (correct me if I am wrong about that part). This should look familiar from the question. Now observe that if we are trying to find a complete orthonormal basis for $M$, one way to do so could be using the basis of Ker$(M)$ and Ker$(M)^{\perp} = $Im$(M)$. Let $f_i$ be basis vectors of Ker$(M)$ and $e_i$ be basis vectors of Im$(M)$, then

$$M = \sum_{i=1}^{\textrm{dim(Im}(M))}(M\cdot e_i)e_i^T + \sum_{j=1}^{\textrm{dim(Ker}(M))}(M\cdot f_j)f_j^T.$$

Notice though that for all $v \in \textrm{Ker}(M), M\cdot v = \textbf{0}$, so

$$M = \sum_{i=1}^{\textrm{dim(Im}(M))}(M\cdot e_i)e_i^T + \sum_{j=1}^{\textrm{dim(Ker}(M))}(M\cdot f_j)f_j^T = \sum_{i=1}^{\textrm{dim(Im}(M))}(M\cdot e_i)e_i^T + \sum_{j=1}^{\textrm{dim(Ker}(M))}(\textbf{0})f_j^T = \sum_{i=1}^{\textrm{dim(Im}(M))}(M\cdot e_i)e_i^T,$$

and

$$M = \sum_{i=1}^{\textrm{dim(Im}(M))}(M\cdot e_i)e_i^T = M\sum_{i=1}^{\textrm{dim(Im}(M))}e_ie_i^T.$$

Thus, if $\left\{e_{\alpha},\tilde{e}_{\beta}\right\}$ form a basis for Im$(M)$, then

$$M = M\left\{\sum_{\alpha=1}^se_{\alpha}e_{\alpha}^T + \sum_{\beta=1}^{\delta}\tilde{e}_{\beta}\tilde{e}_{\beta}^T\right\}.$$

Edit: I think this is a pretty cool result! Generally, given a matrix $M$ (perhaps has to be square?) if $\left\{e_i : 1 \leq i \leq \textrm{dim(Im}(M)\right\}$ is an orthonormal basis for Im$(M)$, then $M = M\sum_ie_ie_i^T$.

Related Question