Characterizing the dual cone of the squares of skew-symmetric matrices

dual-conelinear algebramatricesmatrix-calculussymmetric matrices

Let $X$ be the set of all real $n \times n$ diagonal matrices $D$ satisfying $\langle D,B^2 \rangle \le 0$ for any (real) skew-symmetric matrix $B$. (I am using the Frobenius Euclidean product here).

$X$ is a convex cone.

Can we give an explicit characterization of $X$?

Comment:

If we denote by $C$ the space of all squares of skew-symmetric matrices, we can characterize its dual cone as follows:

Since every square of a skew-symmetric matrix is symmetric, and the symmetric and the skew-symmetric matrices are orthogonal, we know that every skew-symmetric matrix belongs to the dual cone of $C$. So, the question whether a given matrix $A$ belongs to the dual cone of $C$ depends solely on the symmetric part of $A$. Since $C$ is invariant under orthogonal conjugation, we can orthogonally diagonalize $\text{sym}(A)$ and deduce that $A$ lies in $C^*$ if and only if the diagonal matrix whose entries are the eigenvalues of $\text{sym}(A)$ is in $C^*$. Thus, the question reduces to determining the case of diagonal matrices.

Edit:

Omnomnomnom proved in this answer that every $D$ in $X$ has at most one negative entry, and the absolute value of the negative entry is less than or equal to the next-smallest entry.

I have a strangely complicated proof for the converse, namely I can prove that every diagonal matrix satisfying the condition above is in $X$.

I would like to find a "direct" proof based on linear algebra\matrix analysis. (my proof is based on rather convoluted variational considerations).

Best Answer

Claim: $D$ has at most one negative eigenvalue, and the absolute value of the negative eigenvalue is less than or equal to the next-smallest eigenvalue.

Proof: Let $E_{ij}$ denote the matrix with a $1$ in the $i,j$ entry and zeros elsewhere.

It suffices to show that if the $i$th and $j$th diagonal entries of $D$ have a negative sum, then $D$ cannot satisfy the criterion. To that end, it suffices to note that there exists a skew-symmetric matrix with $B^2 = -(E_{ii} + E_{jj})$ (take $B = E_{ij} - E_{ji}$ for instance). $\square$

I am not sure whether this condition is equivalent to your inequality.


We can also prove that the condition above is sufficient as follows. Suppose that $D$ has at most one negative eigenvalue, and the absolute value of the negative eigenvalue is less than or equal to the next-smallest eigenvalue.

We first note that every matrix of the form $M = B^2$ for a skew-symmetric $B$ can be written in the form $$ M = -[a_1 \, (x_1x_1^T + y_1y_1^T) + \cdots + a_k \, (x_kx_k^T + y_ky_k^T)]. $$ where the coefficients $a_i$ are non-negative and $x_i,y_i$ are a pair of orthonormal unit vectors for all $i$. So, it suffices to show that $\langle D,M\rangle \leq 0$ where $M = -(xx^T + yy^T)$ for some orthonormal $x,y$.

Now, let $v_1,\dots,v_n$ be an orthonormal basis for $\Bbb R^n$ such that $x = v_1$ and $y = v_2$. Let $V$ be the orthogonal matrix whose columns are $v_1,\dots,v_n$, and let $A = V^TDV$. We now note that $$ \langle D, xx^T + yy^T \rangle = x^TDx + y^TDy = a_{11} + a_{22}. $$ From here, it suffices to apply the $(\implies)$ direction of the Schur-Horn theorem to $-A$ in order to conclude that $a_{11} + a_{22} \geq \lambda_{n}(D) + \lambda_{n-1}(D)$.


About the squares of skew-symmetric matrices: by the spectral theorem, there exists a unitary $U$ with columns $u_1,u_2,\dots,u_n$ such that $$ B = U \pmatrix{i \lambda_1 \\ & - i\lambda_1 \\ && \ddots \\ &&& i \lambda_k \\ &&&& - i \lambda_k \\ &&&&& 0 } U^* \\ = \lambda_1 i \ [u_1u_1^* - u_2 u_2^*] + \cdots + i\lambda_{k}\ [u_{2k-1}u_{2k-1}^* - u_{2k}u_{2k}^*] $$ where each $\lambda_i$ is positive. Thus, squaring $B$ yields $$ B^2 = -(\lambda_1^2 \ [u_1u_1^* + u_2 u_2^*] + \cdots + \lambda_{k}^2\ [u_{2k-1}u_{2k-1}^* + u_{2k}u_{2k}^*]). $$ We could equivalently have used the canonical form (with a real, orthogonal $U$) $$ B = U \pmatrix{0 & -\lambda_1 \\ \lambda_1 & 0 \\ && \ddots \\ &&& 0 & -\lambda_k \\ &&& \lambda_k & 0 \\ &&&&& 0 } U^T \\ = \lambda_1 \ [u_2u_1^T - u_1 u_2^T] + \cdots + \lambda_{k}\ [u_{2k}u_{2k-1}^T - u_{2k-1}u_{2k}^T] $$

Related Question