I am involved in a project where i need to construct some examples using non-negative Multivariate Gaussian Random variables(log-normal random variables). Part of the computation requires calculating the covariance matrices. Since i do not have any data and i am plucking numbers out of the air i was hoping if some one would be kind to verify that the following algorithm is valid or not.
- I assume a bunch of positive log values as standard deviations for my variables.
Lets call these $$s =[\sigma_1,\cdots, \sigma_n]$$ - I then compute $$s^2 = s^{T}s.$$
- I assume a correlation matrix $\rho$, which are all non-negative and this matrix is symmetric.
- Multiply $\rho$ with $s^2$ element wise. This is a non standard operation, what i am suggesting here is the $(i, j)$th element of $s^2$ is multiplied by $(i, j)$th element of $\rho$. The matrix produced when all the element-wise operations are complete is the covariance matrix $\Sigma$.
I have verified in mat lab that $\Sigma$ values of my examples are Choleski decomposable, as in i get a valid upper triangle matrix as result. Hence i am inferring from that they are positive-definite. These matrices are also symmetric.
Can we infer based on the process and Choleski factorization, that covariance matrices are valid ?
Best Answer
(i) Where you have $s^\top s$ did you mean $ss^\top$?
(ii) Where you say multiply element-wise, do you exclude the diagonal?If the answer is 'yes' to
bothitem (i), you should get a covariance matrix corresponding to pairwise correlations ofρ$\rho_{ij}$ for all pairs.If you can compute a full Choleski it will be a valid covariance matrix - regardless.
Edit: it appears you meant $\rho_{ij}$ where you wrote $\rho$, so point (ii) is irrelevant. I have edited to reflect that.