[Math] Proof that $\frac{(\bar X-\mu)}{\sigma}$ and $\sum_{i=1}^n\frac{(X_i-\bar X)^2}{\sigma^2}$ are independent

generating-functionsprobabilitysamplingstatistical-inferencestatistics

Let $X_i\sim N(\mu,\sigma^2)$ ; where$[i=1,2,\ldots,n]$

$Z_i\sim N(0,1)$ ; where$[i=1,2,\ldots,n]$

Proof that $\bar Z=\frac{(\bar X-\mu)}{\sigma}$ and $\sum_{i=1}^{n}(Z_i-\bar Z)^2=\sum_{i=1}^n\frac{(X_i-\bar X)^2}{\sigma^2}$
are independent, which implies $\bar X$ and $\sum_{i=1}^n(X_i-\bar X)^2$ are independent.

MY ATTEMPT:

I considered $n=2$, and

$$M_{X_1+X_2}(t_1)=M_{X_1}(t_1)M_{X_2}(t_1)$$

$\ast$I did so for the proof but does the statement $X_i\sim N(\mu,\sigma^2)$ ; where$[i=1,2,\ldots,n]$ imply that $X_i's$ are independent?

Moment Generating Function of $X_1+X_2$ and $X_1-X_2$ are
$$M_{X_1+X_2}(t_1)=e^{2\mu t_1+\sigma^2 t_1^2}$$
$$M_{X_1-X_2}(t_2)=e^{\sigma^2 t_2^2}$$
respectively.

Also,

$$M_{X_1+X_2,X_1-X_2}(t_1,t_2)=e^{2\mu t_1+\sigma^2 t_1^2}e^{\sigma^2 t_2^2}=M_{X_1+X_2}(t_1)M_{X_1-X_2}(t_2)$$

since the joint moment generating function factors into the product of the marginal moment generating functions, so $X_1+X_2$ and $X_1-X_2$ are independent which implies that:

$\bullet$ Since $\bar Z=\frac{(\bar X-\mu)}{\sigma}$ is only a function of $X_1+X_2$ and $\sum_{i=1}^{2}(Z_i-\bar Z)^2=\sum_{i=1}^2\frac{(X_i-\bar X)^2}{\sigma^2}$ is only a function of $X_1-X_2$ ,so

$\bar Z=\frac{(\bar X-\mu)}{\sigma}$ and $\sum_{i=1}^{n}(Z_i-\bar Z)^2=\sum_{i=1}^n\frac{(X_i-\bar X)^2}{\sigma^2}$
are independent.

$\bullet$ Since $\bar X$ is only a function of $X_1+X_2$ and $S^2=\frac{1}{2-1}\sum_{i=1}^2(X_i-\bar X)^2$ is only a function of $X_1-X_2$ ,so

Sample mean,$\bar X$ and sample variance,$S^2$ are independent.

$\diamond\diamond\diamond$ We accept the above independence for any arbitrary $n$.

Is the procedure correct?

Best Answer

The way to do this is to show that $\bar X=(X_1+\cdots+X_n)/n$ is independent of $$ (X_1-\bar X, \ldots, X_n-\bar X). $$ Notice that the orthogonal projection of $(X_1,\ldots,X_n)$ onto the $1$-dimensional space $\{ (x,x,x,\ldots,x) : a \in \mathbb R\}$ is $(\bar X,\ldots, \bar X)$, and the orthogonal projection onto the $(n-1)$-dimensional space $x_1+\cdots+x_n=0$ is $(X_1-\bar X, \ldots, X_n-\bar X)$, and those two spaces are orthogonal to each other. Notice also that the probability distribution of $(X_1,\ldots,X_n)$ is spherically symmetric about the point $(\mu,\mu,\ldots,\mu)$, which is in the $1$-dimensional space mentioned above, and which gets mapped to $(0,0,0,\ldots,0)$ by the second projection.

Another way to look at it is this: The random vector $(X_1,\ldots,X_n)$ is jointly normally distributed, i.e. it is so distributed that every linear combination $a_1X_1+\cdots+a_nX_n$ has a $1$-dimensional normal distribution. That implies that any two linear combinations of $X_1,\ldots,X_n$ that are uncorrelated are independent. So then notice that $$ \operatorname{cov}\left( \bar X, \begin{bmatrix} X_1 - \bar X \\ \vdots \\ X_n - \bar X\end{bmatrix}\right) = \begin{bmatrix} 0 \\ \vdots \\ 0 \end{bmatrix}. $$

Related Question