Normal Distribution – Sufficiency of Mean Statistic When Variance is Unknown

normal distributionsufficient-statistics

According to the PDF here: https://www.math.arizona.edu/~tgk/466/sufficient.pdf, the sum of a sample of data is not a sufficient statistic for the normal distribution when the variance is unknown. Instead, the sum and sum of squares are jointly sufficient statistics. Now, let's say we have a game with two players. Both of them know that five samples are drawn from some normal distribution. None of them know the parameters of the normal distribution used to generate the data. The goal of the game is to estimate the mean of the normal distribution. The player that comes closer to the true mean wins 1\$ (absolute difference between estimated value and actual value is the objective function).

While the first player is given all five samples, the second one is given just the sum of the samples (and he knows there were five of them).

If the sum alone is not a sufficient statistic when the variance is unknown, then what strategy can the first player apply that uses all five of the data points to win money over multiple such games?

As a follow-up, let's say the distribution isn't normal anymore and the players know what it is (without knowing the parameters). And they still have to estimate the real mean from the information they're given. Are there examples of distributions where the first player would have a substantial advantage?

And finally, what if both players had no idea what distribution was used to generate the samples? Now would the first player have any advantage?

Best Answer

$\bar X$ is not a sufficient statistic because it does not contain all the information about $(\mu,\sigma^2)$, which is what it would mean for it to be sufficient.

However, $\bar X$ does contain all the information about $\mu$ in the sample, whether or not $\sigma^2$ is known. For example, $\bar X$ attains the Cramèr-Rao bound. Similarly, if $\mu$ is not known, $s^2$ contains all the information about $\sigma^2$ (though not if $\mu$ is known since $(\mu-\bar X)^2$ has information about $\sigma^2$). Having all the information about parts of a parameter is a more complicated property than sufficiency, though it has been studied (see e.g., Sprott 1975).