Solved – When do we use S.E.M. and when do we use S.D.

distributionserrormathematical-statisticsmeansample

Put simply, the standard error of the sample mean is an estimate of how far the sample mean is likely to be from the population mean, whereas the standard deviation of the sample is the degree to which individuals within the sample differ from the sample mean. If the population standard deviation is finite, the standard error of the mean of the sample will tend to zero with increasing sample size, because the estimate of the population mean will improve, while the standard deviation of the sample will tend to approximate the population standard deviation as the sample size increases.

When do we use S.E.M. and when do we use S.D.?

Quote Source: Wikipedia

Best Answer

Suppose you want to know the average height of adults in a given country. Also suppose that if you were able to get all height measurements, the distribution of the data would follow a normal distribution. This distribution has then two important features, that is the mean $\mu$, which is the center of the distribution, and the standard deviation $\sigma$, which is a measure of spread around the center of that distribution. In this scenario, one standard deviation around the mean would capture in 68% of the data points; two standard deviations would capture 95% of the data points; and three standard deviations would capture 99.7% of the data points (see here).

However, in almost all cases, you are not able to measure all members of a given population, instead and you have to rely on taking random samples from that distribution to estimate how far the sample mean is from the true population mean $\mu$. If this is your goal, then you calculate the standard error of the mean. One standard error of the mean is then the interval in which the true population mean would fall 68% of the time if sampling was repeated over and over again. Usually in statistics a 95% confidence interval is used, which you can get by multiplying the standard error with 2 (see link above). Given the formula for the standard error of the mean, it is also apparent that if the sample size goes up, the interval tends to zero and you are closing in on your population mean $\mu$ (as in your quote above). Thus, the standard error of the mean is a tool in inferential statistics, that is inferring from the distribution of a random sample (observed data) to properties of an underlying unknown distribution, or the population.

The standard deviation on the other hand is used to describe the variability in the observed data only (i.e. the sample) without making any inferences with respect to properties of the underlying unknown distribution. The standard deviation is commonly used in descriptive statistics.

Now depending on whether you want to infer properties of an unknown distribution from a random sample (which is what we are mostly interested in when doing statistics), or whether you want to simply describe the variability in your sample, you should use the standard error of the mean and the standard deviation, respectively.

Related Question