Regression Analysis – Understanding Why Residual Standard Error is Not Referred to as RMSE

rregression

As I understand it, in the specific context of linear regression, the R output "residual standard error" is an estimate of $\sigma$, the standard deviation of the distribution of the residuals. You can compute $\sigma^2$ by the MSE, which in general is the mean of the squared errors, but in regression the denominator is the residual degrees of freedom $n-p$ ($p$ representing the number of parameters including the intercept).

$$MSE=\frac{\sum_{i=1}^{n} y_i – \hat{y}_i}{n-p}$$

Why isn't residual standard error just called RMSE, when $\sigma^2$ is the MSE? Or is $\hat{\sigma}^2$ often called the residual variance (making the terms for $\sigma^2$ and $\sigma$ line up)?

Best Answer

Referring to “the” MSE is probably a mistake, since there are reasonable arguments for multiple calculations (an $n$ denominator and and $n−p$ denominator both make sense). I would want to define explicitly what I mean if there is any ambiguity about it.

However, the R decision to call $\hat\sigma$ the standard error makes no sense to me, because standard errors are associated with parameters being estimated. To which parameter does $\hat\sigma$ correspond? (I don’t have an answer, which is why I have yet to see why R calls $\hat\sigma$ a residual standard error.)

I do not really have a clean name for $\hat\sigma=\sqrt{ \frac{ \sum\left( y_i-\hat y_i \right)^2 }{ n-p} }$. While the expression inside the square root is unbiased for error variance (assuming fairly typical assumptions like the Gauss-Markov conditions), Jensen’s inequality means that $\hat\sigma$ is biased for the error standard deviation, so “unbiased error standard deviation” is not correct.

EDIT

The R developers appear to know that residual standard "error" is a misnomer and lament that it crept into the documentation for functions to such an extent that correcting this becomes difficult.