Recall the model for simple linear regression
$$
y_i = \beta_0 + \beta_1 x_i + \varepsilon_i.
$$
I am reading up on the standard error of the coefficients $\beta_0$ and $\beta_1$. As an experiment I generated some linear data using $\beta_0 = 1$ and $\beta_1 = 2$ and added some Gaussian noise with unit variance. So then when I fit the data the lm
function and used the summary
function to examine the model I have the following output:
\begin{align}
\hat \beta_0 & = 1.21054 \quad \text{with Std. Error} = 0.11508, \\
\hat \beta_1 & = 1.87723 \quad \text{with Std. Error} = 0.09844.
\end{align}
So how do I interpret the standard error values? For instance, take $\hat \beta_0$, precisely what is $0.11508$ telling me?
Obviously if I ran the simulation a second time, this time adding Gaussian noise with a higher amount of variance, the standard error would increase as the extra variance in the noise shows up as an increase in the standard error of the coefficients. But, if we consider the first simulation in isolation, then what does this value of $0.11508$ mean?
Best Answer
The standard error is the square root of an estimate of the sampling variability of $\hat\beta_j$ as an estimator of $\beta_j$, or $\sqrt{\widehat{Var}(\hat\beta_j)}$.
As this is many things in one sentence, step-by-step: