Solved – Confusion about interpreting log likelihood (and likelihood ratio test) output

lme4-nlmemaximum likelihood

I am using the nlme package in R to use mixed effects models to analyze multilevel (i.e., students nested within classrooms) data.

I am confused about interpreting the log likelihood output from these functions, and while there are other questions that address why the log is taken of the likelihood statistic and why it is negative, it is not clear how these apply to interpreting these values.

In particular, a negative log likelihood is output. Since estimation methods maximize the log likelihood, I understand a larger value indicates better fit.

Does this interpretation apply to the (negative) log likelihood output, so that a larger (i.e., closer to 0) value indicates better fit, so that if I were comparing, for example, log likelihood values of -3146.9 and -2931.41 using the anova() function, and a likelihood ratio test indicated a test statistic associated with a p-value less than .05, then the model with the log likelihood of -2931.41 demonstrates better fit?

Best Answer

We take the log-likelihood because each case in the dataset gets a likelihood, and the log-likelihood is the product of these likelihoods. But each of these likelihoods is less than 1, and when you multiply lots of numbers less than 1 together you tend to get really, really small numbers.

Nothing wrong with those really small numbers - except that we hit precision limits on the computers that we use. Try exponentiating your log-likelihoods of -3000 and -2000. I suspect that your computer will say that they are both zero - so the likelihood of both models is zero - and they are indistinguishable.

So instead of trying to multiply lots of values togehter, it is easier to take their logs, and add them together. So we don't get a likelihood, we get a log-likelihood, and we don't hit precision limits.

Then @PlayStarCraftOkLetsGo's answer.

Related Question