Solved – the main difference between Multiple R squared and Adjusted R squared

r-squaredregression

How can we use this as a basis to decide the best regression fit model? Not many question posts included the concept of Adjusted R-squared for understanding.

Best Answer

I wont go into the real maths of it (as I don't understand it myself), but I can explain it in more general terms.

Multiple R squared is simply a measure of Rsquared for models that have multiple predictor variables. Therefore it measures the amount of variation in the response variable that can be explained by the predictor variables. The fundamental point is that when you add predictors to your model, the multiple Rsquared will always increase, as a predictor will always explain some portion of the variance.

Adjusted Rsquared controls against this increase, and adds penalties for the number of predictors in the model. Therefore it shows a balance between the most parsimonious model, and the best fitting model. Generally, if you have a large difference between your multiple and your adjusted Rsquared that indicates you may have overfit your model.

Hope this helps. Hopefully someone may come along and explain this more in depth.