Solved – Interpreting significance of predictor vs significance of predictor coeffs in multinomial logistic regression

logisticstatistical significance

I have a multinomial logistic regression with dependent variable valued in {-1,0,1} (reference category is 0) and a number of continuous and discrete predictors. After running the regression a continuous predictor of interest ('size') has a Type 3 analysis of effects p-value of 0.0683, and the two coefficients (corresponding to outcomes of -1 and 1) have p-values of 0.8786 and 0.0220 respectively.

I read somewhere that one should only look at the significance of the coefficients if the predictor itself is significant at the chosen level. Is this right? My naive sense is that the predictor is borderline (taking alpha=0.05 for argument's sake), and that 'size' has a significant relationship to outcome=1 but not to outcome = -1. I would say that the significance of the relationship to outcome=1 is not terribly strong, but that is ok for the application in mind (or at least, with the indirect data I am forced to use)

Best Answer

The p-value itself cannot tell you how strong the relationship is, because the p-value is so influenced by sample size, among other things. But assuming your N is something on the order of 100-150, I'd say there's a reasonably strong effect involving Size whereby as Size increases, the log of the odds of Y being 1 is notably different from the log of the odds of Y being 0. As you indicate, the same cannot be said of the comparison of Y values of -1 and 0.

You are right in viewing all of this as somewhat invalidated by the overall nonsignificance of Size (depending on your alpha, or criterion for significance). You wouldn't get too many arguments if you simply declared Size a nonfactor due to its high p-value. But then again, if your N is sufficiently small--perhaps below 80 or 100--then your design affords low power for detecting effects, and you might make a case for taking seriously the specific effect that managed to show up anyway.

A way around the problem of relying on p-values involves two steps. First, decide what range of odds ratios would constitute an effect worth bothering with, or worth calling substantial. (The trick there is in being facile enough with odds to recognize what they mean for the more intuitive metric of probability.) Then construct a confidence interval for the odds ratio associated with each coefficient and consider it in light of your hypothetical range. Regardless of statistical significance, does the effect have practical significance?

Related Question