I once heard that
log transformation is the most popular one for right-skewed distributions in linear regression or quantile regression
I would like to know is there any reason underlying this statement? Why is the log transformation suitable for a right-skewed distribution?
How about a left-skewed distribution?
Best Answer
Economists (like me) love the log transformation. We especially love it in regression models, like this: \begin{align} \ln{Y_i} &= \beta_1 + \beta_2 \ln{X_i} + \epsilon_i \end{align}
Why do we love it so much? Here is the list of reasons I give students when I lecture on it:
Statisticians generally find economists over-enthusiastic about this particular transformation of the data. This, I think, is because they judge my point 8 and the second half of my point 3 to be very important. Thus, in cases where the data are not log-normally distributed or where logging the data does not result in the transformed data having equal variance across observations, a statistician will tend not to like the transformation very much. The economist is likely to plunge ahead anyway since what we really like about the transformation are points 1,2,and 4-7.