Solved – What’s the relationship between initial eigenvalues and sums of squared loadings in factor analysis

eigenvaluesfactor analysisfactor-rotationpcaspss

On the one hand I read in a comment here that:

You can't speak of "eigenvalues" after rotation, even orthogonal
rotation. Perhaps you mean sum of squared loadings for a principal
component, after rotation. When rotation is oblique, this sum of
squares tells nothing about the amount of variance explained, because
components aren't orthogonal anymore. So, you shouldn't report any
percentage of variance explained.

On the other hand, I sometimes read in books people saying things like:

The eigenvalues associated with each factor represent the variance
explained by that particular factor; SPSS also displays the eigenvalue
in terms of the percentage of variance explained (so factor 1 explains
31.696% of total variance). The first few factors explain relatively large amounts of variance (especially factor 1), whereas subsequent
factors explain only small amounts of variance. SPSS then extracts all
factors with eigenvalues greater than 1, which leaves us with four
factors. The eigenvalues associated with these factors are again
displayed (and the percentage of variance explained) in the columns
labelled Extraction Sums of Squared Loadings.

That text is from Field (2013) Discovering statistics using IBM SPSS, and this diagram accompanies it.

Field example

I'm wondering

  1. Who is correct about whether it's possible to speak of eigenvalues after rotation? Would it matter if it was an oblique or orthogonal rotation?
  2. Why are the "initial eigenvalues" different from the "extraction sums of squared loadings"? Which is a better measure of total variation explained by the factors (or principal components or whatever method is used)? Should I say that the first four factors explain 50.317% of variation, or 40.477%?

Best Answer

The two citations do not generally contradict each other and both look to me correct. The only underwork is in Perhaps you mean sum of squared loadings for a principal component, after rotation one should better drop word "principal" since rotated components or factors are not "principal" anymore, to be rigorous. Also (important!) the second citation is correct only when "factor analysis" is actually PCA method (like it is in SPSS by default) and so factors are just principal components. But the table you present is not after PCA, and I wonder whether they are from the same text and wasn't there some misprint.

In the extraction summary table you display there was 23 variables analyzed. Eigenvalues of their correlation matrix are shown in the left section "Initial eigenvalues". No factors have been extracted yet. These eigenvalues correspond to the variances of Principal components (i.e. PCA was performed), not of factors. Adjective "initial" means "at the initiation point of the analysis" and does not imply that there must be some "final" eigenvalues.

The (default in SPSS) Kaiser rule "eigenvalues>1" was used to decide how many factors to extract, so, 4 factors will come. The "eigenvalues>1" rule is based on PCA's eigenvalues (i.e. the eigenvalues of the intact, input correlation matrix).

Extraction of them was done by Principal axis method and the matrix of loadings obtained. Sums of squared loadings in the matrix columns are the factors' variances after extraction. These values appear in the middle section of your table.

These numbers, generally, should not be called eigenvalues because factor extractions not necessarily are based right on the eigendecomposition of the input data - they are specific algorithms on their own. Even Principal axis method which does involve eigenvalues deal with eigenvalues of a repeatedly "trained" matrix, not an original correlation matrix.

But if you had been doing PCA instead of FA then the 4 numbers in the middle column would have been the 4 first eigenvalues identical to the 4 largest ones on the left: in PCA, no fitting take place and the extracted "latent variables" are the PCs themselves, which eigenvalues are their variances.

In the right section, sums of squared loading after rotaion of the factors are shown. The variances of these new, rotated factors. Please read more about rotated factors (or components), especially footnote 4, and that they are neither "principal" anymore nor this-one-to-that-one correspondent to the extracted ones. After rotation, "2nd" factor, for example, is not "2nd extracted factor, rotated". And it also could have greater variance than the "1st" one.

So,

  1. No, you can't speak of eigenvalues after rotation. No matter be it orthogonal or oblique.
  2. You can't even say - at least should better avoid - of eigenvalues even after extraction of factors unless these factors are principal components$^1$. (An instructive example showing confusion similar to yours with ML factor extraction.) Variances of factors are SS loadings, not eigenvalues, generally.
  3. Rotated factors don't correspond one-to-one to the extracted ones.
  4. The % of total variation explained by the factors is 40.477% in your example, not 50.317%. The first number is less because FA factors explain all the assumed common variation which is less than the portion of total variation skimmed by the same number of PCs. May say in your report, "The 4-factor solution is responsible for the common variance constituting 40.5% of the total variance; while 4 principal components would account for 50.3% of the total variance".

$^1$ (Before factor rotation) variances of factors (pr. components) are the eigenvalues of the correlation/covariance matrix of the data if the FA is PCA method; variances of factors are the eigenvalues of the reduced correlation/covariance matrix with final communalities on the diagonal, if the FA is PAF method of extraction; variances of factors do not correspond to eigenvalues of correlation/covariance matrix in other FA methods such as ML, ULS, GLS (see). In all cases, variances of orthogonal factors are the SS of the extracted/rotated - final - loadings.