Solved – How should perplexity of LDA behave as value of the latent variable k increases

latent-dirichlet-alloclatent-variableperplexity

When increasing the value of the latent variable k for LDA (latent Dirichlet allocation), how should perplexity behave:

  1. On the training set?
  2. On the testing set?

Best Answer

The original paper on LDA gives some insights into this:

In particular, we computed the perplexity of a held-out test set to evaluate the models. The perplexity, used by convention in language modeling, is monotonically decreasing in the likelihood of the test data, and is algebraicly equivalent to the inverse of the geometric mean per-word likelihood. A lower perplexity score indicates better generalization performance.

This should be the behavior on test data. Here is a result from paper: Perplexity as a function of k