Perplexity
- Perplexity is an Intrinsic Evaluation
- Using Extrinsic Evaluation or downstream tasks is slow and hard
- So people use Perplexity to compare different models and with other research
- Perplexity tells us how good the model is generating sentences, or how close the sentence is to the pre-training data distribution
- Less Perplexity is better
- or it means that current text is highly similar to the pre-training data
Perplexity from Likelihood
Perplexity from Cross Entropy