perplexity
scroll ↓ to Resources
Note
- Intrinsic LLM evaluation method: a geometric average of the inverse probability of words predicted by the model.
- Intuitively, perplexity means to be surprised. We measure how much the model is surprised by seeing new data. The lower the perplexity, the better the training is.
- Another common measure is the cross-entropy, which is the Logarithm (base 2) of perplexity.
- As a thumb rule, a reduction of 10-20% in perplexity is noteworthy.
Resources
Links to this File
table file.inlinks, file.outlinks from [[]] and !outgoing([[]]) AND -"Changelog"