๐ Perplexity Viewer
Visualize per-token perplexity using color gradients.
- Red: High perplexity (model is uncertain)
- Green: Low perplexity (model is confident)
Choose between decoder models (like GPT) for true perplexity or encoder models (like BERT) for pseudo-perplexity via MLM.
Model Name
Select a model or enter a custom HuggingFace model name
Detailed Token Results
Click on an example to try it out:
| Input Text | Model Name | Model Type | Mask Probability | Min Samples per Token |
|---|
๐ How it works:
- Decoder Models (GPT, etc.): Calculate true perplexity by measuring how well the model predicts the next token
- Encoder Models (BERT, etc.): Calculate pseudo-perplexity using statistical sampling with multiple token masking
- Mask Probability: For encoder models, controls what fraction of tokens get masked in each iteration
- Min Samples: Minimum number of perplexity measurements collected per token for robust statistics
- Color Coding: Red = High perplexity (uncertain), Green = Low perplexity (confident)
โ ๏ธ Notes:
- First model load may take some time
- Models are cached after first use
- Very long texts are truncated to 512 tokens
- GPU acceleration is used when available
- Encoder models use Monte Carlo sampling for robust perplexity estimates
- Higher min samples = more accurate but slower analysis