๐Ÿ“ˆ Perplexity Viewer

Visualize per-token perplexity using color gradients.

  • Red: High perplexity (model is uncertain)
  • Green: Low perplexity (model is confident)

Choose between decoder models (like GPT) for true perplexity or encoder models (like BERT) for pseudo-perplexity via MLM.

Model Name

Select a model or enter a custom HuggingFace model name

Model Type

Decoder for causal LM, Encoder for masked LM

Detailed Token Results

Click on an example to try it out:
Input Text Model Name Model Type Mask Probability Min Samples per Token

๐Ÿ“Š How it works:

  • Decoder Models (GPT, etc.): Calculate true perplexity by measuring how well the model predicts the next token
  • Encoder Models (BERT, etc.): Calculate pseudo-perplexity using statistical sampling with multiple token masking
  • Mask Probability: For encoder models, controls what fraction of tokens get masked in each iteration
  • Min Samples: Minimum number of perplexity measurements collected per token for robust statistics
  • Color Coding: Red = High perplexity (uncertain), Green = Low perplexity (confident)

โš ๏ธ Notes:

  • First model load may take some time
  • Models are cached after first use
  • Very long texts are truncated to 512 tokens
  • GPU acceleration is used when available
  • Encoder models use Monte Carlo sampling for robust perplexity estimates
  • Higher min samples = more accurate but slower analysis