Skip to content
OpenCatalogcurated by FLOSSK
AI & Machine Learning

Pythia (Hugging Face)

EleutherAI’s public scaling suite: matched GPT-NeoX–architecture models from 70M–12B with public datasets for interpretability research.

Why it is included

Pythia checkpoints remain heavily downloaded for `text-generation`—gold standard for mechanistic interpretability baselines.

Best for

Researchers studying training dynamics, memorization, and layer-wise behavior.

Strengths

  • Public intermediate checkpoints
  • Documented data
  • Reproducible ladder

Limitations

  • Not competitive with frontier chat models for products

Good alternatives

OLMo · GPT-NeoX · TinyLlama

Related tools