Evaluation datasets maintained by EleutherAI
AI & ML interests
Large language models, scaling laws, AI Alignment, democratization of DL
Recent Activity
Pythia is the first LLM suite designed specifically to enable scientific research on LLMs. To learn more see https://github.com/EleutherAI/pythia
-
Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
Paper • 2304.01373 • Published • 9 -
EleutherAI/pythia-14m-deduped
Text Generation • 39.2M • Updated • 13.5k • 29 -
EleutherAI/pythia-70m
95.6M • Updated • 647k • 81 -
EleutherAI/pythia-160m
Text Generation • Updated • 3.17M • 41
A collection of datasets and finetuned models that can be used for Eliciting Latent Knowledge (ELK) research.
Models and datasets for benchmarking ELK methods. NOTE: These models correspond to our first arxiv release. See our newer Quirky Models collection.
SAEs are tools for understanding the internal representations of neural networks. These can be loaded using https://github.com/EleutherAI/sae
This collection contains the model and data artifacts from O'Brien et al. (2025). https://deepignorance.ai
-
Deep Ignorance: Filtering Pretraining Data Builds Tamper-Resistant Safeguards into Open-Weight LLMs
Paper • 2508.06601 • Published • 7 -
EleutherAI/deep-ignorance-unfiltered
Text Generation • 7B • Updated • 1.59k • 5 -
EleutherAI/deep-ignorance-e2e-strong-filter
Text Generation • 7B • Updated • 3.27k • 1 -
EleutherAI/deep-ignorance-strong-filter-pt-weak-filter-anneal
Text Generation • 7B • Updated • 121 • 1
T5 trained on the Pile with Llama Tokenizer
T5x Checkpoints for Pile-T5
Evaluation datasets maintained by EleutherAI
This collection contains the model and data artifacts from O'Brien et al. (2025). https://deepignorance.ai
-
Deep Ignorance: Filtering Pretraining Data Builds Tamper-Resistant Safeguards into Open-Weight LLMs
Paper • 2508.06601 • Published • 7 -
EleutherAI/deep-ignorance-unfiltered
Text Generation • 7B • Updated • 1.59k • 5 -
EleutherAI/deep-ignorance-e2e-strong-filter
Text Generation • 7B • Updated • 3.27k • 1 -
EleutherAI/deep-ignorance-strong-filter-pt-weak-filter-anneal
Text Generation • 7B • Updated • 121 • 1
Pythia is the first LLM suite designed specifically to enable scientific research on LLMs. To learn more see https://github.com/EleutherAI/pythia
-
Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
Paper • 2304.01373 • Published • 9 -
EleutherAI/pythia-14m-deduped
Text Generation • 39.2M • Updated • 13.5k • 29 -
EleutherAI/pythia-70m
95.6M • Updated • 647k • 81 -
EleutherAI/pythia-160m
Text Generation • Updated • 3.17M • 41
A collection of datasets and finetuned models that can be used for Eliciting Latent Knowledge (ELK) research.
T5 trained on the Pile with Llama Tokenizer
Models and datasets for benchmarking ELK methods. NOTE: These models correspond to our first arxiv release. See our newer Quirky Models collection.
T5x Checkpoints for Pile-T5
SAEs are tools for understanding the internal representations of neural networks. These can be loaded using https://github.com/EleutherAI/sae