# Interpretability research
Pythia 12b Deduped
Apache-2.0
Pythia-12B-deduped is a large language model with 12B parameters developed by EleutherAI, designed specifically for interpretability research and trained on the deduplicated Pile dataset.
Large Language Model
Transformers English

P
EleutherAI
4,708
52
Pythia 2.8b
Apache-2.0
Pythia - 2.8 billion is a member of the scalable language model suite developed by EleutherAI, designed specifically to promote the interpretability research of large language models. This model is based on the Transformer architecture and is trained on the The Pile dataset, with 2.8 billion parameters.
Large Language Model
Transformers English

P
EleutherAI
40.38k
30
Pythia 160m
Apache-2.0
Pythia-160M is a language model dedicated to interpretability research developed by EleutherAI. It belongs to the 160M parameter scale version in the Pythia suite and is based on the Transformer architecture, trained on the Pile dataset.
Large Language Model
Transformers English

P
EleutherAI
163.75k
31
Featured Recommended AI Models