Olmo2 8B SuperBPE T180k
O
Olmo2 8B SuperBPE T180k
Developed by UW
An 8-billion-parameter large language model featuring the innovative SuperBPE tokenizer, achieving 27% higher efficiency than traditional BPE models
Downloads 160
Release Time : 3/19/2025
Model Overview
A large language model built on the OLMo2 7B architecture, utilizing SuperBPE tokenization for more efficient text encoding
Model Features
SuperBPE Tokenizer
Innovative tokenization algorithm combining subwords and superword tokens, achieving 27% higher efficiency than traditional BPE
Efficient Encoding
3000-token context length matches the actual byte size of traditional BPE models with 4096 tokens
Large-Scale Training
Trained on 331 billion tokens with a vocabulary of 200,000 words
Model Capabilities
Text Generation
Efficient Text Encoding
Use Cases
Natural Language Processing
Efficient Text Processing
Processing long texts using SuperBPE technology
27% more efficient than traditional BPE models
Featured Recommended AI Models