O

Olmo2 8B SuperBPE T180k

Developed by UW
An 8-billion-parameter large language model featuring the innovative SuperBPE tokenizer, achieving 27% higher efficiency than traditional BPE models
Downloads 160
Release Time : 3/19/2025

Model Overview

A large language model built on the OLMo2 7B architecture, utilizing SuperBPE tokenization for more efficient text encoding

Model Features

SuperBPE Tokenizer
Innovative tokenization algorithm combining subwords and superword tokens, achieving 27% higher efficiency than traditional BPE
Efficient Encoding
3000-token context length matches the actual byte size of traditional BPE models with 4096 tokens
Large-Scale Training
Trained on 331 billion tokens with a vocabulary of 200,000 words

Model Capabilities

Text Generation
Efficient Text Encoding

Use Cases

Natural Language Processing
Efficient Text Processing
Processing long texts using SuperBPE technology
27% more efficient than traditional BPE models
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase