Olmo2 8B SuperBPE T160k
O
Olmo2 8B SuperBPE T160k
Developed by UW
An 8-billion-parameter model featuring the innovative SuperBPE tokenizer, combining subword and super tokens, achieving 30% higher inference efficiency than traditional BPE models.
Downloads 28
Release Time : 3/19/2025
Model Overview
A large language model built on the OLMo2 7B architecture, utilizing the SuperBPE tokenizer for more efficient text encoding and generation.
Model Features
SuperBPE Tokenizer
Combines traditional subword tokens with innovative super tokens that span across word boundaries, significantly improving encoding efficiency.
Efficient Inference
Achieves 30% higher average efficiency during inference compared to traditional BPE models.
Large Vocabulary
Vocabulary size of 200K, including 160K subword tokens and 40K super tokens.
Long Context Support
Supports context lengths of 2,884 tokens, equivalent to the actual byte count of 4,096 tokens in traditional BPE models.
Model Capabilities
Text Generation
Efficient Text Encoding
Use Cases
Natural Language Processing
Text Generation
Generates coherent, contextually relevant text content.
High-quality text output with 30% efficiency improvement.
Text Encoding
Efficiently encodes long texts, reducing token count.
Fewer tokens required to encode equivalent text.
Featured Recommended AI Models
Ā© 2025AIbase