Olmo2 11B SuperBPE T180k
O
Olmo2 11B SuperBPE T180k
Developed by UW
An 11-billion parameter large language model trained with the innovative SuperBPE tokenizer, supporting superword unit recognition and subword tokenization capabilities.
Downloads 29
Release Time : 3/19/2025
Model Overview
A large language model extended from the OLMo2-7B architecture, trained with the SuperBPE tokenizer, featuring enhanced text comprehension and generation capabilities.
Model Features
SuperBPE Tokenizer
Innovatively introduces superword units (capable of spanning word boundaries) while retaining subword tokenization capabilities.
Efficient Context Processing
3000-token context window, equivalent to the byte-level context capacity of a 4096-token BPE model.
Large-scale Training
Trained on 238 billion tokens with a vocabulary size of 200,000.
Model Capabilities
Text generation
Natural language understanding
Use Cases
Text generation
Creative writing
Generate coherent and creative text content.
Code generation
Assist in generating programming code snippets.
Natural language processing
Text summarization
Automatically generate concise summaries of text.
Question answering systems
Build intelligent question-answering systems.
Featured Recommended AI Models
Ā© 2025AIbase