O

Olmo2 11B SuperBPE T180k

Developed by UW
An 11-billion parameter large language model trained with the innovative SuperBPE tokenizer, supporting superword unit recognition and subword tokenization capabilities.
Downloads 29
Release Time : 3/19/2025

Model Overview

A large language model extended from the OLMo2-7B architecture, trained with the SuperBPE tokenizer, featuring enhanced text comprehension and generation capabilities.

Model Features

SuperBPE Tokenizer
Innovatively introduces superword units (capable of spanning word boundaries) while retaining subword tokenization capabilities.
Efficient Context Processing
3000-token context window, equivalent to the byte-level context capacity of a 4096-token BPE model.
Large-scale Training
Trained on 238 billion tokens with a vocabulary size of 200,000.

Model Capabilities

Text generation
Natural language understanding

Use Cases

Text generation
Creative writing
Generate coherent and creative text content.
Code generation
Assist in generating programming code snippets.
Natural language processing
Text summarization
Automatically generate concise summaries of text.
Question answering systems
Build intelligent question-answering systems.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Ā© 2025AIbase