O

Olmo2 8B SuperBPE T160k

Developed by UW
An 8-billion-parameter model featuring the innovative SuperBPE tokenizer, combining subword and super tokens, achieving 30% higher inference efficiency than traditional BPE models.
Downloads 28
Release Time : 3/19/2025

Model Overview

A large language model built on the OLMo2 7B architecture, utilizing the SuperBPE tokenizer for more efficient text encoding and generation.

Model Features

SuperBPE Tokenizer
Combines traditional subword tokens with innovative super tokens that span across word boundaries, significantly improving encoding efficiency.
Efficient Inference
Achieves 30% higher average efficiency during inference compared to traditional BPE models.
Large Vocabulary
Vocabulary size of 200K, including 160K subword tokens and 40K super tokens.
Long Context Support
Supports context lengths of 2,884 tokens, equivalent to the actual byte count of 4,096 tokens in traditional BPE models.

Model Capabilities

Text Generation
Efficient Text Encoding

Use Cases

Natural Language Processing
Text Generation
Generates coherent, contextually relevant text content.
High-quality text output with 30% efficiency improvement.
Text Encoding
Efficiently encodes long texts, reducing token count.
Fewer tokens required to encode equivalent text.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Ā© 2025AIbase