O

Olmo 2 0425 1B Instruct

Developed by allenai
OLMo 2 1B is a post-training variant of the allenai/OLMo-2-0425-1B-RLVR1 model, undergoing supervised fine-tuning, DPO training, and RLVR training, aiming to achieve state-of-the-art performance across multiple tasks.
Downloads 5,127
Release Time : 4/29/2025

Model Overview

OLMo 2 1B is an open language model primarily designed for text generation tasks, supporting English, and suitable for research and educational purposes.

Model Features

Open Language Model
The OLMo series aims to advance scientific research in language models by publicly releasing all code, checkpoints, logs, and related training details.
Multi-task Performance
Excels in various tasks such as MATH, GSM8K, and IFEval.
RLVR Training
Trained with RLVR, providing intermediate checkpoints to facilitate RL fine-tuning research.

Model Capabilities

Text Generation
Multi-task Handling
Instruction Following

Use Cases

Education
Math Problem Solving
Solves math problems, such as those in the GSM8K and MATH datasets.
Achieves 68.3 on GSM8K and 20.7 on MATH.
Research
Language Model Research
Used for fine-tuning and performance studies of language models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase