O

Olmo 2 0425 1B SFT

Developed by allenai
OLMo 2 1B SFT is a supervised fine-tuned version of the OLMo-2-0425-1B model, trained on the Tulu 3 dataset, designed to achieve state-of-the-art performance across multiple tasks.
Downloads 1,759
Release Time : 4/24/2025

Model Overview

This is an open language model primarily for English text generation tasks, released under the Apache 2.0 license, suitable for research and educational purposes.

Model Features

Open Language Model
Fully open-source with weights and training data, advancing scientific research in language models
Multi-task Optimization
Fine-tuned on the Tulu 3 dataset, optimized for performance in tasks like MATH, GSM8K, and IFEval
Transparent Training
All training code, checkpoints, and logs are publicly available to ensure research reproducibility

Model Capabilities

Text Generation
Dialogue Systems
Mathematical Problem Solving
Factual Question Answering

Use Cases

Education
Mathematical Problem Solving
Solving math problems from GSM8K and MATH datasets
Achieved 52.1 on GSM8K and 13.2 on MATH
Research
Language Model Research
Used for studying language model behavior and capabilities
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase