A

AMD OLMo 1B

Developed by amd
AMD-OLMo is a series of 1-billion-parameter language models trained from scratch by AMD on the AMD Instinct™ MI250 GPU.
Downloads 4,419
Release Time : 10/31/2024

Model Overview

AMD-OLMo is a 1-billion-parameter language model based on the OLMo architecture. It has been trained through multiple stages, including pre-training, supervised fine-tuning (SFT), and DPO alignment, and demonstrates high training throughput and multi-task performance.

Model Features

High-performance training
Trained using AMD Instinct™ MI250 GPUs, with high training throughput (12,200 tokens/second/GPU).
Multi-stage fine-tuning
Through multiple stages such as pre-training, supervised fine-tuning (SFT), and DPO alignment, the model performance is improved.
Excellent multi-task performance
Performs excellently in multiple standard benchmarks, with good common-sense reasoning, multi-task understanding, and responsible AI capabilities.

Model Capabilities

Text generation
Instruction following
Multi-round dialogue
Common-sense reasoning
Multi-task understanding

Use Cases

Research
Language model research
Used to study the performance and behavior of language models.
Chat applications
Chatbot
Used to build multi-round dialogue chatbots.
Scored 4.35 on MT-Bench (AMD-OLMo-1B-SFT)
Instruction following
Instruction generation
Generate responses based on user instructions.
Win rate of 54.22% on AlpacaEval 1 (AMD-OLMo-1B-SFT-DPO)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase