Tinyllama 1.1B Intermediate Step 1431k 3T
TinyLlama is a 1.1B parameter Llama model pretrained on 3 trillion tokens, designed to provide compact and efficient text generation capabilities.
Downloads 25.04k
Release Time : 12/28/2023
Model Overview
The TinyLlama project aims to pretrain a 1.1B parameter Llama model using 3 trillion tokens. Through optimization, the training can be completed in 90 days using 16 A100-40G GPUs.
Model Features
Efficient Pretraining
Pretrained on 3 trillion tokens, optimized to complete training within 90 days.
Compact Model
Only 1.1B parameters, suitable for applications with computational and memory constraints.
Compatibility
Adopts the same architecture and tokenizer as Llama 2, enabling plug-and-play integration into many Llama-based open-source projects.
Model Capabilities
Text Generation
Reasoning Tasks
Question Answering Systems
Use Cases
Natural Language Processing
AI2 Reasoning Challenge
Used to solve problems in the AI2 Reasoning Challenge
Standardized accuracy 33.87
HellaSwag
Used for text generation tasks on the HellaSwag dataset
Standardized accuracy 60.31
Education
MMLU
Used for multi-task language understanding evaluation
Accuracy 26.04
Featured Recommended AI Models
Š 2025AIbase