Sheared LLaMA 1.3B
Sheared-LLaMA-1.3B is an efficient language model obtained through structured pruning and continual pre-training based on LLaMA-2-7B
Downloads 11.09k
Release Time : 10/10/2023
Model Overview
This model achieves superior performance compared to similar models under a 50B token budget through dynamic loading of the RedPajama dataset for pruning and continual pre-training
Model Features
Efficient pruning technique
Uses only 0.4B tokens for pruning, significantly reducing computational costs
Continual pre-training
Uses 50B tokens for continual pre-training of the pruned model to maintain performance
Compatibility
Shares the same vocabulary as LLaMA1 and LLaMA2, facilitating migration and usage
Model Capabilities
Text generation
Language understanding
Reasoning tasks
Reading comprehension
Knowledge-intensive task processing
Use Cases
Natural language processing
Language model benchmarking
Excellent performance on benchmarks such as ARC and HellaSwag
Average performance of 51.0, surpassing other 1.3B parameter models
Knowledge QA
Handles knowledge-intensive question answering tasks
Achieved 37.14 on TruthfulQA
Featured Recommended AI Models
Š 2025AIbase