Sheared LLaMA 2.7B
S
Sheared LLaMA 2.7B
Developed by princeton-nlp
Sheared-LLaMA-2.7B is a lightweight language model derived from Llama-2-7b through pruning and continued pretraining, consuming only a 50B token budget.
Downloads 1,131
Release Time : 10/10/2023
Model Overview
This model is compressed from Llama-2-7b using structured pruning techniques, retaining the core capabilities of the original model while excelling in multiple downstream tasks.
Model Features
Efficient Pruning
Uses only 0.4B tokens for pruning, significantly reducing model size
Efficient Training
Achieves excellent performance with only 50B tokens for continued pretraining
Superior Performance
Outperforms models of the same scale at both 1.3B and 2.7B sizes
Model Capabilities
Text generation
Language understanding
Reasoning tasks
Reading comprehension
Knowledge-intensive tasks
Use Cases
Natural Language Processing
Language Modeling
Used for generating coherent text content
Performs excellently in language modeling tasks
Question Answering Systems
Building knowledge-intensive Q&A applications
Performs well in reading comprehension tasks
Featured Recommended AI Models
Š 2025AIbase