S

Sheared LLaMA 2.7B

Developed by princeton-nlp
Sheared-LLaMA-2.7B is a lightweight language model derived from Llama-2-7b through pruning and continued pretraining, consuming only a 50B token budget.
Downloads 1,131
Release Time : 10/10/2023

Model Overview

This model is compressed from Llama-2-7b using structured pruning techniques, retaining the core capabilities of the original model while excelling in multiple downstream tasks.

Model Features

Efficient Pruning
Uses only 0.4B tokens for pruning, significantly reducing model size
Efficient Training
Achieves excellent performance with only 50B tokens for continued pretraining
Superior Performance
Outperforms models of the same scale at both 1.3B and 2.7B sizes

Model Capabilities

Text generation
Language understanding
Reasoning tasks
Reading comprehension
Knowledge-intensive tasks

Use Cases

Natural Language Processing
Language Modeling
Used for generating coherent text content
Performs excellently in language modeling tasks
Question Answering Systems
Building knowledge-intensive Q&A applications
Performs well in reading comprehension tasks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase