S

Sheared LLaMA 1.3B

Developed by princeton-nlp
Sheared-LLaMA-1.3B is an efficient language model obtained through structured pruning and continual pre-training based on LLaMA-2-7B
Downloads 11.09k
Release Time : 10/10/2023

Model Overview

This model achieves superior performance compared to similar models under a 50B token budget through dynamic loading of the RedPajama dataset for pruning and continual pre-training

Model Features

Efficient pruning technique
Uses only 0.4B tokens for pruning, significantly reducing computational costs
Continual pre-training
Uses 50B tokens for continual pre-training of the pruned model to maintain performance
Compatibility
Shares the same vocabulary as LLaMA1 and LLaMA2, facilitating migration and usage

Model Capabilities

Text generation
Language understanding
Reasoning tasks
Reading comprehension
Knowledge-intensive task processing

Use Cases

Natural language processing
Language model benchmarking
Excellent performance on benchmarks such as ARC and HellaSwag
Average performance of 51.0, surpassing other 1.3B parameter models
Knowledge QA
Handles knowledge-intensive question answering tasks
Achieved 37.14 on TruthfulQA
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase