Distilbert Base Uncased Sparse 90 Unstructured Pruneofa
Apache-2.0
This is a sparse pre-trained model achieving 90% weight sparsity through one-shot pruning, suitable for fine-tuning on various language tasks.
Large Language Model
Transformers English