B

Bert Large Uncased Sparse 90 Unstructured Pruneofa

Developed by Intel
This model is a sparse pre-trained model achieving 90% sparsity through weight pruning techniques, suitable for fine-tuning on various language tasks.
Downloads 13
Release Time : 3/2/2022

Model Overview

A sparse language model based on BERT-Large architecture, employing a one-shot pruning general method to retain important weight information while reducing computational overhead.

Model Features

High-Sparsity Design
Achieves efficient computation through 90% weight pruning while maintaining model performance.
One-shot Pruning General Method
Requires only a single pruning session to adapt to multiple downstream tasks without repeated pruning.
Knowledge Retention Technology
Uses a teacher-student framework to ensure critical knowledge is preserved after pruning.

Model Capabilities

Text Understanding
Language Feature Extraction
Downstream Task Fine-tuning

Use Cases

Natural Language Processing
Question Answering Systems
Can be fine-tuned to build efficient QA systems
Achieves 83.35 EM/90.20 F1 on SQuADv1.1
Text Classification
Suitable for classification tasks like sentiment analysis
Achieves 92.95% accuracy on SST-2
Natural Language Inference
Applicable to genre natural language inference tasks
Achieves 83.74%/84.20% accuracy on MNLI-m/mm respectively
Featured Recommended AI Models
ยฉ 2025AIbase