Bert Base Uncased Squad1.1 Block Sparse 0.20 V1
This is a pruned and optimized BERT Q&A model, retaining 38.1% of the original model's weights, fine-tuned on the SQuAD1.1 dataset, supporting English Q&A tasks.
Downloads 15
Release Time : 3/2/2022
Model Overview
A block-sparse Q&A model based on BERT-base uncased architecture, compressed using Movement Pruning, with inference speed 1.39x faster than the original model, suitable for English Q&A systems.
Model Features
Efficient Block Sparse Structure
Retains only 20.2% of linear layer weights, preserving 38.1% of the original model parameters, significantly reducing model size
Fast Inference
Using block-sparse runtime is 1.39x faster than dense networks
Attention Head Optimization
Removed 62.5% of attention heads (90 out of 144), optimizing computational efficiency
Knowledge Distillation
Distilled from csarron/bert-base-uncased-squad-v1 model, maintaining high accuracy
Model Capabilities
English Q&A
Text Understanding
Answer Extraction
Use Cases
Customer Support
Product Knowledge Q&A
Building automated Q&A systems based on product documentation
Can accurately answer user questions about product features
Educational Applications
Learning Assistant Q&A
Helping students quickly find answers in textbooks
Can accurately extract relevant information from textbook texts
Featured Recommended AI Models