B

Bert Large Uncased Wwm Squadv2 X2.15 F83.2 D25 Hybrid V1

Developed by madlag
This model is pruned using the nn_pruning library, retaining 32% of the original weights, with a processing speed 2.15 times faster than the original version and an F1 score of 83.22
Downloads 21
Release Time : 3/2/2022

Model Overview

A question answering system model based on the BERT-large architecture, fine-tuned for the SQuAD 2.0 dataset, utilizing whole-word masking technology, suitable for English question answering tasks

Model Features

Efficient Pruning Technique
Structured pruning achieved via the nn_pruning library, retaining 25% of weights in linear layers and 32% overall
Accelerated Inference
Processing speed reaches 2.15 times that of the original BERT-large
Attention Head Optimization
155 out of 384 attention heads (40.4%) were pruned to improve computational efficiency

Model Capabilities

English Question Answering
Reading Comprehension
Text Understanding

Use Cases

Education
Learning Assistance System
Helps students quickly obtain answers from textbooks
Accuracy with F1 score of 83.22
Customer Support
FAQ Auto-Response
Automatically retrieves answers from a knowledge base
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase