Bert Large Uncased Whole Word Masking Squad Int8 0001
BERT-large English Q&A model pre-trained with whole word masking and fine-tuned on SQuAD v1.1, quantized to INT8 precision
Downloads 23
Release Time : 3/2/2022
Model Overview
This model is specifically designed for English Q&A tasks, taking text and questions as input and outputting the answer's position range in the text.
Model Features
Whole Word Masking Pre-training
Pre-trained using whole word masking technology to enhance the model's understanding of complete vocabulary.
Quantization-Aware Fine-Tuning
Quantization-aware fine-tuning via NNCF reduces model precision from FP32 to INT8.
Efficient Inference
The INT8 quantized version significantly improves inference speed while maintaining high accuracy.
Model Capabilities
English Q&A
Text Understanding
Answer Position Prediction
Use Cases
Smart Customer Service
Automated Q&A System
Document-based automated Q&A service
Quickly and accurately extracts answers from documents.
Education
Learning Assistance
Helps students quickly find answers from textbooks
Featured Recommended AI Models