B

Bert Large Uncased Whole Word Masking Squad Int8 0001

Developed by dkurt
BERT-large English Q&A model pre-trained with whole word masking and fine-tuned on SQuAD v1.1, quantized to INT8 precision
Downloads 23
Release Time : 3/2/2022

Model Overview

This model is specifically designed for English Q&A tasks, taking text and questions as input and outputting the answer's position range in the text.

Model Features

Whole Word Masking Pre-training
Pre-trained using whole word masking technology to enhance the model's understanding of complete vocabulary.
Quantization-Aware Fine-Tuning
Quantization-aware fine-tuning via NNCF reduces model precision from FP32 to INT8.
Efficient Inference
The INT8 quantized version significantly improves inference speed while maintaining high accuracy.

Model Capabilities

English Q&A
Text Understanding
Answer Position Prediction

Use Cases

Smart Customer Service
Automated Q&A System
Document-based automated Q&A service
Quickly and accurately extracts answers from documents.
Education
Learning Assistance
Helps students quickly find answers from textbooks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase