D

Distilbert Base Uncased Distilled Squad Int8 Static Inc

Developed by Intel
This is the INT8 quantized version of the DistilBERT base uncased model, specifically designed for question answering tasks, optimized for model size and inference speed through post-training static quantization.
Downloads 1,737
Release Time : 8/4/2022

Model Overview

This model is the INT8 quantized version of the DistilBERT base uncased model, fine-tuned on the Stanford Question Answering Dataset (SQuAD). The quantization process utilized Hugging Face's Optimum-Intel toolkit and Intel® Neural Compressor technology, aiming to significantly reduce model size and inference latency while maintaining high accuracy.

Model Features

INT8 Quantization
Converts the model from FP32 precision to INT8 through post-training static quantization, significantly reducing model size and inference latency.
Efficient Inference
The optimized model is suitable for deployment in resource-constrained environments, providing low-latency question answering capabilities.
High Accuracy Retention
Strives to retain the original model's accuracy during quantization, ensuring efficient execution of question answering tasks.

Model Capabilities

Question Answering
Text Understanding
Context Analysis

Use Cases

Question Answering Systems
Context-based Question Answering
Answers questions based on given contextual paragraphs, suitable for scenarios such as knowledge base queries and customer service systems.
F1 Score: 86.1069 (INT8 PyTorch version)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase