Vit Base Patch16 224 Int8 Static Inc
This is an INT8 PyTorch model statically quantized using Intel® Neural Compressor post-training, based on Google's ViT model fine-tuning, significantly reducing model size while maintaining high accuracy.
Downloads 82
Release Time : 9/6/2022
Model Overview
This model is a quantized version of Vision Transformer (ViT), suitable for image classification tasks, specifically optimized for the imagenet-1k dataset.
Model Features
Efficient quantization
Post-training static quantization using Intel® Neural Compressor, compressing the model from FP32 to INT8, reducing volume by approximately 71%
Precision control
Selectively reverting specific linear modules to FP32 precision, ensuring accuracy loss is controlled within 1%
Optimized calibration
Using training set data loader for calibration, default sampling of 1000 samples (corresponding to 1000 classes)
Model Capabilities
Image classification
Efficient inference
Low memory footprint
Use Cases
Computer vision
Image classification system
Can be used to build efficient image classification systems, especially for general image classification of 1000 categories
Achieves 80.576% accuracy on imagenet-1k
Edge device deployment
Suitable for deployment on resource-constrained edge devices for image classification tasks
Model size is only 94MB, much smaller than the original FP32 model
Featured Recommended AI Models