D

Deit Base Distilled Patch16 224

Developed by facebook
The distilled version of the Efficient Data Image Transformer (DeiT) model was pre-trained and fine-tuned on ImageNet-1k at 224x224 resolution, extracting knowledge from a teacher model through distillation learning.
Downloads 35.53k
Release Time : 3/2/2022

Model Overview

This model is a distilled version of the Vision Transformer (ViT), learning from a teacher CNN model using distillation tokens, suitable for image classification tasks.

Model Features

Distillation Learning
Learns from a teacher CNN model using distillation tokens to enhance model performance.
Efficient Training
Trained for 3 days on a single 8-GPU node at a resolution of 224x224.
High-Resolution Support
Supports 384x384 resolution, further improving classification accuracy.

Model Capabilities

Image Classification
Visual Feature Extraction

Use Cases

Computer Vision
ImageNet Image Classification
Classifies images into one of the 1000 ImageNet categories.
Top-1 accuracy 83.4%, Top-5 accuracy 96.5%.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase