V

Vit Base Patch16 224 In21k Bantai Vitv1

Developed by AykeeSalazar
This model is an image classification model based on the Google Vision Transformer (ViT) architecture, achieving an accuracy of 86.36% after fine-tuning on the image_folder dataset.
Downloads 17
Release Time : 4/2/2022

Model Overview

This is an image classification model based on the ViT architecture, suitable for general image recognition tasks. The model performs excellently on standard image classification tasks, achieving an accuracy of 86.36%.

Model Features

High Accuracy
Achieves a classification accuracy of 86.36% on the evaluation set.
ViT-based Architecture
Utilizes the Vision Transformer architecture, leveraging self-attention mechanisms for image processing.
Transfer Learning
Fine-tuned based on the pre-trained google/vit-base-patch16-224-in21k model.

Model Capabilities

Image classification
Visual feature extraction

Use Cases

Computer Vision
General Image Classification
Classifies various types of images.
Achieves 86.36% accuracy on the test set.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase