Vit L 14 CLIPA 336 Datacomp1b
CLIPA-v2 model, an efficient contrastive image-text model, focused on zero-shot image classification tasks.
Downloads 239
Release Time : 10/17/2023
Model Overview
This model is a contrastive image-text model based on the CLIPA-v2 architecture, primarily used for zero-shot image classification. It is trained on large-scale datasets and can classify images without fine-tuning.
Model Features
Efficient Zero-shot Classification
Capable of performing image classification tasks directly without the need for fine-tuning.
Large-scale Training Data
Trained using the mlfoundations/datacomp_1b dataset, covering a wide range of visual concepts.
Inverse Scaling Optimization
Utilizes CLIPA-v2's inverse scaling training method to achieve efficient performance.
Model Capabilities
Zero-shot image classification
Image-text contrastive learning
Multi-label image recognition
Use Cases
Image Classification
Zero-shot Image Classification
Classifies images without training, suitable for rapid deployment scenarios.
Achieves 81.1% zero-shot accuracy on ImageNet
Content Understanding
Image-Text Matching
Determines the degree of match between an image and a text description.
Featured Recommended AI Models
Š 2025AIbase