V

Vit L 14 CLIPA 336 Datacomp1b

Developed by UCSC-VLAA
CLIPA-v2 model, an efficient contrastive image-text model, focused on zero-shot image classification tasks.
Downloads 239
Release Time : 10/17/2023

Model Overview

This model is a contrastive image-text model based on the CLIPA-v2 architecture, primarily used for zero-shot image classification. It is trained on large-scale datasets and can classify images without fine-tuning.

Model Features

Efficient Zero-shot Classification
Capable of performing image classification tasks directly without the need for fine-tuning.
Large-scale Training Data
Trained using the mlfoundations/datacomp_1b dataset, covering a wide range of visual concepts.
Inverse Scaling Optimization
Utilizes CLIPA-v2's inverse scaling training method to achieve efficient performance.

Model Capabilities

Zero-shot image classification
Image-text contrastive learning
Multi-label image recognition

Use Cases

Image Classification
Zero-shot Image Classification
Classifies images without training, suitable for rapid deployment scenarios.
Achieves 81.1% zero-shot accuracy on ImageNet
Content Understanding
Image-Text Matching
Determines the degree of match between an image and a text description.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase