V

Vit L 14 CLIPA Datacomp1b

Developed by UCSC-VLAA
CLIPA-v2 model, an efficient contrastive image-text model, focused on zero-shot image classification tasks.
Downloads 278
Release Time : 10/17/2023

Model Overview

This model is a contrastive image-text model based on the CLIPA-v2 architecture, primarily used for zero-shot image classification. Trained on large-scale datasets, it can directly classify images without task-specific fine-tuning.

Model Features

Efficient Zero-shot Classification
Capable of directly classifying images without task-specific fine-tuning.
Large-scale Dataset Training
Trained using the mlfoundations/datacomp_1b dataset, exhibiting strong generalization capabilities.
Low Cost, High Performance
Achieves high-accuracy zero-shot image classification with relatively low training costs.

Model Capabilities

Zero-shot Image Classification
Image-Text Contrastive Learning
Multilingual Support

Use Cases

Image Classification
Zero-shot Image Classification
Directly classify images of unseen categories without fine-tuning.
Achieves 81.1% accuracy on zero-shot ImageNet classification tasks.
Multimodal Applications
Image-Text Retrieval
Retrieve relevant images based on text descriptions or generate relevant text descriptions from images.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase