Tecoa4 ViT B 32 Laion2b S34b B79k
T
Tecoa4 ViT B 32 Laion2b S34b B79k
Developed by chs20
A robust perceptual metric model based on CLIP, enhanced through adversarial fine-tuning for improved performance in perceptual similarity tasks
Downloads 21
Release Time : 8/14/2024
Model Overview
This model is a vision-language model based on the CLIP architecture, adversarially fine-tuned using the TeCoA method on the ImageNet dataset, specifically enhancing robustness in perceptual similarity tasks.
Model Features
Adversarial Fine-tuning
Adversarial fine-tuning on ImageNet using the TeCoA method improves the model's robustness against adversarial attacks
High Robustness
Maintains high performance under both L-infinity norm and L2 norm attacks
Perceptual Similarity Metrics
Specifically optimized for performance in perceptual similarity tasks
Model Capabilities
Zero-shot image classification
Perceptual similarity metrics
Robust classification under adversarial attacks
Use Cases
Computer Vision
Image Classification
Reliable image classification in adversarial attack environments
Maintains 79.1% performance under L-infinity norm attacks on the NIGHTS dataset
Perceptual Similarity Evaluation
Evaluates perceptual similarity between images
Achieves 91.0% clean data performance on the NIGHTS dataset
Featured Recommended AI Models
Š 2025AIbase