FARE4 Convnext Base W Laion2b S13b B82k Augreg
F
FARE4 Convnext Base W Laion2b S13b B82k Augreg
Developed by chs20
A robust perceptual metric model based on CLIP, enhanced for adversarial robustness through adversarial fine-tuning
Downloads 39
Release Time : 8/14/2024
Model Overview
This model is a vision-language model based on the CLIP architecture, adversarially fine-tuned using the FARE method on ImageNet, specifically designed to improve robustness in perceptual similarity tasks.
Model Features
Adversarial Robustness
Utilizes the FARE method for adversarial fine-tuning on ImageNet to enhance resistance against adversarial examples
Perceptual Similarity Metrics
Specially optimized for perceptual similarity tasks, demonstrating excellent performance on the NIGHTS dataset
CLIP-based Architecture
Built upon the powerful CLIP model, inheriting its outstanding vision-language alignment capabilities
Model Capabilities
Zero-shot image classification
Perceptual similarity metrics
Adversarial example recognition
Use Cases
Computer Vision
Image Similarity Comparison
Used to evaluate the similarity between two images at a human perceptual level
Achieves 90.6% accuracy on the NIGHTS dataset
Adversarial Example Detection
Identifies images modified by adversarial attacks
Maintains 74.3% performance under L-infinity norm attacks
Featured Recommended AI Models
Š 2025AIbase