F

FARE4 Convnext Base W Laion2b S13b B82k Augreg

Developed by chs20
A robust perceptual metric model based on CLIP, enhanced for adversarial robustness through adversarial fine-tuning
Downloads 39
Release Time : 8/14/2024

Model Overview

This model is a vision-language model based on the CLIP architecture, adversarially fine-tuned using the FARE method on ImageNet, specifically designed to improve robustness in perceptual similarity tasks.

Model Features

Adversarial Robustness
Utilizes the FARE method for adversarial fine-tuning on ImageNet to enhance resistance against adversarial examples
Perceptual Similarity Metrics
Specially optimized for perceptual similarity tasks, demonstrating excellent performance on the NIGHTS dataset
CLIP-based Architecture
Built upon the powerful CLIP model, inheriting its outstanding vision-language alignment capabilities

Model Capabilities

Zero-shot image classification
Perceptual similarity metrics
Adversarial example recognition

Use Cases

Computer Vision
Image Similarity Comparison
Used to evaluate the similarity between two images at a human perceptual level
Achieves 90.6% accuracy on the NIGHTS dataset
Adversarial Example Detection
Identifies images modified by adversarial attacks
Maintains 74.3% performance under L-infinity norm attacks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase