M

Marqo Fashionclip

Developed by Marqo
Marqo-FashionCLIP is a fashion-domain multimodal retrieval model based on the CLIP architecture, achieving state-of-the-art performance in fashion product search tasks through generalized contrastive learning.
Downloads 8,376
Release Time : 8/8/2024

Model Overview

This model is specifically optimized for the fashion domain, capable of processing both image and text inputs for zero-shot image classification and multimodal retrieval tasks. It surpasses previous SOTA models on multiple fashion datasets.

Model Features

Generalized Contrastive Learning
Utilizes GCL method to train not only on text descriptions but also on multi-dimensional features such as categories, styles, and colors.
Fashion Domain Optimization
Specifically fine-tuned for fashion product search tasks, demonstrating excellent performance on multiple fashion datasets.
Multi-framework Support
Supports various usage methods including Hugging Face, OpenCLIP, and Transformers.js.

Model Capabilities

Zero-shot image classification
Text-to-image retrieval
Image-to-text retrieval
Multimodal feature extraction

Use Cases

E-commerce
Fashion Product Search
Find relevant fashion products based on text descriptions or categories
Surpasses previous state-of-the-art models on multiple fashion datasets
Visual Similarity Search
Find visually similar fashion products based on images
Content Management
Automatic Product Tagging
Automatically generate labels and descriptions for fashion product images
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase