O

Owlv2 Base Patch16 Finetuned

Developed by google
OWLv2 is a zero-shot text-conditioned object detection model that can retrieve objects in images through text queries.
Downloads 2,698
Release Time : 10/13/2023

Model Overview

OWLv2 is an open-world localization model based on CLIP backbone network, capable of detecting objects in images through text queries without requiring category-specific training data.

Model Features

Zero-shot detection capability
Detects novel category objects directly through text queries without requiring category-specific training data
Open-vocabulary recognition
Capable of recognizing category names not present in training data
Multi-query detection
Supports simultaneous object detection using multiple text queries

Model Capabilities

Object detection in images
Text-conditioned object localization
Open-vocabulary recognition
Zero-shot learning

Use Cases

Computer vision research
Zero-shot object detection research
Used to study model's detection capability on unseen categories
Cross-domain applications
Specialized domain object recognition
Identifying rare objects in specialized fields like healthcare and agriculture
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase