O

Owlv2 Base Patch16

Developed by google
OWLv2 is a zero-shot text-conditioned object detection model that can retrieve objects in images through text queries.
Downloads 15.42k
Release Time : 10/13/2023

Model Overview

OWLv2 is an open-world localization model based on the CLIP backbone network, supporting zero-shot object detection via text queries.

Model Features

Zero-shot detection
Detects new objects through text queries without category-specific training
Open-vocabulary classification
Enables detection of arbitrary text categories by replacing classification layer weights
Multi-query support
Supports simultaneous search for objects matching multiple text descriptions in a single image

Model Capabilities

Image object detection
Text-conditioned search
Open-vocabulary recognition

Use Cases

Computer vision research
Zero-shot detection research
Exploring the model's recognition capability for unseen categories
Interdisciplinary applications
Specialized domain object recognition
Performing object detection in domains lacking annotated data (e.g., medical images)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase