B

Biomedclip PubMedBERT 256 Vit Base Patch16 224

Developed by microsoft
BiomedCLIP is a biomedical vision-language foundation model pre-trained via contrastive learning on the PMC-15M dataset, supporting cross-modal retrieval, image classification, visual question answering, and other tasks.
Downloads 137.39k
Release Time : 4/5/2023

Model Overview

This model employs PubMedBERT as the text encoder and Vision Transformer as the image encoder, specifically optimized for the biomedical domain to handle diverse types of biomedical images.

Model Features

Specialized for Biomedical Domain
Specifically optimized for the biomedical domain, capable of handling diverse biomedical image types such as microscopy, radiology, and histology.
Large-scale Pre-training
Pre-trained on the PMC-15M dataset containing 15 million image-caption pairs, covering a wide range of biomedical image types.
Multi-task Support
Supports various vision-language processing tasks including cross-modal retrieval, image classification, and visual question answering.

Model Capabilities

Biomedical image classification
Cross-modal retrieval
Visual question answering
Zero-shot learning

Use Cases

Medical Imaging Analysis
Histopathology Image Classification
Identifying different histopathology image types such as adenocarcinoma and squamous cell carcinoma
Achieved state-of-the-art performance on standard datasets
Radiology Image Analysis
Identifying radiological features such as pleural effusion
Medical Research
Medical Literature Image Retrieval
Retrieving relevant medical images based on textual descriptions
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase