S

Segformer B3 Finetuned Ade 512 512

Developed by nvidia
This SegFormer model was fine-tuned on the ADE20k dataset at 512x512 resolution, featuring a hierarchical Transformer encoder and lightweight all-MLP decoder head architecture.
Downloads 13.13k
Release Time : 3/2/2022

Model Overview

SegFormer is a Transformer-based semantic segmentation model suitable for image segmentation tasks, demonstrating excellent performance on benchmarks like ADE20K and Cityscapes.

Model Features

Efficient design
Utilizes hierarchical Transformer encoder and lightweight all-MLP decoder head architecture for concise and efficient semantic segmentation.
High performance
Demonstrates excellent performance on semantic segmentation benchmarks like ADE20K and Cityscapes.
Pre-training + Fine-tuning
The hierarchical Transformer is first pre-trained on ImageNet-1k, then the decoder head is added and jointly fine-tuned on downstream datasets.

Model Capabilities

Image semantic segmentation
Scene parsing

Use Cases

Computer vision
House scene parsing
Performs semantic segmentation on house images to identify different objects and regions.
Castle scene parsing
Performs semantic segmentation on castle images to identify different architectural structures and environmental elements.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase