D

Dinat Mini In1k 224

Developed by shi-labs
DiNAT-Mini is a hierarchical vision Transformer model based on neighborhood attention mechanism, specifically designed for image classification tasks.
Downloads 462
Release Time : 11/14/2022

Model Overview

This model employs Dilated Neighborhood Attention (DiNA) and is trained on the ImageNet-1K dataset, suitable for image classification tasks at 224x224 resolution.

Model Features

Neighborhood Attention Mechanism
Uses a constrained self-attention mechanism where each token's receptive field is limited to its nearest neighboring pixels, preserving translation equivariance.
Dilated Neighborhood Attention
Extends the receptive field through dilated variants (DiNA), forming a flexible sliding window attention pattern.
Hierarchical Structure
Adopts a hierarchical vision Transformer architecture, suitable for processing visual features at different scales.

Model Capabilities

Image Classification
Visual Feature Extraction

Use Cases

Computer Vision
ImageNet Image Classification
Classifies input images into one of 1000 ImageNet categories
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase