# Fast inference
Qwen3 0.6B 8bit
Apache-2.0
Qwen3-0.6B-8bit is an 8-bit quantized version converted from Qwen/Qwen3-0.6B, a text generation model suitable for the MLX framework.
Large Language Model
Q
mlx-community
2,625
3
Rubert Mini Frida
MIT
A lightweight and fast modified version of the FRIDA model for computing embedding vectors of Russian and English sentences
Text Embedding
Transformers Supports Multiple Languages

R
sergeyzh
1,203
9
Lite Whisper Large V3 Fast
Apache-2.0
Lite-Whisper is a lightweight version of OpenAI Whisper compressed using LiteASR technology, significantly reducing model size while maintaining high recognition accuracy.
Speech Recognition
Transformers

L
efficient-speech
25
1
Kokoro V1 0
Apache-2.0
Kokoro is an open-source text-to-speech model with 82 million parameters, achieving sound quality comparable to large models with a lightweight architecture while improving generation speed and reducing computational costs.
Speech Synthesis English
K
kiriyamaX
18
1
Ai Image Detector Dev Deploy
This is an auto-trained image classification model capable of recognizing multiple common object categories
Image Classification
TensorBoard

A
haywoodsloan
59
1
SD3.5 Large Fp8
Other
FP8 quantized version of Stable Diffusion 3.5 Large for text-to-image generation tasks.
Image Generation
S
dyedd
88
2
Sana 1600M 1024px MultiLing
Sana is an efficient text-to-image framework capable of generating images with resolutions up to 4096×4096, supporting multilingual input.
Text-to-Image Supports Multiple Languages
S
Efficient-Large-Model
111
24
Depthpro Hf
DepthPro is a foundational model for zero-shot metric monocular depth estimation, capable of generating high-resolution, high-precision depth maps.
3D Vision
Transformers

D
apple
13.96k
52
Midjourney Mini Openvino
MIT
This is an OpenVINO-optimized midjourney-mini model for text-to-image generation tasks.
Image Generation Supports Multiple Languages
M
hsuwill000
26
1
Sana 1600M 1024px
Sana is an efficient text-to-image framework capable of generating images up to 4096×4096 resolution, deployable on laptop GPUs.
Image Generation Supports Multiple Languages
S
Efficient-Large-Model
2,327
206
SD2.1 Nitro
Apache-2.0
An efficient text-to-image model series distilled from mainstream diffusion models on AMD Instinct™ GPUs
Image Generation
S
amd
117
6
Vit Gigantic Patch14 Clip 224.metaclip 2pt5b
A dual-framework compatible vision model trained on MetaCLIP-2.5B dataset, supporting both OpenCLIP and timm frameworks
Image Classification
V
timm
444
0
Vit Huge Patch14 Clip 224.metaclip 2pt5b
A dual-purpose vision-language model trained on the MetaCLIP-2.5B dataset, supporting zero-shot image classification tasks
Image Classification
V
timm
3,173
0
Vit Large Patch14 Clip 224.metaclip 2pt5b
A dual-framework compatible vision model trained on MetaCLIP-2.5B dataset, supporting zero-shot image classification tasks
Image Classification
V
timm
2,648
0
Vit Large Patch14 Clip 224.metaclip 400m
Vision Transformer model trained on MetaCLIP-400M dataset, supporting zero-shot image classification tasks
Image Classification
V
timm
294
0
Vit Base Patch16 Clip 224.metaclip 2pt5b
A dual-framework compatible vision model trained on the MetaCLIP-2.5B dataset, supporting both OpenCLIP and timm frameworks
Image Classification
V
timm
889
1
Vit Base Patch32 Clip 224.metaclip 2pt5b
A vision Transformer model trained on the MetaCLIP-2.5B dataset, compatible with both open_clip and timm frameworks
Image Classification
V
timm
5,571
0
Molmo 7B O Bnb 4bit
Apache-2.0
The 4-bit quantized version of Molmo-7B-O, significantly reducing the memory requirement and suitable for environments with limited resources.
Large Language Model
Transformers

M
cyan2k
2,467
11
Historiccolorsoonr Schnell
Apache-2.0
A versatile vision+text generation model, particularly suitable for generating realistic images that simulate color film photography, presenting a wide range of visual paradigms from Autochrome to Kodachrome to Fujifilm and other iconic photographic technologies.
Image Generation English
H
AlekseyCalvin
37
1
Mlx FLUX.1 Schnell 4bit Quantized
Apache-2.0
A 4-bit quantized text-to-image generation model optimized for the MLX framework, supporting efficient image generation
Text-to-Image English
M
argmaxinc
1,644
16
Biomed Right
A zero-shot classification model based on the Transformers library, capable of performing classification tasks without task-specific training data.
Text Classification
Transformers

B
gritli
15
0
Protgpt2 Distilled Tiny
Apache-2.0
A distilled version of ProtGPT2, compressed into a more efficient small model through knowledge distillation, maintaining performance while improving inference speed
Protein Model
Transformers

P
littleworth
157
4
Llama 3 8b Quantized
Other
The 4-bit quantized version of the Llama 3 model, which optimizes memory usage and speeds up inference, suitable for environments with limited computing resources.
Large Language Model
Transformers English

L
SweatyCrayfish
2,037
11
Cat Vs Dog Classification
Apache-2.0
An image classification model fine-tuned on the cats_vs_dogs dataset using Google's ViT model, designed to distinguish between images of cats and dogs.
Image Classification
Transformers

C
kazuma313
42
1
Tinyllama 1.1B Chat V1.0 GGUF
Apache-2.0
TinyLlama is a lightweight 1.1B-parameter Llama model optimized for chat and programming assistance tasks.
Large Language Model English
T
andrijdavid
117
2
Vitforimageclassification
Apache-2.0
This model is a fine-tuned image classification model based on google/vit-base-patch16-224-in21k on the CIFAR10 dataset, achieving an accuracy of 96.78%.
Image Classification
Transformers

V
Andron00e
43
2
X Ray Ai Detection
An X-ray image detection model fine-tuned based on AI-image-detector, achieving 99.83% accuracy
Image Classification
Transformers

X
Artef
22
3
Sdxl Chinese Ink Lora
Openrail
A Chinese ink painting style generation model fine-tuned based on the Stable Diffusion XL framework
Image Generation
S
ming-yang
100
9
Lcm Lora Sdv1 5
This is a LoRA adapter designed for Stable Diffusion v1-5, reducing inference steps to just 2-8 steps, significantly improving generation speed.
Image Generation
L
latent-consistency
127.41k
499
Vit Finetuned Vanilla Cifar10 0
Apache-2.0
An image classification model fine-tuned on the CIFAR-10 dataset based on the Vision Transformer (ViT) architecture, achieving 99.2% accuracy
Image Classification
Transformers

V
02shanky
68
1
Voidnoisecore R0829
Other
A text-to-image generation model based on Stable Diffusion that can generate high-quality images according to text descriptions.
Image Generation
V
digiplay
48.27k
3
Fantasticmix2.5d V4.5
Other
This is a text-to-image generation model based on Stable Diffusion, capable of generating high-quality images according to text descriptions.
Image Generation
F
digiplay
123
1
Bk Sdm Small
Openrail
BK-SDM is a stable diffusion model compressed through architecture compression, used for efficient and general text-to-image synthesis. It achieves lightweight design by removing residual and attention blocks in the U-Net.
Image Generation
B
nota-ai
745
31
Mousetrap ButterflyGenerator
MIT
This is an unconditional image generation model based on the diffusion model, specifically designed to generate cute butterfly images.
Image Generation
M
MouseTrap
2
0
Codexmd
MIT
Large Language Model
Transformers

C
Gouletf
126
1
Vegam Whisper Medium Ml
MIT
This is a version of thennal/whisper-medium-ml converted to the CTranslate2 model format for Malayalam speech recognition
Speech Recognition Other
V
smcproject
83
5
Ct2fast Opus Mt ROMANCE En
Apache-2.0
This is a CTranslate2-optimized multilingual translation model that supports fast translation from multiple Romance languages to English.
Machine Translation
Transformers

C
michaelfeil
74
1
Ct2fast Opus Mt De En
Apache-2.0
This is a quantized version of the Helsinki-NLP/opus-mt-de-en model, enabling fast inference through CTranslate2, supporting German to English machine translation.
Machine Translation
Transformers

C
michaelfeil
72
2
Deit Tiny Patch16 224 Finetuned Main Gpu 20e Final
Apache-2.0
Lightweight image classification model based on DeiT-tiny architecture, achieving 98.56% validation accuracy after fine-tuning on a custom image dataset
Image Classification
Transformers

D
Gokulapriyan
15
0
Convnext Tiny 224 Finetuned Aiornot
Apache-2.0
A computer vision model based on ConvNeXt-Tiny architecture, fine-tuned on specific datasets for image classification tasks
Image Classification
Transformers

C
kanak8278
16
0
- 1
- 2
Featured Recommended AI Models