# 4-bit Quantized Inference
Qwen3 0.6B 4bit
Apache-2.0
This is a 4-bit quantized version converted from the Qwen/Qwen3-0.6B model, suitable for efficient inference on the MLX framework.
Large Language Model
Q
mlx-community
6,015
5
Philosophy Model
Apache-2.0
This is a Mistral-7B instruction fine-tuned model optimized using Unsloth and Huggingface TRL library, achieving 2x training speed improvement
Large Language Model
Transformers English

P
raak-16
54
2
Llama 3.2 Vision Instruct Bpmncoder
Apache-2.0
Llama 3.2 11B vision instruction fine-tuned model optimized with Unsloth, using 4-bit quantization technology, achieving 2x faster training speed
Text-to-Image
Transformers English

L
utkarshkingh
40
1
Qwen 2 VL 7B OCR
Apache-2.0
A fine-tuned version of the Qwen2-VL-7B model, trained using Unsloth and Huggingface's TRL library, achieving a 2x speed improvement.
Text-to-Image
Transformers English

Q
Swapnik
103
1
Llama Bodo Translation Model
Apache-2.0
A 4-bit quantized version of Meta-Llama-3.1-8B fine-tuned for bidirectional Bodo-English translation, optimized with Unsloth for faster training
Large Language Model
Transformers Supports Multiple Languages

L
Luson045
27
1
Cogito Maximus
Apache-2.0
Cogito-Maximus is an advanced text generation model optimized based on the Qwen2.5-72B instruction fine-tuning model. It uses Unsloth for accelerated training and the TRL fine-tuning framework, and is suitable for various text generation scenarios.
Large Language Model
Transformers

C
Daemontatox
694
2
Mistral 7B Summarizer SFT GGUF
MIT
A text summarization model based on the Mistral 7B architecture, optimized for efficiency and performance using LoRA technology.
Text Generation English
M
SURESHBEEKHANI
65
0
Aria Sequential Mlp Bnb Nf4
Apache-2.0
A BitsAndBytes NF4 quantized version based on Aria-sequential_mlp, suitable for image-to-text tasks with approximately 15.5 GB VRAM requirement.
Image-to-Text
Transformers

A
leon-se
76
11
Text2cypher Gemma 2 9b It Finetuned 2024v1
Apache-2.0
This model is a Text2Cypher model fine-tuned based on google/gemma-2-9b-it, capable of converting natural language questions into Cypher query statements for Neo4j graph databases.
Knowledge Graph English
T
neo4j
2,093
22
Acegpt 13B Chat AWQ
The AWQ quantized version of AceGPT 13B Chat, supporting English and Arabic, designed for general GPU users, offering efficient 4-bit quantized inference capabilities.
Large Language Model
Transformers Supports Multiple Languages

A
MohamedRashad
37
3
Featured Recommended AI Models