D

Dorna Llama3 8B Instruct Quantized4Bit

Developed by amirMohammadi
4-bit quantized version of Dorna-Llama3-8B-Instruct, optimized for Persian language with Flash Attention 2 technology for enhanced inference efficiency
Downloads 22
Release Time : 6/8/2024

Model Overview

This is an 8B-parameter large language model based on the Llama3 architecture, specifically fine-tuned for Persian language data and optimized for memory usage through 4-bit quantization, suitable for Persian text generation tasks

Model Features

Memory optimization
4-bit quantization significantly reduces memory requirements, suitable for resource-constrained environments
Inference acceleration
Integrated Flash Attention 2 technology improves processing speed
Persian optimization
Specifically trained/fine-tuned for Persian language data
Easy deployment
Ready to use out-of-the-box without additional libraries like LlamaCPP or Candle

Model Capabilities

Persian text generation
English text generation
Dialogue systems
Question answering systems
Text summarization

Use Cases

Language services
Persian intelligent assistant
For building Persian dialogue systems
Achieved 55.77% win rate against Persian Mind model in human evaluations
Cross-language QA system
Supports question answering services in Persian and English
Excellent performance in news QA tasks
Education
Language learning tool
Assists Persian language learners in practice
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase