E

EXAONE Deep 32B GGUF

Developed by Mungert
EXAONE-Deep-32B is a 32B-parameter large language model supporting English and Korean, specifically designed for text generation tasks.
Downloads 2,249
Release Time : 3/25/2025

Model Overview

This model is fine-tuned based on LGAI-EXAONE/EXAONE-3.5-32B-Instruct, utilizing IQ-DynamicGate ultra-low-bit quantization technology to optimize memory efficiency and inference speed.

Model Features

IQ-DynamicGate Ultra-Low-Bit Quantization
Employs adaptive precision quantization technology, significantly improving model accuracy at 1-2 bit quantization while maintaining high memory efficiency.
Dynamic Precision Allocation
The first and last 25% of layers use IQ4_XS, while the middle 50% use IQ2_XXS/IQ3_S, optimizing model performance.
Key Component Protection
Embedding and output layers use Q5_K quantization to reduce error propagation and enhance overall model accuracy.

Model Capabilities

Text generation
Multilingual support (English, Korean)
Ultra-low-bit quantized inference

Use Cases

Memory-constrained environments
GPU VRAM Adaptation
Running large language models on GPUs with limited VRAM.
Quantization technology significantly reduces model VRAM usage.
CPU and Edge Device Deployment
Running models on low-power devices.
Ultra-low-bit quantization enables model operation on resource-constrained devices.
Research and Applications
Ultra-Low-Bit Quantization Research
Studying the impact of 1-2 bit quantization on model performance.
DynamicGate quantization technology significantly improves model accuracy at low-bit quantization.
Featured Recommended AI Models
ยฉ 2025AIbase