G

GLM Z1 9B 0414 GGUF

Developed by Mungert
GLM-Z1-9B-0414 is a bilingual text generation model supporting both Chinese and English, utilizing the GGUF format and suitable for various quantization levels, from BF16 to ultra-low-bit quantization (1-2 bits).
Downloads 1,598
Release Time : 4/26/2025

Model Overview

This model is based on the GLM architecture, supporting bilingual text generation tasks in Chinese and English, and is suitable for various hardware environments, including CPU and GPU.

Model Features

IQ-DynamicGate Ultra-low-bit Quantization
Supports 1-2 bit ultra-low-bit quantization, improving accuracy while maintaining memory efficiency through dynamic precision allocation and key component protection.
Multi-quantization Level Support
Offers various quantization levels from BF16, F16 to Q4_K, Q6_K, Q8_0, etc., catering to different hardware requirements.
Bilingual Support
The model supports text generation tasks in both Chinese and English.

Model Capabilities

Text generation
Bilingual processing
Low-memory inference

Use Cases

Memory-constrained Deployment
Edge Device Inference
Run text generation tasks on edge devices with limited memory.
Achieves ultimate memory efficiency through ultra-low-bit quantization (e.g., IQ3_XS).
Research
Ultra-low-bit Quantization Research
Study the impact of 1-2 bit quantization on model performance.
Provides benchmark data for various quantization levels.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase