GLM 4 32B 0414 4bit DWQ
This is the MLX format version of the THUDM/GLM-4-32B-0414 model, processed with 4-bit DWQ quantization, suitable for efficient inference on Apple silicon devices.
Downloads 156
Release Time : 5/22/2025
Model Overview
An MLX-adapted version based on Tsinghua University's GLM-4-32B large language model, supporting Chinese and English text generation tasks, optimized for Apple M-series chips.
Model Features
Apple silicon optimization
MLX format specifically optimized for Apple M-series chips, providing efficient local inference capabilities
4-bit quantization
Uses DWQ (Dynamic Weight Quantization) technology to compress the model to 4-bit precision, reducing memory usage
Bilingual support
Native support for Chinese and English text generation tasks
Model Capabilities
Text generation
Dialogue systems
Content creation
Question answering systems
Use Cases
Intelligent assistants
Chatbot
Build fluent Chinese-English dialogue systems
Natural and smooth conversational experience
Content creation
Article generation
Automatically generates coherent text content based on prompts
High-quality long-form text output
Featured Recommended AI Models