G

GLM 4 32B 0414 4bit DWQ

Developed by mlx-community
This is the MLX format version of the THUDM/GLM-4-32B-0414 model, processed with 4-bit DWQ quantization, suitable for efficient inference on Apple silicon devices.
Downloads 156
Release Time : 5/22/2025

Model Overview

An MLX-adapted version based on Tsinghua University's GLM-4-32B large language model, supporting Chinese and English text generation tasks, optimized for Apple M-series chips.

Model Features

Apple silicon optimization
MLX format specifically optimized for Apple M-series chips, providing efficient local inference capabilities
4-bit quantization
Uses DWQ (Dynamic Weight Quantization) technology to compress the model to 4-bit precision, reducing memory usage
Bilingual support
Native support for Chinese and English text generation tasks

Model Capabilities

Text generation
Dialogue systems
Content creation
Question answering systems

Use Cases

Intelligent assistants
Chatbot
Build fluent Chinese-English dialogue systems
Natural and smooth conversational experience
Content creation
Article generation
Automatically generates coherent text content based on prompts
High-quality long-form text output
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase