Qwen1.5 MoE A2.7B GGUF
The Mixture of Experts (MoE) model of Tongyi Qianwen version 1.5, with a parameter scale of 2.7B, providing GGUF format files in multiple quantization versions.
Downloads 163
Release Time : 11/11/2024
Model Overview
This is a Mixture of Experts model based on the Qwen1.5 architecture, providing GGUF format files in multiple quantization versions, suitable for local inference scenarios.
Model Features
Multiple Quantization Options
Provide model files with a total of 12 different quantization levels from Q2_K to Q8_0 to meet the needs of different scenarios.
Efficient Inference
The Mixture of Experts architecture improves inference efficiency while maintaining model performance.
llama.cpp Compatibility
All model files are compatible with llama.cpp, facilitating local deployment and use.
Model Capabilities
Chinese Text Generation
Dialogue System
Text Understanding
Use Cases
Dialogue System
Intelligent Customer Service
Deployed as an online customer service system to answer user questions.
Content Creation
Text Generation
Assist in creating articles, stories and other content.
Featured Recommended AI Models
Š 2025AIbase