Q

Qwen1.5 MoE A2.7B GGUF

Developed by tensorblock
The Mixture of Experts (MoE) model of Tongyi Qianwen version 1.5, with a parameter scale of 2.7B, providing GGUF format files in multiple quantization versions.
Downloads 163
Release Time : 11/11/2024

Model Overview

This is a Mixture of Experts model based on the Qwen1.5 architecture, providing GGUF format files in multiple quantization versions, suitable for local inference scenarios.

Model Features

Multiple Quantization Options
Provide model files with a total of 12 different quantization levels from Q2_K to Q8_0 to meet the needs of different scenarios.
Efficient Inference
The Mixture of Experts architecture improves inference efficiency while maintaining model performance.
llama.cpp Compatibility
All model files are compatible with llama.cpp, facilitating local deployment and use.

Model Capabilities

Chinese Text Generation
Dialogue System
Text Understanding

Use Cases

Dialogue System
Intelligent Customer Service
Deployed as an online customer service system to answer user questions.
Content Creation
Text Generation
Assist in creating articles, stories and other content.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase