Minicpm MoE 8x2B
MiniCPM-MoE-8x2B is a Transformer-based Mixture of Experts (MoE) language model, designed with 8 expert modules where each token activates 2 experts for processing.
Downloads 6,377
Release Time : 4/7/2024
Model Overview
A pure decoder generative language model, fine-tuned with instructions but without RLHF methods, suitable for natural language processing tasks.
Model Features
Mixture of Experts Architecture
Designed with MoE architecture, each layer contains 8 expert modules, with each token activating 2 experts for processing, improving model efficiency.
Instruction Fine-tuning
The model is optimized through instruction fine-tuning but without RLHF methods, making it suitable for specific task processing.
Efficient Inference
Supports inference using the vLLM framework, providing higher throughput.
Model Capabilities
Text Generation
Question Answering System
Dialogue System
Use Cases
Intelligent Q&A
Geographical Knowledge Q&A
Answer complex questions about geographical knowledge, such as comparing the heights of different mountains.
Can accurately answer that Mount Tai is the highest mountain in Shandong Province and compare its height difference with Huangshan Mountain.
Dialogue System
Open-domain Dialogue
Engage in natural and fluent open-domain conversations.
Featured Recommended AI Models