Q

Qwen3 30B A3B Base

Developed by Qwen
Qwen3-30B-A3B-Base is the latest 30.5B parameter-scale Mixture of Experts (MoE) large language model in the Qwen series, supporting 119 languages and 32k context length.
Downloads 9,745
Release Time : 4/28/2025

Model Overview

A MoE architecture large language model developed based on a three-stage pre-training framework, focusing on general language modeling, STEM/programming capability enhancement, and long-context understanding.

Model Features

Multilingual Coverage
Pre-training data covers 36 trillion tokens across 119 languages, with language coverage three times that of the previous generation.
MoE Architecture Optimization
Utilizes global batch load balancing loss and qk layer normalization techniques to improve training stability.
Three-stage Pre-training
Phased enhancement of general capabilities, STEM/programming/reasoning abilities, and long-context understanding.

Model Capabilities

Multilingual text generation
Code generation and completion
Complex logical reasoning
Long document understanding
Mathematical problem solving

Use Cases

Development Assistance
Code Auto-completion
Supports code generation and error fixing for multiple programming languages
Education & Research
STEM Problem Solving
Solves complex problems in mathematics, physics, and other disciplines
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase