Q

Qwen3 14B Base

Developed by unsloth
Qwen3-14B-Base is the latest generation of the Tongyi series of large language models, providing a comprehensive set of dense and mixture-of-experts (MoE) models with significant improvements in training data, model architecture, and optimization techniques.
Downloads 4,693
Release Time : 4/28/2025

Model Overview

Qwen3-14B-Base is a pre-trained causal language model that supports multiple languages and tasks, with strong reasoning and long-context understanding capabilities.

Model Features

Extended high-quality pre-training corpus
Pre-trained on 36 trillion tokens in 119 languages, with a language coverage three times that of Qwen2.5 and containing richer high-quality data.
Improvements in training technology and model architecture
Adopts global batch load balancing loss and qk layer normalization to improve stability and overall performance.
Three-stage pre-training
The first stage focuses on language modeling and general knowledge acquisition, the second stage improves reasoning ability, and the third stage enhances long-context understanding ability.
Hyperparameter tuning based on scaling laws
Through comprehensive scaling law research, key hyperparameters are systematically adjusted to achieve better training dynamics and final performance.

Model Capabilities

Text generation
Language modeling
Logical reasoning
Long-context understanding
Multilingual support

Use Cases

Natural language processing
Text generation
Generate high-quality natural language text, suitable for scenarios such as writing assistance and content creation.
Logical reasoning
Solve complex logical reasoning problems, suitable for problem-solving in the STEM field.
Multilingual applications
Multilingual translation
Supports translation tasks between multiple languages.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase