Q

Qwen3 1.7B Base

Developed by Qwen
Qwen3-1.7B is the latest 1.7 billion parameter base language model in the Qwen series, featuring a three-stage pre-training system and supporting 32k context length.
Downloads 19.24k
Release Time : 4/28/2025

Model Overview

A causal language model pre-trained on 36 trillion tokens of multilingual data, focusing on general language modeling and specialized capability enhancement.

Model Features

Multi-stage pre-training system
Three-stage training process: General language modeling → STEM/coding specialization → 32k long-context adaptation
Ultra-large-scale corpus
Based on 36 trillion tokens of data in 119 languages, covering multiple domains including programming/STEM/reasoning
Training technology innovation
Utilizes global batch load balancing loss and qk layer normalization techniques to improve training stability

Model Capabilities

Multilingual text generation
Long-text understanding (32k context)
Programming code generation
Logical reasoning
STEM problem solving

Use Cases

Intelligent assistant
Multi-turn dialogue system
Dialogue applications with long-context memory support
Education
STEM problem-solving tutoring
Step-by-step solutions for math/science problems
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase