Qwen3 1.7B Base Unsloth Bnb 4bit
Qwen3-1.7B-Base is the latest generation large language model in the Qwen series, providing high-quality pretrained language modeling capabilities.
Downloads 689
Release Time : 4/28/2025
Model Overview
Qwen3-1.7B-Base is a 1.7 billion parameter causal language model focused on general language modeling and knowledge acquisition, supporting 32k ultra-long context understanding.
Model Features
High-Quality Pretraining Corpus
Covers 36 trillion tokens across 119 languages, significantly increasing the proportion of high-value content such as programming, STEM, and reasoning.
Training Techniques and Architecture Optimization
Employs innovative techniques like MoE model global batch load balancing loss and full-model qk layer normalization to enhance training stability and performance.
Three-Stage Pretraining System
Phased reinforcement of general language modeling, STEM/programming/logical reasoning capabilities, and long-context understanding.
Ultra-Long Context Support
Supports 32k ultra-long context training, enhancing long-text comprehension capabilities.
Model Capabilities
Text Generation
Language Understanding
Programming Capability
Logical Reasoning
Long-Text Processing
Use Cases
Natural Language Processing
Text Generation
Generate high-quality natural language text
Programming Assistance
Assist in writing and optimizing code
Education
STEM Education
Assist in learning and teaching in STEM fields
Featured Recommended AI Models
Š 2025AIbase