Q

Qwen3 0.6B Base Unsloth Bnb 4bit

Developed by unsloth
Qwen3-0.6B-Base is the latest generation of large language models in the Tongyi series. It has a parameter scale of 0.6B, supports 119 languages, and has a context length of up to 32,768 tokens.
Downloads 10.84k
Release Time : 4/28/2025

Model Overview

Qwen3-0.6B-Base is a pre-trained causal language model that focuses on extensive language modeling and general knowledge acquisition, while also having inference capabilities and long context understanding capabilities.

Model Features

Multilingual support
Pre-trained on 36 trillion tokens in 119 languages, with a wide language coverage.
Three-stage pre-training
The first stage focuses on language modeling and general knowledge acquisition; the second stage improves inference capabilities; the third stage enhances long context understanding capabilities.
Optimized training techniques
Adopts techniques such as global batch load balancing loss and qk layer normalization to improve model stability and performance.
Long context understanding
Supports a context length of up to 32,768 tokens, suitable for processing long text tasks.

Model Capabilities

Text generation
Language modeling
Multilingual processing
Long context understanding
Logical reasoning

Use Cases

Natural language processing
Multilingual text generation
Generate coherent text in multiple languages
Supports fluent generation in 119 languages
Long document summarization
Process and understand the content of long documents and generate summaries
Benefits from the support of a 32k token context length
Education
STEM question answering
Answer questions related to science, technology, engineering, and mathematics
The STEM capabilities are specifically strengthened in the second stage of pre-training
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase