Qwen2.5 Math 7B RoPE 300k
Apache-2.0
Qwen2.5-Math-7B-RoPE-300k is a variant based on Qwen2.5-Math-7B, which extends the context length to 32k tokens by adjusting the base frequency of Rotary Position Encoding (RoPE).
Large Language Model
Transformers English