Q

Qwenlong L1 32B

Developed by Tongyi-Zhiwen
QwenLong-L1 is a long-context reasoning model trained with reinforcement learning, demonstrating excellent performance across seven long-context document QA benchmarks.
Downloads 683
Release Time : 5/23/2025

Model Overview

QwenLong-L1 is a long-context reasoning model trained via reinforcement learning, specializing in long-document QA tasks with robust reasoning and contextual understanding capabilities.

Model Features

Reinforcement learning training
Trained using a reinforcement learning framework, including warm-up supervised fine-tuning, curriculum-guided RL phases, and difficulty-aware retrospective sampling mechanisms.
Long-context processing
Supports context lengths up to 131,072 tokens through YaRN-based RoPE scaling.
Multi-domain reasoning
Excels in mathematical reasoning, logical reasoning, and multi-hop reasoning across various domains.

Model Capabilities

Long-text comprehension
Complex reasoning
Multi-hop QA
Mathematical reasoning
Logical analysis

Use Cases

Document QA
Financial report analysis
Extract key information from lengthy financial reports and answer related questions.
Performs excellently on the DocMath benchmark.
Legal document comprehension
Parse complex legal documents and answer relevant questions.
Demonstrates strong performance in legal-domain logical reasoning tasks.
Multi-hop reasoning
Cross-document information integration
Integrate information from multiple related documents to answer complex questions.
Performs well on MultiHopRAG and Musique datasets.
Featured Recommended AI Models
ยฉ 2025AIbase