M1 32b
M
M1 32b
Developed by Can111
M1-32B is a 32-billion-parameter large language model fine-tuned from Qwen2.5-32B-Instruct, specifically optimized to enhance reasoning, discussion, and decision-making capabilities in multi-agent systems.
Downloads 179
Release Time : 3/11/2025
Model Overview
This model is trained through multi-agent collaborative reasoning to improve reasoning abilities and role-aware dialogue generation in complex tasks, making it suitable for research and applications in Multi-Agent Systems (MAS).
Model Features
Enhanced Collaborative Reasoning
Trained on real multi-agent interaction trajectories, covering diverse roles such as expert recruiters, problem solvers, and evaluators.
Role-aware Dialogue Generation
Learns structured prompting to reason and respond from different expert perspectives.
Multi-agent System Optimization
Features adaptive collaboration and token budget management, making it an excellent MAS agent.
Model Capabilities
Multi-agent collaborative reasoning
Mathematical problem-solving
Programming task resolution
Cross-language text generation
Role-aware dialogue generation
Use Cases
Academic Research
Multi-agent System Research
Used to study multi-agent collaborative reasoning mechanisms and decision-making processes
Achieved performance comparable to o3-mini and DeepSeek-R1 on MATH-500 and MBPP-S tasks
Education
Mathematical Problem-solving
Assists students in solving complex mathematical problems
Demonstrated excellent performance in AIME2024 and MATH-500 tests
Featured Recommended AI Models
ยฉ 2025AIbase