M

M1 32b

Developed by Can111
M1-32B is a 32-billion-parameter large language model fine-tuned from Qwen2.5-32B-Instruct, specifically optimized to enhance reasoning, discussion, and decision-making capabilities in multi-agent systems.
Downloads 179
Release Time : 3/11/2025

Model Overview

This model is trained through multi-agent collaborative reasoning to improve reasoning abilities and role-aware dialogue generation in complex tasks, making it suitable for research and applications in Multi-Agent Systems (MAS).

Model Features

Enhanced Collaborative Reasoning
Trained on real multi-agent interaction trajectories, covering diverse roles such as expert recruiters, problem solvers, and evaluators.
Role-aware Dialogue Generation
Learns structured prompting to reason and respond from different expert perspectives.
Multi-agent System Optimization
Features adaptive collaboration and token budget management, making it an excellent MAS agent.

Model Capabilities

Multi-agent collaborative reasoning
Mathematical problem-solving
Programming task resolution
Cross-language text generation
Role-aware dialogue generation

Use Cases

Academic Research
Multi-agent System Research
Used to study multi-agent collaborative reasoning mechanisms and decision-making processes
Achieved performance comparable to o3-mini and DeepSeek-R1 on MATH-500 and MBPP-S tasks
Education
Mathematical Problem-solving
Assists students in solving complex mathematical problems
Demonstrated excellent performance in AIME2024 and MATH-500 tests
Featured Recommended AI Models
ยฉ 2025AIbase