M

Mimo 7B RL 0530

Developed by XiaomiMiMo
MiMo is a series of 7B parameter models trained from scratch for inference tasks. Through optimized pre-training and post-training strategies, it performs excellently in mathematical and code reasoning tasks.
Downloads 319
Release Time : 5/30/2025

Model Overview

The MiMo series of models focuses on unlocking the reasoning potential of language models. Through innovative pre-training and post-training strategies, it outperforms many larger 32B models in mathematical and code reasoning tasks.

Model Features

Optimized pre-training strategy
Adopt a three-stage data mixing strategy and an enhanced data preprocessing process to increase the density of reasoning patterns
Innovative multi-token prediction
Introduce multi-token prediction (MTP) as an additional training objective to improve performance and accelerate inference
Reinforcement learning training
Use carefully curated mathematical and code problems for reinforcement learning training and introduce a code reward mechanism
Efficient inference infrastructure
Develop a seamless scrolling engine to support MTP and enhance the robustness of the inference engine

Model Capabilities

Solving mathematical problems
Code generation and understanding
Handling complex reasoning tasks
Answering STEM questions
General language understanding

Use Cases

Education
Solving math competition problems
Solve complex problems in math competitions such as AIME
Achieved an 80.1% Pass@1 accuracy on AIME 2024
Assisting programming education
Help students understand and generate programming code
Achieved a 60.9% Pass@1 accuracy on LiveCodeBench v5
Research
Answering scientific questions
Answer questions in scientific knowledge tests such as GPQA
Achieved a 60.6% Pass@1 accuracy on GPQA Diamond
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase