M

Math LLaVA

Developed by Zhiqiang007
Math-LLaVA-13B is an open-source multimodal large language model fine-tuned on the MathV360K dataset based on LLaVA-1.5-13B, suitable for scenarios such as multimodal reasoning and Q&A.
Downloads 106
Release Time : 6/24/2024

Model Overview

Math-LLaVA-13B is an open-source multimodal large language model obtained by fine-tuning LLaVA-1.5-13B on the carefully selected MathV360K data, mainly used for research on multimodal large language models, multimodal reasoning, and Q&A.

Model Features

Open-source model
Math-LLaVA is an open-source multimodal large language model (MLLM).
Data fine-tuning
Based on LLaVA-1.5-13B, fine-tuned on the carefully selected MathV360K data assisted by GPT4-Vision.
Wide range of application scenarios
Mainly used for research on multimodal large language models, multimodal reasoning, and Q&A.

Model Capabilities

Multimodal reasoning
Multimodal Q&A
Vision-language understanding

Use Cases

Education
Mathematical problem solving
Solve mathematical problems through a combination of images and text
Research
Research on multimodal large language models
Used to explore the performance and capabilities of multimodal large language models
Featured Recommended AI Models
ยฉ 2025AIbase