Math LLaVA
M
Math LLaVA
Developed by Zhiqiang007
Math-LLaVA-13B is an open-source multimodal large language model fine-tuned on the MathV360K dataset based on LLaVA-1.5-13B, suitable for scenarios such as multimodal reasoning and Q&A.
Downloads 106
Release Time : 6/24/2024
Model Overview
Math-LLaVA-13B is an open-source multimodal large language model obtained by fine-tuning LLaVA-1.5-13B on the carefully selected MathV360K data, mainly used for research on multimodal large language models, multimodal reasoning, and Q&A.
Model Features
Open-source model
Math-LLaVA is an open-source multimodal large language model (MLLM).
Data fine-tuning
Based on LLaVA-1.5-13B, fine-tuned on the carefully selected MathV360K data assisted by GPT4-Vision.
Wide range of application scenarios
Mainly used for research on multimodal large language models, multimodal reasoning, and Q&A.
Model Capabilities
Multimodal reasoning
Multimodal Q&A
Vision-language understanding
Use Cases
Education
Mathematical problem solving
Solve mathematical problems through a combination of images and text
Research
Research on multimodal large language models
Used to explore the performance and capabilities of multimodal large language models
Featured Recommended AI Models
ยฉ 2025AIbase