L

Llava V1.6 Mistral 7b

Developed by liuhaotian
LLaVA is an open-source multimodal chatbot, trained by fine-tuning large language models on multimodal instruction-following data.
Downloads 27.45k
Release Time : 1/31/2024

Model Overview

LLaVA is an autoregressive language model based on the transformer architecture, capable of processing both image and text inputs to generate text outputs.

Model Features

Multimodal Capability
Can process both image and text inputs to generate relevant text outputs.
Instruction Following
Specifically trained on multimodal instruction-following data, capable of understanding and executing complex instructions.
Open-Source Model
Fully open-source, allowing researchers and developers to freely use and modify it.

Model Capabilities

Image Understanding
Multimodal Dialogue
Visual Question Answering
Instruction Following
Text Generation

Use Cases

Research
Multimodal Model Research
Used to study the behavior and capabilities of large multimodal models.
Education
Visual-Assisted Learning
Helps students learn knowledge through interactive image and text methods.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase