L

Llava V1.5 13B AWQ

Developed by TheBloke
LLaVA is an open-source multimodal chatbot, fine-tuned on GPT-generated multimodal instruction-following data based on LLaMA/Vicuna.
Downloads 141
Release Time : 10/15/2023

Model Overview

LLaVA is an autoregressive language model based on the transformer architecture, capable of understanding and generating text related to images.

Model Features

Multimodal Understanding
Capable of processing both image and text inputs, understanding the relationship between them.
Instruction Following
Can execute tasks by following complex multimodal instructions.
Open-source
The model is fully open-source, available for both research and commercial use.

Model Capabilities

Visual Question Answering
Image Caption Generation
Multimodal Dialogue
Instruction Following

Use Cases

Research
Multimodal Model Research
Used to study the behavior and capabilities of large multimodal models.
Education
Visual-assisted Learning
Helps students understand complex concepts through images.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase