L

Liuhaotian Llava V1.5 13b GGUF

Developed by PsiPi
LLaVA is an open-source multimodal chatbot, based on the LLaMA/Vicuna architecture, fine-tuned with multimodal instruction-following data.
Downloads 1,225
Release Time : 12/1/2023

Model Overview

LLaVA is a research-oriented large multimodal model primarily used in the fields of computer vision, natural language processing, and artificial intelligence research.

Model Features

Multimodal Capability
Capable of processing both image and text inputs for cross-modal understanding
Instruction Following
Specially trained to follow multimodal instructions
Open-source Model
Released under an open-source license for research and development use
End-to-End Inference
Supports dependency-free inference via llama.cpp

Model Capabilities

Image-text dialogue
Visual question answering
Image caption generation
Multimodal instruction following
Cross-modal understanding

Use Cases

Academic Research
Multimodal Model Research
Used to study the performance and capability boundaries of large multimodal models
Human-Computer Interaction Research
Explores multimodal human-computer interaction based on vision and language
Educational Applications
Visual-Assisted Learning
Helps students understand complex visual content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase