Llava V1.5 13b Lora
LLaVA is an open-source multimodal chatbot, fine-tuned from LLaMA/Vicuna and trained on GPT-generated multimodal instruction-following data.
Downloads 143
Release Time : 10/26/2023
Model Overview
LLaVA is an autoregressive language model based on the Transformer architecture, primarily used for research on large multimodal models and chatbots.
Model Features
Multimodal Capability
Combines visual and language understanding to handle joint tasks involving images and text.
Instruction Following
Capable of understanding and executing complex instructions based on multimodal inputs.
Open-source Model
Developed on an open-source framework, available for research and improvement.
Model Capabilities
Image understanding
Multimodal dialogue
Visual question answering
Image caption generation
Multimodal instruction understanding
Use Cases
Academic Research
Multimodal Model Research
Used for studying joint visual-language representations and cross-modal understanding.
Chatbot Development
Serves as a foundational model for multimodal chatbots.
Educational Applications
Visual-assisted Learning
Helps students understand complex concepts through images.
Featured Recommended AI Models
Š 2025AIbase