L

Llava 1.5 7b Llara D Inbc Aux B VIMA 80k

Developed by variante
LLaRA is an open-source visual motion strategy model, fine-tuned from LLaVA-7b-v1.5 on instruction-following data and auxiliary datasets, primarily used for robotics research.
Downloads 390
Release Time : 7/13/2024

Model Overview

LLaRA is a large multimodal model for robotics, capable of processing visual and language inputs to generate motion strategies.

Model Features

Multimodal Processing Capability
Can simultaneously process visual and text inputs to generate corresponding motion strategies.
Specialized for Robotics
Designed and optimized specifically for robotic applications, suitable for vision-language strategy research.
Open-Source Model
Open-sourced under Apache 2.0 license, facilitating research and extension.

Model Capabilities

Vision-Language Understanding
Motion Strategy Generation
Multimodal Instruction Following

Use Cases

Robotics
Visual Instruction Following
Generates robot motion strategies based on visual and language instructions.
Multimodal Task Planning
Performs complex task planning by combining visual and language inputs.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase