O

Openvla V01 7b

Developed by openvla
OpenVLA v0.1 7B is an open-source vision-language-action model trained on the Open X-Embodiment dataset, supporting various robot controls.
Downloads 30
Release Time : 6/10/2024

Model Overview

OpenVLA v0.1 7B is a vision-language-action model that takes language instructions and camera images as input to generate robot actions. It supports out-of-the-box control for multiple robots and can quickly adapt to new robot domains through fine-tuning.

Model Features

Multi-robot Support
Out-of-the-box control for various robots included in the pre-training data
Efficient Fine-tuning
Can be efficiently fine-tuned with minimal demonstration data to adapt to new tasks and robot setups
Open Source
All checkpoints and training codebases are released under the MIT license

Model Capabilities

Robot Action Prediction
Vision-Language Understanding
Multimodal Input Processing

Use Cases

Robot Control
Zero-shot Robot Control
Execute instructions on robot setups included in the pre-training data with zero-shot capability
Can control robots like Widow-X included in the pre-training data
New Domain Adaptation
Quickly adapt to new robot domains through fine-tuning
Requires collecting demonstration datasets for the target setup
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase