Nora
Nora is an open-source vision-language-action model trained on Qwen 2.5 VL - 3B, capable of generating robot actions based on language instructions and camera images.
Downloads 7,063
Release Time : 4/28/2025
Model Overview
Nora is a vision-language-action model that takes language instructions and camera images as inputs and predicts robot actions composed of 7-DOF end-effector increments.
Model Features
Vision-language-action integration
Capable of simultaneously processing visual inputs (camera images) and language instructions to output robot actions
Open-source availability
All checkpoints and training codebases are publicly available under the MIT license
Trained on large-scale data
Trained using robot manipulation segments from the Open X-Embodiment dataset
7-DOF action prediction
Capable of predicting 7-DOF robot actions including position and pose
Model Capabilities
Vision-language understanding
Robot action prediction
Instruction following
Zero-shot learning
Use Cases
Robot control
Instruction-based robot operation
Control the robot to perform specific tasks based on natural language instructions
Capable of generating 7-DOF actions suitable for robot execution
Zero-shot instruction following
Perform tasks under unseen instructions and scenarios
Featured Recommended AI Models
Š 2025AIbase