N

Nora

Developed by declare-lab
Nora is an open-source vision-language-action model trained on Qwen 2.5 VL - 3B, capable of generating robot actions based on language instructions and camera images.
Downloads 7,063
Release Time : 4/28/2025

Model Overview

Nora is a vision-language-action model that takes language instructions and camera images as inputs and predicts robot actions composed of 7-DOF end-effector increments.

Model Features

Vision-language-action integration
Capable of simultaneously processing visual inputs (camera images) and language instructions to output robot actions
Open-source availability
All checkpoints and training codebases are publicly available under the MIT license
Trained on large-scale data
Trained using robot manipulation segments from the Open X-Embodiment dataset
7-DOF action prediction
Capable of predicting 7-DOF robot actions including position and pose

Model Capabilities

Vision-language understanding
Robot action prediction
Instruction following
Zero-shot learning

Use Cases

Robot control
Instruction-based robot operation
Control the robot to perform specific tasks based on natural language instructions
Capable of generating 7-DOF actions suitable for robot execution
Zero-shot instruction following
Perform tasks under unseen instructions and scenarios
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase