P

Poseless 3B

Developed by Menlo
Poseless-3B is a vision-language model (VLM)-based robotic hand control framework that directly maps 2D images to joint angles without explicit pose estimation.
Downloads 65
Release Time : 3/3/2025

Model Overview

This model leverages projective representations and synthetic training data to achieve zero-shot generalization in real-world scenarios and cross-morphology transfer from robotic hands to human hands. By projecting visual inputs and employing a Transformer-based decoder, PoseLess addresses challenges such as depth ambiguity and data scarcity while enabling robust, low-latency control.

Model Features

Depth-free Visual-to-Joint Control
Directly maps 2D images to joint angles via projective representations without explicit pose estimation.
Synthetic Data Generation
Utilizes synthetic training data generated from random joint configurations, reducing reliance on expensive annotated datasets.
Cross-morphology Generalization
Trained solely on robotic hand data, it can mimic human hand movements, demonstrating cross-morphology generalization capabilities.
Low-latency Control
Employs a Transformer-based decoder for robust, low-latency control.

Model Capabilities

Image-to-joint angle mapping
Robotic hand control
Cross-morphology generalization
Depth-free visual processing

Use Cases

Robotic Control
Robotic Hand Pose Control
Directly controls robotic hand joint angles via monocular images.
Without relying on any manually annotated datasets, the model achieves competitive accuracy in joint angle prediction.
Human-Robot Interaction
Human Hand Pose Imitation
Mimics human hand movements using training data from robotic hands.
Demonstrates the model's potential for cross-morphology generalization.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase