Update README.md
Browse files
README.md
CHANGED
@@ -7,8 +7,8 @@ tags: []
|
|
7 |
|
8 |
<!-- Provide a quick summary of what the model is/does. -->
|
9 |
|
10 |
-
Nora is an open vision-language-action model trained on
|
11 |
-
All Nora checkpoints, as well as our training codebase are released under an MIT License.
|
12 |
|
13 |
|
14 |
|
@@ -24,15 +24,18 @@ All Nora checkpoints, as well as our training codebase are released under an MIT
|
|
24 |
- **License:** MIT
|
25 |
- **Finetuned from model :** Qwen 2.5 VL-3B
|
26 |
|
27 |
-
### Model Sources
|
28 |
|
29 |
<!-- Provide the basic links for the model. -->
|
30 |
|
31 |
- **Repository:** https://github.com/declare-lab/nora
|
32 |
-
- **Paper
|
33 |
-
- **Demo
|
34 |
|
|
|
35 |
|
|
|
|
|
36 |
|
37 |
|
38 |
|
|
|
7 |
|
8 |
<!-- Provide a quick summary of what the model is/does. -->
|
9 |
|
10 |
+
Nora is an open vision-language-action model trained on robot manipulation episodes from the [Open X-Embodiment](https://robotics-transformer-x.github.io/) dataset. The model takes language instructions and camera images as input and generates robot actions. Nora is trained directly from Qwen 2.5 VL-3B.
|
11 |
+
All Nora checkpoints, as well as our [training codebase](https://github.com/declare-lab/nora) are released under an MIT License.
|
12 |
|
13 |
|
14 |
|
|
|
24 |
- **License:** MIT
|
25 |
- **Finetuned from model :** Qwen 2.5 VL-3B
|
26 |
|
27 |
+
### Model Sources
|
28 |
|
29 |
<!-- Provide the basic links for the model. -->
|
30 |
|
31 |
- **Repository:** https://github.com/declare-lab/nora
|
32 |
+
- **Paper :** https://www.arxiv.org/abs/2504.19854
|
33 |
+
- **Demo:** https://declare-lab.github.io/nora
|
34 |
|
35 |
+
## Usage
|
36 |
|
37 |
+
Nora take a language instruction and a camera image of a robot workspace as input, and predict (normalized) robot actions consisting of 7-DoF end-effector deltas of the form (x, y, z, roll, pitch, yaw, gripper).
|
38 |
+
To execute on an actual robot platform, actions need to be un-normalized subject to statistics computed on a per-robot, per-dataset basis.
|
39 |
|
40 |
|
41 |
|