hungchiayu commited on
Commit
35900b7
·
verified ·
1 Parent(s): 0240a71

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -7,8 +7,8 @@ tags: []
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
- Nora is an open vision-language-action model trained on robot manipulation episodes from the Open X-Embodiment dataset. The model takes language instructions and camera images as input and generates robot actions. Nora is trained directly from Qwen 2.5 VL-3B.
11
- All Nora checkpoints, as well as our training codebase are released under an MIT License.
12
 
13
 
14
 
@@ -24,15 +24,18 @@ All Nora checkpoints, as well as our training codebase are released under an MIT
24
  - **License:** MIT
25
  - **Finetuned from model :** Qwen 2.5 VL-3B
26
 
27
- ### Model Sources [optional]
28
 
29
  <!-- Provide the basic links for the model. -->
30
 
31
  - **Repository:** https://github.com/declare-lab/nora
32
- - **Paper [optional]:** https://www.arxiv.org/abs/2504.19854
33
- - **Demo [optional]:** https://declare-lab.github.io/nora
34
 
 
35
 
 
 
36
 
37
 
38
 
 
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ Nora is an open vision-language-action model trained on robot manipulation episodes from the [Open X-Embodiment](https://robotics-transformer-x.github.io/) dataset. The model takes language instructions and camera images as input and generates robot actions. Nora is trained directly from Qwen 2.5 VL-3B.
11
+ All Nora checkpoints, as well as our [training codebase](https://github.com/declare-lab/nora) are released under an MIT License.
12
 
13
 
14
 
 
24
  - **License:** MIT
25
  - **Finetuned from model :** Qwen 2.5 VL-3B
26
 
27
+ ### Model Sources
28
 
29
  <!-- Provide the basic links for the model. -->
30
 
31
  - **Repository:** https://github.com/declare-lab/nora
32
+ - **Paper :** https://www.arxiv.org/abs/2504.19854
33
+ - **Demo:** https://declare-lab.github.io/nora
34
 
35
+ ## Usage
36
 
37
+ Nora take a language instruction and a camera image of a robot workspace as input, and predict (normalized) robot actions consisting of 7-DoF end-effector deltas of the form (x, y, z, roll, pitch, yaw, gripper).
38
+ To execute on an actual robot platform, actions need to be un-normalized subject to statistics computed on a per-robot, per-dataset basis.
39
 
40
 
41