Update README.md
Browse files
README.md
CHANGED
@@ -36,6 +36,18 @@ We introduce **Being-H0**, the first dexterous Vision-Language-Action model pret
|
|
36 |
- **[2025-08-02]**: We release the **Being-H0** codebase and pretrained models! Check our [Hugging Face Model Hub](https://huggingface.co/BeingBeyond/Being-H0) for more details. π₯π₯π₯
|
37 |
- **[2025-07-21]**: We publish **Being-H0**! Check our paper [here](https://arxiv.org/abs/2507.15597). πππ
|
38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
## Setup
|
40 |
|
41 |
### Clone repository
|
@@ -65,18 +77,6 @@ pip install git+https://github.com/mattloper/chumpy.git
|
|
65 |
- Download Models and Code (the downloaded file should have the format `mano_v*_*.zip`). Note that all code and data from this download falls under the [MANO license](http://mano.is.tue.mpg.de/license).
|
66 |
- unzip and copy the contents in `mano_v*_*/` folder to the `beingvla/models/motion/mano/` folder
|
67 |
|
68 |
-
## Model Checkpoints
|
69 |
-
|
70 |
-
Download pre-trained models from Hugging Face:
|
71 |
-
|
72 |
-
| Model Type | Model Name | Parameters | Description |
|
73 |
-
|------------|------------|------------|-------------|
|
74 |
-
| **Motion Model** | [Being-H0-GRVQ-8K](https://huggingface.co/BeingBeyond/Being-H0-GRVQ-8K) | - | Motion tokenizer |
|
75 |
-
| **VLA Pre-trained** | [Being-H0-1B-2508](https://huggingface.co/BeingBeyond/Being-H0-1B-2508) | 1B | Base vision-language-action model |
|
76 |
-
| **VLA Pre-trained** | [Being-H0-8B-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-2508) | 8B | Base vision-language-action model |
|
77 |
-
| **VLA Pre-trained** | [Being-H0-14B-2508](https://huggingface.co/BeingBeyond/Being-H0-14B-2508) | 14B | Base vision-language-action model |
|
78 |
-
| **VLA Post-trained** | [Being-H0-8B-Align-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-Align-2508) | 8B | Fine-tuned for robot alignment |
|
79 |
-
|
80 |
## Inference
|
81 |
|
82 |
### Motion Generation
|
@@ -114,14 +114,9 @@ python -m beingvla.models.motion.m2m.aligner.eval_policy \
|
|
114 |
--action-chunk-length 16
|
115 |
```
|
116 |
|
117 |
-
##
|
118 |
-
|
119 |
-
The following features are planned for future implementation:
|
120 |
|
121 |
-
|
122 |
-
- [ ] Training code and scripts.
|
123 |
-
- [ ] Hugging Face library compatibility.
|
124 |
-
- [x] Inference code and scripts.
|
125 |
|
126 |
## Citation
|
127 |
If you find our work useful, please consider citing us and give a star to our repository! πππ
|
|
|
36 |
- **[2025-08-02]**: We release the **Being-H0** codebase and pretrained models! Check our [Hugging Face Model Hub](https://huggingface.co/BeingBeyond/Being-H0) for more details. π₯π₯π₯
|
37 |
- **[2025-07-21]**: We publish **Being-H0**! Check our paper [here](https://arxiv.org/abs/2507.15597). πππ
|
38 |
|
39 |
+
## Model Checkpoints
|
40 |
+
|
41 |
+
Download pre-trained models from Hugging Face:
|
42 |
+
|
43 |
+
| Model Type | Model Name | Parameters | Description |
|
44 |
+
|------------|------------|------------|-------------|
|
45 |
+
| **Motion Model** | [Being-H0-GRVQ-8K](https://huggingface.co/BeingBeyond/Being-H0-GRVQ-8K) | - | Motion tokenizer |
|
46 |
+
| **VLA Pre-trained** | [Being-H0-1B-2508](https://huggingface.co/BeingBeyond/Being-H0-1B-2508) | 1B | Base vision-language-action model |
|
47 |
+
| **VLA Pre-trained** | [Being-H0-8B-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-2508) | 8B | Base vision-language-action model |
|
48 |
+
| **VLA Pre-trained** | [Being-H0-14B-2508](https://huggingface.co/BeingBeyond/Being-H0-14B-2508) | 14B | Base vision-language-action model |
|
49 |
+
| **VLA Post-trained** | [Being-H0-8B-Align-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-Align-2508) | 8B | Fine-tuned for robot alignment |
|
50 |
+
|
51 |
## Setup
|
52 |
|
53 |
### Clone repository
|
|
|
77 |
- Download Models and Code (the downloaded file should have the format `mano_v*_*.zip`). Note that all code and data from this download falls under the [MANO license](http://mano.is.tue.mpg.de/license).
|
78 |
- unzip and copy the contents in `mano_v*_*/` folder to the `beingvla/models/motion/mano/` folder
|
79 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
80 |
## Inference
|
81 |
|
82 |
### Motion Generation
|
|
|
114 |
--action-chunk-length 16
|
115 |
```
|
116 |
|
117 |
+
## Contributing and Building on Being-H0
|
|
|
|
|
118 |
|
119 |
+
We encourage researchers and practitioners to leverage Being-H0 as a foundation for their own creative experiments and applications. Whether you're adapting Being-H0 to new robotic platforms, exploring novel hand manipulation tasks, or extending the model to new domains, our modular codebase is designed to support your innovations. We welcome contributions of all kinds - from bug fixes and documentation improvements to new features and model architectures. By building on Being-H0 together, we can advance the field of dexterous vision-language-action modeling and enable robots to understand and replicate the rich complexity of human hand movements. Join us in making robotic manipulation more intuitive, capable, and accessible to all.
|
|
|
|
|
|
|
120 |
|
121 |
## Citation
|
122 |
If you find our work useful, please consider citing us and give a star to our repository! πππ
|