zawnpn commited on
Commit
89193b0
Β·
verified Β·
1 Parent(s): d09ed64

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -19
README.md CHANGED
@@ -36,6 +36,18 @@ We introduce **Being-H0**, the first dexterous Vision-Language-Action model pret
36
  - **[2025-08-02]**: We release the **Being-H0** codebase and pretrained models! Check our [Hugging Face Model Hub](https://huggingface.co/BeingBeyond/Being-H0) for more details. πŸ”₯πŸ”₯πŸ”₯
37
  - **[2025-07-21]**: We publish **Being-H0**! Check our paper [here](https://arxiv.org/abs/2507.15597). 🌟🌟🌟
38
 
 
 
 
 
 
 
 
 
 
 
 
 
39
  ## Setup
40
 
41
  ### Clone repository
@@ -65,18 +77,6 @@ pip install git+https://github.com/mattloper/chumpy.git
65
  - Download Models and Code (the downloaded file should have the format `mano_v*_*.zip`). Note that all code and data from this download falls under the [MANO license](http://mano.is.tue.mpg.de/license).
66
  - unzip and copy the contents in `mano_v*_*/` folder to the `beingvla/models/motion/mano/` folder
67
 
68
- ## Model Checkpoints
69
-
70
- Download pre-trained models from Hugging Face:
71
-
72
- | Model Type | Model Name | Parameters | Description |
73
- |------------|------------|------------|-------------|
74
- | **Motion Model** | [Being-H0-GRVQ-8K](https://huggingface.co/BeingBeyond/Being-H0-GRVQ-8K) | - | Motion tokenizer |
75
- | **VLA Pre-trained** | [Being-H0-1B-2508](https://huggingface.co/BeingBeyond/Being-H0-1B-2508) | 1B | Base vision-language-action model |
76
- | **VLA Pre-trained** | [Being-H0-8B-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-2508) | 8B | Base vision-language-action model |
77
- | **VLA Pre-trained** | [Being-H0-14B-2508](https://huggingface.co/BeingBeyond/Being-H0-14B-2508) | 14B | Base vision-language-action model |
78
- | **VLA Post-trained** | [Being-H0-8B-Align-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-Align-2508) | 8B | Fine-tuned for robot alignment |
79
-
80
  ## Inference
81
 
82
  ### Motion Generation
@@ -114,14 +114,9 @@ python -m beingvla.models.motion.m2m.aligner.eval_policy \
114
  --action-chunk-length 16
115
  ```
116
 
117
- ## TODO: Training and Data Features
118
-
119
- The following features are planned for future implementation:
120
 
121
- ### Data Preparation.
122
- - [ ] Training code and scripts.
123
- - [ ] Hugging Face library compatibility.
124
- - [x] Inference code and scripts.
125
 
126
  ## Citation
127
  If you find our work useful, please consider citing us and give a star to our repository! 🌟🌟🌟
 
36
  - **[2025-08-02]**: We release the **Being-H0** codebase and pretrained models! Check our [Hugging Face Model Hub](https://huggingface.co/BeingBeyond/Being-H0) for more details. πŸ”₯πŸ”₯πŸ”₯
37
  - **[2025-07-21]**: We publish **Being-H0**! Check our paper [here](https://arxiv.org/abs/2507.15597). 🌟🌟🌟
38
 
39
+ ## Model Checkpoints
40
+
41
+ Download pre-trained models from Hugging Face:
42
+
43
+ | Model Type | Model Name | Parameters | Description |
44
+ |------------|------------|------------|-------------|
45
+ | **Motion Model** | [Being-H0-GRVQ-8K](https://huggingface.co/BeingBeyond/Being-H0-GRVQ-8K) | - | Motion tokenizer |
46
+ | **VLA Pre-trained** | [Being-H0-1B-2508](https://huggingface.co/BeingBeyond/Being-H0-1B-2508) | 1B | Base vision-language-action model |
47
+ | **VLA Pre-trained** | [Being-H0-8B-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-2508) | 8B | Base vision-language-action model |
48
+ | **VLA Pre-trained** | [Being-H0-14B-2508](https://huggingface.co/BeingBeyond/Being-H0-14B-2508) | 14B | Base vision-language-action model |
49
+ | **VLA Post-trained** | [Being-H0-8B-Align-2508](https://huggingface.co/BeingBeyond/Being-H0-8B-Align-2508) | 8B | Fine-tuned for robot alignment |
50
+
51
  ## Setup
52
 
53
  ### Clone repository
 
77
  - Download Models and Code (the downloaded file should have the format `mano_v*_*.zip`). Note that all code and data from this download falls under the [MANO license](http://mano.is.tue.mpg.de/license).
78
  - unzip and copy the contents in `mano_v*_*/` folder to the `beingvla/models/motion/mano/` folder
79
 
 
 
 
 
 
 
 
 
 
 
 
 
80
  ## Inference
81
 
82
  ### Motion Generation
 
114
  --action-chunk-length 16
115
  ```
116
 
117
+ ## Contributing and Building on Being-H0
 
 
118
 
119
+ We encourage researchers and practitioners to leverage Being-H0 as a foundation for their own creative experiments and applications. Whether you're adapting Being-H0 to new robotic platforms, exploring novel hand manipulation tasks, or extending the model to new domains, our modular codebase is designed to support your innovations. We welcome contributions of all kinds - from bug fixes and documentation improvements to new features and model architectures. By building on Being-H0 together, we can advance the field of dexterous vision-language-action modeling and enable robots to understand and replicate the rich complexity of human hand movements. Join us in making robotic manipulation more intuitive, capable, and accessible to all.
 
 
 
120
 
121
  ## Citation
122
  If you find our work useful, please consider citing us and give a star to our repository! 🌟🌟🌟