hqfang commited on
Commit
b1258a4
Β·
verified Β·
1 Parent(s): 6f8bccf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -24,15 +24,16 @@ tags:
24
  MolmoAct is a fully open-source action reasoning model for robotic manipulation developed by the Allen Institute for AI, as described in their paper [MolmoAct: Action Reasoning Models that can Reason in Space](https://huggingface.co/papers/2508.07917).
25
 
26
  MolmoAct is trained on a subset of OXE and MolmoAct Dataset, a dataset with 10k high-quality trajectories of a single-arm Franka robot performing 93 unique manipulation tasks in both home and tabletop environments. It has state-of-the-art performance among vision-language-action models on multiple benchmarks while being fully open-source. You can find all models in the MolmoAct family [here](https://huggingface.co/collections/allenai/molmoact-689697591a3936fba38174d7).
 
27
 
28
  **MolmoAct 7B-D Pretrain RT-1** is based on [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) and uses [SigLip2](https://huggingface.co/google/siglip2-so400m-patch14-384) as the vision backbone, which is initialized using Molmo's pre-training approach. It is first pre-trained on MolmoAct's [Pre-training Mixture](https://huggingface.co/datasets/allenai/MolmoAct-Pretraining-Mixture), and then fine-tuned on RT-1 data using the same configuration of mid-training. This model is intended to be used for replicating our fine-tuned results on SimplerEnv (Google Robot).
29
 
30
  This checkpoint is a **preview** of the MolmoAct release. All artifacts used in creating MolmoAct (data, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
31
 
32
  Quick links:
33
- - πŸ“ƒ [Paper](https://huggingface.co/papers/2508.07917)
 
34
  - πŸ’» [Code](https://github.com/allenai/MolmoAct)
35
- - πŸ“ [Blog Post / Project Page](https://allenai.org/blog/molmoact)
36
  - πŸ“‚ [All Models](https://huggingface.co/collections/allenai/molmoact-689697591a3936fba38174d7)
37
  - πŸ“‚ [All Data](https://huggingface.co/collections/allenai/molmoact-data-mixture-6897e583e13b6c2cf3ea2b80)
38
  - πŸŽ₯ [Video](https://youtu.be/-_wag1X25OE?si=Xi_kUaJTmcQBx1f6)
 
24
  MolmoAct is a fully open-source action reasoning model for robotic manipulation developed by the Allen Institute for AI, as described in their paper [MolmoAct: Action Reasoning Models that can Reason in Space](https://huggingface.co/papers/2508.07917).
25
 
26
  MolmoAct is trained on a subset of OXE and MolmoAct Dataset, a dataset with 10k high-quality trajectories of a single-arm Franka robot performing 93 unique manipulation tasks in both home and tabletop environments. It has state-of-the-art performance among vision-language-action models on multiple benchmarks while being fully open-source. You can find all models in the MolmoAct family [here](https://huggingface.co/collections/allenai/molmoact-689697591a3936fba38174d7).
27
+ **Learn more about MolmoAct** in our announcement [blog post](https://allenai.org/blog/molmoact) or the [paper](https://arxiv.org/abs/2508.07917).
28
 
29
  **MolmoAct 7B-D Pretrain RT-1** is based on [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) and uses [SigLip2](https://huggingface.co/google/siglip2-so400m-patch14-384) as the vision backbone, which is initialized using Molmo's pre-training approach. It is first pre-trained on MolmoAct's [Pre-training Mixture](https://huggingface.co/datasets/allenai/MolmoAct-Pretraining-Mixture), and then fine-tuned on RT-1 data using the same configuration of mid-training. This model is intended to be used for replicating our fine-tuned results on SimplerEnv (Google Robot).
30
 
31
  This checkpoint is a **preview** of the MolmoAct release. All artifacts used in creating MolmoAct (data, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
32
 
33
  Quick links:
34
+ - πŸ“ [Blog Post](https://allenai.org/blog/molmoact)
35
+ - πŸ“ƒ [Paper](https://arxiv.org/abs/2508.07917)
36
  - πŸ’» [Code](https://github.com/allenai/MolmoAct)
 
37
  - πŸ“‚ [All Models](https://huggingface.co/collections/allenai/molmoact-689697591a3936fba38174d7)
38
  - πŸ“‚ [All Data](https://huggingface.co/collections/allenai/molmoact-data-mixture-6897e583e13b6c2cf3ea2b80)
39
  - πŸŽ₯ [Video](https://youtu.be/-_wag1X25OE?si=Xi_kUaJTmcQBx1f6)