Yukang commited on
Commit
c6db779
·
verified ·
1 Parent(s): 99eaa0a

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-7B-Instruct
3
- datasets: open-r1/OpenR1-Math-220k
4
  library_name: transformers
5
  model_name: Qwen2.5-7B-Open-R1-GRPO
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - grpo
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for Qwen2.5-7B-Open-R1-GRPO
15
 
16
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the [open-r1/OpenR1-Math-220k](https://huggingface.co/datasets/open-r1/OpenR1-Math-220k) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/chenyukang2020-nvidia/huggingface/runs/1ev28g0h)
33
 
34
 
35
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
1
  ---
2
  base_model: Qwen/Qwen2.5-7B-Instruct
 
3
  library_name: transformers
4
  model_name: Qwen2.5-7B-Open-R1-GRPO
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - grpo
9
  licence: license
 
11
 
12
  # Model Card for Qwen2.5-7B-Open-R1-GRPO
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/chenyukang2020-nvidia/huggingface/runs/5jg0las6)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.302693507703102e-05,
4
- "train_runtime": 483.3009,
5
  "train_samples": 93733,
6
- "train_samples_per_second": 193.943,
7
- "train_steps_per_second": 12.123
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 2.3079139061880094e-05,
4
+ "train_runtime": 485.9244,
5
  "train_samples": 93733,
6
+ "train_samples_per_second": 192.896,
7
+ "train_steps_per_second": 12.057
8
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:959738c12da49a51f7a460cbae335d3edfda5c0ce6d04f634404eb1f8986424c
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23f5aa810626a63b560cf7cd27d2bf02323878315cb3e2308653432591c922ef
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8a3a04625954332916483f50026c077127ed95c997fd01337e3cf532492713f
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83c0f9e741d69b86f126910ab16bb1f55610bf9be6a884357b3dc0e5b16a8bd5
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c33c1e57350eca13054c0f8382f178bc3ad2e9d8ac33d0e14033ab693f05149
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f979882f7b714428dfdebb4df29c5a2129581b59494fb54b78ac204d17163fd1
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1472ee2f3d1890b6d608e3943eba50bc2be7dfa444d5ea09a035c4362517abb9
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c98ea4d90ca9138f3515081c703162a207a613d8bf29a531c5ea05e46c5b3ce
3
  size 1089994880
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.302693507703102e-05,
4
- "train_runtime": 483.3009,
5
  "train_samples": 93733,
6
- "train_samples_per_second": 193.943,
7
- "train_steps_per_second": 12.123
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 2.3079139061880094e-05,
4
+ "train_runtime": 485.9244,
5
  "train_samples": 93733,
6
+ "train_samples_per_second": 192.896,
7
+ "train_steps_per_second": 12.057
8
  }
trainer_state.json CHANGED
@@ -175524,7 +175524,7 @@
175524
  "completions/min_terminated_length": 0.0,
175525
  "epoch": 0.9987198088247845,
175526
  "frac_reward_zero_std": 1.0,
175527
- "grad_norm": 0.08476304346639023,
175528
  "kl": 0.42578125,
175529
  "learning_rate": 1.4376004014216017e-10,
175530
  "loss": 0.017,
@@ -175554,8 +175554,8 @@
175554
  "completions/min_terminated_length": 0.0,
175555
  "epoch": 0.9988905009814799,
175556
  "frac_reward_zero_std": 1.0,
175557
- "grad_norm": 0.05497211542384655,
175558
- "kl": 0.42333984375,
175559
  "learning_rate": 1.1358823698404131e-10,
175560
  "loss": 0.0169,
175561
  "num_tokens": 3061130200.0,
@@ -175584,10 +175584,10 @@
175584
  "completions/min_terminated_length": 0.0,
175585
  "epoch": 0.9990611931381753,
175586
  "frac_reward_zero_std": 1.0,
175587
- "grad_norm": 0.09253000509661986,
175588
- "kl": 0.4189453125,
175589
  "learning_rate": 8.696603252866808e-11,
175590
- "loss": 0.0168,
175591
  "num_tokens": 3061696328.0,
175592
  "reward": 0.0,
175593
  "reward_std": 0.0,
@@ -175614,10 +175614,10 @@
175614
  "completions/min_terminated_length": 0.0,
175615
  "epoch": 0.9992318852948707,
175616
  "frac_reward_zero_std": 1.0,
175617
- "grad_norm": 0.051989839520072835,
175618
- "kl": 0.44775390625,
175619
  "learning_rate": 6.389343622403844e-11,
175620
- "loss": 0.0179,
175621
  "num_tokens": 3062258328.0,
175622
  "reward": 0.0,
175623
  "reward_std": 0.0,
@@ -175644,10 +175644,10 @@
175644
  "completions/min_terminated_length": 0.0,
175645
  "epoch": 0.9994025774515661,
175646
  "frac_reward_zero_std": 1.0,
175647
- "grad_norm": 0.12960677022550027,
175648
- "kl": 0.4130859375,
175649
  "learning_rate": 4.437045625915737e-11,
175650
- "loss": 0.0165,
175651
  "num_tokens": 3062819496.0,
175652
  "reward": 0.0,
175653
  "reward_std": 0.0,
@@ -175674,10 +175674,10 @@
175674
  "completions/min_terminated_length": 0.0,
175675
  "epoch": 0.9995732696082615,
175676
  "frac_reward_zero_std": 1.0,
175677
- "grad_norm": 0.09159924740428128,
175678
- "kl": 0.41845703125,
175679
  "learning_rate": 2.839709956625747e-11,
175680
- "loss": 0.0167,
175681
  "num_tokens": 3063382760.0,
175682
  "reward": 0.0,
175683
  "reward_std": 0.0,
@@ -175704,10 +175704,10 @@
175704
  "completions/min_terminated_length": 0.0,
175705
  "epoch": 0.999743961764957,
175706
  "frac_reward_zero_std": 1.0,
175707
- "grad_norm": 0.12075724493713264,
175708
- "kl": 0.40185546875,
175709
  "learning_rate": 1.5973371813027273e-11,
175710
- "loss": 0.0161,
175711
  "num_tokens": 3063952840.0,
175712
  "reward": 0.0,
175713
  "reward_std": 0.0,
@@ -175734,10 +175734,10 @@
175734
  "completions/min_terminated_length": 0.0,
175735
  "epoch": 0.9999146539216524,
175736
  "frac_reward_zero_std": 1.0,
175737
- "grad_norm": 0.29393920311022953,
175738
- "kl": 0.4228515625,
175739
  "learning_rate": 7.099277411493077e-12,
175740
- "loss": 0.0169,
175741
  "num_tokens": 3064519096.0,
175742
  "reward": 0.0,
175743
  "reward_std": 0.0,
@@ -175753,10 +175753,10 @@
175753
  "epoch": 0.9999146539216524,
175754
  "step": 5858,
175755
  "total_flos": 0.0,
175756
- "train_loss": 2.302693507703102e-05,
175757
- "train_runtime": 483.3009,
175758
- "train_samples_per_second": 193.943,
175759
- "train_steps_per_second": 12.123
175760
  }
175761
  ],
175762
  "logging_steps": 1,
 
175524
  "completions/min_terminated_length": 0.0,
175525
  "epoch": 0.9987198088247845,
175526
  "frac_reward_zero_std": 1.0,
175527
+ "grad_norm": 0.08476367046645225,
175528
  "kl": 0.42578125,
175529
  "learning_rate": 1.4376004014216017e-10,
175530
  "loss": 0.017,
 
175554
  "completions/min_terminated_length": 0.0,
175555
  "epoch": 0.9988905009814799,
175556
  "frac_reward_zero_std": 1.0,
175557
+ "grad_norm": 0.055872854179061585,
175558
+ "kl": 0.4228515625,
175559
  "learning_rate": 1.1358823698404131e-10,
175560
  "loss": 0.0169,
175561
  "num_tokens": 3061130200.0,
 
175584
  "completions/min_terminated_length": 0.0,
175585
  "epoch": 0.9990611931381753,
175586
  "frac_reward_zero_std": 1.0,
175587
+ "grad_norm": 0.14358791242151217,
175588
+ "kl": 0.408203125,
175589
  "learning_rate": 8.696603252866808e-11,
175590
+ "loss": 0.0163,
175591
  "num_tokens": 3061696328.0,
175592
  "reward": 0.0,
175593
  "reward_std": 0.0,
 
175614
  "completions/min_terminated_length": 0.0,
175615
  "epoch": 0.9992318852948707,
175616
  "frac_reward_zero_std": 1.0,
175617
+ "grad_norm": 0.07870414983242291,
175618
+ "kl": 0.4404296875,
175619
  "learning_rate": 6.389343622403844e-11,
175620
+ "loss": 0.0176,
175621
  "num_tokens": 3062258328.0,
175622
  "reward": 0.0,
175623
  "reward_std": 0.0,
 
175644
  "completions/min_terminated_length": 0.0,
175645
  "epoch": 0.9994025774515661,
175646
  "frac_reward_zero_std": 1.0,
175647
+ "grad_norm": 0.16919090255532782,
175648
+ "kl": 0.41943359375,
175649
  "learning_rate": 4.437045625915737e-11,
175650
+ "loss": 0.0168,
175651
  "num_tokens": 3062819496.0,
175652
  "reward": 0.0,
175653
  "reward_std": 0.0,
 
175674
  "completions/min_terminated_length": 0.0,
175675
  "epoch": 0.9995732696082615,
175676
  "frac_reward_zero_std": 1.0,
175677
+ "grad_norm": 0.260652148611717,
175678
+ "kl": 0.4287109375,
175679
  "learning_rate": 2.839709956625747e-11,
175680
+ "loss": 0.0172,
175681
  "num_tokens": 3063382760.0,
175682
  "reward": 0.0,
175683
  "reward_std": 0.0,
 
175704
  "completions/min_terminated_length": 0.0,
175705
  "epoch": 0.999743961764957,
175706
  "frac_reward_zero_std": 1.0,
175707
+ "grad_norm": 0.046304060159445125,
175708
+ "kl": 0.3916015625,
175709
  "learning_rate": 1.5973371813027273e-11,
175710
+ "loss": 0.0156,
175711
  "num_tokens": 3063952840.0,
175712
  "reward": 0.0,
175713
  "reward_std": 0.0,
 
175734
  "completions/min_terminated_length": 0.0,
175735
  "epoch": 0.9999146539216524,
175736
  "frac_reward_zero_std": 1.0,
175737
+ "grad_norm": 0.3306328504937249,
175738
+ "kl": 0.44287109375,
175739
  "learning_rate": 7.099277411493077e-12,
175740
+ "loss": 0.0177,
175741
  "num_tokens": 3064519096.0,
175742
  "reward": 0.0,
175743
  "reward_std": 0.0,
 
175753
  "epoch": 0.9999146539216524,
175754
  "step": 5858,
175755
  "total_flos": 0.0,
175756
+ "train_loss": 2.3079139061880094e-05,
175757
+ "train_runtime": 485.9244,
175758
+ "train_samples_per_second": 192.896,
175759
+ "train_steps_per_second": 12.057
175760
  }
175761
  ],
175762
  "logging_steps": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5afb542630ee6b0422d7310f11680454d6f498ab075bef12a859c2be322da28b
3
  size 8440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b01049e2e10045d3bc2aab78145e486cb1e2eb9a88196800173dcb1c1fe37e2
3
  size 8440