hamedkharazmi commited on
Commit
ca6c5ea
·
verified ·
1 Parent(s): 7f411c8

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Gensyn/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mangy_padded_cow
5
  tags:
@@ -14,7 +14,7 @@ licence: license
14
 
15
  # Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mangy_padded_cow
16
 
17
- This model is a fine-tuned version of [Gensyn/Qwen2.5-0.5B-Instruct](https://huggingface.co/Gensyn/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
@@ -37,10 +37,10 @@ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing
37
 
38
  ### Framework versions
39
 
40
- - TRL: 0.15.2
41
  - Transformers: 4.51.3
42
- - Pytorch: 2.6.0
43
- - Datasets: 3.5.0
44
  - Tokenizers: 0.21.1
45
 
46
  ## Citations
@@ -62,7 +62,7 @@ Cite TRL as:
62
  ```bibtex
63
  @misc{vonwerra2022trl,
64
  title = {{TRL: Transformer Reinforcement Learning}},
65
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
66
  year = 2020,
67
  journal = {GitHub repository},
68
  publisher = {GitHub},
 
1
  ---
2
+ base_model: unsloth/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mangy_padded_cow
5
  tags:
 
14
 
15
  # Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mangy_padded_cow
16
 
17
+ This model is a fine-tuned version of [unsloth/Qwen2.5-0.5B-Instruct](https://huggingface.co/unsloth/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
 
37
 
38
  ### Framework versions
39
 
40
+ - TRL: 0.17.0
41
  - Transformers: 4.51.3
42
+ - Pytorch: 2.7.0
43
+ - Datasets: 3.5.1
44
  - Tokenizers: 0.21.1
45
 
46
  ## Citations
 
62
  ```bibtex
63
  @misc{vonwerra2022trl,
64
  title = {{TRL: Transformer Reinforcement Learning}},
65
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
66
  year = 2020,
67
  journal = {GitHub repository},
68
  publisher = {GitHub},
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.0,
4
- "train_runtime": 863.957,
5
- "train_samples": 12,
6
- "train_samples_per_second": 0.093,
7
- "train_steps_per_second": 0.006
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.27013154824574787,
4
+ "train_runtime": 1786.8594,
5
+ "train_samples": 23,
6
+ "train_samples_per_second": 0.013,
7
+ "train_steps_per_second": 0.002
8
  }
config.json CHANGED
@@ -3,7 +3,6 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 896,
@@ -15,13 +14,15 @@
15
  "num_attention_heads": 14,
16
  "num_hidden_layers": 24,
17
  "num_key_value_heads": 2,
 
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
- "sliding_window": 32768,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.3",
 
25
  "use_cache": true,
26
  "use_sliding_window": false,
27
  "vocab_size": 151936
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
 
6
  "eos_token_id": 151645,
7
  "hidden_act": "silu",
8
  "hidden_size": 896,
 
14
  "num_attention_heads": 14,
15
  "num_hidden_layers": 24,
16
  "num_key_value_heads": 2,
17
+ "pad_token_id": 151654,
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.3",
25
+ "unsloth_fixed": true,
26
  "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
generation_config.json CHANGED
@@ -5,7 +5,8 @@
5
  151645,
6
  151643
7
  ],
8
- "pad_token_id": 151643,
 
9
  "repetition_penalty": 1.1,
10
  "temperature": 0.7,
11
  "top_k": 20,
 
5
  151645,
6
  151643
7
  ],
8
+ "max_length": 32768,
9
+ "pad_token_id": 151654,
10
  "repetition_penalty": 1.1,
11
  "temperature": 0.7,
12
  "top_k": 20,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc6798eb79dc628e086c92abff71e217c8353df86dcc554b4d9ae22c5af6daa6
3
  size 1976163472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:042df9bab814993de332623fa5c40266568096ef8b903104ca3c676525f23d13
3
  size 1976163472
special_tokens_map.json CHANGED
@@ -22,7 +22,7 @@
22
  "single_word": false
23
  },
24
  "pad_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
 
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|vision_pad|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eee858c5123a4279c3e1f7b81247343f356ac767940b2692a928ad929543214
3
- size 11422063
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e71213db910f5cafa86d35091f37393dcc344b1bbc34091d1b3eed4cca01d5
3
+ size 11422064
tokenizer_config.json CHANGED
@@ -200,8 +200,9 @@
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
- "model_max_length": 131072,
204
- "pad_token": "<|endoftext|>",
 
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
 
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
+ "model_max_length": 32768,
204
+ "pad_token": "<|vision_pad|>",
205
+ "padding_side": "left",
206
  "split_special_tokens": false,
207
  "tokenizer_class": "Qwen2Tokenizer",
208
  "unk_token": null
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.0,
4
- "train_runtime": 863.957,
5
- "train_samples": 12,
6
- "train_samples_per_second": 0.093,
7
- "train_steps_per_second": 0.006
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.27013154824574787,
4
+ "train_runtime": 1786.8594,
5
+ "train_samples": 23,
6
+ "train_samples_per_second": 0.013,
7
+ "train_steps_per_second": 0.002
8
  }
trainer_state.json CHANGED
@@ -2,77 +2,98 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 2.6666666666666665,
6
  "eval_steps": 500,
7
- "global_step": 5,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "completion_length": 163.66666666666666,
14
- "epoch": 1.0,
15
- "grad_norm": 3.171876650220895e-10,
 
 
 
 
 
 
 
 
 
 
 
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
- "loss": 0.0,
19
- "reward": 7.75,
20
- "reward_std": 0.0,
21
- "rewards/concensus_correctness_reward_func": 2.5,
22
- "rewards/consensus_reward_func": 2.0,
23
- "rewards/cumulative_reward_2": 0.0,
24
- "rewards/final_correctness_reward_func": 0.5,
25
- "rewards/question_recreation_reward_func": 1.0,
26
- "rewards/soft_format_reward_func": 0.0,
27
- "rewards/strict_format_reward_func": 0.5,
28
- "rewards/xmlcount_reward_func": 1.25,
 
 
 
 
 
 
 
 
 
29
  "step": 2
30
  },
31
  {
32
- "completion_length": 161.16666666666666,
33
- "epoch": 2.0,
34
- "grad_norm": 5.25152259456263e-08,
35
- "kl": 0.0,
36
- "learning_rate": 2.5e-07,
37
- "loss": 0.0,
38
- "reward": 7.75,
39
- "reward_std": 0.0,
40
- "rewards/concensus_correctness_reward_func": 2.5,
41
- "rewards/consensus_reward_func": 2.0,
42
- "rewards/cumulative_reward_2": 0.0,
43
- "rewards/final_correctness_reward_func": 0.5,
44
- "rewards/question_recreation_reward_func": 1.0,
45
- "rewards/soft_format_reward_func": 0.0,
46
- "rewards/strict_format_reward_func": 0.5,
47
- "rewards/xmlcount_reward_func": 1.25,
48
- "step": 4
49
- },
50
- {
51
- "completion_length": 160.0,
52
- "epoch": 2.6666666666666665,
53
- "kl": 0.0,
54
- "reward": 8.25,
55
- "reward_std": 0.0,
56
- "rewards/concensus_correctness_reward_func": 2.75,
57
- "rewards/consensus_reward_func": 2.0,
58
- "rewards/cumulative_reward_2": 0.0,
59
- "rewards/final_correctness_reward_func": 0.75,
60
- "rewards/question_recreation_reward_func": 1.0,
61
- "rewards/soft_format_reward_func": 0.0,
62
- "rewards/strict_format_reward_func": 0.5,
63
- "rewards/xmlcount_reward_func": 1.25,
64
- "step": 5,
 
65
  "total_flos": 0.0,
66
- "train_loss": 0.0,
67
- "train_runtime": 863.957,
68
- "train_samples_per_second": 0.093,
69
- "train_steps_per_second": 0.006
70
  }
71
  ],
72
  "logging_steps": 2,
73
- "max_steps": 5,
74
- "num_input_tokens_seen": 0,
75
- "num_train_epochs": 5,
76
  "save_steps": 25,
77
  "stateful_callbacks": {
78
  "TrainerControl": {
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.5217391304347826,
6
  "eval_steps": 500,
7
+ "global_step": 3,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "clip_ratio/high_max": 0.0,
14
+ "clip_ratio/high_mean": 0.0,
15
+ "clip_ratio/low_mean": 0.0,
16
+ "clip_ratio/low_min": 0.0,
17
+ "clip_ratio/region_mean": 0.0,
18
+ "completions/clipped_ratio": 0.125,
19
+ "completions/max_length": 759.0,
20
+ "completions/max_terminated_length": 544.0,
21
+ "completions/mean_length": 344.3125,
22
+ "completions/mean_terminated_length": 246.29166412353516,
23
+ "completions/min_length": 64.5,
24
+ "completions/min_terminated_length": 64.5,
25
+ "epoch": 0.34782608695652173,
26
+ "grad_norm": 6.1356120109558105,
27
  "kl": 0.0,
28
  "learning_rate": 5e-07,
29
+ "loss": 0.1231,
30
+ "num_tokens": 9605.0,
31
+ "reward": 0.0872948095202446,
32
+ "reward_std": 0.036231483216397464,
33
+ "rewards/concensus_correctness_reward_func/mean": 0.0,
34
+ "rewards/concensus_correctness_reward_func/std": 0.0,
35
+ "rewards/consensus_reward_func/mean": 0.0,
36
+ "rewards/consensus_reward_func/std": 0.0,
37
+ "rewards/cumulative_reward_2/mean": 0.0,
38
+ "rewards/cumulative_reward_2/std": 0.0,
39
+ "rewards/final_correctness_reward_func/mean": 0.0,
40
+ "rewards/final_correctness_reward_func/std": 0.0,
41
+ "rewards/question_recreation_reward_func/mean": 0.0872948169708252,
42
+ "rewards/question_recreation_reward_func/std": 0.042022028006613255,
43
+ "rewards/soft_format_reward_func/mean": 0.0,
44
+ "rewards/soft_format_reward_func/std": 0.0,
45
+ "rewards/strict_format_reward_func/mean": 0.0,
46
+ "rewards/strict_format_reward_func/std": 0.0,
47
+ "rewards/xmlcount_reward_func/mean": 0.0,
48
+ "rewards/xmlcount_reward_func/std": 0.0,
49
  "step": 2
50
  },
51
  {
52
+ "clip_ratio/high_max": 0.0,
53
+ "clip_ratio/high_mean": 0.0,
54
+ "clip_ratio/low_mean": 0.0,
55
+ "clip_ratio/low_min": 0.0,
56
+ "clip_ratio/region_mean": 0.0,
57
+ "completions/clipped_ratio": 0.0,
58
+ "completions/max_length": 369.0,
59
+ "completions/max_terminated_length": 369.0,
60
+ "completions/mean_length": 126.25,
61
+ "completions/mean_terminated_length": 126.25,
62
+ "completions/min_length": 2.0,
63
+ "completions/min_terminated_length": 2.0,
64
+ "epoch": 0.5217391304347826,
65
+ "kl": 0.0009940905874827877,
66
+ "num_tokens": 12663.0,
67
+ "reward": 2.516864776611328,
68
+ "reward_std": 3.5382814407348633,
69
+ "rewards/concensus_correctness_reward_func/mean": 2.5,
70
+ "rewards/concensus_correctness_reward_func/std": 7.071068286895752,
71
+ "rewards/consensus_reward_func/mean": 0.0,
72
+ "rewards/consensus_reward_func/std": 0.0,
73
+ "rewards/cumulative_reward_2/mean": 0.0,
74
+ "rewards/cumulative_reward_2/std": 0.0,
75
+ "rewards/final_correctness_reward_func/mean": 0.0,
76
+ "rewards/final_correctness_reward_func/std": 0.0,
77
+ "rewards/question_recreation_reward_func/mean": 0.01686476171016693,
78
+ "rewards/question_recreation_reward_func/std": 0.01196382101625204,
79
+ "rewards/soft_format_reward_func/mean": 0.0,
80
+ "rewards/soft_format_reward_func/std": 0.0,
81
+ "rewards/strict_format_reward_func/mean": 0.0,
82
+ "rewards/strict_format_reward_func/std": 0.0,
83
+ "rewards/xmlcount_reward_func/mean": 0.0,
84
+ "rewards/xmlcount_reward_func/std": 0.0,
85
+ "step": 3,
86
  "total_flos": 0.0,
87
+ "train_loss": 0.27013154824574787,
88
+ "train_runtime": 1786.8594,
89
+ "train_samples_per_second": 0.013,
90
+ "train_steps_per_second": 0.002
91
  }
92
  ],
93
  "logging_steps": 2,
94
+ "max_steps": 3,
95
+ "num_input_tokens_seen": 12663,
96
+ "num_train_epochs": 1,
97
  "save_steps": 25,
98
  "stateful_callbacks": {
99
  "TrainerControl": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85368bab1dd51be2cbdf5e0109659802ffed3c70a6bb03cb0591e9dcae7e152d
3
- size 5944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a71230f32e732eea5fcb1a0593b7bc74ff3864c9b7b760cda833ba20848a0ed
3
+ size 6929