Iedha commited on
Commit
5665c22
·
verified ·
1 Parent(s): 36feafc

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Gensyn/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lethal_tawny_deer
5
  tags:
@@ -14,7 +14,7 @@ licence: license
14
 
15
  # Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lethal_tawny_deer
16
 
17
- This model is a fine-tuned version of [Gensyn/Qwen2.5-0.5B-Instruct](https://huggingface.co/Gensyn/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
 
1
  ---
2
+ base_model: unsloth/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lethal_tawny_deer
5
  tags:
 
14
 
15
  # Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-lethal_tawny_deer
16
 
17
+ This model is a fine-tuned version of [unsloth/Qwen2.5-0.5B-Instruct](https://huggingface.co/unsloth/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 4.9965383652761375e-06,
4
- "train_runtime": 5282.5247,
5
- "train_samples": 17,
6
- "train_samples_per_second": 0.03,
7
- "train_steps_per_second": 0.002
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 7.331371307373047e-07,
4
+ "train_runtime": 1043.6808,
5
+ "train_samples": 3,
6
+ "train_samples_per_second": 0.019,
7
+ "train_steps_per_second": 0.005
8
  }
config.json CHANGED
@@ -3,7 +3,6 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 896,
@@ -15,13 +14,15 @@
15
  "num_attention_heads": 14,
16
  "num_hidden_layers": 24,
17
  "num_key_value_heads": 2,
 
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
- "sliding_window": 32768,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.2",
 
25
  "use_cache": true,
26
  "use_sliding_window": false,
27
  "vocab_size": 151936
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
 
6
  "eos_token_id": 151645,
7
  "hidden_act": "silu",
8
  "hidden_size": 896,
 
14
  "num_attention_heads": 14,
15
  "num_hidden_layers": 24,
16
  "num_key_value_heads": 2,
17
+ "pad_token_id": 151654,
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.2",
25
+ "unsloth_fixed": true,
26
  "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
generation_config.json CHANGED
@@ -5,7 +5,8 @@
5
  151645,
6
  151643
7
  ],
8
- "pad_token_id": 151643,
 
9
  "repetition_penalty": 1.1,
10
  "temperature": 0.7,
11
  "top_k": 20,
 
5
  151645,
6
  151643
7
  ],
8
+ "max_length": 32768,
9
+ "pad_token_id": 151654,
10
  "repetition_penalty": 1.1,
11
  "temperature": 0.7,
12
  "top_k": 20,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fe9160b49ab9da979f8e5a9e1b761aa0608f531a05651ad604686e8252caf69
3
  size 1976163472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e573dc601c408ba8b5ae9d6c63e6cb6686e6332c73231c4810ba5c2bf425bfac
3
  size 1976163472
special_tokens_map.json CHANGED
@@ -22,7 +22,7 @@
22
  "single_word": false
23
  },
24
  "pad_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
 
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|vision_pad|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eee858c5123a4279c3e1f7b81247343f356ac767940b2692a928ad929543214
3
- size 11422063
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e71213db910f5cafa86d35091f37393dcc344b1bbc34091d1b3eed4cca01d5
3
+ size 11422064
tokenizer_config.json CHANGED
@@ -200,8 +200,9 @@
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
- "model_max_length": 131072,
204
- "pad_token": "<|endoftext|>",
 
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
 
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
+ "model_max_length": 32768,
204
+ "pad_token": "<|vision_pad|>",
205
+ "padding_side": "left",
206
  "split_special_tokens": false,
207
  "tokenizer_class": "Qwen2Tokenizer",
208
  "unk_token": null
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 4.9965383652761375e-06,
4
- "train_runtime": 5282.5247,
5
- "train_samples": 17,
6
- "train_samples_per_second": 0.03,
7
- "train_steps_per_second": 0.002
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 7.331371307373047e-07,
4
+ "train_runtime": 1043.6808,
5
+ "train_samples": 3,
6
+ "train_samples_per_second": 0.019,
7
+ "train_steps_per_second": 0.005
8
  }
trainer_state.json CHANGED
@@ -2,123 +2,78 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 3.4705882352941178,
6
  "eval_steps": 500,
7
- "global_step": 10,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "completion_length": 105.875,
14
- "epoch": 0.9411764705882353,
15
- "grad_norm": 0.0,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
- "loss": 0.0,
19
- "reward": 7.015919391065836,
20
- "reward_std": 0.010885396040976048,
21
- "rewards/concensus_correctness_reward_func": 2.176249995827675,
22
- "rewards/consensus_reward_func": 1.875,
23
  "rewards/cumulative_reward_2": 0.0,
24
- "rewards/final_correctness_reward_func": 0.375,
25
- "rewards/question_recreation_reward_func": 0.9412319990806282,
26
  "rewards/soft_format_reward_func": 0.0,
27
- "rewards/strict_format_reward_func": 0.46875,
28
- "rewards/xmlcount_reward_func": 1.1796875,
29
  "step": 2
30
  },
31
  {
32
- "completion_length": 104.61111111111111,
33
- "epoch": 1.4705882352941178,
34
- "grad_norm": 5.037790775299072,
35
- "kl": 0.0018976248730445554,
36
- "learning_rate": 4.415111107797445e-07,
37
- "loss": 0.0,
38
- "reward": 6.546576562854979,
39
- "reward_std": 0.0986187085768001,
40
- "rewards/concensus_correctness_reward_func": 2.162222226460775,
41
- "rewards/consensus_reward_func": 1.5555555555555556,
42
- "rewards/cumulative_reward_2": 0.0,
43
- "rewards/final_correctness_reward_func": 0.6666666666666666,
44
- "rewards/question_recreation_reward_func": 0.7940766894155078,
45
- "rewards/soft_format_reward_func": 0.0,
46
- "rewards/strict_format_reward_func": 0.3611111111111111,
47
- "rewards/xmlcount_reward_func": 1.0069444444444444,
48
- "step": 4
49
- },
50
- {
51
- "completion_length": 121.55555555555556,
52
  "epoch": 2.0,
53
- "grad_norm": 1.8580118417739868,
54
- "kl": 0.0006738082559220655,
55
- "learning_rate": 2.934120444167326e-07,
56
  "loss": 0.0,
57
- "reward": 5.957985914415783,
58
- "reward_std": 0.022622719407081604,
59
- "rewards/concensus_correctness_reward_func": 1.7066666550106473,
60
- "rewards/consensus_reward_func": 1.7777777777777777,
61
  "rewards/cumulative_reward_2": 0.0,
62
  "rewards/final_correctness_reward_func": 0.0,
63
- "rewards/question_recreation_reward_func": 0.9040971166557736,
64
  "rewards/soft_format_reward_func": 0.0,
65
- "rewards/strict_format_reward_func": 0.4444444444444444,
66
- "rewards/xmlcount_reward_func": 1.125,
67
- "step": 6
68
- },
69
- {
70
- "completion_length": 95.625,
71
- "epoch": 2.9411764705882355,
72
- "grad_norm": 3.880513906478882,
73
- "kl": 0.013746480780241654,
74
- "learning_rate": 1.2500000000000005e-07,
75
- "loss": 0.0,
76
- "reward": 6.583770731464028,
77
- "reward_std": 0.0842309154686518,
78
- "rewards/concensus_correctness_reward_func": 2.056499995291233,
79
- "rewards/consensus_reward_func": 1.75,
80
- "rewards/cumulative_reward_2": 0.0,
81
- "rewards/final_correctness_reward_func": 0.375,
82
- "rewards/question_recreation_reward_func": 0.8952083382755518,
83
- "rewards/soft_format_reward_func": 0.0,
84
- "rewards/strict_format_reward_func": 0.40625,
85
- "rewards/xmlcount_reward_func": 1.1008124947547913,
86
- "step": 8
87
  },
88
  {
89
- "completion_length": 106.05555555555556,
90
- "epoch": 3.4705882352941178,
91
- "grad_norm": 3.5325653553009033,
92
- "kl": 0.015616837632579619,
93
- "learning_rate": 1.507684480352292e-08,
94
- "loss": 0.0,
95
- "reward": 5.260379309455554,
96
- "reward_std": 0.027002629927463002,
97
- "rewards/concensus_correctness_reward_func": 1.4937777784135606,
98
- "rewards/consensus_reward_func": 1.5555555555555556,
99
  "rewards/cumulative_reward_2": 0.0,
100
  "rewards/final_correctness_reward_func": 0.0,
101
- "rewards/question_recreation_reward_func": 0.808268278837204,
102
  "rewards/soft_format_reward_func": 0.0,
103
- "rewards/strict_format_reward_func": 0.3888888888888889,
104
- "rewards/xmlcount_reward_func": 1.0138888888888888,
105
- "step": 10
106
- },
107
- {
108
- "epoch": 3.4705882352941178,
109
- "step": 10,
110
  "total_flos": 0.0,
111
- "train_loss": 4.9965383652761375e-06,
112
- "train_runtime": 5282.5247,
113
- "train_samples_per_second": 0.03,
114
- "train_steps_per_second": 0.002
115
  }
116
  ],
117
  "logging_steps": 2,
118
- "max_steps": 10,
119
  "num_input_tokens_seen": 0,
120
  "num_train_epochs": 5,
121
- "save_steps": 25,
122
  "stateful_callbacks": {
123
  "TrainerControl": {
124
  "args": {
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.6666666666666665,
6
  "eval_steps": 500,
7
+ "global_step": 5,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "completion_length": 127.16666666666667,
14
+ "epoch": 1.0,
15
+ "grad_norm": 10.25474739074707,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
+ "loss": -0.0,
19
+ "reward": 0.16262217486898103,
20
+ "reward_std": 0.08637613849714398,
21
+ "rewards/concensus_correctness_reward_func": 0.0,
22
+ "rewards/consensus_reward_func": 0.0,
23
  "rewards/cumulative_reward_2": 0.0,
24
+ "rewards/final_correctness_reward_func": 0.0,
25
+ "rewards/question_recreation_reward_func": 0.16262217486898103,
26
  "rewards/soft_format_reward_func": 0.0,
27
+ "rewards/strict_format_reward_func": 0.0,
28
+ "rewards/xmlcount_reward_func": 0.0,
29
  "step": 2
30
  },
31
  {
32
+ "completion_length": 187.5,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  "epoch": 2.0,
34
+ "grad_norm": 8.673498153686523,
35
+ "kl": 0.0010691387384819488,
36
+ "learning_rate": 2.5e-07,
37
  "loss": 0.0,
38
+ "reward": 0.10594786455233891,
39
+ "reward_std": 0.033774735095600285,
40
+ "rewards/concensus_correctness_reward_func": 0.0,
41
+ "rewards/consensus_reward_func": 0.0,
42
  "rewards/cumulative_reward_2": 0.0,
43
  "rewards/final_correctness_reward_func": 0.0,
44
+ "rewards/question_recreation_reward_func": 0.10594786455233891,
45
  "rewards/soft_format_reward_func": 0.0,
46
+ "rewards/strict_format_reward_func": 0.0,
47
+ "rewards/xmlcount_reward_func": 0.0,
48
+ "step": 4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
  },
50
  {
51
+ "completion_length": 126.25,
52
+ "epoch": 2.6666666666666665,
53
+ "kl": 0.0008855987689457834,
54
+ "reward": 0.1095019057393074,
55
+ "reward_std": 0.005491654243087396,
56
+ "rewards/concensus_correctness_reward_func": 0.0,
57
+ "rewards/consensus_reward_func": 0.0,
 
 
 
58
  "rewards/cumulative_reward_2": 0.0,
59
  "rewards/final_correctness_reward_func": 0.0,
60
+ "rewards/question_recreation_reward_func": 0.1095019057393074,
61
  "rewards/soft_format_reward_func": 0.0,
62
+ "rewards/strict_format_reward_func": 0.0,
63
+ "rewards/xmlcount_reward_func": 0.0,
64
+ "step": 5,
 
 
 
 
65
  "total_flos": 0.0,
66
+ "train_loss": 7.331371307373047e-07,
67
+ "train_runtime": 1043.6808,
68
+ "train_samples_per_second": 0.019,
69
+ "train_steps_per_second": 0.005
70
  }
71
  ],
72
  "logging_steps": 2,
73
+ "max_steps": 5,
74
  "num_input_tokens_seen": 0,
75
  "num_train_epochs": 5,
76
+ "save_steps": 10,
77
  "stateful_callbacks": {
78
  "TrainerControl": {
79
  "args": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69815f1dc525e3403e905adf7230c3cee33a2c72d178fea037f37dbff247a476
3
  size 5944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6e8bda82286f4ad9312e40a24819e6419d65b09f5d5840cb46a13ab0b6e2aa3
3
  size 5944