minpeter commited on
Commit
319fdad
·
verified ·
1 Parent(s): b8585f4

End of training

Browse files
Files changed (1) hide show
  1. README.md +103 -8
README.md CHANGED
@@ -5,7 +5,15 @@ tags:
5
  - axolotl
6
  - generated_from_trainer
7
  datasets:
 
8
  - lemon-mint/smol-koreantalk
 
 
 
 
 
 
 
9
  model-index:
10
  - name: tiny-ko-124m-sft
11
  results: []
@@ -33,6 +41,14 @@ strict: false
33
 
34
  chat_template: chatml
35
  datasets:
 
 
 
 
 
 
 
 
36
  - path: lemon-mint/smol-koreantalk
37
  type: chat_template
38
  split: train
@@ -41,6 +57,64 @@ datasets:
41
  role: role
42
  content: content
43
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  dataset_prepared_path: last_run_prepared
45
  val_set_size: 0.001
46
  save_safetensors: true
@@ -94,9 +168,9 @@ fsdp_config:
94
 
95
  # tiny-ko-124m-sft
96
 
97
- This model is a fine-tuned version of [minpeter/tiny-ko-124m-base](https://huggingface.co/minpeter/tiny-ko-124m-base) on the lemon-mint/smol-koreantalk dataset.
98
  It achieves the following results on the evaluation set:
99
- - Loss: 1.8151
100
 
101
  ## Model description
102
 
@@ -127,17 +201,38 @@ The following hyperparameters were used during training:
127
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
128
  - lr_scheduler_type: cosine
129
  - lr_scheduler_warmup_steps: 20
130
- - training_steps: 887
131
 
132
  ### Training results
133
 
134
  | Training Loss | Epoch | Step | Validation Loss |
135
  |:-------------:|:------:|:----:|:---------------:|
136
- | No log | 0 | 0 | 2.8035 |
137
- | 2.0195 | 0.2256 | 200 | 1.9871 |
138
- | 1.8857 | 0.4513 | 400 | 1.8815 |
139
- | 1.8013 | 0.6769 | 600 | 1.8270 |
140
- | 1.8489 | 0.9026 | 800 | 1.8151 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
141
 
142
 
143
  ### Framework versions
 
5
  - axolotl
6
  - generated_from_trainer
7
  datasets:
8
+ - lemon-mint/Korean-FineTome-100k
9
  - lemon-mint/smol-koreantalk
10
+ - heegyu/open-korean-instructions-v20231020
11
+ - trillionlabs/multisystem-curated
12
+ - allenai/tulu-3-sft-personas-instruction-following
13
+ - coastral/korean-writing-style-instruct
14
+ - devngho/korean-instruction-mix
15
+ - youjunhyeok/Magpie-Pro-300K-Filtered-ko
16
+ - youjunhyeok/smoltalk-ko-translate
17
  model-index:
18
  - name: tiny-ko-124m-sft
19
  results: []
 
41
 
42
  chat_template: chatml
43
  datasets:
44
+ - path: lemon-mint/Korean-FineTome-100k
45
+ type: chat_template
46
+ split: train
47
+ field_messages: messages
48
+ message_property_mappings:
49
+ role: role
50
+ content: content
51
+
52
  - path: lemon-mint/smol-koreantalk
53
  type: chat_template
54
  split: train
 
57
  role: role
58
  content: content
59
 
60
+ - path: heegyu/open-korean-instructions-v20231020
61
+ type: chat_template
62
+ split: train
63
+ field_messages: conversations
64
+ message_property_mappings:
65
+ role: from
66
+ content: value
67
+ roles:
68
+ user: ["human", "user"]
69
+ assistant: ["gpt", "assistant", "bot"]
70
+ system: ["system", "input"]
71
+ - path: trillionlabs/multisystem-curated
72
+ type: chat_template
73
+ split: train
74
+ field_messages: messages
75
+ message_property_mappings:
76
+ role: role
77
+ content: content
78
+ - path: allenai/tulu-3-sft-personas-instruction-following
79
+ type: chat_template
80
+ split: train
81
+ field_messages: messages
82
+ message_property_mappings:
83
+ role: role
84
+ content: content
85
+ - path: coastral/korean-writing-style-instruct
86
+ type: chat_template
87
+ split: train
88
+ field_messages: conversations
89
+ message_property_mappings:
90
+ role: from
91
+ content: value
92
+
93
+ - path: devngho/korean-instruction-mix
94
+ type: chat_template
95
+ split: train
96
+ field_messages: messages
97
+ message_property_mappings:
98
+ role: from
99
+ content: value
100
+
101
+ - path: youjunhyeok/Magpie-Pro-300K-Filtered-ko
102
+ type: chat_template
103
+ split: train
104
+ field_messages: conversations
105
+ message_property_mappings:
106
+ role: from
107
+ content: value
108
+
109
+ - path: youjunhyeok/smoltalk-ko-translate
110
+ type: chat_template
111
+ split: train
112
+ name: merge_filtered
113
+ field_messages: conversations
114
+ message_property_mappings:
115
+ role: role
116
+ content: content
117
+
118
  dataset_prepared_path: last_run_prepared
119
  val_set_size: 0.001
120
  save_safetensors: true
 
168
 
169
  # tiny-ko-124m-sft
170
 
171
+ This model is a fine-tuned version of [minpeter/tiny-ko-124m-base](https://huggingface.co/minpeter/tiny-ko-124m-base) on the lemon-mint/Korean-FineTome-100k, the lemon-mint/smol-koreantalk, the heegyu/open-korean-instructions-v20231020, the trillionlabs/multisystem-curated, the allenai/tulu-3-sft-personas-instruction-following, the coastral/korean-writing-style-instruct, the devngho/korean-instruction-mix, the youjunhyeok/Magpie-Pro-300K-Filtered-ko and the youjunhyeok/smoltalk-ko-translate datasets.
172
  It achieves the following results on the evaluation set:
173
+ - Loss: 1.7098
174
 
175
  ## Model description
176
 
 
201
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
202
  - lr_scheduler_type: cosine
203
  - lr_scheduler_warmup_steps: 20
204
+ - training_steps: 5042
205
 
206
  ### Training results
207
 
208
  | Training Loss | Epoch | Step | Validation Loss |
209
  |:-------------:|:------:|:----:|:---------------:|
210
+ | No log | 0 | 0 | 2.7016 |
211
+ | 2.1419 | 0.0397 | 200 | 2.1320 |
212
+ | 2.0675 | 0.0793 | 400 | 2.0446 |
213
+ | 2.0252 | 0.1190 | 600 | 1.9864 |
214
+ | 1.9304 | 0.1587 | 800 | 1.9468 |
215
+ | 1.9536 | 0.1983 | 1000 | 1.9145 |
216
+ | 1.8692 | 0.2380 | 1200 | 1.8879 |
217
+ | 1.8556 | 0.2777 | 1400 | 1.8645 |
218
+ | 1.8421 | 0.3174 | 1600 | 1.8433 |
219
+ | 1.9118 | 0.3570 | 1800 | 1.8256 |
220
+ | 1.7791 | 0.3967 | 2000 | 1.8090 |
221
+ | 1.8162 | 0.4364 | 2200 | 1.7934 |
222
+ | 1.796 | 0.4760 | 2400 | 1.7795 |
223
+ | 1.749 | 0.5157 | 2600 | 1.7661 |
224
+ | 1.7536 | 0.5554 | 2800 | 1.7540 |
225
+ | 1.7672 | 0.5950 | 3000 | 1.7432 |
226
+ | 1.7523 | 0.6347 | 3200 | 1.7336 |
227
+ | 1.7074 | 0.6744 | 3400 | 1.7259 |
228
+ | 1.7218 | 0.7141 | 3600 | 1.7202 |
229
+ | 1.6928 | 0.7537 | 3800 | 1.7158 |
230
+ | 1.7184 | 0.7934 | 4000 | 1.7127 |
231
+ | 1.761 | 0.8331 | 4200 | 1.7109 |
232
+ | 1.7481 | 0.8727 | 4400 | 1.7101 |
233
+ | 1.7245 | 0.9124 | 4600 | 1.7098 |
234
+ | 1.7076 | 0.9521 | 4800 | 1.7097 |
235
+ | 1.7403 | 0.9917 | 5000 | 1.7098 |
236
 
237
 
238
  ### Framework versions