diff --git a/.gitattributes b/.gitattributes
index a6344aac8c09253b3b630fb776ae94478aa0275b..e02373cbcfee06582e361869ad41fe06719c0cdd 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -33,3 +33,11 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text
+checkpoint-10536/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+checkpoint-12292/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+checkpoint-1756/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+checkpoint-3512/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+checkpoint-5268/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+checkpoint-7024/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+checkpoint-8780/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/checkpoint-10536/config.json b/checkpoint-10536/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-10536/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-10536/model.safetensors b/checkpoint-10536/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f93bd96e79bc1d1ef4dc55d378ec6622bba98879
--- /dev/null
+++ b/checkpoint-10536/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6440a058003a17dd1e8e9f37290457a8d15df03bda5f80db000d0ec55c3b65a3
+size 1109863956
diff --git a/checkpoint-10536/optimizer.pt b/checkpoint-10536/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..963aef4c83eb8f5de1b53ff2cacef00677ec8ed6
--- /dev/null
+++ b/checkpoint-10536/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e5aebcdf200787a3d36ad26e97baf2f95a1edbcaa27dfb7de4afd552f457a17d
+size 2219846202
diff --git a/checkpoint-10536/rng_state.pth b/checkpoint-10536/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..872c885addb0a6f48e4aac33446fb62fb68a98ce
--- /dev/null
+++ b/checkpoint-10536/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:083825a219aea1d69cbec6e4b693fbcf2e11dad1312750efc78367126c76fbe2
+size 14244
diff --git a/checkpoint-10536/scaler.pt b/checkpoint-10536/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..be2d579c9eb1e892fbc0a2f6f7d7c95ae0e90e91
--- /dev/null
+++ b/checkpoint-10536/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:10d19f6e3641d81395be9b8e86101ef67c4f91a8f20888a2ad5331c197d7e6e6
+size 988
diff --git a/checkpoint-10536/scheduler.pt b/checkpoint-10536/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8fee7f3d2cc2ea24e201571ceb19ccf1a2072e53
--- /dev/null
+++ b/checkpoint-10536/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f4cd02bf0a083f5bb6fbdae03e00fa68b9cfc022b252c7ffd71dcb9e74d3421a
+size 1064
diff --git a/checkpoint-10536/special_tokens_map.json b/checkpoint-10536/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-10536/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-10536/tokenizer.json b/checkpoint-10536/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-10536/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-10536/tokenizer_config.json b/checkpoint-10536/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-10536/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-10536/trainer_state.json b/checkpoint-10536/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..cd9919e6e7196ce2d3a6226b5522f8d04a5f19a8
--- /dev/null
+++ b/checkpoint-10536/trainer_state.json
@@ -0,0 +1,1570 @@
+{
+ "best_global_step": 10536,
+ "best_metric": 0.9557566616390145,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-10536",
+ "epoch": 6.0,
+ "eval_steps": 500,
+ "global_step": 10536,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ },
+ {
+ "epoch": 1.0250569476082005,
+ "grad_norm": 4.280463695526123,
+ "learning_rate": 1.708102831109665e-05,
+ "loss": 0.0673,
+ "step": 1800
+ },
+ {
+ "epoch": 1.0535307517084282,
+ "grad_norm": 0.06692535430192947,
+ "learning_rate": 1.6999674585096e-05,
+ "loss": 0.0403,
+ "step": 1850
+ },
+ {
+ "epoch": 1.082004555808656,
+ "grad_norm": 0.03789375349879265,
+ "learning_rate": 1.691832085909535e-05,
+ "loss": 0.0496,
+ "step": 1900
+ },
+ {
+ "epoch": 1.1104783599088839,
+ "grad_norm": 3.698348045349121,
+ "learning_rate": 1.6836967133094698e-05,
+ "loss": 0.0381,
+ "step": 1950
+ },
+ {
+ "epoch": 1.1389521640091116,
+ "grad_norm": 2.753032684326172,
+ "learning_rate": 1.6755613407094048e-05,
+ "loss": 0.0432,
+ "step": 2000
+ },
+ {
+ "epoch": 1.1674259681093395,
+ "grad_norm": 0.06336130946874619,
+ "learning_rate": 1.6674259681093398e-05,
+ "loss": 0.0482,
+ "step": 2050
+ },
+ {
+ "epoch": 1.1958997722095672,
+ "grad_norm": 3.9503254890441895,
+ "learning_rate": 1.6592905955092744e-05,
+ "loss": 0.0436,
+ "step": 2100
+ },
+ {
+ "epoch": 1.224373576309795,
+ "grad_norm": 0.08679741621017456,
+ "learning_rate": 1.6511552229092093e-05,
+ "loss": 0.0346,
+ "step": 2150
+ },
+ {
+ "epoch": 1.2528473804100229,
+ "grad_norm": 2.1995127201080322,
+ "learning_rate": 1.6430198503091443e-05,
+ "loss": 0.0312,
+ "step": 2200
+ },
+ {
+ "epoch": 1.2813211845102506,
+ "grad_norm": 5.7709808349609375,
+ "learning_rate": 1.634884477709079e-05,
+ "loss": 0.032,
+ "step": 2250
+ },
+ {
+ "epoch": 1.3097949886104785,
+ "grad_norm": 0.2031829059123993,
+ "learning_rate": 1.626749105109014e-05,
+ "loss": 0.053,
+ "step": 2300
+ },
+ {
+ "epoch": 1.3382687927107062,
+ "grad_norm": 4.292157173156738,
+ "learning_rate": 1.6186137325089492e-05,
+ "loss": 0.0413,
+ "step": 2350
+ },
+ {
+ "epoch": 1.366742596810934,
+ "grad_norm": 0.07821401953697205,
+ "learning_rate": 1.6104783599088842e-05,
+ "loss": 0.0304,
+ "step": 2400
+ },
+ {
+ "epoch": 1.3952164009111616,
+ "grad_norm": 0.5930206179618835,
+ "learning_rate": 1.6023429873088188e-05,
+ "loss": 0.0525,
+ "step": 2450
+ },
+ {
+ "epoch": 1.4236902050113895,
+ "grad_norm": 1.6114712953567505,
+ "learning_rate": 1.5942076147087538e-05,
+ "loss": 0.0391,
+ "step": 2500
+ },
+ {
+ "epoch": 1.4521640091116172,
+ "grad_norm": 0.28632065653800964,
+ "learning_rate": 1.5860722421086887e-05,
+ "loss": 0.0491,
+ "step": 2550
+ },
+ {
+ "epoch": 1.4806378132118452,
+ "grad_norm": 1.0017669200897217,
+ "learning_rate": 1.5779368695086237e-05,
+ "loss": 0.0395,
+ "step": 2600
+ },
+ {
+ "epoch": 1.5091116173120729,
+ "grad_norm": 0.03409272059798241,
+ "learning_rate": 1.5698014969085583e-05,
+ "loss": 0.0399,
+ "step": 2650
+ },
+ {
+ "epoch": 1.5375854214123006,
+ "grad_norm": 3.5108702182769775,
+ "learning_rate": 1.5616661243084933e-05,
+ "loss": 0.0424,
+ "step": 2700
+ },
+ {
+ "epoch": 1.5660592255125285,
+ "grad_norm": 3.5875349044799805,
+ "learning_rate": 1.5535307517084283e-05,
+ "loss": 0.0497,
+ "step": 2750
+ },
+ {
+ "epoch": 1.5945330296127562,
+ "grad_norm": 0.10498251020908356,
+ "learning_rate": 1.5453953791083632e-05,
+ "loss": 0.0461,
+ "step": 2800
+ },
+ {
+ "epoch": 1.6230068337129842,
+ "grad_norm": 7.618963718414307,
+ "learning_rate": 1.5374227139602996e-05,
+ "loss": 0.0394,
+ "step": 2850
+ },
+ {
+ "epoch": 1.6514806378132119,
+ "grad_norm": 3.3920793533325195,
+ "learning_rate": 1.5292873413602346e-05,
+ "loss": 0.0339,
+ "step": 2900
+ },
+ {
+ "epoch": 1.6799544419134396,
+ "grad_norm": 0.04736129939556122,
+ "learning_rate": 1.5211519687601692e-05,
+ "loss": 0.0358,
+ "step": 2950
+ },
+ {
+ "epoch": 1.7084282460136673,
+ "grad_norm": 0.17463508248329163,
+ "learning_rate": 1.5130165961601042e-05,
+ "loss": 0.034,
+ "step": 3000
+ },
+ {
+ "epoch": 1.7369020501138952,
+ "grad_norm": 0.058148205280303955,
+ "learning_rate": 1.5048812235600392e-05,
+ "loss": 0.0317,
+ "step": 3050
+ },
+ {
+ "epoch": 1.7653758542141231,
+ "grad_norm": 3.5262815952301025,
+ "learning_rate": 1.4967458509599741e-05,
+ "loss": 0.0477,
+ "step": 3100
+ },
+ {
+ "epoch": 1.7938496583143508,
+ "grad_norm": 2.971937894821167,
+ "learning_rate": 1.488610478359909e-05,
+ "loss": 0.0364,
+ "step": 3150
+ },
+ {
+ "epoch": 1.8223234624145785,
+ "grad_norm": 0.05181724950671196,
+ "learning_rate": 1.4804751057598439e-05,
+ "loss": 0.0394,
+ "step": 3200
+ },
+ {
+ "epoch": 1.8507972665148062,
+ "grad_norm": 0.2712370455265045,
+ "learning_rate": 1.4723397331597789e-05,
+ "loss": 0.0456,
+ "step": 3250
+ },
+ {
+ "epoch": 1.8792710706150342,
+ "grad_norm": 1.608622431755066,
+ "learning_rate": 1.4642043605597138e-05,
+ "loss": 0.05,
+ "step": 3300
+ },
+ {
+ "epoch": 1.907744874715262,
+ "grad_norm": 5.2423529624938965,
+ "learning_rate": 1.4560689879596486e-05,
+ "loss": 0.0371,
+ "step": 3350
+ },
+ {
+ "epoch": 1.9362186788154898,
+ "grad_norm": 0.04786211624741554,
+ "learning_rate": 1.4479336153595836e-05,
+ "loss": 0.0329,
+ "step": 3400
+ },
+ {
+ "epoch": 1.9646924829157175,
+ "grad_norm": 0.5264931917190552,
+ "learning_rate": 1.4397982427595185e-05,
+ "loss": 0.0379,
+ "step": 3450
+ },
+ {
+ "epoch": 1.9931662870159452,
+ "grad_norm": 0.236006498336792,
+ "learning_rate": 1.4316628701594535e-05,
+ "loss": 0.0253,
+ "step": 3500
+ },
+ {
+ "epoch": 2.0,
+ "eval_f1": 0.9406438631790744,
+ "eval_loss": 0.05223705992102623,
+ "eval_precision": 0.9371867691279653,
+ "eval_recall": 0.9441265567149109,
+ "eval_runtime": 3.7109,
+ "eval_samples_per_second": 875.791,
+ "eval_steps_per_second": 109.676,
+ "step": 3512
+ },
+ {
+ "epoch": 2.021640091116173,
+ "grad_norm": 4.225963115692139,
+ "learning_rate": 1.4235274975593883e-05,
+ "loss": 0.0373,
+ "step": 3550
+ },
+ {
+ "epoch": 2.050113895216401,
+ "grad_norm": 0.15388111770153046,
+ "learning_rate": 1.4153921249593233e-05,
+ "loss": 0.0258,
+ "step": 3600
+ },
+ {
+ "epoch": 2.078587699316629,
+ "grad_norm": 6.3911895751953125,
+ "learning_rate": 1.4072567523592582e-05,
+ "loss": 0.0217,
+ "step": 3650
+ },
+ {
+ "epoch": 2.1070615034168565,
+ "grad_norm": 0.03075530007481575,
+ "learning_rate": 1.3991213797591932e-05,
+ "loss": 0.019,
+ "step": 3700
+ },
+ {
+ "epoch": 2.135535307517084,
+ "grad_norm": 0.16612432897090912,
+ "learning_rate": 1.3909860071591278e-05,
+ "loss": 0.0175,
+ "step": 3750
+ },
+ {
+ "epoch": 2.164009111617312,
+ "grad_norm": 0.14543022215366364,
+ "learning_rate": 1.382850634559063e-05,
+ "loss": 0.0289,
+ "step": 3800
+ },
+ {
+ "epoch": 2.19248291571754,
+ "grad_norm": 2.9880359172821045,
+ "learning_rate": 1.374715261958998e-05,
+ "loss": 0.0208,
+ "step": 3850
+ },
+ {
+ "epoch": 2.2209567198177678,
+ "grad_norm": 9.878133773803711,
+ "learning_rate": 1.3665798893589329e-05,
+ "loss": 0.021,
+ "step": 3900
+ },
+ {
+ "epoch": 2.2494305239179955,
+ "grad_norm": 9.684176445007324,
+ "learning_rate": 1.3584445167588675e-05,
+ "loss": 0.0267,
+ "step": 3950
+ },
+ {
+ "epoch": 2.277904328018223,
+ "grad_norm": 33.15859603881836,
+ "learning_rate": 1.3503091441588025e-05,
+ "loss": 0.0295,
+ "step": 4000
+ },
+ {
+ "epoch": 2.306378132118451,
+ "grad_norm": 2.559544086456299,
+ "learning_rate": 1.3421737715587374e-05,
+ "loss": 0.0292,
+ "step": 4050
+ },
+ {
+ "epoch": 2.334851936218679,
+ "grad_norm": 0.07091552764177322,
+ "learning_rate": 1.3340383989586724e-05,
+ "loss": 0.0262,
+ "step": 4100
+ },
+ {
+ "epoch": 2.3633257403189067,
+ "grad_norm": 1.8119585514068604,
+ "learning_rate": 1.3259030263586072e-05,
+ "loss": 0.03,
+ "step": 4150
+ },
+ {
+ "epoch": 2.3917995444191344,
+ "grad_norm": 0.1476636379957199,
+ "learning_rate": 1.3177676537585422e-05,
+ "loss": 0.0196,
+ "step": 4200
+ },
+ {
+ "epoch": 2.420273348519362,
+ "grad_norm": 0.04482650011777878,
+ "learning_rate": 1.3096322811584771e-05,
+ "loss": 0.0263,
+ "step": 4250
+ },
+ {
+ "epoch": 2.44874715261959,
+ "grad_norm": 0.09201560169458389,
+ "learning_rate": 1.3014969085584121e-05,
+ "loss": 0.0172,
+ "step": 4300
+ },
+ {
+ "epoch": 2.477220956719818,
+ "grad_norm": 0.02215876244008541,
+ "learning_rate": 1.2933615359583469e-05,
+ "loss": 0.0149,
+ "step": 4350
+ },
+ {
+ "epoch": 2.5056947608200457,
+ "grad_norm": 0.6993932127952576,
+ "learning_rate": 1.2852261633582819e-05,
+ "loss": 0.0331,
+ "step": 4400
+ },
+ {
+ "epoch": 2.5341685649202734,
+ "grad_norm": 0.030453965067863464,
+ "learning_rate": 1.2770907907582168e-05,
+ "loss": 0.0307,
+ "step": 4450
+ },
+ {
+ "epoch": 2.562642369020501,
+ "grad_norm": 0.22211593389511108,
+ "learning_rate": 1.2689554181581518e-05,
+ "loss": 0.0183,
+ "step": 4500
+ },
+ {
+ "epoch": 2.591116173120729,
+ "grad_norm": 1.2842738628387451,
+ "learning_rate": 1.2608200455580866e-05,
+ "loss": 0.0226,
+ "step": 4550
+ },
+ {
+ "epoch": 2.619589977220957,
+ "grad_norm": 3.9048423767089844,
+ "learning_rate": 1.2526846729580216e-05,
+ "loss": 0.0239,
+ "step": 4600
+ },
+ {
+ "epoch": 2.6480637813211843,
+ "grad_norm": 0.9887784719467163,
+ "learning_rate": 1.2445493003579565e-05,
+ "loss": 0.0271,
+ "step": 4650
+ },
+ {
+ "epoch": 2.6765375854214124,
+ "grad_norm": 4.759728908538818,
+ "learning_rate": 1.2364139277578915e-05,
+ "loss": 0.0201,
+ "step": 4700
+ },
+ {
+ "epoch": 2.70501138952164,
+ "grad_norm": 0.2671191096305847,
+ "learning_rate": 1.2282785551578263e-05,
+ "loss": 0.0212,
+ "step": 4750
+ },
+ {
+ "epoch": 2.733485193621868,
+ "grad_norm": 0.036626849323511124,
+ "learning_rate": 1.2201431825577612e-05,
+ "loss": 0.0155,
+ "step": 4800
+ },
+ {
+ "epoch": 2.7619589977220955,
+ "grad_norm": 6.877408027648926,
+ "learning_rate": 1.2120078099576962e-05,
+ "loss": 0.0266,
+ "step": 4850
+ },
+ {
+ "epoch": 2.7904328018223232,
+ "grad_norm": 4.725896835327148,
+ "learning_rate": 1.2038724373576312e-05,
+ "loss": 0.0163,
+ "step": 4900
+ },
+ {
+ "epoch": 2.8189066059225514,
+ "grad_norm": 0.11965059489011765,
+ "learning_rate": 1.195737064757566e-05,
+ "loss": 0.0246,
+ "step": 4950
+ },
+ {
+ "epoch": 2.847380410022779,
+ "grad_norm": 0.6634079813957214,
+ "learning_rate": 1.187601692157501e-05,
+ "loss": 0.0304,
+ "step": 5000
+ },
+ {
+ "epoch": 2.875854214123007,
+ "grad_norm": 3.952694892883301,
+ "learning_rate": 1.1794663195574359e-05,
+ "loss": 0.0183,
+ "step": 5050
+ },
+ {
+ "epoch": 2.9043280182232345,
+ "grad_norm": 19.788942337036133,
+ "learning_rate": 1.1713309469573709e-05,
+ "loss": 0.0193,
+ "step": 5100
+ },
+ {
+ "epoch": 2.932801822323462,
+ "grad_norm": 4.989261627197266,
+ "learning_rate": 1.1631955743573055e-05,
+ "loss": 0.023,
+ "step": 5150
+ },
+ {
+ "epoch": 2.9612756264236904,
+ "grad_norm": 9.020853996276855,
+ "learning_rate": 1.1550602017572406e-05,
+ "loss": 0.0231,
+ "step": 5200
+ },
+ {
+ "epoch": 2.989749430523918,
+ "grad_norm": 0.9151140451431274,
+ "learning_rate": 1.1469248291571756e-05,
+ "loss": 0.0314,
+ "step": 5250
+ },
+ {
+ "epoch": 3.0,
+ "eval_f1": 0.9470420646891236,
+ "eval_loss": 0.04870199039578438,
+ "eval_precision": 0.943090787716956,
+ "eval_recall": 0.9510265903736116,
+ "eval_runtime": 3.7494,
+ "eval_samples_per_second": 866.805,
+ "eval_steps_per_second": 108.551,
+ "step": 5268
+ },
+ {
+ "epoch": 3.0182232346241458,
+ "grad_norm": 8.62992000579834,
+ "learning_rate": 1.1387894565571106e-05,
+ "loss": 0.0194,
+ "step": 5300
+ },
+ {
+ "epoch": 3.0466970387243735,
+ "grad_norm": 0.5615554451942444,
+ "learning_rate": 1.1306540839570452e-05,
+ "loss": 0.0114,
+ "step": 5350
+ },
+ {
+ "epoch": 3.075170842824601,
+ "grad_norm": 2.5735936164855957,
+ "learning_rate": 1.1225187113569801e-05,
+ "loss": 0.0163,
+ "step": 5400
+ },
+ {
+ "epoch": 3.1036446469248293,
+ "grad_norm": 6.504094123840332,
+ "learning_rate": 1.1143833387569151e-05,
+ "loss": 0.0135,
+ "step": 5450
+ },
+ {
+ "epoch": 3.132118451025057,
+ "grad_norm": 14.477537155151367,
+ "learning_rate": 1.10624796615685e-05,
+ "loss": 0.0129,
+ "step": 5500
+ },
+ {
+ "epoch": 3.1605922551252847,
+ "grad_norm": 0.04929841682314873,
+ "learning_rate": 1.0981125935567849e-05,
+ "loss": 0.0263,
+ "step": 5550
+ },
+ {
+ "epoch": 3.1890660592255125,
+ "grad_norm": 0.08875144273042679,
+ "learning_rate": 1.0899772209567198e-05,
+ "loss": 0.0134,
+ "step": 5600
+ },
+ {
+ "epoch": 3.21753986332574,
+ "grad_norm": 0.6238455772399902,
+ "learning_rate": 1.0818418483566548e-05,
+ "loss": 0.0107,
+ "step": 5650
+ },
+ {
+ "epoch": 3.2460136674259683,
+ "grad_norm": 0.004948179703205824,
+ "learning_rate": 1.0737064757565898e-05,
+ "loss": 0.0167,
+ "step": 5700
+ },
+ {
+ "epoch": 3.274487471526196,
+ "grad_norm": 0.017031218856573105,
+ "learning_rate": 1.0655711031565246e-05,
+ "loss": 0.0085,
+ "step": 5750
+ },
+ {
+ "epoch": 3.3029612756264237,
+ "grad_norm": 2.9840469360351562,
+ "learning_rate": 1.0574357305564595e-05,
+ "loss": 0.0158,
+ "step": 5800
+ },
+ {
+ "epoch": 3.3314350797266514,
+ "grad_norm": 1.7477540969848633,
+ "learning_rate": 1.0493003579563945e-05,
+ "loss": 0.0157,
+ "step": 5850
+ },
+ {
+ "epoch": 3.359908883826879,
+ "grad_norm": 0.03963543102145195,
+ "learning_rate": 1.0411649853563295e-05,
+ "loss": 0.0201,
+ "step": 5900
+ },
+ {
+ "epoch": 3.3883826879271073,
+ "grad_norm": 0.16669808328151703,
+ "learning_rate": 1.0330296127562643e-05,
+ "loss": 0.0159,
+ "step": 5950
+ },
+ {
+ "epoch": 3.416856492027335,
+ "grad_norm": 0.01953568309545517,
+ "learning_rate": 1.0248942401561992e-05,
+ "loss": 0.0117,
+ "step": 6000
+ },
+ {
+ "epoch": 3.4453302961275627,
+ "grad_norm": 0.07708246260881424,
+ "learning_rate": 1.0167588675561342e-05,
+ "loss": 0.0117,
+ "step": 6050
+ },
+ {
+ "epoch": 3.4738041002277904,
+ "grad_norm": 2.423590898513794,
+ "learning_rate": 1.0086234949560691e-05,
+ "loss": 0.0111,
+ "step": 6100
+ },
+ {
+ "epoch": 3.502277904328018,
+ "grad_norm": 0.006337775848805904,
+ "learning_rate": 1.000488122356004e-05,
+ "loss": 0.0052,
+ "step": 6150
+ },
+ {
+ "epoch": 3.5307517084282463,
+ "grad_norm": 0.01373753696680069,
+ "learning_rate": 9.923527497559389e-06,
+ "loss": 0.009,
+ "step": 6200
+ },
+ {
+ "epoch": 3.559225512528474,
+ "grad_norm": 0.11260247975587845,
+ "learning_rate": 9.842173771558739e-06,
+ "loss": 0.0208,
+ "step": 6250
+ },
+ {
+ "epoch": 3.5876993166287017,
+ "grad_norm": 1.240822196006775,
+ "learning_rate": 9.760820045558087e-06,
+ "loss": 0.0157,
+ "step": 6300
+ },
+ {
+ "epoch": 3.6161731207289294,
+ "grad_norm": 0.05271293595433235,
+ "learning_rate": 9.679466319557436e-06,
+ "loss": 0.0175,
+ "step": 6350
+ },
+ {
+ "epoch": 3.644646924829157,
+ "grad_norm": 24.546663284301758,
+ "learning_rate": 9.598112593556786e-06,
+ "loss": 0.015,
+ "step": 6400
+ },
+ {
+ "epoch": 3.6731207289293852,
+ "grad_norm": 5.120173931121826,
+ "learning_rate": 9.516758867556136e-06,
+ "loss": 0.017,
+ "step": 6450
+ },
+ {
+ "epoch": 3.7015945330296125,
+ "grad_norm": 1.1922008991241455,
+ "learning_rate": 9.435405141555484e-06,
+ "loss": 0.0194,
+ "step": 6500
+ },
+ {
+ "epoch": 3.7300683371298406,
+ "grad_norm": 0.039439987391233444,
+ "learning_rate": 9.354051415554833e-06,
+ "loss": 0.0309,
+ "step": 6550
+ },
+ {
+ "epoch": 3.7585421412300684,
+ "grad_norm": 0.04055279493331909,
+ "learning_rate": 9.272697689554181e-06,
+ "loss": 0.0195,
+ "step": 6600
+ },
+ {
+ "epoch": 3.787015945330296,
+ "grad_norm": 0.5971085429191589,
+ "learning_rate": 9.191343963553533e-06,
+ "loss": 0.0156,
+ "step": 6650
+ },
+ {
+ "epoch": 3.8154897494305238,
+ "grad_norm": 0.2191866636276245,
+ "learning_rate": 9.10999023755288e-06,
+ "loss": 0.0174,
+ "step": 6700
+ },
+ {
+ "epoch": 3.8439635535307515,
+ "grad_norm": 0.033559828996658325,
+ "learning_rate": 9.02863651155223e-06,
+ "loss": 0.0149,
+ "step": 6750
+ },
+ {
+ "epoch": 3.8724373576309796,
+ "grad_norm": 3.7128634452819824,
+ "learning_rate": 8.947282785551578e-06,
+ "loss": 0.0221,
+ "step": 6800
+ },
+ {
+ "epoch": 3.9009111617312073,
+ "grad_norm": 0.08547580987215042,
+ "learning_rate": 8.865929059550928e-06,
+ "loss": 0.0153,
+ "step": 6850
+ },
+ {
+ "epoch": 3.929384965831435,
+ "grad_norm": 0.058124080300331116,
+ "learning_rate": 8.784575333550277e-06,
+ "loss": 0.0161,
+ "step": 6900
+ },
+ {
+ "epoch": 3.9578587699316627,
+ "grad_norm": 0.015579139813780785,
+ "learning_rate": 8.703221607549627e-06,
+ "loss": 0.011,
+ "step": 6950
+ },
+ {
+ "epoch": 3.9863325740318905,
+ "grad_norm": 0.12764935195446014,
+ "learning_rate": 8.621867881548975e-06,
+ "loss": 0.0166,
+ "step": 7000
+ },
+ {
+ "epoch": 4.0,
+ "eval_f1": 0.9525329748802823,
+ "eval_loss": 0.04983380436897278,
+ "eval_precision": 0.9510149303808086,
+ "eval_recall": 0.9540558734432851,
+ "eval_runtime": 3.7495,
+ "eval_samples_per_second": 866.79,
+ "eval_steps_per_second": 108.549,
+ "step": 7024
+ },
+ {
+ "epoch": 4.014806378132119,
+ "grad_norm": 0.04521024227142334,
+ "learning_rate": 8.540514155548325e-06,
+ "loss": 0.0102,
+ "step": 7050
+ },
+ {
+ "epoch": 4.043280182232346,
+ "grad_norm": 0.015535669401288033,
+ "learning_rate": 8.459160429547674e-06,
+ "loss": 0.0068,
+ "step": 7100
+ },
+ {
+ "epoch": 4.071753986332574,
+ "grad_norm": 12.566010475158691,
+ "learning_rate": 8.377806703547024e-06,
+ "loss": 0.0151,
+ "step": 7150
+ },
+ {
+ "epoch": 4.100227790432802,
+ "grad_norm": 0.027761396020650864,
+ "learning_rate": 8.296452977546372e-06,
+ "loss": 0.0053,
+ "step": 7200
+ },
+ {
+ "epoch": 4.128701594533029,
+ "grad_norm": 1.8846938610076904,
+ "learning_rate": 8.215099251545722e-06,
+ "loss": 0.0077,
+ "step": 7250
+ },
+ {
+ "epoch": 4.157175398633258,
+ "grad_norm": 0.005013479385524988,
+ "learning_rate": 8.13374552554507e-06,
+ "loss": 0.0073,
+ "step": 7300
+ },
+ {
+ "epoch": 4.185649202733485,
+ "grad_norm": 0.32696235179901123,
+ "learning_rate": 8.052391799544421e-06,
+ "loss": 0.0092,
+ "step": 7350
+ },
+ {
+ "epoch": 4.214123006833713,
+ "grad_norm": 0.014357910491526127,
+ "learning_rate": 7.971038073543769e-06,
+ "loss": 0.0164,
+ "step": 7400
+ },
+ {
+ "epoch": 4.242596810933941,
+ "grad_norm": 0.010845329612493515,
+ "learning_rate": 7.889684347543118e-06,
+ "loss": 0.0123,
+ "step": 7450
+ },
+ {
+ "epoch": 4.271070615034168,
+ "grad_norm": 0.6969354748725891,
+ "learning_rate": 7.808330621542466e-06,
+ "loss": 0.0103,
+ "step": 7500
+ },
+ {
+ "epoch": 4.2995444191343966,
+ "grad_norm": 0.3575742840766907,
+ "learning_rate": 7.726976895541816e-06,
+ "loss": 0.0161,
+ "step": 7550
+ },
+ {
+ "epoch": 4.328018223234624,
+ "grad_norm": 0.008076228201389313,
+ "learning_rate": 7.645623169541166e-06,
+ "loss": 0.0122,
+ "step": 7600
+ },
+ {
+ "epoch": 4.356492027334852,
+ "grad_norm": 0.029101597145199776,
+ "learning_rate": 7.564269443540515e-06,
+ "loss": 0.0071,
+ "step": 7650
+ },
+ {
+ "epoch": 4.38496583143508,
+ "grad_norm": 0.06841142475605011,
+ "learning_rate": 7.482915717539863e-06,
+ "loss": 0.0091,
+ "step": 7700
+ },
+ {
+ "epoch": 4.413439635535307,
+ "grad_norm": 0.04133535176515579,
+ "learning_rate": 7.401561991539213e-06,
+ "loss": 0.0073,
+ "step": 7750
+ },
+ {
+ "epoch": 4.4419134396355355,
+ "grad_norm": 0.03571132943034172,
+ "learning_rate": 7.320208265538562e-06,
+ "loss": 0.0131,
+ "step": 7800
+ },
+ {
+ "epoch": 4.470387243735763,
+ "grad_norm": 2.8286924362182617,
+ "learning_rate": 7.2388545395379114e-06,
+ "loss": 0.0043,
+ "step": 7850
+ },
+ {
+ "epoch": 4.498861047835991,
+ "grad_norm": 1.3785158395767212,
+ "learning_rate": 7.15750081353726e-06,
+ "loss": 0.0103,
+ "step": 7900
+ },
+ {
+ "epoch": 4.527334851936219,
+ "grad_norm": 0.05979786813259125,
+ "learning_rate": 7.07614708753661e-06,
+ "loss": 0.006,
+ "step": 7950
+ },
+ {
+ "epoch": 4.555808656036446,
+ "grad_norm": 1.3434393405914307,
+ "learning_rate": 6.994793361535959e-06,
+ "loss": 0.0148,
+ "step": 8000
+ },
+ {
+ "epoch": 4.5842824601366745,
+ "grad_norm": 0.019605603069067,
+ "learning_rate": 6.913439635535308e-06,
+ "loss": 0.0068,
+ "step": 8050
+ },
+ {
+ "epoch": 4.612756264236902,
+ "grad_norm": 0.0028891051188111305,
+ "learning_rate": 6.832085909534657e-06,
+ "loss": 0.0075,
+ "step": 8100
+ },
+ {
+ "epoch": 4.64123006833713,
+ "grad_norm": 0.014856363646686077,
+ "learning_rate": 6.750732183534007e-06,
+ "loss": 0.0073,
+ "step": 8150
+ },
+ {
+ "epoch": 4.669703872437358,
+ "grad_norm": 0.0037149768322706223,
+ "learning_rate": 6.669378457533356e-06,
+ "loss": 0.0066,
+ "step": 8200
+ },
+ {
+ "epoch": 4.698177676537585,
+ "grad_norm": 0.008534993045032024,
+ "learning_rate": 6.588024731532705e-06,
+ "loss": 0.0121,
+ "step": 8250
+ },
+ {
+ "epoch": 4.7266514806378135,
+ "grad_norm": 0.013920712284743786,
+ "learning_rate": 6.506671005532054e-06,
+ "loss": 0.013,
+ "step": 8300
+ },
+ {
+ "epoch": 4.755125284738041,
+ "grad_norm": 0.10121840238571167,
+ "learning_rate": 6.425317279531404e-06,
+ "loss": 0.0087,
+ "step": 8350
+ },
+ {
+ "epoch": 4.783599088838269,
+ "grad_norm": 0.07937771081924438,
+ "learning_rate": 6.343963553530752e-06,
+ "loss": 0.0086,
+ "step": 8400
+ },
+ {
+ "epoch": 4.812072892938497,
+ "grad_norm": 0.07674901187419891,
+ "learning_rate": 6.262609827530101e-06,
+ "loss": 0.0052,
+ "step": 8450
+ },
+ {
+ "epoch": 4.840546697038724,
+ "grad_norm": 0.06683178246021271,
+ "learning_rate": 6.18125610152945e-06,
+ "loss": 0.015,
+ "step": 8500
+ },
+ {
+ "epoch": 4.8690205011389525,
+ "grad_norm": 0.08861212432384491,
+ "learning_rate": 6.0999023755288e-06,
+ "loss": 0.0109,
+ "step": 8550
+ },
+ {
+ "epoch": 4.89749430523918,
+ "grad_norm": 0.006198623217642307,
+ "learning_rate": 6.018548649528149e-06,
+ "loss": 0.0181,
+ "step": 8600
+ },
+ {
+ "epoch": 4.925968109339408,
+ "grad_norm": 0.009029284119606018,
+ "learning_rate": 5.937194923527498e-06,
+ "loss": 0.0043,
+ "step": 8650
+ },
+ {
+ "epoch": 4.954441913439636,
+ "grad_norm": 1.1225602626800537,
+ "learning_rate": 5.855841197526847e-06,
+ "loss": 0.0121,
+ "step": 8700
+ },
+ {
+ "epoch": 4.982915717539863,
+ "grad_norm": 0.01218325924128294,
+ "learning_rate": 5.774487471526197e-06,
+ "loss": 0.0115,
+ "step": 8750
+ },
+ {
+ "epoch": 5.0,
+ "eval_f1": 0.9533444816053511,
+ "eval_loss": 0.04722925275564194,
+ "eval_precision": 0.9473246925889,
+ "eval_recall": 0.9594412655671491,
+ "eval_runtime": 3.7358,
+ "eval_samples_per_second": 869.968,
+ "eval_steps_per_second": 108.947,
+ "step": 8780
+ },
+ {
+ "epoch": 5.011389521640091,
+ "grad_norm": 0.8161097764968872,
+ "learning_rate": 5.6931337455255455e-06,
+ "loss": 0.0127,
+ "step": 8800
+ },
+ {
+ "epoch": 5.039863325740319,
+ "grad_norm": 0.048089127987623215,
+ "learning_rate": 5.611780019524895e-06,
+ "loss": 0.0116,
+ "step": 8850
+ },
+ {
+ "epoch": 5.068337129840547,
+ "grad_norm": 1.7576072216033936,
+ "learning_rate": 5.530426293524244e-06,
+ "loss": 0.0093,
+ "step": 8900
+ },
+ {
+ "epoch": 5.096810933940774,
+ "grad_norm": 0.07270015776157379,
+ "learning_rate": 5.449072567523594e-06,
+ "loss": 0.0043,
+ "step": 8950
+ },
+ {
+ "epoch": 5.125284738041002,
+ "grad_norm": 0.2580782175064087,
+ "learning_rate": 5.367718841522942e-06,
+ "loss": 0.0058,
+ "step": 9000
+ },
+ {
+ "epoch": 5.15375854214123,
+ "grad_norm": 0.12649740278720856,
+ "learning_rate": 5.286365115522292e-06,
+ "loss": 0.0033,
+ "step": 9050
+ },
+ {
+ "epoch": 5.182232346241458,
+ "grad_norm": 0.06850716471672058,
+ "learning_rate": 5.20501138952164e-06,
+ "loss": 0.003,
+ "step": 9100
+ },
+ {
+ "epoch": 5.210706150341686,
+ "grad_norm": 0.01197959017008543,
+ "learning_rate": 5.12365766352099e-06,
+ "loss": 0.0029,
+ "step": 9150
+ },
+ {
+ "epoch": 5.239179954441913,
+ "grad_norm": 0.10818086564540863,
+ "learning_rate": 5.0423039375203385e-06,
+ "loss": 0.0025,
+ "step": 9200
+ },
+ {
+ "epoch": 5.267653758542141,
+ "grad_norm": 0.035460665822029114,
+ "learning_rate": 4.960950211519688e-06,
+ "loss": 0.0047,
+ "step": 9250
+ },
+ {
+ "epoch": 5.296127562642369,
+ "grad_norm": 4.532562255859375,
+ "learning_rate": 4.879596485519037e-06,
+ "loss": 0.0086,
+ "step": 9300
+ },
+ {
+ "epoch": 5.324601366742597,
+ "grad_norm": 0.013236219063401222,
+ "learning_rate": 4.7982427595183866e-06,
+ "loss": 0.0027,
+ "step": 9350
+ },
+ {
+ "epoch": 5.353075170842825,
+ "grad_norm": 0.018059909343719482,
+ "learning_rate": 4.716889033517735e-06,
+ "loss": 0.0065,
+ "step": 9400
+ },
+ {
+ "epoch": 5.381548974943052,
+ "grad_norm": 0.019096272066235542,
+ "learning_rate": 4.635535307517084e-06,
+ "loss": 0.0053,
+ "step": 9450
+ },
+ {
+ "epoch": 5.41002277904328,
+ "grad_norm": 0.10389436781406403,
+ "learning_rate": 4.554181581516434e-06,
+ "loss": 0.0029,
+ "step": 9500
+ },
+ {
+ "epoch": 5.438496583143508,
+ "grad_norm": 0.0029252381063997746,
+ "learning_rate": 4.472827855515783e-06,
+ "loss": 0.0045,
+ "step": 9550
+ },
+ {
+ "epoch": 5.466970387243736,
+ "grad_norm": 0.004542892333120108,
+ "learning_rate": 4.391474129515132e-06,
+ "loss": 0.0102,
+ "step": 9600
+ },
+ {
+ "epoch": 5.495444191343964,
+ "grad_norm": 0.130916565656662,
+ "learning_rate": 4.310120403514481e-06,
+ "loss": 0.0064,
+ "step": 9650
+ },
+ {
+ "epoch": 5.523917995444191,
+ "grad_norm": 0.008013393729925156,
+ "learning_rate": 4.22876667751383e-06,
+ "loss": 0.005,
+ "step": 9700
+ },
+ {
+ "epoch": 5.552391799544419,
+ "grad_norm": 0.012782514095306396,
+ "learning_rate": 4.1474129515131795e-06,
+ "loss": 0.0122,
+ "step": 9750
+ },
+ {
+ "epoch": 5.5808656036446465,
+ "grad_norm": 0.9617053866386414,
+ "learning_rate": 4.066059225512528e-06,
+ "loss": 0.0063,
+ "step": 9800
+ },
+ {
+ "epoch": 5.609339407744875,
+ "grad_norm": 0.014953136444091797,
+ "learning_rate": 3.984705499511878e-06,
+ "loss": 0.0046,
+ "step": 9850
+ },
+ {
+ "epoch": 5.637813211845103,
+ "grad_norm": 1.0275633335113525,
+ "learning_rate": 3.903351773511227e-06,
+ "loss": 0.0075,
+ "step": 9900
+ },
+ {
+ "epoch": 5.66628701594533,
+ "grad_norm": 0.007999264635145664,
+ "learning_rate": 3.8219980475105764e-06,
+ "loss": 0.0031,
+ "step": 9950
+ },
+ {
+ "epoch": 5.694760820045558,
+ "grad_norm": 0.008241960778832436,
+ "learning_rate": 3.7406443215099252e-06,
+ "loss": 0.0065,
+ "step": 10000
+ },
+ {
+ "epoch": 5.723234624145785,
+ "grad_norm": 2.578386068344116,
+ "learning_rate": 3.6592905955092745e-06,
+ "loss": 0.0057,
+ "step": 10050
+ },
+ {
+ "epoch": 5.751708428246014,
+ "grad_norm": 2.8855443000793457,
+ "learning_rate": 3.5779368695086237e-06,
+ "loss": 0.0038,
+ "step": 10100
+ },
+ {
+ "epoch": 5.780182232346242,
+ "grad_norm": 0.016262667253613472,
+ "learning_rate": 3.496583143507973e-06,
+ "loss": 0.0128,
+ "step": 10150
+ },
+ {
+ "epoch": 5.808656036446469,
+ "grad_norm": 0.045168060809373856,
+ "learning_rate": 3.415229417507322e-06,
+ "loss": 0.0036,
+ "step": 10200
+ },
+ {
+ "epoch": 5.837129840546697,
+ "grad_norm": 0.1192622259259224,
+ "learning_rate": 3.3355027660266842e-06,
+ "loss": 0.0111,
+ "step": 10250
+ },
+ {
+ "epoch": 5.865603644646924,
+ "grad_norm": 0.08544190227985382,
+ "learning_rate": 3.2541490400260335e-06,
+ "loss": 0.0051,
+ "step": 10300
+ },
+ {
+ "epoch": 5.894077448747153,
+ "grad_norm": 3.493161678314209,
+ "learning_rate": 3.1727953140253827e-06,
+ "loss": 0.0054,
+ "step": 10350
+ },
+ {
+ "epoch": 5.922551252847381,
+ "grad_norm": 0.022789066657423973,
+ "learning_rate": 3.0914415880247315e-06,
+ "loss": 0.0033,
+ "step": 10400
+ },
+ {
+ "epoch": 5.951025056947608,
+ "grad_norm": 12.131625175476074,
+ "learning_rate": 3.0100878620240807e-06,
+ "loss": 0.0023,
+ "step": 10450
+ },
+ {
+ "epoch": 5.979498861047836,
+ "grad_norm": 0.03329641371965408,
+ "learning_rate": 2.92873413602343e-06,
+ "loss": 0.0058,
+ "step": 10500
+ },
+ {
+ "epoch": 6.0,
+ "eval_f1": 0.9557566616390145,
+ "eval_loss": 0.05410688370466232,
+ "eval_precision": 0.9517690253671562,
+ "eval_recall": 0.9597778525748906,
+ "eval_runtime": 3.6572,
+ "eval_samples_per_second": 888.662,
+ "eval_steps_per_second": 111.288,
+ "step": 10536
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2072296586745846.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-10536/training_args.bin b/checkpoint-10536/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-10536/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/checkpoint-12292/config.json b/checkpoint-12292/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-12292/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-12292/model.safetensors b/checkpoint-12292/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9138b6f1834a9abd518339bbab61b5b8559c4607
--- /dev/null
+++ b/checkpoint-12292/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9f3aa01fa5d49602728eca2190fb610b286d4348c8e470e053f83959dfd91049
+size 1109863956
diff --git a/checkpoint-12292/optimizer.pt b/checkpoint-12292/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..195893e68ae6f690522c151b09ff1d4bc11b6dfa
--- /dev/null
+++ b/checkpoint-12292/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:328145ba117ef4052abfc88cfdc5ed7cb46c854f712b8a60beb15d66d8807396
+size 2219846202
diff --git a/checkpoint-12292/rng_state.pth b/checkpoint-12292/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..576367a78de95019cc8bd62176d2c16e3a27f30e
--- /dev/null
+++ b/checkpoint-12292/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ab0a0c2e05c9b19ae2fb7f37ade343bd0d98b553cbd00cc47aa20772252004c6
+size 14244
diff --git a/checkpoint-12292/scaler.pt b/checkpoint-12292/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d61a5c4f84a88a4510f13223a1f7f6d2db3f658b
--- /dev/null
+++ b/checkpoint-12292/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6509928afa01d668bf56b569c4f6a2e6c2fde9763d17681c72b55f19ab92eb20
+size 988
diff --git a/checkpoint-12292/scheduler.pt b/checkpoint-12292/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b88ec99899c104ace02c2c6be36cd3d4d4b59aaa
--- /dev/null
+++ b/checkpoint-12292/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8f42e34b774989a80cf2635a24272d37af726ec5ad847c7648079f6697e6b0d3
+size 1064
diff --git a/checkpoint-12292/special_tokens_map.json b/checkpoint-12292/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-12292/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-12292/tokenizer.json b/checkpoint-12292/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-12292/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-12292/tokenizer_config.json b/checkpoint-12292/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-12292/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-12292/trainer_state.json b/checkpoint-12292/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..c688846b952266f6a1bc7372a95872d5586258e5
--- /dev/null
+++ b/checkpoint-12292/trainer_state.json
@@ -0,0 +1,1826 @@
+{
+ "best_global_step": 12292,
+ "best_metric": 0.9563834240267894,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-12292",
+ "epoch": 7.0,
+ "eval_steps": 500,
+ "global_step": 12292,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ },
+ {
+ "epoch": 1.0250569476082005,
+ "grad_norm": 4.280463695526123,
+ "learning_rate": 1.708102831109665e-05,
+ "loss": 0.0673,
+ "step": 1800
+ },
+ {
+ "epoch": 1.0535307517084282,
+ "grad_norm": 0.06692535430192947,
+ "learning_rate": 1.6999674585096e-05,
+ "loss": 0.0403,
+ "step": 1850
+ },
+ {
+ "epoch": 1.082004555808656,
+ "grad_norm": 0.03789375349879265,
+ "learning_rate": 1.691832085909535e-05,
+ "loss": 0.0496,
+ "step": 1900
+ },
+ {
+ "epoch": 1.1104783599088839,
+ "grad_norm": 3.698348045349121,
+ "learning_rate": 1.6836967133094698e-05,
+ "loss": 0.0381,
+ "step": 1950
+ },
+ {
+ "epoch": 1.1389521640091116,
+ "grad_norm": 2.753032684326172,
+ "learning_rate": 1.6755613407094048e-05,
+ "loss": 0.0432,
+ "step": 2000
+ },
+ {
+ "epoch": 1.1674259681093395,
+ "grad_norm": 0.06336130946874619,
+ "learning_rate": 1.6674259681093398e-05,
+ "loss": 0.0482,
+ "step": 2050
+ },
+ {
+ "epoch": 1.1958997722095672,
+ "grad_norm": 3.9503254890441895,
+ "learning_rate": 1.6592905955092744e-05,
+ "loss": 0.0436,
+ "step": 2100
+ },
+ {
+ "epoch": 1.224373576309795,
+ "grad_norm": 0.08679741621017456,
+ "learning_rate": 1.6511552229092093e-05,
+ "loss": 0.0346,
+ "step": 2150
+ },
+ {
+ "epoch": 1.2528473804100229,
+ "grad_norm": 2.1995127201080322,
+ "learning_rate": 1.6430198503091443e-05,
+ "loss": 0.0312,
+ "step": 2200
+ },
+ {
+ "epoch": 1.2813211845102506,
+ "grad_norm": 5.7709808349609375,
+ "learning_rate": 1.634884477709079e-05,
+ "loss": 0.032,
+ "step": 2250
+ },
+ {
+ "epoch": 1.3097949886104785,
+ "grad_norm": 0.2031829059123993,
+ "learning_rate": 1.626749105109014e-05,
+ "loss": 0.053,
+ "step": 2300
+ },
+ {
+ "epoch": 1.3382687927107062,
+ "grad_norm": 4.292157173156738,
+ "learning_rate": 1.6186137325089492e-05,
+ "loss": 0.0413,
+ "step": 2350
+ },
+ {
+ "epoch": 1.366742596810934,
+ "grad_norm": 0.07821401953697205,
+ "learning_rate": 1.6104783599088842e-05,
+ "loss": 0.0304,
+ "step": 2400
+ },
+ {
+ "epoch": 1.3952164009111616,
+ "grad_norm": 0.5930206179618835,
+ "learning_rate": 1.6023429873088188e-05,
+ "loss": 0.0525,
+ "step": 2450
+ },
+ {
+ "epoch": 1.4236902050113895,
+ "grad_norm": 1.6114712953567505,
+ "learning_rate": 1.5942076147087538e-05,
+ "loss": 0.0391,
+ "step": 2500
+ },
+ {
+ "epoch": 1.4521640091116172,
+ "grad_norm": 0.28632065653800964,
+ "learning_rate": 1.5860722421086887e-05,
+ "loss": 0.0491,
+ "step": 2550
+ },
+ {
+ "epoch": 1.4806378132118452,
+ "grad_norm": 1.0017669200897217,
+ "learning_rate": 1.5779368695086237e-05,
+ "loss": 0.0395,
+ "step": 2600
+ },
+ {
+ "epoch": 1.5091116173120729,
+ "grad_norm": 0.03409272059798241,
+ "learning_rate": 1.5698014969085583e-05,
+ "loss": 0.0399,
+ "step": 2650
+ },
+ {
+ "epoch": 1.5375854214123006,
+ "grad_norm": 3.5108702182769775,
+ "learning_rate": 1.5616661243084933e-05,
+ "loss": 0.0424,
+ "step": 2700
+ },
+ {
+ "epoch": 1.5660592255125285,
+ "grad_norm": 3.5875349044799805,
+ "learning_rate": 1.5535307517084283e-05,
+ "loss": 0.0497,
+ "step": 2750
+ },
+ {
+ "epoch": 1.5945330296127562,
+ "grad_norm": 0.10498251020908356,
+ "learning_rate": 1.5453953791083632e-05,
+ "loss": 0.0461,
+ "step": 2800
+ },
+ {
+ "epoch": 1.6230068337129842,
+ "grad_norm": 7.618963718414307,
+ "learning_rate": 1.5374227139602996e-05,
+ "loss": 0.0394,
+ "step": 2850
+ },
+ {
+ "epoch": 1.6514806378132119,
+ "grad_norm": 3.3920793533325195,
+ "learning_rate": 1.5292873413602346e-05,
+ "loss": 0.0339,
+ "step": 2900
+ },
+ {
+ "epoch": 1.6799544419134396,
+ "grad_norm": 0.04736129939556122,
+ "learning_rate": 1.5211519687601692e-05,
+ "loss": 0.0358,
+ "step": 2950
+ },
+ {
+ "epoch": 1.7084282460136673,
+ "grad_norm": 0.17463508248329163,
+ "learning_rate": 1.5130165961601042e-05,
+ "loss": 0.034,
+ "step": 3000
+ },
+ {
+ "epoch": 1.7369020501138952,
+ "grad_norm": 0.058148205280303955,
+ "learning_rate": 1.5048812235600392e-05,
+ "loss": 0.0317,
+ "step": 3050
+ },
+ {
+ "epoch": 1.7653758542141231,
+ "grad_norm": 3.5262815952301025,
+ "learning_rate": 1.4967458509599741e-05,
+ "loss": 0.0477,
+ "step": 3100
+ },
+ {
+ "epoch": 1.7938496583143508,
+ "grad_norm": 2.971937894821167,
+ "learning_rate": 1.488610478359909e-05,
+ "loss": 0.0364,
+ "step": 3150
+ },
+ {
+ "epoch": 1.8223234624145785,
+ "grad_norm": 0.05181724950671196,
+ "learning_rate": 1.4804751057598439e-05,
+ "loss": 0.0394,
+ "step": 3200
+ },
+ {
+ "epoch": 1.8507972665148062,
+ "grad_norm": 0.2712370455265045,
+ "learning_rate": 1.4723397331597789e-05,
+ "loss": 0.0456,
+ "step": 3250
+ },
+ {
+ "epoch": 1.8792710706150342,
+ "grad_norm": 1.608622431755066,
+ "learning_rate": 1.4642043605597138e-05,
+ "loss": 0.05,
+ "step": 3300
+ },
+ {
+ "epoch": 1.907744874715262,
+ "grad_norm": 5.2423529624938965,
+ "learning_rate": 1.4560689879596486e-05,
+ "loss": 0.0371,
+ "step": 3350
+ },
+ {
+ "epoch": 1.9362186788154898,
+ "grad_norm": 0.04786211624741554,
+ "learning_rate": 1.4479336153595836e-05,
+ "loss": 0.0329,
+ "step": 3400
+ },
+ {
+ "epoch": 1.9646924829157175,
+ "grad_norm": 0.5264931917190552,
+ "learning_rate": 1.4397982427595185e-05,
+ "loss": 0.0379,
+ "step": 3450
+ },
+ {
+ "epoch": 1.9931662870159452,
+ "grad_norm": 0.236006498336792,
+ "learning_rate": 1.4316628701594535e-05,
+ "loss": 0.0253,
+ "step": 3500
+ },
+ {
+ "epoch": 2.0,
+ "eval_f1": 0.9406438631790744,
+ "eval_loss": 0.05223705992102623,
+ "eval_precision": 0.9371867691279653,
+ "eval_recall": 0.9441265567149109,
+ "eval_runtime": 3.7109,
+ "eval_samples_per_second": 875.791,
+ "eval_steps_per_second": 109.676,
+ "step": 3512
+ },
+ {
+ "epoch": 2.021640091116173,
+ "grad_norm": 4.225963115692139,
+ "learning_rate": 1.4235274975593883e-05,
+ "loss": 0.0373,
+ "step": 3550
+ },
+ {
+ "epoch": 2.050113895216401,
+ "grad_norm": 0.15388111770153046,
+ "learning_rate": 1.4153921249593233e-05,
+ "loss": 0.0258,
+ "step": 3600
+ },
+ {
+ "epoch": 2.078587699316629,
+ "grad_norm": 6.3911895751953125,
+ "learning_rate": 1.4072567523592582e-05,
+ "loss": 0.0217,
+ "step": 3650
+ },
+ {
+ "epoch": 2.1070615034168565,
+ "grad_norm": 0.03075530007481575,
+ "learning_rate": 1.3991213797591932e-05,
+ "loss": 0.019,
+ "step": 3700
+ },
+ {
+ "epoch": 2.135535307517084,
+ "grad_norm": 0.16612432897090912,
+ "learning_rate": 1.3909860071591278e-05,
+ "loss": 0.0175,
+ "step": 3750
+ },
+ {
+ "epoch": 2.164009111617312,
+ "grad_norm": 0.14543022215366364,
+ "learning_rate": 1.382850634559063e-05,
+ "loss": 0.0289,
+ "step": 3800
+ },
+ {
+ "epoch": 2.19248291571754,
+ "grad_norm": 2.9880359172821045,
+ "learning_rate": 1.374715261958998e-05,
+ "loss": 0.0208,
+ "step": 3850
+ },
+ {
+ "epoch": 2.2209567198177678,
+ "grad_norm": 9.878133773803711,
+ "learning_rate": 1.3665798893589329e-05,
+ "loss": 0.021,
+ "step": 3900
+ },
+ {
+ "epoch": 2.2494305239179955,
+ "grad_norm": 9.684176445007324,
+ "learning_rate": 1.3584445167588675e-05,
+ "loss": 0.0267,
+ "step": 3950
+ },
+ {
+ "epoch": 2.277904328018223,
+ "grad_norm": 33.15859603881836,
+ "learning_rate": 1.3503091441588025e-05,
+ "loss": 0.0295,
+ "step": 4000
+ },
+ {
+ "epoch": 2.306378132118451,
+ "grad_norm": 2.559544086456299,
+ "learning_rate": 1.3421737715587374e-05,
+ "loss": 0.0292,
+ "step": 4050
+ },
+ {
+ "epoch": 2.334851936218679,
+ "grad_norm": 0.07091552764177322,
+ "learning_rate": 1.3340383989586724e-05,
+ "loss": 0.0262,
+ "step": 4100
+ },
+ {
+ "epoch": 2.3633257403189067,
+ "grad_norm": 1.8119585514068604,
+ "learning_rate": 1.3259030263586072e-05,
+ "loss": 0.03,
+ "step": 4150
+ },
+ {
+ "epoch": 2.3917995444191344,
+ "grad_norm": 0.1476636379957199,
+ "learning_rate": 1.3177676537585422e-05,
+ "loss": 0.0196,
+ "step": 4200
+ },
+ {
+ "epoch": 2.420273348519362,
+ "grad_norm": 0.04482650011777878,
+ "learning_rate": 1.3096322811584771e-05,
+ "loss": 0.0263,
+ "step": 4250
+ },
+ {
+ "epoch": 2.44874715261959,
+ "grad_norm": 0.09201560169458389,
+ "learning_rate": 1.3014969085584121e-05,
+ "loss": 0.0172,
+ "step": 4300
+ },
+ {
+ "epoch": 2.477220956719818,
+ "grad_norm": 0.02215876244008541,
+ "learning_rate": 1.2933615359583469e-05,
+ "loss": 0.0149,
+ "step": 4350
+ },
+ {
+ "epoch": 2.5056947608200457,
+ "grad_norm": 0.6993932127952576,
+ "learning_rate": 1.2852261633582819e-05,
+ "loss": 0.0331,
+ "step": 4400
+ },
+ {
+ "epoch": 2.5341685649202734,
+ "grad_norm": 0.030453965067863464,
+ "learning_rate": 1.2770907907582168e-05,
+ "loss": 0.0307,
+ "step": 4450
+ },
+ {
+ "epoch": 2.562642369020501,
+ "grad_norm": 0.22211593389511108,
+ "learning_rate": 1.2689554181581518e-05,
+ "loss": 0.0183,
+ "step": 4500
+ },
+ {
+ "epoch": 2.591116173120729,
+ "grad_norm": 1.2842738628387451,
+ "learning_rate": 1.2608200455580866e-05,
+ "loss": 0.0226,
+ "step": 4550
+ },
+ {
+ "epoch": 2.619589977220957,
+ "grad_norm": 3.9048423767089844,
+ "learning_rate": 1.2526846729580216e-05,
+ "loss": 0.0239,
+ "step": 4600
+ },
+ {
+ "epoch": 2.6480637813211843,
+ "grad_norm": 0.9887784719467163,
+ "learning_rate": 1.2445493003579565e-05,
+ "loss": 0.0271,
+ "step": 4650
+ },
+ {
+ "epoch": 2.6765375854214124,
+ "grad_norm": 4.759728908538818,
+ "learning_rate": 1.2364139277578915e-05,
+ "loss": 0.0201,
+ "step": 4700
+ },
+ {
+ "epoch": 2.70501138952164,
+ "grad_norm": 0.2671191096305847,
+ "learning_rate": 1.2282785551578263e-05,
+ "loss": 0.0212,
+ "step": 4750
+ },
+ {
+ "epoch": 2.733485193621868,
+ "grad_norm": 0.036626849323511124,
+ "learning_rate": 1.2201431825577612e-05,
+ "loss": 0.0155,
+ "step": 4800
+ },
+ {
+ "epoch": 2.7619589977220955,
+ "grad_norm": 6.877408027648926,
+ "learning_rate": 1.2120078099576962e-05,
+ "loss": 0.0266,
+ "step": 4850
+ },
+ {
+ "epoch": 2.7904328018223232,
+ "grad_norm": 4.725896835327148,
+ "learning_rate": 1.2038724373576312e-05,
+ "loss": 0.0163,
+ "step": 4900
+ },
+ {
+ "epoch": 2.8189066059225514,
+ "grad_norm": 0.11965059489011765,
+ "learning_rate": 1.195737064757566e-05,
+ "loss": 0.0246,
+ "step": 4950
+ },
+ {
+ "epoch": 2.847380410022779,
+ "grad_norm": 0.6634079813957214,
+ "learning_rate": 1.187601692157501e-05,
+ "loss": 0.0304,
+ "step": 5000
+ },
+ {
+ "epoch": 2.875854214123007,
+ "grad_norm": 3.952694892883301,
+ "learning_rate": 1.1794663195574359e-05,
+ "loss": 0.0183,
+ "step": 5050
+ },
+ {
+ "epoch": 2.9043280182232345,
+ "grad_norm": 19.788942337036133,
+ "learning_rate": 1.1713309469573709e-05,
+ "loss": 0.0193,
+ "step": 5100
+ },
+ {
+ "epoch": 2.932801822323462,
+ "grad_norm": 4.989261627197266,
+ "learning_rate": 1.1631955743573055e-05,
+ "loss": 0.023,
+ "step": 5150
+ },
+ {
+ "epoch": 2.9612756264236904,
+ "grad_norm": 9.020853996276855,
+ "learning_rate": 1.1550602017572406e-05,
+ "loss": 0.0231,
+ "step": 5200
+ },
+ {
+ "epoch": 2.989749430523918,
+ "grad_norm": 0.9151140451431274,
+ "learning_rate": 1.1469248291571756e-05,
+ "loss": 0.0314,
+ "step": 5250
+ },
+ {
+ "epoch": 3.0,
+ "eval_f1": 0.9470420646891236,
+ "eval_loss": 0.04870199039578438,
+ "eval_precision": 0.943090787716956,
+ "eval_recall": 0.9510265903736116,
+ "eval_runtime": 3.7494,
+ "eval_samples_per_second": 866.805,
+ "eval_steps_per_second": 108.551,
+ "step": 5268
+ },
+ {
+ "epoch": 3.0182232346241458,
+ "grad_norm": 8.62992000579834,
+ "learning_rate": 1.1387894565571106e-05,
+ "loss": 0.0194,
+ "step": 5300
+ },
+ {
+ "epoch": 3.0466970387243735,
+ "grad_norm": 0.5615554451942444,
+ "learning_rate": 1.1306540839570452e-05,
+ "loss": 0.0114,
+ "step": 5350
+ },
+ {
+ "epoch": 3.075170842824601,
+ "grad_norm": 2.5735936164855957,
+ "learning_rate": 1.1225187113569801e-05,
+ "loss": 0.0163,
+ "step": 5400
+ },
+ {
+ "epoch": 3.1036446469248293,
+ "grad_norm": 6.504094123840332,
+ "learning_rate": 1.1143833387569151e-05,
+ "loss": 0.0135,
+ "step": 5450
+ },
+ {
+ "epoch": 3.132118451025057,
+ "grad_norm": 14.477537155151367,
+ "learning_rate": 1.10624796615685e-05,
+ "loss": 0.0129,
+ "step": 5500
+ },
+ {
+ "epoch": 3.1605922551252847,
+ "grad_norm": 0.04929841682314873,
+ "learning_rate": 1.0981125935567849e-05,
+ "loss": 0.0263,
+ "step": 5550
+ },
+ {
+ "epoch": 3.1890660592255125,
+ "grad_norm": 0.08875144273042679,
+ "learning_rate": 1.0899772209567198e-05,
+ "loss": 0.0134,
+ "step": 5600
+ },
+ {
+ "epoch": 3.21753986332574,
+ "grad_norm": 0.6238455772399902,
+ "learning_rate": 1.0818418483566548e-05,
+ "loss": 0.0107,
+ "step": 5650
+ },
+ {
+ "epoch": 3.2460136674259683,
+ "grad_norm": 0.004948179703205824,
+ "learning_rate": 1.0737064757565898e-05,
+ "loss": 0.0167,
+ "step": 5700
+ },
+ {
+ "epoch": 3.274487471526196,
+ "grad_norm": 0.017031218856573105,
+ "learning_rate": 1.0655711031565246e-05,
+ "loss": 0.0085,
+ "step": 5750
+ },
+ {
+ "epoch": 3.3029612756264237,
+ "grad_norm": 2.9840469360351562,
+ "learning_rate": 1.0574357305564595e-05,
+ "loss": 0.0158,
+ "step": 5800
+ },
+ {
+ "epoch": 3.3314350797266514,
+ "grad_norm": 1.7477540969848633,
+ "learning_rate": 1.0493003579563945e-05,
+ "loss": 0.0157,
+ "step": 5850
+ },
+ {
+ "epoch": 3.359908883826879,
+ "grad_norm": 0.03963543102145195,
+ "learning_rate": 1.0411649853563295e-05,
+ "loss": 0.0201,
+ "step": 5900
+ },
+ {
+ "epoch": 3.3883826879271073,
+ "grad_norm": 0.16669808328151703,
+ "learning_rate": 1.0330296127562643e-05,
+ "loss": 0.0159,
+ "step": 5950
+ },
+ {
+ "epoch": 3.416856492027335,
+ "grad_norm": 0.01953568309545517,
+ "learning_rate": 1.0248942401561992e-05,
+ "loss": 0.0117,
+ "step": 6000
+ },
+ {
+ "epoch": 3.4453302961275627,
+ "grad_norm": 0.07708246260881424,
+ "learning_rate": 1.0167588675561342e-05,
+ "loss": 0.0117,
+ "step": 6050
+ },
+ {
+ "epoch": 3.4738041002277904,
+ "grad_norm": 2.423590898513794,
+ "learning_rate": 1.0086234949560691e-05,
+ "loss": 0.0111,
+ "step": 6100
+ },
+ {
+ "epoch": 3.502277904328018,
+ "grad_norm": 0.006337775848805904,
+ "learning_rate": 1.000488122356004e-05,
+ "loss": 0.0052,
+ "step": 6150
+ },
+ {
+ "epoch": 3.5307517084282463,
+ "grad_norm": 0.01373753696680069,
+ "learning_rate": 9.923527497559389e-06,
+ "loss": 0.009,
+ "step": 6200
+ },
+ {
+ "epoch": 3.559225512528474,
+ "grad_norm": 0.11260247975587845,
+ "learning_rate": 9.842173771558739e-06,
+ "loss": 0.0208,
+ "step": 6250
+ },
+ {
+ "epoch": 3.5876993166287017,
+ "grad_norm": 1.240822196006775,
+ "learning_rate": 9.760820045558087e-06,
+ "loss": 0.0157,
+ "step": 6300
+ },
+ {
+ "epoch": 3.6161731207289294,
+ "grad_norm": 0.05271293595433235,
+ "learning_rate": 9.679466319557436e-06,
+ "loss": 0.0175,
+ "step": 6350
+ },
+ {
+ "epoch": 3.644646924829157,
+ "grad_norm": 24.546663284301758,
+ "learning_rate": 9.598112593556786e-06,
+ "loss": 0.015,
+ "step": 6400
+ },
+ {
+ "epoch": 3.6731207289293852,
+ "grad_norm": 5.120173931121826,
+ "learning_rate": 9.516758867556136e-06,
+ "loss": 0.017,
+ "step": 6450
+ },
+ {
+ "epoch": 3.7015945330296125,
+ "grad_norm": 1.1922008991241455,
+ "learning_rate": 9.435405141555484e-06,
+ "loss": 0.0194,
+ "step": 6500
+ },
+ {
+ "epoch": 3.7300683371298406,
+ "grad_norm": 0.039439987391233444,
+ "learning_rate": 9.354051415554833e-06,
+ "loss": 0.0309,
+ "step": 6550
+ },
+ {
+ "epoch": 3.7585421412300684,
+ "grad_norm": 0.04055279493331909,
+ "learning_rate": 9.272697689554181e-06,
+ "loss": 0.0195,
+ "step": 6600
+ },
+ {
+ "epoch": 3.787015945330296,
+ "grad_norm": 0.5971085429191589,
+ "learning_rate": 9.191343963553533e-06,
+ "loss": 0.0156,
+ "step": 6650
+ },
+ {
+ "epoch": 3.8154897494305238,
+ "grad_norm": 0.2191866636276245,
+ "learning_rate": 9.10999023755288e-06,
+ "loss": 0.0174,
+ "step": 6700
+ },
+ {
+ "epoch": 3.8439635535307515,
+ "grad_norm": 0.033559828996658325,
+ "learning_rate": 9.02863651155223e-06,
+ "loss": 0.0149,
+ "step": 6750
+ },
+ {
+ "epoch": 3.8724373576309796,
+ "grad_norm": 3.7128634452819824,
+ "learning_rate": 8.947282785551578e-06,
+ "loss": 0.0221,
+ "step": 6800
+ },
+ {
+ "epoch": 3.9009111617312073,
+ "grad_norm": 0.08547580987215042,
+ "learning_rate": 8.865929059550928e-06,
+ "loss": 0.0153,
+ "step": 6850
+ },
+ {
+ "epoch": 3.929384965831435,
+ "grad_norm": 0.058124080300331116,
+ "learning_rate": 8.784575333550277e-06,
+ "loss": 0.0161,
+ "step": 6900
+ },
+ {
+ "epoch": 3.9578587699316627,
+ "grad_norm": 0.015579139813780785,
+ "learning_rate": 8.703221607549627e-06,
+ "loss": 0.011,
+ "step": 6950
+ },
+ {
+ "epoch": 3.9863325740318905,
+ "grad_norm": 0.12764935195446014,
+ "learning_rate": 8.621867881548975e-06,
+ "loss": 0.0166,
+ "step": 7000
+ },
+ {
+ "epoch": 4.0,
+ "eval_f1": 0.9525329748802823,
+ "eval_loss": 0.04983380436897278,
+ "eval_precision": 0.9510149303808086,
+ "eval_recall": 0.9540558734432851,
+ "eval_runtime": 3.7495,
+ "eval_samples_per_second": 866.79,
+ "eval_steps_per_second": 108.549,
+ "step": 7024
+ },
+ {
+ "epoch": 4.014806378132119,
+ "grad_norm": 0.04521024227142334,
+ "learning_rate": 8.540514155548325e-06,
+ "loss": 0.0102,
+ "step": 7050
+ },
+ {
+ "epoch": 4.043280182232346,
+ "grad_norm": 0.015535669401288033,
+ "learning_rate": 8.459160429547674e-06,
+ "loss": 0.0068,
+ "step": 7100
+ },
+ {
+ "epoch": 4.071753986332574,
+ "grad_norm": 12.566010475158691,
+ "learning_rate": 8.377806703547024e-06,
+ "loss": 0.0151,
+ "step": 7150
+ },
+ {
+ "epoch": 4.100227790432802,
+ "grad_norm": 0.027761396020650864,
+ "learning_rate": 8.296452977546372e-06,
+ "loss": 0.0053,
+ "step": 7200
+ },
+ {
+ "epoch": 4.128701594533029,
+ "grad_norm": 1.8846938610076904,
+ "learning_rate": 8.215099251545722e-06,
+ "loss": 0.0077,
+ "step": 7250
+ },
+ {
+ "epoch": 4.157175398633258,
+ "grad_norm": 0.005013479385524988,
+ "learning_rate": 8.13374552554507e-06,
+ "loss": 0.0073,
+ "step": 7300
+ },
+ {
+ "epoch": 4.185649202733485,
+ "grad_norm": 0.32696235179901123,
+ "learning_rate": 8.052391799544421e-06,
+ "loss": 0.0092,
+ "step": 7350
+ },
+ {
+ "epoch": 4.214123006833713,
+ "grad_norm": 0.014357910491526127,
+ "learning_rate": 7.971038073543769e-06,
+ "loss": 0.0164,
+ "step": 7400
+ },
+ {
+ "epoch": 4.242596810933941,
+ "grad_norm": 0.010845329612493515,
+ "learning_rate": 7.889684347543118e-06,
+ "loss": 0.0123,
+ "step": 7450
+ },
+ {
+ "epoch": 4.271070615034168,
+ "grad_norm": 0.6969354748725891,
+ "learning_rate": 7.808330621542466e-06,
+ "loss": 0.0103,
+ "step": 7500
+ },
+ {
+ "epoch": 4.2995444191343966,
+ "grad_norm": 0.3575742840766907,
+ "learning_rate": 7.726976895541816e-06,
+ "loss": 0.0161,
+ "step": 7550
+ },
+ {
+ "epoch": 4.328018223234624,
+ "grad_norm": 0.008076228201389313,
+ "learning_rate": 7.645623169541166e-06,
+ "loss": 0.0122,
+ "step": 7600
+ },
+ {
+ "epoch": 4.356492027334852,
+ "grad_norm": 0.029101597145199776,
+ "learning_rate": 7.564269443540515e-06,
+ "loss": 0.0071,
+ "step": 7650
+ },
+ {
+ "epoch": 4.38496583143508,
+ "grad_norm": 0.06841142475605011,
+ "learning_rate": 7.482915717539863e-06,
+ "loss": 0.0091,
+ "step": 7700
+ },
+ {
+ "epoch": 4.413439635535307,
+ "grad_norm": 0.04133535176515579,
+ "learning_rate": 7.401561991539213e-06,
+ "loss": 0.0073,
+ "step": 7750
+ },
+ {
+ "epoch": 4.4419134396355355,
+ "grad_norm": 0.03571132943034172,
+ "learning_rate": 7.320208265538562e-06,
+ "loss": 0.0131,
+ "step": 7800
+ },
+ {
+ "epoch": 4.470387243735763,
+ "grad_norm": 2.8286924362182617,
+ "learning_rate": 7.2388545395379114e-06,
+ "loss": 0.0043,
+ "step": 7850
+ },
+ {
+ "epoch": 4.498861047835991,
+ "grad_norm": 1.3785158395767212,
+ "learning_rate": 7.15750081353726e-06,
+ "loss": 0.0103,
+ "step": 7900
+ },
+ {
+ "epoch": 4.527334851936219,
+ "grad_norm": 0.05979786813259125,
+ "learning_rate": 7.07614708753661e-06,
+ "loss": 0.006,
+ "step": 7950
+ },
+ {
+ "epoch": 4.555808656036446,
+ "grad_norm": 1.3434393405914307,
+ "learning_rate": 6.994793361535959e-06,
+ "loss": 0.0148,
+ "step": 8000
+ },
+ {
+ "epoch": 4.5842824601366745,
+ "grad_norm": 0.019605603069067,
+ "learning_rate": 6.913439635535308e-06,
+ "loss": 0.0068,
+ "step": 8050
+ },
+ {
+ "epoch": 4.612756264236902,
+ "grad_norm": 0.0028891051188111305,
+ "learning_rate": 6.832085909534657e-06,
+ "loss": 0.0075,
+ "step": 8100
+ },
+ {
+ "epoch": 4.64123006833713,
+ "grad_norm": 0.014856363646686077,
+ "learning_rate": 6.750732183534007e-06,
+ "loss": 0.0073,
+ "step": 8150
+ },
+ {
+ "epoch": 4.669703872437358,
+ "grad_norm": 0.0037149768322706223,
+ "learning_rate": 6.669378457533356e-06,
+ "loss": 0.0066,
+ "step": 8200
+ },
+ {
+ "epoch": 4.698177676537585,
+ "grad_norm": 0.008534993045032024,
+ "learning_rate": 6.588024731532705e-06,
+ "loss": 0.0121,
+ "step": 8250
+ },
+ {
+ "epoch": 4.7266514806378135,
+ "grad_norm": 0.013920712284743786,
+ "learning_rate": 6.506671005532054e-06,
+ "loss": 0.013,
+ "step": 8300
+ },
+ {
+ "epoch": 4.755125284738041,
+ "grad_norm": 0.10121840238571167,
+ "learning_rate": 6.425317279531404e-06,
+ "loss": 0.0087,
+ "step": 8350
+ },
+ {
+ "epoch": 4.783599088838269,
+ "grad_norm": 0.07937771081924438,
+ "learning_rate": 6.343963553530752e-06,
+ "loss": 0.0086,
+ "step": 8400
+ },
+ {
+ "epoch": 4.812072892938497,
+ "grad_norm": 0.07674901187419891,
+ "learning_rate": 6.262609827530101e-06,
+ "loss": 0.0052,
+ "step": 8450
+ },
+ {
+ "epoch": 4.840546697038724,
+ "grad_norm": 0.06683178246021271,
+ "learning_rate": 6.18125610152945e-06,
+ "loss": 0.015,
+ "step": 8500
+ },
+ {
+ "epoch": 4.8690205011389525,
+ "grad_norm": 0.08861212432384491,
+ "learning_rate": 6.0999023755288e-06,
+ "loss": 0.0109,
+ "step": 8550
+ },
+ {
+ "epoch": 4.89749430523918,
+ "grad_norm": 0.006198623217642307,
+ "learning_rate": 6.018548649528149e-06,
+ "loss": 0.0181,
+ "step": 8600
+ },
+ {
+ "epoch": 4.925968109339408,
+ "grad_norm": 0.009029284119606018,
+ "learning_rate": 5.937194923527498e-06,
+ "loss": 0.0043,
+ "step": 8650
+ },
+ {
+ "epoch": 4.954441913439636,
+ "grad_norm": 1.1225602626800537,
+ "learning_rate": 5.855841197526847e-06,
+ "loss": 0.0121,
+ "step": 8700
+ },
+ {
+ "epoch": 4.982915717539863,
+ "grad_norm": 0.01218325924128294,
+ "learning_rate": 5.774487471526197e-06,
+ "loss": 0.0115,
+ "step": 8750
+ },
+ {
+ "epoch": 5.0,
+ "eval_f1": 0.9533444816053511,
+ "eval_loss": 0.04722925275564194,
+ "eval_precision": 0.9473246925889,
+ "eval_recall": 0.9594412655671491,
+ "eval_runtime": 3.7358,
+ "eval_samples_per_second": 869.968,
+ "eval_steps_per_second": 108.947,
+ "step": 8780
+ },
+ {
+ "epoch": 5.011389521640091,
+ "grad_norm": 0.8161097764968872,
+ "learning_rate": 5.6931337455255455e-06,
+ "loss": 0.0127,
+ "step": 8800
+ },
+ {
+ "epoch": 5.039863325740319,
+ "grad_norm": 0.048089127987623215,
+ "learning_rate": 5.611780019524895e-06,
+ "loss": 0.0116,
+ "step": 8850
+ },
+ {
+ "epoch": 5.068337129840547,
+ "grad_norm": 1.7576072216033936,
+ "learning_rate": 5.530426293524244e-06,
+ "loss": 0.0093,
+ "step": 8900
+ },
+ {
+ "epoch": 5.096810933940774,
+ "grad_norm": 0.07270015776157379,
+ "learning_rate": 5.449072567523594e-06,
+ "loss": 0.0043,
+ "step": 8950
+ },
+ {
+ "epoch": 5.125284738041002,
+ "grad_norm": 0.2580782175064087,
+ "learning_rate": 5.367718841522942e-06,
+ "loss": 0.0058,
+ "step": 9000
+ },
+ {
+ "epoch": 5.15375854214123,
+ "grad_norm": 0.12649740278720856,
+ "learning_rate": 5.286365115522292e-06,
+ "loss": 0.0033,
+ "step": 9050
+ },
+ {
+ "epoch": 5.182232346241458,
+ "grad_norm": 0.06850716471672058,
+ "learning_rate": 5.20501138952164e-06,
+ "loss": 0.003,
+ "step": 9100
+ },
+ {
+ "epoch": 5.210706150341686,
+ "grad_norm": 0.01197959017008543,
+ "learning_rate": 5.12365766352099e-06,
+ "loss": 0.0029,
+ "step": 9150
+ },
+ {
+ "epoch": 5.239179954441913,
+ "grad_norm": 0.10818086564540863,
+ "learning_rate": 5.0423039375203385e-06,
+ "loss": 0.0025,
+ "step": 9200
+ },
+ {
+ "epoch": 5.267653758542141,
+ "grad_norm": 0.035460665822029114,
+ "learning_rate": 4.960950211519688e-06,
+ "loss": 0.0047,
+ "step": 9250
+ },
+ {
+ "epoch": 5.296127562642369,
+ "grad_norm": 4.532562255859375,
+ "learning_rate": 4.879596485519037e-06,
+ "loss": 0.0086,
+ "step": 9300
+ },
+ {
+ "epoch": 5.324601366742597,
+ "grad_norm": 0.013236219063401222,
+ "learning_rate": 4.7982427595183866e-06,
+ "loss": 0.0027,
+ "step": 9350
+ },
+ {
+ "epoch": 5.353075170842825,
+ "grad_norm": 0.018059909343719482,
+ "learning_rate": 4.716889033517735e-06,
+ "loss": 0.0065,
+ "step": 9400
+ },
+ {
+ "epoch": 5.381548974943052,
+ "grad_norm": 0.019096272066235542,
+ "learning_rate": 4.635535307517084e-06,
+ "loss": 0.0053,
+ "step": 9450
+ },
+ {
+ "epoch": 5.41002277904328,
+ "grad_norm": 0.10389436781406403,
+ "learning_rate": 4.554181581516434e-06,
+ "loss": 0.0029,
+ "step": 9500
+ },
+ {
+ "epoch": 5.438496583143508,
+ "grad_norm": 0.0029252381063997746,
+ "learning_rate": 4.472827855515783e-06,
+ "loss": 0.0045,
+ "step": 9550
+ },
+ {
+ "epoch": 5.466970387243736,
+ "grad_norm": 0.004542892333120108,
+ "learning_rate": 4.391474129515132e-06,
+ "loss": 0.0102,
+ "step": 9600
+ },
+ {
+ "epoch": 5.495444191343964,
+ "grad_norm": 0.130916565656662,
+ "learning_rate": 4.310120403514481e-06,
+ "loss": 0.0064,
+ "step": 9650
+ },
+ {
+ "epoch": 5.523917995444191,
+ "grad_norm": 0.008013393729925156,
+ "learning_rate": 4.22876667751383e-06,
+ "loss": 0.005,
+ "step": 9700
+ },
+ {
+ "epoch": 5.552391799544419,
+ "grad_norm": 0.012782514095306396,
+ "learning_rate": 4.1474129515131795e-06,
+ "loss": 0.0122,
+ "step": 9750
+ },
+ {
+ "epoch": 5.5808656036446465,
+ "grad_norm": 0.9617053866386414,
+ "learning_rate": 4.066059225512528e-06,
+ "loss": 0.0063,
+ "step": 9800
+ },
+ {
+ "epoch": 5.609339407744875,
+ "grad_norm": 0.014953136444091797,
+ "learning_rate": 3.984705499511878e-06,
+ "loss": 0.0046,
+ "step": 9850
+ },
+ {
+ "epoch": 5.637813211845103,
+ "grad_norm": 1.0275633335113525,
+ "learning_rate": 3.903351773511227e-06,
+ "loss": 0.0075,
+ "step": 9900
+ },
+ {
+ "epoch": 5.66628701594533,
+ "grad_norm": 0.007999264635145664,
+ "learning_rate": 3.8219980475105764e-06,
+ "loss": 0.0031,
+ "step": 9950
+ },
+ {
+ "epoch": 5.694760820045558,
+ "grad_norm": 0.008241960778832436,
+ "learning_rate": 3.7406443215099252e-06,
+ "loss": 0.0065,
+ "step": 10000
+ },
+ {
+ "epoch": 5.723234624145785,
+ "grad_norm": 2.578386068344116,
+ "learning_rate": 3.6592905955092745e-06,
+ "loss": 0.0057,
+ "step": 10050
+ },
+ {
+ "epoch": 5.751708428246014,
+ "grad_norm": 2.8855443000793457,
+ "learning_rate": 3.5779368695086237e-06,
+ "loss": 0.0038,
+ "step": 10100
+ },
+ {
+ "epoch": 5.780182232346242,
+ "grad_norm": 0.016262667253613472,
+ "learning_rate": 3.496583143507973e-06,
+ "loss": 0.0128,
+ "step": 10150
+ },
+ {
+ "epoch": 5.808656036446469,
+ "grad_norm": 0.045168060809373856,
+ "learning_rate": 3.415229417507322e-06,
+ "loss": 0.0036,
+ "step": 10200
+ },
+ {
+ "epoch": 5.837129840546697,
+ "grad_norm": 0.1192622259259224,
+ "learning_rate": 3.3355027660266842e-06,
+ "loss": 0.0111,
+ "step": 10250
+ },
+ {
+ "epoch": 5.865603644646924,
+ "grad_norm": 0.08544190227985382,
+ "learning_rate": 3.2541490400260335e-06,
+ "loss": 0.0051,
+ "step": 10300
+ },
+ {
+ "epoch": 5.894077448747153,
+ "grad_norm": 3.493161678314209,
+ "learning_rate": 3.1727953140253827e-06,
+ "loss": 0.0054,
+ "step": 10350
+ },
+ {
+ "epoch": 5.922551252847381,
+ "grad_norm": 0.022789066657423973,
+ "learning_rate": 3.0914415880247315e-06,
+ "loss": 0.0033,
+ "step": 10400
+ },
+ {
+ "epoch": 5.951025056947608,
+ "grad_norm": 12.131625175476074,
+ "learning_rate": 3.0100878620240807e-06,
+ "loss": 0.0023,
+ "step": 10450
+ },
+ {
+ "epoch": 5.979498861047836,
+ "grad_norm": 0.03329641371965408,
+ "learning_rate": 2.92873413602343e-06,
+ "loss": 0.0058,
+ "step": 10500
+ },
+ {
+ "epoch": 6.0,
+ "eval_f1": 0.9557566616390145,
+ "eval_loss": 0.05410688370466232,
+ "eval_precision": 0.9517690253671562,
+ "eval_recall": 0.9597778525748906,
+ "eval_runtime": 3.6572,
+ "eval_samples_per_second": 888.662,
+ "eval_steps_per_second": 111.288,
+ "step": 10536
+ },
+ {
+ "epoch": 6.007972665148063,
+ "grad_norm": 0.008557640947401524,
+ "learning_rate": 2.847380410022779e-06,
+ "loss": 0.0152,
+ "step": 10550
+ },
+ {
+ "epoch": 6.0364464692482915,
+ "grad_norm": 0.08105529844760895,
+ "learning_rate": 2.7660266840221284e-06,
+ "loss": 0.0025,
+ "step": 10600
+ },
+ {
+ "epoch": 6.06492027334852,
+ "grad_norm": 1.100066065788269,
+ "learning_rate": 2.6846729580214776e-06,
+ "loss": 0.0015,
+ "step": 10650
+ },
+ {
+ "epoch": 6.093394077448747,
+ "grad_norm": 1.96909761428833,
+ "learning_rate": 2.603319232020827e-06,
+ "loss": 0.0021,
+ "step": 10700
+ },
+ {
+ "epoch": 6.121867881548975,
+ "grad_norm": 0.00583766121417284,
+ "learning_rate": 2.5219655060201757e-06,
+ "loss": 0.0127,
+ "step": 10750
+ },
+ {
+ "epoch": 6.150341685649202,
+ "grad_norm": 0.026862677186727524,
+ "learning_rate": 2.440611780019525e-06,
+ "loss": 0.004,
+ "step": 10800
+ },
+ {
+ "epoch": 6.1788154897494305,
+ "grad_norm": 0.010042566806077957,
+ "learning_rate": 2.359258054018874e-06,
+ "loss": 0.0033,
+ "step": 10850
+ },
+ {
+ "epoch": 6.207289293849659,
+ "grad_norm": 0.8956929445266724,
+ "learning_rate": 2.2779043280182233e-06,
+ "loss": 0.0031,
+ "step": 10900
+ },
+ {
+ "epoch": 6.235763097949886,
+ "grad_norm": 0.009118441492319107,
+ "learning_rate": 2.1965506020175726e-06,
+ "loss": 0.0039,
+ "step": 10950
+ },
+ {
+ "epoch": 6.264236902050114,
+ "grad_norm": 0.22793921828269958,
+ "learning_rate": 2.115196876016922e-06,
+ "loss": 0.0013,
+ "step": 11000
+ },
+ {
+ "epoch": 6.292710706150341,
+ "grad_norm": 0.015608682297170162,
+ "learning_rate": 2.033843150016271e-06,
+ "loss": 0.0021,
+ "step": 11050
+ },
+ {
+ "epoch": 6.3211845102505695,
+ "grad_norm": 0.004031027667224407,
+ "learning_rate": 1.95248942401562e-06,
+ "loss": 0.0011,
+ "step": 11100
+ },
+ {
+ "epoch": 6.349658314350798,
+ "grad_norm": 0.008949857205152512,
+ "learning_rate": 1.8711356980149693e-06,
+ "loss": 0.0049,
+ "step": 11150
+ },
+ {
+ "epoch": 6.378132118451025,
+ "grad_norm": 0.018670039251446724,
+ "learning_rate": 1.7897819720143183e-06,
+ "loss": 0.003,
+ "step": 11200
+ },
+ {
+ "epoch": 6.406605922551253,
+ "grad_norm": 0.032393742352724075,
+ "learning_rate": 1.7084282460136675e-06,
+ "loss": 0.0051,
+ "step": 11250
+ },
+ {
+ "epoch": 6.43507972665148,
+ "grad_norm": 0.11851054430007935,
+ "learning_rate": 1.6270745200130167e-06,
+ "loss": 0.0046,
+ "step": 11300
+ },
+ {
+ "epoch": 6.4635535307517085,
+ "grad_norm": 0.09247086197137833,
+ "learning_rate": 1.5457207940123657e-06,
+ "loss": 0.0049,
+ "step": 11350
+ },
+ {
+ "epoch": 6.492027334851937,
+ "grad_norm": 0.002555207349359989,
+ "learning_rate": 1.464367068011715e-06,
+ "loss": 0.0026,
+ "step": 11400
+ },
+ {
+ "epoch": 6.520501138952164,
+ "grad_norm": 3.7294840812683105,
+ "learning_rate": 1.3830133420110642e-06,
+ "loss": 0.0014,
+ "step": 11450
+ },
+ {
+ "epoch": 6.548974943052392,
+ "grad_norm": 0.00703430688008666,
+ "learning_rate": 1.3016596160104134e-06,
+ "loss": 0.0035,
+ "step": 11500
+ },
+ {
+ "epoch": 6.577448747152619,
+ "grad_norm": 12.061240196228027,
+ "learning_rate": 1.2203058900097624e-06,
+ "loss": 0.0078,
+ "step": 11550
+ },
+ {
+ "epoch": 6.605922551252847,
+ "grad_norm": 0.011964640580117702,
+ "learning_rate": 1.1389521640091117e-06,
+ "loss": 0.0024,
+ "step": 11600
+ },
+ {
+ "epoch": 6.634396355353076,
+ "grad_norm": 0.014603933319449425,
+ "learning_rate": 1.057598438008461e-06,
+ "loss": 0.0072,
+ "step": 11650
+ },
+ {
+ "epoch": 6.662870159453303,
+ "grad_norm": 0.026642296463251114,
+ "learning_rate": 9.7624471200781e-07,
+ "loss": 0.0007,
+ "step": 11700
+ },
+ {
+ "epoch": 6.691343963553531,
+ "grad_norm": 3.3646230697631836,
+ "learning_rate": 8.948909860071591e-07,
+ "loss": 0.0046,
+ "step": 11750
+ },
+ {
+ "epoch": 6.719817767653758,
+ "grad_norm": 0.027053840458393097,
+ "learning_rate": 8.135372600065084e-07,
+ "loss": 0.0031,
+ "step": 11800
+ },
+ {
+ "epoch": 6.748291571753986,
+ "grad_norm": 2.6146697998046875,
+ "learning_rate": 7.321835340058575e-07,
+ "loss": 0.003,
+ "step": 11850
+ },
+ {
+ "epoch": 6.776765375854215,
+ "grad_norm": 8.964512825012207,
+ "learning_rate": 6.508298080052067e-07,
+ "loss": 0.0069,
+ "step": 11900
+ },
+ {
+ "epoch": 6.805239179954442,
+ "grad_norm": 0.3086203932762146,
+ "learning_rate": 5.694760820045558e-07,
+ "loss": 0.0035,
+ "step": 11950
+ },
+ {
+ "epoch": 6.83371298405467,
+ "grad_norm": 0.018831729888916016,
+ "learning_rate": 4.88122356003905e-07,
+ "loss": 0.0033,
+ "step": 12000
+ },
+ {
+ "epoch": 6.862186788154897,
+ "grad_norm": 0.0018483272287994623,
+ "learning_rate": 4.067686300032542e-07,
+ "loss": 0.0022,
+ "step": 12050
+ },
+ {
+ "epoch": 6.890660592255125,
+ "grad_norm": 0.003242627950385213,
+ "learning_rate": 3.2541490400260336e-07,
+ "loss": 0.0043,
+ "step": 12100
+ },
+ {
+ "epoch": 6.9191343963553535,
+ "grad_norm": 0.002707740291953087,
+ "learning_rate": 2.440611780019525e-07,
+ "loss": 0.0017,
+ "step": 12150
+ },
+ {
+ "epoch": 6.947608200455581,
+ "grad_norm": 0.21824024617671967,
+ "learning_rate": 1.6270745200130168e-07,
+ "loss": 0.0069,
+ "step": 12200
+ },
+ {
+ "epoch": 6.976082004555809,
+ "grad_norm": 0.008130647242069244,
+ "learning_rate": 8.135372600065084e-08,
+ "loss": 0.0047,
+ "step": 12250
+ },
+ {
+ "epoch": 7.0,
+ "eval_f1": 0.9563834240267894,
+ "eval_loss": 0.052909377962350845,
+ "eval_precision": 0.9515242378810594,
+ "eval_recall": 0.9612924941097274,
+ "eval_runtime": 3.7426,
+ "eval_samples_per_second": 868.392,
+ "eval_steps_per_second": 108.749,
+ "step": 12292
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2418024281348142.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-12292/training_args.bin b/checkpoint-12292/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-12292/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/checkpoint-1756/config.json b/checkpoint-1756/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-1756/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-1756/model.safetensors b/checkpoint-1756/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f121f2424aa7a655a0871d95aaa5fad0f5d8d335
--- /dev/null
+++ b/checkpoint-1756/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b877b1e180a68933eb90e92f3711348a19892e9e1c0e85f1c0c0ac31c900843d
+size 1109863956
diff --git a/checkpoint-1756/optimizer.pt b/checkpoint-1756/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..45d234243e949de2365398c42097e7367d2a74a6
--- /dev/null
+++ b/checkpoint-1756/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e2787580e6c0160af6a4dc99a4e9ce84ae377ab3d6296a6416012dc8df5bfbdf
+size 2219846202
diff --git a/checkpoint-1756/rng_state.pth b/checkpoint-1756/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2e4a785ea487beecc8e27e32e602c17c6df96196
--- /dev/null
+++ b/checkpoint-1756/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8c3bcda270a9bbdc01c80b4119b83fd5b30aca083ec25b4059eed0da02a64608
+size 14244
diff --git a/checkpoint-1756/scaler.pt b/checkpoint-1756/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..34c11e9fa5110e65e1f00e540985dc9a22ea37ae
--- /dev/null
+++ b/checkpoint-1756/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d81f85e057ecf3b4e4481a5b7b37c25917c66b4c3f8f132282d8a277499dbbb3
+size 988
diff --git a/checkpoint-1756/scheduler.pt b/checkpoint-1756/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..80b73ee68a100ec723df1b6bbbe216ea6c9f871e
--- /dev/null
+++ b/checkpoint-1756/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9284fedcb9e147844c4eb634bfce0b54551832557512f611ba8e58c50ee1d23f
+size 1064
diff --git a/checkpoint-1756/special_tokens_map.json b/checkpoint-1756/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-1756/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-1756/tokenizer.json b/checkpoint-1756/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-1756/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-1756/tokenizer_config.json b/checkpoint-1756/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-1756/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-1756/trainer_state.json b/checkpoint-1756/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..51d204ab54aff0dca513bb07660dbc448980a364
--- /dev/null
+++ b/checkpoint-1756/trainer_state.json
@@ -0,0 +1,290 @@
+{
+ "best_global_step": 1756,
+ "best_metric": 0.9242246437552389,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-1756",
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 1756,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 344842189594686.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-1756/training_args.bin b/checkpoint-1756/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-1756/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/checkpoint-3512/config.json b/checkpoint-3512/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-3512/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-3512/model.safetensors b/checkpoint-3512/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..313a8e9862204486b6fea2181fe31337e42c5e75
--- /dev/null
+++ b/checkpoint-3512/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ddbcca6c4d038a5447fa9bf8ac31d433d751923346e36fedb24008cd64521f50
+size 1109863956
diff --git a/checkpoint-3512/optimizer.pt b/checkpoint-3512/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e8d01dba71e58c0890bd4580995e74344c17db9f
--- /dev/null
+++ b/checkpoint-3512/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f038c2d0551d74e2318c48245eab9294694194b187f0a6c8635336b32a4a3845
+size 2219846202
diff --git a/checkpoint-3512/rng_state.pth b/checkpoint-3512/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..93182b1d7b742d2ace5e6d0ef54bed007f67ce0b
--- /dev/null
+++ b/checkpoint-3512/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0da10cefa2cb8a2be8114fc291a5c55cb96e1047f75b823fc7fb82422f79ce2f
+size 14244
diff --git a/checkpoint-3512/scaler.pt b/checkpoint-3512/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7fb15881a7d7c21f1415df4343a484353802a2b0
--- /dev/null
+++ b/checkpoint-3512/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:34bcbd0d9500d8ef0ef7916a5d9e57b44de8cfee243837f363bfae8f0865c9c6
+size 988
diff --git a/checkpoint-3512/scheduler.pt b/checkpoint-3512/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..984b272c57f991e685ebad10e1bb86ee8f11287f
--- /dev/null
+++ b/checkpoint-3512/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cbe09b89711b4ddfda2986487d23873a7e0ee48f39bf17155748cec806279077
+size 1064
diff --git a/checkpoint-3512/special_tokens_map.json b/checkpoint-3512/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-3512/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-3512/tokenizer.json b/checkpoint-3512/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-3512/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-3512/tokenizer_config.json b/checkpoint-3512/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-3512/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-3512/trainer_state.json b/checkpoint-3512/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..2fedb337bdaacc0efd9713b1d6dc2389907d8e06
--- /dev/null
+++ b/checkpoint-3512/trainer_state.json
@@ -0,0 +1,546 @@
+{
+ "best_global_step": 3512,
+ "best_metric": 0.9406438631790744,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-3512",
+ "epoch": 2.0,
+ "eval_steps": 500,
+ "global_step": 3512,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ },
+ {
+ "epoch": 1.0250569476082005,
+ "grad_norm": 4.280463695526123,
+ "learning_rate": 1.708102831109665e-05,
+ "loss": 0.0673,
+ "step": 1800
+ },
+ {
+ "epoch": 1.0535307517084282,
+ "grad_norm": 0.06692535430192947,
+ "learning_rate": 1.6999674585096e-05,
+ "loss": 0.0403,
+ "step": 1850
+ },
+ {
+ "epoch": 1.082004555808656,
+ "grad_norm": 0.03789375349879265,
+ "learning_rate": 1.691832085909535e-05,
+ "loss": 0.0496,
+ "step": 1900
+ },
+ {
+ "epoch": 1.1104783599088839,
+ "grad_norm": 3.698348045349121,
+ "learning_rate": 1.6836967133094698e-05,
+ "loss": 0.0381,
+ "step": 1950
+ },
+ {
+ "epoch": 1.1389521640091116,
+ "grad_norm": 2.753032684326172,
+ "learning_rate": 1.6755613407094048e-05,
+ "loss": 0.0432,
+ "step": 2000
+ },
+ {
+ "epoch": 1.1674259681093395,
+ "grad_norm": 0.06336130946874619,
+ "learning_rate": 1.6674259681093398e-05,
+ "loss": 0.0482,
+ "step": 2050
+ },
+ {
+ "epoch": 1.1958997722095672,
+ "grad_norm": 3.9503254890441895,
+ "learning_rate": 1.6592905955092744e-05,
+ "loss": 0.0436,
+ "step": 2100
+ },
+ {
+ "epoch": 1.224373576309795,
+ "grad_norm": 0.08679741621017456,
+ "learning_rate": 1.6511552229092093e-05,
+ "loss": 0.0346,
+ "step": 2150
+ },
+ {
+ "epoch": 1.2528473804100229,
+ "grad_norm": 2.1995127201080322,
+ "learning_rate": 1.6430198503091443e-05,
+ "loss": 0.0312,
+ "step": 2200
+ },
+ {
+ "epoch": 1.2813211845102506,
+ "grad_norm": 5.7709808349609375,
+ "learning_rate": 1.634884477709079e-05,
+ "loss": 0.032,
+ "step": 2250
+ },
+ {
+ "epoch": 1.3097949886104785,
+ "grad_norm": 0.2031829059123993,
+ "learning_rate": 1.626749105109014e-05,
+ "loss": 0.053,
+ "step": 2300
+ },
+ {
+ "epoch": 1.3382687927107062,
+ "grad_norm": 4.292157173156738,
+ "learning_rate": 1.6186137325089492e-05,
+ "loss": 0.0413,
+ "step": 2350
+ },
+ {
+ "epoch": 1.366742596810934,
+ "grad_norm": 0.07821401953697205,
+ "learning_rate": 1.6104783599088842e-05,
+ "loss": 0.0304,
+ "step": 2400
+ },
+ {
+ "epoch": 1.3952164009111616,
+ "grad_norm": 0.5930206179618835,
+ "learning_rate": 1.6023429873088188e-05,
+ "loss": 0.0525,
+ "step": 2450
+ },
+ {
+ "epoch": 1.4236902050113895,
+ "grad_norm": 1.6114712953567505,
+ "learning_rate": 1.5942076147087538e-05,
+ "loss": 0.0391,
+ "step": 2500
+ },
+ {
+ "epoch": 1.4521640091116172,
+ "grad_norm": 0.28632065653800964,
+ "learning_rate": 1.5860722421086887e-05,
+ "loss": 0.0491,
+ "step": 2550
+ },
+ {
+ "epoch": 1.4806378132118452,
+ "grad_norm": 1.0017669200897217,
+ "learning_rate": 1.5779368695086237e-05,
+ "loss": 0.0395,
+ "step": 2600
+ },
+ {
+ "epoch": 1.5091116173120729,
+ "grad_norm": 0.03409272059798241,
+ "learning_rate": 1.5698014969085583e-05,
+ "loss": 0.0399,
+ "step": 2650
+ },
+ {
+ "epoch": 1.5375854214123006,
+ "grad_norm": 3.5108702182769775,
+ "learning_rate": 1.5616661243084933e-05,
+ "loss": 0.0424,
+ "step": 2700
+ },
+ {
+ "epoch": 1.5660592255125285,
+ "grad_norm": 3.5875349044799805,
+ "learning_rate": 1.5535307517084283e-05,
+ "loss": 0.0497,
+ "step": 2750
+ },
+ {
+ "epoch": 1.5945330296127562,
+ "grad_norm": 0.10498251020908356,
+ "learning_rate": 1.5453953791083632e-05,
+ "loss": 0.0461,
+ "step": 2800
+ },
+ {
+ "epoch": 1.6230068337129842,
+ "grad_norm": 7.618963718414307,
+ "learning_rate": 1.5374227139602996e-05,
+ "loss": 0.0394,
+ "step": 2850
+ },
+ {
+ "epoch": 1.6514806378132119,
+ "grad_norm": 3.3920793533325195,
+ "learning_rate": 1.5292873413602346e-05,
+ "loss": 0.0339,
+ "step": 2900
+ },
+ {
+ "epoch": 1.6799544419134396,
+ "grad_norm": 0.04736129939556122,
+ "learning_rate": 1.5211519687601692e-05,
+ "loss": 0.0358,
+ "step": 2950
+ },
+ {
+ "epoch": 1.7084282460136673,
+ "grad_norm": 0.17463508248329163,
+ "learning_rate": 1.5130165961601042e-05,
+ "loss": 0.034,
+ "step": 3000
+ },
+ {
+ "epoch": 1.7369020501138952,
+ "grad_norm": 0.058148205280303955,
+ "learning_rate": 1.5048812235600392e-05,
+ "loss": 0.0317,
+ "step": 3050
+ },
+ {
+ "epoch": 1.7653758542141231,
+ "grad_norm": 3.5262815952301025,
+ "learning_rate": 1.4967458509599741e-05,
+ "loss": 0.0477,
+ "step": 3100
+ },
+ {
+ "epoch": 1.7938496583143508,
+ "grad_norm": 2.971937894821167,
+ "learning_rate": 1.488610478359909e-05,
+ "loss": 0.0364,
+ "step": 3150
+ },
+ {
+ "epoch": 1.8223234624145785,
+ "grad_norm": 0.05181724950671196,
+ "learning_rate": 1.4804751057598439e-05,
+ "loss": 0.0394,
+ "step": 3200
+ },
+ {
+ "epoch": 1.8507972665148062,
+ "grad_norm": 0.2712370455265045,
+ "learning_rate": 1.4723397331597789e-05,
+ "loss": 0.0456,
+ "step": 3250
+ },
+ {
+ "epoch": 1.8792710706150342,
+ "grad_norm": 1.608622431755066,
+ "learning_rate": 1.4642043605597138e-05,
+ "loss": 0.05,
+ "step": 3300
+ },
+ {
+ "epoch": 1.907744874715262,
+ "grad_norm": 5.2423529624938965,
+ "learning_rate": 1.4560689879596486e-05,
+ "loss": 0.0371,
+ "step": 3350
+ },
+ {
+ "epoch": 1.9362186788154898,
+ "grad_norm": 0.04786211624741554,
+ "learning_rate": 1.4479336153595836e-05,
+ "loss": 0.0329,
+ "step": 3400
+ },
+ {
+ "epoch": 1.9646924829157175,
+ "grad_norm": 0.5264931917190552,
+ "learning_rate": 1.4397982427595185e-05,
+ "loss": 0.0379,
+ "step": 3450
+ },
+ {
+ "epoch": 1.9931662870159452,
+ "grad_norm": 0.236006498336792,
+ "learning_rate": 1.4316628701594535e-05,
+ "loss": 0.0253,
+ "step": 3500
+ },
+ {
+ "epoch": 2.0,
+ "eval_f1": 0.9406438631790744,
+ "eval_loss": 0.05223705992102623,
+ "eval_precision": 0.9371867691279653,
+ "eval_recall": 0.9441265567149109,
+ "eval_runtime": 3.7109,
+ "eval_samples_per_second": 875.791,
+ "eval_steps_per_second": 109.676,
+ "step": 3512
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 688219085312226.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-3512/training_args.bin b/checkpoint-3512/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-3512/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/checkpoint-5268/config.json b/checkpoint-5268/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-5268/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-5268/model.safetensors b/checkpoint-5268/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6278018a10dbc950921e5aff5aa6d52d55d72631
--- /dev/null
+++ b/checkpoint-5268/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6b786779ce4d72c2e34b9d4fc244b422821fb82aa9e9496830a4f8a217e8633f
+size 1109863956
diff --git a/checkpoint-5268/optimizer.pt b/checkpoint-5268/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6f47adbdea25d744f76f78fa72edb2103c8316c5
--- /dev/null
+++ b/checkpoint-5268/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:05d411b3ef7b4d64bdb57d539abb0b4f6f1d8c12b5b8383a2c868aac083b2256
+size 2219846202
diff --git a/checkpoint-5268/rng_state.pth b/checkpoint-5268/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..487bb1bad91d9f8e0bac712686bc5cb8b807c6a0
--- /dev/null
+++ b/checkpoint-5268/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:773a0f008a7607f68cb3da99962235e5d2984b9344734097cf3b25025846c9bc
+size 14244
diff --git a/checkpoint-5268/scaler.pt b/checkpoint-5268/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..380b7d46ced03a6d1881df6ad74a8dbc8579698b
--- /dev/null
+++ b/checkpoint-5268/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5a3533b1927eb7f3b5998624c4eb9b6aa2b02868d4fa466b42a5c27cecc77725
+size 988
diff --git a/checkpoint-5268/scheduler.pt b/checkpoint-5268/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..63eb71b1f0c4efc69d0ed226e84cf0db661e5459
--- /dev/null
+++ b/checkpoint-5268/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3242de081a409069272234d53e25d7d851ad8ccef68c36b808d8d26d28a55496
+size 1064
diff --git a/checkpoint-5268/special_tokens_map.json b/checkpoint-5268/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-5268/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-5268/tokenizer.json b/checkpoint-5268/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-5268/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-5268/tokenizer_config.json b/checkpoint-5268/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-5268/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-5268/trainer_state.json b/checkpoint-5268/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..55e6b1c68f49b3a0394295fbe4b375ef114f56fe
--- /dev/null
+++ b/checkpoint-5268/trainer_state.json
@@ -0,0 +1,802 @@
+{
+ "best_global_step": 5268,
+ "best_metric": 0.9470420646891236,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-5268",
+ "epoch": 3.0,
+ "eval_steps": 500,
+ "global_step": 5268,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ },
+ {
+ "epoch": 1.0250569476082005,
+ "grad_norm": 4.280463695526123,
+ "learning_rate": 1.708102831109665e-05,
+ "loss": 0.0673,
+ "step": 1800
+ },
+ {
+ "epoch": 1.0535307517084282,
+ "grad_norm": 0.06692535430192947,
+ "learning_rate": 1.6999674585096e-05,
+ "loss": 0.0403,
+ "step": 1850
+ },
+ {
+ "epoch": 1.082004555808656,
+ "grad_norm": 0.03789375349879265,
+ "learning_rate": 1.691832085909535e-05,
+ "loss": 0.0496,
+ "step": 1900
+ },
+ {
+ "epoch": 1.1104783599088839,
+ "grad_norm": 3.698348045349121,
+ "learning_rate": 1.6836967133094698e-05,
+ "loss": 0.0381,
+ "step": 1950
+ },
+ {
+ "epoch": 1.1389521640091116,
+ "grad_norm": 2.753032684326172,
+ "learning_rate": 1.6755613407094048e-05,
+ "loss": 0.0432,
+ "step": 2000
+ },
+ {
+ "epoch": 1.1674259681093395,
+ "grad_norm": 0.06336130946874619,
+ "learning_rate": 1.6674259681093398e-05,
+ "loss": 0.0482,
+ "step": 2050
+ },
+ {
+ "epoch": 1.1958997722095672,
+ "grad_norm": 3.9503254890441895,
+ "learning_rate": 1.6592905955092744e-05,
+ "loss": 0.0436,
+ "step": 2100
+ },
+ {
+ "epoch": 1.224373576309795,
+ "grad_norm": 0.08679741621017456,
+ "learning_rate": 1.6511552229092093e-05,
+ "loss": 0.0346,
+ "step": 2150
+ },
+ {
+ "epoch": 1.2528473804100229,
+ "grad_norm": 2.1995127201080322,
+ "learning_rate": 1.6430198503091443e-05,
+ "loss": 0.0312,
+ "step": 2200
+ },
+ {
+ "epoch": 1.2813211845102506,
+ "grad_norm": 5.7709808349609375,
+ "learning_rate": 1.634884477709079e-05,
+ "loss": 0.032,
+ "step": 2250
+ },
+ {
+ "epoch": 1.3097949886104785,
+ "grad_norm": 0.2031829059123993,
+ "learning_rate": 1.626749105109014e-05,
+ "loss": 0.053,
+ "step": 2300
+ },
+ {
+ "epoch": 1.3382687927107062,
+ "grad_norm": 4.292157173156738,
+ "learning_rate": 1.6186137325089492e-05,
+ "loss": 0.0413,
+ "step": 2350
+ },
+ {
+ "epoch": 1.366742596810934,
+ "grad_norm": 0.07821401953697205,
+ "learning_rate": 1.6104783599088842e-05,
+ "loss": 0.0304,
+ "step": 2400
+ },
+ {
+ "epoch": 1.3952164009111616,
+ "grad_norm": 0.5930206179618835,
+ "learning_rate": 1.6023429873088188e-05,
+ "loss": 0.0525,
+ "step": 2450
+ },
+ {
+ "epoch": 1.4236902050113895,
+ "grad_norm": 1.6114712953567505,
+ "learning_rate": 1.5942076147087538e-05,
+ "loss": 0.0391,
+ "step": 2500
+ },
+ {
+ "epoch": 1.4521640091116172,
+ "grad_norm": 0.28632065653800964,
+ "learning_rate": 1.5860722421086887e-05,
+ "loss": 0.0491,
+ "step": 2550
+ },
+ {
+ "epoch": 1.4806378132118452,
+ "grad_norm": 1.0017669200897217,
+ "learning_rate": 1.5779368695086237e-05,
+ "loss": 0.0395,
+ "step": 2600
+ },
+ {
+ "epoch": 1.5091116173120729,
+ "grad_norm": 0.03409272059798241,
+ "learning_rate": 1.5698014969085583e-05,
+ "loss": 0.0399,
+ "step": 2650
+ },
+ {
+ "epoch": 1.5375854214123006,
+ "grad_norm": 3.5108702182769775,
+ "learning_rate": 1.5616661243084933e-05,
+ "loss": 0.0424,
+ "step": 2700
+ },
+ {
+ "epoch": 1.5660592255125285,
+ "grad_norm": 3.5875349044799805,
+ "learning_rate": 1.5535307517084283e-05,
+ "loss": 0.0497,
+ "step": 2750
+ },
+ {
+ "epoch": 1.5945330296127562,
+ "grad_norm": 0.10498251020908356,
+ "learning_rate": 1.5453953791083632e-05,
+ "loss": 0.0461,
+ "step": 2800
+ },
+ {
+ "epoch": 1.6230068337129842,
+ "grad_norm": 7.618963718414307,
+ "learning_rate": 1.5374227139602996e-05,
+ "loss": 0.0394,
+ "step": 2850
+ },
+ {
+ "epoch": 1.6514806378132119,
+ "grad_norm": 3.3920793533325195,
+ "learning_rate": 1.5292873413602346e-05,
+ "loss": 0.0339,
+ "step": 2900
+ },
+ {
+ "epoch": 1.6799544419134396,
+ "grad_norm": 0.04736129939556122,
+ "learning_rate": 1.5211519687601692e-05,
+ "loss": 0.0358,
+ "step": 2950
+ },
+ {
+ "epoch": 1.7084282460136673,
+ "grad_norm": 0.17463508248329163,
+ "learning_rate": 1.5130165961601042e-05,
+ "loss": 0.034,
+ "step": 3000
+ },
+ {
+ "epoch": 1.7369020501138952,
+ "grad_norm": 0.058148205280303955,
+ "learning_rate": 1.5048812235600392e-05,
+ "loss": 0.0317,
+ "step": 3050
+ },
+ {
+ "epoch": 1.7653758542141231,
+ "grad_norm": 3.5262815952301025,
+ "learning_rate": 1.4967458509599741e-05,
+ "loss": 0.0477,
+ "step": 3100
+ },
+ {
+ "epoch": 1.7938496583143508,
+ "grad_norm": 2.971937894821167,
+ "learning_rate": 1.488610478359909e-05,
+ "loss": 0.0364,
+ "step": 3150
+ },
+ {
+ "epoch": 1.8223234624145785,
+ "grad_norm": 0.05181724950671196,
+ "learning_rate": 1.4804751057598439e-05,
+ "loss": 0.0394,
+ "step": 3200
+ },
+ {
+ "epoch": 1.8507972665148062,
+ "grad_norm": 0.2712370455265045,
+ "learning_rate": 1.4723397331597789e-05,
+ "loss": 0.0456,
+ "step": 3250
+ },
+ {
+ "epoch": 1.8792710706150342,
+ "grad_norm": 1.608622431755066,
+ "learning_rate": 1.4642043605597138e-05,
+ "loss": 0.05,
+ "step": 3300
+ },
+ {
+ "epoch": 1.907744874715262,
+ "grad_norm": 5.2423529624938965,
+ "learning_rate": 1.4560689879596486e-05,
+ "loss": 0.0371,
+ "step": 3350
+ },
+ {
+ "epoch": 1.9362186788154898,
+ "grad_norm": 0.04786211624741554,
+ "learning_rate": 1.4479336153595836e-05,
+ "loss": 0.0329,
+ "step": 3400
+ },
+ {
+ "epoch": 1.9646924829157175,
+ "grad_norm": 0.5264931917190552,
+ "learning_rate": 1.4397982427595185e-05,
+ "loss": 0.0379,
+ "step": 3450
+ },
+ {
+ "epoch": 1.9931662870159452,
+ "grad_norm": 0.236006498336792,
+ "learning_rate": 1.4316628701594535e-05,
+ "loss": 0.0253,
+ "step": 3500
+ },
+ {
+ "epoch": 2.0,
+ "eval_f1": 0.9406438631790744,
+ "eval_loss": 0.05223705992102623,
+ "eval_precision": 0.9371867691279653,
+ "eval_recall": 0.9441265567149109,
+ "eval_runtime": 3.7109,
+ "eval_samples_per_second": 875.791,
+ "eval_steps_per_second": 109.676,
+ "step": 3512
+ },
+ {
+ "epoch": 2.021640091116173,
+ "grad_norm": 4.225963115692139,
+ "learning_rate": 1.4235274975593883e-05,
+ "loss": 0.0373,
+ "step": 3550
+ },
+ {
+ "epoch": 2.050113895216401,
+ "grad_norm": 0.15388111770153046,
+ "learning_rate": 1.4153921249593233e-05,
+ "loss": 0.0258,
+ "step": 3600
+ },
+ {
+ "epoch": 2.078587699316629,
+ "grad_norm": 6.3911895751953125,
+ "learning_rate": 1.4072567523592582e-05,
+ "loss": 0.0217,
+ "step": 3650
+ },
+ {
+ "epoch": 2.1070615034168565,
+ "grad_norm": 0.03075530007481575,
+ "learning_rate": 1.3991213797591932e-05,
+ "loss": 0.019,
+ "step": 3700
+ },
+ {
+ "epoch": 2.135535307517084,
+ "grad_norm": 0.16612432897090912,
+ "learning_rate": 1.3909860071591278e-05,
+ "loss": 0.0175,
+ "step": 3750
+ },
+ {
+ "epoch": 2.164009111617312,
+ "grad_norm": 0.14543022215366364,
+ "learning_rate": 1.382850634559063e-05,
+ "loss": 0.0289,
+ "step": 3800
+ },
+ {
+ "epoch": 2.19248291571754,
+ "grad_norm": 2.9880359172821045,
+ "learning_rate": 1.374715261958998e-05,
+ "loss": 0.0208,
+ "step": 3850
+ },
+ {
+ "epoch": 2.2209567198177678,
+ "grad_norm": 9.878133773803711,
+ "learning_rate": 1.3665798893589329e-05,
+ "loss": 0.021,
+ "step": 3900
+ },
+ {
+ "epoch": 2.2494305239179955,
+ "grad_norm": 9.684176445007324,
+ "learning_rate": 1.3584445167588675e-05,
+ "loss": 0.0267,
+ "step": 3950
+ },
+ {
+ "epoch": 2.277904328018223,
+ "grad_norm": 33.15859603881836,
+ "learning_rate": 1.3503091441588025e-05,
+ "loss": 0.0295,
+ "step": 4000
+ },
+ {
+ "epoch": 2.306378132118451,
+ "grad_norm": 2.559544086456299,
+ "learning_rate": 1.3421737715587374e-05,
+ "loss": 0.0292,
+ "step": 4050
+ },
+ {
+ "epoch": 2.334851936218679,
+ "grad_norm": 0.07091552764177322,
+ "learning_rate": 1.3340383989586724e-05,
+ "loss": 0.0262,
+ "step": 4100
+ },
+ {
+ "epoch": 2.3633257403189067,
+ "grad_norm": 1.8119585514068604,
+ "learning_rate": 1.3259030263586072e-05,
+ "loss": 0.03,
+ "step": 4150
+ },
+ {
+ "epoch": 2.3917995444191344,
+ "grad_norm": 0.1476636379957199,
+ "learning_rate": 1.3177676537585422e-05,
+ "loss": 0.0196,
+ "step": 4200
+ },
+ {
+ "epoch": 2.420273348519362,
+ "grad_norm": 0.04482650011777878,
+ "learning_rate": 1.3096322811584771e-05,
+ "loss": 0.0263,
+ "step": 4250
+ },
+ {
+ "epoch": 2.44874715261959,
+ "grad_norm": 0.09201560169458389,
+ "learning_rate": 1.3014969085584121e-05,
+ "loss": 0.0172,
+ "step": 4300
+ },
+ {
+ "epoch": 2.477220956719818,
+ "grad_norm": 0.02215876244008541,
+ "learning_rate": 1.2933615359583469e-05,
+ "loss": 0.0149,
+ "step": 4350
+ },
+ {
+ "epoch": 2.5056947608200457,
+ "grad_norm": 0.6993932127952576,
+ "learning_rate": 1.2852261633582819e-05,
+ "loss": 0.0331,
+ "step": 4400
+ },
+ {
+ "epoch": 2.5341685649202734,
+ "grad_norm": 0.030453965067863464,
+ "learning_rate": 1.2770907907582168e-05,
+ "loss": 0.0307,
+ "step": 4450
+ },
+ {
+ "epoch": 2.562642369020501,
+ "grad_norm": 0.22211593389511108,
+ "learning_rate": 1.2689554181581518e-05,
+ "loss": 0.0183,
+ "step": 4500
+ },
+ {
+ "epoch": 2.591116173120729,
+ "grad_norm": 1.2842738628387451,
+ "learning_rate": 1.2608200455580866e-05,
+ "loss": 0.0226,
+ "step": 4550
+ },
+ {
+ "epoch": 2.619589977220957,
+ "grad_norm": 3.9048423767089844,
+ "learning_rate": 1.2526846729580216e-05,
+ "loss": 0.0239,
+ "step": 4600
+ },
+ {
+ "epoch": 2.6480637813211843,
+ "grad_norm": 0.9887784719467163,
+ "learning_rate": 1.2445493003579565e-05,
+ "loss": 0.0271,
+ "step": 4650
+ },
+ {
+ "epoch": 2.6765375854214124,
+ "grad_norm": 4.759728908538818,
+ "learning_rate": 1.2364139277578915e-05,
+ "loss": 0.0201,
+ "step": 4700
+ },
+ {
+ "epoch": 2.70501138952164,
+ "grad_norm": 0.2671191096305847,
+ "learning_rate": 1.2282785551578263e-05,
+ "loss": 0.0212,
+ "step": 4750
+ },
+ {
+ "epoch": 2.733485193621868,
+ "grad_norm": 0.036626849323511124,
+ "learning_rate": 1.2201431825577612e-05,
+ "loss": 0.0155,
+ "step": 4800
+ },
+ {
+ "epoch": 2.7619589977220955,
+ "grad_norm": 6.877408027648926,
+ "learning_rate": 1.2120078099576962e-05,
+ "loss": 0.0266,
+ "step": 4850
+ },
+ {
+ "epoch": 2.7904328018223232,
+ "grad_norm": 4.725896835327148,
+ "learning_rate": 1.2038724373576312e-05,
+ "loss": 0.0163,
+ "step": 4900
+ },
+ {
+ "epoch": 2.8189066059225514,
+ "grad_norm": 0.11965059489011765,
+ "learning_rate": 1.195737064757566e-05,
+ "loss": 0.0246,
+ "step": 4950
+ },
+ {
+ "epoch": 2.847380410022779,
+ "grad_norm": 0.6634079813957214,
+ "learning_rate": 1.187601692157501e-05,
+ "loss": 0.0304,
+ "step": 5000
+ },
+ {
+ "epoch": 2.875854214123007,
+ "grad_norm": 3.952694892883301,
+ "learning_rate": 1.1794663195574359e-05,
+ "loss": 0.0183,
+ "step": 5050
+ },
+ {
+ "epoch": 2.9043280182232345,
+ "grad_norm": 19.788942337036133,
+ "learning_rate": 1.1713309469573709e-05,
+ "loss": 0.0193,
+ "step": 5100
+ },
+ {
+ "epoch": 2.932801822323462,
+ "grad_norm": 4.989261627197266,
+ "learning_rate": 1.1631955743573055e-05,
+ "loss": 0.023,
+ "step": 5150
+ },
+ {
+ "epoch": 2.9612756264236904,
+ "grad_norm": 9.020853996276855,
+ "learning_rate": 1.1550602017572406e-05,
+ "loss": 0.0231,
+ "step": 5200
+ },
+ {
+ "epoch": 2.989749430523918,
+ "grad_norm": 0.9151140451431274,
+ "learning_rate": 1.1469248291571756e-05,
+ "loss": 0.0314,
+ "step": 5250
+ },
+ {
+ "epoch": 3.0,
+ "eval_f1": 0.9470420646891236,
+ "eval_loss": 0.04870199039578438,
+ "eval_precision": 0.943090787716956,
+ "eval_recall": 0.9510265903736116,
+ "eval_runtime": 3.7494,
+ "eval_samples_per_second": 866.805,
+ "eval_steps_per_second": 108.551,
+ "step": 5268
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1034617415983686.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-5268/training_args.bin b/checkpoint-5268/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-5268/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/checkpoint-7024/config.json b/checkpoint-7024/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-7024/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-7024/model.safetensors b/checkpoint-7024/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8b6cd3509a45ea182bbc7582fb59dabff55d943f
--- /dev/null
+++ b/checkpoint-7024/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f6598e075d96ee3d6f97747af06fa07f2c3911ddaed25abfb7de1fd7721141c8
+size 1109863956
diff --git a/checkpoint-7024/optimizer.pt b/checkpoint-7024/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..76c231a740cac6d8eff8f04aac1274adcc676984
--- /dev/null
+++ b/checkpoint-7024/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a117e163d94c08fd1f1f9af39bd82f4b3716f00ac1182ace0c59a87904392071
+size 2219846202
diff --git a/checkpoint-7024/rng_state.pth b/checkpoint-7024/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0b6af7a7910ac3ba6f06785c64a3a3ee803a68ea
--- /dev/null
+++ b/checkpoint-7024/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:812e0e72dade8665b357d04e5e8e4e32884d5be5e0a27b18a52cc0cb56ff2a99
+size 14244
diff --git a/checkpoint-7024/scaler.pt b/checkpoint-7024/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..369ecafb3ebf9577cec81398c91b2d5a39fe82e9
--- /dev/null
+++ b/checkpoint-7024/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d9000913a2cd4b9138ab4ca6ca075392e853b97a40a57530ed47e619f75669ea
+size 988
diff --git a/checkpoint-7024/scheduler.pt b/checkpoint-7024/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..2cac5bc5613e1621b097e0c3320bc4e2b82b52dc
--- /dev/null
+++ b/checkpoint-7024/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:35960966f24baf5111ea581dacf8142ebf55786fd138b781b8da1cd15ea40080
+size 1064
diff --git a/checkpoint-7024/special_tokens_map.json b/checkpoint-7024/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-7024/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-7024/tokenizer.json b/checkpoint-7024/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-7024/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-7024/tokenizer_config.json b/checkpoint-7024/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-7024/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-7024/trainer_state.json b/checkpoint-7024/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..26f3fc85eb672424322acb0b19f86e0cb3c15a84
--- /dev/null
+++ b/checkpoint-7024/trainer_state.json
@@ -0,0 +1,1058 @@
+{
+ "best_global_step": 7024,
+ "best_metric": 0.9525329748802823,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-7024",
+ "epoch": 4.0,
+ "eval_steps": 500,
+ "global_step": 7024,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ },
+ {
+ "epoch": 1.0250569476082005,
+ "grad_norm": 4.280463695526123,
+ "learning_rate": 1.708102831109665e-05,
+ "loss": 0.0673,
+ "step": 1800
+ },
+ {
+ "epoch": 1.0535307517084282,
+ "grad_norm": 0.06692535430192947,
+ "learning_rate": 1.6999674585096e-05,
+ "loss": 0.0403,
+ "step": 1850
+ },
+ {
+ "epoch": 1.082004555808656,
+ "grad_norm": 0.03789375349879265,
+ "learning_rate": 1.691832085909535e-05,
+ "loss": 0.0496,
+ "step": 1900
+ },
+ {
+ "epoch": 1.1104783599088839,
+ "grad_norm": 3.698348045349121,
+ "learning_rate": 1.6836967133094698e-05,
+ "loss": 0.0381,
+ "step": 1950
+ },
+ {
+ "epoch": 1.1389521640091116,
+ "grad_norm": 2.753032684326172,
+ "learning_rate": 1.6755613407094048e-05,
+ "loss": 0.0432,
+ "step": 2000
+ },
+ {
+ "epoch": 1.1674259681093395,
+ "grad_norm": 0.06336130946874619,
+ "learning_rate": 1.6674259681093398e-05,
+ "loss": 0.0482,
+ "step": 2050
+ },
+ {
+ "epoch": 1.1958997722095672,
+ "grad_norm": 3.9503254890441895,
+ "learning_rate": 1.6592905955092744e-05,
+ "loss": 0.0436,
+ "step": 2100
+ },
+ {
+ "epoch": 1.224373576309795,
+ "grad_norm": 0.08679741621017456,
+ "learning_rate": 1.6511552229092093e-05,
+ "loss": 0.0346,
+ "step": 2150
+ },
+ {
+ "epoch": 1.2528473804100229,
+ "grad_norm": 2.1995127201080322,
+ "learning_rate": 1.6430198503091443e-05,
+ "loss": 0.0312,
+ "step": 2200
+ },
+ {
+ "epoch": 1.2813211845102506,
+ "grad_norm": 5.7709808349609375,
+ "learning_rate": 1.634884477709079e-05,
+ "loss": 0.032,
+ "step": 2250
+ },
+ {
+ "epoch": 1.3097949886104785,
+ "grad_norm": 0.2031829059123993,
+ "learning_rate": 1.626749105109014e-05,
+ "loss": 0.053,
+ "step": 2300
+ },
+ {
+ "epoch": 1.3382687927107062,
+ "grad_norm": 4.292157173156738,
+ "learning_rate": 1.6186137325089492e-05,
+ "loss": 0.0413,
+ "step": 2350
+ },
+ {
+ "epoch": 1.366742596810934,
+ "grad_norm": 0.07821401953697205,
+ "learning_rate": 1.6104783599088842e-05,
+ "loss": 0.0304,
+ "step": 2400
+ },
+ {
+ "epoch": 1.3952164009111616,
+ "grad_norm": 0.5930206179618835,
+ "learning_rate": 1.6023429873088188e-05,
+ "loss": 0.0525,
+ "step": 2450
+ },
+ {
+ "epoch": 1.4236902050113895,
+ "grad_norm": 1.6114712953567505,
+ "learning_rate": 1.5942076147087538e-05,
+ "loss": 0.0391,
+ "step": 2500
+ },
+ {
+ "epoch": 1.4521640091116172,
+ "grad_norm": 0.28632065653800964,
+ "learning_rate": 1.5860722421086887e-05,
+ "loss": 0.0491,
+ "step": 2550
+ },
+ {
+ "epoch": 1.4806378132118452,
+ "grad_norm": 1.0017669200897217,
+ "learning_rate": 1.5779368695086237e-05,
+ "loss": 0.0395,
+ "step": 2600
+ },
+ {
+ "epoch": 1.5091116173120729,
+ "grad_norm": 0.03409272059798241,
+ "learning_rate": 1.5698014969085583e-05,
+ "loss": 0.0399,
+ "step": 2650
+ },
+ {
+ "epoch": 1.5375854214123006,
+ "grad_norm": 3.5108702182769775,
+ "learning_rate": 1.5616661243084933e-05,
+ "loss": 0.0424,
+ "step": 2700
+ },
+ {
+ "epoch": 1.5660592255125285,
+ "grad_norm": 3.5875349044799805,
+ "learning_rate": 1.5535307517084283e-05,
+ "loss": 0.0497,
+ "step": 2750
+ },
+ {
+ "epoch": 1.5945330296127562,
+ "grad_norm": 0.10498251020908356,
+ "learning_rate": 1.5453953791083632e-05,
+ "loss": 0.0461,
+ "step": 2800
+ },
+ {
+ "epoch": 1.6230068337129842,
+ "grad_norm": 7.618963718414307,
+ "learning_rate": 1.5374227139602996e-05,
+ "loss": 0.0394,
+ "step": 2850
+ },
+ {
+ "epoch": 1.6514806378132119,
+ "grad_norm": 3.3920793533325195,
+ "learning_rate": 1.5292873413602346e-05,
+ "loss": 0.0339,
+ "step": 2900
+ },
+ {
+ "epoch": 1.6799544419134396,
+ "grad_norm": 0.04736129939556122,
+ "learning_rate": 1.5211519687601692e-05,
+ "loss": 0.0358,
+ "step": 2950
+ },
+ {
+ "epoch": 1.7084282460136673,
+ "grad_norm": 0.17463508248329163,
+ "learning_rate": 1.5130165961601042e-05,
+ "loss": 0.034,
+ "step": 3000
+ },
+ {
+ "epoch": 1.7369020501138952,
+ "grad_norm": 0.058148205280303955,
+ "learning_rate": 1.5048812235600392e-05,
+ "loss": 0.0317,
+ "step": 3050
+ },
+ {
+ "epoch": 1.7653758542141231,
+ "grad_norm": 3.5262815952301025,
+ "learning_rate": 1.4967458509599741e-05,
+ "loss": 0.0477,
+ "step": 3100
+ },
+ {
+ "epoch": 1.7938496583143508,
+ "grad_norm": 2.971937894821167,
+ "learning_rate": 1.488610478359909e-05,
+ "loss": 0.0364,
+ "step": 3150
+ },
+ {
+ "epoch": 1.8223234624145785,
+ "grad_norm": 0.05181724950671196,
+ "learning_rate": 1.4804751057598439e-05,
+ "loss": 0.0394,
+ "step": 3200
+ },
+ {
+ "epoch": 1.8507972665148062,
+ "grad_norm": 0.2712370455265045,
+ "learning_rate": 1.4723397331597789e-05,
+ "loss": 0.0456,
+ "step": 3250
+ },
+ {
+ "epoch": 1.8792710706150342,
+ "grad_norm": 1.608622431755066,
+ "learning_rate": 1.4642043605597138e-05,
+ "loss": 0.05,
+ "step": 3300
+ },
+ {
+ "epoch": 1.907744874715262,
+ "grad_norm": 5.2423529624938965,
+ "learning_rate": 1.4560689879596486e-05,
+ "loss": 0.0371,
+ "step": 3350
+ },
+ {
+ "epoch": 1.9362186788154898,
+ "grad_norm": 0.04786211624741554,
+ "learning_rate": 1.4479336153595836e-05,
+ "loss": 0.0329,
+ "step": 3400
+ },
+ {
+ "epoch": 1.9646924829157175,
+ "grad_norm": 0.5264931917190552,
+ "learning_rate": 1.4397982427595185e-05,
+ "loss": 0.0379,
+ "step": 3450
+ },
+ {
+ "epoch": 1.9931662870159452,
+ "grad_norm": 0.236006498336792,
+ "learning_rate": 1.4316628701594535e-05,
+ "loss": 0.0253,
+ "step": 3500
+ },
+ {
+ "epoch": 2.0,
+ "eval_f1": 0.9406438631790744,
+ "eval_loss": 0.05223705992102623,
+ "eval_precision": 0.9371867691279653,
+ "eval_recall": 0.9441265567149109,
+ "eval_runtime": 3.7109,
+ "eval_samples_per_second": 875.791,
+ "eval_steps_per_second": 109.676,
+ "step": 3512
+ },
+ {
+ "epoch": 2.021640091116173,
+ "grad_norm": 4.225963115692139,
+ "learning_rate": 1.4235274975593883e-05,
+ "loss": 0.0373,
+ "step": 3550
+ },
+ {
+ "epoch": 2.050113895216401,
+ "grad_norm": 0.15388111770153046,
+ "learning_rate": 1.4153921249593233e-05,
+ "loss": 0.0258,
+ "step": 3600
+ },
+ {
+ "epoch": 2.078587699316629,
+ "grad_norm": 6.3911895751953125,
+ "learning_rate": 1.4072567523592582e-05,
+ "loss": 0.0217,
+ "step": 3650
+ },
+ {
+ "epoch": 2.1070615034168565,
+ "grad_norm": 0.03075530007481575,
+ "learning_rate": 1.3991213797591932e-05,
+ "loss": 0.019,
+ "step": 3700
+ },
+ {
+ "epoch": 2.135535307517084,
+ "grad_norm": 0.16612432897090912,
+ "learning_rate": 1.3909860071591278e-05,
+ "loss": 0.0175,
+ "step": 3750
+ },
+ {
+ "epoch": 2.164009111617312,
+ "grad_norm": 0.14543022215366364,
+ "learning_rate": 1.382850634559063e-05,
+ "loss": 0.0289,
+ "step": 3800
+ },
+ {
+ "epoch": 2.19248291571754,
+ "grad_norm": 2.9880359172821045,
+ "learning_rate": 1.374715261958998e-05,
+ "loss": 0.0208,
+ "step": 3850
+ },
+ {
+ "epoch": 2.2209567198177678,
+ "grad_norm": 9.878133773803711,
+ "learning_rate": 1.3665798893589329e-05,
+ "loss": 0.021,
+ "step": 3900
+ },
+ {
+ "epoch": 2.2494305239179955,
+ "grad_norm": 9.684176445007324,
+ "learning_rate": 1.3584445167588675e-05,
+ "loss": 0.0267,
+ "step": 3950
+ },
+ {
+ "epoch": 2.277904328018223,
+ "grad_norm": 33.15859603881836,
+ "learning_rate": 1.3503091441588025e-05,
+ "loss": 0.0295,
+ "step": 4000
+ },
+ {
+ "epoch": 2.306378132118451,
+ "grad_norm": 2.559544086456299,
+ "learning_rate": 1.3421737715587374e-05,
+ "loss": 0.0292,
+ "step": 4050
+ },
+ {
+ "epoch": 2.334851936218679,
+ "grad_norm": 0.07091552764177322,
+ "learning_rate": 1.3340383989586724e-05,
+ "loss": 0.0262,
+ "step": 4100
+ },
+ {
+ "epoch": 2.3633257403189067,
+ "grad_norm": 1.8119585514068604,
+ "learning_rate": 1.3259030263586072e-05,
+ "loss": 0.03,
+ "step": 4150
+ },
+ {
+ "epoch": 2.3917995444191344,
+ "grad_norm": 0.1476636379957199,
+ "learning_rate": 1.3177676537585422e-05,
+ "loss": 0.0196,
+ "step": 4200
+ },
+ {
+ "epoch": 2.420273348519362,
+ "grad_norm": 0.04482650011777878,
+ "learning_rate": 1.3096322811584771e-05,
+ "loss": 0.0263,
+ "step": 4250
+ },
+ {
+ "epoch": 2.44874715261959,
+ "grad_norm": 0.09201560169458389,
+ "learning_rate": 1.3014969085584121e-05,
+ "loss": 0.0172,
+ "step": 4300
+ },
+ {
+ "epoch": 2.477220956719818,
+ "grad_norm": 0.02215876244008541,
+ "learning_rate": 1.2933615359583469e-05,
+ "loss": 0.0149,
+ "step": 4350
+ },
+ {
+ "epoch": 2.5056947608200457,
+ "grad_norm": 0.6993932127952576,
+ "learning_rate": 1.2852261633582819e-05,
+ "loss": 0.0331,
+ "step": 4400
+ },
+ {
+ "epoch": 2.5341685649202734,
+ "grad_norm": 0.030453965067863464,
+ "learning_rate": 1.2770907907582168e-05,
+ "loss": 0.0307,
+ "step": 4450
+ },
+ {
+ "epoch": 2.562642369020501,
+ "grad_norm": 0.22211593389511108,
+ "learning_rate": 1.2689554181581518e-05,
+ "loss": 0.0183,
+ "step": 4500
+ },
+ {
+ "epoch": 2.591116173120729,
+ "grad_norm": 1.2842738628387451,
+ "learning_rate": 1.2608200455580866e-05,
+ "loss": 0.0226,
+ "step": 4550
+ },
+ {
+ "epoch": 2.619589977220957,
+ "grad_norm": 3.9048423767089844,
+ "learning_rate": 1.2526846729580216e-05,
+ "loss": 0.0239,
+ "step": 4600
+ },
+ {
+ "epoch": 2.6480637813211843,
+ "grad_norm": 0.9887784719467163,
+ "learning_rate": 1.2445493003579565e-05,
+ "loss": 0.0271,
+ "step": 4650
+ },
+ {
+ "epoch": 2.6765375854214124,
+ "grad_norm": 4.759728908538818,
+ "learning_rate": 1.2364139277578915e-05,
+ "loss": 0.0201,
+ "step": 4700
+ },
+ {
+ "epoch": 2.70501138952164,
+ "grad_norm": 0.2671191096305847,
+ "learning_rate": 1.2282785551578263e-05,
+ "loss": 0.0212,
+ "step": 4750
+ },
+ {
+ "epoch": 2.733485193621868,
+ "grad_norm": 0.036626849323511124,
+ "learning_rate": 1.2201431825577612e-05,
+ "loss": 0.0155,
+ "step": 4800
+ },
+ {
+ "epoch": 2.7619589977220955,
+ "grad_norm": 6.877408027648926,
+ "learning_rate": 1.2120078099576962e-05,
+ "loss": 0.0266,
+ "step": 4850
+ },
+ {
+ "epoch": 2.7904328018223232,
+ "grad_norm": 4.725896835327148,
+ "learning_rate": 1.2038724373576312e-05,
+ "loss": 0.0163,
+ "step": 4900
+ },
+ {
+ "epoch": 2.8189066059225514,
+ "grad_norm": 0.11965059489011765,
+ "learning_rate": 1.195737064757566e-05,
+ "loss": 0.0246,
+ "step": 4950
+ },
+ {
+ "epoch": 2.847380410022779,
+ "grad_norm": 0.6634079813957214,
+ "learning_rate": 1.187601692157501e-05,
+ "loss": 0.0304,
+ "step": 5000
+ },
+ {
+ "epoch": 2.875854214123007,
+ "grad_norm": 3.952694892883301,
+ "learning_rate": 1.1794663195574359e-05,
+ "loss": 0.0183,
+ "step": 5050
+ },
+ {
+ "epoch": 2.9043280182232345,
+ "grad_norm": 19.788942337036133,
+ "learning_rate": 1.1713309469573709e-05,
+ "loss": 0.0193,
+ "step": 5100
+ },
+ {
+ "epoch": 2.932801822323462,
+ "grad_norm": 4.989261627197266,
+ "learning_rate": 1.1631955743573055e-05,
+ "loss": 0.023,
+ "step": 5150
+ },
+ {
+ "epoch": 2.9612756264236904,
+ "grad_norm": 9.020853996276855,
+ "learning_rate": 1.1550602017572406e-05,
+ "loss": 0.0231,
+ "step": 5200
+ },
+ {
+ "epoch": 2.989749430523918,
+ "grad_norm": 0.9151140451431274,
+ "learning_rate": 1.1469248291571756e-05,
+ "loss": 0.0314,
+ "step": 5250
+ },
+ {
+ "epoch": 3.0,
+ "eval_f1": 0.9470420646891236,
+ "eval_loss": 0.04870199039578438,
+ "eval_precision": 0.943090787716956,
+ "eval_recall": 0.9510265903736116,
+ "eval_runtime": 3.7494,
+ "eval_samples_per_second": 866.805,
+ "eval_steps_per_second": 108.551,
+ "step": 5268
+ },
+ {
+ "epoch": 3.0182232346241458,
+ "grad_norm": 8.62992000579834,
+ "learning_rate": 1.1387894565571106e-05,
+ "loss": 0.0194,
+ "step": 5300
+ },
+ {
+ "epoch": 3.0466970387243735,
+ "grad_norm": 0.5615554451942444,
+ "learning_rate": 1.1306540839570452e-05,
+ "loss": 0.0114,
+ "step": 5350
+ },
+ {
+ "epoch": 3.075170842824601,
+ "grad_norm": 2.5735936164855957,
+ "learning_rate": 1.1225187113569801e-05,
+ "loss": 0.0163,
+ "step": 5400
+ },
+ {
+ "epoch": 3.1036446469248293,
+ "grad_norm": 6.504094123840332,
+ "learning_rate": 1.1143833387569151e-05,
+ "loss": 0.0135,
+ "step": 5450
+ },
+ {
+ "epoch": 3.132118451025057,
+ "grad_norm": 14.477537155151367,
+ "learning_rate": 1.10624796615685e-05,
+ "loss": 0.0129,
+ "step": 5500
+ },
+ {
+ "epoch": 3.1605922551252847,
+ "grad_norm": 0.04929841682314873,
+ "learning_rate": 1.0981125935567849e-05,
+ "loss": 0.0263,
+ "step": 5550
+ },
+ {
+ "epoch": 3.1890660592255125,
+ "grad_norm": 0.08875144273042679,
+ "learning_rate": 1.0899772209567198e-05,
+ "loss": 0.0134,
+ "step": 5600
+ },
+ {
+ "epoch": 3.21753986332574,
+ "grad_norm": 0.6238455772399902,
+ "learning_rate": 1.0818418483566548e-05,
+ "loss": 0.0107,
+ "step": 5650
+ },
+ {
+ "epoch": 3.2460136674259683,
+ "grad_norm": 0.004948179703205824,
+ "learning_rate": 1.0737064757565898e-05,
+ "loss": 0.0167,
+ "step": 5700
+ },
+ {
+ "epoch": 3.274487471526196,
+ "grad_norm": 0.017031218856573105,
+ "learning_rate": 1.0655711031565246e-05,
+ "loss": 0.0085,
+ "step": 5750
+ },
+ {
+ "epoch": 3.3029612756264237,
+ "grad_norm": 2.9840469360351562,
+ "learning_rate": 1.0574357305564595e-05,
+ "loss": 0.0158,
+ "step": 5800
+ },
+ {
+ "epoch": 3.3314350797266514,
+ "grad_norm": 1.7477540969848633,
+ "learning_rate": 1.0493003579563945e-05,
+ "loss": 0.0157,
+ "step": 5850
+ },
+ {
+ "epoch": 3.359908883826879,
+ "grad_norm": 0.03963543102145195,
+ "learning_rate": 1.0411649853563295e-05,
+ "loss": 0.0201,
+ "step": 5900
+ },
+ {
+ "epoch": 3.3883826879271073,
+ "grad_norm": 0.16669808328151703,
+ "learning_rate": 1.0330296127562643e-05,
+ "loss": 0.0159,
+ "step": 5950
+ },
+ {
+ "epoch": 3.416856492027335,
+ "grad_norm": 0.01953568309545517,
+ "learning_rate": 1.0248942401561992e-05,
+ "loss": 0.0117,
+ "step": 6000
+ },
+ {
+ "epoch": 3.4453302961275627,
+ "grad_norm": 0.07708246260881424,
+ "learning_rate": 1.0167588675561342e-05,
+ "loss": 0.0117,
+ "step": 6050
+ },
+ {
+ "epoch": 3.4738041002277904,
+ "grad_norm": 2.423590898513794,
+ "learning_rate": 1.0086234949560691e-05,
+ "loss": 0.0111,
+ "step": 6100
+ },
+ {
+ "epoch": 3.502277904328018,
+ "grad_norm": 0.006337775848805904,
+ "learning_rate": 1.000488122356004e-05,
+ "loss": 0.0052,
+ "step": 6150
+ },
+ {
+ "epoch": 3.5307517084282463,
+ "grad_norm": 0.01373753696680069,
+ "learning_rate": 9.923527497559389e-06,
+ "loss": 0.009,
+ "step": 6200
+ },
+ {
+ "epoch": 3.559225512528474,
+ "grad_norm": 0.11260247975587845,
+ "learning_rate": 9.842173771558739e-06,
+ "loss": 0.0208,
+ "step": 6250
+ },
+ {
+ "epoch": 3.5876993166287017,
+ "grad_norm": 1.240822196006775,
+ "learning_rate": 9.760820045558087e-06,
+ "loss": 0.0157,
+ "step": 6300
+ },
+ {
+ "epoch": 3.6161731207289294,
+ "grad_norm": 0.05271293595433235,
+ "learning_rate": 9.679466319557436e-06,
+ "loss": 0.0175,
+ "step": 6350
+ },
+ {
+ "epoch": 3.644646924829157,
+ "grad_norm": 24.546663284301758,
+ "learning_rate": 9.598112593556786e-06,
+ "loss": 0.015,
+ "step": 6400
+ },
+ {
+ "epoch": 3.6731207289293852,
+ "grad_norm": 5.120173931121826,
+ "learning_rate": 9.516758867556136e-06,
+ "loss": 0.017,
+ "step": 6450
+ },
+ {
+ "epoch": 3.7015945330296125,
+ "grad_norm": 1.1922008991241455,
+ "learning_rate": 9.435405141555484e-06,
+ "loss": 0.0194,
+ "step": 6500
+ },
+ {
+ "epoch": 3.7300683371298406,
+ "grad_norm": 0.039439987391233444,
+ "learning_rate": 9.354051415554833e-06,
+ "loss": 0.0309,
+ "step": 6550
+ },
+ {
+ "epoch": 3.7585421412300684,
+ "grad_norm": 0.04055279493331909,
+ "learning_rate": 9.272697689554181e-06,
+ "loss": 0.0195,
+ "step": 6600
+ },
+ {
+ "epoch": 3.787015945330296,
+ "grad_norm": 0.5971085429191589,
+ "learning_rate": 9.191343963553533e-06,
+ "loss": 0.0156,
+ "step": 6650
+ },
+ {
+ "epoch": 3.8154897494305238,
+ "grad_norm": 0.2191866636276245,
+ "learning_rate": 9.10999023755288e-06,
+ "loss": 0.0174,
+ "step": 6700
+ },
+ {
+ "epoch": 3.8439635535307515,
+ "grad_norm": 0.033559828996658325,
+ "learning_rate": 9.02863651155223e-06,
+ "loss": 0.0149,
+ "step": 6750
+ },
+ {
+ "epoch": 3.8724373576309796,
+ "grad_norm": 3.7128634452819824,
+ "learning_rate": 8.947282785551578e-06,
+ "loss": 0.0221,
+ "step": 6800
+ },
+ {
+ "epoch": 3.9009111617312073,
+ "grad_norm": 0.08547580987215042,
+ "learning_rate": 8.865929059550928e-06,
+ "loss": 0.0153,
+ "step": 6850
+ },
+ {
+ "epoch": 3.929384965831435,
+ "grad_norm": 0.058124080300331116,
+ "learning_rate": 8.784575333550277e-06,
+ "loss": 0.0161,
+ "step": 6900
+ },
+ {
+ "epoch": 3.9578587699316627,
+ "grad_norm": 0.015579139813780785,
+ "learning_rate": 8.703221607549627e-06,
+ "loss": 0.011,
+ "step": 6950
+ },
+ {
+ "epoch": 3.9863325740318905,
+ "grad_norm": 0.12764935195446014,
+ "learning_rate": 8.621867881548975e-06,
+ "loss": 0.0166,
+ "step": 7000
+ },
+ {
+ "epoch": 4.0,
+ "eval_f1": 0.9525329748802823,
+ "eval_loss": 0.04983380436897278,
+ "eval_precision": 0.9510149303808086,
+ "eval_recall": 0.9540558734432851,
+ "eval_runtime": 3.7495,
+ "eval_samples_per_second": 866.79,
+ "eval_steps_per_second": 108.549,
+ "step": 7024
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1381501115682372.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-7024/training_args.bin b/checkpoint-7024/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-7024/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/checkpoint-8780/config.json b/checkpoint-8780/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/checkpoint-8780/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/checkpoint-8780/model.safetensors b/checkpoint-8780/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..db41ce94478898eb99f37de35ed6a1e939782b7d
--- /dev/null
+++ b/checkpoint-8780/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d615235602767f4cfde51b1a188a54b5083d657570a6052102b84397d87dad48
+size 1109863956
diff --git a/checkpoint-8780/optimizer.pt b/checkpoint-8780/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..9f1f397648a065b1410a62541096b5ea2ebde34e
--- /dev/null
+++ b/checkpoint-8780/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4274569bea9bc0a483b31f96a00b4d3062f2158c82027a814a4ac525cdd5e55e
+size 2219846202
diff --git a/checkpoint-8780/rng_state.pth b/checkpoint-8780/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..427922bb24fe67d1a43bf9a809f882bb59b71987
--- /dev/null
+++ b/checkpoint-8780/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8d107323f4a2ea07f33aada80a6882ba4b85809248993851a21076ef56f91ddd
+size 14244
diff --git a/checkpoint-8780/scaler.pt b/checkpoint-8780/scaler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..bf937a7044c4e59d9b1e2b1257ff4802be72cc15
--- /dev/null
+++ b/checkpoint-8780/scaler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:20f0698ad3d85f8bd4cbb53cb7a94efb3a21a83c224036fe2f15ec6148d3c6c6
+size 988
diff --git a/checkpoint-8780/scheduler.pt b/checkpoint-8780/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e2deded1f5ac35b82db5e1ec771d12adbd08c1da
--- /dev/null
+++ b/checkpoint-8780/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:652940fd31d647532843bea47875013a7ecffe9de1753ce5cb1605324da74495
+size 1064
diff --git a/checkpoint-8780/special_tokens_map.json b/checkpoint-8780/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/checkpoint-8780/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-8780/tokenizer.json b/checkpoint-8780/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/checkpoint-8780/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/checkpoint-8780/tokenizer_config.json b/checkpoint-8780/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/checkpoint-8780/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-8780/trainer_state.json b/checkpoint-8780/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..2bb81a9b70edce46c7d8e0ff48208c9929d12701
--- /dev/null
+++ b/checkpoint-8780/trainer_state.json
@@ -0,0 +1,1314 @@
+{
+ "best_global_step": 8780,
+ "best_metric": 0.9533444816053511,
+ "best_model_checkpoint": "C:\\Users\\Shara\\projects\\models\\run_16-lr_2e-05-acc_1-wd_0.01-bs_8-ep_7\\checkpoint-8780",
+ "epoch": 5.0,
+ "eval_steps": 500,
+ "global_step": 8780,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.02847380410022779,
+ "grad_norm": 4.8302321434021,
+ "learning_rate": 1.992840872111943e-05,
+ "loss": 0.9994,
+ "step": 50
+ },
+ {
+ "epoch": 0.05694760820045558,
+ "grad_norm": 3.4978649616241455,
+ "learning_rate": 1.9847054995118778e-05,
+ "loss": 0.3939,
+ "step": 100
+ },
+ {
+ "epoch": 0.08542141230068337,
+ "grad_norm": 8.353672981262207,
+ "learning_rate": 1.9765701269118128e-05,
+ "loss": 0.2882,
+ "step": 150
+ },
+ {
+ "epoch": 0.11389521640091116,
+ "grad_norm": 9.822871208190918,
+ "learning_rate": 1.9684347543117477e-05,
+ "loss": 0.1728,
+ "step": 200
+ },
+ {
+ "epoch": 0.14236902050113895,
+ "grad_norm": 3.1411221027374268,
+ "learning_rate": 1.9602993817116827e-05,
+ "loss": 0.1251,
+ "step": 250
+ },
+ {
+ "epoch": 0.17084282460136674,
+ "grad_norm": 8.493424415588379,
+ "learning_rate": 1.9521640091116173e-05,
+ "loss": 0.1405,
+ "step": 300
+ },
+ {
+ "epoch": 0.19931662870159453,
+ "grad_norm": 4.757007122039795,
+ "learning_rate": 1.9440286365115523e-05,
+ "loss": 0.1431,
+ "step": 350
+ },
+ {
+ "epoch": 0.22779043280182232,
+ "grad_norm": 4.339049816131592,
+ "learning_rate": 1.9358932639114873e-05,
+ "loss": 0.0912,
+ "step": 400
+ },
+ {
+ "epoch": 0.25626423690205014,
+ "grad_norm": 14.540772438049316,
+ "learning_rate": 1.9277578913114222e-05,
+ "loss": 0.1048,
+ "step": 450
+ },
+ {
+ "epoch": 0.2847380410022779,
+ "grad_norm": 10.503073692321777,
+ "learning_rate": 1.9196225187113572e-05,
+ "loss": 0.1224,
+ "step": 500
+ },
+ {
+ "epoch": 0.3132118451025057,
+ "grad_norm": 2.738131284713745,
+ "learning_rate": 1.911487146111292e-05,
+ "loss": 0.1025,
+ "step": 550
+ },
+ {
+ "epoch": 0.3416856492027335,
+ "grad_norm": 6.681102752685547,
+ "learning_rate": 1.903351773511227e-05,
+ "loss": 0.0869,
+ "step": 600
+ },
+ {
+ "epoch": 0.3701594533029613,
+ "grad_norm": 1.4330339431762695,
+ "learning_rate": 1.895216400911162e-05,
+ "loss": 0.052,
+ "step": 650
+ },
+ {
+ "epoch": 0.39863325740318906,
+ "grad_norm": 2.581470012664795,
+ "learning_rate": 1.8870810283110967e-05,
+ "loss": 0.0808,
+ "step": 700
+ },
+ {
+ "epoch": 0.4271070615034169,
+ "grad_norm": 5.825100898742676,
+ "learning_rate": 1.8789456557110317e-05,
+ "loss": 0.0716,
+ "step": 750
+ },
+ {
+ "epoch": 0.45558086560364464,
+ "grad_norm": 1.74166738986969,
+ "learning_rate": 1.8708102831109666e-05,
+ "loss": 0.0806,
+ "step": 800
+ },
+ {
+ "epoch": 0.48405466970387245,
+ "grad_norm": 7.3097405433654785,
+ "learning_rate": 1.8626749105109016e-05,
+ "loss": 0.0666,
+ "step": 850
+ },
+ {
+ "epoch": 0.5125284738041003,
+ "grad_norm": 2.220766067504883,
+ "learning_rate": 1.8545395379108362e-05,
+ "loss": 0.0768,
+ "step": 900
+ },
+ {
+ "epoch": 0.541002277904328,
+ "grad_norm": 4.0566511154174805,
+ "learning_rate": 1.8464041653107715e-05,
+ "loss": 0.0707,
+ "step": 950
+ },
+ {
+ "epoch": 0.5694760820045558,
+ "grad_norm": 0.2833240330219269,
+ "learning_rate": 1.8382687927107065e-05,
+ "loss": 0.0858,
+ "step": 1000
+ },
+ {
+ "epoch": 0.5979498861047836,
+ "grad_norm": 21.131481170654297,
+ "learning_rate": 1.8301334201106415e-05,
+ "loss": 0.0595,
+ "step": 1050
+ },
+ {
+ "epoch": 0.6264236902050114,
+ "grad_norm": 27.655433654785156,
+ "learning_rate": 1.821998047510576e-05,
+ "loss": 0.0466,
+ "step": 1100
+ },
+ {
+ "epoch": 0.6548974943052391,
+ "grad_norm": 3.4721717834472656,
+ "learning_rate": 1.813862674910511e-05,
+ "loss": 0.0776,
+ "step": 1150
+ },
+ {
+ "epoch": 0.683371298405467,
+ "grad_norm": 2.4811441898345947,
+ "learning_rate": 1.805727302310446e-05,
+ "loss": 0.0647,
+ "step": 1200
+ },
+ {
+ "epoch": 0.7118451025056948,
+ "grad_norm": 6.169066429138184,
+ "learning_rate": 1.797591929710381e-05,
+ "loss": 0.0724,
+ "step": 1250
+ },
+ {
+ "epoch": 0.7403189066059226,
+ "grad_norm": 8.984748840332031,
+ "learning_rate": 1.7894565571103156e-05,
+ "loss": 0.0838,
+ "step": 1300
+ },
+ {
+ "epoch": 0.7687927107061503,
+ "grad_norm": 5.944505214691162,
+ "learning_rate": 1.7813211845102506e-05,
+ "loss": 0.0683,
+ "step": 1350
+ },
+ {
+ "epoch": 0.7972665148063781,
+ "grad_norm": 0.28812381625175476,
+ "learning_rate": 1.7731858119101856e-05,
+ "loss": 0.0614,
+ "step": 1400
+ },
+ {
+ "epoch": 0.8257403189066059,
+ "grad_norm": 6.176011085510254,
+ "learning_rate": 1.7650504393101205e-05,
+ "loss": 0.0615,
+ "step": 1450
+ },
+ {
+ "epoch": 0.8542141230068337,
+ "grad_norm": 2.5244405269622803,
+ "learning_rate": 1.7569150667100555e-05,
+ "loss": 0.0714,
+ "step": 1500
+ },
+ {
+ "epoch": 0.8826879271070615,
+ "grad_norm": 3.3508074283599854,
+ "learning_rate": 1.7487796941099904e-05,
+ "loss": 0.0563,
+ "step": 1550
+ },
+ {
+ "epoch": 0.9111617312072893,
+ "grad_norm": 9.980842590332031,
+ "learning_rate": 1.7406443215099254e-05,
+ "loss": 0.0647,
+ "step": 1600
+ },
+ {
+ "epoch": 0.9396355353075171,
+ "grad_norm": 3.6959774494171143,
+ "learning_rate": 1.7325089489098604e-05,
+ "loss": 0.0682,
+ "step": 1650
+ },
+ {
+ "epoch": 0.9681093394077449,
+ "grad_norm": 0.626620888710022,
+ "learning_rate": 1.724373576309795e-05,
+ "loss": 0.0499,
+ "step": 1700
+ },
+ {
+ "epoch": 0.9965831435079726,
+ "grad_norm": 7.145672798156738,
+ "learning_rate": 1.71623820370973e-05,
+ "loss": 0.0569,
+ "step": 1750
+ },
+ {
+ "epoch": 1.0,
+ "eval_f1": 0.9242246437552389,
+ "eval_loss": 0.05600914731621742,
+ "eval_precision": 0.9206746826987308,
+ "eval_recall": 0.927802086839448,
+ "eval_runtime": 3.7538,
+ "eval_samples_per_second": 865.798,
+ "eval_steps_per_second": 108.425,
+ "step": 1756
+ },
+ {
+ "epoch": 1.0250569476082005,
+ "grad_norm": 4.280463695526123,
+ "learning_rate": 1.708102831109665e-05,
+ "loss": 0.0673,
+ "step": 1800
+ },
+ {
+ "epoch": 1.0535307517084282,
+ "grad_norm": 0.06692535430192947,
+ "learning_rate": 1.6999674585096e-05,
+ "loss": 0.0403,
+ "step": 1850
+ },
+ {
+ "epoch": 1.082004555808656,
+ "grad_norm": 0.03789375349879265,
+ "learning_rate": 1.691832085909535e-05,
+ "loss": 0.0496,
+ "step": 1900
+ },
+ {
+ "epoch": 1.1104783599088839,
+ "grad_norm": 3.698348045349121,
+ "learning_rate": 1.6836967133094698e-05,
+ "loss": 0.0381,
+ "step": 1950
+ },
+ {
+ "epoch": 1.1389521640091116,
+ "grad_norm": 2.753032684326172,
+ "learning_rate": 1.6755613407094048e-05,
+ "loss": 0.0432,
+ "step": 2000
+ },
+ {
+ "epoch": 1.1674259681093395,
+ "grad_norm": 0.06336130946874619,
+ "learning_rate": 1.6674259681093398e-05,
+ "loss": 0.0482,
+ "step": 2050
+ },
+ {
+ "epoch": 1.1958997722095672,
+ "grad_norm": 3.9503254890441895,
+ "learning_rate": 1.6592905955092744e-05,
+ "loss": 0.0436,
+ "step": 2100
+ },
+ {
+ "epoch": 1.224373576309795,
+ "grad_norm": 0.08679741621017456,
+ "learning_rate": 1.6511552229092093e-05,
+ "loss": 0.0346,
+ "step": 2150
+ },
+ {
+ "epoch": 1.2528473804100229,
+ "grad_norm": 2.1995127201080322,
+ "learning_rate": 1.6430198503091443e-05,
+ "loss": 0.0312,
+ "step": 2200
+ },
+ {
+ "epoch": 1.2813211845102506,
+ "grad_norm": 5.7709808349609375,
+ "learning_rate": 1.634884477709079e-05,
+ "loss": 0.032,
+ "step": 2250
+ },
+ {
+ "epoch": 1.3097949886104785,
+ "grad_norm": 0.2031829059123993,
+ "learning_rate": 1.626749105109014e-05,
+ "loss": 0.053,
+ "step": 2300
+ },
+ {
+ "epoch": 1.3382687927107062,
+ "grad_norm": 4.292157173156738,
+ "learning_rate": 1.6186137325089492e-05,
+ "loss": 0.0413,
+ "step": 2350
+ },
+ {
+ "epoch": 1.366742596810934,
+ "grad_norm": 0.07821401953697205,
+ "learning_rate": 1.6104783599088842e-05,
+ "loss": 0.0304,
+ "step": 2400
+ },
+ {
+ "epoch": 1.3952164009111616,
+ "grad_norm": 0.5930206179618835,
+ "learning_rate": 1.6023429873088188e-05,
+ "loss": 0.0525,
+ "step": 2450
+ },
+ {
+ "epoch": 1.4236902050113895,
+ "grad_norm": 1.6114712953567505,
+ "learning_rate": 1.5942076147087538e-05,
+ "loss": 0.0391,
+ "step": 2500
+ },
+ {
+ "epoch": 1.4521640091116172,
+ "grad_norm": 0.28632065653800964,
+ "learning_rate": 1.5860722421086887e-05,
+ "loss": 0.0491,
+ "step": 2550
+ },
+ {
+ "epoch": 1.4806378132118452,
+ "grad_norm": 1.0017669200897217,
+ "learning_rate": 1.5779368695086237e-05,
+ "loss": 0.0395,
+ "step": 2600
+ },
+ {
+ "epoch": 1.5091116173120729,
+ "grad_norm": 0.03409272059798241,
+ "learning_rate": 1.5698014969085583e-05,
+ "loss": 0.0399,
+ "step": 2650
+ },
+ {
+ "epoch": 1.5375854214123006,
+ "grad_norm": 3.5108702182769775,
+ "learning_rate": 1.5616661243084933e-05,
+ "loss": 0.0424,
+ "step": 2700
+ },
+ {
+ "epoch": 1.5660592255125285,
+ "grad_norm": 3.5875349044799805,
+ "learning_rate": 1.5535307517084283e-05,
+ "loss": 0.0497,
+ "step": 2750
+ },
+ {
+ "epoch": 1.5945330296127562,
+ "grad_norm": 0.10498251020908356,
+ "learning_rate": 1.5453953791083632e-05,
+ "loss": 0.0461,
+ "step": 2800
+ },
+ {
+ "epoch": 1.6230068337129842,
+ "grad_norm": 7.618963718414307,
+ "learning_rate": 1.5374227139602996e-05,
+ "loss": 0.0394,
+ "step": 2850
+ },
+ {
+ "epoch": 1.6514806378132119,
+ "grad_norm": 3.3920793533325195,
+ "learning_rate": 1.5292873413602346e-05,
+ "loss": 0.0339,
+ "step": 2900
+ },
+ {
+ "epoch": 1.6799544419134396,
+ "grad_norm": 0.04736129939556122,
+ "learning_rate": 1.5211519687601692e-05,
+ "loss": 0.0358,
+ "step": 2950
+ },
+ {
+ "epoch": 1.7084282460136673,
+ "grad_norm": 0.17463508248329163,
+ "learning_rate": 1.5130165961601042e-05,
+ "loss": 0.034,
+ "step": 3000
+ },
+ {
+ "epoch": 1.7369020501138952,
+ "grad_norm": 0.058148205280303955,
+ "learning_rate": 1.5048812235600392e-05,
+ "loss": 0.0317,
+ "step": 3050
+ },
+ {
+ "epoch": 1.7653758542141231,
+ "grad_norm": 3.5262815952301025,
+ "learning_rate": 1.4967458509599741e-05,
+ "loss": 0.0477,
+ "step": 3100
+ },
+ {
+ "epoch": 1.7938496583143508,
+ "grad_norm": 2.971937894821167,
+ "learning_rate": 1.488610478359909e-05,
+ "loss": 0.0364,
+ "step": 3150
+ },
+ {
+ "epoch": 1.8223234624145785,
+ "grad_norm": 0.05181724950671196,
+ "learning_rate": 1.4804751057598439e-05,
+ "loss": 0.0394,
+ "step": 3200
+ },
+ {
+ "epoch": 1.8507972665148062,
+ "grad_norm": 0.2712370455265045,
+ "learning_rate": 1.4723397331597789e-05,
+ "loss": 0.0456,
+ "step": 3250
+ },
+ {
+ "epoch": 1.8792710706150342,
+ "grad_norm": 1.608622431755066,
+ "learning_rate": 1.4642043605597138e-05,
+ "loss": 0.05,
+ "step": 3300
+ },
+ {
+ "epoch": 1.907744874715262,
+ "grad_norm": 5.2423529624938965,
+ "learning_rate": 1.4560689879596486e-05,
+ "loss": 0.0371,
+ "step": 3350
+ },
+ {
+ "epoch": 1.9362186788154898,
+ "grad_norm": 0.04786211624741554,
+ "learning_rate": 1.4479336153595836e-05,
+ "loss": 0.0329,
+ "step": 3400
+ },
+ {
+ "epoch": 1.9646924829157175,
+ "grad_norm": 0.5264931917190552,
+ "learning_rate": 1.4397982427595185e-05,
+ "loss": 0.0379,
+ "step": 3450
+ },
+ {
+ "epoch": 1.9931662870159452,
+ "grad_norm": 0.236006498336792,
+ "learning_rate": 1.4316628701594535e-05,
+ "loss": 0.0253,
+ "step": 3500
+ },
+ {
+ "epoch": 2.0,
+ "eval_f1": 0.9406438631790744,
+ "eval_loss": 0.05223705992102623,
+ "eval_precision": 0.9371867691279653,
+ "eval_recall": 0.9441265567149109,
+ "eval_runtime": 3.7109,
+ "eval_samples_per_second": 875.791,
+ "eval_steps_per_second": 109.676,
+ "step": 3512
+ },
+ {
+ "epoch": 2.021640091116173,
+ "grad_norm": 4.225963115692139,
+ "learning_rate": 1.4235274975593883e-05,
+ "loss": 0.0373,
+ "step": 3550
+ },
+ {
+ "epoch": 2.050113895216401,
+ "grad_norm": 0.15388111770153046,
+ "learning_rate": 1.4153921249593233e-05,
+ "loss": 0.0258,
+ "step": 3600
+ },
+ {
+ "epoch": 2.078587699316629,
+ "grad_norm": 6.3911895751953125,
+ "learning_rate": 1.4072567523592582e-05,
+ "loss": 0.0217,
+ "step": 3650
+ },
+ {
+ "epoch": 2.1070615034168565,
+ "grad_norm": 0.03075530007481575,
+ "learning_rate": 1.3991213797591932e-05,
+ "loss": 0.019,
+ "step": 3700
+ },
+ {
+ "epoch": 2.135535307517084,
+ "grad_norm": 0.16612432897090912,
+ "learning_rate": 1.3909860071591278e-05,
+ "loss": 0.0175,
+ "step": 3750
+ },
+ {
+ "epoch": 2.164009111617312,
+ "grad_norm": 0.14543022215366364,
+ "learning_rate": 1.382850634559063e-05,
+ "loss": 0.0289,
+ "step": 3800
+ },
+ {
+ "epoch": 2.19248291571754,
+ "grad_norm": 2.9880359172821045,
+ "learning_rate": 1.374715261958998e-05,
+ "loss": 0.0208,
+ "step": 3850
+ },
+ {
+ "epoch": 2.2209567198177678,
+ "grad_norm": 9.878133773803711,
+ "learning_rate": 1.3665798893589329e-05,
+ "loss": 0.021,
+ "step": 3900
+ },
+ {
+ "epoch": 2.2494305239179955,
+ "grad_norm": 9.684176445007324,
+ "learning_rate": 1.3584445167588675e-05,
+ "loss": 0.0267,
+ "step": 3950
+ },
+ {
+ "epoch": 2.277904328018223,
+ "grad_norm": 33.15859603881836,
+ "learning_rate": 1.3503091441588025e-05,
+ "loss": 0.0295,
+ "step": 4000
+ },
+ {
+ "epoch": 2.306378132118451,
+ "grad_norm": 2.559544086456299,
+ "learning_rate": 1.3421737715587374e-05,
+ "loss": 0.0292,
+ "step": 4050
+ },
+ {
+ "epoch": 2.334851936218679,
+ "grad_norm": 0.07091552764177322,
+ "learning_rate": 1.3340383989586724e-05,
+ "loss": 0.0262,
+ "step": 4100
+ },
+ {
+ "epoch": 2.3633257403189067,
+ "grad_norm": 1.8119585514068604,
+ "learning_rate": 1.3259030263586072e-05,
+ "loss": 0.03,
+ "step": 4150
+ },
+ {
+ "epoch": 2.3917995444191344,
+ "grad_norm": 0.1476636379957199,
+ "learning_rate": 1.3177676537585422e-05,
+ "loss": 0.0196,
+ "step": 4200
+ },
+ {
+ "epoch": 2.420273348519362,
+ "grad_norm": 0.04482650011777878,
+ "learning_rate": 1.3096322811584771e-05,
+ "loss": 0.0263,
+ "step": 4250
+ },
+ {
+ "epoch": 2.44874715261959,
+ "grad_norm": 0.09201560169458389,
+ "learning_rate": 1.3014969085584121e-05,
+ "loss": 0.0172,
+ "step": 4300
+ },
+ {
+ "epoch": 2.477220956719818,
+ "grad_norm": 0.02215876244008541,
+ "learning_rate": 1.2933615359583469e-05,
+ "loss": 0.0149,
+ "step": 4350
+ },
+ {
+ "epoch": 2.5056947608200457,
+ "grad_norm": 0.6993932127952576,
+ "learning_rate": 1.2852261633582819e-05,
+ "loss": 0.0331,
+ "step": 4400
+ },
+ {
+ "epoch": 2.5341685649202734,
+ "grad_norm": 0.030453965067863464,
+ "learning_rate": 1.2770907907582168e-05,
+ "loss": 0.0307,
+ "step": 4450
+ },
+ {
+ "epoch": 2.562642369020501,
+ "grad_norm": 0.22211593389511108,
+ "learning_rate": 1.2689554181581518e-05,
+ "loss": 0.0183,
+ "step": 4500
+ },
+ {
+ "epoch": 2.591116173120729,
+ "grad_norm": 1.2842738628387451,
+ "learning_rate": 1.2608200455580866e-05,
+ "loss": 0.0226,
+ "step": 4550
+ },
+ {
+ "epoch": 2.619589977220957,
+ "grad_norm": 3.9048423767089844,
+ "learning_rate": 1.2526846729580216e-05,
+ "loss": 0.0239,
+ "step": 4600
+ },
+ {
+ "epoch": 2.6480637813211843,
+ "grad_norm": 0.9887784719467163,
+ "learning_rate": 1.2445493003579565e-05,
+ "loss": 0.0271,
+ "step": 4650
+ },
+ {
+ "epoch": 2.6765375854214124,
+ "grad_norm": 4.759728908538818,
+ "learning_rate": 1.2364139277578915e-05,
+ "loss": 0.0201,
+ "step": 4700
+ },
+ {
+ "epoch": 2.70501138952164,
+ "grad_norm": 0.2671191096305847,
+ "learning_rate": 1.2282785551578263e-05,
+ "loss": 0.0212,
+ "step": 4750
+ },
+ {
+ "epoch": 2.733485193621868,
+ "grad_norm": 0.036626849323511124,
+ "learning_rate": 1.2201431825577612e-05,
+ "loss": 0.0155,
+ "step": 4800
+ },
+ {
+ "epoch": 2.7619589977220955,
+ "grad_norm": 6.877408027648926,
+ "learning_rate": 1.2120078099576962e-05,
+ "loss": 0.0266,
+ "step": 4850
+ },
+ {
+ "epoch": 2.7904328018223232,
+ "grad_norm": 4.725896835327148,
+ "learning_rate": 1.2038724373576312e-05,
+ "loss": 0.0163,
+ "step": 4900
+ },
+ {
+ "epoch": 2.8189066059225514,
+ "grad_norm": 0.11965059489011765,
+ "learning_rate": 1.195737064757566e-05,
+ "loss": 0.0246,
+ "step": 4950
+ },
+ {
+ "epoch": 2.847380410022779,
+ "grad_norm": 0.6634079813957214,
+ "learning_rate": 1.187601692157501e-05,
+ "loss": 0.0304,
+ "step": 5000
+ },
+ {
+ "epoch": 2.875854214123007,
+ "grad_norm": 3.952694892883301,
+ "learning_rate": 1.1794663195574359e-05,
+ "loss": 0.0183,
+ "step": 5050
+ },
+ {
+ "epoch": 2.9043280182232345,
+ "grad_norm": 19.788942337036133,
+ "learning_rate": 1.1713309469573709e-05,
+ "loss": 0.0193,
+ "step": 5100
+ },
+ {
+ "epoch": 2.932801822323462,
+ "grad_norm": 4.989261627197266,
+ "learning_rate": 1.1631955743573055e-05,
+ "loss": 0.023,
+ "step": 5150
+ },
+ {
+ "epoch": 2.9612756264236904,
+ "grad_norm": 9.020853996276855,
+ "learning_rate": 1.1550602017572406e-05,
+ "loss": 0.0231,
+ "step": 5200
+ },
+ {
+ "epoch": 2.989749430523918,
+ "grad_norm": 0.9151140451431274,
+ "learning_rate": 1.1469248291571756e-05,
+ "loss": 0.0314,
+ "step": 5250
+ },
+ {
+ "epoch": 3.0,
+ "eval_f1": 0.9470420646891236,
+ "eval_loss": 0.04870199039578438,
+ "eval_precision": 0.943090787716956,
+ "eval_recall": 0.9510265903736116,
+ "eval_runtime": 3.7494,
+ "eval_samples_per_second": 866.805,
+ "eval_steps_per_second": 108.551,
+ "step": 5268
+ },
+ {
+ "epoch": 3.0182232346241458,
+ "grad_norm": 8.62992000579834,
+ "learning_rate": 1.1387894565571106e-05,
+ "loss": 0.0194,
+ "step": 5300
+ },
+ {
+ "epoch": 3.0466970387243735,
+ "grad_norm": 0.5615554451942444,
+ "learning_rate": 1.1306540839570452e-05,
+ "loss": 0.0114,
+ "step": 5350
+ },
+ {
+ "epoch": 3.075170842824601,
+ "grad_norm": 2.5735936164855957,
+ "learning_rate": 1.1225187113569801e-05,
+ "loss": 0.0163,
+ "step": 5400
+ },
+ {
+ "epoch": 3.1036446469248293,
+ "grad_norm": 6.504094123840332,
+ "learning_rate": 1.1143833387569151e-05,
+ "loss": 0.0135,
+ "step": 5450
+ },
+ {
+ "epoch": 3.132118451025057,
+ "grad_norm": 14.477537155151367,
+ "learning_rate": 1.10624796615685e-05,
+ "loss": 0.0129,
+ "step": 5500
+ },
+ {
+ "epoch": 3.1605922551252847,
+ "grad_norm": 0.04929841682314873,
+ "learning_rate": 1.0981125935567849e-05,
+ "loss": 0.0263,
+ "step": 5550
+ },
+ {
+ "epoch": 3.1890660592255125,
+ "grad_norm": 0.08875144273042679,
+ "learning_rate": 1.0899772209567198e-05,
+ "loss": 0.0134,
+ "step": 5600
+ },
+ {
+ "epoch": 3.21753986332574,
+ "grad_norm": 0.6238455772399902,
+ "learning_rate": 1.0818418483566548e-05,
+ "loss": 0.0107,
+ "step": 5650
+ },
+ {
+ "epoch": 3.2460136674259683,
+ "grad_norm": 0.004948179703205824,
+ "learning_rate": 1.0737064757565898e-05,
+ "loss": 0.0167,
+ "step": 5700
+ },
+ {
+ "epoch": 3.274487471526196,
+ "grad_norm": 0.017031218856573105,
+ "learning_rate": 1.0655711031565246e-05,
+ "loss": 0.0085,
+ "step": 5750
+ },
+ {
+ "epoch": 3.3029612756264237,
+ "grad_norm": 2.9840469360351562,
+ "learning_rate": 1.0574357305564595e-05,
+ "loss": 0.0158,
+ "step": 5800
+ },
+ {
+ "epoch": 3.3314350797266514,
+ "grad_norm": 1.7477540969848633,
+ "learning_rate": 1.0493003579563945e-05,
+ "loss": 0.0157,
+ "step": 5850
+ },
+ {
+ "epoch": 3.359908883826879,
+ "grad_norm": 0.03963543102145195,
+ "learning_rate": 1.0411649853563295e-05,
+ "loss": 0.0201,
+ "step": 5900
+ },
+ {
+ "epoch": 3.3883826879271073,
+ "grad_norm": 0.16669808328151703,
+ "learning_rate": 1.0330296127562643e-05,
+ "loss": 0.0159,
+ "step": 5950
+ },
+ {
+ "epoch": 3.416856492027335,
+ "grad_norm": 0.01953568309545517,
+ "learning_rate": 1.0248942401561992e-05,
+ "loss": 0.0117,
+ "step": 6000
+ },
+ {
+ "epoch": 3.4453302961275627,
+ "grad_norm": 0.07708246260881424,
+ "learning_rate": 1.0167588675561342e-05,
+ "loss": 0.0117,
+ "step": 6050
+ },
+ {
+ "epoch": 3.4738041002277904,
+ "grad_norm": 2.423590898513794,
+ "learning_rate": 1.0086234949560691e-05,
+ "loss": 0.0111,
+ "step": 6100
+ },
+ {
+ "epoch": 3.502277904328018,
+ "grad_norm": 0.006337775848805904,
+ "learning_rate": 1.000488122356004e-05,
+ "loss": 0.0052,
+ "step": 6150
+ },
+ {
+ "epoch": 3.5307517084282463,
+ "grad_norm": 0.01373753696680069,
+ "learning_rate": 9.923527497559389e-06,
+ "loss": 0.009,
+ "step": 6200
+ },
+ {
+ "epoch": 3.559225512528474,
+ "grad_norm": 0.11260247975587845,
+ "learning_rate": 9.842173771558739e-06,
+ "loss": 0.0208,
+ "step": 6250
+ },
+ {
+ "epoch": 3.5876993166287017,
+ "grad_norm": 1.240822196006775,
+ "learning_rate": 9.760820045558087e-06,
+ "loss": 0.0157,
+ "step": 6300
+ },
+ {
+ "epoch": 3.6161731207289294,
+ "grad_norm": 0.05271293595433235,
+ "learning_rate": 9.679466319557436e-06,
+ "loss": 0.0175,
+ "step": 6350
+ },
+ {
+ "epoch": 3.644646924829157,
+ "grad_norm": 24.546663284301758,
+ "learning_rate": 9.598112593556786e-06,
+ "loss": 0.015,
+ "step": 6400
+ },
+ {
+ "epoch": 3.6731207289293852,
+ "grad_norm": 5.120173931121826,
+ "learning_rate": 9.516758867556136e-06,
+ "loss": 0.017,
+ "step": 6450
+ },
+ {
+ "epoch": 3.7015945330296125,
+ "grad_norm": 1.1922008991241455,
+ "learning_rate": 9.435405141555484e-06,
+ "loss": 0.0194,
+ "step": 6500
+ },
+ {
+ "epoch": 3.7300683371298406,
+ "grad_norm": 0.039439987391233444,
+ "learning_rate": 9.354051415554833e-06,
+ "loss": 0.0309,
+ "step": 6550
+ },
+ {
+ "epoch": 3.7585421412300684,
+ "grad_norm": 0.04055279493331909,
+ "learning_rate": 9.272697689554181e-06,
+ "loss": 0.0195,
+ "step": 6600
+ },
+ {
+ "epoch": 3.787015945330296,
+ "grad_norm": 0.5971085429191589,
+ "learning_rate": 9.191343963553533e-06,
+ "loss": 0.0156,
+ "step": 6650
+ },
+ {
+ "epoch": 3.8154897494305238,
+ "grad_norm": 0.2191866636276245,
+ "learning_rate": 9.10999023755288e-06,
+ "loss": 0.0174,
+ "step": 6700
+ },
+ {
+ "epoch": 3.8439635535307515,
+ "grad_norm": 0.033559828996658325,
+ "learning_rate": 9.02863651155223e-06,
+ "loss": 0.0149,
+ "step": 6750
+ },
+ {
+ "epoch": 3.8724373576309796,
+ "grad_norm": 3.7128634452819824,
+ "learning_rate": 8.947282785551578e-06,
+ "loss": 0.0221,
+ "step": 6800
+ },
+ {
+ "epoch": 3.9009111617312073,
+ "grad_norm": 0.08547580987215042,
+ "learning_rate": 8.865929059550928e-06,
+ "loss": 0.0153,
+ "step": 6850
+ },
+ {
+ "epoch": 3.929384965831435,
+ "grad_norm": 0.058124080300331116,
+ "learning_rate": 8.784575333550277e-06,
+ "loss": 0.0161,
+ "step": 6900
+ },
+ {
+ "epoch": 3.9578587699316627,
+ "grad_norm": 0.015579139813780785,
+ "learning_rate": 8.703221607549627e-06,
+ "loss": 0.011,
+ "step": 6950
+ },
+ {
+ "epoch": 3.9863325740318905,
+ "grad_norm": 0.12764935195446014,
+ "learning_rate": 8.621867881548975e-06,
+ "loss": 0.0166,
+ "step": 7000
+ },
+ {
+ "epoch": 4.0,
+ "eval_f1": 0.9525329748802823,
+ "eval_loss": 0.04983380436897278,
+ "eval_precision": 0.9510149303808086,
+ "eval_recall": 0.9540558734432851,
+ "eval_runtime": 3.7495,
+ "eval_samples_per_second": 866.79,
+ "eval_steps_per_second": 108.549,
+ "step": 7024
+ },
+ {
+ "epoch": 4.014806378132119,
+ "grad_norm": 0.04521024227142334,
+ "learning_rate": 8.540514155548325e-06,
+ "loss": 0.0102,
+ "step": 7050
+ },
+ {
+ "epoch": 4.043280182232346,
+ "grad_norm": 0.015535669401288033,
+ "learning_rate": 8.459160429547674e-06,
+ "loss": 0.0068,
+ "step": 7100
+ },
+ {
+ "epoch": 4.071753986332574,
+ "grad_norm": 12.566010475158691,
+ "learning_rate": 8.377806703547024e-06,
+ "loss": 0.0151,
+ "step": 7150
+ },
+ {
+ "epoch": 4.100227790432802,
+ "grad_norm": 0.027761396020650864,
+ "learning_rate": 8.296452977546372e-06,
+ "loss": 0.0053,
+ "step": 7200
+ },
+ {
+ "epoch": 4.128701594533029,
+ "grad_norm": 1.8846938610076904,
+ "learning_rate": 8.215099251545722e-06,
+ "loss": 0.0077,
+ "step": 7250
+ },
+ {
+ "epoch": 4.157175398633258,
+ "grad_norm": 0.005013479385524988,
+ "learning_rate": 8.13374552554507e-06,
+ "loss": 0.0073,
+ "step": 7300
+ },
+ {
+ "epoch": 4.185649202733485,
+ "grad_norm": 0.32696235179901123,
+ "learning_rate": 8.052391799544421e-06,
+ "loss": 0.0092,
+ "step": 7350
+ },
+ {
+ "epoch": 4.214123006833713,
+ "grad_norm": 0.014357910491526127,
+ "learning_rate": 7.971038073543769e-06,
+ "loss": 0.0164,
+ "step": 7400
+ },
+ {
+ "epoch": 4.242596810933941,
+ "grad_norm": 0.010845329612493515,
+ "learning_rate": 7.889684347543118e-06,
+ "loss": 0.0123,
+ "step": 7450
+ },
+ {
+ "epoch": 4.271070615034168,
+ "grad_norm": 0.6969354748725891,
+ "learning_rate": 7.808330621542466e-06,
+ "loss": 0.0103,
+ "step": 7500
+ },
+ {
+ "epoch": 4.2995444191343966,
+ "grad_norm": 0.3575742840766907,
+ "learning_rate": 7.726976895541816e-06,
+ "loss": 0.0161,
+ "step": 7550
+ },
+ {
+ "epoch": 4.328018223234624,
+ "grad_norm": 0.008076228201389313,
+ "learning_rate": 7.645623169541166e-06,
+ "loss": 0.0122,
+ "step": 7600
+ },
+ {
+ "epoch": 4.356492027334852,
+ "grad_norm": 0.029101597145199776,
+ "learning_rate": 7.564269443540515e-06,
+ "loss": 0.0071,
+ "step": 7650
+ },
+ {
+ "epoch": 4.38496583143508,
+ "grad_norm": 0.06841142475605011,
+ "learning_rate": 7.482915717539863e-06,
+ "loss": 0.0091,
+ "step": 7700
+ },
+ {
+ "epoch": 4.413439635535307,
+ "grad_norm": 0.04133535176515579,
+ "learning_rate": 7.401561991539213e-06,
+ "loss": 0.0073,
+ "step": 7750
+ },
+ {
+ "epoch": 4.4419134396355355,
+ "grad_norm": 0.03571132943034172,
+ "learning_rate": 7.320208265538562e-06,
+ "loss": 0.0131,
+ "step": 7800
+ },
+ {
+ "epoch": 4.470387243735763,
+ "grad_norm": 2.8286924362182617,
+ "learning_rate": 7.2388545395379114e-06,
+ "loss": 0.0043,
+ "step": 7850
+ },
+ {
+ "epoch": 4.498861047835991,
+ "grad_norm": 1.3785158395767212,
+ "learning_rate": 7.15750081353726e-06,
+ "loss": 0.0103,
+ "step": 7900
+ },
+ {
+ "epoch": 4.527334851936219,
+ "grad_norm": 0.05979786813259125,
+ "learning_rate": 7.07614708753661e-06,
+ "loss": 0.006,
+ "step": 7950
+ },
+ {
+ "epoch": 4.555808656036446,
+ "grad_norm": 1.3434393405914307,
+ "learning_rate": 6.994793361535959e-06,
+ "loss": 0.0148,
+ "step": 8000
+ },
+ {
+ "epoch": 4.5842824601366745,
+ "grad_norm": 0.019605603069067,
+ "learning_rate": 6.913439635535308e-06,
+ "loss": 0.0068,
+ "step": 8050
+ },
+ {
+ "epoch": 4.612756264236902,
+ "grad_norm": 0.0028891051188111305,
+ "learning_rate": 6.832085909534657e-06,
+ "loss": 0.0075,
+ "step": 8100
+ },
+ {
+ "epoch": 4.64123006833713,
+ "grad_norm": 0.014856363646686077,
+ "learning_rate": 6.750732183534007e-06,
+ "loss": 0.0073,
+ "step": 8150
+ },
+ {
+ "epoch": 4.669703872437358,
+ "grad_norm": 0.0037149768322706223,
+ "learning_rate": 6.669378457533356e-06,
+ "loss": 0.0066,
+ "step": 8200
+ },
+ {
+ "epoch": 4.698177676537585,
+ "grad_norm": 0.008534993045032024,
+ "learning_rate": 6.588024731532705e-06,
+ "loss": 0.0121,
+ "step": 8250
+ },
+ {
+ "epoch": 4.7266514806378135,
+ "grad_norm": 0.013920712284743786,
+ "learning_rate": 6.506671005532054e-06,
+ "loss": 0.013,
+ "step": 8300
+ },
+ {
+ "epoch": 4.755125284738041,
+ "grad_norm": 0.10121840238571167,
+ "learning_rate": 6.425317279531404e-06,
+ "loss": 0.0087,
+ "step": 8350
+ },
+ {
+ "epoch": 4.783599088838269,
+ "grad_norm": 0.07937771081924438,
+ "learning_rate": 6.343963553530752e-06,
+ "loss": 0.0086,
+ "step": 8400
+ },
+ {
+ "epoch": 4.812072892938497,
+ "grad_norm": 0.07674901187419891,
+ "learning_rate": 6.262609827530101e-06,
+ "loss": 0.0052,
+ "step": 8450
+ },
+ {
+ "epoch": 4.840546697038724,
+ "grad_norm": 0.06683178246021271,
+ "learning_rate": 6.18125610152945e-06,
+ "loss": 0.015,
+ "step": 8500
+ },
+ {
+ "epoch": 4.8690205011389525,
+ "grad_norm": 0.08861212432384491,
+ "learning_rate": 6.0999023755288e-06,
+ "loss": 0.0109,
+ "step": 8550
+ },
+ {
+ "epoch": 4.89749430523918,
+ "grad_norm": 0.006198623217642307,
+ "learning_rate": 6.018548649528149e-06,
+ "loss": 0.0181,
+ "step": 8600
+ },
+ {
+ "epoch": 4.925968109339408,
+ "grad_norm": 0.009029284119606018,
+ "learning_rate": 5.937194923527498e-06,
+ "loss": 0.0043,
+ "step": 8650
+ },
+ {
+ "epoch": 4.954441913439636,
+ "grad_norm": 1.1225602626800537,
+ "learning_rate": 5.855841197526847e-06,
+ "loss": 0.0121,
+ "step": 8700
+ },
+ {
+ "epoch": 4.982915717539863,
+ "grad_norm": 0.01218325924128294,
+ "learning_rate": 5.774487471526197e-06,
+ "loss": 0.0115,
+ "step": 8750
+ },
+ {
+ "epoch": 5.0,
+ "eval_f1": 0.9533444816053511,
+ "eval_loss": 0.04722925275564194,
+ "eval_precision": 0.9473246925889,
+ "eval_recall": 0.9594412655671491,
+ "eval_runtime": 3.7358,
+ "eval_samples_per_second": 869.968,
+ "eval_steps_per_second": 108.947,
+ "step": 8780
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 12292,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 7,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1726265217515580.0,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-8780/training_args.bin b/checkpoint-8780/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/checkpoint-8780/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/config.json b/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f744ab25e17fedaffadd5fcb3ae10a7b9e4083bc
--- /dev/null
+++ b/config.json
@@ -0,0 +1,49 @@
+{
+ "architectures": [
+ "XLMRobertaForTokenClassification"
+ ],
+ "attention_probs_dropout_prob": 0.1,
+ "bos_token_id": 0,
+ "classifier_dropout": null,
+ "eos_token_id": 2,
+ "hidden_act": "gelu",
+ "hidden_dropout_prob": 0.1,
+ "hidden_size": 768,
+ "id2label": {
+ "0": "O",
+ "1": "B-PER",
+ "2": "I-PER",
+ "3": "B-ORG",
+ "4": "I-ORG",
+ "5": "B-LOC",
+ "6": "I-LOC",
+ "7": "B-MISC",
+ "8": "I-MISC"
+ },
+ "initializer_range": 0.02,
+ "intermediate_size": 3072,
+ "label2id": {
+ "B-LOC": 5,
+ "B-MISC": 7,
+ "B-ORG": 3,
+ "B-PER": 1,
+ "I-LOC": 6,
+ "I-MISC": 8,
+ "I-ORG": 4,
+ "I-PER": 2,
+ "O": 0
+ },
+ "layer_norm_eps": 1e-05,
+ "max_position_embeddings": 514,
+ "model_type": "xlm-roberta",
+ "num_attention_heads": 12,
+ "num_hidden_layers": 12,
+ "output_past": true,
+ "pad_token_id": 1,
+ "position_embedding_type": "absolute",
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "type_vocab_size": 1,
+ "use_cache": true,
+ "vocab_size": 250002
+}
diff --git a/model.safetensors b/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9138b6f1834a9abd518339bbab61b5b8559c4607
--- /dev/null
+++ b/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9f3aa01fa5d49602728eca2190fb610b286d4348c8e470e053f83959dfd91049
+size 1109863956
diff --git a/special_tokens_map.json b/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..d79c0fe029062eb03708803e72c3483296e2ea97
--- /dev/null
+++ b/special_tokens_map.json
@@ -0,0 +1,15 @@
+{
+ "bos_token": "",
+ "cls_token": "",
+ "eos_token": "",
+ "mask_token": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "sep_token": "",
+ "unk_token": ""
+}
diff --git a/tokenizer.json b/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..7126b1d228f6e5f725b6ee02a88b1118c725aeed
--- /dev/null
+++ b/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a56def25aa40facc030ea8b0b87f3688e4b3c39eb8b45d5702b3a1300fe2a20
+size 17082734
diff --git a/tokenizer_config.json b/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..d297107d53da62483767139822cf5ed3a6975e24
--- /dev/null
+++ b/tokenizer_config.json
@@ -0,0 +1,55 @@
+{
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "3": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "250001": {
+ "content": "",
+ "lstrip": true,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "cls_token": "",
+ "eos_token": "",
+ "extra_special_tokens": {},
+ "mask_token": "",
+ "model_max_length": 512,
+ "pad_token": "",
+ "sep_token": "",
+ "tokenizer_class": "XLMRobertaTokenizer",
+ "unk_token": ""
+}
diff --git a/training_args.bin b/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5883d1e2ddb42c11dd6ad1fddc79d9ffe7a246f3
--- /dev/null
+++ b/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:15e08f60c5260f286c88c6054e532ee0950cb5316081a3a51d07b57f6a266f71
+size 5368
diff --git a/training_log.txt b/training_log.txt
new file mode 100644
index 0000000000000000000000000000000000000000..602c6d92b9b71a6668f77754a882b5cb41b8d37d
--- /dev/null
+++ b/training_log.txt
@@ -0,0 +1,18 @@
+Run 16 Hyperparameters:
+ Learning Rate: 2e-05
+ Gradient Accumulation Steps: 1
+ Weight Decay: 0.01
+ Batch Size: 8
+ Epochs: 7
+
+Evaluation Metrics:
+ Eval Loss: 0.052909377962350845
+ Eval Precision: 0.9515242378810594
+ Eval Recall: 0.9612924941097274
+ Eval F1: 0.9563834240267894
+
+Test Set Evaluation Results for Run 16:
+ Test Loss: 0.19122998416423798
+ Test Precision: 0.9052686671318911
+ Test Recall: 0.9187322946175638
+ Test F1: 0.9119507908611598