barek2k2's picture
Upload fine-tuned BERT model for HIPAA-sensitive DB schema detection
1284eab
{
"best_metric": 3.619581912062131e-05,
"best_model_checkpoint": "./bert_sensitive_columns/checkpoint-2200",
"epoch": 10.0,
"eval_steps": 500,
"global_step": 2200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.045454545454545456,
"grad_norm": 8.354640007019043,
"learning_rate": 2.9863636363636365e-05,
"loss": 0.6464,
"step": 10
},
{
"epoch": 0.09090909090909091,
"grad_norm": 4.68574857711792,
"learning_rate": 2.972727272727273e-05,
"loss": 0.6782,
"step": 20
},
{
"epoch": 0.13636363636363635,
"grad_norm": 18.99250602722168,
"learning_rate": 2.959090909090909e-05,
"loss": 0.5558,
"step": 30
},
{
"epoch": 0.18181818181818182,
"grad_norm": 6.562917709350586,
"learning_rate": 2.9454545454545456e-05,
"loss": 0.4202,
"step": 40
},
{
"epoch": 0.22727272727272727,
"grad_norm": 23.24289321899414,
"learning_rate": 2.931818181818182e-05,
"loss": 0.3803,
"step": 50
},
{
"epoch": 0.2727272727272727,
"grad_norm": 6.368680953979492,
"learning_rate": 2.9181818181818185e-05,
"loss": 0.436,
"step": 60
},
{
"epoch": 0.3181818181818182,
"grad_norm": 7.490790367126465,
"learning_rate": 2.9045454545454546e-05,
"loss": 0.359,
"step": 70
},
{
"epoch": 0.36363636363636365,
"grad_norm": 14.990336418151855,
"learning_rate": 2.890909090909091e-05,
"loss": 0.351,
"step": 80
},
{
"epoch": 0.4090909090909091,
"grad_norm": 16.408206939697266,
"learning_rate": 2.8772727272727272e-05,
"loss": 0.2892,
"step": 90
},
{
"epoch": 0.45454545454545453,
"grad_norm": 4.340272903442383,
"learning_rate": 2.8636363636363637e-05,
"loss": 0.3178,
"step": 100
},
{
"epoch": 0.5,
"grad_norm": 1.4607642889022827,
"learning_rate": 2.8499999999999998e-05,
"loss": 0.1817,
"step": 110
},
{
"epoch": 0.5454545454545454,
"grad_norm": 70.05758666992188,
"learning_rate": 2.8363636363636363e-05,
"loss": 0.1249,
"step": 120
},
{
"epoch": 0.5909090909090909,
"grad_norm": 3.1135473251342773,
"learning_rate": 2.8227272727272727e-05,
"loss": 0.2626,
"step": 130
},
{
"epoch": 0.6363636363636364,
"grad_norm": 11.004677772521973,
"learning_rate": 2.8090909090909092e-05,
"loss": 0.2237,
"step": 140
},
{
"epoch": 0.6818181818181818,
"grad_norm": 3.861924886703491,
"learning_rate": 2.7954545454545453e-05,
"loss": 0.1652,
"step": 150
},
{
"epoch": 0.7272727272727273,
"grad_norm": 0.520849883556366,
"learning_rate": 2.7818181818181818e-05,
"loss": 0.11,
"step": 160
},
{
"epoch": 0.7727272727272727,
"grad_norm": 0.3427947163581848,
"learning_rate": 2.7681818181818183e-05,
"loss": 0.0663,
"step": 170
},
{
"epoch": 0.8181818181818182,
"grad_norm": 2.3007407188415527,
"learning_rate": 2.7545454545454547e-05,
"loss": 0.302,
"step": 180
},
{
"epoch": 0.8636363636363636,
"grad_norm": 0.3552773594856262,
"learning_rate": 2.7409090909090912e-05,
"loss": 0.1245,
"step": 190
},
{
"epoch": 0.9090909090909091,
"grad_norm": 0.14678223431110382,
"learning_rate": 2.7272727272727273e-05,
"loss": 0.1086,
"step": 200
},
{
"epoch": 0.9545454545454546,
"grad_norm": 0.7700904607772827,
"learning_rate": 2.7136363636363638e-05,
"loss": 0.1189,
"step": 210
},
{
"epoch": 1.0,
"grad_norm": 1.1396052837371826,
"learning_rate": 2.7000000000000002e-05,
"loss": 0.1262,
"step": 220
},
{
"epoch": 1.0,
"eval_loss": 0.08161866664886475,
"eval_runtime": 1.4104,
"eval_samples_per_second": 622.523,
"eval_steps_per_second": 38.996,
"step": 220
},
{
"epoch": 1.0454545454545454,
"grad_norm": 5.328164577484131,
"learning_rate": 2.6863636363636367e-05,
"loss": 0.0575,
"step": 230
},
{
"epoch": 1.0909090909090908,
"grad_norm": 0.3553941547870636,
"learning_rate": 2.6727272727272728e-05,
"loss": 0.1262,
"step": 240
},
{
"epoch": 1.1363636363636362,
"grad_norm": 0.1090177446603775,
"learning_rate": 2.6590909090909093e-05,
"loss": 0.0168,
"step": 250
},
{
"epoch": 1.1818181818181819,
"grad_norm": 3.0335617065429688,
"learning_rate": 2.6454545454545454e-05,
"loss": 0.0829,
"step": 260
},
{
"epoch": 1.2272727272727273,
"grad_norm": 40.902191162109375,
"learning_rate": 2.631818181818182e-05,
"loss": 0.0295,
"step": 270
},
{
"epoch": 1.2727272727272727,
"grad_norm": 0.07629093527793884,
"learning_rate": 2.618181818181818e-05,
"loss": 0.0022,
"step": 280
},
{
"epoch": 1.3181818181818181,
"grad_norm": 9.37717056274414,
"learning_rate": 2.6045454545454545e-05,
"loss": 0.1626,
"step": 290
},
{
"epoch": 1.3636363636363638,
"grad_norm": 0.09757604449987411,
"learning_rate": 2.590909090909091e-05,
"loss": 0.0302,
"step": 300
},
{
"epoch": 1.4090909090909092,
"grad_norm": 2.172060966491699,
"learning_rate": 2.5772727272727274e-05,
"loss": 0.0301,
"step": 310
},
{
"epoch": 1.4545454545454546,
"grad_norm": 0.06304845958948135,
"learning_rate": 2.5636363636363635e-05,
"loss": 0.0411,
"step": 320
},
{
"epoch": 1.5,
"grad_norm": 0.07031694054603577,
"learning_rate": 2.55e-05,
"loss": 0.0139,
"step": 330
},
{
"epoch": 1.5454545454545454,
"grad_norm": 3.3001744747161865,
"learning_rate": 2.5363636363636364e-05,
"loss": 0.1424,
"step": 340
},
{
"epoch": 1.5909090909090908,
"grad_norm": 0.08083586394786835,
"learning_rate": 2.522727272727273e-05,
"loss": 0.0291,
"step": 350
},
{
"epoch": 1.6363636363636362,
"grad_norm": 34.89284133911133,
"learning_rate": 2.509090909090909e-05,
"loss": 0.0757,
"step": 360
},
{
"epoch": 1.6818181818181817,
"grad_norm": 0.082089863717556,
"learning_rate": 2.4954545454545455e-05,
"loss": 0.0143,
"step": 370
},
{
"epoch": 1.7272727272727273,
"grad_norm": 0.2883528172969818,
"learning_rate": 2.481818181818182e-05,
"loss": 0.0769,
"step": 380
},
{
"epoch": 1.7727272727272727,
"grad_norm": 11.552708625793457,
"learning_rate": 2.4681818181818184e-05,
"loss": 0.0487,
"step": 390
},
{
"epoch": 1.8181818181818183,
"grad_norm": 0.09678918868303299,
"learning_rate": 2.454545454545455e-05,
"loss": 0.0612,
"step": 400
},
{
"epoch": 1.8636363636363638,
"grad_norm": 0.08423396944999695,
"learning_rate": 2.440909090909091e-05,
"loss": 0.0214,
"step": 410
},
{
"epoch": 1.9090909090909092,
"grad_norm": 97.11893463134766,
"learning_rate": 2.4272727272727275e-05,
"loss": 0.1287,
"step": 420
},
{
"epoch": 1.9545454545454546,
"grad_norm": 0.04146511107683182,
"learning_rate": 2.4136363636363636e-05,
"loss": 0.0202,
"step": 430
},
{
"epoch": 2.0,
"grad_norm": 0.03929423540830612,
"learning_rate": 2.4e-05,
"loss": 0.0047,
"step": 440
},
{
"epoch": 2.0,
"eval_loss": 0.010260566137731075,
"eval_runtime": 0.9845,
"eval_samples_per_second": 891.815,
"eval_steps_per_second": 55.865,
"step": 440
},
{
"epoch": 2.0454545454545454,
"grad_norm": 0.10950633883476257,
"learning_rate": 2.3863636363636362e-05,
"loss": 0.0509,
"step": 450
},
{
"epoch": 2.090909090909091,
"grad_norm": 0.02540852129459381,
"learning_rate": 2.3727272727272726e-05,
"loss": 0.0476,
"step": 460
},
{
"epoch": 2.1363636363636362,
"grad_norm": 0.024377569556236267,
"learning_rate": 2.359090909090909e-05,
"loss": 0.0356,
"step": 470
},
{
"epoch": 2.1818181818181817,
"grad_norm": 0.025098495185375214,
"learning_rate": 2.3454545454545456e-05,
"loss": 0.0009,
"step": 480
},
{
"epoch": 2.227272727272727,
"grad_norm": 0.06944375485181808,
"learning_rate": 2.3318181818181817e-05,
"loss": 0.0193,
"step": 490
},
{
"epoch": 2.2727272727272725,
"grad_norm": 28.133596420288086,
"learning_rate": 2.318181818181818e-05,
"loss": 0.0339,
"step": 500
},
{
"epoch": 2.3181818181818183,
"grad_norm": 0.025236543267965317,
"learning_rate": 2.3045454545454546e-05,
"loss": 0.0462,
"step": 510
},
{
"epoch": 2.3636363636363638,
"grad_norm": 0.022095683962106705,
"learning_rate": 2.290909090909091e-05,
"loss": 0.0248,
"step": 520
},
{
"epoch": 2.409090909090909,
"grad_norm": 0.02965674363076687,
"learning_rate": 2.2772727272727272e-05,
"loss": 0.0023,
"step": 530
},
{
"epoch": 2.4545454545454546,
"grad_norm": 0.02505609020590782,
"learning_rate": 2.2636363636363637e-05,
"loss": 0.031,
"step": 540
},
{
"epoch": 2.5,
"grad_norm": 0.0254733357578516,
"learning_rate": 2.25e-05,
"loss": 0.0019,
"step": 550
},
{
"epoch": 2.5454545454545454,
"grad_norm": 0.018397022038698196,
"learning_rate": 2.2363636363636366e-05,
"loss": 0.0006,
"step": 560
},
{
"epoch": 2.590909090909091,
"grad_norm": 0.01716865971684456,
"learning_rate": 2.222727272727273e-05,
"loss": 0.0517,
"step": 570
},
{
"epoch": 2.6363636363636362,
"grad_norm": 0.016052110120654106,
"learning_rate": 2.2090909090909092e-05,
"loss": 0.0005,
"step": 580
},
{
"epoch": 2.6818181818181817,
"grad_norm": 0.025269588455557823,
"learning_rate": 2.1954545454545457e-05,
"loss": 0.002,
"step": 590
},
{
"epoch": 2.7272727272727275,
"grad_norm": 0.04568961635231972,
"learning_rate": 2.1818181818181818e-05,
"loss": 0.036,
"step": 600
},
{
"epoch": 2.7727272727272725,
"grad_norm": 0.01690821908414364,
"learning_rate": 2.1681818181818182e-05,
"loss": 0.0239,
"step": 610
},
{
"epoch": 2.8181818181818183,
"grad_norm": 0.02976076677441597,
"learning_rate": 2.1545454545454544e-05,
"loss": 0.0005,
"step": 620
},
{
"epoch": 2.8636363636363638,
"grad_norm": 0.15770655870437622,
"learning_rate": 2.140909090909091e-05,
"loss": 0.0006,
"step": 630
},
{
"epoch": 2.909090909090909,
"grad_norm": 0.011741632595658302,
"learning_rate": 2.1272727272727273e-05,
"loss": 0.0004,
"step": 640
},
{
"epoch": 2.9545454545454546,
"grad_norm": 0.012714399956166744,
"learning_rate": 2.1136363636363638e-05,
"loss": 0.0004,
"step": 650
},
{
"epoch": 3.0,
"grad_norm": 0.016626961529254913,
"learning_rate": 2.1e-05,
"loss": 0.0383,
"step": 660
},
{
"epoch": 3.0,
"eval_loss": 0.0007602364639751613,
"eval_runtime": 1.0081,
"eval_samples_per_second": 870.902,
"eval_steps_per_second": 54.555,
"step": 660
},
{
"epoch": 3.0454545454545454,
"grad_norm": 0.010502061806619167,
"learning_rate": 2.0863636363636363e-05,
"loss": 0.0004,
"step": 670
},
{
"epoch": 3.090909090909091,
"grad_norm": 0.011059875600039959,
"learning_rate": 2.0727272727272728e-05,
"loss": 0.0003,
"step": 680
},
{
"epoch": 3.1363636363636362,
"grad_norm": 0.013830793090164661,
"learning_rate": 2.0590909090909093e-05,
"loss": 0.0005,
"step": 690
},
{
"epoch": 3.1818181818181817,
"grad_norm": 0.016489654779434204,
"learning_rate": 2.0454545454545454e-05,
"loss": 0.0014,
"step": 700
},
{
"epoch": 3.227272727272727,
"grad_norm": 0.011767825111746788,
"learning_rate": 2.031818181818182e-05,
"loss": 0.0004,
"step": 710
},
{
"epoch": 3.2727272727272725,
"grad_norm": 0.01013511698693037,
"learning_rate": 2.0181818181818183e-05,
"loss": 0.0003,
"step": 720
},
{
"epoch": 3.3181818181818183,
"grad_norm": 0.016494890674948692,
"learning_rate": 2.0045454545454548e-05,
"loss": 0.0941,
"step": 730
},
{
"epoch": 3.3636363636363638,
"grad_norm": 0.028399920091032982,
"learning_rate": 1.9909090909090913e-05,
"loss": 0.0003,
"step": 740
},
{
"epoch": 3.409090909090909,
"grad_norm": 0.006911724805831909,
"learning_rate": 1.9772727272727274e-05,
"loss": 0.0003,
"step": 750
},
{
"epoch": 3.4545454545454546,
"grad_norm": 0.009757892228662968,
"learning_rate": 1.963636363636364e-05,
"loss": 0.0003,
"step": 760
},
{
"epoch": 3.5,
"grad_norm": 0.014759697020053864,
"learning_rate": 1.95e-05,
"loss": 0.0702,
"step": 770
},
{
"epoch": 3.5454545454545454,
"grad_norm": 0.018820617347955704,
"learning_rate": 1.9363636363636364e-05,
"loss": 0.0023,
"step": 780
},
{
"epoch": 3.590909090909091,
"grad_norm": 0.008335668593645096,
"learning_rate": 1.9227272727272726e-05,
"loss": 0.099,
"step": 790
},
{
"epoch": 3.6363636363636362,
"grad_norm": 0.007782892789691687,
"learning_rate": 1.909090909090909e-05,
"loss": 0.0005,
"step": 800
},
{
"epoch": 3.6818181818181817,
"grad_norm": 0.007866962812840939,
"learning_rate": 1.8954545454545455e-05,
"loss": 0.0007,
"step": 810
},
{
"epoch": 3.7272727272727275,
"grad_norm": 0.00743032805621624,
"learning_rate": 1.881818181818182e-05,
"loss": 0.0003,
"step": 820
},
{
"epoch": 3.7727272727272725,
"grad_norm": 0.007409967016428709,
"learning_rate": 1.868181818181818e-05,
"loss": 0.0128,
"step": 830
},
{
"epoch": 3.8181818181818183,
"grad_norm": 0.008645043708384037,
"learning_rate": 1.8545454545454545e-05,
"loss": 0.0002,
"step": 840
},
{
"epoch": 3.8636363636363638,
"grad_norm": 0.006393834948539734,
"learning_rate": 1.840909090909091e-05,
"loss": 0.0002,
"step": 850
},
{
"epoch": 3.909090909090909,
"grad_norm": 0.008199839852750301,
"learning_rate": 1.8272727272727275e-05,
"loss": 0.0003,
"step": 860
},
{
"epoch": 3.9545454545454546,
"grad_norm": 0.007581517565995455,
"learning_rate": 1.8136363636363636e-05,
"loss": 0.0002,
"step": 870
},
{
"epoch": 4.0,
"grad_norm": 0.007007090840488672,
"learning_rate": 1.8e-05,
"loss": 0.0002,
"step": 880
},
{
"epoch": 4.0,
"eval_loss": 0.00015143574273679405,
"eval_runtime": 0.9926,
"eval_samples_per_second": 884.547,
"eval_steps_per_second": 55.41,
"step": 880
},
{
"epoch": 4.045454545454546,
"grad_norm": 81.15371704101562,
"learning_rate": 1.7863636363636365e-05,
"loss": 0.006,
"step": 890
},
{
"epoch": 4.090909090909091,
"grad_norm": 0.01149928942322731,
"learning_rate": 1.772727272727273e-05,
"loss": 0.0002,
"step": 900
},
{
"epoch": 4.136363636363637,
"grad_norm": 0.0050843264907598495,
"learning_rate": 1.759090909090909e-05,
"loss": 0.0002,
"step": 910
},
{
"epoch": 4.181818181818182,
"grad_norm": 0.005241791717708111,
"learning_rate": 1.7454545454545456e-05,
"loss": 0.0002,
"step": 920
},
{
"epoch": 4.2272727272727275,
"grad_norm": 0.004743785131722689,
"learning_rate": 1.731818181818182e-05,
"loss": 0.0002,
"step": 930
},
{
"epoch": 4.2727272727272725,
"grad_norm": 0.004982436075806618,
"learning_rate": 1.718181818181818e-05,
"loss": 0.0002,
"step": 940
},
{
"epoch": 4.318181818181818,
"grad_norm": 0.005252942908555269,
"learning_rate": 1.7045454545454546e-05,
"loss": 0.0002,
"step": 950
},
{
"epoch": 4.363636363636363,
"grad_norm": 0.00490264967083931,
"learning_rate": 1.6909090909090907e-05,
"loss": 0.0002,
"step": 960
},
{
"epoch": 4.409090909090909,
"grad_norm": 0.003635741537436843,
"learning_rate": 1.6772727272727272e-05,
"loss": 0.0001,
"step": 970
},
{
"epoch": 4.454545454545454,
"grad_norm": 0.005204927641898394,
"learning_rate": 1.6636363636363637e-05,
"loss": 0.0001,
"step": 980
},
{
"epoch": 4.5,
"grad_norm": 0.009996837005019188,
"learning_rate": 1.65e-05,
"loss": 0.0002,
"step": 990
},
{
"epoch": 4.545454545454545,
"grad_norm": 0.017751624807715416,
"learning_rate": 1.6363636363636363e-05,
"loss": 0.0315,
"step": 1000
},
{
"epoch": 4.590909090909091,
"grad_norm": 0.00822280440479517,
"learning_rate": 1.6227272727272727e-05,
"loss": 0.0001,
"step": 1010
},
{
"epoch": 4.636363636363637,
"grad_norm": 0.005263584200292826,
"learning_rate": 1.6090909090909092e-05,
"loss": 0.0001,
"step": 1020
},
{
"epoch": 4.681818181818182,
"grad_norm": 0.004647717345505953,
"learning_rate": 1.5954545454545456e-05,
"loss": 0.0002,
"step": 1030
},
{
"epoch": 4.7272727272727275,
"grad_norm": 0.0030076594557613134,
"learning_rate": 1.5818181818181818e-05,
"loss": 0.0001,
"step": 1040
},
{
"epoch": 4.7727272727272725,
"grad_norm": 0.0052589308470487595,
"learning_rate": 1.5681818181818182e-05,
"loss": 0.0001,
"step": 1050
},
{
"epoch": 4.818181818181818,
"grad_norm": 0.003165638307109475,
"learning_rate": 1.5545454545454547e-05,
"loss": 0.0001,
"step": 1060
},
{
"epoch": 4.863636363636363,
"grad_norm": 0.0037285718135535717,
"learning_rate": 1.540909090909091e-05,
"loss": 0.0001,
"step": 1070
},
{
"epoch": 4.909090909090909,
"grad_norm": 0.004402661230415106,
"learning_rate": 1.5272727272727273e-05,
"loss": 0.0001,
"step": 1080
},
{
"epoch": 4.954545454545455,
"grad_norm": 0.003948619589209557,
"learning_rate": 1.5136363636363636e-05,
"loss": 0.0001,
"step": 1090
},
{
"epoch": 5.0,
"grad_norm": 0.0029597911052405834,
"learning_rate": 1.5e-05,
"loss": 0.0001,
"step": 1100
},
{
"epoch": 5.0,
"eval_loss": 8.946753951022401e-05,
"eval_runtime": 0.9926,
"eval_samples_per_second": 884.51,
"eval_steps_per_second": 55.408,
"step": 1100
},
{
"epoch": 5.045454545454546,
"grad_norm": 0.00478377053514123,
"learning_rate": 1.4863636363636365e-05,
"loss": 0.0001,
"step": 1110
},
{
"epoch": 5.090909090909091,
"grad_norm": 0.0036518580745905638,
"learning_rate": 1.4727272727272728e-05,
"loss": 0.0001,
"step": 1120
},
{
"epoch": 5.136363636363637,
"grad_norm": 0.004087444860488176,
"learning_rate": 1.4590909090909093e-05,
"loss": 0.0001,
"step": 1130
},
{
"epoch": 5.181818181818182,
"grad_norm": 0.004169765394181013,
"learning_rate": 1.4454545454545456e-05,
"loss": 0.0001,
"step": 1140
},
{
"epoch": 5.2272727272727275,
"grad_norm": 0.00350973685272038,
"learning_rate": 1.4318181818181818e-05,
"loss": 0.0001,
"step": 1150
},
{
"epoch": 5.2727272727272725,
"grad_norm": 0.0037286856677383184,
"learning_rate": 1.4181818181818181e-05,
"loss": 0.0001,
"step": 1160
},
{
"epoch": 5.318181818181818,
"grad_norm": 0.003630703780800104,
"learning_rate": 1.4045454545454546e-05,
"loss": 0.0001,
"step": 1170
},
{
"epoch": 5.363636363636363,
"grad_norm": 0.0036072884686291218,
"learning_rate": 1.3909090909090909e-05,
"loss": 0.0001,
"step": 1180
},
{
"epoch": 5.409090909090909,
"grad_norm": 0.004187653306871653,
"learning_rate": 1.3772727272727274e-05,
"loss": 0.0001,
"step": 1190
},
{
"epoch": 5.454545454545454,
"grad_norm": 0.00530035886913538,
"learning_rate": 1.3636363636363637e-05,
"loss": 0.0001,
"step": 1200
},
{
"epoch": 5.5,
"grad_norm": 0.0027139252051711082,
"learning_rate": 1.3500000000000001e-05,
"loss": 0.0001,
"step": 1210
},
{
"epoch": 5.545454545454545,
"grad_norm": 0.0034846195485442877,
"learning_rate": 1.3363636363636364e-05,
"loss": 0.0001,
"step": 1220
},
{
"epoch": 5.590909090909091,
"grad_norm": 0.003408952383324504,
"learning_rate": 1.3227272727272727e-05,
"loss": 0.0001,
"step": 1230
},
{
"epoch": 5.636363636363637,
"grad_norm": 0.0027936245314776897,
"learning_rate": 1.309090909090909e-05,
"loss": 0.0001,
"step": 1240
},
{
"epoch": 5.681818181818182,
"grad_norm": 0.002841574139893055,
"learning_rate": 1.2954545454545455e-05,
"loss": 0.0001,
"step": 1250
},
{
"epoch": 5.7272727272727275,
"grad_norm": 0.26475799083709717,
"learning_rate": 1.2818181818181818e-05,
"loss": 0.0001,
"step": 1260
},
{
"epoch": 5.7727272727272725,
"grad_norm": 0.0034859974402934313,
"learning_rate": 1.2681818181818182e-05,
"loss": 0.0001,
"step": 1270
},
{
"epoch": 5.818181818181818,
"grad_norm": 0.0028127585537731647,
"learning_rate": 1.2545454545454545e-05,
"loss": 0.0001,
"step": 1280
},
{
"epoch": 5.863636363636363,
"grad_norm": 0.002384282648563385,
"learning_rate": 1.240909090909091e-05,
"loss": 0.0001,
"step": 1290
},
{
"epoch": 5.909090909090909,
"grad_norm": 0.0030948910862207413,
"learning_rate": 1.2272727272727274e-05,
"loss": 0.0001,
"step": 1300
},
{
"epoch": 5.954545454545455,
"grad_norm": 0.002442040015012026,
"learning_rate": 1.2136363636363637e-05,
"loss": 0.0001,
"step": 1310
},
{
"epoch": 6.0,
"grad_norm": 0.0018502968596294522,
"learning_rate": 1.2e-05,
"loss": 0.0001,
"step": 1320
},
{
"epoch": 6.0,
"eval_loss": 6.235863111214712e-05,
"eval_runtime": 0.9927,
"eval_samples_per_second": 884.48,
"eval_steps_per_second": 55.406,
"step": 1320
},
{
"epoch": 6.045454545454546,
"grad_norm": 0.002548688091337681,
"learning_rate": 1.1863636363636363e-05,
"loss": 0.0001,
"step": 1330
},
{
"epoch": 6.090909090909091,
"grad_norm": 0.002689023967832327,
"learning_rate": 1.1727272727272728e-05,
"loss": 0.0001,
"step": 1340
},
{
"epoch": 6.136363636363637,
"grad_norm": 0.002400546334683895,
"learning_rate": 1.159090909090909e-05,
"loss": 0.0001,
"step": 1350
},
{
"epoch": 6.181818181818182,
"grad_norm": 0.0029753774870187044,
"learning_rate": 1.1454545454545455e-05,
"loss": 0.0001,
"step": 1360
},
{
"epoch": 6.2272727272727275,
"grad_norm": 0.0020764051005244255,
"learning_rate": 1.1318181818181818e-05,
"loss": 0.0001,
"step": 1370
},
{
"epoch": 6.2727272727272725,
"grad_norm": 0.00242880592122674,
"learning_rate": 1.1181818181818183e-05,
"loss": 0.0004,
"step": 1380
},
{
"epoch": 6.318181818181818,
"grad_norm": 0.013365722261369228,
"learning_rate": 1.1045454545454546e-05,
"loss": 0.0001,
"step": 1390
},
{
"epoch": 6.363636363636363,
"grad_norm": 0.0019247201271355152,
"learning_rate": 1.0909090909090909e-05,
"loss": 0.0001,
"step": 1400
},
{
"epoch": 6.409090909090909,
"grad_norm": 0.0019971681758761406,
"learning_rate": 1.0772727272727272e-05,
"loss": 0.0001,
"step": 1410
},
{
"epoch": 6.454545454545454,
"grad_norm": 0.002282345201820135,
"learning_rate": 1.0636363636363636e-05,
"loss": 0.0001,
"step": 1420
},
{
"epoch": 6.5,
"grad_norm": 0.0025554117746651173,
"learning_rate": 1.05e-05,
"loss": 0.0001,
"step": 1430
},
{
"epoch": 6.545454545454545,
"grad_norm": 0.003260772442445159,
"learning_rate": 1.0363636363636364e-05,
"loss": 0.0001,
"step": 1440
},
{
"epoch": 6.590909090909091,
"grad_norm": 0.0019251167541369796,
"learning_rate": 1.0227272727272727e-05,
"loss": 0.0001,
"step": 1450
},
{
"epoch": 6.636363636363637,
"grad_norm": 0.002389734610915184,
"learning_rate": 1.0090909090909092e-05,
"loss": 0.0001,
"step": 1460
},
{
"epoch": 6.681818181818182,
"grad_norm": 0.00212781666778028,
"learning_rate": 9.954545454545456e-06,
"loss": 0.0001,
"step": 1470
},
{
"epoch": 6.7272727272727275,
"grad_norm": 0.003619167488068342,
"learning_rate": 9.81818181818182e-06,
"loss": 0.0001,
"step": 1480
},
{
"epoch": 6.7727272727272725,
"grad_norm": 0.002498344052582979,
"learning_rate": 9.681818181818182e-06,
"loss": 0.0001,
"step": 1490
},
{
"epoch": 6.818181818181818,
"grad_norm": 0.0023231736849993467,
"learning_rate": 9.545454545454545e-06,
"loss": 0.0001,
"step": 1500
},
{
"epoch": 6.863636363636363,
"grad_norm": 0.0021667128894478083,
"learning_rate": 9.40909090909091e-06,
"loss": 0.003,
"step": 1510
},
{
"epoch": 6.909090909090909,
"grad_norm": 0.0022279066033661366,
"learning_rate": 9.272727272727273e-06,
"loss": 0.0001,
"step": 1520
},
{
"epoch": 6.954545454545455,
"grad_norm": 0.00177993334364146,
"learning_rate": 9.136363636363637e-06,
"loss": 0.0001,
"step": 1530
},
{
"epoch": 7.0,
"grad_norm": 0.0025669343303889036,
"learning_rate": 9e-06,
"loss": 0.0001,
"step": 1540
},
{
"epoch": 7.0,
"eval_loss": 4.8654597776476294e-05,
"eval_runtime": 0.981,
"eval_samples_per_second": 895.028,
"eval_steps_per_second": 56.067,
"step": 1540
},
{
"epoch": 7.045454545454546,
"grad_norm": 0.0023670855443924665,
"learning_rate": 8.863636363636365e-06,
"loss": 0.0001,
"step": 1550
},
{
"epoch": 7.090909090909091,
"grad_norm": 0.0023604007437825203,
"learning_rate": 8.727272727272728e-06,
"loss": 0.0001,
"step": 1560
},
{
"epoch": 7.136363636363637,
"grad_norm": 0.0018116602441295981,
"learning_rate": 8.59090909090909e-06,
"loss": 0.0001,
"step": 1570
},
{
"epoch": 7.181818181818182,
"grad_norm": 0.00249605649150908,
"learning_rate": 8.454545454545454e-06,
"loss": 0.0001,
"step": 1580
},
{
"epoch": 7.2272727272727275,
"grad_norm": 0.001895196153782308,
"learning_rate": 8.318181818181818e-06,
"loss": 0.0001,
"step": 1590
},
{
"epoch": 7.2727272727272725,
"grad_norm": 0.0017933849012479186,
"learning_rate": 8.181818181818181e-06,
"loss": 0.0429,
"step": 1600
},
{
"epoch": 7.318181818181818,
"grad_norm": 0.0016959038330242038,
"learning_rate": 8.045454545454546e-06,
"loss": 0.0001,
"step": 1610
},
{
"epoch": 7.363636363636363,
"grad_norm": 0.0016535187605768442,
"learning_rate": 7.909090909090909e-06,
"loss": 0.0001,
"step": 1620
},
{
"epoch": 7.409090909090909,
"grad_norm": 0.0020366287790238857,
"learning_rate": 7.772727272727273e-06,
"loss": 0.0001,
"step": 1630
},
{
"epoch": 7.454545454545454,
"grad_norm": 0.0017039361409842968,
"learning_rate": 7.636363636363636e-06,
"loss": 0.0001,
"step": 1640
},
{
"epoch": 7.5,
"grad_norm": 0.0018439262639731169,
"learning_rate": 7.5e-06,
"loss": 0.0001,
"step": 1650
},
{
"epoch": 7.545454545454545,
"grad_norm": 0.0017576288664713502,
"learning_rate": 7.363636363636364e-06,
"loss": 0.0001,
"step": 1660
},
{
"epoch": 7.590909090909091,
"grad_norm": 0.0015827094903215766,
"learning_rate": 7.227272727272728e-06,
"loss": 0.0001,
"step": 1670
},
{
"epoch": 7.636363636363637,
"grad_norm": 0.0021957652643322945,
"learning_rate": 7.090909090909091e-06,
"loss": 0.0001,
"step": 1680
},
{
"epoch": 7.681818181818182,
"grad_norm": 0.0018625753000378609,
"learning_rate": 6.9545454545454545e-06,
"loss": 0.0001,
"step": 1690
},
{
"epoch": 7.7272727272727275,
"grad_norm": 0.003254745388403535,
"learning_rate": 6.818181818181818e-06,
"loss": 0.0001,
"step": 1700
},
{
"epoch": 7.7727272727272725,
"grad_norm": 0.0016077288892120123,
"learning_rate": 6.681818181818182e-06,
"loss": 0.0001,
"step": 1710
},
{
"epoch": 7.818181818181818,
"grad_norm": 0.0023993789218366146,
"learning_rate": 6.545454545454545e-06,
"loss": 0.0001,
"step": 1720
},
{
"epoch": 7.863636363636363,
"grad_norm": 0.001893221982754767,
"learning_rate": 6.409090909090909e-06,
"loss": 0.0001,
"step": 1730
},
{
"epoch": 7.909090909090909,
"grad_norm": 0.0018822376150637865,
"learning_rate": 6.272727272727273e-06,
"loss": 0.0001,
"step": 1740
},
{
"epoch": 7.954545454545455,
"grad_norm": 0.009597906842827797,
"learning_rate": 6.136363636363637e-06,
"loss": 0.0001,
"step": 1750
},
{
"epoch": 8.0,
"grad_norm": 0.0014578086556866765,
"learning_rate": 6e-06,
"loss": 0.0001,
"step": 1760
},
{
"epoch": 8.0,
"eval_loss": 4.155210262979381e-05,
"eval_runtime": 1.017,
"eval_samples_per_second": 863.319,
"eval_steps_per_second": 54.08,
"step": 1760
},
{
"epoch": 8.045454545454545,
"grad_norm": 0.001562977209687233,
"learning_rate": 5.863636363636364e-06,
"loss": 0.0001,
"step": 1770
},
{
"epoch": 8.090909090909092,
"grad_norm": 0.0019722983706742525,
"learning_rate": 5.727272727272728e-06,
"loss": 0.0001,
"step": 1780
},
{
"epoch": 8.136363636363637,
"grad_norm": 0.0016468315152451396,
"learning_rate": 5.5909090909090915e-06,
"loss": 0.0001,
"step": 1790
},
{
"epoch": 8.181818181818182,
"grad_norm": 0.0018590294057503343,
"learning_rate": 5.4545454545454545e-06,
"loss": 0.0001,
"step": 1800
},
{
"epoch": 8.227272727272727,
"grad_norm": 0.002041436033323407,
"learning_rate": 5.318181818181818e-06,
"loss": 0.0001,
"step": 1810
},
{
"epoch": 8.272727272727273,
"grad_norm": 0.0021510140504688025,
"learning_rate": 5.181818181818182e-06,
"loss": 0.0001,
"step": 1820
},
{
"epoch": 8.318181818181818,
"grad_norm": 0.0015175098087638617,
"learning_rate": 5.045454545454546e-06,
"loss": 0.0001,
"step": 1830
},
{
"epoch": 8.363636363636363,
"grad_norm": 0.001754813943989575,
"learning_rate": 4.90909090909091e-06,
"loss": 0.0001,
"step": 1840
},
{
"epoch": 8.409090909090908,
"grad_norm": 0.001608902239240706,
"learning_rate": 4.7727272727272725e-06,
"loss": 0.0001,
"step": 1850
},
{
"epoch": 8.454545454545455,
"grad_norm": 0.002168968552723527,
"learning_rate": 4.636363636363636e-06,
"loss": 0.0001,
"step": 1860
},
{
"epoch": 8.5,
"grad_norm": 0.001456632511690259,
"learning_rate": 4.5e-06,
"loss": 0.0,
"step": 1870
},
{
"epoch": 8.545454545454545,
"grad_norm": 0.0017024242551997304,
"learning_rate": 4.363636363636364e-06,
"loss": 0.0001,
"step": 1880
},
{
"epoch": 8.590909090909092,
"grad_norm": 0.00176154519431293,
"learning_rate": 4.227272727272727e-06,
"loss": 0.0001,
"step": 1890
},
{
"epoch": 8.636363636363637,
"grad_norm": 0.0019339303253218532,
"learning_rate": 4.090909090909091e-06,
"loss": 0.0001,
"step": 1900
},
{
"epoch": 8.681818181818182,
"grad_norm": 0.0019142951350659132,
"learning_rate": 3.954545454545454e-06,
"loss": 0.0001,
"step": 1910
},
{
"epoch": 8.727272727272727,
"grad_norm": 0.0015304730040952563,
"learning_rate": 3.818181818181818e-06,
"loss": 0.0001,
"step": 1920
},
{
"epoch": 8.772727272727273,
"grad_norm": 0.0021663594525307417,
"learning_rate": 3.681818181818182e-06,
"loss": 0.0001,
"step": 1930
},
{
"epoch": 8.818181818181818,
"grad_norm": 0.001586704864166677,
"learning_rate": 3.5454545454545454e-06,
"loss": 0.0001,
"step": 1940
},
{
"epoch": 8.863636363636363,
"grad_norm": 0.001986338524147868,
"learning_rate": 3.409090909090909e-06,
"loss": 0.0,
"step": 1950
},
{
"epoch": 8.909090909090908,
"grad_norm": 0.001552366535179317,
"learning_rate": 3.2727272727272725e-06,
"loss": 0.0033,
"step": 1960
},
{
"epoch": 8.954545454545455,
"grad_norm": 0.0018597301095724106,
"learning_rate": 3.1363636363636363e-06,
"loss": 0.0001,
"step": 1970
},
{
"epoch": 9.0,
"grad_norm": 0.002290609758347273,
"learning_rate": 3e-06,
"loss": 0.0,
"step": 1980
},
{
"epoch": 9.0,
"eval_loss": 3.7486017390619963e-05,
"eval_runtime": 1.0378,
"eval_samples_per_second": 846.047,
"eval_steps_per_second": 52.998,
"step": 1980
},
{
"epoch": 9.045454545454545,
"grad_norm": 0.0014905119314789772,
"learning_rate": 2.863636363636364e-06,
"loss": 0.0,
"step": 1990
},
{
"epoch": 9.090909090909092,
"grad_norm": 0.0013228630414232612,
"learning_rate": 2.7272727272727272e-06,
"loss": 0.0001,
"step": 2000
},
{
"epoch": 9.136363636363637,
"grad_norm": 0.00189464061986655,
"learning_rate": 2.590909090909091e-06,
"loss": 0.0,
"step": 2010
},
{
"epoch": 9.181818181818182,
"grad_norm": 0.0020432292949408293,
"learning_rate": 2.454545454545455e-06,
"loss": 0.0001,
"step": 2020
},
{
"epoch": 9.227272727272727,
"grad_norm": 0.0014936975203454494,
"learning_rate": 2.318181818181818e-06,
"loss": 0.0001,
"step": 2030
},
{
"epoch": 9.272727272727273,
"grad_norm": 0.0020053344778716564,
"learning_rate": 2.181818181818182e-06,
"loss": 0.0,
"step": 2040
},
{
"epoch": 9.318181818181818,
"grad_norm": 0.0015693982131779194,
"learning_rate": 2.0454545454545453e-06,
"loss": 0.0,
"step": 2050
},
{
"epoch": 9.363636363636363,
"grad_norm": 0.0016571198357269168,
"learning_rate": 1.909090909090909e-06,
"loss": 0.0001,
"step": 2060
},
{
"epoch": 9.409090909090908,
"grad_norm": 0.0013354700058698654,
"learning_rate": 1.7727272727272727e-06,
"loss": 0.0,
"step": 2070
},
{
"epoch": 9.454545454545455,
"grad_norm": 0.0013917312026023865,
"learning_rate": 1.6363636363636363e-06,
"loss": 0.0,
"step": 2080
},
{
"epoch": 9.5,
"grad_norm": 0.0014988429611548781,
"learning_rate": 1.5e-06,
"loss": 0.0,
"step": 2090
},
{
"epoch": 9.545454545454545,
"grad_norm": 0.0014679876621812582,
"learning_rate": 1.3636363636363636e-06,
"loss": 0.0,
"step": 2100
},
{
"epoch": 9.590909090909092,
"grad_norm": 0.0018639364279806614,
"learning_rate": 1.2272727272727274e-06,
"loss": 0.0001,
"step": 2110
},
{
"epoch": 9.636363636363637,
"grad_norm": 0.001416134531609714,
"learning_rate": 1.090909090909091e-06,
"loss": 0.0,
"step": 2120
},
{
"epoch": 9.681818181818182,
"grad_norm": 0.0018406022572889924,
"learning_rate": 9.545454545454546e-07,
"loss": 0.0001,
"step": 2130
},
{
"epoch": 9.727272727272727,
"grad_norm": 0.0014891604660078883,
"learning_rate": 8.181818181818181e-07,
"loss": 0.0001,
"step": 2140
},
{
"epoch": 9.772727272727273,
"grad_norm": 0.0017427564598619938,
"learning_rate": 6.818181818181818e-07,
"loss": 0.0,
"step": 2150
},
{
"epoch": 9.818181818181818,
"grad_norm": 0.0014797528274357319,
"learning_rate": 5.454545454545455e-07,
"loss": 0.0,
"step": 2160
},
{
"epoch": 9.863636363636363,
"grad_norm": 0.0015343551058322191,
"learning_rate": 4.0909090909090906e-07,
"loss": 0.0001,
"step": 2170
},
{
"epoch": 9.909090909090908,
"grad_norm": 0.0014587711775675416,
"learning_rate": 2.7272727272727274e-07,
"loss": 0.0,
"step": 2180
},
{
"epoch": 9.954545454545455,
"grad_norm": 0.0013577837962657213,
"learning_rate": 1.3636363636363637e-07,
"loss": 0.0353,
"step": 2190
},
{
"epoch": 10.0,
"grad_norm": 0.0017256715800613165,
"learning_rate": 0.0,
"loss": 0.0001,
"step": 2200
},
{
"epoch": 10.0,
"eval_loss": 3.619581912062131e-05,
"eval_runtime": 1.0359,
"eval_samples_per_second": 847.586,
"eval_steps_per_second": 53.095,
"step": 2200
}
],
"logging_steps": 10,
"max_steps": 2200,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 234554255855400.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}