{ "best_metric": 0.9976791739463806, "best_model_checkpoint": "./mobilenetv2-typecoffee\\checkpoint-145300", "epoch": 100.0, "eval_steps": 500, "global_step": 581200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008602890571231933, "grad_norm": 13.874606132507324, "learning_rate": 4.999569855471439e-05, "loss": 1.6641, "step": 50 }, { "epoch": 0.017205781142463867, "grad_norm": 14.468774795532227, "learning_rate": 4.999139710942877e-05, "loss": 1.61, "step": 100 }, { "epoch": 0.025808671713695802, "grad_norm": 12.805009841918945, "learning_rate": 4.998709566414315e-05, "loss": 1.589, "step": 150 }, { "epoch": 0.034411562284927734, "grad_norm": 13.371192932128906, "learning_rate": 4.9982794218857534e-05, "loss": 1.5607, "step": 200 }, { "epoch": 0.04301445285615967, "grad_norm": 16.11214256286621, "learning_rate": 4.997849277357192e-05, "loss": 1.5587, "step": 250 }, { "epoch": 0.051617343427391604, "grad_norm": 15.734381675720215, "learning_rate": 4.997419132828631e-05, "loss": 1.5423, "step": 300 }, { "epoch": 0.060220233998623536, "grad_norm": 12.164186477661133, "learning_rate": 4.996988988300069e-05, "loss": 1.5365, "step": 350 }, { "epoch": 0.06882312456985547, "grad_norm": 13.708253860473633, "learning_rate": 4.996558843771508e-05, "loss": 1.5312, "step": 400 }, { "epoch": 0.0774260151410874, "grad_norm": 13.897035598754883, "learning_rate": 4.996128699242946e-05, "loss": 1.5117, "step": 450 }, { "epoch": 0.08602890571231935, "grad_norm": 16.94737434387207, "learning_rate": 4.995698554714384e-05, "loss": 1.5149, "step": 500 }, { "epoch": 0.09463179628355127, "grad_norm": 14.74962043762207, "learning_rate": 4.995268410185823e-05, "loss": 1.5345, "step": 550 }, { "epoch": 0.10323468685478321, "grad_norm": 14.384037017822266, "learning_rate": 4.994838265657261e-05, "loss": 1.4901, "step": 600 }, { "epoch": 0.11183757742601515, "grad_norm": 13.840729713439941, "learning_rate": 4.9944081211287e-05, "loss": 1.4641, "step": 650 }, { "epoch": 0.12044046799724707, "grad_norm": 10.810714721679688, "learning_rate": 4.993977976600138e-05, "loss": 1.4785, "step": 700 }, { "epoch": 0.129043358568479, "grad_norm": 16.095022201538086, "learning_rate": 4.993547832071576e-05, "loss": 1.5185, "step": 750 }, { "epoch": 0.13764624913971094, "grad_norm": 13.53649616241455, "learning_rate": 4.993117687543015e-05, "loss": 1.4713, "step": 800 }, { "epoch": 0.1462491397109429, "grad_norm": 10.76221752166748, "learning_rate": 4.992687543014453e-05, "loss": 1.4364, "step": 850 }, { "epoch": 0.1548520302821748, "grad_norm": 11.265481948852539, "learning_rate": 4.992257398485892e-05, "loss": 1.4525, "step": 900 }, { "epoch": 0.16345492085340674, "grad_norm": 13.075508117675781, "learning_rate": 4.99182725395733e-05, "loss": 1.4521, "step": 950 }, { "epoch": 0.1720578114246387, "grad_norm": 9.532099723815918, "learning_rate": 4.991397109428768e-05, "loss": 1.444, "step": 1000 }, { "epoch": 0.18066070199587062, "grad_norm": 11.907129287719727, "learning_rate": 4.990966964900207e-05, "loss": 1.4165, "step": 1050 }, { "epoch": 0.18926359256710254, "grad_norm": 13.140506744384766, "learning_rate": 4.990536820371645e-05, "loss": 1.4361, "step": 1100 }, { "epoch": 0.1978664831383345, "grad_norm": 13.249748229980469, "learning_rate": 4.990106675843084e-05, "loss": 1.4483, "step": 1150 }, { "epoch": 0.20646937370956642, "grad_norm": 14.576303482055664, "learning_rate": 4.9896765313145225e-05, "loss": 1.3993, "step": 1200 }, { "epoch": 0.21507226428079834, "grad_norm": 13.937341690063477, "learning_rate": 4.98924638678596e-05, "loss": 1.4076, "step": 1250 }, { "epoch": 0.2236751548520303, "grad_norm": 13.786910057067871, "learning_rate": 4.988816242257399e-05, "loss": 1.4144, "step": 1300 }, { "epoch": 0.23227804542326222, "grad_norm": 13.71471881866455, "learning_rate": 4.988386097728837e-05, "loss": 1.4166, "step": 1350 }, { "epoch": 0.24088093599449414, "grad_norm": 12.28220272064209, "learning_rate": 4.9879559532002757e-05, "loss": 1.4395, "step": 1400 }, { "epoch": 0.2494838265657261, "grad_norm": 11.353926658630371, "learning_rate": 4.9875258086717144e-05, "loss": 1.4289, "step": 1450 }, { "epoch": 0.258086717136958, "grad_norm": 12.339083671569824, "learning_rate": 4.987095664143152e-05, "loss": 1.4136, "step": 1500 }, { "epoch": 0.26668960770818995, "grad_norm": 13.849453926086426, "learning_rate": 4.986665519614591e-05, "loss": 1.3704, "step": 1550 }, { "epoch": 0.27529249827942187, "grad_norm": 14.709623336791992, "learning_rate": 4.986235375086029e-05, "loss": 1.4428, "step": 1600 }, { "epoch": 0.2838953888506538, "grad_norm": 11.732614517211914, "learning_rate": 4.9858052305574676e-05, "loss": 1.3912, "step": 1650 }, { "epoch": 0.2924982794218858, "grad_norm": 15.952874183654785, "learning_rate": 4.9853750860289064e-05, "loss": 1.3933, "step": 1700 }, { "epoch": 0.3011011699931177, "grad_norm": 10.65099048614502, "learning_rate": 4.984944941500344e-05, "loss": 1.3909, "step": 1750 }, { "epoch": 0.3097040605643496, "grad_norm": 13.565245628356934, "learning_rate": 4.9845147969717826e-05, "loss": 1.4064, "step": 1800 }, { "epoch": 0.31830695113558155, "grad_norm": 12.298967361450195, "learning_rate": 4.9840846524432214e-05, "loss": 1.3851, "step": 1850 }, { "epoch": 0.3269098417068135, "grad_norm": 12.037237167358398, "learning_rate": 4.9836545079146595e-05, "loss": 1.4064, "step": 1900 }, { "epoch": 0.3355127322780454, "grad_norm": 15.738449096679688, "learning_rate": 4.983224363386098e-05, "loss": 1.4148, "step": 1950 }, { "epoch": 0.3441156228492774, "grad_norm": 18.89379119873047, "learning_rate": 4.9827942188575364e-05, "loss": 1.3896, "step": 2000 }, { "epoch": 0.3527185134205093, "grad_norm": 15.500096321105957, "learning_rate": 4.9823640743289746e-05, "loss": 1.4046, "step": 2050 }, { "epoch": 0.36132140399174123, "grad_norm": 10.018769264221191, "learning_rate": 4.9819339298004134e-05, "loss": 1.3773, "step": 2100 }, { "epoch": 0.36992429456297315, "grad_norm": 17.35185432434082, "learning_rate": 4.9815037852718515e-05, "loss": 1.3906, "step": 2150 }, { "epoch": 0.3785271851342051, "grad_norm": 11.627890586853027, "learning_rate": 4.98107364074329e-05, "loss": 1.428, "step": 2200 }, { "epoch": 0.387130075705437, "grad_norm": 11.006091117858887, "learning_rate": 4.9806434962147284e-05, "loss": 1.4305, "step": 2250 }, { "epoch": 0.395732966276669, "grad_norm": 12.06324577331543, "learning_rate": 4.9802133516861665e-05, "loss": 1.3869, "step": 2300 }, { "epoch": 0.4043358568479009, "grad_norm": 11.566773414611816, "learning_rate": 4.979783207157605e-05, "loss": 1.3826, "step": 2350 }, { "epoch": 0.41293874741913283, "grad_norm": 11.53071403503418, "learning_rate": 4.9793530626290434e-05, "loss": 1.4024, "step": 2400 }, { "epoch": 0.42154163799036476, "grad_norm": 10.548382759094238, "learning_rate": 4.978922918100482e-05, "loss": 1.3514, "step": 2450 }, { "epoch": 0.4301445285615967, "grad_norm": 12.574633598327637, "learning_rate": 4.97849277357192e-05, "loss": 1.375, "step": 2500 }, { "epoch": 0.4387474191328286, "grad_norm": 13.270973205566406, "learning_rate": 4.978062629043359e-05, "loss": 1.346, "step": 2550 }, { "epoch": 0.4473503097040606, "grad_norm": 10.755953788757324, "learning_rate": 4.977632484514797e-05, "loss": 1.4104, "step": 2600 }, { "epoch": 0.4559532002752925, "grad_norm": 12.99761962890625, "learning_rate": 4.9772023399862354e-05, "loss": 1.3712, "step": 2650 }, { "epoch": 0.46455609084652444, "grad_norm": 11.830048561096191, "learning_rate": 4.976772195457674e-05, "loss": 1.3256, "step": 2700 }, { "epoch": 0.47315898141775636, "grad_norm": 12.662361145019531, "learning_rate": 4.976342050929112e-05, "loss": 1.3343, "step": 2750 }, { "epoch": 0.4817618719889883, "grad_norm": 9.659152030944824, "learning_rate": 4.975911906400551e-05, "loss": 1.3307, "step": 2800 }, { "epoch": 0.4903647625602202, "grad_norm": 13.964424133300781, "learning_rate": 4.975481761871989e-05, "loss": 1.3648, "step": 2850 }, { "epoch": 0.4989676531314522, "grad_norm": 11.565299034118652, "learning_rate": 4.975051617343427e-05, "loss": 1.3452, "step": 2900 }, { "epoch": 0.5075705437026841, "grad_norm": 20.211776733398438, "learning_rate": 4.974621472814866e-05, "loss": 1.3515, "step": 2950 }, { "epoch": 0.516173434273916, "grad_norm": 12.26140022277832, "learning_rate": 4.974191328286304e-05, "loss": 1.365, "step": 3000 }, { "epoch": 0.524776324845148, "grad_norm": 11.490259170532227, "learning_rate": 4.973761183757743e-05, "loss": 1.3193, "step": 3050 }, { "epoch": 0.5333792154163799, "grad_norm": 10.181940078735352, "learning_rate": 4.973331039229181e-05, "loss": 1.3855, "step": 3100 }, { "epoch": 0.5419821059876119, "grad_norm": 11.861502647399902, "learning_rate": 4.972900894700619e-05, "loss": 1.3521, "step": 3150 }, { "epoch": 0.5505849965588437, "grad_norm": 12.692533493041992, "learning_rate": 4.972470750172058e-05, "loss": 1.2813, "step": 3200 }, { "epoch": 0.5591878871300757, "grad_norm": 11.342544555664062, "learning_rate": 4.972040605643497e-05, "loss": 1.3131, "step": 3250 }, { "epoch": 0.5677907777013076, "grad_norm": 12.48619270324707, "learning_rate": 4.971610461114935e-05, "loss": 1.3128, "step": 3300 }, { "epoch": 0.5763936682725396, "grad_norm": 11.72147274017334, "learning_rate": 4.971180316586374e-05, "loss": 1.2796, "step": 3350 }, { "epoch": 0.5849965588437716, "grad_norm": 12.354662895202637, "learning_rate": 4.970750172057811e-05, "loss": 1.3176, "step": 3400 }, { "epoch": 0.5935994494150034, "grad_norm": 10.477376937866211, "learning_rate": 4.97032002752925e-05, "loss": 1.2993, "step": 3450 }, { "epoch": 0.6022023399862354, "grad_norm": 9.37282657623291, "learning_rate": 4.969889883000689e-05, "loss": 1.3226, "step": 3500 }, { "epoch": 0.6108052305574673, "grad_norm": 11.707993507385254, "learning_rate": 4.969459738472127e-05, "loss": 1.3294, "step": 3550 }, { "epoch": 0.6194081211286993, "grad_norm": 11.999861717224121, "learning_rate": 4.969029593943566e-05, "loss": 1.2978, "step": 3600 }, { "epoch": 0.6280110116999312, "grad_norm": 11.146305084228516, "learning_rate": 4.968599449415003e-05, "loss": 1.3599, "step": 3650 }, { "epoch": 0.6366139022711631, "grad_norm": 10.04617691040039, "learning_rate": 4.968169304886442e-05, "loss": 1.3172, "step": 3700 }, { "epoch": 0.6452167928423951, "grad_norm": 11.032917022705078, "learning_rate": 4.967739160357881e-05, "loss": 1.3336, "step": 3750 }, { "epoch": 0.653819683413627, "grad_norm": 12.053647994995117, "learning_rate": 4.967309015829319e-05, "loss": 1.274, "step": 3800 }, { "epoch": 0.6624225739848589, "grad_norm": 12.994665145874023, "learning_rate": 4.9668788713007576e-05, "loss": 1.307, "step": 3850 }, { "epoch": 0.6710254645560908, "grad_norm": 8.859834671020508, "learning_rate": 4.966448726772196e-05, "loss": 1.3533, "step": 3900 }, { "epoch": 0.6796283551273228, "grad_norm": 9.247245788574219, "learning_rate": 4.966018582243634e-05, "loss": 1.3313, "step": 3950 }, { "epoch": 0.6882312456985548, "grad_norm": 9.9243803024292, "learning_rate": 4.9655884377150727e-05, "loss": 1.3054, "step": 4000 }, { "epoch": 0.6968341362697866, "grad_norm": 10.985905647277832, "learning_rate": 4.965158293186511e-05, "loss": 1.3256, "step": 4050 }, { "epoch": 0.7054370268410186, "grad_norm": 15.668318748474121, "learning_rate": 4.9647281486579496e-05, "loss": 1.3446, "step": 4100 }, { "epoch": 0.7140399174122505, "grad_norm": 12.085655212402344, "learning_rate": 4.964298004129388e-05, "loss": 1.3055, "step": 4150 }, { "epoch": 0.7226428079834825, "grad_norm": 12.59761905670166, "learning_rate": 4.963867859600826e-05, "loss": 1.3043, "step": 4200 }, { "epoch": 0.7312456985547143, "grad_norm": 9.371703147888184, "learning_rate": 4.9634377150722646e-05, "loss": 1.3477, "step": 4250 }, { "epoch": 0.7398485891259463, "grad_norm": 11.015939712524414, "learning_rate": 4.963007570543703e-05, "loss": 1.3172, "step": 4300 }, { "epoch": 0.7484514796971783, "grad_norm": 12.30109977722168, "learning_rate": 4.9625774260151415e-05, "loss": 1.3187, "step": 4350 }, { "epoch": 0.7570543702684102, "grad_norm": 8.654661178588867, "learning_rate": 4.9621472814865796e-05, "loss": 1.3257, "step": 4400 }, { "epoch": 0.7656572608396421, "grad_norm": 10.325228691101074, "learning_rate": 4.961717136958018e-05, "loss": 1.3091, "step": 4450 }, { "epoch": 0.774260151410874, "grad_norm": 12.7671480178833, "learning_rate": 4.9612869924294565e-05, "loss": 1.26, "step": 4500 }, { "epoch": 0.782863041982106, "grad_norm": 13.846147537231445, "learning_rate": 4.9608568479008947e-05, "loss": 1.3029, "step": 4550 }, { "epoch": 0.791465932553338, "grad_norm": 12.797240257263184, "learning_rate": 4.9604267033723335e-05, "loss": 1.3127, "step": 4600 }, { "epoch": 0.8000688231245698, "grad_norm": 11.94734001159668, "learning_rate": 4.9599965588437716e-05, "loss": 1.2985, "step": 4650 }, { "epoch": 0.8086717136958018, "grad_norm": 12.71529769897461, "learning_rate": 4.9595664143152104e-05, "loss": 1.3609, "step": 4700 }, { "epoch": 0.8172746042670337, "grad_norm": 11.306694030761719, "learning_rate": 4.9591362697866485e-05, "loss": 1.3007, "step": 4750 }, { "epoch": 0.8258774948382657, "grad_norm": 9.800276756286621, "learning_rate": 4.9587061252580866e-05, "loss": 1.2885, "step": 4800 }, { "epoch": 0.8344803854094975, "grad_norm": 21.473785400390625, "learning_rate": 4.9582759807295254e-05, "loss": 1.3211, "step": 4850 }, { "epoch": 0.8430832759807295, "grad_norm": 13.846829414367676, "learning_rate": 4.957845836200964e-05, "loss": 1.3046, "step": 4900 }, { "epoch": 0.8516861665519615, "grad_norm": 8.907471656799316, "learning_rate": 4.957415691672402e-05, "loss": 1.3047, "step": 4950 }, { "epoch": 0.8602890571231934, "grad_norm": 8.968548774719238, "learning_rate": 4.9569855471438404e-05, "loss": 1.2917, "step": 5000 }, { "epoch": 0.8688919476944253, "grad_norm": 10.980504989624023, "learning_rate": 4.9565554026152785e-05, "loss": 1.2779, "step": 5050 }, { "epoch": 0.8774948382656572, "grad_norm": 11.495574951171875, "learning_rate": 4.956125258086717e-05, "loss": 1.3079, "step": 5100 }, { "epoch": 0.8860977288368892, "grad_norm": 10.425631523132324, "learning_rate": 4.955695113558156e-05, "loss": 1.2716, "step": 5150 }, { "epoch": 0.8947006194081212, "grad_norm": 13.97283935546875, "learning_rate": 4.955264969029594e-05, "loss": 1.285, "step": 5200 }, { "epoch": 0.903303509979353, "grad_norm": 9.34812068939209, "learning_rate": 4.9548348245010324e-05, "loss": 1.2621, "step": 5250 }, { "epoch": 0.911906400550585, "grad_norm": 14.970287322998047, "learning_rate": 4.9544046799724705e-05, "loss": 1.2797, "step": 5300 }, { "epoch": 0.9205092911218169, "grad_norm": 8.435407638549805, "learning_rate": 4.953974535443909e-05, "loss": 1.2756, "step": 5350 }, { "epoch": 0.9291121816930489, "grad_norm": 7.786146640777588, "learning_rate": 4.953544390915348e-05, "loss": 1.305, "step": 5400 }, { "epoch": 0.9377150722642807, "grad_norm": 10.196477890014648, "learning_rate": 4.953114246386786e-05, "loss": 1.2982, "step": 5450 }, { "epoch": 0.9463179628355127, "grad_norm": 11.93211841583252, "learning_rate": 4.952684101858225e-05, "loss": 1.2873, "step": 5500 }, { "epoch": 0.9549208534067447, "grad_norm": 13.924481391906738, "learning_rate": 4.952253957329663e-05, "loss": 1.2348, "step": 5550 }, { "epoch": 0.9635237439779766, "grad_norm": 11.79347038269043, "learning_rate": 4.951823812801101e-05, "loss": 1.2764, "step": 5600 }, { "epoch": 0.9721266345492086, "grad_norm": 11.686305046081543, "learning_rate": 4.95139366827254e-05, "loss": 1.2877, "step": 5650 }, { "epoch": 0.9807295251204404, "grad_norm": 11.770838737487793, "learning_rate": 4.950963523743978e-05, "loss": 1.2673, "step": 5700 }, { "epoch": 0.9893324156916724, "grad_norm": 12.244403839111328, "learning_rate": 4.950533379215417e-05, "loss": 1.2657, "step": 5750 }, { "epoch": 0.9979353062629044, "grad_norm": 17.923946380615234, "learning_rate": 4.950103234686855e-05, "loss": 1.2625, "step": 5800 }, { "epoch": 1.0, "eval_accuracy": 0.4037170882808467, "eval_f1": 0.3557606203123941, "eval_f1_DuraRiadoRio_16x16": 0.28837787445618396, "eval_f1_Mole_16x16": 0.20045045045045046, "eval_f1_Quebrado_16x16": 0.5594719471947195, "eval_f1_RiadoRio_16x16": 0.41658274588788186, "eval_f1_RioFechado_16x16": 0.3139200835727344, "eval_loss": 1.6691983938217163, "eval_precision": 0.478157029188541, "eval_precision_DuraRiadoRio_16x16": 0.5076586433260394, "eval_precision_Mole_16x16": 0.7416666666666667, "eval_precision_Quebrado_16x16": 0.4020110036046291, "eval_precision_RiadoRio_16x16": 0.3519568916619399, "eval_precision_RioFechado_16x16": 0.38749194068343007, "eval_recall": 0.4022173382314434, "eval_recall_DuraRiadoRio_16x16": 0.2013888888888889, "eval_recall_Mole_16x16": 0.11588541666666667, "eval_recall_Quebrado_16x16": 0.9197048611111112, "eval_recall_RiadoRio_16x16": 0.5102796052631579, "eval_recall_RioFechado_16x16": 0.26382791922739246, "eval_runtime": 43.3243, "eval_samples_per_second": 268.256, "eval_steps_per_second": 16.78, "step": 5812 }, { "epoch": 1.0065381968341363, "grad_norm": 11.09004020690918, "learning_rate": 4.949673090158293e-05, "loss": 1.2581, "step": 5850 }, { "epoch": 1.0151410874053681, "grad_norm": 6.4218950271606445, "learning_rate": 4.949242945629732e-05, "loss": 1.2598, "step": 5900 }, { "epoch": 1.0237439779766002, "grad_norm": 10.536774635314941, "learning_rate": 4.94881280110117e-05, "loss": 1.2802, "step": 5950 }, { "epoch": 1.032346868547832, "grad_norm": 11.318628311157227, "learning_rate": 4.948382656572609e-05, "loss": 1.2267, "step": 6000 }, { "epoch": 1.040949759119064, "grad_norm": 10.506241798400879, "learning_rate": 4.947952512044047e-05, "loss": 1.2322, "step": 6050 }, { "epoch": 1.049552649690296, "grad_norm": 10.276853561401367, "learning_rate": 4.947522367515485e-05, "loss": 1.2503, "step": 6100 }, { "epoch": 1.058155540261528, "grad_norm": 9.759271621704102, "learning_rate": 4.947092222986924e-05, "loss": 1.2705, "step": 6150 }, { "epoch": 1.0667584308327598, "grad_norm": 15.14022159576416, "learning_rate": 4.946662078458362e-05, "loss": 1.2485, "step": 6200 }, { "epoch": 1.0753613214039917, "grad_norm": 9.084566116333008, "learning_rate": 4.946231933929801e-05, "loss": 1.2308, "step": 6250 }, { "epoch": 1.0839642119752237, "grad_norm": 14.45799446105957, "learning_rate": 4.9458017894012396e-05, "loss": 1.2969, "step": 6300 }, { "epoch": 1.0925671025464556, "grad_norm": 12.430665016174316, "learning_rate": 4.945371644872677e-05, "loss": 1.2745, "step": 6350 }, { "epoch": 1.1011699931176875, "grad_norm": 10.0015287399292, "learning_rate": 4.944941500344116e-05, "loss": 1.2471, "step": 6400 }, { "epoch": 1.1097728836889196, "grad_norm": 12.208296775817871, "learning_rate": 4.944511355815554e-05, "loss": 1.2899, "step": 6450 }, { "epoch": 1.1183757742601514, "grad_norm": 9.420799255371094, "learning_rate": 4.944081211286993e-05, "loss": 1.2483, "step": 6500 }, { "epoch": 1.1269786648313833, "grad_norm": 9.979134559631348, "learning_rate": 4.9436510667584315e-05, "loss": 1.3112, "step": 6550 }, { "epoch": 1.1355815554026152, "grad_norm": 10.230840682983398, "learning_rate": 4.943220922229869e-05, "loss": 1.2841, "step": 6600 }, { "epoch": 1.1441844459738473, "grad_norm": 25.69493293762207, "learning_rate": 4.942790777701308e-05, "loss": 1.2632, "step": 6650 }, { "epoch": 1.1527873365450791, "grad_norm": 12.315206527709961, "learning_rate": 4.942360633172746e-05, "loss": 1.2107, "step": 6700 }, { "epoch": 1.161390227116311, "grad_norm": 16.225040435791016, "learning_rate": 4.941930488644185e-05, "loss": 1.2472, "step": 6750 }, { "epoch": 1.169993117687543, "grad_norm": 9.702127456665039, "learning_rate": 4.9415003441156235e-05, "loss": 1.2709, "step": 6800 }, { "epoch": 1.178596008258775, "grad_norm": 13.141595840454102, "learning_rate": 4.9410701995870616e-05, "loss": 1.2973, "step": 6850 }, { "epoch": 1.1871988988300068, "grad_norm": 9.340795516967773, "learning_rate": 4.9406400550585e-05, "loss": 1.2827, "step": 6900 }, { "epoch": 1.195801789401239, "grad_norm": 10.446883201599121, "learning_rate": 4.9402099105299385e-05, "loss": 1.3047, "step": 6950 }, { "epoch": 1.2044046799724708, "grad_norm": 11.482792854309082, "learning_rate": 4.9397797660013766e-05, "loss": 1.2405, "step": 7000 }, { "epoch": 1.2130075705437027, "grad_norm": 16.536739349365234, "learning_rate": 4.9393496214728154e-05, "loss": 1.2274, "step": 7050 }, { "epoch": 1.2216104611149345, "grad_norm": 11.198225021362305, "learning_rate": 4.9389194769442535e-05, "loss": 1.2288, "step": 7100 }, { "epoch": 1.2302133516861666, "grad_norm": 10.817582130432129, "learning_rate": 4.9384893324156917e-05, "loss": 1.2892, "step": 7150 }, { "epoch": 1.2388162422573985, "grad_norm": 11.8817777633667, "learning_rate": 4.9380591878871305e-05, "loss": 1.2388, "step": 7200 }, { "epoch": 1.2474191328286304, "grad_norm": 11.52081298828125, "learning_rate": 4.9376290433585686e-05, "loss": 1.2437, "step": 7250 }, { "epoch": 1.2560220233998622, "grad_norm": 9.285940170288086, "learning_rate": 4.9371988988300074e-05, "loss": 1.2473, "step": 7300 }, { "epoch": 1.2646249139710943, "grad_norm": 11.291216850280762, "learning_rate": 4.9367687543014455e-05, "loss": 1.2418, "step": 7350 }, { "epoch": 1.2732278045423262, "grad_norm": 12.938750267028809, "learning_rate": 4.9363386097728836e-05, "loss": 1.2661, "step": 7400 }, { "epoch": 1.281830695113558, "grad_norm": 11.961236953735352, "learning_rate": 4.9359084652443224e-05, "loss": 1.2156, "step": 7450 }, { "epoch": 1.2904335856847902, "grad_norm": 11.83258056640625, "learning_rate": 4.9354783207157605e-05, "loss": 1.2198, "step": 7500 }, { "epoch": 1.299036476256022, "grad_norm": 10.678873062133789, "learning_rate": 4.935048176187199e-05, "loss": 1.2658, "step": 7550 }, { "epoch": 1.307639366827254, "grad_norm": 10.375088691711426, "learning_rate": 4.9346180316586374e-05, "loss": 1.2202, "step": 7600 }, { "epoch": 1.316242257398486, "grad_norm": 12.464033126831055, "learning_rate": 4.934187887130076e-05, "loss": 1.2514, "step": 7650 }, { "epoch": 1.3248451479697179, "grad_norm": 10.885096549987793, "learning_rate": 4.933757742601514e-05, "loss": 1.1898, "step": 7700 }, { "epoch": 1.3334480385409497, "grad_norm": 14.115507125854492, "learning_rate": 4.9333275980729525e-05, "loss": 1.2577, "step": 7750 }, { "epoch": 1.3420509291121818, "grad_norm": 9.37861156463623, "learning_rate": 4.932897453544391e-05, "loss": 1.2777, "step": 7800 }, { "epoch": 1.3506538196834137, "grad_norm": 16.28864288330078, "learning_rate": 4.9324673090158294e-05, "loss": 1.2598, "step": 7850 }, { "epoch": 1.3592567102546456, "grad_norm": 15.778472900390625, "learning_rate": 4.932037164487268e-05, "loss": 1.2151, "step": 7900 }, { "epoch": 1.3678596008258774, "grad_norm": 8.507882118225098, "learning_rate": 4.931607019958706e-05, "loss": 1.2451, "step": 7950 }, { "epoch": 1.3764624913971093, "grad_norm": 13.908921241760254, "learning_rate": 4.9311768754301444e-05, "loss": 1.2968, "step": 8000 }, { "epoch": 1.3850653819683414, "grad_norm": 10.563688278198242, "learning_rate": 4.930746730901583e-05, "loss": 1.289, "step": 8050 }, { "epoch": 1.3936682725395733, "grad_norm": 10.757795333862305, "learning_rate": 4.930316586373021e-05, "loss": 1.2438, "step": 8100 }, { "epoch": 1.4022711631108051, "grad_norm": 9.429142951965332, "learning_rate": 4.92988644184446e-05, "loss": 1.2442, "step": 8150 }, { "epoch": 1.4108740536820372, "grad_norm": 9.320662498474121, "learning_rate": 4.929456297315899e-05, "loss": 1.2738, "step": 8200 }, { "epoch": 1.419476944253269, "grad_norm": 9.128153800964355, "learning_rate": 4.929026152787336e-05, "loss": 1.2098, "step": 8250 }, { "epoch": 1.428079834824501, "grad_norm": 10.14345645904541, "learning_rate": 4.928596008258775e-05, "loss": 1.2721, "step": 8300 }, { "epoch": 1.436682725395733, "grad_norm": 10.856945991516113, "learning_rate": 4.928165863730214e-05, "loss": 1.22, "step": 8350 }, { "epoch": 1.445285615966965, "grad_norm": 10.88556957244873, "learning_rate": 4.927735719201652e-05, "loss": 1.2837, "step": 8400 }, { "epoch": 1.4538885065381968, "grad_norm": 11.498441696166992, "learning_rate": 4.927305574673091e-05, "loss": 1.2249, "step": 8450 }, { "epoch": 1.4624913971094289, "grad_norm": 8.089730262756348, "learning_rate": 4.926875430144528e-05, "loss": 1.2314, "step": 8500 }, { "epoch": 1.4710942876806608, "grad_norm": 8.864239692687988, "learning_rate": 4.926445285615967e-05, "loss": 1.2225, "step": 8550 }, { "epoch": 1.4796971782518926, "grad_norm": 10.722824096679688, "learning_rate": 4.926015141087406e-05, "loss": 1.3006, "step": 8600 }, { "epoch": 1.4883000688231245, "grad_norm": 10.375917434692383, "learning_rate": 4.925584996558844e-05, "loss": 1.2226, "step": 8650 }, { "epoch": 1.4969029593943566, "grad_norm": 13.003503799438477, "learning_rate": 4.925154852030283e-05, "loss": 1.2223, "step": 8700 }, { "epoch": 1.5055058499655884, "grad_norm": 11.364338874816895, "learning_rate": 4.92472470750172e-05, "loss": 1.2548, "step": 8750 }, { "epoch": 1.5141087405368203, "grad_norm": 11.156951904296875, "learning_rate": 4.924294562973159e-05, "loss": 1.3141, "step": 8800 }, { "epoch": 1.5227116311080522, "grad_norm": 9.266166687011719, "learning_rate": 4.923864418444598e-05, "loss": 1.2373, "step": 8850 }, { "epoch": 1.5313145216792843, "grad_norm": 10.958678245544434, "learning_rate": 4.923434273916036e-05, "loss": 1.2272, "step": 8900 }, { "epoch": 1.5399174122505161, "grad_norm": 10.486472129821777, "learning_rate": 4.923004129387475e-05, "loss": 1.229, "step": 8950 }, { "epoch": 1.548520302821748, "grad_norm": 7.6122918128967285, "learning_rate": 4.922573984858913e-05, "loss": 1.2358, "step": 9000 }, { "epoch": 1.55712319339298, "grad_norm": 8.866777420043945, "learning_rate": 4.922143840330351e-05, "loss": 1.2251, "step": 9050 }, { "epoch": 1.565726083964212, "grad_norm": 13.207948684692383, "learning_rate": 4.92171369580179e-05, "loss": 1.239, "step": 9100 }, { "epoch": 1.5743289745354438, "grad_norm": 11.939790725708008, "learning_rate": 4.921283551273228e-05, "loss": 1.2249, "step": 9150 }, { "epoch": 1.582931865106676, "grad_norm": 10.389349937438965, "learning_rate": 4.920853406744667e-05, "loss": 1.2171, "step": 9200 }, { "epoch": 1.5915347556779078, "grad_norm": 8.054533004760742, "learning_rate": 4.920423262216105e-05, "loss": 1.2616, "step": 9250 }, { "epoch": 1.6001376462491397, "grad_norm": 7.838688373565674, "learning_rate": 4.919993117687543e-05, "loss": 1.2335, "step": 9300 }, { "epoch": 1.6087405368203718, "grad_norm": 10.445599555969238, "learning_rate": 4.919562973158982e-05, "loss": 1.255, "step": 9350 }, { "epoch": 1.6173434273916034, "grad_norm": 11.625615119934082, "learning_rate": 4.91913282863042e-05, "loss": 1.2369, "step": 9400 }, { "epoch": 1.6259463179628355, "grad_norm": 8.240401268005371, "learning_rate": 4.9187026841018586e-05, "loss": 1.3119, "step": 9450 }, { "epoch": 1.6345492085340676, "grad_norm": 11.563984870910645, "learning_rate": 4.918272539573297e-05, "loss": 1.2074, "step": 9500 }, { "epoch": 1.6431520991052992, "grad_norm": 7.806783199310303, "learning_rate": 4.9178423950447355e-05, "loss": 1.2775, "step": 9550 }, { "epoch": 1.6517549896765313, "grad_norm": 10.189064025878906, "learning_rate": 4.9174122505161736e-05, "loss": 1.2477, "step": 9600 }, { "epoch": 1.6603578802477632, "grad_norm": 9.787823677062988, "learning_rate": 4.916982105987612e-05, "loss": 1.2101, "step": 9650 }, { "epoch": 1.668960770818995, "grad_norm": 10.610733032226562, "learning_rate": 4.9165519614590505e-05, "loss": 1.2117, "step": 9700 }, { "epoch": 1.6775636613902272, "grad_norm": 12.896672248840332, "learning_rate": 4.9161218169304893e-05, "loss": 1.2634, "step": 9750 }, { "epoch": 1.686166551961459, "grad_norm": 10.037786483764648, "learning_rate": 4.9156916724019275e-05, "loss": 1.2622, "step": 9800 }, { "epoch": 1.694769442532691, "grad_norm": 8.214544296264648, "learning_rate": 4.9152615278733656e-05, "loss": 1.2465, "step": 9850 }, { "epoch": 1.703372333103923, "grad_norm": 7.726459980010986, "learning_rate": 4.914831383344804e-05, "loss": 1.2258, "step": 9900 }, { "epoch": 1.7119752236751549, "grad_norm": 8.272558212280273, "learning_rate": 4.9144012388162425e-05, "loss": 1.187, "step": 9950 }, { "epoch": 1.7205781142463867, "grad_norm": 7.534845352172852, "learning_rate": 4.913971094287681e-05, "loss": 1.2149, "step": 10000 }, { "epoch": 1.7291810048176188, "grad_norm": 6.756843090057373, "learning_rate": 4.9135409497591194e-05, "loss": 1.1891, "step": 10050 }, { "epoch": 1.7377838953888507, "grad_norm": 9.385401725769043, "learning_rate": 4.9131108052305575e-05, "loss": 1.2322, "step": 10100 }, { "epoch": 1.7463867859600826, "grad_norm": 9.431879997253418, "learning_rate": 4.9126806607019956e-05, "loss": 1.225, "step": 10150 }, { "epoch": 1.7549896765313147, "grad_norm": 9.035143852233887, "learning_rate": 4.9122505161734344e-05, "loss": 1.199, "step": 10200 }, { "epoch": 1.7635925671025463, "grad_norm": 9.286103248596191, "learning_rate": 4.911820371644873e-05, "loss": 1.2473, "step": 10250 }, { "epoch": 1.7721954576737784, "grad_norm": 9.58206558227539, "learning_rate": 4.9113902271163113e-05, "loss": 1.239, "step": 10300 }, { "epoch": 1.7807983482450105, "grad_norm": 8.583107948303223, "learning_rate": 4.91096008258775e-05, "loss": 1.214, "step": 10350 }, { "epoch": 1.7894012388162421, "grad_norm": 9.411613464355469, "learning_rate": 4.910529938059188e-05, "loss": 1.1843, "step": 10400 }, { "epoch": 1.7980041293874742, "grad_norm": 13.505537986755371, "learning_rate": 4.9100997935306264e-05, "loss": 1.154, "step": 10450 }, { "epoch": 1.806607019958706, "grad_norm": 11.143860816955566, "learning_rate": 4.909669649002065e-05, "loss": 1.1872, "step": 10500 }, { "epoch": 1.815209910529938, "grad_norm": 13.327857971191406, "learning_rate": 4.909239504473503e-05, "loss": 1.2308, "step": 10550 }, { "epoch": 1.82381280110117, "grad_norm": 17.38473892211914, "learning_rate": 4.908809359944942e-05, "loss": 1.1924, "step": 10600 }, { "epoch": 1.832415691672402, "grad_norm": 9.832144737243652, "learning_rate": 4.90837921541638e-05, "loss": 1.2751, "step": 10650 }, { "epoch": 1.8410185822436338, "grad_norm": 10.646970748901367, "learning_rate": 4.907949070887818e-05, "loss": 1.209, "step": 10700 }, { "epoch": 1.8496214728148659, "grad_norm": 8.377860069274902, "learning_rate": 4.907518926359257e-05, "loss": 1.2069, "step": 10750 }, { "epoch": 1.8582243633860978, "grad_norm": 9.223164558410645, "learning_rate": 4.907088781830695e-05, "loss": 1.2617, "step": 10800 }, { "epoch": 1.8668272539573296, "grad_norm": 9.218362808227539, "learning_rate": 4.906658637302134e-05, "loss": 1.1984, "step": 10850 }, { "epoch": 1.8754301445285617, "grad_norm": 9.784517288208008, "learning_rate": 4.906228492773572e-05, "loss": 1.2175, "step": 10900 }, { "epoch": 1.8840330350997936, "grad_norm": 11.078054428100586, "learning_rate": 4.90579834824501e-05, "loss": 1.2044, "step": 10950 }, { "epoch": 1.8926359256710255, "grad_norm": 9.21428108215332, "learning_rate": 4.905368203716449e-05, "loss": 1.2249, "step": 11000 }, { "epoch": 1.9012388162422575, "grad_norm": 9.764015197753906, "learning_rate": 4.904938059187887e-05, "loss": 1.2804, "step": 11050 }, { "epoch": 1.9098417068134892, "grad_norm": 6.741626739501953, "learning_rate": 4.904507914659326e-05, "loss": 1.2209, "step": 11100 }, { "epoch": 1.9184445973847213, "grad_norm": 10.178824424743652, "learning_rate": 4.904077770130764e-05, "loss": 1.2391, "step": 11150 }, { "epoch": 1.9270474879559532, "grad_norm": 10.501123428344727, "learning_rate": 4.903647625602202e-05, "loss": 1.2006, "step": 11200 }, { "epoch": 1.935650378527185, "grad_norm": 12.410333633422852, "learning_rate": 4.903217481073641e-05, "loss": 1.1914, "step": 11250 }, { "epoch": 1.9442532690984171, "grad_norm": 8.225683212280273, "learning_rate": 4.902787336545079e-05, "loss": 1.2342, "step": 11300 }, { "epoch": 1.952856159669649, "grad_norm": 15.866887092590332, "learning_rate": 4.902357192016518e-05, "loss": 1.1684, "step": 11350 }, { "epoch": 1.9614590502408809, "grad_norm": 12.385305404663086, "learning_rate": 4.901927047487957e-05, "loss": 1.227, "step": 11400 }, { "epoch": 1.970061940812113, "grad_norm": 14.774800300598145, "learning_rate": 4.901496902959394e-05, "loss": 1.2445, "step": 11450 }, { "epoch": 1.9786648313833448, "grad_norm": 7.357580184936523, "learning_rate": 4.901066758430833e-05, "loss": 1.2444, "step": 11500 }, { "epoch": 1.9872677219545767, "grad_norm": 10.391231536865234, "learning_rate": 4.900636613902271e-05, "loss": 1.1746, "step": 11550 }, { "epoch": 1.9958706125258088, "grad_norm": 6.618278503417969, "learning_rate": 4.90020646937371e-05, "loss": 1.2046, "step": 11600 }, { "epoch": 2.0, "eval_accuracy": 0.45250387196695924, "eval_f1": 0.41124147375341674, "eval_f1_DuraRiadoRio_16x16": 0.4296074018504626, "eval_f1_Mole_16x16": 0.18077803203661327, "eval_f1_Quebrado_16x16": 0.6089679715302491, "eval_f1_RiadoRio_16x16": 0.4397188682645522, "eval_f1_RioFechado_16x16": 0.3971350950852062, "eval_loss": 1.4499568939208984, "eval_precision": 0.5101064889383735, "eval_precision_DuraRiadoRio_16x16": 0.5067846607669616, "eval_precision_Mole_16x16": 0.7452830188679245, "eval_precision_Quebrado_16x16": 0.4530819741580174, "eval_precision_RiadoRio_16x16": 0.3914019890920757, "eval_precision_RioFechado_16x16": 0.45398080180688877, "eval_recall": 0.45173315488476096, "eval_recall_DuraRiadoRio_16x16": 0.3728298611111111, "eval_recall_Mole_16x16": 0.10286458333333333, "eval_recall_Quebrado_16x16": 0.9283854166666666, "eval_recall_RiadoRio_16x16": 0.5016447368421053, "eval_recall_RioFechado_16x16": 0.35294117647058826, "eval_runtime": 43.9797, "eval_samples_per_second": 264.258, "eval_steps_per_second": 16.53, "step": 11624 }, { "epoch": 2.0044735030970404, "grad_norm": 9.90128231048584, "learning_rate": 4.8997763248451486e-05, "loss": 1.2428, "step": 11650 }, { "epoch": 2.0130763936682725, "grad_norm": 7.467269420623779, "learning_rate": 4.899346180316587e-05, "loss": 1.1699, "step": 11700 }, { "epoch": 2.0216792842395046, "grad_norm": 14.753714561462402, "learning_rate": 4.898916035788025e-05, "loss": 1.2478, "step": 11750 }, { "epoch": 2.0302821748107363, "grad_norm": 7.708162307739258, "learning_rate": 4.898485891259464e-05, "loss": 1.2202, "step": 11800 }, { "epoch": 2.0388850653819683, "grad_norm": 13.251511573791504, "learning_rate": 4.898055746730902e-05, "loss": 1.238, "step": 11850 }, { "epoch": 2.0474879559532004, "grad_norm": 10.625846862792969, "learning_rate": 4.8976256022023406e-05, "loss": 1.228, "step": 11900 }, { "epoch": 2.056090846524432, "grad_norm": 8.894543647766113, "learning_rate": 4.897195457673779e-05, "loss": 1.1949, "step": 11950 }, { "epoch": 2.064693737095664, "grad_norm": 22.649932861328125, "learning_rate": 4.896765313145217e-05, "loss": 1.1932, "step": 12000 }, { "epoch": 2.0732966276668963, "grad_norm": 6.768689155578613, "learning_rate": 4.8963351686166556e-05, "loss": 1.2601, "step": 12050 }, { "epoch": 2.081899518238128, "grad_norm": 7.9566545486450195, "learning_rate": 4.895905024088094e-05, "loss": 1.2395, "step": 12100 }, { "epoch": 2.09050240880936, "grad_norm": 7.108730792999268, "learning_rate": 4.8954748795595325e-05, "loss": 1.212, "step": 12150 }, { "epoch": 2.099105299380592, "grad_norm": 12.010674476623535, "learning_rate": 4.8950447350309706e-05, "loss": 1.2019, "step": 12200 }, { "epoch": 2.1077081899518237, "grad_norm": 11.981115341186523, "learning_rate": 4.894614590502409e-05, "loss": 1.2361, "step": 12250 }, { "epoch": 2.116311080523056, "grad_norm": 11.860133171081543, "learning_rate": 4.8941844459738476e-05, "loss": 1.2031, "step": 12300 }, { "epoch": 2.1249139710942875, "grad_norm": 10.547916412353516, "learning_rate": 4.893754301445286e-05, "loss": 1.1851, "step": 12350 }, { "epoch": 2.1335168616655196, "grad_norm": 8.951492309570312, "learning_rate": 4.8933241569167245e-05, "loss": 1.2318, "step": 12400 }, { "epoch": 2.1421197522367517, "grad_norm": 8.861294746398926, "learning_rate": 4.8928940123881626e-05, "loss": 1.2046, "step": 12450 }, { "epoch": 2.1507226428079833, "grad_norm": 10.411590576171875, "learning_rate": 4.8924638678596014e-05, "loss": 1.1884, "step": 12500 }, { "epoch": 2.1593255333792154, "grad_norm": 23.359643936157227, "learning_rate": 4.8920337233310395e-05, "loss": 1.1704, "step": 12550 }, { "epoch": 2.1679284239504475, "grad_norm": 9.794258117675781, "learning_rate": 4.8916035788024776e-05, "loss": 1.2388, "step": 12600 }, { "epoch": 2.176531314521679, "grad_norm": 10.289375305175781, "learning_rate": 4.8911734342739164e-05, "loss": 1.2278, "step": 12650 }, { "epoch": 2.1851342050929112, "grad_norm": 10.67758560180664, "learning_rate": 4.8907432897453545e-05, "loss": 1.1624, "step": 12700 }, { "epoch": 2.1937370956641433, "grad_norm": 8.621283531188965, "learning_rate": 4.890313145216793e-05, "loss": 1.1965, "step": 12750 }, { "epoch": 2.202339986235375, "grad_norm": 7.182273864746094, "learning_rate": 4.8898830006882314e-05, "loss": 1.2305, "step": 12800 }, { "epoch": 2.210942876806607, "grad_norm": 9.711100578308105, "learning_rate": 4.8894528561596696e-05, "loss": 1.1906, "step": 12850 }, { "epoch": 2.219545767377839, "grad_norm": 10.993975639343262, "learning_rate": 4.8890227116311083e-05, "loss": 1.2012, "step": 12900 }, { "epoch": 2.228148657949071, "grad_norm": 7.881711006164551, "learning_rate": 4.8885925671025465e-05, "loss": 1.2498, "step": 12950 }, { "epoch": 2.236751548520303, "grad_norm": 8.517545700073242, "learning_rate": 4.888162422573985e-05, "loss": 1.1605, "step": 13000 }, { "epoch": 2.2453544390915345, "grad_norm": 10.774834632873535, "learning_rate": 4.887732278045424e-05, "loss": 1.1687, "step": 13050 }, { "epoch": 2.2539573296627666, "grad_norm": 9.796731948852539, "learning_rate": 4.8873021335168615e-05, "loss": 1.1972, "step": 13100 }, { "epoch": 2.2625602202339987, "grad_norm": 8.634834289550781, "learning_rate": 4.8868719889883e-05, "loss": 1.1747, "step": 13150 }, { "epoch": 2.2711631108052304, "grad_norm": 12.028709411621094, "learning_rate": 4.8864418444597384e-05, "loss": 1.2334, "step": 13200 }, { "epoch": 2.2797660013764625, "grad_norm": 15.203660011291504, "learning_rate": 4.886011699931177e-05, "loss": 1.1555, "step": 13250 }, { "epoch": 2.2883688919476946, "grad_norm": 12.072103500366211, "learning_rate": 4.885581555402616e-05, "loss": 1.1194, "step": 13300 }, { "epoch": 2.296971782518926, "grad_norm": 7.439048767089844, "learning_rate": 4.8851514108740534e-05, "loss": 1.2094, "step": 13350 }, { "epoch": 2.3055746730901583, "grad_norm": 11.489731788635254, "learning_rate": 4.884721266345492e-05, "loss": 1.183, "step": 13400 }, { "epoch": 2.3141775636613904, "grad_norm": 12.077587127685547, "learning_rate": 4.884291121816931e-05, "loss": 1.2455, "step": 13450 }, { "epoch": 2.322780454232622, "grad_norm": 8.671648979187012, "learning_rate": 4.883860977288369e-05, "loss": 1.1679, "step": 13500 }, { "epoch": 2.331383344803854, "grad_norm": 12.40151596069336, "learning_rate": 4.883430832759808e-05, "loss": 1.174, "step": 13550 }, { "epoch": 2.339986235375086, "grad_norm": 9.201745986938477, "learning_rate": 4.8830006882312454e-05, "loss": 1.2174, "step": 13600 }, { "epoch": 2.348589125946318, "grad_norm": 9.285476684570312, "learning_rate": 4.882570543702684e-05, "loss": 1.1953, "step": 13650 }, { "epoch": 2.35719201651755, "grad_norm": 7.375884532928467, "learning_rate": 4.882140399174123e-05, "loss": 1.2523, "step": 13700 }, { "epoch": 2.365794907088782, "grad_norm": 9.531774520874023, "learning_rate": 4.881710254645561e-05, "loss": 1.2128, "step": 13750 }, { "epoch": 2.3743977976600137, "grad_norm": 10.347550392150879, "learning_rate": 4.881280110117e-05, "loss": 1.2161, "step": 13800 }, { "epoch": 2.383000688231246, "grad_norm": 11.81945514678955, "learning_rate": 4.880849965588438e-05, "loss": 1.1574, "step": 13850 }, { "epoch": 2.391603578802478, "grad_norm": 10.93974494934082, "learning_rate": 4.880419821059876e-05, "loss": 1.2063, "step": 13900 }, { "epoch": 2.4002064693737095, "grad_norm": 13.004681587219238, "learning_rate": 4.879989676531315e-05, "loss": 1.1783, "step": 13950 }, { "epoch": 2.4088093599449416, "grad_norm": 11.95052433013916, "learning_rate": 4.879559532002753e-05, "loss": 1.1866, "step": 14000 }, { "epoch": 2.4174122505161733, "grad_norm": 10.22706413269043, "learning_rate": 4.879129387474192e-05, "loss": 1.1904, "step": 14050 }, { "epoch": 2.4260151410874053, "grad_norm": 8.098734855651855, "learning_rate": 4.87869924294563e-05, "loss": 1.1951, "step": 14100 }, { "epoch": 2.4346180316586374, "grad_norm": 10.463532447814941, "learning_rate": 4.878269098417068e-05, "loss": 1.1914, "step": 14150 }, { "epoch": 2.443220922229869, "grad_norm": 12.285577774047852, "learning_rate": 4.877838953888507e-05, "loss": 1.1624, "step": 14200 }, { "epoch": 2.451823812801101, "grad_norm": 7.861328601837158, "learning_rate": 4.877408809359945e-05, "loss": 1.1802, "step": 14250 }, { "epoch": 2.4604267033723333, "grad_norm": 7.734827995300293, "learning_rate": 4.876978664831384e-05, "loss": 1.1635, "step": 14300 }, { "epoch": 2.469029593943565, "grad_norm": 12.312912940979004, "learning_rate": 4.876548520302822e-05, "loss": 1.1833, "step": 14350 }, { "epoch": 2.477632484514797, "grad_norm": 8.662589073181152, "learning_rate": 4.87611837577426e-05, "loss": 1.1922, "step": 14400 }, { "epoch": 2.486235375086029, "grad_norm": 9.126067161560059, "learning_rate": 4.875688231245699e-05, "loss": 1.1939, "step": 14450 }, { "epoch": 2.4948382656572607, "grad_norm": 8.491592407226562, "learning_rate": 4.875258086717137e-05, "loss": 1.22, "step": 14500 }, { "epoch": 2.503441156228493, "grad_norm": 7.438276290893555, "learning_rate": 4.874827942188576e-05, "loss": 1.1709, "step": 14550 }, { "epoch": 2.5120440467997245, "grad_norm": 10.518980026245117, "learning_rate": 4.874397797660014e-05, "loss": 1.1826, "step": 14600 }, { "epoch": 2.5206469373709566, "grad_norm": 10.452248573303223, "learning_rate": 4.8739676531314526e-05, "loss": 1.1964, "step": 14650 }, { "epoch": 2.5292498279421887, "grad_norm": 7.407390117645264, "learning_rate": 4.873537508602891e-05, "loss": 1.1818, "step": 14700 }, { "epoch": 2.5378527185134203, "grad_norm": 10.052775382995605, "learning_rate": 4.873107364074329e-05, "loss": 1.2173, "step": 14750 }, { "epoch": 2.5464556090846524, "grad_norm": 8.482536315917969, "learning_rate": 4.8726772195457676e-05, "loss": 1.2112, "step": 14800 }, { "epoch": 2.5550584996558845, "grad_norm": 8.581706047058105, "learning_rate": 4.8722470750172064e-05, "loss": 1.1907, "step": 14850 }, { "epoch": 2.563661390227116, "grad_norm": 11.216682434082031, "learning_rate": 4.8718169304886446e-05, "loss": 1.2213, "step": 14900 }, { "epoch": 2.5722642807983482, "grad_norm": 8.719677925109863, "learning_rate": 4.871386785960083e-05, "loss": 1.2067, "step": 14950 }, { "epoch": 2.5808671713695803, "grad_norm": 8.045714378356934, "learning_rate": 4.870956641431521e-05, "loss": 1.1634, "step": 15000 }, { "epoch": 2.589470061940812, "grad_norm": 7.104714870452881, "learning_rate": 4.8705264969029596e-05, "loss": 1.165, "step": 15050 }, { "epoch": 2.598072952512044, "grad_norm": 12.666434288024902, "learning_rate": 4.8700963523743984e-05, "loss": 1.1605, "step": 15100 }, { "epoch": 2.606675843083276, "grad_norm": 8.917975425720215, "learning_rate": 4.8696662078458365e-05, "loss": 1.1859, "step": 15150 }, { "epoch": 2.615278733654508, "grad_norm": 8.190360069274902, "learning_rate": 4.869236063317275e-05, "loss": 1.1334, "step": 15200 }, { "epoch": 2.62388162422574, "grad_norm": 10.583536148071289, "learning_rate": 4.868805918788713e-05, "loss": 1.183, "step": 15250 }, { "epoch": 2.632484514796972, "grad_norm": 10.741872787475586, "learning_rate": 4.8683757742601515e-05, "loss": 1.2187, "step": 15300 }, { "epoch": 2.6410874053682036, "grad_norm": 9.788289070129395, "learning_rate": 4.86794562973159e-05, "loss": 1.196, "step": 15350 }, { "epoch": 2.6496902959394357, "grad_norm": 7.5520429611206055, "learning_rate": 4.8675154852030284e-05, "loss": 1.1885, "step": 15400 }, { "epoch": 2.658293186510668, "grad_norm": 7.831338882446289, "learning_rate": 4.867085340674467e-05, "loss": 1.1769, "step": 15450 }, { "epoch": 2.6668960770818995, "grad_norm": 11.354333877563477, "learning_rate": 4.8666551961459053e-05, "loss": 1.1847, "step": 15500 }, { "epoch": 2.6754989676531316, "grad_norm": 9.93191909790039, "learning_rate": 4.8662250516173435e-05, "loss": 1.1948, "step": 15550 }, { "epoch": 2.6841018582243636, "grad_norm": 8.438941955566406, "learning_rate": 4.865794907088782e-05, "loss": 1.1561, "step": 15600 }, { "epoch": 2.6927047487955953, "grad_norm": 11.943193435668945, "learning_rate": 4.8653647625602204e-05, "loss": 1.2152, "step": 15650 }, { "epoch": 2.7013076393668274, "grad_norm": 10.366228103637695, "learning_rate": 4.864934618031659e-05, "loss": 1.178, "step": 15700 }, { "epoch": 2.709910529938059, "grad_norm": 10.72542953491211, "learning_rate": 4.864504473503097e-05, "loss": 1.2019, "step": 15750 }, { "epoch": 2.718513420509291, "grad_norm": 9.819501876831055, "learning_rate": 4.8640743289745354e-05, "loss": 1.1758, "step": 15800 }, { "epoch": 2.727116311080523, "grad_norm": 7.744460582733154, "learning_rate": 4.863644184445974e-05, "loss": 1.1798, "step": 15850 }, { "epoch": 2.735719201651755, "grad_norm": 10.292325973510742, "learning_rate": 4.863214039917412e-05, "loss": 1.1809, "step": 15900 }, { "epoch": 2.744322092222987, "grad_norm": 6.738109588623047, "learning_rate": 4.862783895388851e-05, "loss": 1.1542, "step": 15950 }, { "epoch": 2.7529249827942186, "grad_norm": 8.823808670043945, "learning_rate": 4.862353750860289e-05, "loss": 1.1174, "step": 16000 }, { "epoch": 2.7615278733654507, "grad_norm": 9.813061714172363, "learning_rate": 4.8619236063317273e-05, "loss": 1.1871, "step": 16050 }, { "epoch": 2.770130763936683, "grad_norm": 7.0893988609313965, "learning_rate": 4.861493461803166e-05, "loss": 1.2091, "step": 16100 }, { "epoch": 2.7787336545079144, "grad_norm": 7.606479644775391, "learning_rate": 4.861063317274604e-05, "loss": 1.1659, "step": 16150 }, { "epoch": 2.7873365450791465, "grad_norm": 6.882772922515869, "learning_rate": 4.860633172746043e-05, "loss": 1.1545, "step": 16200 }, { "epoch": 2.7959394356503786, "grad_norm": 12.652734756469727, "learning_rate": 4.860203028217482e-05, "loss": 1.1627, "step": 16250 }, { "epoch": 2.8045423262216103, "grad_norm": 9.29095458984375, "learning_rate": 4.859772883688919e-05, "loss": 1.1365, "step": 16300 }, { "epoch": 2.8131452167928424, "grad_norm": 11.364659309387207, "learning_rate": 4.859342739160358e-05, "loss": 1.1955, "step": 16350 }, { "epoch": 2.8217481073640744, "grad_norm": 8.329474449157715, "learning_rate": 4.858912594631796e-05, "loss": 1.1422, "step": 16400 }, { "epoch": 2.830350997935306, "grad_norm": 11.45989990234375, "learning_rate": 4.858482450103235e-05, "loss": 1.1689, "step": 16450 }, { "epoch": 2.838953888506538, "grad_norm": 10.32991886138916, "learning_rate": 4.858052305574674e-05, "loss": 1.1583, "step": 16500 }, { "epoch": 2.8475567790777703, "grad_norm": 7.74367618560791, "learning_rate": 4.857622161046111e-05, "loss": 1.21, "step": 16550 }, { "epoch": 2.856159669649002, "grad_norm": 8.507062911987305, "learning_rate": 4.85719201651755e-05, "loss": 1.1764, "step": 16600 }, { "epoch": 2.864762560220234, "grad_norm": 10.237994194030762, "learning_rate": 4.856761871988988e-05, "loss": 1.2159, "step": 16650 }, { "epoch": 2.873365450791466, "grad_norm": 10.500870704650879, "learning_rate": 4.856331727460427e-05, "loss": 1.1578, "step": 16700 }, { "epoch": 2.8819683413626977, "grad_norm": 7.6299147605896, "learning_rate": 4.855901582931866e-05, "loss": 1.1754, "step": 16750 }, { "epoch": 2.89057123193393, "grad_norm": 8.785295486450195, "learning_rate": 4.855471438403304e-05, "loss": 1.1609, "step": 16800 }, { "epoch": 2.899174122505162, "grad_norm": 9.508923530578613, "learning_rate": 4.855041293874742e-05, "loss": 1.1863, "step": 16850 }, { "epoch": 2.9077770130763936, "grad_norm": 8.817083358764648, "learning_rate": 4.854611149346181e-05, "loss": 1.2111, "step": 16900 }, { "epoch": 2.9163799036476257, "grad_norm": 11.38044548034668, "learning_rate": 4.854181004817619e-05, "loss": 1.1941, "step": 16950 }, { "epoch": 2.9249827942188578, "grad_norm": 10.149999618530273, "learning_rate": 4.853750860289058e-05, "loss": 1.1947, "step": 17000 }, { "epoch": 2.9335856847900894, "grad_norm": 10.993596076965332, "learning_rate": 4.853320715760496e-05, "loss": 1.2319, "step": 17050 }, { "epoch": 2.9421885753613215, "grad_norm": 9.793594360351562, "learning_rate": 4.852890571231934e-05, "loss": 1.1696, "step": 17100 }, { "epoch": 2.9507914659325536, "grad_norm": 9.54121208190918, "learning_rate": 4.852460426703373e-05, "loss": 1.1789, "step": 17150 }, { "epoch": 2.9593943565037852, "grad_norm": 9.454851150512695, "learning_rate": 4.852030282174811e-05, "loss": 1.1706, "step": 17200 }, { "epoch": 2.9679972470750173, "grad_norm": 15.321932792663574, "learning_rate": 4.8516001376462496e-05, "loss": 1.182, "step": 17250 }, { "epoch": 2.976600137646249, "grad_norm": 9.780303955078125, "learning_rate": 4.851169993117688e-05, "loss": 1.2448, "step": 17300 }, { "epoch": 2.985203028217481, "grad_norm": 8.188822746276855, "learning_rate": 4.8507398485891265e-05, "loss": 1.1379, "step": 17350 }, { "epoch": 2.993805918788713, "grad_norm": 10.298974990844727, "learning_rate": 4.8503097040605646e-05, "loss": 1.1274, "step": 17400 }, { "epoch": 3.0, "eval_accuracy": 0.5297711237308552, "eval_f1": 0.5182019607181394, "eval_f1_DuraRiadoRio_16x16": 0.4614599285349668, "eval_f1_Mole_16x16": 0.5296356715606307, "eval_f1_Quebrado_16x16": 0.6322133494048516, "eval_f1_RiadoRio_16x16": 0.46561797752808987, "eval_f1_RioFechado_16x16": 0.5020828765621574, "eval_loss": 1.2716028690338135, "eval_precision": 0.5536164126683033, "eval_precision_DuraRiadoRio_16x16": 0.5600991325898389, "eval_precision_Mole_16x16": 0.7088791848617176, "eval_precision_Quebrado_16x16": 0.4841910916224325, "eval_precision_RiadoRio_16x16": 0.5133795837462835, "eval_precision_RioFechado_16x16": 0.501533070521244, "eval_recall": 0.5308630351852708, "eval_recall_DuraRiadoRio_16x16": 0.3923611111111111, "eval_recall_Mole_16x16": 0.4227430555555556, "eval_recall_Quebrado_16x16": 0.9105902777777778, "eval_recall_RiadoRio_16x16": 0.42598684210526316, "eval_recall_RioFechado_16x16": 0.5026338893766462, "eval_runtime": 43.7001, "eval_samples_per_second": 265.949, "eval_steps_per_second": 16.636, "step": 17436 }, { "epoch": 3.002408809359945, "grad_norm": 6.4393134117126465, "learning_rate": 4.849879559532003e-05, "loss": 1.2116, "step": 17450 }, { "epoch": 3.011011699931177, "grad_norm": 13.954197883605957, "learning_rate": 4.8494494150034416e-05, "loss": 1.1748, "step": 17500 }, { "epoch": 3.019614590502409, "grad_norm": 7.564427375793457, "learning_rate": 4.84901927047488e-05, "loss": 1.174, "step": 17550 }, { "epoch": 3.0282174810736406, "grad_norm": 14.441926002502441, "learning_rate": 4.8485891259463185e-05, "loss": 1.1941, "step": 17600 }, { "epoch": 3.0368203716448727, "grad_norm": 7.774544715881348, "learning_rate": 4.8481589814177566e-05, "loss": 1.1586, "step": 17650 }, { "epoch": 3.045423262216105, "grad_norm": 9.607131958007812, "learning_rate": 4.847728836889195e-05, "loss": 1.1671, "step": 17700 }, { "epoch": 3.0540261527873365, "grad_norm": 13.690775871276855, "learning_rate": 4.8472986923606335e-05, "loss": 1.1666, "step": 17750 }, { "epoch": 3.0626290433585686, "grad_norm": 14.38357925415039, "learning_rate": 4.8468685478320716e-05, "loss": 1.1591, "step": 17800 }, { "epoch": 3.0712319339298, "grad_norm": 7.010596752166748, "learning_rate": 4.8464384033035104e-05, "loss": 1.1738, "step": 17850 }, { "epoch": 3.0798348245010323, "grad_norm": 8.213446617126465, "learning_rate": 4.8460082587749485e-05, "loss": 1.12, "step": 17900 }, { "epoch": 3.0884377150722644, "grad_norm": 12.153669357299805, "learning_rate": 4.8455781142463866e-05, "loss": 1.1745, "step": 17950 }, { "epoch": 3.097040605643496, "grad_norm": 12.54633617401123, "learning_rate": 4.8451479697178254e-05, "loss": 1.13, "step": 18000 }, { "epoch": 3.105643496214728, "grad_norm": 9.615897178649902, "learning_rate": 4.8447178251892636e-05, "loss": 1.1883, "step": 18050 }, { "epoch": 3.11424638678596, "grad_norm": 10.751274108886719, "learning_rate": 4.8442876806607024e-05, "loss": 1.2019, "step": 18100 }, { "epoch": 3.122849277357192, "grad_norm": 7.151867866516113, "learning_rate": 4.843857536132141e-05, "loss": 1.1784, "step": 18150 }, { "epoch": 3.131452167928424, "grad_norm": 11.280678749084473, "learning_rate": 4.8434273916035786e-05, "loss": 1.1881, "step": 18200 }, { "epoch": 3.140055058499656, "grad_norm": 14.192291259765625, "learning_rate": 4.8429972470750174e-05, "loss": 1.1673, "step": 18250 }, { "epoch": 3.1486579490708877, "grad_norm": 7.704963684082031, "learning_rate": 4.842567102546456e-05, "loss": 1.1819, "step": 18300 }, { "epoch": 3.15726083964212, "grad_norm": 7.99170446395874, "learning_rate": 4.842136958017894e-05, "loss": 1.0945, "step": 18350 }, { "epoch": 3.165863730213352, "grad_norm": 11.118195533752441, "learning_rate": 4.841706813489333e-05, "loss": 1.1567, "step": 18400 }, { "epoch": 3.1744666207845835, "grad_norm": 11.197527885437012, "learning_rate": 4.8412766689607705e-05, "loss": 1.1624, "step": 18450 }, { "epoch": 3.1830695113558156, "grad_norm": 7.612173080444336, "learning_rate": 4.840846524432209e-05, "loss": 1.1289, "step": 18500 }, { "epoch": 3.1916724019270477, "grad_norm": 6.330729961395264, "learning_rate": 4.840416379903648e-05, "loss": 1.1351, "step": 18550 }, { "epoch": 3.2002752924982794, "grad_norm": 14.436402320861816, "learning_rate": 4.839986235375086e-05, "loss": 1.1715, "step": 18600 }, { "epoch": 3.2088781830695114, "grad_norm": 13.629887580871582, "learning_rate": 4.839556090846525e-05, "loss": 1.1569, "step": 18650 }, { "epoch": 3.217481073640743, "grad_norm": 11.088851928710938, "learning_rate": 4.839125946317963e-05, "loss": 1.1787, "step": 18700 }, { "epoch": 3.226083964211975, "grad_norm": 7.983802318572998, "learning_rate": 4.838695801789401e-05, "loss": 1.1962, "step": 18750 }, { "epoch": 3.2346868547832073, "grad_norm": 8.49528980255127, "learning_rate": 4.83826565726084e-05, "loss": 1.1812, "step": 18800 }, { "epoch": 3.243289745354439, "grad_norm": 10.093893051147461, "learning_rate": 4.837835512732278e-05, "loss": 1.1924, "step": 18850 }, { "epoch": 3.251892635925671, "grad_norm": 8.878253936767578, "learning_rate": 4.837405368203717e-05, "loss": 1.1648, "step": 18900 }, { "epoch": 3.260495526496903, "grad_norm": 11.365337371826172, "learning_rate": 4.836975223675155e-05, "loss": 1.1976, "step": 18950 }, { "epoch": 3.2690984170681348, "grad_norm": 9.513070106506348, "learning_rate": 4.836545079146593e-05, "loss": 1.2366, "step": 19000 }, { "epoch": 3.277701307639367, "grad_norm": 7.4674787521362305, "learning_rate": 4.836114934618032e-05, "loss": 1.156, "step": 19050 }, { "epoch": 3.286304198210599, "grad_norm": 12.183306694030762, "learning_rate": 4.83568479008947e-05, "loss": 1.1643, "step": 19100 }, { "epoch": 3.2949070887818306, "grad_norm": 6.3620285987854, "learning_rate": 4.835254645560909e-05, "loss": 1.184, "step": 19150 }, { "epoch": 3.3035099793530627, "grad_norm": 10.080217361450195, "learning_rate": 4.834824501032347e-05, "loss": 1.1409, "step": 19200 }, { "epoch": 3.3121128699242943, "grad_norm": 6.786382675170898, "learning_rate": 4.834394356503785e-05, "loss": 1.146, "step": 19250 }, { "epoch": 3.3207157604955264, "grad_norm": 8.675495147705078, "learning_rate": 4.833964211975224e-05, "loss": 1.1251, "step": 19300 }, { "epoch": 3.3293186510667585, "grad_norm": 15.38300609588623, "learning_rate": 4.833534067446662e-05, "loss": 1.1735, "step": 19350 }, { "epoch": 3.33792154163799, "grad_norm": 9.91826057434082, "learning_rate": 4.833103922918101e-05, "loss": 1.164, "step": 19400 }, { "epoch": 3.3465244322092222, "grad_norm": 7.503796100616455, "learning_rate": 4.832673778389539e-05, "loss": 1.1183, "step": 19450 }, { "epoch": 3.3551273227804543, "grad_norm": 9.662027359008789, "learning_rate": 4.832243633860978e-05, "loss": 1.1463, "step": 19500 }, { "epoch": 3.363730213351686, "grad_norm": 6.198797702789307, "learning_rate": 4.831813489332416e-05, "loss": 1.1733, "step": 19550 }, { "epoch": 3.372333103922918, "grad_norm": 10.325329780578613, "learning_rate": 4.831383344803854e-05, "loss": 1.1427, "step": 19600 }, { "epoch": 3.38093599449415, "grad_norm": 13.664487838745117, "learning_rate": 4.830953200275293e-05, "loss": 1.1609, "step": 19650 }, { "epoch": 3.389538885065382, "grad_norm": 9.900769233703613, "learning_rate": 4.8305230557467316e-05, "loss": 1.1511, "step": 19700 }, { "epoch": 3.398141775636614, "grad_norm": 9.172723770141602, "learning_rate": 4.83009291121817e-05, "loss": 1.155, "step": 19750 }, { "epoch": 3.406744666207846, "grad_norm": 13.414734840393066, "learning_rate": 4.829662766689608e-05, "loss": 1.1442, "step": 19800 }, { "epoch": 3.4153475567790776, "grad_norm": 8.490633964538574, "learning_rate": 4.829232622161046e-05, "loss": 1.1579, "step": 19850 }, { "epoch": 3.4239504473503097, "grad_norm": 8.174596786499023, "learning_rate": 4.828802477632485e-05, "loss": 1.2031, "step": 19900 }, { "epoch": 3.432553337921542, "grad_norm": 8.734251022338867, "learning_rate": 4.8283723331039235e-05, "loss": 1.2054, "step": 19950 }, { "epoch": 3.4411562284927735, "grad_norm": 6.56998872756958, "learning_rate": 4.8279421885753616e-05, "loss": 1.1839, "step": 20000 }, { "epoch": 3.4497591190640056, "grad_norm": 8.091642379760742, "learning_rate": 4.8275120440468e-05, "loss": 1.1454, "step": 20050 }, { "epoch": 3.4583620096352377, "grad_norm": 9.90014934539795, "learning_rate": 4.827081899518238e-05, "loss": 1.1267, "step": 20100 }, { "epoch": 3.4669649002064693, "grad_norm": 9.621966361999512, "learning_rate": 4.826651754989677e-05, "loss": 1.1931, "step": 20150 }, { "epoch": 3.4755677907777014, "grad_norm": 13.209905624389648, "learning_rate": 4.8262216104611155e-05, "loss": 1.1524, "step": 20200 }, { "epoch": 3.4841706813489335, "grad_norm": 8.266006469726562, "learning_rate": 4.8257914659325536e-05, "loss": 1.2345, "step": 20250 }, { "epoch": 3.492773571920165, "grad_norm": 11.675355911254883, "learning_rate": 4.8253613214039924e-05, "loss": 1.1361, "step": 20300 }, { "epoch": 3.5013764624913972, "grad_norm": 20.15110969543457, "learning_rate": 4.8249311768754305e-05, "loss": 1.1796, "step": 20350 }, { "epoch": 3.5099793530626293, "grad_norm": 11.621223449707031, "learning_rate": 4.8245010323468686e-05, "loss": 1.1466, "step": 20400 }, { "epoch": 3.518582243633861, "grad_norm": 8.29684066772461, "learning_rate": 4.8240708878183074e-05, "loss": 1.1748, "step": 20450 }, { "epoch": 3.527185134205093, "grad_norm": 13.84134578704834, "learning_rate": 4.8236407432897455e-05, "loss": 1.1379, "step": 20500 }, { "epoch": 3.5357880247763247, "grad_norm": 9.757213592529297, "learning_rate": 4.823210598761184e-05, "loss": 1.2056, "step": 20550 }, { "epoch": 3.544390915347557, "grad_norm": 7.77637243270874, "learning_rate": 4.8227804542326224e-05, "loss": 1.1542, "step": 20600 }, { "epoch": 3.5529938059187884, "grad_norm": 8.835380554199219, "learning_rate": 4.8223503097040606e-05, "loss": 1.1293, "step": 20650 }, { "epoch": 3.5615966964900205, "grad_norm": 9.331871032714844, "learning_rate": 4.8219201651754994e-05, "loss": 1.2057, "step": 20700 }, { "epoch": 3.5701995870612526, "grad_norm": 6.393391132354736, "learning_rate": 4.8214900206469375e-05, "loss": 1.1484, "step": 20750 }, { "epoch": 3.5788024776324843, "grad_norm": 8.29530143737793, "learning_rate": 4.821059876118376e-05, "loss": 1.1164, "step": 20800 }, { "epoch": 3.5874053682037164, "grad_norm": 8.601723670959473, "learning_rate": 4.8206297315898144e-05, "loss": 1.1353, "step": 20850 }, { "epoch": 3.5960082587749485, "grad_norm": 13.755023956298828, "learning_rate": 4.8201995870612525e-05, "loss": 1.1036, "step": 20900 }, { "epoch": 3.60461114934618, "grad_norm": 10.134452819824219, "learning_rate": 4.819769442532691e-05, "loss": 1.1921, "step": 20950 }, { "epoch": 3.613214039917412, "grad_norm": 9.995678901672363, "learning_rate": 4.8193392980041294e-05, "loss": 1.1544, "step": 21000 }, { "epoch": 3.6218169304886443, "grad_norm": 7.881764888763428, "learning_rate": 4.818909153475568e-05, "loss": 1.2083, "step": 21050 }, { "epoch": 3.630419821059876, "grad_norm": 8.395647048950195, "learning_rate": 4.818479008947006e-05, "loss": 1.1475, "step": 21100 }, { "epoch": 3.639022711631108, "grad_norm": 9.572067260742188, "learning_rate": 4.8180488644184444e-05, "loss": 1.1898, "step": 21150 }, { "epoch": 3.64762560220234, "grad_norm": 7.93234920501709, "learning_rate": 4.817618719889883e-05, "loss": 1.1305, "step": 21200 }, { "epoch": 3.6562284927735718, "grad_norm": 8.246980667114258, "learning_rate": 4.8171885753613214e-05, "loss": 1.1355, "step": 21250 }, { "epoch": 3.664831383344804, "grad_norm": 8.441601753234863, "learning_rate": 4.81675843083276e-05, "loss": 1.2023, "step": 21300 }, { "epoch": 3.673434273916036, "grad_norm": 7.168173789978027, "learning_rate": 4.816328286304199e-05, "loss": 1.1855, "step": 21350 }, { "epoch": 3.6820371644872676, "grad_norm": 8.349146842956543, "learning_rate": 4.8158981417756364e-05, "loss": 1.1822, "step": 21400 }, { "epoch": 3.6906400550584997, "grad_norm": 7.697701930999756, "learning_rate": 4.815467997247075e-05, "loss": 1.1059, "step": 21450 }, { "epoch": 3.6992429456297318, "grad_norm": 12.39500904083252, "learning_rate": 4.815037852718513e-05, "loss": 1.1255, "step": 21500 }, { "epoch": 3.7078458362009634, "grad_norm": 10.624629020690918, "learning_rate": 4.814607708189952e-05, "loss": 1.1391, "step": 21550 }, { "epoch": 3.7164487267721955, "grad_norm": 11.993627548217773, "learning_rate": 4.814177563661391e-05, "loss": 1.1807, "step": 21600 }, { "epoch": 3.7250516173434276, "grad_norm": 6.292657375335693, "learning_rate": 4.813747419132829e-05, "loss": 1.1356, "step": 21650 }, { "epoch": 3.7336545079146592, "grad_norm": 7.952933311462402, "learning_rate": 4.813317274604267e-05, "loss": 1.1895, "step": 21700 }, { "epoch": 3.7422573984858913, "grad_norm": 13.37735366821289, "learning_rate": 4.812887130075705e-05, "loss": 1.1757, "step": 21750 }, { "epoch": 3.7508602890571234, "grad_norm": 7.6209282875061035, "learning_rate": 4.812456985547144e-05, "loss": 1.1813, "step": 21800 }, { "epoch": 3.759463179628355, "grad_norm": 5.755713939666748, "learning_rate": 4.812026841018583e-05, "loss": 1.1539, "step": 21850 }, { "epoch": 3.768066070199587, "grad_norm": 7.485439777374268, "learning_rate": 4.811596696490021e-05, "loss": 1.1447, "step": 21900 }, { "epoch": 3.7766689607708193, "grad_norm": 8.496212005615234, "learning_rate": 4.811166551961459e-05, "loss": 1.1374, "step": 21950 }, { "epoch": 3.785271851342051, "grad_norm": 7.141066074371338, "learning_rate": 4.810736407432898e-05, "loss": 1.1237, "step": 22000 }, { "epoch": 3.793874741913283, "grad_norm": 8.594643592834473, "learning_rate": 4.810306262904336e-05, "loss": 1.1451, "step": 22050 }, { "epoch": 3.8024776324845146, "grad_norm": 6.973026275634766, "learning_rate": 4.809876118375775e-05, "loss": 1.1381, "step": 22100 }, { "epoch": 3.8110805230557467, "grad_norm": 9.671688079833984, "learning_rate": 4.809445973847213e-05, "loss": 1.1461, "step": 22150 }, { "epoch": 3.819683413626979, "grad_norm": 10.278329849243164, "learning_rate": 4.809015829318651e-05, "loss": 1.1943, "step": 22200 }, { "epoch": 3.8282863041982105, "grad_norm": 7.608288764953613, "learning_rate": 4.80858568479009e-05, "loss": 1.13, "step": 22250 }, { "epoch": 3.8368891947694426, "grad_norm": 9.3580322265625, "learning_rate": 4.808155540261528e-05, "loss": 1.165, "step": 22300 }, { "epoch": 3.845492085340674, "grad_norm": 9.355257987976074, "learning_rate": 4.807725395732967e-05, "loss": 1.1437, "step": 22350 }, { "epoch": 3.8540949759119063, "grad_norm": 11.491572380065918, "learning_rate": 4.807295251204405e-05, "loss": 1.172, "step": 22400 }, { "epoch": 3.8626978664831384, "grad_norm": 10.758626937866211, "learning_rate": 4.8068651066758436e-05, "loss": 1.1002, "step": 22450 }, { "epoch": 3.87130075705437, "grad_norm": 12.047676086425781, "learning_rate": 4.806434962147282e-05, "loss": 1.1575, "step": 22500 }, { "epoch": 3.879903647625602, "grad_norm": 5.910977363586426, "learning_rate": 4.80600481761872e-05, "loss": 1.1926, "step": 22550 }, { "epoch": 3.8885065381968342, "grad_norm": 16.768482208251953, "learning_rate": 4.8055746730901587e-05, "loss": 1.0975, "step": 22600 }, { "epoch": 3.897109428768066, "grad_norm": 10.903653144836426, "learning_rate": 4.805144528561597e-05, "loss": 1.1569, "step": 22650 }, { "epoch": 3.905712319339298, "grad_norm": 10.464058876037598, "learning_rate": 4.8047143840330356e-05, "loss": 1.1052, "step": 22700 }, { "epoch": 3.91431520991053, "grad_norm": 8.284379959106445, "learning_rate": 4.804284239504474e-05, "loss": 1.1739, "step": 22750 }, { "epoch": 3.9229181004817617, "grad_norm": 10.824211120605469, "learning_rate": 4.803854094975912e-05, "loss": 1.1579, "step": 22800 }, { "epoch": 3.931520991052994, "grad_norm": 7.8123779296875, "learning_rate": 4.8034239504473506e-05, "loss": 1.1543, "step": 22850 }, { "epoch": 3.940123881624226, "grad_norm": 7.661468505859375, "learning_rate": 4.802993805918789e-05, "loss": 1.2166, "step": 22900 }, { "epoch": 3.9487267721954575, "grad_norm": 8.579504013061523, "learning_rate": 4.8025636613902275e-05, "loss": 1.1207, "step": 22950 }, { "epoch": 3.9573296627666896, "grad_norm": 11.790451049804688, "learning_rate": 4.802133516861666e-05, "loss": 1.1341, "step": 23000 }, { "epoch": 3.9659325533379217, "grad_norm": 11.202832221984863, "learning_rate": 4.801703372333104e-05, "loss": 1.134, "step": 23050 }, { "epoch": 3.9745354439091534, "grad_norm": 8.850720405578613, "learning_rate": 4.8012732278045425e-05, "loss": 1.1084, "step": 23100 }, { "epoch": 3.9831383344803855, "grad_norm": 10.917728424072266, "learning_rate": 4.8008430832759807e-05, "loss": 1.1833, "step": 23150 }, { "epoch": 3.9917412250516175, "grad_norm": 10.134353637695312, "learning_rate": 4.8004129387474194e-05, "loss": 1.138, "step": 23200 }, { "epoch": 4.0, "eval_accuracy": 0.5511099638616417, "eval_f1": 0.5411148429306765, "eval_f1_DuraRiadoRio_16x16": 0.5052178234222295, "eval_f1_Mole_16x16": 0.5798381162619574, "eval_f1_Quebrado_16x16": 0.7198340874811463, "eval_f1_RiadoRio_16x16": 0.48203404076486656, "eval_f1_RioFechado_16x16": 0.41865014672318224, "eval_loss": 1.1628681421279907, "eval_precision": 0.6036232311946086, "eval_precision_DuraRiadoRio_16x16": 0.4085186177337262, "eval_precision_Mole_16x16": 0.6666666666666666, "eval_precision_Quebrado_16x16": 0.6363333333333333, "eval_precision_RiadoRio_16x16": 0.49290932531155995, "eval_precision_RioFechado_16x16": 0.8136882129277566, "eval_recall": 0.5513853349994096, "eval_recall_DuraRiadoRio_16x16": 0.6618923611111112, "eval_recall_Mole_16x16": 0.5130208333333334, "eval_recall_Quebrado_16x16": 0.8285590277777778, "eval_recall_RiadoRio_16x16": 0.47162828947368424, "eval_recall_RioFechado_16x16": 0.28182616330114135, "eval_runtime": 43.0206, "eval_samples_per_second": 270.15, "eval_steps_per_second": 16.899, "step": 23248 }, { "epoch": 4.000344115622849, "grad_norm": 13.33624267578125, "learning_rate": 4.799982794218858e-05, "loss": 1.1544, "step": 23250 }, { "epoch": 4.008947006194081, "grad_norm": 8.64680004119873, "learning_rate": 4.799552649690296e-05, "loss": 1.1301, "step": 23300 }, { "epoch": 4.017549896765313, "grad_norm": 9.224178314208984, "learning_rate": 4.7991225051617345e-05, "loss": 1.1372, "step": 23350 }, { "epoch": 4.026152787336545, "grad_norm": 7.879807949066162, "learning_rate": 4.798692360633173e-05, "loss": 1.1707, "step": 23400 }, { "epoch": 4.034755677907777, "grad_norm": 14.496654510498047, "learning_rate": 4.7982622161046114e-05, "loss": 1.1462, "step": 23450 }, { "epoch": 4.043358568479009, "grad_norm": 9.417076110839844, "learning_rate": 4.79783207157605e-05, "loss": 1.1151, "step": 23500 }, { "epoch": 4.051961459050241, "grad_norm": 8.171489715576172, "learning_rate": 4.7974019270474876e-05, "loss": 1.1361, "step": 23550 }, { "epoch": 4.0605643496214725, "grad_norm": 5.8768510818481445, "learning_rate": 4.7969717825189264e-05, "loss": 1.138, "step": 23600 }, { "epoch": 4.069167240192705, "grad_norm": 9.739823341369629, "learning_rate": 4.796541637990365e-05, "loss": 1.1176, "step": 23650 }, { "epoch": 4.077770130763937, "grad_norm": 10.253618240356445, "learning_rate": 4.796111493461803e-05, "loss": 1.1778, "step": 23700 }, { "epoch": 4.086373021335168, "grad_norm": 8.859175682067871, "learning_rate": 4.795681348933242e-05, "loss": 1.1574, "step": 23750 }, { "epoch": 4.094975911906401, "grad_norm": 9.802985191345215, "learning_rate": 4.79525120440468e-05, "loss": 1.1415, "step": 23800 }, { "epoch": 4.1035788024776325, "grad_norm": 9.9966402053833, "learning_rate": 4.7948210598761184e-05, "loss": 1.164, "step": 23850 }, { "epoch": 4.112181693048864, "grad_norm": 9.745949745178223, "learning_rate": 4.794390915347557e-05, "loss": 1.1374, "step": 23900 }, { "epoch": 4.120784583620097, "grad_norm": 9.957430839538574, "learning_rate": 4.793960770818995e-05, "loss": 1.1162, "step": 23950 }, { "epoch": 4.129387474191328, "grad_norm": 8.747591972351074, "learning_rate": 4.793530626290434e-05, "loss": 1.1478, "step": 24000 }, { "epoch": 4.13799036476256, "grad_norm": 7.125631809234619, "learning_rate": 4.793100481761872e-05, "loss": 1.1119, "step": 24050 }, { "epoch": 4.1465932553337925, "grad_norm": 6.554769515991211, "learning_rate": 4.79267033723331e-05, "loss": 1.1515, "step": 24100 }, { "epoch": 4.155196145905024, "grad_norm": 11.683444023132324, "learning_rate": 4.792240192704749e-05, "loss": 1.1665, "step": 24150 }, { "epoch": 4.163799036476256, "grad_norm": 9.807790756225586, "learning_rate": 4.791810048176187e-05, "loss": 1.1724, "step": 24200 }, { "epoch": 4.172401927047488, "grad_norm": 8.38243293762207, "learning_rate": 4.791379903647626e-05, "loss": 1.1584, "step": 24250 }, { "epoch": 4.18100481761872, "grad_norm": 11.718547821044922, "learning_rate": 4.790949759119064e-05, "loss": 1.1177, "step": 24300 }, { "epoch": 4.189607708189952, "grad_norm": 8.948811531066895, "learning_rate": 4.790519614590503e-05, "loss": 1.1509, "step": 24350 }, { "epoch": 4.198210598761184, "grad_norm": 8.120148658752441, "learning_rate": 4.790089470061941e-05, "loss": 1.1159, "step": 24400 }, { "epoch": 4.206813489332416, "grad_norm": 8.107816696166992, "learning_rate": 4.789659325533379e-05, "loss": 1.1725, "step": 24450 }, { "epoch": 4.2154163799036475, "grad_norm": 10.2937593460083, "learning_rate": 4.789229181004818e-05, "loss": 1.1936, "step": 24500 }, { "epoch": 4.22401927047488, "grad_norm": 9.562525749206543, "learning_rate": 4.788799036476256e-05, "loss": 1.1717, "step": 24550 }, { "epoch": 4.232622161046112, "grad_norm": 15.684385299682617, "learning_rate": 4.788368891947695e-05, "loss": 1.1175, "step": 24600 }, { "epoch": 4.241225051617343, "grad_norm": 7.266592502593994, "learning_rate": 4.787938747419133e-05, "loss": 1.0955, "step": 24650 }, { "epoch": 4.249827942188575, "grad_norm": 13.350434303283691, "learning_rate": 4.787508602890571e-05, "loss": 1.1685, "step": 24700 }, { "epoch": 4.2584308327598075, "grad_norm": 7.553900718688965, "learning_rate": 4.78707845836201e-05, "loss": 1.1665, "step": 24750 }, { "epoch": 4.267033723331039, "grad_norm": 8.88724136352539, "learning_rate": 4.786648313833449e-05, "loss": 1.1641, "step": 24800 }, { "epoch": 4.275636613902271, "grad_norm": 10.17748737335205, "learning_rate": 4.786218169304887e-05, "loss": 1.1394, "step": 24850 }, { "epoch": 4.284239504473503, "grad_norm": 12.761473655700684, "learning_rate": 4.785788024776325e-05, "loss": 1.1239, "step": 24900 }, { "epoch": 4.292842395044735, "grad_norm": 9.519500732421875, "learning_rate": 4.785357880247763e-05, "loss": 1.2028, "step": 24950 }, { "epoch": 4.301445285615967, "grad_norm": 12.717300415039062, "learning_rate": 4.784927735719202e-05, "loss": 1.128, "step": 25000 }, { "epoch": 4.310048176187199, "grad_norm": 9.871108055114746, "learning_rate": 4.7844975911906406e-05, "loss": 1.1416, "step": 25050 }, { "epoch": 4.318651066758431, "grad_norm": 9.110209465026855, "learning_rate": 4.784067446662079e-05, "loss": 1.1876, "step": 25100 }, { "epoch": 4.3272539573296624, "grad_norm": 8.76669979095459, "learning_rate": 4.7836373021335175e-05, "loss": 1.1153, "step": 25150 }, { "epoch": 4.335856847900895, "grad_norm": 8.31628131866455, "learning_rate": 4.783207157604955e-05, "loss": 1.1213, "step": 25200 }, { "epoch": 4.344459738472127, "grad_norm": 11.543647766113281, "learning_rate": 4.782777013076394e-05, "loss": 1.1121, "step": 25250 }, { "epoch": 4.353062629043358, "grad_norm": 6.9339518547058105, "learning_rate": 4.7823468685478326e-05, "loss": 1.1164, "step": 25300 }, { "epoch": 4.361665519614591, "grad_norm": 10.87055492401123, "learning_rate": 4.781916724019271e-05, "loss": 1.1323, "step": 25350 }, { "epoch": 4.3702684101858225, "grad_norm": 9.360238075256348, "learning_rate": 4.7814865794907095e-05, "loss": 1.0968, "step": 25400 }, { "epoch": 4.378871300757054, "grad_norm": 7.358870983123779, "learning_rate": 4.7810564349621476e-05, "loss": 1.1416, "step": 25450 }, { "epoch": 4.387474191328287, "grad_norm": 7.85824728012085, "learning_rate": 4.780626290433586e-05, "loss": 1.1412, "step": 25500 }, { "epoch": 4.396077081899518, "grad_norm": 7.462183475494385, "learning_rate": 4.7801961459050245e-05, "loss": 1.1525, "step": 25550 }, { "epoch": 4.40467997247075, "grad_norm": 7.159233093261719, "learning_rate": 4.7797660013764626e-05, "loss": 1.1336, "step": 25600 }, { "epoch": 4.4132828630419825, "grad_norm": 15.847599983215332, "learning_rate": 4.7793358568479014e-05, "loss": 1.1172, "step": 25650 }, { "epoch": 4.421885753613214, "grad_norm": 7.956804275512695, "learning_rate": 4.7789057123193395e-05, "loss": 1.1422, "step": 25700 }, { "epoch": 4.430488644184446, "grad_norm": 8.91761589050293, "learning_rate": 4.7784755677907777e-05, "loss": 1.1374, "step": 25750 }, { "epoch": 4.439091534755678, "grad_norm": 8.245267868041992, "learning_rate": 4.7780454232622165e-05, "loss": 1.1088, "step": 25800 }, { "epoch": 4.44769442532691, "grad_norm": 9.835065841674805, "learning_rate": 4.7776152787336546e-05, "loss": 1.1551, "step": 25850 }, { "epoch": 4.456297315898142, "grad_norm": 12.182549476623535, "learning_rate": 4.7771851342050934e-05, "loss": 1.1447, "step": 25900 }, { "epoch": 4.464900206469373, "grad_norm": 10.502915382385254, "learning_rate": 4.7767549896765315e-05, "loss": 1.1348, "step": 25950 }, { "epoch": 4.473503097040606, "grad_norm": 7.5070624351501465, "learning_rate": 4.7763248451479696e-05, "loss": 1.0998, "step": 26000 }, { "epoch": 4.482105987611837, "grad_norm": 8.96928882598877, "learning_rate": 4.7758947006194084e-05, "loss": 1.1687, "step": 26050 }, { "epoch": 4.490708878183069, "grad_norm": 9.457311630249023, "learning_rate": 4.7754645560908465e-05, "loss": 1.1163, "step": 26100 }, { "epoch": 4.499311768754302, "grad_norm": 9.039990425109863, "learning_rate": 4.775034411562285e-05, "loss": 1.1482, "step": 26150 }, { "epoch": 4.507914659325533, "grad_norm": 9.14122200012207, "learning_rate": 4.774604267033724e-05, "loss": 1.1371, "step": 26200 }, { "epoch": 4.516517549896765, "grad_norm": 7.104044437408447, "learning_rate": 4.7741741225051615e-05, "loss": 1.1027, "step": 26250 }, { "epoch": 4.525120440467997, "grad_norm": 10.128216743469238, "learning_rate": 4.7737439779766e-05, "loss": 1.1377, "step": 26300 }, { "epoch": 4.533723331039229, "grad_norm": 9.200722694396973, "learning_rate": 4.7733138334480385e-05, "loss": 1.1261, "step": 26350 }, { "epoch": 4.542326221610461, "grad_norm": 9.21136474609375, "learning_rate": 4.772883688919477e-05, "loss": 1.1123, "step": 26400 }, { "epoch": 4.550929112181693, "grad_norm": 9.981544494628906, "learning_rate": 4.772453544390916e-05, "loss": 1.1115, "step": 26450 }, { "epoch": 4.559532002752925, "grad_norm": 9.89134693145752, "learning_rate": 4.772023399862354e-05, "loss": 1.1792, "step": 26500 }, { "epoch": 4.568134893324157, "grad_norm": 9.986018180847168, "learning_rate": 4.771593255333792e-05, "loss": 1.111, "step": 26550 }, { "epoch": 4.576737783895389, "grad_norm": 9.210308074951172, "learning_rate": 4.7711631108052304e-05, "loss": 1.0511, "step": 26600 }, { "epoch": 4.585340674466621, "grad_norm": 6.926953315734863, "learning_rate": 4.770732966276669e-05, "loss": 1.1806, "step": 26650 }, { "epoch": 4.593943565037852, "grad_norm": 11.342245101928711, "learning_rate": 4.770302821748108e-05, "loss": 1.1567, "step": 26700 }, { "epoch": 4.602546455609085, "grad_norm": 9.028667449951172, "learning_rate": 4.769872677219546e-05, "loss": 1.0878, "step": 26750 }, { "epoch": 4.611149346180317, "grad_norm": 8.811992645263672, "learning_rate": 4.769442532690984e-05, "loss": 1.1063, "step": 26800 }, { "epoch": 4.619752236751548, "grad_norm": 9.186092376708984, "learning_rate": 4.769012388162423e-05, "loss": 1.1356, "step": 26850 }, { "epoch": 4.628355127322781, "grad_norm": 7.792421340942383, "learning_rate": 4.768582243633861e-05, "loss": 1.0664, "step": 26900 }, { "epoch": 4.636958017894012, "grad_norm": 8.25204849243164, "learning_rate": 4.7681520991053e-05, "loss": 1.1039, "step": 26950 }, { "epoch": 4.645560908465244, "grad_norm": 8.27835750579834, "learning_rate": 4.767721954576738e-05, "loss": 1.1683, "step": 27000 }, { "epoch": 4.654163799036477, "grad_norm": 9.29047966003418, "learning_rate": 4.767291810048176e-05, "loss": 1.1286, "step": 27050 }, { "epoch": 4.662766689607708, "grad_norm": 9.111126899719238, "learning_rate": 4.766861665519615e-05, "loss": 1.1246, "step": 27100 }, { "epoch": 4.67136958017894, "grad_norm": 8.470370292663574, "learning_rate": 4.766431520991053e-05, "loss": 1.132, "step": 27150 }, { "epoch": 4.679972470750172, "grad_norm": 6.801020622253418, "learning_rate": 4.766001376462492e-05, "loss": 1.1126, "step": 27200 }, { "epoch": 4.688575361321404, "grad_norm": 6.8773345947265625, "learning_rate": 4.76557123193393e-05, "loss": 1.1197, "step": 27250 }, { "epoch": 4.697178251892636, "grad_norm": 11.976003646850586, "learning_rate": 4.765141087405369e-05, "loss": 1.178, "step": 27300 }, { "epoch": 4.705781142463868, "grad_norm": 9.004233360290527, "learning_rate": 4.764710942876807e-05, "loss": 1.1258, "step": 27350 }, { "epoch": 4.7143840330351, "grad_norm": 10.555312156677246, "learning_rate": 4.764280798348245e-05, "loss": 1.1352, "step": 27400 }, { "epoch": 4.7229869236063315, "grad_norm": 10.352596282958984, "learning_rate": 4.763850653819684e-05, "loss": 1.1213, "step": 27450 }, { "epoch": 4.731589814177564, "grad_norm": 7.926731586456299, "learning_rate": 4.763420509291122e-05, "loss": 1.1442, "step": 27500 }, { "epoch": 4.740192704748796, "grad_norm": 10.087555885314941, "learning_rate": 4.762990364762561e-05, "loss": 1.094, "step": 27550 }, { "epoch": 4.748795595320027, "grad_norm": 10.202659606933594, "learning_rate": 4.762560220233999e-05, "loss": 1.2037, "step": 27600 }, { "epoch": 4.75739848589126, "grad_norm": 9.314948081970215, "learning_rate": 4.762130075705437e-05, "loss": 1.1037, "step": 27650 }, { "epoch": 4.766001376462492, "grad_norm": 6.5578718185424805, "learning_rate": 4.761699931176876e-05, "loss": 1.1993, "step": 27700 }, { "epoch": 4.774604267033723, "grad_norm": 12.517261505126953, "learning_rate": 4.761269786648314e-05, "loss": 1.1171, "step": 27750 }, { "epoch": 4.783207157604956, "grad_norm": 9.939400672912598, "learning_rate": 4.7608396421197527e-05, "loss": 1.1675, "step": 27800 }, { "epoch": 4.791810048176187, "grad_norm": 12.914031982421875, "learning_rate": 4.760409497591191e-05, "loss": 1.1248, "step": 27850 }, { "epoch": 4.800412938747419, "grad_norm": 7.1246018409729, "learning_rate": 4.759979353062629e-05, "loss": 1.2084, "step": 27900 }, { "epoch": 4.809015829318651, "grad_norm": 11.916221618652344, "learning_rate": 4.759549208534068e-05, "loss": 1.1485, "step": 27950 }, { "epoch": 4.817618719889883, "grad_norm": 8.748719215393066, "learning_rate": 4.759119064005506e-05, "loss": 1.1366, "step": 28000 }, { "epoch": 4.826221610461115, "grad_norm": 10.020893096923828, "learning_rate": 4.7586889194769446e-05, "loss": 1.1104, "step": 28050 }, { "epoch": 4.8348245010323465, "grad_norm": 11.644490242004395, "learning_rate": 4.7582587749483834e-05, "loss": 1.112, "step": 28100 }, { "epoch": 4.843427391603579, "grad_norm": 7.254745006561279, "learning_rate": 4.757828630419821e-05, "loss": 1.1067, "step": 28150 }, { "epoch": 4.852030282174811, "grad_norm": 8.130575180053711, "learning_rate": 4.7573984858912596e-05, "loss": 1.1528, "step": 28200 }, { "epoch": 4.860633172746042, "grad_norm": 12.890424728393555, "learning_rate": 4.7569683413626984e-05, "loss": 1.1019, "step": 28250 }, { "epoch": 4.869236063317275, "grad_norm": 7.213662147521973, "learning_rate": 4.7565381968341365e-05, "loss": 1.1452, "step": 28300 }, { "epoch": 4.8778389538885065, "grad_norm": 13.076166152954102, "learning_rate": 4.756108052305575e-05, "loss": 1.1093, "step": 28350 }, { "epoch": 4.886441844459738, "grad_norm": 13.714181900024414, "learning_rate": 4.755677907777013e-05, "loss": 1.133, "step": 28400 }, { "epoch": 4.895044735030971, "grad_norm": 10.559218406677246, "learning_rate": 4.7552477632484516e-05, "loss": 1.1389, "step": 28450 }, { "epoch": 4.903647625602202, "grad_norm": 6.398659706115723, "learning_rate": 4.7548176187198904e-05, "loss": 1.1314, "step": 28500 }, { "epoch": 4.912250516173434, "grad_norm": 8.338431358337402, "learning_rate": 4.7543874741913285e-05, "loss": 1.1445, "step": 28550 }, { "epoch": 4.9208534067446665, "grad_norm": 9.172647476196289, "learning_rate": 4.753957329662767e-05, "loss": 1.1072, "step": 28600 }, { "epoch": 4.929456297315898, "grad_norm": 9.009549140930176, "learning_rate": 4.7535271851342054e-05, "loss": 1.1082, "step": 28650 }, { "epoch": 4.93805918788713, "grad_norm": 9.52446460723877, "learning_rate": 4.7530970406056435e-05, "loss": 1.1471, "step": 28700 }, { "epoch": 4.946662078458362, "grad_norm": 12.027637481689453, "learning_rate": 4.752666896077082e-05, "loss": 1.1629, "step": 28750 }, { "epoch": 4.955264969029594, "grad_norm": 7.422645092010498, "learning_rate": 4.7522367515485204e-05, "loss": 1.1052, "step": 28800 }, { "epoch": 4.963867859600826, "grad_norm": 6.467534065246582, "learning_rate": 4.751806607019959e-05, "loss": 1.1565, "step": 28850 }, { "epoch": 4.972470750172058, "grad_norm": 7.959805011749268, "learning_rate": 4.751376462491397e-05, "loss": 1.1226, "step": 28900 }, { "epoch": 4.98107364074329, "grad_norm": 7.971855640411377, "learning_rate": 4.7509463179628355e-05, "loss": 1.093, "step": 28950 }, { "epoch": 4.9896765313145215, "grad_norm": 12.263150215148926, "learning_rate": 4.750516173434274e-05, "loss": 1.1312, "step": 29000 }, { "epoch": 4.998279421885753, "grad_norm": 15.601908683776855, "learning_rate": 4.7500860289057124e-05, "loss": 1.1274, "step": 29050 }, { "epoch": 5.0, "eval_accuracy": 0.4207537429013939, "eval_f1": 0.39630635732407943, "eval_f1_DuraRiadoRio_16x16": 0.4295411349443389, "eval_f1_Mole_16x16": 0.366025641025641, "eval_f1_Quebrado_16x16": 0.493328922703716, "eval_f1_RiadoRio_16x16": 0.41063666581436975, "eval_f1_RioFechado_16x16": 0.2819994221323317, "eval_loss": 1.8811869621276855, "eval_precision": 0.5118953047659291, "eval_precision_DuraRiadoRio_16x16": 0.5736040609137056, "eval_precision_Mole_16x16": 0.6997549019607843, "eval_precision_Quebrado_16x16": 0.3306725794530673, "eval_precision_RiadoRio_16x16": 0.5429344151453684, "eval_precision_RioFechado_16x16": 0.4125105663567202, "eval_recall": 0.42129397918174866, "eval_recall_DuraRiadoRio_16x16": 0.3433159722222222, "eval_recall_Mole_16x16": 0.2478298611111111, "eval_recall_Quebrado_16x16": 0.9709201388888888, "eval_recall_RiadoRio_16x16": 0.3301809210526316, "eval_recall_RioFechado_16x16": 0.21422300263388938, "eval_runtime": 43.997, "eval_samples_per_second": 264.155, "eval_steps_per_second": 16.524, "step": 29060 }, { "epoch": 5.006882312456986, "grad_norm": 8.273503303527832, "learning_rate": 4.749655884377151e-05, "loss": 1.1489, "step": 29100 }, { "epoch": 5.015485203028217, "grad_norm": 8.736411094665527, "learning_rate": 4.749225739848589e-05, "loss": 1.1507, "step": 29150 }, { "epoch": 5.02408809359945, "grad_norm": 9.551499366760254, "learning_rate": 4.7487955953200274e-05, "loss": 1.129, "step": 29200 }, { "epoch": 5.0326909841706815, "grad_norm": 8.612555503845215, "learning_rate": 4.748365450791466e-05, "loss": 1.1573, "step": 29250 }, { "epoch": 5.041293874741913, "grad_norm": 7.090129375457764, "learning_rate": 4.747935306262904e-05, "loss": 1.0914, "step": 29300 }, { "epoch": 5.049896765313145, "grad_norm": 7.175063610076904, "learning_rate": 4.747505161734343e-05, "loss": 1.1327, "step": 29350 }, { "epoch": 5.058499655884377, "grad_norm": 9.44810962677002, "learning_rate": 4.747075017205781e-05, "loss": 1.1467, "step": 29400 }, { "epoch": 5.067102546455609, "grad_norm": 7.202795505523682, "learning_rate": 4.74664487267722e-05, "loss": 1.084, "step": 29450 }, { "epoch": 5.075705437026841, "grad_norm": 9.322400093078613, "learning_rate": 4.746214728148658e-05, "loss": 1.1022, "step": 29500 }, { "epoch": 5.084308327598073, "grad_norm": 11.811029434204102, "learning_rate": 4.745784583620096e-05, "loss": 1.1251, "step": 29550 }, { "epoch": 5.092911218169305, "grad_norm": 11.32391357421875, "learning_rate": 4.745354439091535e-05, "loss": 1.0892, "step": 29600 }, { "epoch": 5.1015141087405365, "grad_norm": 11.291163444519043, "learning_rate": 4.744924294562973e-05, "loss": 1.2079, "step": 29650 }, { "epoch": 5.110116999311769, "grad_norm": 6.380284786224365, "learning_rate": 4.744494150034412e-05, "loss": 1.0893, "step": 29700 }, { "epoch": 5.118719889883001, "grad_norm": 9.654666900634766, "learning_rate": 4.74406400550585e-05, "loss": 1.1061, "step": 29750 }, { "epoch": 5.127322780454232, "grad_norm": 8.087433815002441, "learning_rate": 4.743633860977288e-05, "loss": 1.1006, "step": 29800 }, { "epoch": 5.135925671025465, "grad_norm": 16.171363830566406, "learning_rate": 4.743203716448727e-05, "loss": 1.0351, "step": 29850 }, { "epoch": 5.1445285615966965, "grad_norm": 10.57519245147705, "learning_rate": 4.742773571920166e-05, "loss": 1.0659, "step": 29900 }, { "epoch": 5.153131452167928, "grad_norm": 9.539327621459961, "learning_rate": 4.742343427391604e-05, "loss": 1.1343, "step": 29950 }, { "epoch": 5.161734342739161, "grad_norm": 8.320158004760742, "learning_rate": 4.741913282863043e-05, "loss": 1.0932, "step": 30000 }, { "epoch": 5.170337233310392, "grad_norm": 8.315125465393066, "learning_rate": 4.74148313833448e-05, "loss": 1.1344, "step": 30050 }, { "epoch": 5.178940123881624, "grad_norm": 7.526221752166748, "learning_rate": 4.741052993805919e-05, "loss": 1.094, "step": 30100 }, { "epoch": 5.1875430144528565, "grad_norm": 7.00205659866333, "learning_rate": 4.740622849277358e-05, "loss": 1.1309, "step": 30150 }, { "epoch": 5.196145905024088, "grad_norm": 8.665617942810059, "learning_rate": 4.740192704748796e-05, "loss": 1.0961, "step": 30200 }, { "epoch": 5.20474879559532, "grad_norm": 7.226852893829346, "learning_rate": 4.7397625602202346e-05, "loss": 1.0834, "step": 30250 }, { "epoch": 5.213351686166552, "grad_norm": 9.005526542663574, "learning_rate": 4.739332415691672e-05, "loss": 1.1344, "step": 30300 }, { "epoch": 5.221954576737784, "grad_norm": 6.575538635253906, "learning_rate": 4.738902271163111e-05, "loss": 1.0885, "step": 30350 }, { "epoch": 5.230557467309016, "grad_norm": 7.571394920349121, "learning_rate": 4.73847212663455e-05, "loss": 1.1893, "step": 30400 }, { "epoch": 5.239160357880248, "grad_norm": 11.684035301208496, "learning_rate": 4.738041982105988e-05, "loss": 1.1118, "step": 30450 }, { "epoch": 5.24776324845148, "grad_norm": 10.263812065124512, "learning_rate": 4.7376118375774266e-05, "loss": 1.0774, "step": 30500 }, { "epoch": 5.256366139022711, "grad_norm": 7.101699352264404, "learning_rate": 4.737181693048865e-05, "loss": 1.1093, "step": 30550 }, { "epoch": 5.264969029593944, "grad_norm": 13.406167030334473, "learning_rate": 4.736751548520303e-05, "loss": 1.1165, "step": 30600 }, { "epoch": 5.273571920165176, "grad_norm": 8.761307716369629, "learning_rate": 4.7363214039917416e-05, "loss": 1.13, "step": 30650 }, { "epoch": 5.282174810736407, "grad_norm": 9.367657661437988, "learning_rate": 4.73589125946318e-05, "loss": 1.1061, "step": 30700 }, { "epoch": 5.29077770130764, "grad_norm": 9.759580612182617, "learning_rate": 4.7354611149346185e-05, "loss": 1.14, "step": 30750 }, { "epoch": 5.2993805918788714, "grad_norm": 8.56114673614502, "learning_rate": 4.7350309704060566e-05, "loss": 1.1032, "step": 30800 }, { "epoch": 5.307983482450103, "grad_norm": 7.02548360824585, "learning_rate": 4.734600825877495e-05, "loss": 1.1758, "step": 30850 }, { "epoch": 5.316586373021336, "grad_norm": 8.80550479888916, "learning_rate": 4.7341706813489335e-05, "loss": 1.1118, "step": 30900 }, { "epoch": 5.325189263592567, "grad_norm": 20.671131134033203, "learning_rate": 4.733740536820372e-05, "loss": 1.1144, "step": 30950 }, { "epoch": 5.333792154163799, "grad_norm": 11.108724594116211, "learning_rate": 4.7333103922918105e-05, "loss": 1.1183, "step": 31000 }, { "epoch": 5.342395044735031, "grad_norm": 11.339166641235352, "learning_rate": 4.7328802477632486e-05, "loss": 1.0968, "step": 31050 }, { "epoch": 5.350997935306263, "grad_norm": 10.62255573272705, "learning_rate": 4.732450103234687e-05, "loss": 1.1034, "step": 31100 }, { "epoch": 5.359600825877495, "grad_norm": 7.00341796875, "learning_rate": 4.7320199587061255e-05, "loss": 1.154, "step": 31150 }, { "epoch": 5.368203716448726, "grad_norm": 9.289557456970215, "learning_rate": 4.7315898141775636e-05, "loss": 1.1015, "step": 31200 }, { "epoch": 5.376806607019959, "grad_norm": 15.561761856079102, "learning_rate": 4.7311596696490024e-05, "loss": 1.1106, "step": 31250 }, { "epoch": 5.385409497591191, "grad_norm": 7.767017841339111, "learning_rate": 4.730729525120441e-05, "loss": 1.1099, "step": 31300 }, { "epoch": 5.394012388162422, "grad_norm": 8.826674461364746, "learning_rate": 4.7302993805918786e-05, "loss": 1.0875, "step": 31350 }, { "epoch": 5.402615278733655, "grad_norm": 11.853226661682129, "learning_rate": 4.7298692360633174e-05, "loss": 1.1062, "step": 31400 }, { "epoch": 5.411218169304886, "grad_norm": 9.526140213012695, "learning_rate": 4.7294390915347555e-05, "loss": 1.0586, "step": 31450 }, { "epoch": 5.419821059876118, "grad_norm": 7.355340003967285, "learning_rate": 4.7290089470061943e-05, "loss": 1.1137, "step": 31500 }, { "epoch": 5.428423950447351, "grad_norm": 11.112641334533691, "learning_rate": 4.728578802477633e-05, "loss": 1.1336, "step": 31550 }, { "epoch": 5.437026841018582, "grad_norm": 10.075203895568848, "learning_rate": 4.728148657949071e-05, "loss": 1.1537, "step": 31600 }, { "epoch": 5.445629731589814, "grad_norm": 8.10478687286377, "learning_rate": 4.7277185134205094e-05, "loss": 1.0798, "step": 31650 }, { "epoch": 5.454232622161046, "grad_norm": 13.16767692565918, "learning_rate": 4.7272883688919475e-05, "loss": 1.1589, "step": 31700 }, { "epoch": 5.462835512732278, "grad_norm": 7.3839240074157715, "learning_rate": 4.726858224363386e-05, "loss": 1.1123, "step": 31750 }, { "epoch": 5.47143840330351, "grad_norm": 13.087346076965332, "learning_rate": 4.726428079834825e-05, "loss": 1.1293, "step": 31800 }, { "epoch": 5.480041293874742, "grad_norm": 8.777517318725586, "learning_rate": 4.725997935306263e-05, "loss": 1.1029, "step": 31850 }, { "epoch": 5.488644184445974, "grad_norm": 11.62597370147705, "learning_rate": 4.725567790777701e-05, "loss": 1.1145, "step": 31900 }, { "epoch": 5.4972470750172056, "grad_norm": 9.100055694580078, "learning_rate": 4.72513764624914e-05, "loss": 1.148, "step": 31950 }, { "epoch": 5.505849965588438, "grad_norm": 11.302828788757324, "learning_rate": 4.724707501720578e-05, "loss": 1.155, "step": 32000 }, { "epoch": 5.51445285615967, "grad_norm": 8.214823722839355, "learning_rate": 4.724277357192017e-05, "loss": 1.0832, "step": 32050 }, { "epoch": 5.523055746730901, "grad_norm": 11.595958709716797, "learning_rate": 4.723847212663455e-05, "loss": 1.1168, "step": 32100 }, { "epoch": 5.531658637302133, "grad_norm": 8.933788299560547, "learning_rate": 4.723417068134894e-05, "loss": 1.0831, "step": 32150 }, { "epoch": 5.540261527873366, "grad_norm": 11.556126594543457, "learning_rate": 4.722986923606332e-05, "loss": 1.0912, "step": 32200 }, { "epoch": 5.548864418444597, "grad_norm": 9.124109268188477, "learning_rate": 4.72255677907777e-05, "loss": 1.108, "step": 32250 }, { "epoch": 5.557467309015829, "grad_norm": 8.962485313415527, "learning_rate": 4.722126634549209e-05, "loss": 1.1362, "step": 32300 }, { "epoch": 5.566070199587061, "grad_norm": 12.160035133361816, "learning_rate": 4.721696490020647e-05, "loss": 1.1782, "step": 32350 }, { "epoch": 5.574673090158293, "grad_norm": 9.00731086730957, "learning_rate": 4.721266345492086e-05, "loss": 1.1296, "step": 32400 }, { "epoch": 5.583275980729525, "grad_norm": 9.513813972473145, "learning_rate": 4.720836200963524e-05, "loss": 1.0944, "step": 32450 }, { "epoch": 5.591878871300757, "grad_norm": 8.684186935424805, "learning_rate": 4.720406056434962e-05, "loss": 1.1467, "step": 32500 }, { "epoch": 5.600481761871989, "grad_norm": 12.06817626953125, "learning_rate": 4.719975911906401e-05, "loss": 1.0873, "step": 32550 }, { "epoch": 5.6090846524432205, "grad_norm": 10.180322647094727, "learning_rate": 4.719545767377839e-05, "loss": 1.1254, "step": 32600 }, { "epoch": 5.617687543014453, "grad_norm": 9.721559524536133, "learning_rate": 4.719115622849278e-05, "loss": 1.0792, "step": 32650 }, { "epoch": 5.626290433585685, "grad_norm": 10.143672943115234, "learning_rate": 4.718685478320716e-05, "loss": 1.1173, "step": 32700 }, { "epoch": 5.634893324156916, "grad_norm": 13.625216484069824, "learning_rate": 4.718255333792154e-05, "loss": 1.0924, "step": 32750 }, { "epoch": 5.643496214728149, "grad_norm": 8.645957946777344, "learning_rate": 4.717825189263593e-05, "loss": 1.0742, "step": 32800 }, { "epoch": 5.6520991052993805, "grad_norm": 9.395225524902344, "learning_rate": 4.717395044735031e-05, "loss": 1.0975, "step": 32850 }, { "epoch": 5.660701995870612, "grad_norm": 9.528785705566406, "learning_rate": 4.71696490020647e-05, "loss": 1.1172, "step": 32900 }, { "epoch": 5.669304886441845, "grad_norm": 10.422342300415039, "learning_rate": 4.7165347556779086e-05, "loss": 1.1317, "step": 32950 }, { "epoch": 5.677907777013076, "grad_norm": 9.828675270080566, "learning_rate": 4.716104611149346e-05, "loss": 1.1366, "step": 33000 }, { "epoch": 5.686510667584308, "grad_norm": 7.4821906089782715, "learning_rate": 4.715674466620785e-05, "loss": 1.1323, "step": 33050 }, { "epoch": 5.6951135581555405, "grad_norm": 9.167981147766113, "learning_rate": 4.715244322092223e-05, "loss": 1.1002, "step": 33100 }, { "epoch": 5.703716448726772, "grad_norm": 11.369945526123047, "learning_rate": 4.714814177563662e-05, "loss": 1.1034, "step": 33150 }, { "epoch": 5.712319339298004, "grad_norm": 10.359912872314453, "learning_rate": 4.7143840330351005e-05, "loss": 1.1103, "step": 33200 }, { "epoch": 5.720922229869236, "grad_norm": 9.701354026794434, "learning_rate": 4.713953888506538e-05, "loss": 1.0999, "step": 33250 }, { "epoch": 5.729525120440468, "grad_norm": 9.83102035522461, "learning_rate": 4.713523743977977e-05, "loss": 1.1338, "step": 33300 }, { "epoch": 5.7381280110117, "grad_norm": 8.784396171569824, "learning_rate": 4.7130935994494155e-05, "loss": 1.1244, "step": 33350 }, { "epoch": 5.746730901582932, "grad_norm": 10.28615951538086, "learning_rate": 4.7126634549208536e-05, "loss": 1.1322, "step": 33400 }, { "epoch": 5.755333792154164, "grad_norm": 11.13105297088623, "learning_rate": 4.7122333103922924e-05, "loss": 1.1421, "step": 33450 }, { "epoch": 5.7639366827253955, "grad_norm": 6.541734218597412, "learning_rate": 4.71180316586373e-05, "loss": 1.0986, "step": 33500 }, { "epoch": 5.772539573296628, "grad_norm": 8.877166748046875, "learning_rate": 4.711373021335169e-05, "loss": 1.1393, "step": 33550 }, { "epoch": 5.78114246386786, "grad_norm": 11.930392265319824, "learning_rate": 4.7109428768066075e-05, "loss": 1.1175, "step": 33600 }, { "epoch": 5.789745354439091, "grad_norm": 10.812129020690918, "learning_rate": 4.7105127322780456e-05, "loss": 1.0834, "step": 33650 }, { "epoch": 5.798348245010324, "grad_norm": 7.650674343109131, "learning_rate": 4.7100825877494844e-05, "loss": 1.0832, "step": 33700 }, { "epoch": 5.8069511355815555, "grad_norm": 10.42587661743164, "learning_rate": 4.7096524432209225e-05, "loss": 1.1172, "step": 33750 }, { "epoch": 5.815554026152787, "grad_norm": 5.958961009979248, "learning_rate": 4.7092222986923606e-05, "loss": 1.0832, "step": 33800 }, { "epoch": 5.82415691672402, "grad_norm": 13.15678596496582, "learning_rate": 4.7087921541637994e-05, "loss": 1.1442, "step": 33850 }, { "epoch": 5.832759807295251, "grad_norm": 21.27756690979004, "learning_rate": 4.7083620096352375e-05, "loss": 1.1468, "step": 33900 }, { "epoch": 5.841362697866483, "grad_norm": 9.111732482910156, "learning_rate": 4.707931865106676e-05, "loss": 1.0687, "step": 33950 }, { "epoch": 5.8499655884377155, "grad_norm": 10.99925708770752, "learning_rate": 4.7075017205781144e-05, "loss": 1.0617, "step": 34000 }, { "epoch": 5.858568479008947, "grad_norm": 8.522407531738281, "learning_rate": 4.7070715760495526e-05, "loss": 1.073, "step": 34050 }, { "epoch": 5.867171369580179, "grad_norm": 8.487279891967773, "learning_rate": 4.7066414315209913e-05, "loss": 1.1193, "step": 34100 }, { "epoch": 5.875774260151411, "grad_norm": 12.61154842376709, "learning_rate": 4.7062112869924295e-05, "loss": 1.1123, "step": 34150 }, { "epoch": 5.884377150722643, "grad_norm": 8.535813331604004, "learning_rate": 4.705781142463868e-05, "loss": 1.1067, "step": 34200 }, { "epoch": 5.892980041293875, "grad_norm": 9.375371932983398, "learning_rate": 4.7053509979353064e-05, "loss": 1.1272, "step": 34250 }, { "epoch": 5.901582931865106, "grad_norm": 7.59044075012207, "learning_rate": 4.704920853406745e-05, "loss": 1.1614, "step": 34300 }, { "epoch": 5.910185822436339, "grad_norm": 6.2995476722717285, "learning_rate": 4.704490708878183e-05, "loss": 1.0755, "step": 34350 }, { "epoch": 5.9187887130075705, "grad_norm": 8.171875953674316, "learning_rate": 4.7040605643496214e-05, "loss": 1.0724, "step": 34400 }, { "epoch": 5.927391603578802, "grad_norm": 9.613368034362793, "learning_rate": 4.70363041982106e-05, "loss": 1.0986, "step": 34450 }, { "epoch": 5.935994494150035, "grad_norm": 7.582942008972168, "learning_rate": 4.703200275292498e-05, "loss": 1.1236, "step": 34500 }, { "epoch": 5.944597384721266, "grad_norm": 7.468239784240723, "learning_rate": 4.702770130763937e-05, "loss": 1.1205, "step": 34550 }, { "epoch": 5.953200275292498, "grad_norm": 11.330022811889648, "learning_rate": 4.702339986235375e-05, "loss": 1.104, "step": 34600 }, { "epoch": 5.9618031658637305, "grad_norm": 14.482731819152832, "learning_rate": 4.7019098417068133e-05, "loss": 1.1074, "step": 34650 }, { "epoch": 5.970406056434962, "grad_norm": 6.603274345397949, "learning_rate": 4.701479697178252e-05, "loss": 1.1087, "step": 34700 }, { "epoch": 5.979008947006194, "grad_norm": 7.899117946624756, "learning_rate": 4.701049552649691e-05, "loss": 1.1604, "step": 34750 }, { "epoch": 5.987611837577426, "grad_norm": 6.954283237457275, "learning_rate": 4.700619408121129e-05, "loss": 1.1361, "step": 34800 }, { "epoch": 5.996214728148658, "grad_norm": 18.264930725097656, "learning_rate": 4.700189263592567e-05, "loss": 1.1279, "step": 34850 }, { "epoch": 6.0, "eval_accuracy": 0.5339872655308897, "eval_f1": 0.5402993157374666, "eval_f1_DuraRiadoRio_16x16": 0.5093745723279047, "eval_f1_Mole_16x16": 0.5589931830099633, "eval_f1_Quebrado_16x16": 0.6554156171284635, "eval_f1_RiadoRio_16x16": 0.4530066815144766, "eval_f1_RioFechado_16x16": 0.5247065247065247, "eval_loss": 1.189704179763794, "eval_precision": 0.6093762711705366, "eval_precision_DuraRiadoRio_16x16": 0.371976813911653, "eval_precision_Mole_16x16": 0.7059602649006622, "eval_precision_Quebrado_16x16": 0.7809123649459784, "eval_precision_RiadoRio_16x16": 0.49416909620991256, "eval_precision_RioFechado_16x16": 0.6938628158844765, "eval_recall": 0.5350210136751741, "eval_recall_DuraRiadoRio_16x16": 0.8077256944444444, "eval_recall_Mole_16x16": 0.4626736111111111, "eval_recall_Quebrado_16x16": 0.5646701388888888, "eval_recall_RiadoRio_16x16": 0.41817434210526316, "eval_recall_RioFechado_16x16": 0.4218612818261633, "eval_runtime": 43.6565, "eval_samples_per_second": 266.214, "eval_steps_per_second": 16.653, "step": 34872 }, { "epoch": 6.00481761871989, "grad_norm": 10.599852561950684, "learning_rate": 4.699759119064005e-05, "loss": 1.0852, "step": 34900 }, { "epoch": 6.013420509291122, "grad_norm": 13.947686195373535, "learning_rate": 4.699328974535444e-05, "loss": 1.1221, "step": 34950 }, { "epoch": 6.022023399862354, "grad_norm": 6.681899070739746, "learning_rate": 4.698898830006883e-05, "loss": 1.1007, "step": 35000 }, { "epoch": 6.0306262904335854, "grad_norm": 10.022741317749023, "learning_rate": 4.698468685478321e-05, "loss": 1.0746, "step": 35050 }, { "epoch": 6.039229181004818, "grad_norm": 12.551980972290039, "learning_rate": 4.69803854094976e-05, "loss": 1.1039, "step": 35100 }, { "epoch": 6.04783207157605, "grad_norm": 9.670503616333008, "learning_rate": 4.697608396421197e-05, "loss": 1.1427, "step": 35150 }, { "epoch": 6.056434962147281, "grad_norm": 12.14493179321289, "learning_rate": 4.697178251892636e-05, "loss": 1.1114, "step": 35200 }, { "epoch": 6.065037852718514, "grad_norm": 8.779290199279785, "learning_rate": 4.696748107364075e-05, "loss": 1.1295, "step": 35250 }, { "epoch": 6.0736407432897455, "grad_norm": 9.879617691040039, "learning_rate": 4.696317962835513e-05, "loss": 1.0684, "step": 35300 }, { "epoch": 6.082243633860977, "grad_norm": 12.795211791992188, "learning_rate": 4.695887818306952e-05, "loss": 1.0879, "step": 35350 }, { "epoch": 6.09084652443221, "grad_norm": 6.466690540313721, "learning_rate": 4.69545767377839e-05, "loss": 1.0434, "step": 35400 }, { "epoch": 6.099449415003441, "grad_norm": 13.476972579956055, "learning_rate": 4.695027529249828e-05, "loss": 1.133, "step": 35450 }, { "epoch": 6.108052305574673, "grad_norm": 12.043290138244629, "learning_rate": 4.694597384721267e-05, "loss": 1.1071, "step": 35500 }, { "epoch": 6.116655196145905, "grad_norm": 14.487146377563477, "learning_rate": 4.694167240192705e-05, "loss": 1.1268, "step": 35550 }, { "epoch": 6.125258086717137, "grad_norm": 8.155052185058594, "learning_rate": 4.693737095664144e-05, "loss": 1.0904, "step": 35600 }, { "epoch": 6.133860977288369, "grad_norm": 7.112220764160156, "learning_rate": 4.693306951135582e-05, "loss": 1.0954, "step": 35650 }, { "epoch": 6.1424638678596, "grad_norm": 9.152868270874023, "learning_rate": 4.69287680660702e-05, "loss": 1.0805, "step": 35700 }, { "epoch": 6.151066758430833, "grad_norm": 13.248756408691406, "learning_rate": 4.692446662078459e-05, "loss": 1.0982, "step": 35750 }, { "epoch": 6.159669649002065, "grad_norm": 9.083317756652832, "learning_rate": 4.692016517549897e-05, "loss": 1.1566, "step": 35800 }, { "epoch": 6.168272539573296, "grad_norm": 7.904919147491455, "learning_rate": 4.6915863730213356e-05, "loss": 1.0798, "step": 35850 }, { "epoch": 6.176875430144529, "grad_norm": 9.128437995910645, "learning_rate": 4.691156228492774e-05, "loss": 1.0698, "step": 35900 }, { "epoch": 6.18547832071576, "grad_norm": 9.93344497680664, "learning_rate": 4.690726083964212e-05, "loss": 1.0693, "step": 35950 }, { "epoch": 6.194081211286992, "grad_norm": 6.662264347076416, "learning_rate": 4.6902959394356506e-05, "loss": 1.102, "step": 36000 }, { "epoch": 6.202684101858225, "grad_norm": 11.871209144592285, "learning_rate": 4.689865794907089e-05, "loss": 1.1315, "step": 36050 }, { "epoch": 6.211286992429456, "grad_norm": 6.72116756439209, "learning_rate": 4.6894356503785276e-05, "loss": 1.1268, "step": 36100 }, { "epoch": 6.219889883000688, "grad_norm": 10.820186614990234, "learning_rate": 4.689005505849966e-05, "loss": 1.1036, "step": 36150 }, { "epoch": 6.22849277357192, "grad_norm": 7.405300140380859, "learning_rate": 4.688575361321404e-05, "loss": 1.0955, "step": 36200 }, { "epoch": 6.237095664143152, "grad_norm": 9.79846477508545, "learning_rate": 4.6881452167928426e-05, "loss": 1.0374, "step": 36250 }, { "epoch": 6.245698554714384, "grad_norm": 12.176464080810547, "learning_rate": 4.687715072264281e-05, "loss": 1.0412, "step": 36300 }, { "epoch": 6.254301445285616, "grad_norm": 10.590193748474121, "learning_rate": 4.6872849277357195e-05, "loss": 1.0975, "step": 36350 }, { "epoch": 6.262904335856848, "grad_norm": 8.75326156616211, "learning_rate": 4.686854783207158e-05, "loss": 1.0447, "step": 36400 }, { "epoch": 6.27150722642808, "grad_norm": 9.55170726776123, "learning_rate": 4.6864246386785964e-05, "loss": 1.0836, "step": 36450 }, { "epoch": 6.280110116999312, "grad_norm": 8.378891944885254, "learning_rate": 4.6859944941500345e-05, "loss": 1.0968, "step": 36500 }, { "epoch": 6.288713007570544, "grad_norm": 9.43307113647461, "learning_rate": 4.6855643496214726e-05, "loss": 1.0949, "step": 36550 }, { "epoch": 6.297315898141775, "grad_norm": 8.4977445602417, "learning_rate": 4.6851342050929114e-05, "loss": 1.086, "step": 36600 }, { "epoch": 6.305918788713008, "grad_norm": 9.52658748626709, "learning_rate": 4.68470406056435e-05, "loss": 1.0284, "step": 36650 }, { "epoch": 6.31452167928424, "grad_norm": 9.090752601623535, "learning_rate": 4.6842739160357883e-05, "loss": 1.0969, "step": 36700 }, { "epoch": 6.323124569855471, "grad_norm": 13.932868957519531, "learning_rate": 4.6838437715072265e-05, "loss": 1.1255, "step": 36750 }, { "epoch": 6.331727460426704, "grad_norm": 8.39207649230957, "learning_rate": 4.683413626978665e-05, "loss": 1.0789, "step": 36800 }, { "epoch": 6.340330350997935, "grad_norm": 17.80738639831543, "learning_rate": 4.6829834824501034e-05, "loss": 1.053, "step": 36850 }, { "epoch": 6.348933241569167, "grad_norm": 9.20669937133789, "learning_rate": 4.682553337921542e-05, "loss": 1.1231, "step": 36900 }, { "epoch": 6.3575361321404, "grad_norm": 8.831192970275879, "learning_rate": 4.68212319339298e-05, "loss": 1.1007, "step": 36950 }, { "epoch": 6.366139022711631, "grad_norm": 9.579017639160156, "learning_rate": 4.6816930488644184e-05, "loss": 1.1473, "step": 37000 }, { "epoch": 6.374741913282863, "grad_norm": 8.42011833190918, "learning_rate": 4.681262904335857e-05, "loss": 1.1003, "step": 37050 }, { "epoch": 6.383344803854095, "grad_norm": 11.460577964782715, "learning_rate": 4.680832759807295e-05, "loss": 1.0514, "step": 37100 }, { "epoch": 6.391947694425327, "grad_norm": 5.769279479980469, "learning_rate": 4.680402615278734e-05, "loss": 1.1013, "step": 37150 }, { "epoch": 6.400550584996559, "grad_norm": 7.965705394744873, "learning_rate": 4.679972470750172e-05, "loss": 1.0507, "step": 37200 }, { "epoch": 6.409153475567791, "grad_norm": 10.24227523803711, "learning_rate": 4.679542326221611e-05, "loss": 1.1041, "step": 37250 }, { "epoch": 6.417756366139023, "grad_norm": 7.9588117599487305, "learning_rate": 4.679112181693049e-05, "loss": 1.0577, "step": 37300 }, { "epoch": 6.4263592567102545, "grad_norm": 8.542363166809082, "learning_rate": 4.678682037164487e-05, "loss": 1.1027, "step": 37350 }, { "epoch": 6.434962147281486, "grad_norm": 9.024065971374512, "learning_rate": 4.678251892635926e-05, "loss": 1.0985, "step": 37400 }, { "epoch": 6.443565037852719, "grad_norm": 7.985350131988525, "learning_rate": 4.677821748107364e-05, "loss": 1.1113, "step": 37450 }, { "epoch": 6.45216792842395, "grad_norm": 9.208552360534668, "learning_rate": 4.677391603578803e-05, "loss": 1.1155, "step": 37500 }, { "epoch": 6.460770818995182, "grad_norm": 8.737655639648438, "learning_rate": 4.676961459050241e-05, "loss": 1.0878, "step": 37550 }, { "epoch": 6.4693737095664146, "grad_norm": 8.260438919067383, "learning_rate": 4.676531314521679e-05, "loss": 1.0784, "step": 37600 }, { "epoch": 6.477976600137646, "grad_norm": 8.496352195739746, "learning_rate": 4.676101169993118e-05, "loss": 1.1064, "step": 37650 }, { "epoch": 6.486579490708878, "grad_norm": 11.730514526367188, "learning_rate": 4.675671025464556e-05, "loss": 1.1234, "step": 37700 }, { "epoch": 6.49518238128011, "grad_norm": 7.518617153167725, "learning_rate": 4.675240880935995e-05, "loss": 1.1421, "step": 37750 }, { "epoch": 6.503785271851342, "grad_norm": 6.3342509269714355, "learning_rate": 4.674810736407434e-05, "loss": 1.0422, "step": 37800 }, { "epoch": 6.512388162422574, "grad_norm": 7.928038120269775, "learning_rate": 4.674380591878871e-05, "loss": 1.0992, "step": 37850 }, { "epoch": 6.520991052993806, "grad_norm": 7.696896553039551, "learning_rate": 4.67395044735031e-05, "loss": 1.0472, "step": 37900 }, { "epoch": 6.529593943565038, "grad_norm": 9.891212463378906, "learning_rate": 4.673520302821748e-05, "loss": 1.1005, "step": 37950 }, { "epoch": 6.5381968341362695, "grad_norm": 8.601638793945312, "learning_rate": 4.673090158293187e-05, "loss": 1.0838, "step": 38000 }, { "epoch": 6.546799724707502, "grad_norm": 6.51720666885376, "learning_rate": 4.6726600137646256e-05, "loss": 1.0624, "step": 38050 }, { "epoch": 6.555402615278734, "grad_norm": 7.004611492156982, "learning_rate": 4.672229869236063e-05, "loss": 1.115, "step": 38100 }, { "epoch": 6.564005505849965, "grad_norm": 9.33319091796875, "learning_rate": 4.671799724707502e-05, "loss": 1.1009, "step": 38150 }, { "epoch": 6.572608396421198, "grad_norm": 7.95089054107666, "learning_rate": 4.67136958017894e-05, "loss": 1.1097, "step": 38200 }, { "epoch": 6.5812112869924295, "grad_norm": 8.482962608337402, "learning_rate": 4.670939435650379e-05, "loss": 1.0876, "step": 38250 }, { "epoch": 6.589814177563661, "grad_norm": 10.062106132507324, "learning_rate": 4.6705092911218176e-05, "loss": 1.1028, "step": 38300 }, { "epoch": 6.598417068134894, "grad_norm": 9.445982933044434, "learning_rate": 4.670079146593255e-05, "loss": 1.0543, "step": 38350 }, { "epoch": 6.607019958706125, "grad_norm": 8.946648597717285, "learning_rate": 4.669649002064694e-05, "loss": 1.0801, "step": 38400 }, { "epoch": 6.615622849277357, "grad_norm": 7.678650856018066, "learning_rate": 4.6692188575361326e-05, "loss": 1.0862, "step": 38450 }, { "epoch": 6.624225739848589, "grad_norm": 10.741484642028809, "learning_rate": 4.668788713007571e-05, "loss": 1.015, "step": 38500 }, { "epoch": 6.632828630419821, "grad_norm": 11.552152633666992, "learning_rate": 4.6683585684790095e-05, "loss": 1.0811, "step": 38550 }, { "epoch": 6.641431520991053, "grad_norm": 11.74350643157959, "learning_rate": 4.6679284239504476e-05, "loss": 1.1386, "step": 38600 }, { "epoch": 6.6500344115622845, "grad_norm": 13.595562934875488, "learning_rate": 4.667498279421886e-05, "loss": 1.0984, "step": 38650 }, { "epoch": 6.658637302133517, "grad_norm": 10.471508026123047, "learning_rate": 4.6670681348933246e-05, "loss": 1.13, "step": 38700 }, { "epoch": 6.667240192704749, "grad_norm": 7.955653190612793, "learning_rate": 4.666637990364763e-05, "loss": 1.0714, "step": 38750 }, { "epoch": 6.67584308327598, "grad_norm": 8.774974822998047, "learning_rate": 4.6662078458362015e-05, "loss": 1.1387, "step": 38800 }, { "epoch": 6.684445973847213, "grad_norm": 9.823657989501953, "learning_rate": 4.6657777013076396e-05, "loss": 1.0983, "step": 38850 }, { "epoch": 6.6930488644184445, "grad_norm": 14.901476860046387, "learning_rate": 4.665347556779078e-05, "loss": 1.0814, "step": 38900 }, { "epoch": 6.701651754989676, "grad_norm": 10.45166301727295, "learning_rate": 4.6649174122505165e-05, "loss": 1.0444, "step": 38950 }, { "epoch": 6.710254645560909, "grad_norm": 7.188843727111816, "learning_rate": 4.6644872677219546e-05, "loss": 1.0963, "step": 39000 }, { "epoch": 6.71885753613214, "grad_norm": 8.655662536621094, "learning_rate": 4.6640571231933934e-05, "loss": 1.1262, "step": 39050 }, { "epoch": 6.727460426703372, "grad_norm": 7.907524108886719, "learning_rate": 4.6636269786648315e-05, "loss": 1.0819, "step": 39100 }, { "epoch": 6.7360633172746045, "grad_norm": 7.586410045623779, "learning_rate": 4.66319683413627e-05, "loss": 1.0976, "step": 39150 }, { "epoch": 6.744666207845836, "grad_norm": 8.85905647277832, "learning_rate": 4.6627666896077084e-05, "loss": 1.1507, "step": 39200 }, { "epoch": 6.753269098417068, "grad_norm": 6.875450134277344, "learning_rate": 4.6623365450791466e-05, "loss": 1.1094, "step": 39250 }, { "epoch": 6.7618719889883, "grad_norm": 7.654621601104736, "learning_rate": 4.6619064005505854e-05, "loss": 1.1239, "step": 39300 }, { "epoch": 6.770474879559532, "grad_norm": 9.164655685424805, "learning_rate": 4.6614762560220235e-05, "loss": 1.1518, "step": 39350 }, { "epoch": 6.779077770130764, "grad_norm": 10.718964576721191, "learning_rate": 4.661046111493462e-05, "loss": 1.0972, "step": 39400 }, { "epoch": 6.787680660701996, "grad_norm": 10.09834098815918, "learning_rate": 4.6606159669649004e-05, "loss": 1.0605, "step": 39450 }, { "epoch": 6.796283551273228, "grad_norm": 10.361448287963867, "learning_rate": 4.6601858224363385e-05, "loss": 1.0642, "step": 39500 }, { "epoch": 6.8048864418444595, "grad_norm": 8.866425514221191, "learning_rate": 4.659755677907777e-05, "loss": 1.1113, "step": 39550 }, { "epoch": 6.813489332415692, "grad_norm": 10.469717025756836, "learning_rate": 4.6593255333792154e-05, "loss": 1.0675, "step": 39600 }, { "epoch": 6.822092222986924, "grad_norm": 10.50460147857666, "learning_rate": 4.658895388850654e-05, "loss": 1.0869, "step": 39650 }, { "epoch": 6.830695113558155, "grad_norm": 8.89022445678711, "learning_rate": 4.658465244322092e-05, "loss": 1.068, "step": 39700 }, { "epoch": 6.839298004129388, "grad_norm": 8.539641380310059, "learning_rate": 4.6580350997935304e-05, "loss": 1.0774, "step": 39750 }, { "epoch": 6.8479008947006195, "grad_norm": 10.698287963867188, "learning_rate": 4.657604955264969e-05, "loss": 1.1351, "step": 39800 }, { "epoch": 6.856503785271851, "grad_norm": 9.181215286254883, "learning_rate": 4.657174810736408e-05, "loss": 1.1115, "step": 39850 }, { "epoch": 6.865106675843084, "grad_norm": 10.15242862701416, "learning_rate": 4.656744666207846e-05, "loss": 1.0789, "step": 39900 }, { "epoch": 6.873709566414315, "grad_norm": 7.583922386169434, "learning_rate": 4.656314521679285e-05, "loss": 1.0735, "step": 39950 }, { "epoch": 6.882312456985547, "grad_norm": 12.133856773376465, "learning_rate": 4.6558843771507224e-05, "loss": 1.0853, "step": 40000 }, { "epoch": 6.8909153475567795, "grad_norm": 8.991165161132812, "learning_rate": 4.655454232622161e-05, "loss": 1.0844, "step": 40050 }, { "epoch": 6.899518238128011, "grad_norm": 10.341141700744629, "learning_rate": 4.6550240880936e-05, "loss": 1.1246, "step": 40100 }, { "epoch": 6.908121128699243, "grad_norm": 7.50656270980835, "learning_rate": 4.654593943565038e-05, "loss": 1.0658, "step": 40150 }, { "epoch": 6.916724019270475, "grad_norm": 10.976350784301758, "learning_rate": 4.654163799036477e-05, "loss": 1.0962, "step": 40200 }, { "epoch": 6.925326909841707, "grad_norm": 8.160745620727539, "learning_rate": 4.653733654507914e-05, "loss": 1.0466, "step": 40250 }, { "epoch": 6.933929800412939, "grad_norm": 7.51157808303833, "learning_rate": 4.653303509979353e-05, "loss": 1.0923, "step": 40300 }, { "epoch": 6.942532690984171, "grad_norm": 9.476041793823242, "learning_rate": 4.652873365450792e-05, "loss": 1.0492, "step": 40350 }, { "epoch": 6.951135581555403, "grad_norm": 7.748928070068359, "learning_rate": 4.65244322092223e-05, "loss": 1.0782, "step": 40400 }, { "epoch": 6.959738472126634, "grad_norm": 8.588224411010742, "learning_rate": 4.652013076393669e-05, "loss": 1.106, "step": 40450 }, { "epoch": 6.968341362697867, "grad_norm": 12.306254386901855, "learning_rate": 4.651582931865107e-05, "loss": 1.0863, "step": 40500 }, { "epoch": 6.976944253269099, "grad_norm": 10.35886287689209, "learning_rate": 4.651152787336545e-05, "loss": 1.0906, "step": 40550 }, { "epoch": 6.98554714384033, "grad_norm": 8.48803424835205, "learning_rate": 4.650722642807984e-05, "loss": 1.0679, "step": 40600 }, { "epoch": 6.994150034411562, "grad_norm": 7.3028645515441895, "learning_rate": 4.650292498279422e-05, "loss": 1.058, "step": 40650 }, { "epoch": 7.0, "eval_accuracy": 0.5188435725348477, "eval_f1": 0.4943336847094926, "eval_f1_DuraRiadoRio_16x16": 0.3907828282828283, "eval_f1_Mole_16x16": 0.3916125221500295, "eval_f1_Quebrado_16x16": 0.6823413350871423, "eval_f1_RiadoRio_16x16": 0.5064808813998704, "eval_f1_RioFechado_16x16": 0.5004508566275925, "eval_loss": 1.5140295028686523, "eval_precision": 0.5621402667332677, "eval_precision_DuraRiadoRio_16x16": 0.7164351851851852, "eval_precision_Mole_16x16": 0.6127541589648798, "eval_precision_Quebrado_16x16": 0.5492323980942297, "eval_precision_RiadoRio_16x16": 0.4179144385026738, "eval_precision_RioFechado_16x16": 0.5143651529193698, "eval_recall": 0.517396341146435, "eval_recall_DuraRiadoRio_16x16": 0.2686631944444444, "eval_recall_Mole_16x16": 0.2877604166666667, "eval_recall_Quebrado_16x16": 0.9006076388888888, "eval_recall_RiadoRio_16x16": 0.6426809210526315, "eval_recall_RioFechado_16x16": 0.4872695346795435, "eval_runtime": 43.8266, "eval_samples_per_second": 265.182, "eval_steps_per_second": 16.588, "step": 40684 }, { "epoch": 7.0027529249827944, "grad_norm": 6.790908336639404, "learning_rate": 4.649862353750861e-05, "loss": 1.0772, "step": 40700 }, { "epoch": 7.011355815554026, "grad_norm": 7.16606330871582, "learning_rate": 4.649432209222299e-05, "loss": 1.0784, "step": 40750 }, { "epoch": 7.019958706125258, "grad_norm": 8.382502555847168, "learning_rate": 4.649002064693737e-05, "loss": 1.0593, "step": 40800 }, { "epoch": 7.02856159669649, "grad_norm": 12.827681541442871, "learning_rate": 4.648571920165176e-05, "loss": 1.0944, "step": 40850 }, { "epoch": 7.037164487267722, "grad_norm": 8.85312557220459, "learning_rate": 4.648141775636614e-05, "loss": 1.0579, "step": 40900 }, { "epoch": 7.045767377838954, "grad_norm": 9.052436828613281, "learning_rate": 4.647711631108053e-05, "loss": 1.1444, "step": 40950 }, { "epoch": 7.054370268410186, "grad_norm": 9.710103034973145, "learning_rate": 4.647281486579491e-05, "loss": 1.0986, "step": 41000 }, { "epoch": 7.062973158981418, "grad_norm": 7.4884185791015625, "learning_rate": 4.646851342050929e-05, "loss": 1.0516, "step": 41050 }, { "epoch": 7.071576049552649, "grad_norm": 6.857085704803467, "learning_rate": 4.646421197522368e-05, "loss": 1.0371, "step": 41100 }, { "epoch": 7.080178940123882, "grad_norm": 8.582945823669434, "learning_rate": 4.645991052993806e-05, "loss": 1.032, "step": 41150 }, { "epoch": 7.088781830695114, "grad_norm": 12.467230796813965, "learning_rate": 4.6455609084652446e-05, "loss": 1.1046, "step": 41200 }, { "epoch": 7.097384721266345, "grad_norm": 5.738219738006592, "learning_rate": 4.6451307639366834e-05, "loss": 1.041, "step": 41250 }, { "epoch": 7.105987611837578, "grad_norm": 8.97633171081543, "learning_rate": 4.6447006194081216e-05, "loss": 1.0601, "step": 41300 }, { "epoch": 7.114590502408809, "grad_norm": 10.43515396118164, "learning_rate": 4.64427047487956e-05, "loss": 1.1512, "step": 41350 }, { "epoch": 7.123193392980041, "grad_norm": 9.317268371582031, "learning_rate": 4.643840330350998e-05, "loss": 1.0558, "step": 41400 }, { "epoch": 7.131796283551274, "grad_norm": 9.622278213500977, "learning_rate": 4.6434101858224366e-05, "loss": 1.1393, "step": 41450 }, { "epoch": 7.140399174122505, "grad_norm": 8.324629783630371, "learning_rate": 4.6429800412938754e-05, "loss": 1.0899, "step": 41500 }, { "epoch": 7.149002064693737, "grad_norm": 9.124373435974121, "learning_rate": 4.6425498967653135e-05, "loss": 1.0665, "step": 41550 }, { "epoch": 7.157604955264969, "grad_norm": 10.592803001403809, "learning_rate": 4.6421197522367516e-05, "loss": 1.0579, "step": 41600 }, { "epoch": 7.166207845836201, "grad_norm": 9.500916481018066, "learning_rate": 4.64168960770819e-05, "loss": 1.0748, "step": 41650 }, { "epoch": 7.174810736407433, "grad_norm": 13.619081497192383, "learning_rate": 4.6412594631796285e-05, "loss": 1.0936, "step": 41700 }, { "epoch": 7.183413626978665, "grad_norm": 7.344498634338379, "learning_rate": 4.640829318651067e-05, "loss": 1.0902, "step": 41750 }, { "epoch": 7.192016517549897, "grad_norm": 9.184757232666016, "learning_rate": 4.6403991741225054e-05, "loss": 1.0642, "step": 41800 }, { "epoch": 7.2006194081211286, "grad_norm": 12.788204193115234, "learning_rate": 4.6399690295939436e-05, "loss": 1.1083, "step": 41850 }, { "epoch": 7.20922229869236, "grad_norm": 9.554183006286621, "learning_rate": 4.6395388850653824e-05, "loss": 1.1017, "step": 41900 }, { "epoch": 7.217825189263593, "grad_norm": 13.410782814025879, "learning_rate": 4.6391087405368205e-05, "loss": 1.0387, "step": 41950 }, { "epoch": 7.226428079834824, "grad_norm": 10.374594688415527, "learning_rate": 4.638678596008259e-05, "loss": 1.1315, "step": 42000 }, { "epoch": 7.235030970406056, "grad_norm": 7.749642372131348, "learning_rate": 4.6382484514796974e-05, "loss": 1.0482, "step": 42050 }, { "epoch": 7.243633860977289, "grad_norm": 13.512876510620117, "learning_rate": 4.637818306951136e-05, "loss": 1.0405, "step": 42100 }, { "epoch": 7.25223675154852, "grad_norm": 8.839202880859375, "learning_rate": 4.637388162422574e-05, "loss": 1.109, "step": 42150 }, { "epoch": 7.260839642119752, "grad_norm": 7.797042369842529, "learning_rate": 4.6369580178940124e-05, "loss": 1.065, "step": 42200 }, { "epoch": 7.269442532690984, "grad_norm": 9.571735382080078, "learning_rate": 4.636527873365451e-05, "loss": 1.0821, "step": 42250 }, { "epoch": 7.278045423262216, "grad_norm": 8.040657043457031, "learning_rate": 4.636097728836889e-05, "loss": 1.0798, "step": 42300 }, { "epoch": 7.286648313833448, "grad_norm": 6.881222724914551, "learning_rate": 4.635667584308328e-05, "loss": 1.1087, "step": 42350 }, { "epoch": 7.29525120440468, "grad_norm": 10.972506523132324, "learning_rate": 4.635237439779766e-05, "loss": 1.0811, "step": 42400 }, { "epoch": 7.303854094975912, "grad_norm": 7.688700199127197, "learning_rate": 4.6348072952512044e-05, "loss": 1.0944, "step": 42450 }, { "epoch": 7.3124569855471435, "grad_norm": 7.879290580749512, "learning_rate": 4.634377150722643e-05, "loss": 1.0609, "step": 42500 }, { "epoch": 7.321059876118376, "grad_norm": 11.565215110778809, "learning_rate": 4.633947006194081e-05, "loss": 1.0808, "step": 42550 }, { "epoch": 7.329662766689608, "grad_norm": 8.037055015563965, "learning_rate": 4.63351686166552e-05, "loss": 1.0386, "step": 42600 }, { "epoch": 7.338265657260839, "grad_norm": 10.586461067199707, "learning_rate": 4.633086717136958e-05, "loss": 1.1163, "step": 42650 }, { "epoch": 7.346868547832072, "grad_norm": 8.485228538513184, "learning_rate": 4.632656572608396e-05, "loss": 1.0758, "step": 42700 }, { "epoch": 7.3554714384033035, "grad_norm": 9.067122459411621, "learning_rate": 4.632226428079835e-05, "loss": 1.0555, "step": 42750 }, { "epoch": 7.364074328974535, "grad_norm": 8.212343215942383, "learning_rate": 4.631796283551273e-05, "loss": 1.0684, "step": 42800 }, { "epoch": 7.372677219545768, "grad_norm": 9.883347511291504, "learning_rate": 4.631366139022712e-05, "loss": 1.0955, "step": 42850 }, { "epoch": 7.381280110116999, "grad_norm": 9.860845565795898, "learning_rate": 4.630935994494151e-05, "loss": 1.0643, "step": 42900 }, { "epoch": 7.389883000688231, "grad_norm": 7.3349385261535645, "learning_rate": 4.630505849965588e-05, "loss": 1.0301, "step": 42950 }, { "epoch": 7.3984858912594635, "grad_norm": 9.053544998168945, "learning_rate": 4.630075705437027e-05, "loss": 1.0424, "step": 43000 }, { "epoch": 7.407088781830695, "grad_norm": 8.288087844848633, "learning_rate": 4.629645560908465e-05, "loss": 1.0707, "step": 43050 }, { "epoch": 7.415691672401927, "grad_norm": 9.007438659667969, "learning_rate": 4.629215416379904e-05, "loss": 1.0616, "step": 43100 }, { "epoch": 7.424294562973159, "grad_norm": 9.21951675415039, "learning_rate": 4.628785271851343e-05, "loss": 1.0113, "step": 43150 }, { "epoch": 7.432897453544391, "grad_norm": 8.351646423339844, "learning_rate": 4.62835512732278e-05, "loss": 1.0881, "step": 43200 }, { "epoch": 7.441500344115623, "grad_norm": 5.9820051193237305, "learning_rate": 4.627924982794219e-05, "loss": 1.0843, "step": 43250 }, { "epoch": 7.450103234686855, "grad_norm": 6.624630451202393, "learning_rate": 4.627494838265658e-05, "loss": 1.0396, "step": 43300 }, { "epoch": 7.458706125258087, "grad_norm": 6.1603569984436035, "learning_rate": 4.627064693737096e-05, "loss": 1.1245, "step": 43350 }, { "epoch": 7.4673090158293185, "grad_norm": 6.145068168640137, "learning_rate": 4.626634549208535e-05, "loss": 1.0393, "step": 43400 }, { "epoch": 7.475911906400551, "grad_norm": 7.875848770141602, "learning_rate": 4.626204404679973e-05, "loss": 1.072, "step": 43450 }, { "epoch": 7.484514796971783, "grad_norm": 10.855006217956543, "learning_rate": 4.625774260151411e-05, "loss": 1.0909, "step": 43500 }, { "epoch": 7.493117687543014, "grad_norm": 9.061909675598145, "learning_rate": 4.62534411562285e-05, "loss": 1.1003, "step": 43550 }, { "epoch": 7.501720578114247, "grad_norm": 7.4084858894348145, "learning_rate": 4.624913971094288e-05, "loss": 1.0883, "step": 43600 }, { "epoch": 7.5103234686854785, "grad_norm": 8.6874418258667, "learning_rate": 4.6244838265657266e-05, "loss": 1.1103, "step": 43650 }, { "epoch": 7.51892635925671, "grad_norm": 12.348627090454102, "learning_rate": 4.624053682037165e-05, "loss": 1.1283, "step": 43700 }, { "epoch": 7.527529249827942, "grad_norm": 10.374614715576172, "learning_rate": 4.623623537508603e-05, "loss": 1.0787, "step": 43750 }, { "epoch": 7.536132140399174, "grad_norm": 9.063237190246582, "learning_rate": 4.6231933929800417e-05, "loss": 1.04, "step": 43800 }, { "epoch": 7.544735030970406, "grad_norm": 10.992244720458984, "learning_rate": 4.62276324845148e-05, "loss": 1.0821, "step": 43850 }, { "epoch": 7.553337921541638, "grad_norm": 8.593378067016602, "learning_rate": 4.6223331039229186e-05, "loss": 1.0423, "step": 43900 }, { "epoch": 7.56194081211287, "grad_norm": 7.959821701049805, "learning_rate": 4.621902959394357e-05, "loss": 1.096, "step": 43950 }, { "epoch": 7.570543702684102, "grad_norm": 6.374561786651611, "learning_rate": 4.621472814865795e-05, "loss": 1.1136, "step": 44000 }, { "epoch": 7.5791465932553335, "grad_norm": 10.092625617980957, "learning_rate": 4.6210426703372336e-05, "loss": 1.0933, "step": 44050 }, { "epoch": 7.587749483826566, "grad_norm": 9.36583137512207, "learning_rate": 4.620612525808672e-05, "loss": 1.103, "step": 44100 }, { "epoch": 7.596352374397798, "grad_norm": 7.236337661743164, "learning_rate": 4.6201823812801105e-05, "loss": 1.0423, "step": 44150 }, { "epoch": 7.604955264969029, "grad_norm": 8.608985900878906, "learning_rate": 4.6197522367515486e-05, "loss": 1.087, "step": 44200 }, { "epoch": 7.613558155540262, "grad_norm": 11.279805183410645, "learning_rate": 4.6193220922229874e-05, "loss": 1.0908, "step": 44250 }, { "epoch": 7.6221610461114935, "grad_norm": 8.045662879943848, "learning_rate": 4.6188919476944255e-05, "loss": 1.0759, "step": 44300 }, { "epoch": 7.630763936682725, "grad_norm": 9.500272750854492, "learning_rate": 4.6184618031658637e-05, "loss": 1.0342, "step": 44350 }, { "epoch": 7.639366827253958, "grad_norm": 9.361437797546387, "learning_rate": 4.6180316586373024e-05, "loss": 1.1099, "step": 44400 }, { "epoch": 7.647969717825189, "grad_norm": 9.500378608703613, "learning_rate": 4.6176015141087406e-05, "loss": 1.0901, "step": 44450 }, { "epoch": 7.656572608396421, "grad_norm": 9.235288619995117, "learning_rate": 4.6171713695801794e-05, "loss": 1.1053, "step": 44500 }, { "epoch": 7.6651754989676535, "grad_norm": 12.323328971862793, "learning_rate": 4.6167412250516175e-05, "loss": 1.0478, "step": 44550 }, { "epoch": 7.673778389538885, "grad_norm": 9.603800773620605, "learning_rate": 4.6163110805230556e-05, "loss": 1.0812, "step": 44600 }, { "epoch": 7.682381280110117, "grad_norm": 8.22758960723877, "learning_rate": 4.6158809359944944e-05, "loss": 1.11, "step": 44650 }, { "epoch": 7.690984170681349, "grad_norm": 12.510063171386719, "learning_rate": 4.6154507914659325e-05, "loss": 1.0981, "step": 44700 }, { "epoch": 7.699587061252581, "grad_norm": 7.120272636413574, "learning_rate": 4.615020646937371e-05, "loss": 1.0667, "step": 44750 }, { "epoch": 7.708189951823813, "grad_norm": 7.990087509155273, "learning_rate": 4.61459050240881e-05, "loss": 1.0914, "step": 44800 }, { "epoch": 7.716792842395044, "grad_norm": 10.369135856628418, "learning_rate": 4.6141603578802475e-05, "loss": 1.0736, "step": 44850 }, { "epoch": 7.725395732966277, "grad_norm": 7.007133483886719, "learning_rate": 4.613730213351686e-05, "loss": 1.0781, "step": 44900 }, { "epoch": 7.733998623537508, "grad_norm": 5.75659704208374, "learning_rate": 4.613300068823125e-05, "loss": 1.0354, "step": 44950 }, { "epoch": 7.74260151410874, "grad_norm": 7.6367669105529785, "learning_rate": 4.612869924294563e-05, "loss": 1.1058, "step": 45000 }, { "epoch": 7.751204404679973, "grad_norm": 9.408629417419434, "learning_rate": 4.612439779766002e-05, "loss": 1.0852, "step": 45050 }, { "epoch": 7.759807295251204, "grad_norm": 8.513847351074219, "learning_rate": 4.6120096352374395e-05, "loss": 1.0383, "step": 45100 }, { "epoch": 7.768410185822436, "grad_norm": 9.073342323303223, "learning_rate": 4.611579490708878e-05, "loss": 1.079, "step": 45150 }, { "epoch": 7.7770130763936685, "grad_norm": 10.37783145904541, "learning_rate": 4.611149346180317e-05, "loss": 1.1194, "step": 45200 }, { "epoch": 7.7856159669649, "grad_norm": 13.978628158569336, "learning_rate": 4.610719201651755e-05, "loss": 1.0686, "step": 45250 }, { "epoch": 7.794218857536132, "grad_norm": 6.687712669372559, "learning_rate": 4.610289057123194e-05, "loss": 1.0653, "step": 45300 }, { "epoch": 7.802821748107364, "grad_norm": 9.787546157836914, "learning_rate": 4.609858912594632e-05, "loss": 1.079, "step": 45350 }, { "epoch": 7.811424638678596, "grad_norm": 7.699717044830322, "learning_rate": 4.60942876806607e-05, "loss": 1.1179, "step": 45400 }, { "epoch": 7.820027529249828, "grad_norm": 18.60771942138672, "learning_rate": 4.608998623537509e-05, "loss": 1.0646, "step": 45450 }, { "epoch": 7.82863041982106, "grad_norm": 9.905989646911621, "learning_rate": 4.608568479008947e-05, "loss": 1.0891, "step": 45500 }, { "epoch": 7.837233310392292, "grad_norm": 9.094120025634766, "learning_rate": 4.608138334480386e-05, "loss": 1.0909, "step": 45550 }, { "epoch": 7.845836200963523, "grad_norm": 7.555301666259766, "learning_rate": 4.607708189951824e-05, "loss": 1.0822, "step": 45600 }, { "epoch": 7.854439091534756, "grad_norm": 11.657919883728027, "learning_rate": 4.607278045423262e-05, "loss": 1.0528, "step": 45650 }, { "epoch": 7.863041982105988, "grad_norm": 15.088217735290527, "learning_rate": 4.606847900894701e-05, "loss": 1.0458, "step": 45700 }, { "epoch": 7.871644872677219, "grad_norm": 8.566354751586914, "learning_rate": 4.606417756366139e-05, "loss": 1.0308, "step": 45750 }, { "epoch": 7.880247763248452, "grad_norm": 7.664295673370361, "learning_rate": 4.605987611837578e-05, "loss": 1.0319, "step": 45800 }, { "epoch": 7.888850653819683, "grad_norm": 8.495899200439453, "learning_rate": 4.605557467309016e-05, "loss": 1.0534, "step": 45850 }, { "epoch": 7.897453544390915, "grad_norm": 9.094196319580078, "learning_rate": 4.605127322780454e-05, "loss": 1.0765, "step": 45900 }, { "epoch": 7.906056434962148, "grad_norm": 9.198328018188477, "learning_rate": 4.604697178251893e-05, "loss": 1.1039, "step": 45950 }, { "epoch": 7.914659325533379, "grad_norm": 8.233175277709961, "learning_rate": 4.604267033723331e-05, "loss": 1.0639, "step": 46000 }, { "epoch": 7.923262216104611, "grad_norm": 7.9715728759765625, "learning_rate": 4.60383688919477e-05, "loss": 1.0724, "step": 46050 }, { "epoch": 7.931865106675843, "grad_norm": 8.191460609436035, "learning_rate": 4.603406744666208e-05, "loss": 1.1239, "step": 46100 }, { "epoch": 7.940467997247075, "grad_norm": 13.209691047668457, "learning_rate": 4.602976600137646e-05, "loss": 1.0728, "step": 46150 }, { "epoch": 7.949070887818307, "grad_norm": 8.51311206817627, "learning_rate": 4.602546455609085e-05, "loss": 1.032, "step": 46200 }, { "epoch": 7.957673778389539, "grad_norm": 7.762655258178711, "learning_rate": 4.602116311080523e-05, "loss": 1.0503, "step": 46250 }, { "epoch": 7.966276668960771, "grad_norm": 6.508998870849609, "learning_rate": 4.601686166551962e-05, "loss": 1.1271, "step": 46300 }, { "epoch": 7.974879559532003, "grad_norm": 9.160560607910156, "learning_rate": 4.6012560220234005e-05, "loss": 1.1471, "step": 46350 }, { "epoch": 7.983482450103235, "grad_norm": 7.9591193199157715, "learning_rate": 4.6008258774948387e-05, "loss": 1.05, "step": 46400 }, { "epoch": 7.992085340674467, "grad_norm": 7.220909595489502, "learning_rate": 4.600395732966277e-05, "loss": 1.0996, "step": 46450 }, { "epoch": 8.0, "eval_accuracy": 0.5762347272414386, "eval_f1": 0.5530189327780126, "eval_f1_DuraRiadoRio_16x16": 0.5288178376104333, "eval_f1_Mole_16x16": 0.6110086100861009, "eval_f1_Quebrado_16x16": 0.773500097713504, "eval_f1_RiadoRio_16x16": 0.31115107913669066, "eval_f1_RioFechado_16x16": 0.5406170393433343, "eval_loss": 1.211322546005249, "eval_precision": 0.6049494112930157, "eval_precision_DuraRiadoRio_16x16": 0.513061224489796, "eval_precision_Mole_16x16": 0.47309523809523807, "eval_precision_Quebrado_16x16": 0.7035193743334518, "eval_precision_RiadoRio_16x16": 0.5741150442477876, "eval_precision_RioFechado_16x16": 0.7609561752988048, "eval_recall": 0.5799118162092017, "eval_recall_DuraRiadoRio_16x16": 0.5455729166666666, "eval_recall_Mole_16x16": 0.8624131944444444, "eval_recall_Quebrado_16x16": 0.8589409722222222, "eval_recall_RiadoRio_16x16": 0.21340460526315788, "eval_recall_RioFechado_16x16": 0.4192273924495171, "eval_runtime": 43.6651, "eval_samples_per_second": 266.162, "eval_steps_per_second": 16.649, "step": 46496 }, { "epoch": 8.000688231245698, "grad_norm": 6.582836151123047, "learning_rate": 4.599965588437715e-05, "loss": 1.1282, "step": 46500 }, { "epoch": 8.009291121816931, "grad_norm": 8.755108833312988, "learning_rate": 4.599535443909154e-05, "loss": 1.0838, "step": 46550 }, { "epoch": 8.017894012388162, "grad_norm": 10.495073318481445, "learning_rate": 4.5991052993805925e-05, "loss": 1.0478, "step": 46600 }, { "epoch": 8.026496902959394, "grad_norm": 9.42346477508545, "learning_rate": 4.5986751548520306e-05, "loss": 1.105, "step": 46650 }, { "epoch": 8.035099793530627, "grad_norm": 8.100893020629883, "learning_rate": 4.598245010323469e-05, "loss": 1.0758, "step": 46700 }, { "epoch": 8.043702684101858, "grad_norm": 11.174704551696777, "learning_rate": 4.597814865794907e-05, "loss": 1.0594, "step": 46750 }, { "epoch": 8.05230557467309, "grad_norm": 6.2121076583862305, "learning_rate": 4.5973847212663456e-05, "loss": 1.0416, "step": 46800 }, { "epoch": 8.060908465244323, "grad_norm": 10.287881851196289, "learning_rate": 4.5969545767377844e-05, "loss": 1.0164, "step": 46850 }, { "epoch": 8.069511355815553, "grad_norm": 6.438128471374512, "learning_rate": 4.5965244322092225e-05, "loss": 1.0574, "step": 46900 }, { "epoch": 8.078114246386786, "grad_norm": 6.6518049240112305, "learning_rate": 4.596094287680661e-05, "loss": 1.017, "step": 46950 }, { "epoch": 8.086717136958018, "grad_norm": 7.970098972320557, "learning_rate": 4.5956641431520995e-05, "loss": 1.0903, "step": 47000 }, { "epoch": 8.09532002752925, "grad_norm": 7.873311996459961, "learning_rate": 4.5952339986235376e-05, "loss": 1.0396, "step": 47050 }, { "epoch": 8.103922918100482, "grad_norm": 7.313480377197266, "learning_rate": 4.5948038540949764e-05, "loss": 1.0529, "step": 47100 }, { "epoch": 8.112525808671714, "grad_norm": 9.713329315185547, "learning_rate": 4.5943737095664145e-05, "loss": 1.0782, "step": 47150 }, { "epoch": 8.121128699242945, "grad_norm": 9.48261833190918, "learning_rate": 4.593943565037853e-05, "loss": 1.1004, "step": 47200 }, { "epoch": 8.129731589814178, "grad_norm": 9.808256149291992, "learning_rate": 4.5935134205092914e-05, "loss": 1.0392, "step": 47250 }, { "epoch": 8.13833448038541, "grad_norm": 8.46033000946045, "learning_rate": 4.5930832759807295e-05, "loss": 1.099, "step": 47300 }, { "epoch": 8.14693737095664, "grad_norm": 10.821392059326172, "learning_rate": 4.592653131452168e-05, "loss": 1.0126, "step": 47350 }, { "epoch": 8.155540261527873, "grad_norm": 9.12646484375, "learning_rate": 4.5922229869236064e-05, "loss": 1.0083, "step": 47400 }, { "epoch": 8.164143152099106, "grad_norm": 8.36214828491211, "learning_rate": 4.591792842395045e-05, "loss": 1.0872, "step": 47450 }, { "epoch": 8.172746042670337, "grad_norm": 12.642848014831543, "learning_rate": 4.591362697866483e-05, "loss": 1.0617, "step": 47500 }, { "epoch": 8.18134893324157, "grad_norm": 8.734383583068848, "learning_rate": 4.5909325533379215e-05, "loss": 1.0394, "step": 47550 }, { "epoch": 8.189951823812802, "grad_norm": 9.485404968261719, "learning_rate": 4.59050240880936e-05, "loss": 1.0468, "step": 47600 }, { "epoch": 8.198554714384032, "grad_norm": 8.559128761291504, "learning_rate": 4.5900722642807984e-05, "loss": 1.0825, "step": 47650 }, { "epoch": 8.207157604955265, "grad_norm": 10.561463356018066, "learning_rate": 4.589642119752237e-05, "loss": 1.049, "step": 47700 }, { "epoch": 8.215760495526498, "grad_norm": 10.665487289428711, "learning_rate": 4.589211975223676e-05, "loss": 1.0611, "step": 47750 }, { "epoch": 8.224363386097728, "grad_norm": 8.244600296020508, "learning_rate": 4.5887818306951134e-05, "loss": 1.0495, "step": 47800 }, { "epoch": 8.23296627666896, "grad_norm": 7.471245765686035, "learning_rate": 4.588351686166552e-05, "loss": 1.0488, "step": 47850 }, { "epoch": 8.241569167240193, "grad_norm": 8.320267677307129, "learning_rate": 4.58792154163799e-05, "loss": 1.0461, "step": 47900 }, { "epoch": 8.250172057811424, "grad_norm": 8.291898727416992, "learning_rate": 4.587491397109429e-05, "loss": 1.0414, "step": 47950 }, { "epoch": 8.258774948382657, "grad_norm": 7.596700668334961, "learning_rate": 4.587061252580868e-05, "loss": 1.0598, "step": 48000 }, { "epoch": 8.26737783895389, "grad_norm": 7.287243843078613, "learning_rate": 4.586631108052305e-05, "loss": 1.0642, "step": 48050 }, { "epoch": 8.27598072952512, "grad_norm": 11.552695274353027, "learning_rate": 4.586200963523744e-05, "loss": 1.0515, "step": 48100 }, { "epoch": 8.284583620096353, "grad_norm": 6.160436630249023, "learning_rate": 4.585770818995182e-05, "loss": 1.0319, "step": 48150 }, { "epoch": 8.293186510667585, "grad_norm": 9.577202796936035, "learning_rate": 4.585340674466621e-05, "loss": 1.0468, "step": 48200 }, { "epoch": 8.301789401238816, "grad_norm": 9.340706825256348, "learning_rate": 4.58491052993806e-05, "loss": 1.0299, "step": 48250 }, { "epoch": 8.310392291810048, "grad_norm": 9.632691383361816, "learning_rate": 4.584480385409497e-05, "loss": 1.0298, "step": 48300 }, { "epoch": 8.31899518238128, "grad_norm": 8.307758331298828, "learning_rate": 4.584050240880936e-05, "loss": 1.0762, "step": 48350 }, { "epoch": 8.327598072952512, "grad_norm": 10.546927452087402, "learning_rate": 4.583620096352375e-05, "loss": 1.0723, "step": 48400 }, { "epoch": 8.336200963523744, "grad_norm": 11.395562171936035, "learning_rate": 4.583189951823813e-05, "loss": 1.0519, "step": 48450 }, { "epoch": 8.344803854094977, "grad_norm": 11.359604835510254, "learning_rate": 4.582759807295252e-05, "loss": 1.0734, "step": 48500 }, { "epoch": 8.353406744666207, "grad_norm": 8.852102279663086, "learning_rate": 4.58232966276669e-05, "loss": 1.0849, "step": 48550 }, { "epoch": 8.36200963523744, "grad_norm": 10.245153427124023, "learning_rate": 4.581899518238128e-05, "loss": 1.0684, "step": 48600 }, { "epoch": 8.370612525808673, "grad_norm": 9.14461898803711, "learning_rate": 4.581469373709567e-05, "loss": 1.0429, "step": 48650 }, { "epoch": 8.379215416379903, "grad_norm": 10.557744026184082, "learning_rate": 4.581039229181005e-05, "loss": 1.045, "step": 48700 }, { "epoch": 8.387818306951136, "grad_norm": 6.793970584869385, "learning_rate": 4.580609084652444e-05, "loss": 1.0363, "step": 48750 }, { "epoch": 8.396421197522368, "grad_norm": 12.062253952026367, "learning_rate": 4.580178940123882e-05, "loss": 1.0318, "step": 48800 }, { "epoch": 8.4050240880936, "grad_norm": 9.768625259399414, "learning_rate": 4.57974879559532e-05, "loss": 1.0601, "step": 48850 }, { "epoch": 8.413626978664832, "grad_norm": 12.249754905700684, "learning_rate": 4.579318651066759e-05, "loss": 1.0409, "step": 48900 }, { "epoch": 8.422229869236062, "grad_norm": 12.913392066955566, "learning_rate": 4.578888506538197e-05, "loss": 1.0527, "step": 48950 }, { "epoch": 8.430832759807295, "grad_norm": 8.205379486083984, "learning_rate": 4.5784583620096357e-05, "loss": 1.0626, "step": 49000 }, { "epoch": 8.439435650378527, "grad_norm": 7.102281093597412, "learning_rate": 4.578028217481074e-05, "loss": 1.0385, "step": 49050 }, { "epoch": 8.44803854094976, "grad_norm": 9.794292449951172, "learning_rate": 4.5775980729525126e-05, "loss": 0.9925, "step": 49100 }, { "epoch": 8.45664143152099, "grad_norm": 9.823112487792969, "learning_rate": 4.577167928423951e-05, "loss": 1.0737, "step": 49150 }, { "epoch": 8.465244322092223, "grad_norm": 8.594245910644531, "learning_rate": 4.576737783895389e-05, "loss": 1.1228, "step": 49200 }, { "epoch": 8.473847212663454, "grad_norm": 9.170352935791016, "learning_rate": 4.5763076393668276e-05, "loss": 1.0884, "step": 49250 }, { "epoch": 8.482450103234687, "grad_norm": 13.118249893188477, "learning_rate": 4.575877494838266e-05, "loss": 1.0395, "step": 49300 }, { "epoch": 8.49105299380592, "grad_norm": 9.796896934509277, "learning_rate": 4.5754473503097045e-05, "loss": 1.055, "step": 49350 }, { "epoch": 8.49965588437715, "grad_norm": 6.871298789978027, "learning_rate": 4.5750172057811426e-05, "loss": 1.0449, "step": 49400 }, { "epoch": 8.508258774948382, "grad_norm": 9.417366981506348, "learning_rate": 4.574587061252581e-05, "loss": 1.0944, "step": 49450 }, { "epoch": 8.516861665519615, "grad_norm": 7.591712474822998, "learning_rate": 4.5741569167240195e-05, "loss": 1.0284, "step": 49500 }, { "epoch": 8.525464556090846, "grad_norm": 7.6349358558654785, "learning_rate": 4.5737267721954577e-05, "loss": 1.107, "step": 49550 }, { "epoch": 8.534067446662078, "grad_norm": 10.136457443237305, "learning_rate": 4.5732966276668965e-05, "loss": 1.0909, "step": 49600 }, { "epoch": 8.54267033723331, "grad_norm": 8.240406036376953, "learning_rate": 4.5728664831383346e-05, "loss": 1.0897, "step": 49650 }, { "epoch": 8.551273227804542, "grad_norm": 8.91602611541748, "learning_rate": 4.572436338609773e-05, "loss": 1.0308, "step": 49700 }, { "epoch": 8.559876118375774, "grad_norm": 10.878846168518066, "learning_rate": 4.5720061940812115e-05, "loss": 1.0623, "step": 49750 }, { "epoch": 8.568479008947007, "grad_norm": 11.309752464294434, "learning_rate": 4.57157604955265e-05, "loss": 1.0081, "step": 49800 }, { "epoch": 8.577081899518237, "grad_norm": 10.526039123535156, "learning_rate": 4.5711459050240884e-05, "loss": 1.0604, "step": 49850 }, { "epoch": 8.58568479008947, "grad_norm": 8.650833129882812, "learning_rate": 4.570715760495527e-05, "loss": 1.0413, "step": 49900 }, { "epoch": 8.594287680660702, "grad_norm": 13.197624206542969, "learning_rate": 4.5702856159669646e-05, "loss": 1.1004, "step": 49950 }, { "epoch": 8.602890571231933, "grad_norm": 10.853604316711426, "learning_rate": 4.5698554714384034e-05, "loss": 1.034, "step": 50000 }, { "epoch": 8.611493461803166, "grad_norm": 5.854525566101074, "learning_rate": 4.569425326909842e-05, "loss": 1.0513, "step": 50050 }, { "epoch": 8.620096352374398, "grad_norm": 10.302093505859375, "learning_rate": 4.56899518238128e-05, "loss": 1.0588, "step": 50100 }, { "epoch": 8.628699242945629, "grad_norm": 9.993099212646484, "learning_rate": 4.568565037852719e-05, "loss": 1.0575, "step": 50150 }, { "epoch": 8.637302133516862, "grad_norm": 10.95141887664795, "learning_rate": 4.5681348933241566e-05, "loss": 1.0899, "step": 50200 }, { "epoch": 8.645905024088094, "grad_norm": 7.9068379402160645, "learning_rate": 4.5677047487955954e-05, "loss": 1.0749, "step": 50250 }, { "epoch": 8.654507914659325, "grad_norm": 10.570439338684082, "learning_rate": 4.567274604267034e-05, "loss": 1.0402, "step": 50300 }, { "epoch": 8.663110805230557, "grad_norm": 10.024449348449707, "learning_rate": 4.566844459738472e-05, "loss": 1.0845, "step": 50350 }, { "epoch": 8.67171369580179, "grad_norm": 10.732593536376953, "learning_rate": 4.566414315209911e-05, "loss": 1.0765, "step": 50400 }, { "epoch": 8.68031658637302, "grad_norm": 11.805743217468262, "learning_rate": 4.565984170681349e-05, "loss": 1.0074, "step": 50450 }, { "epoch": 8.688919476944253, "grad_norm": 8.290992736816406, "learning_rate": 4.565554026152787e-05, "loss": 1.1365, "step": 50500 }, { "epoch": 8.697522367515486, "grad_norm": 8.844246864318848, "learning_rate": 4.565123881624226e-05, "loss": 1.0695, "step": 50550 }, { "epoch": 8.706125258086717, "grad_norm": 8.396626472473145, "learning_rate": 4.564693737095664e-05, "loss": 1.0251, "step": 50600 }, { "epoch": 8.714728148657949, "grad_norm": 10.907661437988281, "learning_rate": 4.564263592567103e-05, "loss": 1.0824, "step": 50650 }, { "epoch": 8.723331039229182, "grad_norm": 12.257349014282227, "learning_rate": 4.563833448038541e-05, "loss": 1.0521, "step": 50700 }, { "epoch": 8.731933929800412, "grad_norm": 8.716961860656738, "learning_rate": 4.563403303509979e-05, "loss": 1.0955, "step": 50750 }, { "epoch": 8.740536820371645, "grad_norm": 14.914793968200684, "learning_rate": 4.562973158981418e-05, "loss": 1.0016, "step": 50800 }, { "epoch": 8.749139710942877, "grad_norm": 14.021855354309082, "learning_rate": 4.562543014452856e-05, "loss": 1.0437, "step": 50850 }, { "epoch": 8.757742601514108, "grad_norm": 8.595991134643555, "learning_rate": 4.562112869924295e-05, "loss": 1.1229, "step": 50900 }, { "epoch": 8.76634549208534, "grad_norm": 10.522969245910645, "learning_rate": 4.561682725395733e-05, "loss": 1.0913, "step": 50950 }, { "epoch": 8.774948382656573, "grad_norm": 9.497330665588379, "learning_rate": 4.561252580867171e-05, "loss": 1.069, "step": 51000 }, { "epoch": 8.783551273227804, "grad_norm": 10.872665405273438, "learning_rate": 4.56082243633861e-05, "loss": 1.0229, "step": 51050 }, { "epoch": 8.792154163799037, "grad_norm": 11.742667198181152, "learning_rate": 4.560392291810048e-05, "loss": 1.1082, "step": 51100 }, { "epoch": 8.80075705437027, "grad_norm": 7.22667121887207, "learning_rate": 4.559962147281487e-05, "loss": 1.0743, "step": 51150 }, { "epoch": 8.8093599449415, "grad_norm": 9.836243629455566, "learning_rate": 4.559532002752926e-05, "loss": 1.1443, "step": 51200 }, { "epoch": 8.817962835512732, "grad_norm": 9.716800689697266, "learning_rate": 4.559101858224364e-05, "loss": 1.0404, "step": 51250 }, { "epoch": 8.826565726083965, "grad_norm": 7.680674076080322, "learning_rate": 4.558671713695802e-05, "loss": 1.0716, "step": 51300 }, { "epoch": 8.835168616655196, "grad_norm": 10.107732772827148, "learning_rate": 4.55824156916724e-05, "loss": 1.0772, "step": 51350 }, { "epoch": 8.843771507226428, "grad_norm": 6.258971214294434, "learning_rate": 4.557811424638679e-05, "loss": 1.1092, "step": 51400 }, { "epoch": 8.85237439779766, "grad_norm": 10.128870010375977, "learning_rate": 4.5573812801101176e-05, "loss": 1.0677, "step": 51450 }, { "epoch": 8.860977288368892, "grad_norm": 10.545123100280762, "learning_rate": 4.556951135581556e-05, "loss": 1.0424, "step": 51500 }, { "epoch": 8.869580178940124, "grad_norm": 23.011104583740234, "learning_rate": 4.556520991052994e-05, "loss": 1.1058, "step": 51550 }, { "epoch": 8.878183069511357, "grad_norm": 10.704680442810059, "learning_rate": 4.556090846524432e-05, "loss": 1.082, "step": 51600 }, { "epoch": 8.886785960082587, "grad_norm": 9.081031799316406, "learning_rate": 4.555660701995871e-05, "loss": 1.0407, "step": 51650 }, { "epoch": 8.89538885065382, "grad_norm": 9.241740226745605, "learning_rate": 4.5552305574673096e-05, "loss": 1.108, "step": 51700 }, { "epoch": 8.903991741225052, "grad_norm": 11.827104568481445, "learning_rate": 4.554800412938748e-05, "loss": 1.0387, "step": 51750 }, { "epoch": 8.912594631796283, "grad_norm": 12.719792366027832, "learning_rate": 4.554370268410186e-05, "loss": 1.0733, "step": 51800 }, { "epoch": 8.921197522367516, "grad_norm": 10.82470989227295, "learning_rate": 4.5539401238816246e-05, "loss": 1.0822, "step": 51850 }, { "epoch": 8.929800412938746, "grad_norm": 7.0495076179504395, "learning_rate": 4.553509979353063e-05, "loss": 1.0895, "step": 51900 }, { "epoch": 8.938403303509979, "grad_norm": 9.935005187988281, "learning_rate": 4.5530798348245015e-05, "loss": 1.048, "step": 51950 }, { "epoch": 8.947006194081212, "grad_norm": 10.314850807189941, "learning_rate": 4.5526496902959396e-05, "loss": 1.0298, "step": 52000 }, { "epoch": 8.955609084652444, "grad_norm": 10.768085479736328, "learning_rate": 4.5522195457673784e-05, "loss": 1.0413, "step": 52050 }, { "epoch": 8.964211975223675, "grad_norm": 11.239855766296387, "learning_rate": 4.5517894012388165e-05, "loss": 1.0842, "step": 52100 }, { "epoch": 8.972814865794907, "grad_norm": 9.263214111328125, "learning_rate": 4.551359256710255e-05, "loss": 1.0552, "step": 52150 }, { "epoch": 8.981417756366138, "grad_norm": 7.623110771179199, "learning_rate": 4.5509291121816935e-05, "loss": 1.095, "step": 52200 }, { "epoch": 8.99002064693737, "grad_norm": 8.095392227172852, "learning_rate": 4.5504989676531316e-05, "loss": 1.0559, "step": 52250 }, { "epoch": 8.998623537508603, "grad_norm": 11.937847137451172, "learning_rate": 4.5500688231245704e-05, "loss": 1.0587, "step": 52300 }, { "epoch": 9.0, "eval_accuracy": 0.4198933057993461, "eval_f1": 0.37835918327344636, "eval_f1_DuraRiadoRio_16x16": 0.320204023802777, "eval_f1_Mole_16x16": 0.30030395136778115, "eval_f1_Quebrado_16x16": 0.5330271511549372, "eval_f1_RiadoRio_16x16": 0.2253782031491201, "eval_f1_RioFechado_16x16": 0.5128825868926163, "eval_loss": 2.0437774658203125, "eval_precision": 0.4449159176430008, "eval_precision_DuraRiadoRio_16x16": 0.46122448979591835, "eval_precision_Mole_16x16": 0.5010141987829615, "eval_precision_Quebrado_16x16": 0.3869386154147872, "eval_precision_RiadoRio_16x16": 0.45229244114002476, "eval_precision_RioFechado_16x16": 0.4231098430813124, "eval_recall": 0.4234128233317417, "eval_recall_DuraRiadoRio_16x16": 0.24522569444444445, "eval_recall_Mole_16x16": 0.2144097222222222, "eval_recall_Quebrado_16x16": 0.8563368055555556, "eval_recall_RiadoRio_16x16": 0.15008223684210525, "eval_recall_RioFechado_16x16": 0.651009657594381, "eval_runtime": 44.2486, "eval_samples_per_second": 262.652, "eval_steps_per_second": 16.43, "step": 52308 }, { "epoch": 9.007226428079834, "grad_norm": 9.922981262207031, "learning_rate": 4.5496386785960085e-05, "loss": 1.0743, "step": 52350 }, { "epoch": 9.015829318651067, "grad_norm": 8.979776382446289, "learning_rate": 4.5492085340674466e-05, "loss": 1.0588, "step": 52400 }, { "epoch": 9.024432209222299, "grad_norm": 8.762675285339355, "learning_rate": 4.5487783895388854e-05, "loss": 1.0164, "step": 52450 }, { "epoch": 9.03303509979353, "grad_norm": 7.24167537689209, "learning_rate": 4.5483482450103235e-05, "loss": 1.0247, "step": 52500 }, { "epoch": 9.041637990364762, "grad_norm": 9.22210693359375, "learning_rate": 4.547918100481762e-05, "loss": 1.0245, "step": 52550 }, { "epoch": 9.050240880935995, "grad_norm": 10.754871368408203, "learning_rate": 4.5474879559532004e-05, "loss": 1.0713, "step": 52600 }, { "epoch": 9.058843771507226, "grad_norm": 8.627971649169922, "learning_rate": 4.5470578114246385e-05, "loss": 1.0055, "step": 52650 }, { "epoch": 9.067446662078458, "grad_norm": 6.284625053405762, "learning_rate": 4.5466276668960773e-05, "loss": 1.0748, "step": 52700 }, { "epoch": 9.07604955264969, "grad_norm": 10.268670082092285, "learning_rate": 4.5461975223675155e-05, "loss": 1.0474, "step": 52750 }, { "epoch": 9.084652443220921, "grad_norm": 7.883617401123047, "learning_rate": 4.545767377838954e-05, "loss": 1.0142, "step": 52800 }, { "epoch": 9.093255333792154, "grad_norm": 9.324201583862305, "learning_rate": 4.545337233310393e-05, "loss": 1.0562, "step": 52850 }, { "epoch": 9.101858224363387, "grad_norm": 10.69708251953125, "learning_rate": 4.5449070887818305e-05, "loss": 1.0408, "step": 52900 }, { "epoch": 9.110461114934617, "grad_norm": 10.481953620910645, "learning_rate": 4.544476944253269e-05, "loss": 1.0917, "step": 52950 }, { "epoch": 9.11906400550585, "grad_norm": 8.789783477783203, "learning_rate": 4.5440467997247074e-05, "loss": 1.0082, "step": 53000 }, { "epoch": 9.127666896077082, "grad_norm": 9.5361328125, "learning_rate": 4.543616655196146e-05, "loss": 1.039, "step": 53050 }, { "epoch": 9.136269786648313, "grad_norm": 12.38455581665039, "learning_rate": 4.543186510667585e-05, "loss": 1.0759, "step": 53100 }, { "epoch": 9.144872677219546, "grad_norm": 8.174059867858887, "learning_rate": 4.5427563661390224e-05, "loss": 1.0123, "step": 53150 }, { "epoch": 9.153475567790778, "grad_norm": 9.85912036895752, "learning_rate": 4.542326221610461e-05, "loss": 1.0848, "step": 53200 }, { "epoch": 9.162078458362009, "grad_norm": 10.66696548461914, "learning_rate": 4.5418960770819e-05, "loss": 1.0107, "step": 53250 }, { "epoch": 9.170681348933241, "grad_norm": 7.345121383666992, "learning_rate": 4.541465932553338e-05, "loss": 1.0487, "step": 53300 }, { "epoch": 9.179284239504474, "grad_norm": 9.421292304992676, "learning_rate": 4.541035788024777e-05, "loss": 1.0731, "step": 53350 }, { "epoch": 9.187887130075705, "grad_norm": 6.793117046356201, "learning_rate": 4.540605643496215e-05, "loss": 1.0622, "step": 53400 }, { "epoch": 9.196490020646937, "grad_norm": 8.584726333618164, "learning_rate": 4.540175498967653e-05, "loss": 1.0923, "step": 53450 }, { "epoch": 9.20509291121817, "grad_norm": 7.966603755950928, "learning_rate": 4.539745354439092e-05, "loss": 1.0506, "step": 53500 }, { "epoch": 9.2136958017894, "grad_norm": 7.739331245422363, "learning_rate": 4.53931520991053e-05, "loss": 1.0291, "step": 53550 }, { "epoch": 9.222298692360633, "grad_norm": 12.210006713867188, "learning_rate": 4.538885065381969e-05, "loss": 1.0126, "step": 53600 }, { "epoch": 9.230901582931866, "grad_norm": 12.468774795532227, "learning_rate": 4.538454920853407e-05, "loss": 0.9583, "step": 53650 }, { "epoch": 9.239504473503096, "grad_norm": 10.049121856689453, "learning_rate": 4.538024776324845e-05, "loss": 1.0254, "step": 53700 }, { "epoch": 9.248107364074329, "grad_norm": 8.355347633361816, "learning_rate": 4.537594631796284e-05, "loss": 1.0632, "step": 53750 }, { "epoch": 9.256710254645562, "grad_norm": 8.189370155334473, "learning_rate": 4.537164487267722e-05, "loss": 1.0077, "step": 53800 }, { "epoch": 9.265313145216792, "grad_norm": 12.704758644104004, "learning_rate": 4.536734342739161e-05, "loss": 1.0789, "step": 53850 }, { "epoch": 9.273916035788025, "grad_norm": 10.87044620513916, "learning_rate": 4.536304198210599e-05, "loss": 1.0757, "step": 53900 }, { "epoch": 9.282518926359257, "grad_norm": 8.700004577636719, "learning_rate": 4.535874053682037e-05, "loss": 1.0523, "step": 53950 }, { "epoch": 9.291121816930488, "grad_norm": 6.160933971405029, "learning_rate": 4.535443909153476e-05, "loss": 1.0477, "step": 54000 }, { "epoch": 9.29972470750172, "grad_norm": 12.407735824584961, "learning_rate": 4.535013764624914e-05, "loss": 1.0185, "step": 54050 }, { "epoch": 9.308327598072953, "grad_norm": 8.268204689025879, "learning_rate": 4.534583620096353e-05, "loss": 1.0102, "step": 54100 }, { "epoch": 9.316930488644184, "grad_norm": 9.77997875213623, "learning_rate": 4.534153475567791e-05, "loss": 1.1007, "step": 54150 }, { "epoch": 9.325533379215416, "grad_norm": 5.467930793762207, "learning_rate": 4.53372333103923e-05, "loss": 1.0431, "step": 54200 }, { "epoch": 9.334136269786649, "grad_norm": 9.13076400756836, "learning_rate": 4.533293186510668e-05, "loss": 1.0186, "step": 54250 }, { "epoch": 9.34273916035788, "grad_norm": 8.1834716796875, "learning_rate": 4.532863041982106e-05, "loss": 1.0519, "step": 54300 }, { "epoch": 9.351342050929112, "grad_norm": 11.334811210632324, "learning_rate": 4.532432897453545e-05, "loss": 1.0791, "step": 54350 }, { "epoch": 9.359944941500345, "grad_norm": 5.853205680847168, "learning_rate": 4.532002752924983e-05, "loss": 1.0592, "step": 54400 }, { "epoch": 9.368547832071576, "grad_norm": 8.090055465698242, "learning_rate": 4.5315726083964216e-05, "loss": 1.0785, "step": 54450 }, { "epoch": 9.377150722642808, "grad_norm": 9.159048080444336, "learning_rate": 4.53114246386786e-05, "loss": 1.0887, "step": 54500 }, { "epoch": 9.38575361321404, "grad_norm": 6.047050952911377, "learning_rate": 4.530712319339298e-05, "loss": 1.0578, "step": 54550 }, { "epoch": 9.394356503785271, "grad_norm": 6.050055027008057, "learning_rate": 4.5302821748107366e-05, "loss": 1.0327, "step": 54600 }, { "epoch": 9.402959394356504, "grad_norm": 9.312603950500488, "learning_rate": 4.529852030282175e-05, "loss": 1.0179, "step": 54650 }, { "epoch": 9.411562284927736, "grad_norm": 7.193454742431641, "learning_rate": 4.5294218857536135e-05, "loss": 1.0798, "step": 54700 }, { "epoch": 9.420165175498967, "grad_norm": 7.983763694763184, "learning_rate": 4.5289917412250523e-05, "loss": 1.0626, "step": 54750 }, { "epoch": 9.4287680660702, "grad_norm": 8.663176536560059, "learning_rate": 4.52856159669649e-05, "loss": 1.085, "step": 54800 }, { "epoch": 9.437370956641432, "grad_norm": 9.066131591796875, "learning_rate": 4.5281314521679286e-05, "loss": 1.052, "step": 54850 }, { "epoch": 9.445973847212663, "grad_norm": 7.8415608406066895, "learning_rate": 4.5277013076393674e-05, "loss": 1.0628, "step": 54900 }, { "epoch": 9.454576737783896, "grad_norm": 9.01369571685791, "learning_rate": 4.5272711631108055e-05, "loss": 1.0075, "step": 54950 }, { "epoch": 9.463179628355128, "grad_norm": 9.185172080993652, "learning_rate": 4.526841018582244e-05, "loss": 1.0682, "step": 55000 }, { "epoch": 9.471782518926359, "grad_norm": 7.79381799697876, "learning_rate": 4.526410874053682e-05, "loss": 1.1048, "step": 55050 }, { "epoch": 9.480385409497591, "grad_norm": 6.931903839111328, "learning_rate": 4.5259807295251205e-05, "loss": 1.0612, "step": 55100 }, { "epoch": 9.488988300068822, "grad_norm": 9.405593872070312, "learning_rate": 4.525550584996559e-05, "loss": 1.0284, "step": 55150 }, { "epoch": 9.497591190640055, "grad_norm": 11.144434928894043, "learning_rate": 4.5251204404679974e-05, "loss": 0.9918, "step": 55200 }, { "epoch": 9.506194081211287, "grad_norm": 8.819583892822266, "learning_rate": 4.524690295939436e-05, "loss": 1.0598, "step": 55250 }, { "epoch": 9.51479697178252, "grad_norm": 12.985445976257324, "learning_rate": 4.524260151410874e-05, "loss": 1.0054, "step": 55300 }, { "epoch": 9.52339986235375, "grad_norm": 9.291825294494629, "learning_rate": 4.5238300068823125e-05, "loss": 1.0166, "step": 55350 }, { "epoch": 9.532002752924983, "grad_norm": 10.232939720153809, "learning_rate": 4.523399862353751e-05, "loss": 1.066, "step": 55400 }, { "epoch": 9.540605643496214, "grad_norm": 8.146832466125488, "learning_rate": 4.5229697178251894e-05, "loss": 1.0422, "step": 55450 }, { "epoch": 9.549208534067446, "grad_norm": 10.787943840026855, "learning_rate": 4.522539573296628e-05, "loss": 1.0566, "step": 55500 }, { "epoch": 9.557811424638679, "grad_norm": 6.148542404174805, "learning_rate": 4.522109428768066e-05, "loss": 1.0251, "step": 55550 }, { "epoch": 9.56641431520991, "grad_norm": 7.921225547790527, "learning_rate": 4.5216792842395044e-05, "loss": 1.0142, "step": 55600 }, { "epoch": 9.575017205781142, "grad_norm": 10.527291297912598, "learning_rate": 4.521249139710943e-05, "loss": 1.0438, "step": 55650 }, { "epoch": 9.583620096352375, "grad_norm": 10.801708221435547, "learning_rate": 4.520818995182381e-05, "loss": 1.0382, "step": 55700 }, { "epoch": 9.592222986923606, "grad_norm": 8.003885269165039, "learning_rate": 4.52038885065382e-05, "loss": 1.0889, "step": 55750 }, { "epoch": 9.600825877494838, "grad_norm": 9.928646087646484, "learning_rate": 4.519958706125258e-05, "loss": 1.0486, "step": 55800 }, { "epoch": 9.60942876806607, "grad_norm": 12.343487739562988, "learning_rate": 4.5195285615966963e-05, "loss": 1.068, "step": 55850 }, { "epoch": 9.618031658637301, "grad_norm": 7.312779426574707, "learning_rate": 4.519098417068135e-05, "loss": 1.0808, "step": 55900 }, { "epoch": 9.626634549208534, "grad_norm": 7.080031871795654, "learning_rate": 4.518668272539573e-05, "loss": 1.0706, "step": 55950 }, { "epoch": 9.635237439779766, "grad_norm": 9.524930953979492, "learning_rate": 4.518238128011012e-05, "loss": 1.041, "step": 56000 }, { "epoch": 9.643840330350997, "grad_norm": 9.400650978088379, "learning_rate": 4.51780798348245e-05, "loss": 1.1053, "step": 56050 }, { "epoch": 9.65244322092223, "grad_norm": 10.766495704650879, "learning_rate": 4.517377838953889e-05, "loss": 1.0171, "step": 56100 }, { "epoch": 9.661046111493462, "grad_norm": 8.815221786499023, "learning_rate": 4.516947694425327e-05, "loss": 1.0872, "step": 56150 }, { "epoch": 9.669649002064693, "grad_norm": 12.727176666259766, "learning_rate": 4.516517549896765e-05, "loss": 1.0765, "step": 56200 }, { "epoch": 9.678251892635926, "grad_norm": 8.453421592712402, "learning_rate": 4.516087405368204e-05, "loss": 1.0412, "step": 56250 }, { "epoch": 9.686854783207158, "grad_norm": 9.545226097106934, "learning_rate": 4.515657260839643e-05, "loss": 1.0258, "step": 56300 }, { "epoch": 9.695457673778389, "grad_norm": 9.387917518615723, "learning_rate": 4.515227116311081e-05, "loss": 1.0618, "step": 56350 }, { "epoch": 9.704060564349621, "grad_norm": 7.760498046875, "learning_rate": 4.514796971782519e-05, "loss": 1.0257, "step": 56400 }, { "epoch": 9.712663454920854, "grad_norm": 6.892874717712402, "learning_rate": 4.514366827253957e-05, "loss": 1.0537, "step": 56450 }, { "epoch": 9.721266345492085, "grad_norm": 9.4236421585083, "learning_rate": 4.513936682725396e-05, "loss": 1.0677, "step": 56500 }, { "epoch": 9.729869236063317, "grad_norm": 8.729511260986328, "learning_rate": 4.513506538196835e-05, "loss": 1.0275, "step": 56550 }, { "epoch": 9.73847212663455, "grad_norm": 10.910252571105957, "learning_rate": 4.513076393668273e-05, "loss": 1.0536, "step": 56600 }, { "epoch": 9.74707501720578, "grad_norm": 8.411504745483398, "learning_rate": 4.512646249139711e-05, "loss": 1.078, "step": 56650 }, { "epoch": 9.755677907777013, "grad_norm": 10.043415069580078, "learning_rate": 4.512216104611149e-05, "loss": 1.1077, "step": 56700 }, { "epoch": 9.764280798348246, "grad_norm": 10.43034839630127, "learning_rate": 4.511785960082588e-05, "loss": 1.0338, "step": 56750 }, { "epoch": 9.772883688919476, "grad_norm": 9.806894302368164, "learning_rate": 4.511355815554027e-05, "loss": 1.0819, "step": 56800 }, { "epoch": 9.781486579490709, "grad_norm": 7.622389316558838, "learning_rate": 4.510925671025465e-05, "loss": 1.0398, "step": 56850 }, { "epoch": 9.790089470061941, "grad_norm": 7.930126190185547, "learning_rate": 4.5104955264969036e-05, "loss": 1.0781, "step": 56900 }, { "epoch": 9.798692360633172, "grad_norm": 9.715886116027832, "learning_rate": 4.510065381968342e-05, "loss": 1.0041, "step": 56950 }, { "epoch": 9.807295251204405, "grad_norm": 7.75253963470459, "learning_rate": 4.50963523743978e-05, "loss": 1.0036, "step": 57000 }, { "epoch": 9.815898141775637, "grad_norm": 7.226025104522705, "learning_rate": 4.5092050929112186e-05, "loss": 1.0417, "step": 57050 }, { "epoch": 9.824501032346868, "grad_norm": 9.54867172241211, "learning_rate": 4.508774948382657e-05, "loss": 1.0029, "step": 57100 }, { "epoch": 9.8331039229181, "grad_norm": 7.994626045227051, "learning_rate": 4.5083448038540955e-05, "loss": 1.021, "step": 57150 }, { "epoch": 9.841706813489333, "grad_norm": 9.072806358337402, "learning_rate": 4.5079146593255336e-05, "loss": 1.012, "step": 57200 }, { "epoch": 9.850309704060564, "grad_norm": 9.010644912719727, "learning_rate": 4.507484514796972e-05, "loss": 1.0516, "step": 57250 }, { "epoch": 9.858912594631796, "grad_norm": 7.591326713562012, "learning_rate": 4.5070543702684106e-05, "loss": 1.0656, "step": 57300 }, { "epoch": 9.867515485203029, "grad_norm": 8.449004173278809, "learning_rate": 4.506624225739849e-05, "loss": 1.0821, "step": 57350 }, { "epoch": 9.87611837577426, "grad_norm": 7.878366947174072, "learning_rate": 4.5061940812112875e-05, "loss": 1.089, "step": 57400 }, { "epoch": 9.884721266345492, "grad_norm": 12.887643814086914, "learning_rate": 4.5057639366827256e-05, "loss": 1.0806, "step": 57450 }, { "epoch": 9.893324156916725, "grad_norm": 9.142376899719238, "learning_rate": 4.505333792154164e-05, "loss": 1.0174, "step": 57500 }, { "epoch": 9.901927047487955, "grad_norm": 12.486419677734375, "learning_rate": 4.5049036476256025e-05, "loss": 1.0842, "step": 57550 }, { "epoch": 9.910529938059188, "grad_norm": 8.338517189025879, "learning_rate": 4.5044735030970406e-05, "loss": 1.0589, "step": 57600 }, { "epoch": 9.91913282863042, "grad_norm": 6.4759416580200195, "learning_rate": 4.5040433585684794e-05, "loss": 1.0499, "step": 57650 }, { "epoch": 9.927735719201651, "grad_norm": 7.6119914054870605, "learning_rate": 4.503613214039918e-05, "loss": 1.0616, "step": 57700 }, { "epoch": 9.936338609772884, "grad_norm": 10.068697929382324, "learning_rate": 4.5031830695113556e-05, "loss": 1.0109, "step": 57750 }, { "epoch": 9.944941500344116, "grad_norm": 8.650705337524414, "learning_rate": 4.5027529249827944e-05, "loss": 1.0333, "step": 57800 }, { "epoch": 9.953544390915347, "grad_norm": 14.173545837402344, "learning_rate": 4.5023227804542326e-05, "loss": 1.0868, "step": 57850 }, { "epoch": 9.96214728148658, "grad_norm": 5.54899263381958, "learning_rate": 4.5018926359256713e-05, "loss": 1.0673, "step": 57900 }, { "epoch": 9.970750172057812, "grad_norm": 8.386231422424316, "learning_rate": 4.50146249139711e-05, "loss": 1.0045, "step": 57950 }, { "epoch": 9.979353062629043, "grad_norm": 8.25656795501709, "learning_rate": 4.5010323468685476e-05, "loss": 1.0794, "step": 58000 }, { "epoch": 9.987955953200276, "grad_norm": 12.856457710266113, "learning_rate": 4.5006022023399864e-05, "loss": 1.0414, "step": 58050 }, { "epoch": 9.996558843771508, "grad_norm": 7.505799770355225, "learning_rate": 4.5001720578114245e-05, "loss": 0.9988, "step": 58100 }, { "epoch": 10.0, "eval_accuracy": 0.5622096024780588, "eval_f1": 0.5670412657096382, "eval_f1_DuraRiadoRio_16x16": 0.5286368048814312, "eval_f1_Mole_16x16": 0.5308282614448266, "eval_f1_Quebrado_16x16": 0.7202380952380952, "eval_f1_RiadoRio_16x16": 0.4176470588235294, "eval_f1_RioFechado_16x16": 0.637856108160309, "eval_loss": 1.2635369300842285, "eval_precision": 0.626876751114677, "eval_precision_DuraRiadoRio_16x16": 0.38842469940900753, "eval_precision_Mole_16x16": 0.68, "eval_precision_Quebrado_16x16": 0.8402777777777778, "eval_precision_RiadoRio_16x16": 0.5169902912621359, "eval_precision_RioFechado_16x16": 0.7086909871244635, "eval_recall": 0.5646037461364488, "eval_recall_DuraRiadoRio_16x16": 0.8272569444444444, "eval_recall_Mole_16x16": 0.4353298611111111, "eval_recall_Quebrado_16x16": 0.6302083333333334, "eval_recall_RiadoRio_16x16": 0.3503289473684211, "eval_recall_RioFechado_16x16": 0.5798946444249341, "eval_runtime": 55.3029, "eval_samples_per_second": 210.152, "eval_steps_per_second": 13.146, "step": 58120 }, { "epoch": 10.005161734342739, "grad_norm": 5.172220706939697, "learning_rate": 4.499741913282863e-05, "loss": 1.013, "step": 58150 }, { "epoch": 10.013764624913971, "grad_norm": 9.825302124023438, "learning_rate": 4.499311768754302e-05, "loss": 1.0025, "step": 58200 }, { "epoch": 10.022367515485204, "grad_norm": 10.08629322052002, "learning_rate": 4.49888162422574e-05, "loss": 1.0373, "step": 58250 }, { "epoch": 10.030970406056435, "grad_norm": 6.852029800415039, "learning_rate": 4.498451479697178e-05, "loss": 1.034, "step": 58300 }, { "epoch": 10.039573296627667, "grad_norm": 10.965917587280273, "learning_rate": 4.498021335168617e-05, "loss": 1.0018, "step": 58350 }, { "epoch": 10.0481761871989, "grad_norm": 7.480066299438477, "learning_rate": 4.497591190640055e-05, "loss": 1.0507, "step": 58400 }, { "epoch": 10.05677907777013, "grad_norm": 9.52606201171875, "learning_rate": 4.497161046111494e-05, "loss": 1.0125, "step": 58450 }, { "epoch": 10.065381968341363, "grad_norm": 15.931929588317871, "learning_rate": 4.496730901582932e-05, "loss": 1.028, "step": 58500 }, { "epoch": 10.073984858912594, "grad_norm": 9.86666202545166, "learning_rate": 4.49630075705437e-05, "loss": 1.0501, "step": 58550 }, { "epoch": 10.082587749483826, "grad_norm": 9.124605178833008, "learning_rate": 4.495870612525809e-05, "loss": 1.067, "step": 58600 }, { "epoch": 10.091190640055059, "grad_norm": 7.959956169128418, "learning_rate": 4.495440467997247e-05, "loss": 1.0117, "step": 58650 }, { "epoch": 10.09979353062629, "grad_norm": 7.6685309410095215, "learning_rate": 4.495010323468686e-05, "loss": 1.0797, "step": 58700 }, { "epoch": 10.108396421197522, "grad_norm": 10.487074851989746, "learning_rate": 4.494580178940124e-05, "loss": 1.07, "step": 58750 }, { "epoch": 10.116999311768755, "grad_norm": 5.104241371154785, "learning_rate": 4.494150034411562e-05, "loss": 0.973, "step": 58800 }, { "epoch": 10.125602202339985, "grad_norm": 6.80340576171875, "learning_rate": 4.493719889883001e-05, "loss": 1.0372, "step": 58850 }, { "epoch": 10.134205092911218, "grad_norm": 10.320247650146484, "learning_rate": 4.493289745354439e-05, "loss": 1.0259, "step": 58900 }, { "epoch": 10.14280798348245, "grad_norm": 7.9220757484436035, "learning_rate": 4.492859600825878e-05, "loss": 1.0123, "step": 58950 }, { "epoch": 10.151410874053681, "grad_norm": 7.803106784820557, "learning_rate": 4.492429456297316e-05, "loss": 1.0417, "step": 59000 }, { "epoch": 10.160013764624914, "grad_norm": 9.78612995147705, "learning_rate": 4.491999311768755e-05, "loss": 0.9997, "step": 59050 }, { "epoch": 10.168616655196146, "grad_norm": 7.560581684112549, "learning_rate": 4.491569167240193e-05, "loss": 1.0306, "step": 59100 }, { "epoch": 10.177219545767377, "grad_norm": 9.90194034576416, "learning_rate": 4.491139022711631e-05, "loss": 1.058, "step": 59150 }, { "epoch": 10.18582243633861, "grad_norm": 9.725329399108887, "learning_rate": 4.49070887818307e-05, "loss": 0.9856, "step": 59200 }, { "epoch": 10.194425326909842, "grad_norm": 9.212682723999023, "learning_rate": 4.490278733654508e-05, "loss": 1.0597, "step": 59250 }, { "epoch": 10.203028217481073, "grad_norm": 7.454093933105469, "learning_rate": 4.489848589125947e-05, "loss": 1.0302, "step": 59300 }, { "epoch": 10.211631108052305, "grad_norm": 6.9897074699401855, "learning_rate": 4.489418444597385e-05, "loss": 1.0619, "step": 59350 }, { "epoch": 10.220233998623538, "grad_norm": 11.047837257385254, "learning_rate": 4.488988300068823e-05, "loss": 1.0343, "step": 59400 }, { "epoch": 10.228836889194769, "grad_norm": 10.540821075439453, "learning_rate": 4.488558155540262e-05, "loss": 1.0177, "step": 59450 }, { "epoch": 10.237439779766001, "grad_norm": 8.58780574798584, "learning_rate": 4.4881280110117e-05, "loss": 1.0635, "step": 59500 }, { "epoch": 10.246042670337234, "grad_norm": 10.894421577453613, "learning_rate": 4.487697866483139e-05, "loss": 1.012, "step": 59550 }, { "epoch": 10.254645560908465, "grad_norm": 9.924884796142578, "learning_rate": 4.487267721954577e-05, "loss": 1.0826, "step": 59600 }, { "epoch": 10.263248451479697, "grad_norm": 9.886469841003418, "learning_rate": 4.486837577426015e-05, "loss": 1.0724, "step": 59650 }, { "epoch": 10.27185134205093, "grad_norm": 9.742155075073242, "learning_rate": 4.486407432897454e-05, "loss": 1.0256, "step": 59700 }, { "epoch": 10.28045423262216, "grad_norm": 15.690170288085938, "learning_rate": 4.4859772883688925e-05, "loss": 1.0268, "step": 59750 }, { "epoch": 10.289057123193393, "grad_norm": 8.029948234558105, "learning_rate": 4.4855471438403306e-05, "loss": 1.0642, "step": 59800 }, { "epoch": 10.297660013764625, "grad_norm": 10.347542762756348, "learning_rate": 4.4851169993117694e-05, "loss": 1.0363, "step": 59850 }, { "epoch": 10.306262904335856, "grad_norm": 6.760985851287842, "learning_rate": 4.484686854783207e-05, "loss": 1.0551, "step": 59900 }, { "epoch": 10.314865794907089, "grad_norm": 12.17372989654541, "learning_rate": 4.484256710254646e-05, "loss": 1.0649, "step": 59950 }, { "epoch": 10.323468685478321, "grad_norm": 10.952040672302246, "learning_rate": 4.4838265657260845e-05, "loss": 1.0579, "step": 60000 }, { "epoch": 10.332071576049552, "grad_norm": 5.882075309753418, "learning_rate": 4.4833964211975226e-05, "loss": 1.0538, "step": 60050 }, { "epoch": 10.340674466620785, "grad_norm": 11.87692928314209, "learning_rate": 4.4829662766689614e-05, "loss": 1.0109, "step": 60100 }, { "epoch": 10.349277357192017, "grad_norm": 6.8704118728637695, "learning_rate": 4.482536132140399e-05, "loss": 1.0396, "step": 60150 }, { "epoch": 10.357880247763248, "grad_norm": 6.353635787963867, "learning_rate": 4.4821059876118376e-05, "loss": 1.0365, "step": 60200 }, { "epoch": 10.36648313833448, "grad_norm": 11.739554405212402, "learning_rate": 4.4816758430832764e-05, "loss": 1.0516, "step": 60250 }, { "epoch": 10.375086028905713, "grad_norm": 8.735274314880371, "learning_rate": 4.4812456985547145e-05, "loss": 1.0151, "step": 60300 }, { "epoch": 10.383688919476944, "grad_norm": 8.460657119750977, "learning_rate": 4.480815554026153e-05, "loss": 1.0525, "step": 60350 }, { "epoch": 10.392291810048176, "grad_norm": 9.63074779510498, "learning_rate": 4.4803854094975914e-05, "loss": 1.0227, "step": 60400 }, { "epoch": 10.400894700619409, "grad_norm": 9.768966674804688, "learning_rate": 4.4799552649690296e-05, "loss": 1.0728, "step": 60450 }, { "epoch": 10.40949759119064, "grad_norm": 7.409395217895508, "learning_rate": 4.4795251204404684e-05, "loss": 1.0509, "step": 60500 }, { "epoch": 10.418100481761872, "grad_norm": 10.930662155151367, "learning_rate": 4.4790949759119065e-05, "loss": 1.043, "step": 60550 }, { "epoch": 10.426703372333105, "grad_norm": 10.579763412475586, "learning_rate": 4.478664831383345e-05, "loss": 0.9801, "step": 60600 }, { "epoch": 10.435306262904335, "grad_norm": 10.48686695098877, "learning_rate": 4.4782346868547834e-05, "loss": 1.0748, "step": 60650 }, { "epoch": 10.443909153475568, "grad_norm": 6.76818323135376, "learning_rate": 4.4778045423262215e-05, "loss": 1.0333, "step": 60700 }, { "epoch": 10.4525120440468, "grad_norm": 8.421025276184082, "learning_rate": 4.47737439779766e-05, "loss": 0.9719, "step": 60750 }, { "epoch": 10.461114934618031, "grad_norm": 10.924844741821289, "learning_rate": 4.4769442532690984e-05, "loss": 1.0808, "step": 60800 }, { "epoch": 10.469717825189264, "grad_norm": 10.769779205322266, "learning_rate": 4.476514108740537e-05, "loss": 1.0548, "step": 60850 }, { "epoch": 10.478320715760496, "grad_norm": 9.575783729553223, "learning_rate": 4.476083964211975e-05, "loss": 1.0318, "step": 60900 }, { "epoch": 10.486923606331727, "grad_norm": 8.061126708984375, "learning_rate": 4.4756538196834134e-05, "loss": 1.0527, "step": 60950 }, { "epoch": 10.49552649690296, "grad_norm": 10.77354907989502, "learning_rate": 4.475223675154852e-05, "loss": 1.0398, "step": 61000 }, { "epoch": 10.504129387474192, "grad_norm": 8.128907203674316, "learning_rate": 4.4747935306262904e-05, "loss": 1.0536, "step": 61050 }, { "epoch": 10.512732278045423, "grad_norm": 15.628767013549805, "learning_rate": 4.474363386097729e-05, "loss": 1.0311, "step": 61100 }, { "epoch": 10.521335168616655, "grad_norm": 8.271849632263184, "learning_rate": 4.473933241569167e-05, "loss": 1.0292, "step": 61150 }, { "epoch": 10.529938059187888, "grad_norm": 11.656572341918945, "learning_rate": 4.473503097040606e-05, "loss": 1.0253, "step": 61200 }, { "epoch": 10.538540949759119, "grad_norm": 9.223893165588379, "learning_rate": 4.473072952512044e-05, "loss": 1.0504, "step": 61250 }, { "epoch": 10.547143840330351, "grad_norm": 11.067399978637695, "learning_rate": 4.472642807983482e-05, "loss": 1.065, "step": 61300 }, { "epoch": 10.555746730901582, "grad_norm": 14.842527389526367, "learning_rate": 4.472212663454921e-05, "loss": 1.0523, "step": 61350 }, { "epoch": 10.564349621472815, "grad_norm": 9.253747940063477, "learning_rate": 4.47178251892636e-05, "loss": 1.0043, "step": 61400 }, { "epoch": 10.572952512044047, "grad_norm": 8.215195655822754, "learning_rate": 4.471352374397798e-05, "loss": 1.0712, "step": 61450 }, { "epoch": 10.58155540261528, "grad_norm": 10.332090377807617, "learning_rate": 4.470922229869236e-05, "loss": 1.0029, "step": 61500 }, { "epoch": 10.59015829318651, "grad_norm": 12.06308650970459, "learning_rate": 4.470492085340674e-05, "loss": 1.0375, "step": 61550 }, { "epoch": 10.598761183757743, "grad_norm": 8.694219589233398, "learning_rate": 4.470061940812113e-05, "loss": 1.0617, "step": 61600 }, { "epoch": 10.607364074328974, "grad_norm": 9.651632308959961, "learning_rate": 4.469631796283552e-05, "loss": 1.0119, "step": 61650 }, { "epoch": 10.615966964900206, "grad_norm": 11.267427444458008, "learning_rate": 4.46920165175499e-05, "loss": 0.9905, "step": 61700 }, { "epoch": 10.624569855471439, "grad_norm": 10.208940505981445, "learning_rate": 4.468771507226429e-05, "loss": 1.0628, "step": 61750 }, { "epoch": 10.633172746042671, "grad_norm": 8.76193904876709, "learning_rate": 4.468341362697867e-05, "loss": 0.9946, "step": 61800 }, { "epoch": 10.641775636613902, "grad_norm": 12.242188453674316, "learning_rate": 4.467911218169305e-05, "loss": 1.0302, "step": 61850 }, { "epoch": 10.650378527185135, "grad_norm": 12.077661514282227, "learning_rate": 4.467481073640744e-05, "loss": 1.0275, "step": 61900 }, { "epoch": 10.658981417756365, "grad_norm": 7.635251522064209, "learning_rate": 4.467050929112182e-05, "loss": 1.0068, "step": 61950 }, { "epoch": 10.667584308327598, "grad_norm": 9.506245613098145, "learning_rate": 4.466620784583621e-05, "loss": 1.0674, "step": 62000 }, { "epoch": 10.67618719889883, "grad_norm": 8.839384078979492, "learning_rate": 4.466190640055059e-05, "loss": 1.0509, "step": 62050 }, { "epoch": 10.684790089470061, "grad_norm": 8.905369758605957, "learning_rate": 4.465760495526497e-05, "loss": 1.0877, "step": 62100 }, { "epoch": 10.693392980041294, "grad_norm": 9.683390617370605, "learning_rate": 4.465330350997936e-05, "loss": 1.0533, "step": 62150 }, { "epoch": 10.701995870612526, "grad_norm": 8.7412109375, "learning_rate": 4.464900206469374e-05, "loss": 1.0386, "step": 62200 }, { "epoch": 10.710598761183757, "grad_norm": 7.5940351486206055, "learning_rate": 4.4644700619408126e-05, "loss": 1.0205, "step": 62250 }, { "epoch": 10.71920165175499, "grad_norm": 8.854721069335938, "learning_rate": 4.464039917412251e-05, "loss": 1.043, "step": 62300 }, { "epoch": 10.727804542326222, "grad_norm": 7.120172500610352, "learning_rate": 4.463609772883689e-05, "loss": 0.9956, "step": 62350 }, { "epoch": 10.736407432897453, "grad_norm": 11.816934585571289, "learning_rate": 4.4631796283551276e-05, "loss": 1.033, "step": 62400 }, { "epoch": 10.745010323468685, "grad_norm": 9.770413398742676, "learning_rate": 4.462749483826566e-05, "loss": 1.008, "step": 62450 }, { "epoch": 10.753613214039918, "grad_norm": 9.53784465789795, "learning_rate": 4.4623193392980046e-05, "loss": 1.0491, "step": 62500 }, { "epoch": 10.762216104611149, "grad_norm": 9.127557754516602, "learning_rate": 4.461889194769443e-05, "loss": 1.035, "step": 62550 }, { "epoch": 10.770818995182381, "grad_norm": 12.451385498046875, "learning_rate": 4.461459050240881e-05, "loss": 1.0618, "step": 62600 }, { "epoch": 10.779421885753614, "grad_norm": 8.760132789611816, "learning_rate": 4.4610289057123196e-05, "loss": 1.0261, "step": 62650 }, { "epoch": 10.788024776324844, "grad_norm": 10.269099235534668, "learning_rate": 4.460598761183758e-05, "loss": 1.0101, "step": 62700 }, { "epoch": 10.796627666896077, "grad_norm": 9.212080001831055, "learning_rate": 4.4601686166551965e-05, "loss": 1.0909, "step": 62750 }, { "epoch": 10.80523055746731, "grad_norm": 7.590916156768799, "learning_rate": 4.459738472126635e-05, "loss": 1.1295, "step": 62800 }, { "epoch": 10.81383344803854, "grad_norm": 8.31568431854248, "learning_rate": 4.459308327598073e-05, "loss": 1.0432, "step": 62850 }, { "epoch": 10.822436338609773, "grad_norm": 7.781918048858643, "learning_rate": 4.4588781830695115e-05, "loss": 1.0306, "step": 62900 }, { "epoch": 10.831039229181005, "grad_norm": 6.58039665222168, "learning_rate": 4.4584480385409496e-05, "loss": 1.0557, "step": 62950 }, { "epoch": 10.839642119752236, "grad_norm": 7.381245136260986, "learning_rate": 4.4580178940123884e-05, "loss": 1.0063, "step": 63000 }, { "epoch": 10.848245010323469, "grad_norm": 11.122227668762207, "learning_rate": 4.457587749483827e-05, "loss": 0.9978, "step": 63050 }, { "epoch": 10.856847900894701, "grad_norm": 10.064472198486328, "learning_rate": 4.457157604955265e-05, "loss": 1.001, "step": 63100 }, { "epoch": 10.865450791465932, "grad_norm": 9.076276779174805, "learning_rate": 4.4567274604267035e-05, "loss": 1.0279, "step": 63150 }, { "epoch": 10.874053682037164, "grad_norm": 10.235466957092285, "learning_rate": 4.4562973158981416e-05, "loss": 1.0602, "step": 63200 }, { "epoch": 10.882656572608397, "grad_norm": 13.201318740844727, "learning_rate": 4.4558671713695804e-05, "loss": 1.0324, "step": 63250 }, { "epoch": 10.891259463179628, "grad_norm": 9.369601249694824, "learning_rate": 4.455437026841019e-05, "loss": 1.1001, "step": 63300 }, { "epoch": 10.89986235375086, "grad_norm": 6.800917148590088, "learning_rate": 4.455006882312457e-05, "loss": 1.0621, "step": 63350 }, { "epoch": 10.908465244322093, "grad_norm": 8.106900215148926, "learning_rate": 4.4545767377838954e-05, "loss": 1.0536, "step": 63400 }, { "epoch": 10.917068134893324, "grad_norm": 13.18602466583252, "learning_rate": 4.454146593255334e-05, "loss": 1.0517, "step": 63450 }, { "epoch": 10.925671025464556, "grad_norm": 9.864371299743652, "learning_rate": 4.453716448726772e-05, "loss": 1.0267, "step": 63500 }, { "epoch": 10.934273916035789, "grad_norm": 7.603129863739014, "learning_rate": 4.453286304198211e-05, "loss": 0.9899, "step": 63550 }, { "epoch": 10.94287680660702, "grad_norm": 7.988731861114502, "learning_rate": 4.452856159669649e-05, "loss": 1.0107, "step": 63600 }, { "epoch": 10.951479697178252, "grad_norm": 7.888236045837402, "learning_rate": 4.4524260151410874e-05, "loss": 1.0367, "step": 63650 }, { "epoch": 10.960082587749485, "grad_norm": 12.226395606994629, "learning_rate": 4.451995870612526e-05, "loss": 1.0133, "step": 63700 }, { "epoch": 10.968685478320715, "grad_norm": 8.571328163146973, "learning_rate": 4.451565726083964e-05, "loss": 1.0738, "step": 63750 }, { "epoch": 10.977288368891948, "grad_norm": 8.605777740478516, "learning_rate": 4.451135581555403e-05, "loss": 1.0427, "step": 63800 }, { "epoch": 10.98589125946318, "grad_norm": 10.815055847167969, "learning_rate": 4.450705437026841e-05, "loss": 1.0558, "step": 63850 }, { "epoch": 10.994494150034411, "grad_norm": 8.889481544494629, "learning_rate": 4.45027529249828e-05, "loss": 1.0754, "step": 63900 }, { "epoch": 11.0, "eval_accuracy": 0.5401824126656342, "eval_f1": 0.5280369986428141, "eval_f1_DuraRiadoRio_16x16": 0.4718034617532105, "eval_f1_Mole_16x16": 0.593859649122807, "eval_f1_Quebrado_16x16": 0.6236398867193322, "eval_f1_RiadoRio_16x16": 0.43830654290242244, "eval_f1_RioFechado_16x16": 0.5125754527162978, "eval_loss": 1.2551488876342773, "eval_precision": 0.5648113498328285, "eval_precision_DuraRiadoRio_16x16": 0.661189358372457, "eval_precision_Mole_16x16": 0.600177304964539, "eval_precision_Quebrado_16x16": 0.47491486946651534, "eval_precision_RiadoRio_16x16": 0.48765743073047857, "eval_precision_RioFechado_16x16": 0.6001177856301532, "eval_recall": 0.541552344540199, "eval_recall_DuraRiadoRio_16x16": 0.3667534722222222, "eval_recall_Mole_16x16": 0.5876736111111112, "eval_recall_Quebrado_16x16": 0.9079861111111112, "eval_recall_RiadoRio_16x16": 0.3980263157894737, "eval_recall_RioFechado_16x16": 0.4473222124670764, "eval_runtime": 45.7056, "eval_samples_per_second": 254.279, "eval_steps_per_second": 15.906, "step": 63932 }, { "epoch": 11.003097040605644, "grad_norm": 12.081048965454102, "learning_rate": 4.449845147969718e-05, "loss": 1.0129, "step": 63950 }, { "epoch": 11.011699931176876, "grad_norm": 8.237885475158691, "learning_rate": 4.449415003441156e-05, "loss": 1.0045, "step": 64000 }, { "epoch": 11.020302821748107, "grad_norm": 12.987425804138184, "learning_rate": 4.448984858912595e-05, "loss": 0.9589, "step": 64050 }, { "epoch": 11.02890571231934, "grad_norm": 16.44765853881836, "learning_rate": 4.448554714384033e-05, "loss": 1.0613, "step": 64100 }, { "epoch": 11.037508602890572, "grad_norm": 7.872641563415527, "learning_rate": 4.448124569855472e-05, "loss": 1.0209, "step": 64150 }, { "epoch": 11.046111493461803, "grad_norm": 8.331313133239746, "learning_rate": 4.44769442532691e-05, "loss": 1.0344, "step": 64200 }, { "epoch": 11.054714384033035, "grad_norm": 7.680593490600586, "learning_rate": 4.447264280798348e-05, "loss": 0.998, "step": 64250 }, { "epoch": 11.063317274604268, "grad_norm": 11.445902824401855, "learning_rate": 4.446834136269787e-05, "loss": 1.0675, "step": 64300 }, { "epoch": 11.071920165175499, "grad_norm": 8.51711368560791, "learning_rate": 4.446403991741225e-05, "loss": 1.0562, "step": 64350 }, { "epoch": 11.080523055746731, "grad_norm": 10.319658279418945, "learning_rate": 4.445973847212664e-05, "loss": 1.0498, "step": 64400 }, { "epoch": 11.089125946317964, "grad_norm": 10.604608535766602, "learning_rate": 4.445543702684102e-05, "loss": 1.0276, "step": 64450 }, { "epoch": 11.097728836889194, "grad_norm": 7.123222351074219, "learning_rate": 4.44511355815554e-05, "loss": 1.057, "step": 64500 }, { "epoch": 11.106331727460427, "grad_norm": 9.22993278503418, "learning_rate": 4.444683413626979e-05, "loss": 0.959, "step": 64550 }, { "epoch": 11.11493461803166, "grad_norm": 7.348292350769043, "learning_rate": 4.444253269098417e-05, "loss": 1.0509, "step": 64600 }, { "epoch": 11.12353750860289, "grad_norm": 11.415627479553223, "learning_rate": 4.443823124569856e-05, "loss": 1.036, "step": 64650 }, { "epoch": 11.132140399174123, "grad_norm": 10.537965774536133, "learning_rate": 4.4433929800412946e-05, "loss": 0.9985, "step": 64700 }, { "epoch": 11.140743289745355, "grad_norm": 9.393211364746094, "learning_rate": 4.442962835512732e-05, "loss": 1.0174, "step": 64750 }, { "epoch": 11.149346180316586, "grad_norm": 14.761847496032715, "learning_rate": 4.442532690984171e-05, "loss": 1.0308, "step": 64800 }, { "epoch": 11.157949070887819, "grad_norm": 9.778112411499023, "learning_rate": 4.4421025464556096e-05, "loss": 1.0309, "step": 64850 }, { "epoch": 11.16655196145905, "grad_norm": 13.531583786010742, "learning_rate": 4.441672401927048e-05, "loss": 1.037, "step": 64900 }, { "epoch": 11.175154852030282, "grad_norm": 10.525081634521484, "learning_rate": 4.4412422573984865e-05, "loss": 1.038, "step": 64950 }, { "epoch": 11.183757742601514, "grad_norm": 7.4119343757629395, "learning_rate": 4.440812112869924e-05, "loss": 1.034, "step": 65000 }, { "epoch": 11.192360633172745, "grad_norm": 9.931920051574707, "learning_rate": 4.440381968341363e-05, "loss": 1.0458, "step": 65050 }, { "epoch": 11.200963523743978, "grad_norm": 11.944510459899902, "learning_rate": 4.4399518238128016e-05, "loss": 1.04, "step": 65100 }, { "epoch": 11.20956641431521, "grad_norm": 9.23837661743164, "learning_rate": 4.43952167928424e-05, "loss": 0.9952, "step": 65150 }, { "epoch": 11.218169304886441, "grad_norm": 8.572166442871094, "learning_rate": 4.4390915347556785e-05, "loss": 1.0136, "step": 65200 }, { "epoch": 11.226772195457674, "grad_norm": 13.648146629333496, "learning_rate": 4.4386613902271166e-05, "loss": 1.0461, "step": 65250 }, { "epoch": 11.235375086028906, "grad_norm": 7.7325119972229, "learning_rate": 4.438231245698555e-05, "loss": 1.0085, "step": 65300 }, { "epoch": 11.243977976600137, "grad_norm": 9.810218811035156, "learning_rate": 4.4378011011699935e-05, "loss": 1.0066, "step": 65350 }, { "epoch": 11.25258086717137, "grad_norm": 6.82174015045166, "learning_rate": 4.4373709566414316e-05, "loss": 0.9925, "step": 65400 }, { "epoch": 11.261183757742602, "grad_norm": 7.154357433319092, "learning_rate": 4.4369408121128704e-05, "loss": 1.0445, "step": 65450 }, { "epoch": 11.269786648313833, "grad_norm": 9.543225288391113, "learning_rate": 4.4365106675843085e-05, "loss": 0.9664, "step": 65500 }, { "epoch": 11.278389538885065, "grad_norm": 8.299501419067383, "learning_rate": 4.4360805230557467e-05, "loss": 1.011, "step": 65550 }, { "epoch": 11.286992429456298, "grad_norm": 8.865137100219727, "learning_rate": 4.4356503785271854e-05, "loss": 1.0134, "step": 65600 }, { "epoch": 11.295595320027529, "grad_norm": 5.773687362670898, "learning_rate": 4.4352202339986236e-05, "loss": 1.0174, "step": 65650 }, { "epoch": 11.304198210598761, "grad_norm": 7.456663131713867, "learning_rate": 4.4347900894700624e-05, "loss": 1.0029, "step": 65700 }, { "epoch": 11.312801101169994, "grad_norm": 13.345144271850586, "learning_rate": 4.4343599449415005e-05, "loss": 0.9657, "step": 65750 }, { "epoch": 11.321403991741224, "grad_norm": 8.289722442626953, "learning_rate": 4.4339298004129386e-05, "loss": 1.0625, "step": 65800 }, { "epoch": 11.330006882312457, "grad_norm": 9.908332824707031, "learning_rate": 4.4334996558843774e-05, "loss": 0.9995, "step": 65850 }, { "epoch": 11.33860977288369, "grad_norm": 11.919471740722656, "learning_rate": 4.4330695113558155e-05, "loss": 1.0749, "step": 65900 }, { "epoch": 11.34721266345492, "grad_norm": 5.770901679992676, "learning_rate": 4.432639366827254e-05, "loss": 1.0074, "step": 65950 }, { "epoch": 11.355815554026153, "grad_norm": 9.393926620483398, "learning_rate": 4.4322092222986924e-05, "loss": 1.0068, "step": 66000 }, { "epoch": 11.364418444597385, "grad_norm": 9.636308670043945, "learning_rate": 4.431779077770131e-05, "loss": 1.0056, "step": 66050 }, { "epoch": 11.373021335168616, "grad_norm": 13.58751392364502, "learning_rate": 4.431348933241569e-05, "loss": 1.0624, "step": 66100 }, { "epoch": 11.381624225739849, "grad_norm": 10.19518756866455, "learning_rate": 4.4309187887130074e-05, "loss": 1.0846, "step": 66150 }, { "epoch": 11.390227116311081, "grad_norm": 9.036636352539062, "learning_rate": 4.430488644184446e-05, "loss": 1.033, "step": 66200 }, { "epoch": 11.398830006882312, "grad_norm": 6.861222267150879, "learning_rate": 4.430058499655885e-05, "loss": 0.9934, "step": 66250 }, { "epoch": 11.407432897453544, "grad_norm": 11.904183387756348, "learning_rate": 4.429628355127323e-05, "loss": 1.0869, "step": 66300 }, { "epoch": 11.416035788024777, "grad_norm": 17.239551544189453, "learning_rate": 4.429198210598761e-05, "loss": 1.1233, "step": 66350 }, { "epoch": 11.424638678596008, "grad_norm": 13.022936820983887, "learning_rate": 4.4287680660701994e-05, "loss": 1.0465, "step": 66400 }, { "epoch": 11.43324156916724, "grad_norm": 14.921419143676758, "learning_rate": 4.428337921541638e-05, "loss": 1.0395, "step": 66450 }, { "epoch": 11.441844459738473, "grad_norm": 8.820733070373535, "learning_rate": 4.427907777013077e-05, "loss": 0.9844, "step": 66500 }, { "epoch": 11.450447350309704, "grad_norm": 8.320117950439453, "learning_rate": 4.427477632484515e-05, "loss": 1.0129, "step": 66550 }, { "epoch": 11.459050240880936, "grad_norm": 9.357397079467773, "learning_rate": 4.427047487955953e-05, "loss": 0.9864, "step": 66600 }, { "epoch": 11.467653131452169, "grad_norm": 6.300539970397949, "learning_rate": 4.426617343427391e-05, "loss": 1.0093, "step": 66650 }, { "epoch": 11.4762560220234, "grad_norm": 10.756897926330566, "learning_rate": 4.42618719889883e-05, "loss": 0.9943, "step": 66700 }, { "epoch": 11.484858912594632, "grad_norm": 6.166489124298096, "learning_rate": 4.425757054370269e-05, "loss": 0.9878, "step": 66750 }, { "epoch": 11.493461803165864, "grad_norm": 7.916852951049805, "learning_rate": 4.425326909841707e-05, "loss": 1.0667, "step": 66800 }, { "epoch": 11.502064693737095, "grad_norm": 14.503311157226562, "learning_rate": 4.424896765313146e-05, "loss": 1.0493, "step": 66850 }, { "epoch": 11.510667584308328, "grad_norm": 11.017976760864258, "learning_rate": 4.424466620784584e-05, "loss": 1.0511, "step": 66900 }, { "epoch": 11.51927047487956, "grad_norm": 11.471643447875977, "learning_rate": 4.424036476256022e-05, "loss": 1.0214, "step": 66950 }, { "epoch": 11.527873365450791, "grad_norm": 9.916298866271973, "learning_rate": 4.423606331727461e-05, "loss": 1.0849, "step": 67000 }, { "epoch": 11.536476256022024, "grad_norm": 6.584139823913574, "learning_rate": 4.423176187198899e-05, "loss": 1.0203, "step": 67050 }, { "epoch": 11.545079146593256, "grad_norm": 6.960028648376465, "learning_rate": 4.422746042670338e-05, "loss": 1.0891, "step": 67100 }, { "epoch": 11.553682037164487, "grad_norm": 9.302091598510742, "learning_rate": 4.422315898141776e-05, "loss": 0.9822, "step": 67150 }, { "epoch": 11.56228492773572, "grad_norm": 9.037643432617188, "learning_rate": 4.421885753613214e-05, "loss": 1.0463, "step": 67200 }, { "epoch": 11.570887818306952, "grad_norm": 7.522223949432373, "learning_rate": 4.421455609084653e-05, "loss": 0.9895, "step": 67250 }, { "epoch": 11.579490708878183, "grad_norm": 13.315877914428711, "learning_rate": 4.421025464556091e-05, "loss": 1.013, "step": 67300 }, { "epoch": 11.588093599449415, "grad_norm": 8.070871353149414, "learning_rate": 4.42059532002753e-05, "loss": 1.0749, "step": 67350 }, { "epoch": 11.596696490020648, "grad_norm": 8.514222145080566, "learning_rate": 4.420165175498968e-05, "loss": 1.0351, "step": 67400 }, { "epoch": 11.605299380591878, "grad_norm": 12.164554595947266, "learning_rate": 4.419735030970406e-05, "loss": 1.0802, "step": 67450 }, { "epoch": 11.613902271163111, "grad_norm": 10.417610168457031, "learning_rate": 4.419304886441845e-05, "loss": 0.9861, "step": 67500 }, { "epoch": 11.622505161734344, "grad_norm": 7.237798690795898, "learning_rate": 4.418874741913283e-05, "loss": 1.0345, "step": 67550 }, { "epoch": 11.631108052305574, "grad_norm": 9.191650390625, "learning_rate": 4.4184445973847217e-05, "loss": 0.9899, "step": 67600 }, { "epoch": 11.639710942876807, "grad_norm": 9.45376205444336, "learning_rate": 4.4180144528561605e-05, "loss": 0.9812, "step": 67650 }, { "epoch": 11.64831383344804, "grad_norm": 8.39510726928711, "learning_rate": 4.417584308327598e-05, "loss": 0.989, "step": 67700 }, { "epoch": 11.65691672401927, "grad_norm": 16.779462814331055, "learning_rate": 4.417154163799037e-05, "loss": 1.0657, "step": 67750 }, { "epoch": 11.665519614590503, "grad_norm": 9.625822067260742, "learning_rate": 4.416724019270475e-05, "loss": 1.0546, "step": 67800 }, { "epoch": 11.674122505161733, "grad_norm": 12.334480285644531, "learning_rate": 4.4162938747419136e-05, "loss": 1.0111, "step": 67850 }, { "epoch": 11.682725395732966, "grad_norm": 8.511658668518066, "learning_rate": 4.4158637302133524e-05, "loss": 1.0423, "step": 67900 }, { "epoch": 11.691328286304199, "grad_norm": 9.60682487487793, "learning_rate": 4.41543358568479e-05, "loss": 1.0634, "step": 67950 }, { "epoch": 11.699931176875431, "grad_norm": 7.805917263031006, "learning_rate": 4.4150034411562286e-05, "loss": 1.0963, "step": 68000 }, { "epoch": 11.708534067446662, "grad_norm": 9.508646965026855, "learning_rate": 4.414573296627667e-05, "loss": 1.0515, "step": 68050 }, { "epoch": 11.717136958017894, "grad_norm": 7.786457538604736, "learning_rate": 4.4141431520991055e-05, "loss": 1.0995, "step": 68100 }, { "epoch": 11.725739848589125, "grad_norm": 7.00074577331543, "learning_rate": 4.413713007570544e-05, "loss": 0.9953, "step": 68150 }, { "epoch": 11.734342739160358, "grad_norm": 17.949260711669922, "learning_rate": 4.4132828630419825e-05, "loss": 1.0393, "step": 68200 }, { "epoch": 11.74294562973159, "grad_norm": 9.419028282165527, "learning_rate": 4.4128527185134206e-05, "loss": 1.0387, "step": 68250 }, { "epoch": 11.751548520302821, "grad_norm": 12.782201766967773, "learning_rate": 4.4124225739848594e-05, "loss": 1.0446, "step": 68300 }, { "epoch": 11.760151410874053, "grad_norm": 5.427596569061279, "learning_rate": 4.4119924294562975e-05, "loss": 1.0692, "step": 68350 }, { "epoch": 11.768754301445286, "grad_norm": 8.391310691833496, "learning_rate": 4.411562284927736e-05, "loss": 1.0548, "step": 68400 }, { "epoch": 11.777357192016517, "grad_norm": 8.454948425292969, "learning_rate": 4.4111321403991744e-05, "loss": 1.0951, "step": 68450 }, { "epoch": 11.78596008258775, "grad_norm": 9.014921188354492, "learning_rate": 4.4107019958706125e-05, "loss": 1.0518, "step": 68500 }, { "epoch": 11.794562973158982, "grad_norm": 7.117879390716553, "learning_rate": 4.410271851342051e-05, "loss": 1.0463, "step": 68550 }, { "epoch": 11.803165863730213, "grad_norm": 7.027888774871826, "learning_rate": 4.4098417068134894e-05, "loss": 1.0268, "step": 68600 }, { "epoch": 11.811768754301445, "grad_norm": 11.803406715393066, "learning_rate": 4.409411562284928e-05, "loss": 1.0342, "step": 68650 }, { "epoch": 11.820371644872678, "grad_norm": 7.135079383850098, "learning_rate": 4.408981417756366e-05, "loss": 1.0168, "step": 68700 }, { "epoch": 11.828974535443908, "grad_norm": 7.295753002166748, "learning_rate": 4.4085512732278045e-05, "loss": 0.986, "step": 68750 }, { "epoch": 11.837577426015141, "grad_norm": 11.762704849243164, "learning_rate": 4.408121128699243e-05, "loss": 0.9977, "step": 68800 }, { "epoch": 11.846180316586373, "grad_norm": 12.376636505126953, "learning_rate": 4.4076909841706814e-05, "loss": 1.0134, "step": 68850 }, { "epoch": 11.854783207157604, "grad_norm": 6.78245210647583, "learning_rate": 4.40726083964212e-05, "loss": 1.0203, "step": 68900 }, { "epoch": 11.863386097728837, "grad_norm": 9.237421989440918, "learning_rate": 4.406830695113558e-05, "loss": 1.0474, "step": 68950 }, { "epoch": 11.87198898830007, "grad_norm": 5.91409969329834, "learning_rate": 4.406400550584997e-05, "loss": 1.0175, "step": 69000 }, { "epoch": 11.8805918788713, "grad_norm": 8.757611274719238, "learning_rate": 4.405970406056435e-05, "loss": 1.0464, "step": 69050 }, { "epoch": 11.889194769442533, "grad_norm": 8.635917663574219, "learning_rate": 4.405540261527873e-05, "loss": 1.0192, "step": 69100 }, { "epoch": 11.897797660013765, "grad_norm": 9.760272026062012, "learning_rate": 4.405110116999312e-05, "loss": 0.9948, "step": 69150 }, { "epoch": 11.906400550584996, "grad_norm": 8.781262397766113, "learning_rate": 4.40467997247075e-05, "loss": 1.0716, "step": 69200 }, { "epoch": 11.915003441156228, "grad_norm": 10.2241849899292, "learning_rate": 4.404249827942189e-05, "loss": 1.0254, "step": 69250 }, { "epoch": 11.923606331727461, "grad_norm": 11.030665397644043, "learning_rate": 4.403819683413627e-05, "loss": 1.0225, "step": 69300 }, { "epoch": 11.932209222298692, "grad_norm": 9.202460289001465, "learning_rate": 4.403389538885065e-05, "loss": 1.0433, "step": 69350 }, { "epoch": 11.940812112869924, "grad_norm": 9.906309127807617, "learning_rate": 4.402959394356504e-05, "loss": 1.054, "step": 69400 }, { "epoch": 11.949415003441157, "grad_norm": 12.063810348510742, "learning_rate": 4.402529249827942e-05, "loss": 1.0195, "step": 69450 }, { "epoch": 11.958017894012388, "grad_norm": 11.338631629943848, "learning_rate": 4.402099105299381e-05, "loss": 1.0723, "step": 69500 }, { "epoch": 11.96662078458362, "grad_norm": 8.463791847229004, "learning_rate": 4.40166896077082e-05, "loss": 1.0046, "step": 69550 }, { "epoch": 11.975223675154853, "grad_norm": 9.627805709838867, "learning_rate": 4.401238816242257e-05, "loss": 1.0373, "step": 69600 }, { "epoch": 11.983826565726083, "grad_norm": 12.623554229736328, "learning_rate": 4.400808671713696e-05, "loss": 1.0228, "step": 69650 }, { "epoch": 11.992429456297316, "grad_norm": 6.710487365722656, "learning_rate": 4.400378527185134e-05, "loss": 0.9792, "step": 69700 }, { "epoch": 12.0, "eval_accuracy": 0.5909482016864567, "eval_f1": 0.5966655416425901, "eval_f1_DuraRiadoRio_16x16": 0.5866561389656534, "eval_f1_Mole_16x16": 0.6632485643970467, "eval_f1_Quebrado_16x16": 0.6247933884297521, "eval_f1_RiadoRio_16x16": 0.4949762030671602, "eval_f1_RioFechado_16x16": 0.6136534133533383, "eval_loss": 1.112992763519287, "eval_precision": 0.6336140773156905, "eval_precision_DuraRiadoRio_16x16": 0.5380159304851557, "eval_precision_Mole_16x16": 0.6286936236391913, "eval_precision_Quebrado_16x16": 0.8552036199095022, "eval_precision_RiadoRio_16x16": 0.4331996297439062, "eval_precision_RioFechado_16x16": 0.7129575828006973, "eval_recall": 0.590981740709507, "eval_recall_DuraRiadoRio_16x16": 0.6449652777777778, "eval_recall_Mole_16x16": 0.7018229166666666, "eval_recall_Quebrado_16x16": 0.4921875, "eval_recall_RiadoRio_16x16": 0.5773026315789473, "eval_recall_RioFechado_16x16": 0.538630377524144, "eval_runtime": 46.1866, "eval_samples_per_second": 251.631, "eval_steps_per_second": 15.74, "step": 69744 }, { "epoch": 12.001032346868548, "grad_norm": 8.323135375976562, "learning_rate": 4.399948382656573e-05, "loss": 1.0329, "step": 69750 }, { "epoch": 12.00963523743978, "grad_norm": 10.818211555480957, "learning_rate": 4.399518238128012e-05, "loss": 1.0315, "step": 69800 }, { "epoch": 12.018238128011012, "grad_norm": 12.659343719482422, "learning_rate": 4.399088093599449e-05, "loss": 0.9772, "step": 69850 }, { "epoch": 12.026841018582244, "grad_norm": 5.7924981117248535, "learning_rate": 4.398657949070888e-05, "loss": 1.0044, "step": 69900 }, { "epoch": 12.035443909153475, "grad_norm": 7.791136741638184, "learning_rate": 4.398227804542327e-05, "loss": 0.9775, "step": 69950 }, { "epoch": 12.044046799724708, "grad_norm": 10.680967330932617, "learning_rate": 4.397797660013765e-05, "loss": 0.9892, "step": 70000 }, { "epoch": 12.05264969029594, "grad_norm": 10.081522941589355, "learning_rate": 4.3973675154852036e-05, "loss": 1.0294, "step": 70050 }, { "epoch": 12.061252580867171, "grad_norm": 9.84035587310791, "learning_rate": 4.396937370956641e-05, "loss": 1.0075, "step": 70100 }, { "epoch": 12.069855471438403, "grad_norm": 9.310629844665527, "learning_rate": 4.39650722642808e-05, "loss": 0.9827, "step": 70150 }, { "epoch": 12.078458362009636, "grad_norm": 9.309300422668457, "learning_rate": 4.3960770818995187e-05, "loss": 0.9923, "step": 70200 }, { "epoch": 12.087061252580867, "grad_norm": 7.871562957763672, "learning_rate": 4.395646937370957e-05, "loss": 1.0148, "step": 70250 }, { "epoch": 12.0956641431521, "grad_norm": 4.9004034996032715, "learning_rate": 4.3952167928423956e-05, "loss": 1.0143, "step": 70300 }, { "epoch": 12.104267033723332, "grad_norm": 7.253734111785889, "learning_rate": 4.394786648313834e-05, "loss": 1.0083, "step": 70350 }, { "epoch": 12.112869924294563, "grad_norm": 16.36159324645996, "learning_rate": 4.394356503785272e-05, "loss": 1.0101, "step": 70400 }, { "epoch": 12.121472814865795, "grad_norm": 7.663978576660156, "learning_rate": 4.3939263592567106e-05, "loss": 1.0301, "step": 70450 }, { "epoch": 12.130075705437028, "grad_norm": 7.737570285797119, "learning_rate": 4.393496214728149e-05, "loss": 1.0489, "step": 70500 }, { "epoch": 12.138678596008258, "grad_norm": 7.040054798126221, "learning_rate": 4.3930660701995875e-05, "loss": 1.0074, "step": 70550 }, { "epoch": 12.147281486579491, "grad_norm": 7.705316066741943, "learning_rate": 4.3926359256710256e-05, "loss": 1.0406, "step": 70600 }, { "epoch": 12.155884377150723, "grad_norm": 8.744702339172363, "learning_rate": 4.392205781142464e-05, "loss": 1.0382, "step": 70650 }, { "epoch": 12.164487267721954, "grad_norm": 8.745141983032227, "learning_rate": 4.3917756366139025e-05, "loss": 1.0048, "step": 70700 }, { "epoch": 12.173090158293187, "grad_norm": 7.582046031951904, "learning_rate": 4.3913454920853407e-05, "loss": 1.0122, "step": 70750 }, { "epoch": 12.18169304886442, "grad_norm": 7.843523979187012, "learning_rate": 4.3909153475567795e-05, "loss": 1.0505, "step": 70800 }, { "epoch": 12.19029593943565, "grad_norm": 11.188311576843262, "learning_rate": 4.3904852030282176e-05, "loss": 0.9755, "step": 70850 }, { "epoch": 12.198898830006883, "grad_norm": 8.463284492492676, "learning_rate": 4.3900550584996564e-05, "loss": 1.0323, "step": 70900 }, { "epoch": 12.207501720578115, "grad_norm": 7.629931926727295, "learning_rate": 4.3896249139710945e-05, "loss": 1.0515, "step": 70950 }, { "epoch": 12.216104611149346, "grad_norm": 18.167449951171875, "learning_rate": 4.3891947694425326e-05, "loss": 0.9793, "step": 71000 }, { "epoch": 12.224707501720578, "grad_norm": 9.327773094177246, "learning_rate": 4.3887646249139714e-05, "loss": 0.9734, "step": 71050 }, { "epoch": 12.23331039229181, "grad_norm": 8.255547523498535, "learning_rate": 4.3883344803854095e-05, "loss": 1.0216, "step": 71100 }, { "epoch": 12.241913282863042, "grad_norm": 7.3647613525390625, "learning_rate": 4.387904335856848e-05, "loss": 0.9806, "step": 71150 }, { "epoch": 12.250516173434274, "grad_norm": 10.188752174377441, "learning_rate": 4.3874741913282864e-05, "loss": 1.0784, "step": 71200 }, { "epoch": 12.259119064005505, "grad_norm": 9.678730964660645, "learning_rate": 4.3870440467997245e-05, "loss": 0.9834, "step": 71250 }, { "epoch": 12.267721954576738, "grad_norm": 11.209688186645508, "learning_rate": 4.386613902271163e-05, "loss": 1.0324, "step": 71300 }, { "epoch": 12.27632484514797, "grad_norm": 7.837023735046387, "learning_rate": 4.386183757742602e-05, "loss": 0.9757, "step": 71350 }, { "epoch": 12.2849277357192, "grad_norm": 8.617661476135254, "learning_rate": 4.38575361321404e-05, "loss": 0.9793, "step": 71400 }, { "epoch": 12.293530626290433, "grad_norm": 9.284584045410156, "learning_rate": 4.3853234686854784e-05, "loss": 1.021, "step": 71450 }, { "epoch": 12.302133516861666, "grad_norm": 9.359448432922363, "learning_rate": 4.3848933241569165e-05, "loss": 0.9422, "step": 71500 }, { "epoch": 12.310736407432897, "grad_norm": 11.321222305297852, "learning_rate": 4.384463179628355e-05, "loss": 1.0177, "step": 71550 }, { "epoch": 12.31933929800413, "grad_norm": 12.167092323303223, "learning_rate": 4.384033035099794e-05, "loss": 1.0206, "step": 71600 }, { "epoch": 12.327942188575362, "grad_norm": 7.627601623535156, "learning_rate": 4.383602890571232e-05, "loss": 0.9961, "step": 71650 }, { "epoch": 12.336545079146592, "grad_norm": 15.104934692382812, "learning_rate": 4.383172746042671e-05, "loss": 1.0504, "step": 71700 }, { "epoch": 12.345147969717825, "grad_norm": 8.173483848571777, "learning_rate": 4.3827426015141084e-05, "loss": 1.0008, "step": 71750 }, { "epoch": 12.353750860289058, "grad_norm": 11.463541030883789, "learning_rate": 4.382312456985547e-05, "loss": 1.0, "step": 71800 }, { "epoch": 12.362353750860288, "grad_norm": 10.826177597045898, "learning_rate": 4.381882312456986e-05, "loss": 1.0129, "step": 71850 }, { "epoch": 12.37095664143152, "grad_norm": 7.169140338897705, "learning_rate": 4.381452167928424e-05, "loss": 1.0506, "step": 71900 }, { "epoch": 12.379559532002753, "grad_norm": 9.245952606201172, "learning_rate": 4.381022023399863e-05, "loss": 1.0221, "step": 71950 }, { "epoch": 12.388162422573984, "grad_norm": 8.805835723876953, "learning_rate": 4.380591878871301e-05, "loss": 1.0214, "step": 72000 }, { "epoch": 12.396765313145217, "grad_norm": 8.561253547668457, "learning_rate": 4.380161734342739e-05, "loss": 0.9948, "step": 72050 }, { "epoch": 12.40536820371645, "grad_norm": 11.316656112670898, "learning_rate": 4.379731589814178e-05, "loss": 1.0281, "step": 72100 }, { "epoch": 12.41397109428768, "grad_norm": 10.27090072631836, "learning_rate": 4.379301445285616e-05, "loss": 1.0188, "step": 72150 }, { "epoch": 12.422573984858913, "grad_norm": 9.137163162231445, "learning_rate": 4.378871300757055e-05, "loss": 1.0332, "step": 72200 }, { "epoch": 12.431176875430145, "grad_norm": 12.43577766418457, "learning_rate": 4.378441156228493e-05, "loss": 1.0151, "step": 72250 }, { "epoch": 12.439779766001376, "grad_norm": 11.198251724243164, "learning_rate": 4.378011011699931e-05, "loss": 1.0116, "step": 72300 }, { "epoch": 12.448382656572608, "grad_norm": 8.049688339233398, "learning_rate": 4.37758086717137e-05, "loss": 1.0232, "step": 72350 }, { "epoch": 12.45698554714384, "grad_norm": 9.955120086669922, "learning_rate": 4.377150722642808e-05, "loss": 1.023, "step": 72400 }, { "epoch": 12.465588437715072, "grad_norm": 9.309581756591797, "learning_rate": 4.376720578114247e-05, "loss": 1.0068, "step": 72450 }, { "epoch": 12.474191328286304, "grad_norm": 8.27065658569336, "learning_rate": 4.376290433585685e-05, "loss": 1.0067, "step": 72500 }, { "epoch": 12.482794218857537, "grad_norm": 6.19020938873291, "learning_rate": 4.375860289057123e-05, "loss": 1.021, "step": 72550 }, { "epoch": 12.491397109428767, "grad_norm": 12.244876861572266, "learning_rate": 4.375430144528562e-05, "loss": 0.9833, "step": 72600 }, { "epoch": 12.5, "grad_norm": 7.269354343414307, "learning_rate": 4.375e-05, "loss": 1.0535, "step": 72650 }, { "epoch": 12.508602890571233, "grad_norm": 8.799257278442383, "learning_rate": 4.374569855471439e-05, "loss": 1.023, "step": 72700 }, { "epoch": 12.517205781142463, "grad_norm": 12.544415473937988, "learning_rate": 4.3741397109428775e-05, "loss": 1.0374, "step": 72750 }, { "epoch": 12.525808671713696, "grad_norm": 10.344293594360352, "learning_rate": 4.373709566414315e-05, "loss": 1.0486, "step": 72800 }, { "epoch": 12.534411562284928, "grad_norm": 8.505773544311523, "learning_rate": 4.373279421885754e-05, "loss": 1.024, "step": 72850 }, { "epoch": 12.54301445285616, "grad_norm": 6.2822723388671875, "learning_rate": 4.372849277357192e-05, "loss": 0.9776, "step": 72900 }, { "epoch": 12.551617343427392, "grad_norm": 5.889500141143799, "learning_rate": 4.372419132828631e-05, "loss": 1.0166, "step": 72950 }, { "epoch": 12.560220233998624, "grad_norm": 7.608059406280518, "learning_rate": 4.3719889883000695e-05, "loss": 1.0055, "step": 73000 }, { "epoch": 12.568823124569855, "grad_norm": 7.996105670928955, "learning_rate": 4.3715588437715076e-05, "loss": 1.0175, "step": 73050 }, { "epoch": 12.577426015141087, "grad_norm": 9.381621360778809, "learning_rate": 4.371128699242946e-05, "loss": 1.0281, "step": 73100 }, { "epoch": 12.58602890571232, "grad_norm": 8.266741752624512, "learning_rate": 4.370698554714384e-05, "loss": 1.0388, "step": 73150 }, { "epoch": 12.59463179628355, "grad_norm": 14.274736404418945, "learning_rate": 4.3702684101858226e-05, "loss": 1.016, "step": 73200 }, { "epoch": 12.603234686854783, "grad_norm": 12.064790725708008, "learning_rate": 4.3698382656572614e-05, "loss": 1.026, "step": 73250 }, { "epoch": 12.611837577426016, "grad_norm": 11.791854858398438, "learning_rate": 4.3694081211286995e-05, "loss": 1.0143, "step": 73300 }, { "epoch": 12.620440467997247, "grad_norm": 7.73874568939209, "learning_rate": 4.368977976600138e-05, "loss": 0.9944, "step": 73350 }, { "epoch": 12.62904335856848, "grad_norm": 10.642823219299316, "learning_rate": 4.3685478320715765e-05, "loss": 1.043, "step": 73400 }, { "epoch": 12.637646249139712, "grad_norm": 7.605630397796631, "learning_rate": 4.3681176875430146e-05, "loss": 1.0522, "step": 73450 }, { "epoch": 12.646249139710942, "grad_norm": 10.598270416259766, "learning_rate": 4.3676875430144534e-05, "loss": 1.0516, "step": 73500 }, { "epoch": 12.654852030282175, "grad_norm": 8.861499786376953, "learning_rate": 4.3672573984858915e-05, "loss": 0.9989, "step": 73550 }, { "epoch": 12.663454920853408, "grad_norm": 10.998734474182129, "learning_rate": 4.3668272539573296e-05, "loss": 0.99, "step": 73600 }, { "epoch": 12.672057811424638, "grad_norm": 9.588033676147461, "learning_rate": 4.3663971094287684e-05, "loss": 1.0224, "step": 73650 }, { "epoch": 12.68066070199587, "grad_norm": 6.149745941162109, "learning_rate": 4.3659669649002065e-05, "loss": 1.0251, "step": 73700 }, { "epoch": 12.689263592567103, "grad_norm": 8.496808052062988, "learning_rate": 4.365536820371645e-05, "loss": 1.0012, "step": 73750 }, { "epoch": 12.697866483138334, "grad_norm": 8.759016036987305, "learning_rate": 4.3651066758430834e-05, "loss": 1.0067, "step": 73800 }, { "epoch": 12.706469373709567, "grad_norm": 7.60256290435791, "learning_rate": 4.364676531314522e-05, "loss": 1.0444, "step": 73850 }, { "epoch": 12.7150722642808, "grad_norm": 11.65133285522461, "learning_rate": 4.3642463867859603e-05, "loss": 1.0629, "step": 73900 }, { "epoch": 12.72367515485203, "grad_norm": 10.100995063781738, "learning_rate": 4.3638162422573985e-05, "loss": 0.9812, "step": 73950 }, { "epoch": 12.732278045423262, "grad_norm": 8.958033561706543, "learning_rate": 4.363386097728837e-05, "loss": 0.9928, "step": 74000 }, { "epoch": 12.740880935994493, "grad_norm": 10.833667755126953, "learning_rate": 4.3629559532002754e-05, "loss": 0.9719, "step": 74050 }, { "epoch": 12.749483826565726, "grad_norm": 9.16603946685791, "learning_rate": 4.362525808671714e-05, "loss": 0.979, "step": 74100 }, { "epoch": 12.758086717136958, "grad_norm": 6.960202693939209, "learning_rate": 4.362095664143152e-05, "loss": 1.1021, "step": 74150 }, { "epoch": 12.76668960770819, "grad_norm": 7.133851528167725, "learning_rate": 4.3616655196145904e-05, "loss": 1.0004, "step": 74200 }, { "epoch": 12.775292498279422, "grad_norm": 9.98944091796875, "learning_rate": 4.361235375086029e-05, "loss": 1.0229, "step": 74250 }, { "epoch": 12.783895388850654, "grad_norm": 9.497576713562012, "learning_rate": 4.360805230557467e-05, "loss": 0.9967, "step": 74300 }, { "epoch": 12.792498279421885, "grad_norm": 9.442399024963379, "learning_rate": 4.360375086028906e-05, "loss": 1.0305, "step": 74350 }, { "epoch": 12.801101169993117, "grad_norm": 8.481595993041992, "learning_rate": 4.359944941500344e-05, "loss": 1.0028, "step": 74400 }, { "epoch": 12.80970406056435, "grad_norm": 14.361505508422852, "learning_rate": 4.3595147969717823e-05, "loss": 0.9688, "step": 74450 }, { "epoch": 12.818306951135582, "grad_norm": 11.282585144042969, "learning_rate": 4.359084652443221e-05, "loss": 1.0186, "step": 74500 }, { "epoch": 12.826909841706813, "grad_norm": 8.780722618103027, "learning_rate": 4.358654507914659e-05, "loss": 1.056, "step": 74550 }, { "epoch": 12.835512732278046, "grad_norm": 10.630980491638184, "learning_rate": 4.358224363386098e-05, "loss": 0.9863, "step": 74600 }, { "epoch": 12.844115622849277, "grad_norm": 7.289377212524414, "learning_rate": 4.357794218857537e-05, "loss": 1.0887, "step": 74650 }, { "epoch": 12.852718513420509, "grad_norm": 7.743929862976074, "learning_rate": 4.357364074328974e-05, "loss": 0.9614, "step": 74700 }, { "epoch": 12.861321403991742, "grad_norm": 10.2237548828125, "learning_rate": 4.356933929800413e-05, "loss": 1.0383, "step": 74750 }, { "epoch": 12.869924294562972, "grad_norm": 11.546788215637207, "learning_rate": 4.356503785271852e-05, "loss": 1.0604, "step": 74800 }, { "epoch": 12.878527185134205, "grad_norm": 11.987998008728027, "learning_rate": 4.35607364074329e-05, "loss": 1.0101, "step": 74850 }, { "epoch": 12.887130075705437, "grad_norm": 10.122774124145508, "learning_rate": 4.355643496214729e-05, "loss": 1.0538, "step": 74900 }, { "epoch": 12.895732966276668, "grad_norm": 9.912205696105957, "learning_rate": 4.355213351686166e-05, "loss": 1.0065, "step": 74950 }, { "epoch": 12.9043358568479, "grad_norm": 11.500368118286133, "learning_rate": 4.354783207157605e-05, "loss": 1.0327, "step": 75000 }, { "epoch": 12.912938747419133, "grad_norm": 10.022310256958008, "learning_rate": 4.354353062629044e-05, "loss": 1.019, "step": 75050 }, { "epoch": 12.921541637990364, "grad_norm": 8.112092018127441, "learning_rate": 4.353922918100482e-05, "loss": 1.014, "step": 75100 }, { "epoch": 12.930144528561597, "grad_norm": 8.137290000915527, "learning_rate": 4.353492773571921e-05, "loss": 1.0266, "step": 75150 }, { "epoch": 12.938747419132829, "grad_norm": 7.911017417907715, "learning_rate": 4.353062629043359e-05, "loss": 0.9959, "step": 75200 }, { "epoch": 12.94735030970406, "grad_norm": 10.406177520751953, "learning_rate": 4.352632484514797e-05, "loss": 1.0023, "step": 75250 }, { "epoch": 12.955953200275292, "grad_norm": 11.668051719665527, "learning_rate": 4.352202339986236e-05, "loss": 1.0259, "step": 75300 }, { "epoch": 12.964556090846525, "grad_norm": 9.252588272094727, "learning_rate": 4.351772195457674e-05, "loss": 1.0866, "step": 75350 }, { "epoch": 12.973158981417756, "grad_norm": 10.886812210083008, "learning_rate": 4.351342050929113e-05, "loss": 1.013, "step": 75400 }, { "epoch": 12.981761871988988, "grad_norm": 10.325173377990723, "learning_rate": 4.350911906400551e-05, "loss": 1.006, "step": 75450 }, { "epoch": 12.99036476256022, "grad_norm": 7.973714351654053, "learning_rate": 4.350481761871989e-05, "loss": 1.0061, "step": 75500 }, { "epoch": 12.998967653131452, "grad_norm": 7.727682113647461, "learning_rate": 4.350051617343428e-05, "loss": 0.9902, "step": 75550 }, { "epoch": 13.0, "eval_accuracy": 0.5068834968163827, "eval_f1": 0.46922045733730877, "eval_f1_DuraRiadoRio_16x16": 0.2967741935483871, "eval_f1_Mole_16x16": 0.3480825958702065, "eval_f1_Quebrado_16x16": 0.6962524654832347, "eval_f1_RiadoRio_16x16": 0.5089413447782547, "eval_f1_RioFechado_16x16": 0.49605168700646085, "eval_loss": 1.7618135213851929, "eval_precision": 0.5645550515275315, "eval_precision_DuraRiadoRio_16x16": 0.6817472698907956, "eval_precision_Mole_16x16": 0.7108433734939759, "eval_precision_Quebrado_16x16": 0.5603174603174603, "eval_precision_RiadoRio_16x16": 0.45031645569620254, "eval_precision_RioFechado_16x16": 0.4195506982392228, "eval_recall": 0.506239474711068, "eval_recall_DuraRiadoRio_16x16": 0.1896701388888889, "eval_recall_Mole_16x16": 0.23046875, "eval_recall_Quebrado_16x16": 0.9192708333333334, "eval_recall_RiadoRio_16x16": 0.5851151315789473, "eval_recall_RioFechado_16x16": 0.6066725197541704, "eval_runtime": 45.6724, "eval_samples_per_second": 254.464, "eval_steps_per_second": 15.918, "step": 75556 }, { "epoch": 13.007570543702684, "grad_norm": 7.921777248382568, "learning_rate": 4.349621472814866e-05, "loss": 1.0631, "step": 75600 }, { "epoch": 13.016173434273917, "grad_norm": 9.266386985778809, "learning_rate": 4.3491913282863046e-05, "loss": 0.9778, "step": 75650 }, { "epoch": 13.024776324845147, "grad_norm": 11.47981071472168, "learning_rate": 4.348761183757743e-05, "loss": 1.0239, "step": 75700 }, { "epoch": 13.03337921541638, "grad_norm": 8.644224166870117, "learning_rate": 4.348331039229181e-05, "loss": 0.9916, "step": 75750 }, { "epoch": 13.041982105987612, "grad_norm": 7.934998512268066, "learning_rate": 4.3479008947006196e-05, "loss": 0.9913, "step": 75800 }, { "epoch": 13.050584996558843, "grad_norm": 11.14716911315918, "learning_rate": 4.347470750172058e-05, "loss": 0.9656, "step": 75850 }, { "epoch": 13.059187887130076, "grad_norm": 8.072668075561523, "learning_rate": 4.3470406056434965e-05, "loss": 1.0199, "step": 75900 }, { "epoch": 13.067790777701308, "grad_norm": 11.748513221740723, "learning_rate": 4.346610461114935e-05, "loss": 1.0548, "step": 75950 }, { "epoch": 13.076393668272539, "grad_norm": 10.173382759094238, "learning_rate": 4.3461803165863735e-05, "loss": 0.9467, "step": 76000 }, { "epoch": 13.084996558843772, "grad_norm": 9.066375732421875, "learning_rate": 4.3457501720578116e-05, "loss": 1.0532, "step": 76050 }, { "epoch": 13.093599449415004, "grad_norm": 8.831941604614258, "learning_rate": 4.34532002752925e-05, "loss": 0.9879, "step": 76100 }, { "epoch": 13.102202339986235, "grad_norm": 7.981929302215576, "learning_rate": 4.3448898830006885e-05, "loss": 1.0233, "step": 76150 }, { "epoch": 13.110805230557467, "grad_norm": 9.813725471496582, "learning_rate": 4.344459738472127e-05, "loss": 1.0306, "step": 76200 }, { "epoch": 13.1194081211287, "grad_norm": 9.232442855834961, "learning_rate": 4.3440295939435654e-05, "loss": 1.0024, "step": 76250 }, { "epoch": 13.12801101169993, "grad_norm": 10.554043769836426, "learning_rate": 4.3435994494150035e-05, "loss": 0.9782, "step": 76300 }, { "epoch": 13.136613902271163, "grad_norm": 10.293919563293457, "learning_rate": 4.3431693048864416e-05, "loss": 1.0062, "step": 76350 }, { "epoch": 13.145216792842396, "grad_norm": 10.598396301269531, "learning_rate": 4.3427391603578804e-05, "loss": 0.9854, "step": 76400 }, { "epoch": 13.153819683413627, "grad_norm": 9.566530227661133, "learning_rate": 4.342309015829319e-05, "loss": 0.9834, "step": 76450 }, { "epoch": 13.162422573984859, "grad_norm": 7.898808479309082, "learning_rate": 4.3418788713007573e-05, "loss": 1.0103, "step": 76500 }, { "epoch": 13.171025464556092, "grad_norm": 8.22989559173584, "learning_rate": 4.341448726772196e-05, "loss": 0.9972, "step": 76550 }, { "epoch": 13.179628355127322, "grad_norm": 12.654022216796875, "learning_rate": 4.3410185822436336e-05, "loss": 0.9703, "step": 76600 }, { "epoch": 13.188231245698555, "grad_norm": 7.087377548217773, "learning_rate": 4.3405884377150724e-05, "loss": 1.0124, "step": 76650 }, { "epoch": 13.196834136269787, "grad_norm": 10.579480171203613, "learning_rate": 4.340158293186511e-05, "loss": 1.0431, "step": 76700 }, { "epoch": 13.205437026841018, "grad_norm": 8.593647956848145, "learning_rate": 4.339728148657949e-05, "loss": 1.0202, "step": 76750 }, { "epoch": 13.21403991741225, "grad_norm": 9.043983459472656, "learning_rate": 4.339298004129388e-05, "loss": 1.0308, "step": 76800 }, { "epoch": 13.222642807983483, "grad_norm": 6.93137788772583, "learning_rate": 4.338867859600826e-05, "loss": 1.0072, "step": 76850 }, { "epoch": 13.231245698554714, "grad_norm": 9.251587867736816, "learning_rate": 4.338437715072264e-05, "loss": 0.9792, "step": 76900 }, { "epoch": 13.239848589125947, "grad_norm": 9.216775894165039, "learning_rate": 4.338007570543703e-05, "loss": 1.0303, "step": 76950 }, { "epoch": 13.248451479697179, "grad_norm": 12.886672973632812, "learning_rate": 4.337577426015141e-05, "loss": 0.9646, "step": 77000 }, { "epoch": 13.25705437026841, "grad_norm": 12.682437896728516, "learning_rate": 4.33714728148658e-05, "loss": 1.036, "step": 77050 }, { "epoch": 13.265657260839642, "grad_norm": 10.836043357849121, "learning_rate": 4.336717136958018e-05, "loss": 1.0265, "step": 77100 }, { "epoch": 13.274260151410875, "grad_norm": 10.687431335449219, "learning_rate": 4.336286992429456e-05, "loss": 0.9722, "step": 77150 }, { "epoch": 13.282863041982106, "grad_norm": 13.255990028381348, "learning_rate": 4.335856847900895e-05, "loss": 1.0215, "step": 77200 }, { "epoch": 13.291465932553338, "grad_norm": 9.26196575164795, "learning_rate": 4.335426703372333e-05, "loss": 1.0179, "step": 77250 }, { "epoch": 13.300068823124569, "grad_norm": 13.192245483398438, "learning_rate": 4.334996558843772e-05, "loss": 1.0674, "step": 77300 }, { "epoch": 13.308671713695801, "grad_norm": 11.920169830322266, "learning_rate": 4.33456641431521e-05, "loss": 0.9734, "step": 77350 }, { "epoch": 13.317274604267034, "grad_norm": 7.744823932647705, "learning_rate": 4.334136269786648e-05, "loss": 1.0755, "step": 77400 }, { "epoch": 13.325877494838267, "grad_norm": 12.281566619873047, "learning_rate": 4.333706125258087e-05, "loss": 0.9703, "step": 77450 }, { "epoch": 13.334480385409497, "grad_norm": 9.489835739135742, "learning_rate": 4.333275980729525e-05, "loss": 1.0384, "step": 77500 }, { "epoch": 13.34308327598073, "grad_norm": 11.044934272766113, "learning_rate": 4.332845836200964e-05, "loss": 0.979, "step": 77550 }, { "epoch": 13.35168616655196, "grad_norm": 9.222172737121582, "learning_rate": 4.332415691672402e-05, "loss": 1.0141, "step": 77600 }, { "epoch": 13.360289057123193, "grad_norm": 8.440628051757812, "learning_rate": 4.33198554714384e-05, "loss": 1.0239, "step": 77650 }, { "epoch": 13.368891947694426, "grad_norm": 9.377918243408203, "learning_rate": 4.331555402615279e-05, "loss": 1.0007, "step": 77700 }, { "epoch": 13.377494838265656, "grad_norm": 10.01707649230957, "learning_rate": 4.331125258086717e-05, "loss": 0.959, "step": 77750 }, { "epoch": 13.386097728836889, "grad_norm": 12.543006896972656, "learning_rate": 4.330695113558156e-05, "loss": 1.0318, "step": 77800 }, { "epoch": 13.394700619408122, "grad_norm": 10.955134391784668, "learning_rate": 4.3302649690295946e-05, "loss": 1.023, "step": 77850 }, { "epoch": 13.403303509979352, "grad_norm": 6.438146591186523, "learning_rate": 4.329834824501032e-05, "loss": 1.0136, "step": 77900 }, { "epoch": 13.411906400550585, "grad_norm": 18.35066795349121, "learning_rate": 4.329404679972471e-05, "loss": 1.0366, "step": 77950 }, { "epoch": 13.420509291121817, "grad_norm": 7.427887916564941, "learning_rate": 4.328974535443909e-05, "loss": 1.0135, "step": 78000 }, { "epoch": 13.429112181693048, "grad_norm": 12.902887344360352, "learning_rate": 4.328544390915348e-05, "loss": 1.0181, "step": 78050 }, { "epoch": 13.43771507226428, "grad_norm": 11.76200008392334, "learning_rate": 4.3281142463867866e-05, "loss": 1.0257, "step": 78100 }, { "epoch": 13.446317962835513, "grad_norm": 11.441340446472168, "learning_rate": 4.327684101858225e-05, "loss": 1.0167, "step": 78150 }, { "epoch": 13.454920853406744, "grad_norm": 8.964686393737793, "learning_rate": 4.327253957329663e-05, "loss": 1.0324, "step": 78200 }, { "epoch": 13.463523743977976, "grad_norm": 10.17509651184082, "learning_rate": 4.3268238128011016e-05, "loss": 1.0505, "step": 78250 }, { "epoch": 13.472126634549209, "grad_norm": 7.324222087860107, "learning_rate": 4.32639366827254e-05, "loss": 1.0305, "step": 78300 }, { "epoch": 13.48072952512044, "grad_norm": 13.370238304138184, "learning_rate": 4.3259635237439785e-05, "loss": 1.0061, "step": 78350 }, { "epoch": 13.489332415691672, "grad_norm": 10.716375350952148, "learning_rate": 4.3255333792154166e-05, "loss": 0.9713, "step": 78400 }, { "epoch": 13.497935306262905, "grad_norm": 13.077316284179688, "learning_rate": 4.325103234686855e-05, "loss": 1.0306, "step": 78450 }, { "epoch": 13.506538196834136, "grad_norm": 17.162185668945312, "learning_rate": 4.3246730901582936e-05, "loss": 0.9961, "step": 78500 }, { "epoch": 13.515141087405368, "grad_norm": 7.333647727966309, "learning_rate": 4.324242945629732e-05, "loss": 0.9746, "step": 78550 }, { "epoch": 13.5237439779766, "grad_norm": 8.915434837341309, "learning_rate": 4.3238128011011705e-05, "loss": 1.084, "step": 78600 }, { "epoch": 13.532346868547831, "grad_norm": 7.551562786102295, "learning_rate": 4.3233826565726086e-05, "loss": 0.9625, "step": 78650 }, { "epoch": 13.540949759119064, "grad_norm": 9.7229585647583, "learning_rate": 4.3229525120440474e-05, "loss": 0.9592, "step": 78700 }, { "epoch": 13.549552649690296, "grad_norm": 10.366869926452637, "learning_rate": 4.3225223675154855e-05, "loss": 0.9896, "step": 78750 }, { "epoch": 13.558155540261527, "grad_norm": 11.212160110473633, "learning_rate": 4.3220922229869236e-05, "loss": 1.0335, "step": 78800 }, { "epoch": 13.56675843083276, "grad_norm": 5.124449729919434, "learning_rate": 4.3216620784583624e-05, "loss": 0.9749, "step": 78850 }, { "epoch": 13.575361321403992, "grad_norm": 8.685343742370605, "learning_rate": 4.3212319339298005e-05, "loss": 1.0072, "step": 78900 }, { "epoch": 13.583964211975223, "grad_norm": 9.860660552978516, "learning_rate": 4.320801789401239e-05, "loss": 1.0099, "step": 78950 }, { "epoch": 13.592567102546456, "grad_norm": 8.450360298156738, "learning_rate": 4.3203716448726774e-05, "loss": 1.0044, "step": 79000 }, { "epoch": 13.601169993117688, "grad_norm": 10.195157051086426, "learning_rate": 4.3199415003441156e-05, "loss": 0.9336, "step": 79050 }, { "epoch": 13.609772883688919, "grad_norm": 13.110608100891113, "learning_rate": 4.3195113558155543e-05, "loss": 1.03, "step": 79100 }, { "epoch": 13.618375774260151, "grad_norm": 10.44467830657959, "learning_rate": 4.3190812112869925e-05, "loss": 1.0169, "step": 79150 }, { "epoch": 13.626978664831384, "grad_norm": 7.20471715927124, "learning_rate": 4.318651066758431e-05, "loss": 0.9675, "step": 79200 }, { "epoch": 13.635581555402615, "grad_norm": 7.771047592163086, "learning_rate": 4.3182209222298694e-05, "loss": 1.0385, "step": 79250 }, { "epoch": 13.644184445973847, "grad_norm": 9.286245346069336, "learning_rate": 4.3177907777013075e-05, "loss": 1.0147, "step": 79300 }, { "epoch": 13.65278733654508, "grad_norm": 10.320023536682129, "learning_rate": 4.317360633172746e-05, "loss": 0.9824, "step": 79350 }, { "epoch": 13.66139022711631, "grad_norm": 12.081183433532715, "learning_rate": 4.3169304886441844e-05, "loss": 1.0956, "step": 79400 }, { "epoch": 13.669993117687543, "grad_norm": 10.386073112487793, "learning_rate": 4.316500344115623e-05, "loss": 0.9783, "step": 79450 }, { "epoch": 13.678596008258776, "grad_norm": 7.451290607452393, "learning_rate": 4.316070199587062e-05, "loss": 0.9852, "step": 79500 }, { "epoch": 13.687198898830006, "grad_norm": 10.668111801147461, "learning_rate": 4.3156400550584994e-05, "loss": 1.0055, "step": 79550 }, { "epoch": 13.695801789401239, "grad_norm": 9.914555549621582, "learning_rate": 4.315209910529938e-05, "loss": 1.0158, "step": 79600 }, { "epoch": 13.704404679972471, "grad_norm": 9.364931106567383, "learning_rate": 4.3147797660013763e-05, "loss": 0.9959, "step": 79650 }, { "epoch": 13.713007570543702, "grad_norm": 8.919195175170898, "learning_rate": 4.314349621472815e-05, "loss": 0.9845, "step": 79700 }, { "epoch": 13.721610461114935, "grad_norm": 14.196260452270508, "learning_rate": 4.313919476944254e-05, "loss": 1.0223, "step": 79750 }, { "epoch": 13.730213351686167, "grad_norm": 12.156351089477539, "learning_rate": 4.3134893324156914e-05, "loss": 1.0254, "step": 79800 }, { "epoch": 13.738816242257398, "grad_norm": 10.462852478027344, "learning_rate": 4.31305918788713e-05, "loss": 1.0131, "step": 79850 }, { "epoch": 13.74741913282863, "grad_norm": 8.378691673278809, "learning_rate": 4.312629043358569e-05, "loss": 0.98, "step": 79900 }, { "epoch": 13.756022023399863, "grad_norm": 9.23403263092041, "learning_rate": 4.312198898830007e-05, "loss": 1.0404, "step": 79950 }, { "epoch": 13.764624913971094, "grad_norm": 10.930680274963379, "learning_rate": 4.311768754301446e-05, "loss": 1.0372, "step": 80000 }, { "epoch": 13.773227804542326, "grad_norm": 7.675738334655762, "learning_rate": 4.311338609772883e-05, "loss": 1.0084, "step": 80050 }, { "epoch": 13.781830695113559, "grad_norm": 9.647958755493164, "learning_rate": 4.310908465244322e-05, "loss": 1.03, "step": 80100 }, { "epoch": 13.79043358568479, "grad_norm": 8.542559623718262, "learning_rate": 4.310478320715761e-05, "loss": 0.9921, "step": 80150 }, { "epoch": 13.799036476256022, "grad_norm": 11.1516695022583, "learning_rate": 4.310048176187199e-05, "loss": 0.9922, "step": 80200 }, { "epoch": 13.807639366827255, "grad_norm": 9.139967918395996, "learning_rate": 4.309618031658638e-05, "loss": 1.0241, "step": 80250 }, { "epoch": 13.816242257398486, "grad_norm": 9.063216209411621, "learning_rate": 4.309187887130076e-05, "loss": 0.9717, "step": 80300 }, { "epoch": 13.824845147969718, "grad_norm": 9.506855010986328, "learning_rate": 4.308757742601514e-05, "loss": 1.0031, "step": 80350 }, { "epoch": 13.83344803854095, "grad_norm": 8.911547660827637, "learning_rate": 4.308327598072953e-05, "loss": 0.9694, "step": 80400 }, { "epoch": 13.842050929112181, "grad_norm": 11.060297966003418, "learning_rate": 4.307897453544391e-05, "loss": 0.9613, "step": 80450 }, { "epoch": 13.850653819683414, "grad_norm": 11.188961029052734, "learning_rate": 4.30746730901583e-05, "loss": 0.9864, "step": 80500 }, { "epoch": 13.859256710254645, "grad_norm": 10.277800559997559, "learning_rate": 4.307037164487268e-05, "loss": 0.9255, "step": 80550 }, { "epoch": 13.867859600825877, "grad_norm": 6.584629058837891, "learning_rate": 4.306607019958706e-05, "loss": 0.994, "step": 80600 }, { "epoch": 13.87646249139711, "grad_norm": 7.1436028480529785, "learning_rate": 4.306176875430145e-05, "loss": 1.0138, "step": 80650 }, { "epoch": 13.885065381968342, "grad_norm": 15.193840980529785, "learning_rate": 4.305746730901583e-05, "loss": 1.0603, "step": 80700 }, { "epoch": 13.893668272539573, "grad_norm": 10.70821475982666, "learning_rate": 4.305316586373022e-05, "loss": 1.0224, "step": 80750 }, { "epoch": 13.902271163110806, "grad_norm": 11.16057014465332, "learning_rate": 4.30488644184446e-05, "loss": 1.0242, "step": 80800 }, { "epoch": 13.910874053682036, "grad_norm": 8.83073616027832, "learning_rate": 4.3044562973158986e-05, "loss": 1.0002, "step": 80850 }, { "epoch": 13.919476944253269, "grad_norm": 11.45268726348877, "learning_rate": 4.304026152787337e-05, "loss": 1.0208, "step": 80900 }, { "epoch": 13.928079834824501, "grad_norm": 9.961888313293457, "learning_rate": 4.303596008258775e-05, "loss": 1.0085, "step": 80950 }, { "epoch": 13.936682725395732, "grad_norm": 11.002728462219238, "learning_rate": 4.3031658637302136e-05, "loss": 1.0801, "step": 81000 }, { "epoch": 13.945285615966965, "grad_norm": 9.819936752319336, "learning_rate": 4.302735719201652e-05, "loss": 1.0011, "step": 81050 }, { "epoch": 13.953888506538197, "grad_norm": 9.240921020507812, "learning_rate": 4.3023055746730906e-05, "loss": 0.9693, "step": 81100 }, { "epoch": 13.962491397109428, "grad_norm": 7.576094627380371, "learning_rate": 4.301875430144529e-05, "loss": 0.9745, "step": 81150 }, { "epoch": 13.97109428768066, "grad_norm": 7.457500457763672, "learning_rate": 4.301445285615967e-05, "loss": 1.0446, "step": 81200 }, { "epoch": 13.979697178251893, "grad_norm": 8.425722122192383, "learning_rate": 4.3010151410874056e-05, "loss": 1.05, "step": 81250 }, { "epoch": 13.988300068823124, "grad_norm": 15.076637268066406, "learning_rate": 4.3005849965588444e-05, "loss": 0.9892, "step": 81300 }, { "epoch": 13.996902959394356, "grad_norm": 12.628260612487793, "learning_rate": 4.3001548520302825e-05, "loss": 1.0104, "step": 81350 }, { "epoch": 14.0, "eval_accuracy": 0.5684907933230081, "eval_f1": 0.5674110860176721, "eval_f1_DuraRiadoRio_16x16": 0.5143546441495778, "eval_f1_Mole_16x16": 0.4875896304467733, "eval_f1_Quebrado_16x16": 0.7590571802706242, "eval_f1_RiadoRio_16x16": 0.511541072640869, "eval_f1_RioFechado_16x16": 0.5645129025805161, "eval_loss": 1.261860966682434, "eval_precision": 0.5930428740726468, "eval_precision_DuraRiadoRio_16x16": 0.579033134166214, "eval_precision_Mole_16x16": 0.6686838124054463, "eval_precision_Quebrado_16x16": 0.7633889376646181, "eval_precision_RiadoRio_16x16": 0.43554913294797687, "eval_precision_RioFechado_16x16": 0.5185593531789783, "eval_recall": 0.5680372125120015, "eval_recall_DuraRiadoRio_16x16": 0.4626736111111111, "eval_recall_Mole_16x16": 0.3836805555555556, "eval_recall_Quebrado_16x16": 0.7547743055555556, "eval_recall_RiadoRio_16x16": 0.6196546052631579, "eval_recall_RioFechado_16x16": 0.6194029850746269, "eval_runtime": 46.2386, "eval_samples_per_second": 251.349, "eval_steps_per_second": 15.723, "step": 81368 }, { "epoch": 14.005505849965589, "grad_norm": 10.575414657592773, "learning_rate": 4.2997247075017206e-05, "loss": 0.9844, "step": 81400 }, { "epoch": 14.01410874053682, "grad_norm": 7.010415554046631, "learning_rate": 4.299294562973159e-05, "loss": 0.9619, "step": 81450 }, { "epoch": 14.022711631108052, "grad_norm": 6.362885475158691, "learning_rate": 4.2988644184445975e-05, "loss": 1.0389, "step": 81500 }, { "epoch": 14.031314521679285, "grad_norm": 13.085395812988281, "learning_rate": 4.298434273916036e-05, "loss": 1.0387, "step": 81550 }, { "epoch": 14.039917412250515, "grad_norm": 10.877254486083984, "learning_rate": 4.2980041293874744e-05, "loss": 0.936, "step": 81600 }, { "epoch": 14.048520302821748, "grad_norm": 10.31653118133545, "learning_rate": 4.297573984858913e-05, "loss": 0.9588, "step": 81650 }, { "epoch": 14.05712319339298, "grad_norm": 7.2779741287231445, "learning_rate": 4.297143840330351e-05, "loss": 0.9807, "step": 81700 }, { "epoch": 14.065726083964211, "grad_norm": 9.304015159606934, "learning_rate": 4.2967136958017895e-05, "loss": 0.984, "step": 81750 }, { "epoch": 14.074328974535444, "grad_norm": 10.372323989868164, "learning_rate": 4.296283551273228e-05, "loss": 0.995, "step": 81800 }, { "epoch": 14.082931865106676, "grad_norm": 10.134417533874512, "learning_rate": 4.2958534067446664e-05, "loss": 1.0015, "step": 81850 }, { "epoch": 14.091534755677907, "grad_norm": 7.266290664672852, "learning_rate": 4.295423262216105e-05, "loss": 0.9724, "step": 81900 }, { "epoch": 14.10013764624914, "grad_norm": 9.774820327758789, "learning_rate": 4.294993117687543e-05, "loss": 1.0459, "step": 81950 }, { "epoch": 14.108740536820372, "grad_norm": 8.744791030883789, "learning_rate": 4.2945629731589814e-05, "loss": 0.9722, "step": 82000 }, { "epoch": 14.117343427391603, "grad_norm": 7.8503499031066895, "learning_rate": 4.29413282863042e-05, "loss": 0.9701, "step": 82050 }, { "epoch": 14.125946317962836, "grad_norm": 13.402193069458008, "learning_rate": 4.293702684101858e-05, "loss": 1.0224, "step": 82100 }, { "epoch": 14.134549208534068, "grad_norm": 7.7425103187561035, "learning_rate": 4.293272539573297e-05, "loss": 1.0015, "step": 82150 }, { "epoch": 14.143152099105299, "grad_norm": 11.520112991333008, "learning_rate": 4.292842395044735e-05, "loss": 0.9719, "step": 82200 }, { "epoch": 14.151754989676531, "grad_norm": 12.74526309967041, "learning_rate": 4.2924122505161734e-05, "loss": 1.0335, "step": 82250 }, { "epoch": 14.160357880247764, "grad_norm": 10.922868728637695, "learning_rate": 4.291982105987612e-05, "loss": 1.0906, "step": 82300 }, { "epoch": 14.168960770818995, "grad_norm": 11.226542472839355, "learning_rate": 4.29155196145905e-05, "loss": 0.9903, "step": 82350 }, { "epoch": 14.177563661390227, "grad_norm": 8.544866561889648, "learning_rate": 4.291121816930489e-05, "loss": 0.9556, "step": 82400 }, { "epoch": 14.18616655196146, "grad_norm": 7.240637302398682, "learning_rate": 4.290691672401927e-05, "loss": 1.0359, "step": 82450 }, { "epoch": 14.19476944253269, "grad_norm": 9.151276588439941, "learning_rate": 4.290261527873365e-05, "loss": 1.0001, "step": 82500 }, { "epoch": 14.203372333103923, "grad_norm": 7.717922687530518, "learning_rate": 4.289831383344804e-05, "loss": 0.9396, "step": 82550 }, { "epoch": 14.211975223675156, "grad_norm": 9.99289608001709, "learning_rate": 4.289401238816242e-05, "loss": 0.9631, "step": 82600 }, { "epoch": 14.220578114246386, "grad_norm": 9.733555793762207, "learning_rate": 4.288971094287681e-05, "loss": 1.0191, "step": 82650 }, { "epoch": 14.229181004817619, "grad_norm": 7.240375995635986, "learning_rate": 4.28854094975912e-05, "loss": 0.9984, "step": 82700 }, { "epoch": 14.237783895388851, "grad_norm": 7.699287414550781, "learning_rate": 4.288110805230557e-05, "loss": 0.9626, "step": 82750 }, { "epoch": 14.246386785960082, "grad_norm": 8.033370971679688, "learning_rate": 4.287680660701996e-05, "loss": 0.9709, "step": 82800 }, { "epoch": 14.254989676531315, "grad_norm": 12.082639694213867, "learning_rate": 4.287250516173434e-05, "loss": 1.0123, "step": 82850 }, { "epoch": 14.263592567102547, "grad_norm": 7.6383137702941895, "learning_rate": 4.286820371644873e-05, "loss": 1.0168, "step": 82900 }, { "epoch": 14.272195457673778, "grad_norm": 7.115789890289307, "learning_rate": 4.286390227116312e-05, "loss": 0.9858, "step": 82950 }, { "epoch": 14.28079834824501, "grad_norm": 8.69091510772705, "learning_rate": 4.28596008258775e-05, "loss": 1.0631, "step": 83000 }, { "epoch": 14.289401238816243, "grad_norm": 8.747679710388184, "learning_rate": 4.285529938059188e-05, "loss": 0.9523, "step": 83050 }, { "epoch": 14.298004129387474, "grad_norm": 9.751111030578613, "learning_rate": 4.285099793530626e-05, "loss": 1.011, "step": 83100 }, { "epoch": 14.306607019958706, "grad_norm": 10.335315704345703, "learning_rate": 4.284669649002065e-05, "loss": 1.0544, "step": 83150 }, { "epoch": 14.315209910529939, "grad_norm": 9.37348747253418, "learning_rate": 4.284239504473504e-05, "loss": 0.9859, "step": 83200 }, { "epoch": 14.32381280110117, "grad_norm": 8.211024284362793, "learning_rate": 4.283809359944942e-05, "loss": 0.9441, "step": 83250 }, { "epoch": 14.332415691672402, "grad_norm": 8.4849853515625, "learning_rate": 4.28337921541638e-05, "loss": 0.982, "step": 83300 }, { "epoch": 14.341018582243635, "grad_norm": 13.986501693725586, "learning_rate": 4.282949070887819e-05, "loss": 1.0108, "step": 83350 }, { "epoch": 14.349621472814865, "grad_norm": 7.808427810668945, "learning_rate": 4.282518926359257e-05, "loss": 0.971, "step": 83400 }, { "epoch": 14.358224363386098, "grad_norm": 8.503271102905273, "learning_rate": 4.2820887818306956e-05, "loss": 0.9778, "step": 83450 }, { "epoch": 14.36682725395733, "grad_norm": 11.719829559326172, "learning_rate": 4.281658637302134e-05, "loss": 1.0097, "step": 83500 }, { "epoch": 14.375430144528561, "grad_norm": 9.728038787841797, "learning_rate": 4.281228492773572e-05, "loss": 1.0388, "step": 83550 }, { "epoch": 14.384033035099794, "grad_norm": 10.552639961242676, "learning_rate": 4.2807983482450106e-05, "loss": 0.9877, "step": 83600 }, { "epoch": 14.392635925671026, "grad_norm": 17.068811416625977, "learning_rate": 4.280368203716449e-05, "loss": 1.0192, "step": 83650 }, { "epoch": 14.401238816242257, "grad_norm": 9.111903190612793, "learning_rate": 4.2799380591878876e-05, "loss": 0.9943, "step": 83700 }, { "epoch": 14.40984170681349, "grad_norm": 14.352967262268066, "learning_rate": 4.279507914659326e-05, "loss": 0.9734, "step": 83750 }, { "epoch": 14.41844459738472, "grad_norm": 12.943285942077637, "learning_rate": 4.2790777701307645e-05, "loss": 0.9333, "step": 83800 }, { "epoch": 14.427047487955953, "grad_norm": 7.155236721038818, "learning_rate": 4.2786476256022026e-05, "loss": 0.9904, "step": 83850 }, { "epoch": 14.435650378527185, "grad_norm": 13.505380630493164, "learning_rate": 4.278217481073641e-05, "loss": 0.9819, "step": 83900 }, { "epoch": 14.444253269098416, "grad_norm": 8.422693252563477, "learning_rate": 4.2777873365450795e-05, "loss": 0.9952, "step": 83950 }, { "epoch": 14.452856159669649, "grad_norm": 7.974184036254883, "learning_rate": 4.2773571920165176e-05, "loss": 0.9974, "step": 84000 }, { "epoch": 14.461459050240881, "grad_norm": 8.051703453063965, "learning_rate": 4.2769270474879564e-05, "loss": 0.995, "step": 84050 }, { "epoch": 14.470061940812112, "grad_norm": 13.322359085083008, "learning_rate": 4.2764969029593945e-05, "loss": 0.9725, "step": 84100 }, { "epoch": 14.478664831383345, "grad_norm": 7.427457809448242, "learning_rate": 4.2760667584308326e-05, "loss": 0.9724, "step": 84150 }, { "epoch": 14.487267721954577, "grad_norm": 7.390655994415283, "learning_rate": 4.2756366139022714e-05, "loss": 1.0191, "step": 84200 }, { "epoch": 14.495870612525808, "grad_norm": 4.5698113441467285, "learning_rate": 4.2752064693737096e-05, "loss": 0.9626, "step": 84250 }, { "epoch": 14.50447350309704, "grad_norm": 10.471458435058594, "learning_rate": 4.2747763248451484e-05, "loss": 1.0179, "step": 84300 }, { "epoch": 14.513076393668273, "grad_norm": 17.320138931274414, "learning_rate": 4.274346180316587e-05, "loss": 0.9776, "step": 84350 }, { "epoch": 14.521679284239504, "grad_norm": 6.85123348236084, "learning_rate": 4.2739160357880246e-05, "loss": 0.9613, "step": 84400 }, { "epoch": 14.530282174810736, "grad_norm": 11.598538398742676, "learning_rate": 4.2734858912594634e-05, "loss": 0.9891, "step": 84450 }, { "epoch": 14.538885065381969, "grad_norm": 11.440105438232422, "learning_rate": 4.2730557467309015e-05, "loss": 0.9606, "step": 84500 }, { "epoch": 14.5474879559532, "grad_norm": 7.069523811340332, "learning_rate": 4.27262560220234e-05, "loss": 1.0068, "step": 84550 }, { "epoch": 14.556090846524432, "grad_norm": 10.174443244934082, "learning_rate": 4.272195457673779e-05, "loss": 1.0106, "step": 84600 }, { "epoch": 14.564693737095665, "grad_norm": 6.262446880340576, "learning_rate": 4.2717653131452165e-05, "loss": 1.0265, "step": 84650 }, { "epoch": 14.573296627666895, "grad_norm": 8.590171813964844, "learning_rate": 4.271335168616655e-05, "loss": 1.0253, "step": 84700 }, { "epoch": 14.581899518238128, "grad_norm": 10.548550605773926, "learning_rate": 4.270905024088094e-05, "loss": 1.0362, "step": 84750 }, { "epoch": 14.59050240880936, "grad_norm": 8.34200668334961, "learning_rate": 4.270474879559532e-05, "loss": 0.9979, "step": 84800 }, { "epoch": 14.599105299380591, "grad_norm": 6.8865485191345215, "learning_rate": 4.270044735030971e-05, "loss": 0.9946, "step": 84850 }, { "epoch": 14.607708189951824, "grad_norm": 10.718092918395996, "learning_rate": 4.2696145905024085e-05, "loss": 1.0051, "step": 84900 }, { "epoch": 14.616311080523056, "grad_norm": 7.440936088562012, "learning_rate": 4.269184445973847e-05, "loss": 0.9296, "step": 84950 }, { "epoch": 14.624913971094287, "grad_norm": 11.764447212219238, "learning_rate": 4.268754301445286e-05, "loss": 0.9596, "step": 85000 }, { "epoch": 14.63351686166552, "grad_norm": 7.805148601531982, "learning_rate": 4.268324156916724e-05, "loss": 1.0973, "step": 85050 }, { "epoch": 14.642119752236752, "grad_norm": 8.961553573608398, "learning_rate": 4.267894012388163e-05, "loss": 0.974, "step": 85100 }, { "epoch": 14.650722642807983, "grad_norm": 10.80073070526123, "learning_rate": 4.267463867859601e-05, "loss": 1.0169, "step": 85150 }, { "epoch": 14.659325533379215, "grad_norm": 10.268904685974121, "learning_rate": 4.267033723331039e-05, "loss": 0.9784, "step": 85200 }, { "epoch": 14.667928423950448, "grad_norm": 11.254559516906738, "learning_rate": 4.266603578802478e-05, "loss": 0.9847, "step": 85250 }, { "epoch": 14.676531314521679, "grad_norm": 7.985254287719727, "learning_rate": 4.266173434273916e-05, "loss": 0.9884, "step": 85300 }, { "epoch": 14.685134205092911, "grad_norm": 8.982976913452148, "learning_rate": 4.265743289745355e-05, "loss": 0.9826, "step": 85350 }, { "epoch": 14.693737095664144, "grad_norm": 8.3786039352417, "learning_rate": 4.265313145216793e-05, "loss": 1.0031, "step": 85400 }, { "epoch": 14.702339986235375, "grad_norm": 11.399736404418945, "learning_rate": 4.264883000688231e-05, "loss": 1.0042, "step": 85450 }, { "epoch": 14.710942876806607, "grad_norm": 6.932462215423584, "learning_rate": 4.26445285615967e-05, "loss": 0.9458, "step": 85500 }, { "epoch": 14.71954576737784, "grad_norm": 11.832301139831543, "learning_rate": 4.264022711631108e-05, "loss": 1.0012, "step": 85550 }, { "epoch": 14.72814865794907, "grad_norm": 10.194280624389648, "learning_rate": 4.263592567102547e-05, "loss": 1.0448, "step": 85600 }, { "epoch": 14.736751548520303, "grad_norm": 9.332804679870605, "learning_rate": 4.263162422573985e-05, "loss": 1.0166, "step": 85650 }, { "epoch": 14.745354439091535, "grad_norm": 10.729357719421387, "learning_rate": 4.262732278045424e-05, "loss": 0.9914, "step": 85700 }, { "epoch": 14.753957329662766, "grad_norm": 7.809874057769775, "learning_rate": 4.262302133516862e-05, "loss": 0.985, "step": 85750 }, { "epoch": 14.762560220233999, "grad_norm": 8.669493675231934, "learning_rate": 4.2618719889883e-05, "loss": 0.9851, "step": 85800 }, { "epoch": 14.771163110805231, "grad_norm": 8.16219711303711, "learning_rate": 4.261441844459739e-05, "loss": 0.9847, "step": 85850 }, { "epoch": 14.779766001376462, "grad_norm": 6.197021484375, "learning_rate": 4.261011699931177e-05, "loss": 0.942, "step": 85900 }, { "epoch": 14.788368891947695, "grad_norm": 8.971457481384277, "learning_rate": 4.260581555402616e-05, "loss": 1.0159, "step": 85950 }, { "epoch": 14.796971782518927, "grad_norm": 12.717209815979004, "learning_rate": 4.260151410874054e-05, "loss": 1.0236, "step": 86000 }, { "epoch": 14.805574673090158, "grad_norm": 8.72926139831543, "learning_rate": 4.259721266345492e-05, "loss": 1.0465, "step": 86050 }, { "epoch": 14.81417756366139, "grad_norm": 8.444025039672852, "learning_rate": 4.259291121816931e-05, "loss": 0.9785, "step": 86100 }, { "epoch": 14.822780454232623, "grad_norm": 10.549696922302246, "learning_rate": 4.258860977288369e-05, "loss": 0.9669, "step": 86150 }, { "epoch": 14.831383344803854, "grad_norm": 14.799835205078125, "learning_rate": 4.2584308327598077e-05, "loss": 1.0188, "step": 86200 }, { "epoch": 14.839986235375086, "grad_norm": 7.419358253479004, "learning_rate": 4.258000688231246e-05, "loss": 0.9933, "step": 86250 }, { "epoch": 14.848589125946319, "grad_norm": 12.537158012390137, "learning_rate": 4.257570543702684e-05, "loss": 0.9915, "step": 86300 }, { "epoch": 14.85719201651755, "grad_norm": 9.664939880371094, "learning_rate": 4.257140399174123e-05, "loss": 1.025, "step": 86350 }, { "epoch": 14.865794907088782, "grad_norm": 6.897673606872559, "learning_rate": 4.2567102546455615e-05, "loss": 1.0039, "step": 86400 }, { "epoch": 14.874397797660015, "grad_norm": 9.900800704956055, "learning_rate": 4.2562801101169996e-05, "loss": 0.9453, "step": 86450 }, { "epoch": 14.883000688231245, "grad_norm": 11.763041496276855, "learning_rate": 4.2558499655884384e-05, "loss": 1.0168, "step": 86500 }, { "epoch": 14.891603578802478, "grad_norm": 8.483213424682617, "learning_rate": 4.255419821059876e-05, "loss": 1.0475, "step": 86550 }, { "epoch": 14.90020646937371, "grad_norm": 7.536127090454102, "learning_rate": 4.2549896765313146e-05, "loss": 1.01, "step": 86600 }, { "epoch": 14.908809359944941, "grad_norm": 9.826441764831543, "learning_rate": 4.2545595320027534e-05, "loss": 1.0098, "step": 86650 }, { "epoch": 14.917412250516174, "grad_norm": 15.838622093200684, "learning_rate": 4.2541293874741915e-05, "loss": 0.9888, "step": 86700 }, { "epoch": 14.926015141087404, "grad_norm": 17.51608657836914, "learning_rate": 4.25369924294563e-05, "loss": 0.9695, "step": 86750 }, { "epoch": 14.934618031658637, "grad_norm": 9.797567367553711, "learning_rate": 4.2532690984170684e-05, "loss": 1.0295, "step": 86800 }, { "epoch": 14.94322092222987, "grad_norm": 8.801209449768066, "learning_rate": 4.2528389538885066e-05, "loss": 0.9614, "step": 86850 }, { "epoch": 14.951823812801102, "grad_norm": 10.259273529052734, "learning_rate": 4.2524088093599454e-05, "loss": 0.9778, "step": 86900 }, { "epoch": 14.960426703372333, "grad_norm": 8.482404708862305, "learning_rate": 4.2519786648313835e-05, "loss": 1.0207, "step": 86950 }, { "epoch": 14.969029593943565, "grad_norm": 7.829961776733398, "learning_rate": 4.251548520302822e-05, "loss": 1.014, "step": 87000 }, { "epoch": 14.977632484514796, "grad_norm": 11.436854362487793, "learning_rate": 4.2511183757742604e-05, "loss": 1.002, "step": 87050 }, { "epoch": 14.986235375086029, "grad_norm": 8.461100578308105, "learning_rate": 4.2506882312456985e-05, "loss": 1.017, "step": 87100 }, { "epoch": 14.994838265657261, "grad_norm": 8.799827575683594, "learning_rate": 4.250258086717137e-05, "loss": 1.0061, "step": 87150 }, { "epoch": 15.0, "eval_accuracy": 0.5458612975391499, "eval_f1": 0.5398741393784909, "eval_f1_DuraRiadoRio_16x16": 0.5575902789662844, "eval_f1_Mole_16x16": 0.3367416496250852, "eval_f1_Quebrado_16x16": 0.7050209205020921, "eval_f1_RiadoRio_16x16": 0.5122962962962962, "eval_f1_RioFechado_16x16": 0.5877215515026971, "eval_loss": 1.2752740383148193, "eval_precision": 0.6480089203625283, "eval_precision_DuraRiadoRio_16x16": 0.4602995196383159, "eval_precision_Mole_16x16": 0.7841269841269841, "eval_precision_Quebrado_16x16": 0.8868421052631579, "eval_precision_RiadoRio_16x16": 0.4004168596572487, "eval_precision_RioFechado_16x16": 0.708359133126935, "eval_recall": 0.5439285648961076, "eval_recall_DuraRiadoRio_16x16": 0.70703125, "eval_recall_Mole_16x16": 0.2144097222222222, "eval_recall_Quebrado_16x16": 0.5850694444444444, "eval_recall_RiadoRio_16x16": 0.7109375, "eval_recall_RioFechado_16x16": 0.5021949078138718, "eval_runtime": 45.9089, "eval_samples_per_second": 253.154, "eval_steps_per_second": 15.836, "step": 87180 }, { "epoch": 15.003441156228492, "grad_norm": 9.547394752502441, "learning_rate": 4.2498279421885754e-05, "loss": 1.0418, "step": 87200 }, { "epoch": 15.012044046799724, "grad_norm": 11.446144104003906, "learning_rate": 4.249397797660014e-05, "loss": 1.0021, "step": 87250 }, { "epoch": 15.020646937370957, "grad_norm": 10.1669282913208, "learning_rate": 4.248967653131452e-05, "loss": 1.0035, "step": 87300 }, { "epoch": 15.029249827942188, "grad_norm": 9.137524604797363, "learning_rate": 4.2485375086028904e-05, "loss": 1.0387, "step": 87350 }, { "epoch": 15.03785271851342, "grad_norm": 9.900424003601074, "learning_rate": 4.248107364074329e-05, "loss": 0.985, "step": 87400 }, { "epoch": 15.046455609084653, "grad_norm": 9.160223960876465, "learning_rate": 4.2476772195457674e-05, "loss": 0.9735, "step": 87450 }, { "epoch": 15.055058499655884, "grad_norm": 9.453089714050293, "learning_rate": 4.247247075017206e-05, "loss": 1.0407, "step": 87500 }, { "epoch": 15.063661390227116, "grad_norm": 7.347286701202393, "learning_rate": 4.246816930488644e-05, "loss": 1.0307, "step": 87550 }, { "epoch": 15.072264280798349, "grad_norm": 13.160804748535156, "learning_rate": 4.2463867859600824e-05, "loss": 1.0105, "step": 87600 }, { "epoch": 15.08086717136958, "grad_norm": 9.098710060119629, "learning_rate": 4.245956641431521e-05, "loss": 0.9578, "step": 87650 }, { "epoch": 15.089470061940812, "grad_norm": 9.719996452331543, "learning_rate": 4.245526496902959e-05, "loss": 0.9941, "step": 87700 }, { "epoch": 15.098072952512045, "grad_norm": 12.104400634765625, "learning_rate": 4.245096352374398e-05, "loss": 1.0174, "step": 87750 }, { "epoch": 15.106675843083275, "grad_norm": 11.1522216796875, "learning_rate": 4.244666207845837e-05, "loss": 0.9504, "step": 87800 }, { "epoch": 15.115278733654508, "grad_norm": 14.281304359436035, "learning_rate": 4.244236063317275e-05, "loss": 0.9482, "step": 87850 }, { "epoch": 15.12388162422574, "grad_norm": 6.7487382888793945, "learning_rate": 4.243805918788713e-05, "loss": 0.9433, "step": 87900 }, { "epoch": 15.132484514796971, "grad_norm": 10.230938911437988, "learning_rate": 4.243375774260151e-05, "loss": 0.9694, "step": 87950 }, { "epoch": 15.141087405368204, "grad_norm": 11.832991600036621, "learning_rate": 4.24294562973159e-05, "loss": 0.9822, "step": 88000 }, { "epoch": 15.149690295939436, "grad_norm": 10.423210144042969, "learning_rate": 4.242515485203029e-05, "loss": 1.0157, "step": 88050 }, { "epoch": 15.158293186510667, "grad_norm": 13.682433128356934, "learning_rate": 4.242085340674467e-05, "loss": 0.9945, "step": 88100 }, { "epoch": 15.1668960770819, "grad_norm": 10.292101860046387, "learning_rate": 4.241655196145905e-05, "loss": 0.9705, "step": 88150 }, { "epoch": 15.175498967653132, "grad_norm": 6.685235023498535, "learning_rate": 4.241225051617343e-05, "loss": 0.9696, "step": 88200 }, { "epoch": 15.184101858224363, "grad_norm": 8.127517700195312, "learning_rate": 4.240794907088782e-05, "loss": 0.9469, "step": 88250 }, { "epoch": 15.192704748795595, "grad_norm": 8.420137405395508, "learning_rate": 4.240364762560221e-05, "loss": 0.9728, "step": 88300 }, { "epoch": 15.201307639366828, "grad_norm": 7.616537094116211, "learning_rate": 4.239934618031659e-05, "loss": 0.9602, "step": 88350 }, { "epoch": 15.209910529938059, "grad_norm": 10.074312210083008, "learning_rate": 4.239504473503097e-05, "loss": 1.0016, "step": 88400 }, { "epoch": 15.218513420509291, "grad_norm": 8.921157836914062, "learning_rate": 4.239074328974536e-05, "loss": 0.9695, "step": 88450 }, { "epoch": 15.227116311080524, "grad_norm": 11.627534866333008, "learning_rate": 4.238644184445974e-05, "loss": 0.9969, "step": 88500 }, { "epoch": 15.235719201651754, "grad_norm": 9.452815055847168, "learning_rate": 4.238214039917413e-05, "loss": 0.9288, "step": 88550 }, { "epoch": 15.244322092222987, "grad_norm": 14.070107460021973, "learning_rate": 4.237783895388851e-05, "loss": 0.9421, "step": 88600 }, { "epoch": 15.25292498279422, "grad_norm": 12.112996101379395, "learning_rate": 4.2373537508602896e-05, "loss": 0.9463, "step": 88650 }, { "epoch": 15.26152787336545, "grad_norm": 10.906972885131836, "learning_rate": 4.236923606331728e-05, "loss": 0.9981, "step": 88700 }, { "epoch": 15.270130763936683, "grad_norm": 8.165678977966309, "learning_rate": 4.236493461803166e-05, "loss": 0.9954, "step": 88750 }, { "epoch": 15.278733654507915, "grad_norm": 7.315967559814453, "learning_rate": 4.2360633172746047e-05, "loss": 1.0135, "step": 88800 }, { "epoch": 15.287336545079146, "grad_norm": 14.247698783874512, "learning_rate": 4.235633172746043e-05, "loss": 0.9577, "step": 88850 }, { "epoch": 15.295939435650379, "grad_norm": 10.015584945678711, "learning_rate": 4.2352030282174816e-05, "loss": 1.0198, "step": 88900 }, { "epoch": 15.304542326221611, "grad_norm": 8.069112777709961, "learning_rate": 4.23477288368892e-05, "loss": 1.02, "step": 88950 }, { "epoch": 15.313145216792842, "grad_norm": 9.667657852172852, "learning_rate": 4.234342739160358e-05, "loss": 1.0043, "step": 89000 }, { "epoch": 15.321748107364074, "grad_norm": 9.18915843963623, "learning_rate": 4.2339125946317966e-05, "loss": 1.0126, "step": 89050 }, { "epoch": 15.330350997935307, "grad_norm": 11.360067367553711, "learning_rate": 4.233482450103235e-05, "loss": 0.9905, "step": 89100 }, { "epoch": 15.338953888506538, "grad_norm": 15.03067398071289, "learning_rate": 4.2330523055746735e-05, "loss": 0.9857, "step": 89150 }, { "epoch": 15.34755677907777, "grad_norm": 8.80513858795166, "learning_rate": 4.2326221610461116e-05, "loss": 1.0059, "step": 89200 }, { "epoch": 15.356159669649003, "grad_norm": 8.210318565368652, "learning_rate": 4.23219201651755e-05, "loss": 0.994, "step": 89250 }, { "epoch": 15.364762560220234, "grad_norm": 7.53951358795166, "learning_rate": 4.2317618719889885e-05, "loss": 0.9705, "step": 89300 }, { "epoch": 15.373365450791466, "grad_norm": 8.793084144592285, "learning_rate": 4.2313317274604267e-05, "loss": 1.005, "step": 89350 }, { "epoch": 15.381968341362699, "grad_norm": 9.579442024230957, "learning_rate": 4.2309015829318654e-05, "loss": 0.9354, "step": 89400 }, { "epoch": 15.39057123193393, "grad_norm": 8.68235969543457, "learning_rate": 4.230471438403304e-05, "loss": 0.977, "step": 89450 }, { "epoch": 15.399174122505162, "grad_norm": 10.671452522277832, "learning_rate": 4.230041293874742e-05, "loss": 0.9971, "step": 89500 }, { "epoch": 15.407777013076394, "grad_norm": 10.293514251708984, "learning_rate": 4.2296111493461805e-05, "loss": 0.9944, "step": 89550 }, { "epoch": 15.416379903647625, "grad_norm": 8.35737133026123, "learning_rate": 4.2291810048176186e-05, "loss": 0.9432, "step": 89600 }, { "epoch": 15.424982794218858, "grad_norm": 13.794173240661621, "learning_rate": 4.2287508602890574e-05, "loss": 1.0278, "step": 89650 }, { "epoch": 15.43358568479009, "grad_norm": 9.694953918457031, "learning_rate": 4.228320715760496e-05, "loss": 1.0418, "step": 89700 }, { "epoch": 15.442188575361321, "grad_norm": 10.165706634521484, "learning_rate": 4.2278905712319336e-05, "loss": 0.9975, "step": 89750 }, { "epoch": 15.450791465932554, "grad_norm": 8.211334228515625, "learning_rate": 4.2274604267033724e-05, "loss": 0.9506, "step": 89800 }, { "epoch": 15.459394356503786, "grad_norm": 8.025797843933105, "learning_rate": 4.227030282174811e-05, "loss": 0.9895, "step": 89850 }, { "epoch": 15.467997247075017, "grad_norm": 13.314275741577148, "learning_rate": 4.226600137646249e-05, "loss": 0.9425, "step": 89900 }, { "epoch": 15.47660013764625, "grad_norm": 9.877361297607422, "learning_rate": 4.226169993117688e-05, "loss": 0.995, "step": 89950 }, { "epoch": 15.48520302821748, "grad_norm": 12.20814037322998, "learning_rate": 4.225739848589126e-05, "loss": 1.0106, "step": 90000 }, { "epoch": 15.493805918788713, "grad_norm": 8.772645950317383, "learning_rate": 4.2253097040605644e-05, "loss": 1.0276, "step": 90050 }, { "epoch": 15.502408809359945, "grad_norm": 10.193516731262207, "learning_rate": 4.224879559532003e-05, "loss": 0.9693, "step": 90100 }, { "epoch": 15.511011699931178, "grad_norm": 13.247690200805664, "learning_rate": 4.224449415003441e-05, "loss": 0.9666, "step": 90150 }, { "epoch": 15.519614590502409, "grad_norm": 9.44120979309082, "learning_rate": 4.22401927047488e-05, "loss": 1.0168, "step": 90200 }, { "epoch": 15.528217481073641, "grad_norm": 14.252159118652344, "learning_rate": 4.223589125946318e-05, "loss": 1.016, "step": 90250 }, { "epoch": 15.536820371644872, "grad_norm": 12.765522956848145, "learning_rate": 4.223158981417756e-05, "loss": 0.9688, "step": 90300 }, { "epoch": 15.545423262216104, "grad_norm": 9.388051986694336, "learning_rate": 4.222728836889195e-05, "loss": 1.0001, "step": 90350 }, { "epoch": 15.554026152787337, "grad_norm": 8.08633041381836, "learning_rate": 4.222298692360633e-05, "loss": 0.9707, "step": 90400 }, { "epoch": 15.562629043358568, "grad_norm": 11.677124977111816, "learning_rate": 4.221868547832072e-05, "loss": 0.967, "step": 90450 }, { "epoch": 15.5712319339298, "grad_norm": 8.775369644165039, "learning_rate": 4.22143840330351e-05, "loss": 0.9658, "step": 90500 }, { "epoch": 15.579834824501033, "grad_norm": 8.545756340026855, "learning_rate": 4.221008258774948e-05, "loss": 0.9724, "step": 90550 }, { "epoch": 15.588437715072264, "grad_norm": 8.198704719543457, "learning_rate": 4.220578114246387e-05, "loss": 0.9837, "step": 90600 }, { "epoch": 15.597040605643496, "grad_norm": 7.527681350708008, "learning_rate": 4.220147969717825e-05, "loss": 0.9794, "step": 90650 }, { "epoch": 15.605643496214729, "grad_norm": 12.296113014221191, "learning_rate": 4.219717825189264e-05, "loss": 1.0016, "step": 90700 }, { "epoch": 15.61424638678596, "grad_norm": 9.177253723144531, "learning_rate": 4.219287680660702e-05, "loss": 1.036, "step": 90750 }, { "epoch": 15.622849277357192, "grad_norm": 8.908042907714844, "learning_rate": 4.218857536132141e-05, "loss": 1.0369, "step": 90800 }, { "epoch": 15.631452167928424, "grad_norm": 12.879129409790039, "learning_rate": 4.218427391603579e-05, "loss": 1.0207, "step": 90850 }, { "epoch": 15.640055058499655, "grad_norm": 7.104382038116455, "learning_rate": 4.217997247075017e-05, "loss": 1.0357, "step": 90900 }, { "epoch": 15.648657949070888, "grad_norm": 7.173849105834961, "learning_rate": 4.217567102546456e-05, "loss": 0.9796, "step": 90950 }, { "epoch": 15.65726083964212, "grad_norm": 7.623749256134033, "learning_rate": 4.217136958017894e-05, "loss": 0.9869, "step": 91000 }, { "epoch": 15.665863730213351, "grad_norm": 9.341753005981445, "learning_rate": 4.216706813489333e-05, "loss": 0.9705, "step": 91050 }, { "epoch": 15.674466620784584, "grad_norm": 9.821338653564453, "learning_rate": 4.216276668960771e-05, "loss": 0.9694, "step": 91100 }, { "epoch": 15.683069511355816, "grad_norm": 9.033586502075195, "learning_rate": 4.215846524432209e-05, "loss": 0.9776, "step": 91150 }, { "epoch": 15.691672401927047, "grad_norm": 13.098190307617188, "learning_rate": 4.215416379903648e-05, "loss": 1.004, "step": 91200 }, { "epoch": 15.70027529249828, "grad_norm": 10.38226318359375, "learning_rate": 4.2149862353750866e-05, "loss": 0.9973, "step": 91250 }, { "epoch": 15.708878183069512, "grad_norm": 9.659680366516113, "learning_rate": 4.214556090846525e-05, "loss": 1.0001, "step": 91300 }, { "epoch": 15.717481073640743, "grad_norm": 8.388848304748535, "learning_rate": 4.2141259463179635e-05, "loss": 0.9482, "step": 91350 }, { "epoch": 15.726083964211975, "grad_norm": 9.265814781188965, "learning_rate": 4.213695801789401e-05, "loss": 0.9277, "step": 91400 }, { "epoch": 15.734686854783208, "grad_norm": 8.860733985900879, "learning_rate": 4.21326565726084e-05, "loss": 0.9807, "step": 91450 }, { "epoch": 15.743289745354438, "grad_norm": 6.265392303466797, "learning_rate": 4.2128355127322786e-05, "loss": 0.9911, "step": 91500 }, { "epoch": 15.751892635925671, "grad_norm": 8.9376802444458, "learning_rate": 4.212405368203717e-05, "loss": 1.0238, "step": 91550 }, { "epoch": 15.760495526496904, "grad_norm": 9.14009952545166, "learning_rate": 4.2119752236751555e-05, "loss": 1.0208, "step": 91600 }, { "epoch": 15.769098417068134, "grad_norm": 8.099481582641602, "learning_rate": 4.211545079146593e-05, "loss": 0.9888, "step": 91650 }, { "epoch": 15.777701307639367, "grad_norm": 9.825207710266113, "learning_rate": 4.211114934618032e-05, "loss": 0.9667, "step": 91700 }, { "epoch": 15.7863041982106, "grad_norm": 7.752381324768066, "learning_rate": 4.2106847900894705e-05, "loss": 0.9725, "step": 91750 }, { "epoch": 15.79490708878183, "grad_norm": 13.358495712280273, "learning_rate": 4.2102546455609086e-05, "loss": 0.9938, "step": 91800 }, { "epoch": 15.803509979353063, "grad_norm": 6.156054973602295, "learning_rate": 4.2098245010323474e-05, "loss": 0.9966, "step": 91850 }, { "epoch": 15.812112869924295, "grad_norm": 7.9332146644592285, "learning_rate": 4.2093943565037855e-05, "loss": 1.0131, "step": 91900 }, { "epoch": 15.820715760495526, "grad_norm": 7.491793155670166, "learning_rate": 4.2089642119752237e-05, "loss": 1.0081, "step": 91950 }, { "epoch": 15.829318651066759, "grad_norm": 10.106185913085938, "learning_rate": 4.2085340674466625e-05, "loss": 1.0134, "step": 92000 }, { "epoch": 15.837921541637991, "grad_norm": 8.89893627166748, "learning_rate": 4.2081039229181006e-05, "loss": 0.9454, "step": 92050 }, { "epoch": 15.846524432209222, "grad_norm": 10.680603981018066, "learning_rate": 4.2076737783895394e-05, "loss": 1.0127, "step": 92100 }, { "epoch": 15.855127322780454, "grad_norm": 10.654461860656738, "learning_rate": 4.2072436338609775e-05, "loss": 0.9329, "step": 92150 }, { "epoch": 15.863730213351687, "grad_norm": 11.977115631103516, "learning_rate": 4.2068134893324156e-05, "loss": 0.9763, "step": 92200 }, { "epoch": 15.872333103922918, "grad_norm": 12.695249557495117, "learning_rate": 4.2063833448038544e-05, "loss": 0.9471, "step": 92250 }, { "epoch": 15.88093599449415, "grad_norm": 8.20704174041748, "learning_rate": 4.2059532002752925e-05, "loss": 0.9999, "step": 92300 }, { "epoch": 15.889538885065383, "grad_norm": 9.531217575073242, "learning_rate": 4.205523055746731e-05, "loss": 1.0188, "step": 92350 }, { "epoch": 15.898141775636613, "grad_norm": 8.231558799743652, "learning_rate": 4.2050929112181694e-05, "loss": 1.007, "step": 92400 }, { "epoch": 15.906744666207846, "grad_norm": 7.937769889831543, "learning_rate": 4.2046627666896075e-05, "loss": 0.9905, "step": 92450 }, { "epoch": 15.915347556779079, "grad_norm": 12.92549991607666, "learning_rate": 4.204232622161046e-05, "loss": 0.9775, "step": 92500 }, { "epoch": 15.92395044735031, "grad_norm": 11.519530296325684, "learning_rate": 4.2038024776324845e-05, "loss": 1.0101, "step": 92550 }, { "epoch": 15.932553337921542, "grad_norm": 8.428977012634277, "learning_rate": 4.203372333103923e-05, "loss": 1.0082, "step": 92600 }, { "epoch": 15.941156228492774, "grad_norm": 9.095731735229492, "learning_rate": 4.202942188575362e-05, "loss": 0.9973, "step": 92650 }, { "epoch": 15.949759119064005, "grad_norm": 10.919301986694336, "learning_rate": 4.2025120440467995e-05, "loss": 1.0095, "step": 92700 }, { "epoch": 15.958362009635238, "grad_norm": 7.59476375579834, "learning_rate": 4.202081899518238e-05, "loss": 0.9964, "step": 92750 }, { "epoch": 15.96696490020647, "grad_norm": 10.800294876098633, "learning_rate": 4.2016517549896764e-05, "loss": 1.0462, "step": 92800 }, { "epoch": 15.975567790777701, "grad_norm": 8.932245254516602, "learning_rate": 4.201221610461115e-05, "loss": 1.0105, "step": 92850 }, { "epoch": 15.984170681348933, "grad_norm": 10.350913047790527, "learning_rate": 4.200791465932554e-05, "loss": 0.9895, "step": 92900 }, { "epoch": 15.992773571920164, "grad_norm": 9.21055793762207, "learning_rate": 4.200361321403992e-05, "loss": 1.0244, "step": 92950 }, { "epoch": 16.0, "eval_accuracy": 0.5975735673722251, "eval_f1": 0.5860454125224587, "eval_f1_DuraRiadoRio_16x16": 0.5360272638753651, "eval_f1_Mole_16x16": 0.6240074731433909, "eval_f1_Quebrado_16x16": 0.7052870354213924, "eval_f1_RiadoRio_16x16": 0.4540455616653574, "eval_f1_RioFechado_16x16": 0.6108597285067874, "eval_loss": 1.1748765707015991, "eval_precision": 0.607183821594019, "eval_precision_DuraRiadoRio_16x16": 0.6103104212860311, "eval_precision_Mole_16x16": 0.6754297269969667, "eval_precision_Quebrado_16x16": 0.5897286256200759, "eval_precision_RiadoRio_16x16": 0.6250901225666907, "eval_precision_RioFechado_16x16": 0.5353602115003304, "eval_recall": 0.6005085351108235, "eval_recall_DuraRiadoRio_16x16": 0.4778645833333333, "eval_recall_Mole_16x16": 0.5798611111111112, "eval_recall_Quebrado_16x16": 0.8771701388888888, "eval_recall_RiadoRio_16x16": 0.35649671052631576, "eval_recall_RioFechado_16x16": 0.7111501316944688, "eval_runtime": 46.0886, "eval_samples_per_second": 252.167, "eval_steps_per_second": 15.774, "step": 92992 }, { "epoch": 16.001376462491397, "grad_norm": 8.783780097961426, "learning_rate": 4.19993117687543e-05, "loss": 1.0407, "step": 93000 }, { "epoch": 16.00997935306263, "grad_norm": 10.254236221313477, "learning_rate": 4.199501032346868e-05, "loss": 0.991, "step": 93050 }, { "epoch": 16.018582243633862, "grad_norm": 5.874866008758545, "learning_rate": 4.199070887818307e-05, "loss": 0.9811, "step": 93100 }, { "epoch": 16.027185134205094, "grad_norm": 11.003693580627441, "learning_rate": 4.198640743289746e-05, "loss": 1.0122, "step": 93150 }, { "epoch": 16.035788024776323, "grad_norm": 9.458943367004395, "learning_rate": 4.198210598761184e-05, "loss": 0.9491, "step": 93200 }, { "epoch": 16.044390915347556, "grad_norm": 15.038995742797852, "learning_rate": 4.197780454232622e-05, "loss": 0.9609, "step": 93250 }, { "epoch": 16.05299380591879, "grad_norm": 12.425536155700684, "learning_rate": 4.197350309704061e-05, "loss": 0.9875, "step": 93300 }, { "epoch": 16.06159669649002, "grad_norm": 13.697285652160645, "learning_rate": 4.196920165175499e-05, "loss": 0.9802, "step": 93350 }, { "epoch": 16.070199587061254, "grad_norm": 9.400041580200195, "learning_rate": 4.196490020646938e-05, "loss": 1.0012, "step": 93400 }, { "epoch": 16.078802477632486, "grad_norm": 17.247560501098633, "learning_rate": 4.196059876118376e-05, "loss": 0.9843, "step": 93450 }, { "epoch": 16.087405368203715, "grad_norm": 9.365756034851074, "learning_rate": 4.195629731589815e-05, "loss": 0.9602, "step": 93500 }, { "epoch": 16.096008258774948, "grad_norm": 9.304664611816406, "learning_rate": 4.195199587061253e-05, "loss": 0.9844, "step": 93550 }, { "epoch": 16.10461114934618, "grad_norm": 9.931727409362793, "learning_rate": 4.194769442532691e-05, "loss": 0.9481, "step": 93600 }, { "epoch": 16.113214039917413, "grad_norm": 8.975923538208008, "learning_rate": 4.19433929800413e-05, "loss": 0.9468, "step": 93650 }, { "epoch": 16.121816930488645, "grad_norm": 13.428293228149414, "learning_rate": 4.193909153475568e-05, "loss": 0.9449, "step": 93700 }, { "epoch": 16.130419821059878, "grad_norm": 15.459061622619629, "learning_rate": 4.193479008947007e-05, "loss": 1.0154, "step": 93750 }, { "epoch": 16.139022711631107, "grad_norm": 11.04698371887207, "learning_rate": 4.193048864418445e-05, "loss": 0.9237, "step": 93800 }, { "epoch": 16.14762560220234, "grad_norm": 13.471234321594238, "learning_rate": 4.192618719889883e-05, "loss": 0.9652, "step": 93850 }, { "epoch": 16.15622849277357, "grad_norm": 12.473538398742676, "learning_rate": 4.192188575361322e-05, "loss": 0.9811, "step": 93900 }, { "epoch": 16.164831383344804, "grad_norm": 8.941057205200195, "learning_rate": 4.19175843083276e-05, "loss": 0.9596, "step": 93950 }, { "epoch": 16.173434273916037, "grad_norm": 9.454408645629883, "learning_rate": 4.191328286304199e-05, "loss": 0.9977, "step": 94000 }, { "epoch": 16.18203716448727, "grad_norm": 12.180450439453125, "learning_rate": 4.190898141775637e-05, "loss": 0.9869, "step": 94050 }, { "epoch": 16.1906400550585, "grad_norm": 11.057807922363281, "learning_rate": 4.190467997247075e-05, "loss": 1.0014, "step": 94100 }, { "epoch": 16.19924294562973, "grad_norm": 5.794293403625488, "learning_rate": 4.190037852718514e-05, "loss": 0.9446, "step": 94150 }, { "epoch": 16.207845836200963, "grad_norm": 11.059802055358887, "learning_rate": 4.189607708189952e-05, "loss": 0.9612, "step": 94200 }, { "epoch": 16.216448726772196, "grad_norm": 10.951423645019531, "learning_rate": 4.1891775636613906e-05, "loss": 0.9441, "step": 94250 }, { "epoch": 16.22505161734343, "grad_norm": 11.639784812927246, "learning_rate": 4.1887474191328294e-05, "loss": 1.0001, "step": 94300 }, { "epoch": 16.23365450791466, "grad_norm": 9.559476852416992, "learning_rate": 4.188317274604267e-05, "loss": 1.0166, "step": 94350 }, { "epoch": 16.24225739848589, "grad_norm": 8.525561332702637, "learning_rate": 4.1878871300757056e-05, "loss": 0.9722, "step": 94400 }, { "epoch": 16.250860289057123, "grad_norm": 11.693270683288574, "learning_rate": 4.187456985547144e-05, "loss": 1.0037, "step": 94450 }, { "epoch": 16.259463179628355, "grad_norm": 9.069439888000488, "learning_rate": 4.1870268410185825e-05, "loss": 1.0136, "step": 94500 }, { "epoch": 16.268066070199588, "grad_norm": 7.162515640258789, "learning_rate": 4.1865966964900213e-05, "loss": 0.9646, "step": 94550 }, { "epoch": 16.27666896077082, "grad_norm": 13.922733306884766, "learning_rate": 4.186166551961459e-05, "loss": 0.9718, "step": 94600 }, { "epoch": 16.285271851342053, "grad_norm": 7.861927509307861, "learning_rate": 4.1857364074328976e-05, "loss": 0.9653, "step": 94650 }, { "epoch": 16.29387474191328, "grad_norm": 6.523833751678467, "learning_rate": 4.185306262904336e-05, "loss": 0.9738, "step": 94700 }, { "epoch": 16.302477632484514, "grad_norm": 10.768519401550293, "learning_rate": 4.1848761183757745e-05, "loss": 1.003, "step": 94750 }, { "epoch": 16.311080523055747, "grad_norm": 14.191226959228516, "learning_rate": 4.184445973847213e-05, "loss": 0.9355, "step": 94800 }, { "epoch": 16.31968341362698, "grad_norm": 8.99422550201416, "learning_rate": 4.184015829318651e-05, "loss": 1.0252, "step": 94850 }, { "epoch": 16.328286304198212, "grad_norm": 9.376070022583008, "learning_rate": 4.1835856847900895e-05, "loss": 0.9639, "step": 94900 }, { "epoch": 16.33688919476944, "grad_norm": 7.48316764831543, "learning_rate": 4.183155540261528e-05, "loss": 1.0146, "step": 94950 }, { "epoch": 16.345492085340673, "grad_norm": 12.62804126739502, "learning_rate": 4.1827253957329664e-05, "loss": 0.972, "step": 95000 }, { "epoch": 16.354094975911906, "grad_norm": 10.568094253540039, "learning_rate": 4.182295251204405e-05, "loss": 0.9422, "step": 95050 }, { "epoch": 16.36269786648314, "grad_norm": 7.841088771820068, "learning_rate": 4.1818651066758433e-05, "loss": 0.9787, "step": 95100 }, { "epoch": 16.37130075705437, "grad_norm": 13.684410095214844, "learning_rate": 4.1814349621472815e-05, "loss": 1.0139, "step": 95150 }, { "epoch": 16.379903647625603, "grad_norm": 8.355493545532227, "learning_rate": 4.18100481761872e-05, "loss": 0.9865, "step": 95200 }, { "epoch": 16.388506538196832, "grad_norm": 9.26408576965332, "learning_rate": 4.1805746730901584e-05, "loss": 0.9985, "step": 95250 }, { "epoch": 16.397109428768065, "grad_norm": 10.581493377685547, "learning_rate": 4.180144528561597e-05, "loss": 0.9897, "step": 95300 }, { "epoch": 16.405712319339298, "grad_norm": 9.668686866760254, "learning_rate": 4.179714384033035e-05, "loss": 0.9865, "step": 95350 }, { "epoch": 16.41431520991053, "grad_norm": 8.222257614135742, "learning_rate": 4.1792842395044734e-05, "loss": 1.0078, "step": 95400 }, { "epoch": 16.422918100481763, "grad_norm": 20.15555191040039, "learning_rate": 4.178854094975912e-05, "loss": 0.9555, "step": 95450 }, { "epoch": 16.431520991052995, "grad_norm": 8.215752601623535, "learning_rate": 4.17842395044735e-05, "loss": 1.0193, "step": 95500 }, { "epoch": 16.440123881624224, "grad_norm": 8.874390602111816, "learning_rate": 4.177993805918789e-05, "loss": 0.9881, "step": 95550 }, { "epoch": 16.448726772195457, "grad_norm": 10.583806037902832, "learning_rate": 4.177563661390227e-05, "loss": 0.9326, "step": 95600 }, { "epoch": 16.45732966276669, "grad_norm": 8.916718482971191, "learning_rate": 4.177133516861666e-05, "loss": 0.9489, "step": 95650 }, { "epoch": 16.46593255333792, "grad_norm": 10.50352668762207, "learning_rate": 4.176703372333104e-05, "loss": 0.976, "step": 95700 }, { "epoch": 16.474535443909154, "grad_norm": 9.302934646606445, "learning_rate": 4.176273227804542e-05, "loss": 1.007, "step": 95750 }, { "epoch": 16.483138334480387, "grad_norm": 8.145527839660645, "learning_rate": 4.175843083275981e-05, "loss": 1.0106, "step": 95800 }, { "epoch": 16.491741225051616, "grad_norm": 8.233237266540527, "learning_rate": 4.175412938747419e-05, "loss": 0.9353, "step": 95850 }, { "epoch": 16.50034411562285, "grad_norm": 9.462848663330078, "learning_rate": 4.174982794218858e-05, "loss": 0.9662, "step": 95900 }, { "epoch": 16.50894700619408, "grad_norm": 7.8347883224487305, "learning_rate": 4.174552649690296e-05, "loss": 1.0686, "step": 95950 }, { "epoch": 16.517549896765313, "grad_norm": 10.611796379089355, "learning_rate": 4.174122505161734e-05, "loss": 0.9437, "step": 96000 }, { "epoch": 16.526152787336546, "grad_norm": 10.238731384277344, "learning_rate": 4.173692360633173e-05, "loss": 0.9156, "step": 96050 }, { "epoch": 16.53475567790778, "grad_norm": 8.670917510986328, "learning_rate": 4.173262216104611e-05, "loss": 0.9965, "step": 96100 }, { "epoch": 16.543358568479007, "grad_norm": 8.068706512451172, "learning_rate": 4.17283207157605e-05, "loss": 1.0128, "step": 96150 }, { "epoch": 16.55196145905024, "grad_norm": 13.267080307006836, "learning_rate": 4.172401927047488e-05, "loss": 0.9851, "step": 96200 }, { "epoch": 16.560564349621473, "grad_norm": 8.340829849243164, "learning_rate": 4.171971782518926e-05, "loss": 0.9714, "step": 96250 }, { "epoch": 16.569167240192705, "grad_norm": 15.639427185058594, "learning_rate": 4.171541637990365e-05, "loss": 0.9562, "step": 96300 }, { "epoch": 16.577770130763938, "grad_norm": 13.510056495666504, "learning_rate": 4.171111493461804e-05, "loss": 0.9643, "step": 96350 }, { "epoch": 16.58637302133517, "grad_norm": 8.290148735046387, "learning_rate": 4.170681348933242e-05, "loss": 0.9321, "step": 96400 }, { "epoch": 16.5949759119064, "grad_norm": 10.448437690734863, "learning_rate": 4.1702512044046806e-05, "loss": 0.9878, "step": 96450 }, { "epoch": 16.60357880247763, "grad_norm": 8.912208557128906, "learning_rate": 4.169821059876118e-05, "loss": 0.917, "step": 96500 }, { "epoch": 16.612181693048864, "grad_norm": 9.932550430297852, "learning_rate": 4.169390915347557e-05, "loss": 0.9413, "step": 96550 }, { "epoch": 16.620784583620097, "grad_norm": 12.270027160644531, "learning_rate": 4.168960770818996e-05, "loss": 0.973, "step": 96600 }, { "epoch": 16.62938747419133, "grad_norm": 11.67715835571289, "learning_rate": 4.168530626290434e-05, "loss": 1.0084, "step": 96650 }, { "epoch": 16.63799036476256, "grad_norm": 10.55839729309082, "learning_rate": 4.1681004817618726e-05, "loss": 0.9833, "step": 96700 }, { "epoch": 16.64659325533379, "grad_norm": 10.612383842468262, "learning_rate": 4.16767033723331e-05, "loss": 0.9696, "step": 96750 }, { "epoch": 16.655196145905023, "grad_norm": 6.986774444580078, "learning_rate": 4.167240192704749e-05, "loss": 0.9284, "step": 96800 }, { "epoch": 16.663799036476256, "grad_norm": 5.798050403594971, "learning_rate": 4.1668100481761876e-05, "loss": 1.0064, "step": 96850 }, { "epoch": 16.67240192704749, "grad_norm": 9.056644439697266, "learning_rate": 4.166379903647626e-05, "loss": 0.9482, "step": 96900 }, { "epoch": 16.68100481761872, "grad_norm": 10.94788932800293, "learning_rate": 4.1659497591190645e-05, "loss": 1.0049, "step": 96950 }, { "epoch": 16.689607708189953, "grad_norm": 9.348265647888184, "learning_rate": 4.1655196145905026e-05, "loss": 0.9334, "step": 97000 }, { "epoch": 16.698210598761182, "grad_norm": 11.30555534362793, "learning_rate": 4.165089470061941e-05, "loss": 1.0223, "step": 97050 }, { "epoch": 16.706813489332415, "grad_norm": 8.413512229919434, "learning_rate": 4.1646593255333795e-05, "loss": 1.0257, "step": 97100 }, { "epoch": 16.715416379903647, "grad_norm": 8.930301666259766, "learning_rate": 4.164229181004818e-05, "loss": 0.9438, "step": 97150 }, { "epoch": 16.72401927047488, "grad_norm": 15.362107276916504, "learning_rate": 4.1637990364762565e-05, "loss": 1.0098, "step": 97200 }, { "epoch": 16.732622161046113, "grad_norm": 15.395135879516602, "learning_rate": 4.1633688919476946e-05, "loss": 0.9475, "step": 97250 }, { "epoch": 16.741225051617345, "grad_norm": 8.0408296585083, "learning_rate": 4.162938747419133e-05, "loss": 0.9541, "step": 97300 }, { "epoch": 16.749827942188574, "grad_norm": 16.18621253967285, "learning_rate": 4.1625086028905715e-05, "loss": 0.9944, "step": 97350 }, { "epoch": 16.758430832759807, "grad_norm": 7.090983867645264, "learning_rate": 4.1620784583620096e-05, "loss": 0.9409, "step": 97400 }, { "epoch": 16.76703372333104, "grad_norm": 11.433194160461426, "learning_rate": 4.1616483138334484e-05, "loss": 1.0018, "step": 97450 }, { "epoch": 16.77563661390227, "grad_norm": 9.820618629455566, "learning_rate": 4.1612181693048865e-05, "loss": 0.9651, "step": 97500 }, { "epoch": 16.784239504473504, "grad_norm": 15.421008110046387, "learning_rate": 4.1607880247763246e-05, "loss": 0.9598, "step": 97550 }, { "epoch": 16.792842395044737, "grad_norm": 9.246977806091309, "learning_rate": 4.1603578802477634e-05, "loss": 0.9443, "step": 97600 }, { "epoch": 16.801445285615966, "grad_norm": 9.49930191040039, "learning_rate": 4.1599277357192015e-05, "loss": 1.0463, "step": 97650 }, { "epoch": 16.8100481761872, "grad_norm": 9.499979019165039, "learning_rate": 4.1594975911906403e-05, "loss": 0.951, "step": 97700 }, { "epoch": 16.81865106675843, "grad_norm": 7.754093170166016, "learning_rate": 4.159067446662079e-05, "loss": 0.9711, "step": 97750 }, { "epoch": 16.827253957329663, "grad_norm": 9.280466079711914, "learning_rate": 4.158637302133517e-05, "loss": 0.9795, "step": 97800 }, { "epoch": 16.835856847900896, "grad_norm": 10.254216194152832, "learning_rate": 4.1582071576049554e-05, "loss": 0.9794, "step": 97850 }, { "epoch": 16.844459738472125, "grad_norm": 11.548324584960938, "learning_rate": 4.1577770130763935e-05, "loss": 0.9819, "step": 97900 }, { "epoch": 16.853062629043357, "grad_norm": 9.561954498291016, "learning_rate": 4.157346868547832e-05, "loss": 0.9757, "step": 97950 }, { "epoch": 16.86166551961459, "grad_norm": 15.23478889465332, "learning_rate": 4.156916724019271e-05, "loss": 1.0123, "step": 98000 }, { "epoch": 16.870268410185822, "grad_norm": 7.57096004486084, "learning_rate": 4.156486579490709e-05, "loss": 0.9769, "step": 98050 }, { "epoch": 16.878871300757055, "grad_norm": 12.005406379699707, "learning_rate": 4.156056434962147e-05, "loss": 0.9525, "step": 98100 }, { "epoch": 16.887474191328288, "grad_norm": 7.99483060836792, "learning_rate": 4.1556262904335854e-05, "loss": 0.9675, "step": 98150 }, { "epoch": 16.89607708189952, "grad_norm": 11.299764633178711, "learning_rate": 4.155196145905024e-05, "loss": 1.0036, "step": 98200 }, { "epoch": 16.90467997247075, "grad_norm": 11.319596290588379, "learning_rate": 4.154766001376463e-05, "loss": 0.9778, "step": 98250 }, { "epoch": 16.91328286304198, "grad_norm": 9.733622550964355, "learning_rate": 4.154335856847901e-05, "loss": 0.9471, "step": 98300 }, { "epoch": 16.921885753613214, "grad_norm": 10.331831932067871, "learning_rate": 4.153905712319339e-05, "loss": 0.9966, "step": 98350 }, { "epoch": 16.930488644184447, "grad_norm": 9.76812744140625, "learning_rate": 4.153475567790778e-05, "loss": 0.9651, "step": 98400 }, { "epoch": 16.93909153475568, "grad_norm": 12.050947189331055, "learning_rate": 4.153045423262216e-05, "loss": 0.9691, "step": 98450 }, { "epoch": 16.947694425326908, "grad_norm": 7.051280975341797, "learning_rate": 4.152615278733655e-05, "loss": 0.9842, "step": 98500 }, { "epoch": 16.95629731589814, "grad_norm": 7.461218357086182, "learning_rate": 4.152185134205093e-05, "loss": 1.0127, "step": 98550 }, { "epoch": 16.964900206469373, "grad_norm": 7.9866204261779785, "learning_rate": 4.151754989676532e-05, "loss": 0.9824, "step": 98600 }, { "epoch": 16.973503097040606, "grad_norm": 8.792472839355469, "learning_rate": 4.15132484514797e-05, "loss": 0.9717, "step": 98650 }, { "epoch": 16.98210598761184, "grad_norm": 10.430839538574219, "learning_rate": 4.150894700619408e-05, "loss": 0.9515, "step": 98700 }, { "epoch": 16.99070887818307, "grad_norm": 11.460681915283203, "learning_rate": 4.150464556090847e-05, "loss": 0.972, "step": 98750 }, { "epoch": 16.9993117687543, "grad_norm": 8.176837921142578, "learning_rate": 4.150034411562285e-05, "loss": 0.9428, "step": 98800 }, { "epoch": 17.0, "eval_accuracy": 0.532266391326794, "eval_f1": 0.5211887303883123, "eval_f1_DuraRiadoRio_16x16": 0.5121056493030081, "eval_f1_Mole_16x16": 0.5043921790875602, "eval_f1_Quebrado_16x16": 0.5993069993069993, "eval_f1_RiadoRio_16x16": 0.4177734846655397, "eval_f1_RioFechado_16x16": 0.5723653395784544, "eval_loss": 1.511399745941162, "eval_precision": 0.5745838395240991, "eval_precision_DuraRiadoRio_16x16": 0.5865546218487395, "eval_precision_Mole_16x16": 0.726530612244898, "eval_precision_Quebrado_16x16": 0.44023620443901446, "eval_precision_RiadoRio_16x16": 0.5061439438267993, "eval_precision_RioFechado_16x16": 0.6134538152610441, "eval_recall": 0.5342379345853293, "eval_recall_DuraRiadoRio_16x16": 0.4544270833333333, "eval_recall_Mole_16x16": 0.3862847222222222, "eval_recall_Quebrado_16x16": 0.9383680555555556, "eval_recall_RiadoRio_16x16": 0.35567434210526316, "eval_recall_RioFechado_16x16": 0.5364354697102721, "eval_runtime": 45.9956, "eval_samples_per_second": 252.677, "eval_steps_per_second": 15.806, "step": 98804 }, { "epoch": 17.007914659325532, "grad_norm": 7.930378437042236, "learning_rate": 4.149604267033724e-05, "loss": 0.9764, "step": 98850 }, { "epoch": 17.016517549896765, "grad_norm": 11.913801193237305, "learning_rate": 4.149174122505162e-05, "loss": 1.0193, "step": 98900 }, { "epoch": 17.025120440467997, "grad_norm": 9.436736106872559, "learning_rate": 4.1487439779766e-05, "loss": 0.9314, "step": 98950 }, { "epoch": 17.03372333103923, "grad_norm": 10.397587776184082, "learning_rate": 4.148313833448039e-05, "loss": 0.9443, "step": 99000 }, { "epoch": 17.042326221610463, "grad_norm": 9.47846508026123, "learning_rate": 4.147883688919477e-05, "loss": 0.9594, "step": 99050 }, { "epoch": 17.05092911218169, "grad_norm": 11.653204917907715, "learning_rate": 4.147453544390916e-05, "loss": 0.9681, "step": 99100 }, { "epoch": 17.059532002752924, "grad_norm": 10.999855995178223, "learning_rate": 4.1470233998623546e-05, "loss": 0.9884, "step": 99150 }, { "epoch": 17.068134893324157, "grad_norm": 9.559646606445312, "learning_rate": 4.146593255333792e-05, "loss": 0.9568, "step": 99200 }, { "epoch": 17.07673778389539, "grad_norm": 14.823806762695312, "learning_rate": 4.146163110805231e-05, "loss": 0.9107, "step": 99250 }, { "epoch": 17.08534067446662, "grad_norm": 9.569289207458496, "learning_rate": 4.145732966276669e-05, "loss": 0.9308, "step": 99300 }, { "epoch": 17.093943565037854, "grad_norm": 7.441787242889404, "learning_rate": 4.145302821748108e-05, "loss": 0.9612, "step": 99350 }, { "epoch": 17.102546455609083, "grad_norm": 15.096323013305664, "learning_rate": 4.1448726772195465e-05, "loss": 1.0155, "step": 99400 }, { "epoch": 17.111149346180316, "grad_norm": 9.355422019958496, "learning_rate": 4.144442532690984e-05, "loss": 0.9485, "step": 99450 }, { "epoch": 17.11975223675155, "grad_norm": 11.885660171508789, "learning_rate": 4.144012388162423e-05, "loss": 0.9187, "step": 99500 }, { "epoch": 17.12835512732278, "grad_norm": 9.792088508605957, "learning_rate": 4.143582243633861e-05, "loss": 0.9467, "step": 99550 }, { "epoch": 17.136958017894013, "grad_norm": 9.259902000427246, "learning_rate": 4.1431520991052996e-05, "loss": 0.9431, "step": 99600 }, { "epoch": 17.145560908465246, "grad_norm": 13.870136260986328, "learning_rate": 4.1427219545767384e-05, "loss": 1.005, "step": 99650 }, { "epoch": 17.154163799036475, "grad_norm": 11.10527229309082, "learning_rate": 4.142291810048176e-05, "loss": 0.9399, "step": 99700 }, { "epoch": 17.162766689607707, "grad_norm": 8.564321517944336, "learning_rate": 4.141861665519615e-05, "loss": 0.9724, "step": 99750 }, { "epoch": 17.17136958017894, "grad_norm": 15.025402069091797, "learning_rate": 4.1414315209910535e-05, "loss": 0.8982, "step": 99800 }, { "epoch": 17.179972470750172, "grad_norm": 20.88810157775879, "learning_rate": 4.1410013764624916e-05, "loss": 0.9585, "step": 99850 }, { "epoch": 17.188575361321405, "grad_norm": 8.815108299255371, "learning_rate": 4.1405712319339304e-05, "loss": 0.9208, "step": 99900 }, { "epoch": 17.197178251892637, "grad_norm": 8.587363243103027, "learning_rate": 4.1401410874053685e-05, "loss": 1.0674, "step": 99950 }, { "epoch": 17.205781142463866, "grad_norm": 6.535338878631592, "learning_rate": 4.1397109428768066e-05, "loss": 0.9315, "step": 100000 }, { "epoch": 17.2143840330351, "grad_norm": 15.138066291809082, "learning_rate": 4.1392807983482454e-05, "loss": 0.9676, "step": 100050 }, { "epoch": 17.22298692360633, "grad_norm": 7.65643835067749, "learning_rate": 4.1388506538196835e-05, "loss": 0.947, "step": 100100 }, { "epoch": 17.231589814177564, "grad_norm": 10.675469398498535, "learning_rate": 4.138420509291122e-05, "loss": 0.9672, "step": 100150 }, { "epoch": 17.240192704748797, "grad_norm": 15.273283958435059, "learning_rate": 4.1379903647625604e-05, "loss": 0.9247, "step": 100200 }, { "epoch": 17.24879559532003, "grad_norm": 14.314194679260254, "learning_rate": 4.1375602202339986e-05, "loss": 0.9727, "step": 100250 }, { "epoch": 17.257398485891258, "grad_norm": 9.593879699707031, "learning_rate": 4.1371300757054373e-05, "loss": 1.021, "step": 100300 }, { "epoch": 17.26600137646249, "grad_norm": 9.599863052368164, "learning_rate": 4.1366999311768755e-05, "loss": 0.9905, "step": 100350 }, { "epoch": 17.274604267033723, "grad_norm": 10.370691299438477, "learning_rate": 4.136269786648314e-05, "loss": 0.942, "step": 100400 }, { "epoch": 17.283207157604956, "grad_norm": 11.906078338623047, "learning_rate": 4.1358396421197524e-05, "loss": 1.0266, "step": 100450 }, { "epoch": 17.29181004817619, "grad_norm": 10.487113952636719, "learning_rate": 4.1354094975911905e-05, "loss": 0.9254, "step": 100500 }, { "epoch": 17.30041293874742, "grad_norm": 8.12554931640625, "learning_rate": 4.134979353062629e-05, "loss": 1.0027, "step": 100550 }, { "epoch": 17.30901582931865, "grad_norm": 7.642636299133301, "learning_rate": 4.1345492085340674e-05, "loss": 1.0301, "step": 100600 }, { "epoch": 17.317618719889882, "grad_norm": 10.440439224243164, "learning_rate": 4.134119064005506e-05, "loss": 0.9369, "step": 100650 }, { "epoch": 17.326221610461115, "grad_norm": 9.187166213989258, "learning_rate": 4.133688919476944e-05, "loss": 0.9819, "step": 100700 }, { "epoch": 17.334824501032347, "grad_norm": 10.180106163024902, "learning_rate": 4.133258774948383e-05, "loss": 0.996, "step": 100750 }, { "epoch": 17.34342739160358, "grad_norm": 6.928048133850098, "learning_rate": 4.132828630419821e-05, "loss": 0.9469, "step": 100800 }, { "epoch": 17.352030282174812, "grad_norm": 8.808256149291992, "learning_rate": 4.1323984858912593e-05, "loss": 0.9823, "step": 100850 }, { "epoch": 17.36063317274604, "grad_norm": 17.26322364807129, "learning_rate": 4.131968341362698e-05, "loss": 0.9745, "step": 100900 }, { "epoch": 17.369236063317274, "grad_norm": 9.68840217590332, "learning_rate": 4.131538196834136e-05, "loss": 0.9659, "step": 100950 }, { "epoch": 17.377838953888507, "grad_norm": 16.597444534301758, "learning_rate": 4.131108052305575e-05, "loss": 1.0071, "step": 101000 }, { "epoch": 17.38644184445974, "grad_norm": 7.641477584838867, "learning_rate": 4.130677907777013e-05, "loss": 1.0058, "step": 101050 }, { "epoch": 17.39504473503097, "grad_norm": 5.73620080947876, "learning_rate": 4.130247763248451e-05, "loss": 0.9284, "step": 101100 }, { "epoch": 17.403647625602204, "grad_norm": 10.337767601013184, "learning_rate": 4.12981761871989e-05, "loss": 0.9916, "step": 101150 }, { "epoch": 17.412250516173433, "grad_norm": 10.649316787719727, "learning_rate": 4.129387474191329e-05, "loss": 0.9497, "step": 101200 }, { "epoch": 17.420853406744666, "grad_norm": 6.522881984710693, "learning_rate": 4.128957329662767e-05, "loss": 0.941, "step": 101250 }, { "epoch": 17.429456297315898, "grad_norm": 9.445099830627441, "learning_rate": 4.128527185134206e-05, "loss": 0.9778, "step": 101300 }, { "epoch": 17.43805918788713, "grad_norm": 12.122477531433105, "learning_rate": 4.128097040605643e-05, "loss": 0.9953, "step": 101350 }, { "epoch": 17.446662078458363, "grad_norm": 9.906142234802246, "learning_rate": 4.127666896077082e-05, "loss": 1.0043, "step": 101400 }, { "epoch": 17.455264969029592, "grad_norm": 6.584541320800781, "learning_rate": 4.127236751548521e-05, "loss": 0.9605, "step": 101450 }, { "epoch": 17.463867859600825, "grad_norm": 12.31583023071289, "learning_rate": 4.126806607019959e-05, "loss": 0.9852, "step": 101500 }, { "epoch": 17.472470750172057, "grad_norm": 9.394309043884277, "learning_rate": 4.126376462491398e-05, "loss": 0.9445, "step": 101550 }, { "epoch": 17.48107364074329, "grad_norm": 8.858049392700195, "learning_rate": 4.125946317962835e-05, "loss": 0.9622, "step": 101600 }, { "epoch": 17.489676531314522, "grad_norm": 7.471407890319824, "learning_rate": 4.125516173434274e-05, "loss": 1.0022, "step": 101650 }, { "epoch": 17.498279421885755, "grad_norm": 10.697000503540039, "learning_rate": 4.125086028905713e-05, "loss": 0.993, "step": 101700 }, { "epoch": 17.506882312456984, "grad_norm": 13.406034469604492, "learning_rate": 4.124655884377151e-05, "loss": 0.9482, "step": 101750 }, { "epoch": 17.515485203028216, "grad_norm": 10.611784934997559, "learning_rate": 4.12422573984859e-05, "loss": 0.9699, "step": 101800 }, { "epoch": 17.52408809359945, "grad_norm": 10.629151344299316, "learning_rate": 4.123795595320028e-05, "loss": 0.9823, "step": 101850 }, { "epoch": 17.53269098417068, "grad_norm": 6.685856342315674, "learning_rate": 4.123365450791466e-05, "loss": 1.0014, "step": 101900 }, { "epoch": 17.541293874741914, "grad_norm": 9.876399993896484, "learning_rate": 4.122935306262905e-05, "loss": 0.9773, "step": 101950 }, { "epoch": 17.549896765313147, "grad_norm": 9.165271759033203, "learning_rate": 4.122505161734343e-05, "loss": 0.9704, "step": 102000 }, { "epoch": 17.558499655884376, "grad_norm": 10.900294303894043, "learning_rate": 4.1220750172057816e-05, "loss": 0.9683, "step": 102050 }, { "epoch": 17.567102546455608, "grad_norm": 9.296416282653809, "learning_rate": 4.12164487267722e-05, "loss": 0.9519, "step": 102100 }, { "epoch": 17.57570543702684, "grad_norm": 13.194743156433105, "learning_rate": 4.121214728148658e-05, "loss": 1.0193, "step": 102150 }, { "epoch": 17.584308327598073, "grad_norm": 7.247140884399414, "learning_rate": 4.1207845836200966e-05, "loss": 0.9549, "step": 102200 }, { "epoch": 17.592911218169306, "grad_norm": 7.088600158691406, "learning_rate": 4.120354439091535e-05, "loss": 0.905, "step": 102250 }, { "epoch": 17.60151410874054, "grad_norm": 10.125738143920898, "learning_rate": 4.1199242945629736e-05, "loss": 0.9597, "step": 102300 }, { "epoch": 17.610116999311767, "grad_norm": 14.845233917236328, "learning_rate": 4.119494150034412e-05, "loss": 0.9538, "step": 102350 }, { "epoch": 17.618719889883, "grad_norm": 11.768908500671387, "learning_rate": 4.11906400550585e-05, "loss": 0.9741, "step": 102400 }, { "epoch": 17.627322780454232, "grad_norm": 8.092485427856445, "learning_rate": 4.1186338609772886e-05, "loss": 1.0051, "step": 102450 }, { "epoch": 17.635925671025465, "grad_norm": 10.519120216369629, "learning_rate": 4.118203716448727e-05, "loss": 0.9387, "step": 102500 }, { "epoch": 17.644528561596697, "grad_norm": 8.400145530700684, "learning_rate": 4.1177735719201655e-05, "loss": 0.9437, "step": 102550 }, { "epoch": 17.65313145216793, "grad_norm": 11.677557945251465, "learning_rate": 4.1173434273916036e-05, "loss": 0.9681, "step": 102600 }, { "epoch": 17.66173434273916, "grad_norm": 8.579787254333496, "learning_rate": 4.1169132828630424e-05, "loss": 0.9679, "step": 102650 }, { "epoch": 17.67033723331039, "grad_norm": 7.702165126800537, "learning_rate": 4.1164831383344805e-05, "loss": 0.9782, "step": 102700 }, { "epoch": 17.678940123881624, "grad_norm": 11.622040748596191, "learning_rate": 4.1160529938059186e-05, "loss": 1.0213, "step": 102750 }, { "epoch": 17.687543014452856, "grad_norm": 7.43345308303833, "learning_rate": 4.1156228492773574e-05, "loss": 1.0301, "step": 102800 }, { "epoch": 17.69614590502409, "grad_norm": 10.249974250793457, "learning_rate": 4.115192704748796e-05, "loss": 0.9356, "step": 102850 }, { "epoch": 17.70474879559532, "grad_norm": 7.040240287780762, "learning_rate": 4.1147625602202344e-05, "loss": 0.9269, "step": 102900 }, { "epoch": 17.71335168616655, "grad_norm": 6.719895839691162, "learning_rate": 4.1143324156916725e-05, "loss": 1.0142, "step": 102950 }, { "epoch": 17.721954576737783, "grad_norm": 11.500811576843262, "learning_rate": 4.1139022711631106e-05, "loss": 0.9975, "step": 103000 }, { "epoch": 17.730557467309016, "grad_norm": 8.29609489440918, "learning_rate": 4.1134721266345494e-05, "loss": 0.9617, "step": 103050 }, { "epoch": 17.739160357880248, "grad_norm": 7.678513526916504, "learning_rate": 4.113041982105988e-05, "loss": 0.9643, "step": 103100 }, { "epoch": 17.74776324845148, "grad_norm": 9.246646881103516, "learning_rate": 4.112611837577426e-05, "loss": 0.9874, "step": 103150 }, { "epoch": 17.756366139022713, "grad_norm": 8.197779655456543, "learning_rate": 4.1121816930488644e-05, "loss": 0.9469, "step": 103200 }, { "epoch": 17.764969029593942, "grad_norm": 10.088048934936523, "learning_rate": 4.111751548520303e-05, "loss": 0.9875, "step": 103250 }, { "epoch": 17.773571920165175, "grad_norm": 10.596574783325195, "learning_rate": 4.111321403991741e-05, "loss": 0.9933, "step": 103300 }, { "epoch": 17.782174810736407, "grad_norm": 8.741995811462402, "learning_rate": 4.11089125946318e-05, "loss": 0.9681, "step": 103350 }, { "epoch": 17.79077770130764, "grad_norm": 7.469597816467285, "learning_rate": 4.110461114934618e-05, "loss": 0.9712, "step": 103400 }, { "epoch": 17.799380591878872, "grad_norm": 10.549139976501465, "learning_rate": 4.110030970406057e-05, "loss": 1.012, "step": 103450 }, { "epoch": 17.807983482450105, "grad_norm": 6.480816841125488, "learning_rate": 4.109600825877495e-05, "loss": 1.071, "step": 103500 }, { "epoch": 17.816586373021334, "grad_norm": 9.317760467529297, "learning_rate": 4.109170681348933e-05, "loss": 0.9921, "step": 103550 }, { "epoch": 17.825189263592566, "grad_norm": 8.323650360107422, "learning_rate": 4.108740536820372e-05, "loss": 0.9915, "step": 103600 }, { "epoch": 17.8337921541638, "grad_norm": 8.624070167541504, "learning_rate": 4.10831039229181e-05, "loss": 0.9534, "step": 103650 }, { "epoch": 17.84239504473503, "grad_norm": 7.949827194213867, "learning_rate": 4.107880247763249e-05, "loss": 0.9788, "step": 103700 }, { "epoch": 17.850997935306264, "grad_norm": 9.888788223266602, "learning_rate": 4.107450103234687e-05, "loss": 0.9807, "step": 103750 }, { "epoch": 17.859600825877497, "grad_norm": 8.421555519104004, "learning_rate": 4.107019958706125e-05, "loss": 0.9941, "step": 103800 }, { "epoch": 17.868203716448726, "grad_norm": 14.440583229064941, "learning_rate": 4.106589814177564e-05, "loss": 0.9996, "step": 103850 }, { "epoch": 17.876806607019958, "grad_norm": 5.6206889152526855, "learning_rate": 4.106159669649002e-05, "loss": 0.9757, "step": 103900 }, { "epoch": 17.88540949759119, "grad_norm": 6.72746467590332, "learning_rate": 4.105729525120441e-05, "loss": 1.0157, "step": 103950 }, { "epoch": 17.894012388162423, "grad_norm": 9.616094589233398, "learning_rate": 4.105299380591879e-05, "loss": 1.0292, "step": 104000 }, { "epoch": 17.902615278733656, "grad_norm": 9.529831886291504, "learning_rate": 4.104869236063317e-05, "loss": 0.9793, "step": 104050 }, { "epoch": 17.911218169304888, "grad_norm": 7.770634651184082, "learning_rate": 4.104439091534756e-05, "loss": 0.943, "step": 104100 }, { "epoch": 17.919821059876117, "grad_norm": 8.204238891601562, "learning_rate": 4.104008947006194e-05, "loss": 0.9853, "step": 104150 }, { "epoch": 17.92842395044735, "grad_norm": 11.048352241516113, "learning_rate": 4.103578802477633e-05, "loss": 1.0038, "step": 104200 }, { "epoch": 17.937026841018582, "grad_norm": 16.757230758666992, "learning_rate": 4.1031486579490716e-05, "loss": 0.9171, "step": 104250 }, { "epoch": 17.945629731589815, "grad_norm": 6.392510414123535, "learning_rate": 4.102718513420509e-05, "loss": 0.9973, "step": 104300 }, { "epoch": 17.954232622161047, "grad_norm": 10.618914604187012, "learning_rate": 4.102288368891948e-05, "loss": 0.9933, "step": 104350 }, { "epoch": 17.962835512732276, "grad_norm": 8.927621841430664, "learning_rate": 4.101858224363386e-05, "loss": 0.985, "step": 104400 }, { "epoch": 17.97143840330351, "grad_norm": 9.383471488952637, "learning_rate": 4.101428079834825e-05, "loss": 0.9435, "step": 104450 }, { "epoch": 17.98004129387474, "grad_norm": 9.853461265563965, "learning_rate": 4.1009979353062636e-05, "loss": 0.9623, "step": 104500 }, { "epoch": 17.988644184445974, "grad_norm": 9.121235847473145, "learning_rate": 4.100567790777701e-05, "loss": 0.8962, "step": 104550 }, { "epoch": 17.997247075017206, "grad_norm": 11.45913028717041, "learning_rate": 4.10013764624914e-05, "loss": 1.01, "step": 104600 }, { "epoch": 18.0, "eval_accuracy": 0.6045431078988126, "eval_f1": 0.596150973812675, "eval_f1_DuraRiadoRio_16x16": 0.5092230296254892, "eval_f1_Mole_16x16": 0.5227537922987164, "eval_f1_Quebrado_16x16": 0.7745827468328976, "eval_f1_RiadoRio_16x16": 0.5431309904153354, "eval_f1_RioFechado_16x16": 0.6310643098909364, "eval_loss": 1.2250785827636719, "eval_precision": 0.6490550791924636, "eval_precision_DuraRiadoRio_16x16": 0.7150706436420722, "eval_precision_Mole_16x16": 0.797153024911032, "eval_precision_Quebrado_16x16": 0.7216185837392282, "eval_precision_RiadoRio_16x16": 0.4594594594594595, "eval_precision_RioFechado_16x16": 0.5519736842105263, "eval_recall": 0.6041798513559653, "eval_recall_DuraRiadoRio_16x16": 0.3953993055555556, "eval_recall_Mole_16x16": 0.3888888888888889, "eval_recall_Quebrado_16x16": 0.8359375, "eval_recall_RiadoRio_16x16": 0.6640625, "eval_recall_RioFechado_16x16": 0.7366110623353819, "eval_runtime": 46.302, "eval_samples_per_second": 251.004, "eval_steps_per_second": 15.701, "step": 104616 }, { "epoch": 18.00584996558844, "grad_norm": 12.367700576782227, "learning_rate": 4.099707501720578e-05, "loss": 0.9503, "step": 104650 }, { "epoch": 18.014452856159668, "grad_norm": 13.641201972961426, "learning_rate": 4.099277357192017e-05, "loss": 0.9777, "step": 104700 }, { "epoch": 18.0230557467309, "grad_norm": 12.543866157531738, "learning_rate": 4.0988472126634555e-05, "loss": 1.013, "step": 104750 }, { "epoch": 18.031658637302133, "grad_norm": 9.424630165100098, "learning_rate": 4.0984170681348936e-05, "loss": 0.9274, "step": 104800 }, { "epoch": 18.040261527873366, "grad_norm": 10.321890830993652, "learning_rate": 4.097986923606332e-05, "loss": 0.9359, "step": 104850 }, { "epoch": 18.048864418444598, "grad_norm": 12.115867614746094, "learning_rate": 4.0975567790777706e-05, "loss": 0.9418, "step": 104900 }, { "epoch": 18.05746730901583, "grad_norm": 7.73814058303833, "learning_rate": 4.097126634549209e-05, "loss": 0.9877, "step": 104950 }, { "epoch": 18.06607019958706, "grad_norm": 9.641854286193848, "learning_rate": 4.0966964900206475e-05, "loss": 0.9606, "step": 105000 }, { "epoch": 18.074673090158292, "grad_norm": 9.52087116241455, "learning_rate": 4.0962663454920856e-05, "loss": 0.9457, "step": 105050 }, { "epoch": 18.083275980729525, "grad_norm": 8.857209205627441, "learning_rate": 4.095836200963524e-05, "loss": 0.9224, "step": 105100 }, { "epoch": 18.091878871300757, "grad_norm": 10.362009048461914, "learning_rate": 4.0954060564349625e-05, "loss": 0.9516, "step": 105150 }, { "epoch": 18.10048176187199, "grad_norm": 8.884369850158691, "learning_rate": 4.0949759119064006e-05, "loss": 1.0034, "step": 105200 }, { "epoch": 18.109084652443222, "grad_norm": 5.8345746994018555, "learning_rate": 4.0945457673778394e-05, "loss": 0.9712, "step": 105250 }, { "epoch": 18.11768754301445, "grad_norm": 12.663044929504395, "learning_rate": 4.0941156228492775e-05, "loss": 0.9817, "step": 105300 }, { "epoch": 18.126290433585684, "grad_norm": 11.078413963317871, "learning_rate": 4.0936854783207156e-05, "loss": 0.9256, "step": 105350 }, { "epoch": 18.134893324156916, "grad_norm": 14.132830619812012, "learning_rate": 4.0932553337921544e-05, "loss": 0.9982, "step": 105400 }, { "epoch": 18.14349621472815, "grad_norm": 7.812006950378418, "learning_rate": 4.0928251892635926e-05, "loss": 0.9016, "step": 105450 }, { "epoch": 18.15209910529938, "grad_norm": 9.118341445922852, "learning_rate": 4.0923950447350314e-05, "loss": 0.9368, "step": 105500 }, { "epoch": 18.160701995870614, "grad_norm": 10.96101188659668, "learning_rate": 4.0919649002064695e-05, "loss": 0.9657, "step": 105550 }, { "epoch": 18.169304886441843, "grad_norm": 7.317468643188477, "learning_rate": 4.091534755677908e-05, "loss": 0.9554, "step": 105600 }, { "epoch": 18.177907777013075, "grad_norm": 8.561816215515137, "learning_rate": 4.0911046111493464e-05, "loss": 0.9725, "step": 105650 }, { "epoch": 18.186510667584308, "grad_norm": 8.644918441772461, "learning_rate": 4.0906744666207845e-05, "loss": 1.0021, "step": 105700 }, { "epoch": 18.19511355815554, "grad_norm": 13.09717082977295, "learning_rate": 4.090244322092223e-05, "loss": 0.9609, "step": 105750 }, { "epoch": 18.203716448726773, "grad_norm": 8.10731029510498, "learning_rate": 4.0898141775636614e-05, "loss": 0.9547, "step": 105800 }, { "epoch": 18.212319339298006, "grad_norm": 11.417423248291016, "learning_rate": 4.0893840330351e-05, "loss": 0.9526, "step": 105850 }, { "epoch": 18.220922229869235, "grad_norm": 8.616067886352539, "learning_rate": 4.088953888506538e-05, "loss": 0.9497, "step": 105900 }, { "epoch": 18.229525120440467, "grad_norm": 8.19046401977539, "learning_rate": 4.0885237439779764e-05, "loss": 0.9645, "step": 105950 }, { "epoch": 18.2381280110117, "grad_norm": 11.13180160522461, "learning_rate": 4.088093599449415e-05, "loss": 0.957, "step": 106000 }, { "epoch": 18.246730901582932, "grad_norm": 9.74319076538086, "learning_rate": 4.0876634549208534e-05, "loss": 0.9411, "step": 106050 }, { "epoch": 18.255333792154165, "grad_norm": 9.87049674987793, "learning_rate": 4.087233310392292e-05, "loss": 0.9825, "step": 106100 }, { "epoch": 18.263936682725397, "grad_norm": 11.513566017150879, "learning_rate": 4.08680316586373e-05, "loss": 0.9476, "step": 106150 }, { "epoch": 18.272539573296626, "grad_norm": 10.411094665527344, "learning_rate": 4.0863730213351684e-05, "loss": 0.9933, "step": 106200 }, { "epoch": 18.28114246386786, "grad_norm": 9.257622718811035, "learning_rate": 4.085942876806607e-05, "loss": 0.9487, "step": 106250 }, { "epoch": 18.28974535443909, "grad_norm": 10.668641090393066, "learning_rate": 4.085512732278046e-05, "loss": 0.912, "step": 106300 }, { "epoch": 18.298348245010324, "grad_norm": 10.471810340881348, "learning_rate": 4.085082587749484e-05, "loss": 0.9714, "step": 106350 }, { "epoch": 18.306951135581556, "grad_norm": 11.700697898864746, "learning_rate": 4.084652443220923e-05, "loss": 0.944, "step": 106400 }, { "epoch": 18.31555402615279, "grad_norm": 9.484747886657715, "learning_rate": 4.08422229869236e-05, "loss": 0.9508, "step": 106450 }, { "epoch": 18.324156916724018, "grad_norm": 12.634302139282227, "learning_rate": 4.083792154163799e-05, "loss": 1.0089, "step": 106500 }, { "epoch": 18.33275980729525, "grad_norm": 9.93751049041748, "learning_rate": 4.083362009635238e-05, "loss": 0.9632, "step": 106550 }, { "epoch": 18.341362697866483, "grad_norm": 11.225348472595215, "learning_rate": 4.082931865106676e-05, "loss": 0.9654, "step": 106600 }, { "epoch": 18.349965588437716, "grad_norm": 9.990254402160645, "learning_rate": 4.082501720578115e-05, "loss": 0.9386, "step": 106650 }, { "epoch": 18.358568479008948, "grad_norm": 14.136207580566406, "learning_rate": 4.082071576049552e-05, "loss": 0.9396, "step": 106700 }, { "epoch": 18.36717136958018, "grad_norm": 11.120823860168457, "learning_rate": 4.081641431520991e-05, "loss": 0.9619, "step": 106750 }, { "epoch": 18.37577426015141, "grad_norm": 6.534911632537842, "learning_rate": 4.08121128699243e-05, "loss": 0.9515, "step": 106800 }, { "epoch": 18.384377150722642, "grad_norm": 9.838722229003906, "learning_rate": 4.080781142463868e-05, "loss": 0.9845, "step": 106850 }, { "epoch": 18.392980041293875, "grad_norm": 7.890871524810791, "learning_rate": 4.080350997935307e-05, "loss": 0.9435, "step": 106900 }, { "epoch": 18.401582931865107, "grad_norm": 8.727521896362305, "learning_rate": 4.079920853406745e-05, "loss": 0.979, "step": 106950 }, { "epoch": 18.41018582243634, "grad_norm": 9.346055030822754, "learning_rate": 4.079490708878183e-05, "loss": 0.9309, "step": 107000 }, { "epoch": 18.418788713007572, "grad_norm": 9.056248664855957, "learning_rate": 4.079060564349622e-05, "loss": 0.9449, "step": 107050 }, { "epoch": 18.4273916035788, "grad_norm": 11.10054874420166, "learning_rate": 4.07863041982106e-05, "loss": 0.9666, "step": 107100 }, { "epoch": 18.435994494150034, "grad_norm": 11.012615203857422, "learning_rate": 4.078200275292499e-05, "loss": 0.9925, "step": 107150 }, { "epoch": 18.444597384721266, "grad_norm": 9.908246040344238, "learning_rate": 4.077770130763937e-05, "loss": 0.9464, "step": 107200 }, { "epoch": 18.4532002752925, "grad_norm": 8.861047744750977, "learning_rate": 4.077339986235375e-05, "loss": 0.9856, "step": 107250 }, { "epoch": 18.46180316586373, "grad_norm": 10.415766716003418, "learning_rate": 4.076909841706814e-05, "loss": 0.9986, "step": 107300 }, { "epoch": 18.47040605643496, "grad_norm": 9.139345169067383, "learning_rate": 4.076479697178252e-05, "loss": 0.9593, "step": 107350 }, { "epoch": 18.479008947006193, "grad_norm": 11.237157821655273, "learning_rate": 4.0760495526496907e-05, "loss": 0.9764, "step": 107400 }, { "epoch": 18.487611837577425, "grad_norm": 11.098730087280273, "learning_rate": 4.075619408121129e-05, "loss": 0.9498, "step": 107450 }, { "epoch": 18.496214728148658, "grad_norm": 18.053956985473633, "learning_rate": 4.075189263592567e-05, "loss": 0.9602, "step": 107500 }, { "epoch": 18.50481761871989, "grad_norm": 6.42105770111084, "learning_rate": 4.074759119064006e-05, "loss": 0.959, "step": 107550 }, { "epoch": 18.513420509291123, "grad_norm": 12.933253288269043, "learning_rate": 4.074328974535444e-05, "loss": 0.9741, "step": 107600 }, { "epoch": 18.522023399862356, "grad_norm": 9.021415710449219, "learning_rate": 4.0738988300068826e-05, "loss": 0.9512, "step": 107650 }, { "epoch": 18.530626290433585, "grad_norm": 8.425719261169434, "learning_rate": 4.0734686854783214e-05, "loss": 0.9554, "step": 107700 }, { "epoch": 18.539229181004817, "grad_norm": 8.922636032104492, "learning_rate": 4.0730385409497595e-05, "loss": 0.9296, "step": 107750 }, { "epoch": 18.54783207157605, "grad_norm": 9.793169975280762, "learning_rate": 4.0726083964211976e-05, "loss": 0.9538, "step": 107800 }, { "epoch": 18.556434962147282, "grad_norm": 7.948843002319336, "learning_rate": 4.072178251892636e-05, "loss": 0.9427, "step": 107850 }, { "epoch": 18.565037852718515, "grad_norm": 9.77756404876709, "learning_rate": 4.0717481073640745e-05, "loss": 0.9609, "step": 107900 }, { "epoch": 18.573640743289744, "grad_norm": 9.256531715393066, "learning_rate": 4.071317962835513e-05, "loss": 1.0253, "step": 107950 }, { "epoch": 18.582243633860976, "grad_norm": 9.133671760559082, "learning_rate": 4.0708878183069514e-05, "loss": 0.9542, "step": 108000 }, { "epoch": 18.59084652443221, "grad_norm": 7.838448524475098, "learning_rate": 4.0704576737783896e-05, "loss": 0.9556, "step": 108050 }, { "epoch": 18.59944941500344, "grad_norm": 8.957901954650879, "learning_rate": 4.070027529249828e-05, "loss": 0.9407, "step": 108100 }, { "epoch": 18.608052305574674, "grad_norm": 8.095242500305176, "learning_rate": 4.0695973847212665e-05, "loss": 0.9904, "step": 108150 }, { "epoch": 18.616655196145906, "grad_norm": 8.104695320129395, "learning_rate": 4.069167240192705e-05, "loss": 0.9108, "step": 108200 }, { "epoch": 18.625258086717135, "grad_norm": 7.878484725952148, "learning_rate": 4.0687370956641434e-05, "loss": 0.9876, "step": 108250 }, { "epoch": 18.633860977288368, "grad_norm": 15.336718559265137, "learning_rate": 4.068306951135582e-05, "loss": 0.9364, "step": 108300 }, { "epoch": 18.6424638678596, "grad_norm": 6.222857475280762, "learning_rate": 4.06787680660702e-05, "loss": 0.9664, "step": 108350 }, { "epoch": 18.651066758430833, "grad_norm": 12.131458282470703, "learning_rate": 4.0674466620784584e-05, "loss": 0.9632, "step": 108400 }, { "epoch": 18.659669649002065, "grad_norm": 6.671494960784912, "learning_rate": 4.067016517549897e-05, "loss": 0.9628, "step": 108450 }, { "epoch": 18.668272539573298, "grad_norm": 8.267560005187988, "learning_rate": 4.066586373021335e-05, "loss": 1.005, "step": 108500 }, { "epoch": 18.676875430144527, "grad_norm": 8.944509506225586, "learning_rate": 4.066156228492774e-05, "loss": 0.9432, "step": 108550 }, { "epoch": 18.68547832071576, "grad_norm": 10.57104206085205, "learning_rate": 4.065726083964212e-05, "loss": 0.9688, "step": 108600 }, { "epoch": 18.694081211286992, "grad_norm": 9.94381332397461, "learning_rate": 4.0652959394356504e-05, "loss": 0.9943, "step": 108650 }, { "epoch": 18.702684101858225, "grad_norm": 7.103782653808594, "learning_rate": 4.064865794907089e-05, "loss": 0.9902, "step": 108700 }, { "epoch": 18.711286992429457, "grad_norm": 9.103118896484375, "learning_rate": 4.064435650378527e-05, "loss": 0.968, "step": 108750 }, { "epoch": 18.71988988300069, "grad_norm": 10.780988693237305, "learning_rate": 4.064005505849966e-05, "loss": 0.9571, "step": 108800 }, { "epoch": 18.72849277357192, "grad_norm": 7.1277923583984375, "learning_rate": 4.063575361321404e-05, "loss": 0.9712, "step": 108850 }, { "epoch": 18.73709566414315, "grad_norm": 5.709153652191162, "learning_rate": 4.063145216792842e-05, "loss": 0.9691, "step": 108900 }, { "epoch": 18.745698554714384, "grad_norm": 10.382944107055664, "learning_rate": 4.062715072264281e-05, "loss": 0.9397, "step": 108950 }, { "epoch": 18.754301445285616, "grad_norm": 9.375112533569336, "learning_rate": 4.062284927735719e-05, "loss": 0.9986, "step": 109000 }, { "epoch": 18.76290433585685, "grad_norm": 10.814908981323242, "learning_rate": 4.061854783207158e-05, "loss": 0.9622, "step": 109050 }, { "epoch": 18.77150722642808, "grad_norm": 14.301888465881348, "learning_rate": 4.061424638678596e-05, "loss": 1.0265, "step": 109100 }, { "epoch": 18.78011011699931, "grad_norm": 8.239998817443848, "learning_rate": 4.060994494150034e-05, "loss": 1.0035, "step": 109150 }, { "epoch": 18.788713007570543, "grad_norm": 7.821311950683594, "learning_rate": 4.060564349621473e-05, "loss": 0.9181, "step": 109200 }, { "epoch": 18.797315898141775, "grad_norm": 9.28254222869873, "learning_rate": 4.060134205092911e-05, "loss": 0.9739, "step": 109250 }, { "epoch": 18.805918788713008, "grad_norm": 8.892488479614258, "learning_rate": 4.05970406056435e-05, "loss": 0.9877, "step": 109300 }, { "epoch": 18.81452167928424, "grad_norm": 10.836039543151855, "learning_rate": 4.059273916035789e-05, "loss": 0.9719, "step": 109350 }, { "epoch": 18.823124569855473, "grad_norm": 9.839016914367676, "learning_rate": 4.058843771507226e-05, "loss": 0.9927, "step": 109400 }, { "epoch": 18.831727460426702, "grad_norm": 7.303333282470703, "learning_rate": 4.058413626978665e-05, "loss": 0.9843, "step": 109450 }, { "epoch": 18.840330350997935, "grad_norm": 9.0081205368042, "learning_rate": 4.057983482450103e-05, "loss": 1.0003, "step": 109500 }, { "epoch": 18.848933241569167, "grad_norm": 6.485849857330322, "learning_rate": 4.057553337921542e-05, "loss": 0.9798, "step": 109550 }, { "epoch": 18.8575361321404, "grad_norm": 9.814563751220703, "learning_rate": 4.057123193392981e-05, "loss": 0.9164, "step": 109600 }, { "epoch": 18.866139022711632, "grad_norm": 10.269583702087402, "learning_rate": 4.056693048864418e-05, "loss": 0.9828, "step": 109650 }, { "epoch": 18.874741913282865, "grad_norm": 8.53067398071289, "learning_rate": 4.056262904335857e-05, "loss": 0.9809, "step": 109700 }, { "epoch": 18.883344803854094, "grad_norm": 8.719897270202637, "learning_rate": 4.055832759807296e-05, "loss": 0.9646, "step": 109750 }, { "epoch": 18.891947694425326, "grad_norm": 6.194845199584961, "learning_rate": 4.055402615278734e-05, "loss": 0.9617, "step": 109800 }, { "epoch": 18.90055058499656, "grad_norm": 7.987786769866943, "learning_rate": 4.0549724707501726e-05, "loss": 0.8983, "step": 109850 }, { "epoch": 18.90915347556779, "grad_norm": 11.944184303283691, "learning_rate": 4.054542326221611e-05, "loss": 0.9497, "step": 109900 }, { "epoch": 18.917756366139024, "grad_norm": 11.536967277526855, "learning_rate": 4.054112181693049e-05, "loss": 0.9497, "step": 109950 }, { "epoch": 18.926359256710256, "grad_norm": 8.380833625793457, "learning_rate": 4.0536820371644877e-05, "loss": 0.9348, "step": 110000 }, { "epoch": 18.934962147281485, "grad_norm": 9.257452011108398, "learning_rate": 4.053251892635926e-05, "loss": 0.9522, "step": 110050 }, { "epoch": 18.943565037852718, "grad_norm": 10.278019905090332, "learning_rate": 4.0528217481073646e-05, "loss": 0.9891, "step": 110100 }, { "epoch": 18.95216792842395, "grad_norm": 9.952164649963379, "learning_rate": 4.052391603578803e-05, "loss": 0.9955, "step": 110150 }, { "epoch": 18.960770818995183, "grad_norm": 7.8329949378967285, "learning_rate": 4.051961459050241e-05, "loss": 0.9518, "step": 110200 }, { "epoch": 18.969373709566415, "grad_norm": 7.4886555671691895, "learning_rate": 4.0515313145216796e-05, "loss": 0.9658, "step": 110250 }, { "epoch": 18.977976600137644, "grad_norm": 11.031951904296875, "learning_rate": 4.051101169993118e-05, "loss": 0.8933, "step": 110300 }, { "epoch": 18.986579490708877, "grad_norm": 8.641465187072754, "learning_rate": 4.0506710254645565e-05, "loss": 0.9387, "step": 110350 }, { "epoch": 18.99518238128011, "grad_norm": 10.682602882385254, "learning_rate": 4.0502408809359946e-05, "loss": 0.9718, "step": 110400 }, { "epoch": 19.0, "eval_accuracy": 0.5229736706246774, "eval_f1": 0.49064137761786386, "eval_f1_DuraRiadoRio_16x16": 0.5466825586015097, "eval_f1_Mole_16x16": 0.342698577870274, "eval_f1_Quebrado_16x16": 0.6999825874978234, "eval_f1_RiadoRio_16x16": 0.33031161473087817, "eval_f1_RioFechado_16x16": 0.5335315493888338, "eval_loss": 1.5228605270385742, "eval_precision": 0.5800726722424746, "eval_precision_DuraRiadoRio_16x16": 0.504029304029304, "eval_precision_Mole_16x16": 0.853195164075993, "eval_precision_Quebrado_16x16": 0.5844722302995057, "eval_precision_RiadoRio_16x16": 0.5309653916211293, "eval_precision_RioFechado_16x16": 0.4277012711864407, "eval_recall": 0.5265406792790434, "eval_recall_DuraRiadoRio_16x16": 0.5972222222222222, "eval_recall_Mole_16x16": 0.2144097222222222, "eval_recall_Quebrado_16x16": 0.8723958333333334, "eval_recall_RiadoRio_16x16": 0.23972039473684212, "eval_recall_RioFechado_16x16": 0.7089552238805971, "eval_runtime": 46.2551, "eval_samples_per_second": 251.259, "eval_steps_per_second": 15.717, "step": 110428 }, { "epoch": 19.003785271851342, "grad_norm": 12.068224906921387, "learning_rate": 4.0498107364074334e-05, "loss": 0.9683, "step": 110450 }, { "epoch": 19.012388162422575, "grad_norm": 8.659614562988281, "learning_rate": 4.0493805918788715e-05, "loss": 0.9694, "step": 110500 }, { "epoch": 19.020991052993807, "grad_norm": 14.0364351272583, "learning_rate": 4.0489504473503097e-05, "loss": 0.9596, "step": 110550 }, { "epoch": 19.029593943565036, "grad_norm": 10.346569061279297, "learning_rate": 4.0485203028217484e-05, "loss": 0.9743, "step": 110600 }, { "epoch": 19.03819683413627, "grad_norm": 8.261329650878906, "learning_rate": 4.0480901582931866e-05, "loss": 1.0115, "step": 110650 }, { "epoch": 19.0467997247075, "grad_norm": 8.60399055480957, "learning_rate": 4.0476600137646254e-05, "loss": 1.0154, "step": 110700 }, { "epoch": 19.055402615278734, "grad_norm": 12.168950080871582, "learning_rate": 4.0472298692360635e-05, "loss": 0.9187, "step": 110750 }, { "epoch": 19.064005505849966, "grad_norm": 6.351713180541992, "learning_rate": 4.0467997247075016e-05, "loss": 0.9074, "step": 110800 }, { "epoch": 19.0726083964212, "grad_norm": 12.15097713470459, "learning_rate": 4.0463695801789404e-05, "loss": 0.9929, "step": 110850 }, { "epoch": 19.081211286992428, "grad_norm": 7.743757247924805, "learning_rate": 4.0459394356503785e-05, "loss": 0.9259, "step": 110900 }, { "epoch": 19.08981417756366, "grad_norm": 12.910379409790039, "learning_rate": 4.045509291121817e-05, "loss": 0.9564, "step": 110950 }, { "epoch": 19.098417068134893, "grad_norm": 9.44820785522461, "learning_rate": 4.0450791465932554e-05, "loss": 0.9231, "step": 111000 }, { "epoch": 19.107019958706125, "grad_norm": 12.301440238952637, "learning_rate": 4.0446490020646935e-05, "loss": 0.9527, "step": 111050 }, { "epoch": 19.115622849277358, "grad_norm": 14.057518005371094, "learning_rate": 4.044218857536132e-05, "loss": 0.9154, "step": 111100 }, { "epoch": 19.12422573984859, "grad_norm": 8.647504806518555, "learning_rate": 4.0437887130075704e-05, "loss": 1.0044, "step": 111150 }, { "epoch": 19.13282863041982, "grad_norm": 9.663602828979492, "learning_rate": 4.043358568479009e-05, "loss": 0.9289, "step": 111200 }, { "epoch": 19.141431520991052, "grad_norm": 12.08297061920166, "learning_rate": 4.042928423950448e-05, "loss": 0.976, "step": 111250 }, { "epoch": 19.150034411562284, "grad_norm": 11.757014274597168, "learning_rate": 4.0424982794218855e-05, "loss": 0.9577, "step": 111300 }, { "epoch": 19.158637302133517, "grad_norm": 8.765030860900879, "learning_rate": 4.042068134893324e-05, "loss": 0.9782, "step": 111350 }, { "epoch": 19.16724019270475, "grad_norm": 8.03034496307373, "learning_rate": 4.041637990364763e-05, "loss": 0.9368, "step": 111400 }, { "epoch": 19.175843083275982, "grad_norm": 8.312366485595703, "learning_rate": 4.041207845836201e-05, "loss": 0.9387, "step": 111450 }, { "epoch": 19.18444597384721, "grad_norm": 15.305963516235352, "learning_rate": 4.04077770130764e-05, "loss": 0.9223, "step": 111500 }, { "epoch": 19.193048864418444, "grad_norm": 11.098335266113281, "learning_rate": 4.0403475567790774e-05, "loss": 0.9537, "step": 111550 }, { "epoch": 19.201651754989676, "grad_norm": 7.413483619689941, "learning_rate": 4.039917412250516e-05, "loss": 0.9877, "step": 111600 }, { "epoch": 19.21025464556091, "grad_norm": 10.356688499450684, "learning_rate": 4.039487267721955e-05, "loss": 0.9854, "step": 111650 }, { "epoch": 19.21885753613214, "grad_norm": 14.177543640136719, "learning_rate": 4.039057123193393e-05, "loss": 0.9177, "step": 111700 }, { "epoch": 19.227460426703374, "grad_norm": 12.351696968078613, "learning_rate": 4.038626978664832e-05, "loss": 0.9467, "step": 111750 }, { "epoch": 19.236063317274603, "grad_norm": 10.412821769714355, "learning_rate": 4.03819683413627e-05, "loss": 0.9398, "step": 111800 }, { "epoch": 19.244666207845835, "grad_norm": 7.852023124694824, "learning_rate": 4.037766689607708e-05, "loss": 0.9514, "step": 111850 }, { "epoch": 19.253269098417068, "grad_norm": 8.996772766113281, "learning_rate": 4.037336545079147e-05, "loss": 0.9157, "step": 111900 }, { "epoch": 19.2618719889883, "grad_norm": 7.786957263946533, "learning_rate": 4.036906400550585e-05, "loss": 0.932, "step": 111950 }, { "epoch": 19.270474879559533, "grad_norm": 7.921579837799072, "learning_rate": 4.036476256022024e-05, "loss": 0.955, "step": 112000 }, { "epoch": 19.279077770130765, "grad_norm": 11.77133560180664, "learning_rate": 4.036046111493462e-05, "loss": 0.9623, "step": 112050 }, { "epoch": 19.287680660701994, "grad_norm": 10.57996654510498, "learning_rate": 4.0356159669649e-05, "loss": 0.9165, "step": 112100 }, { "epoch": 19.296283551273227, "grad_norm": 16.695514678955078, "learning_rate": 4.035185822436339e-05, "loss": 0.9877, "step": 112150 }, { "epoch": 19.30488644184446, "grad_norm": 8.64188003540039, "learning_rate": 4.034755677907777e-05, "loss": 0.9749, "step": 112200 }, { "epoch": 19.313489332415692, "grad_norm": 15.355863571166992, "learning_rate": 4.034325533379216e-05, "loss": 0.9114, "step": 112250 }, { "epoch": 19.322092222986925, "grad_norm": 11.471567153930664, "learning_rate": 4.033895388850654e-05, "loss": 0.9156, "step": 112300 }, { "epoch": 19.330695113558157, "grad_norm": 7.943390846252441, "learning_rate": 4.033465244322092e-05, "loss": 0.9734, "step": 112350 }, { "epoch": 19.339298004129386, "grad_norm": 11.120708465576172, "learning_rate": 4.033035099793531e-05, "loss": 0.9626, "step": 112400 }, { "epoch": 19.34790089470062, "grad_norm": 11.899245262145996, "learning_rate": 4.032604955264969e-05, "loss": 1.0042, "step": 112450 }, { "epoch": 19.35650378527185, "grad_norm": 15.016937255859375, "learning_rate": 4.032174810736408e-05, "loss": 0.9306, "step": 112500 }, { "epoch": 19.365106675843084, "grad_norm": 8.887939453125, "learning_rate": 4.031744666207846e-05, "loss": 0.8994, "step": 112550 }, { "epoch": 19.373709566414316, "grad_norm": 11.643440246582031, "learning_rate": 4.0313145216792847e-05, "loss": 1.0142, "step": 112600 }, { "epoch": 19.38231245698555, "grad_norm": 11.27357006072998, "learning_rate": 4.030884377150723e-05, "loss": 0.9228, "step": 112650 }, { "epoch": 19.390915347556778, "grad_norm": 7.445782661437988, "learning_rate": 4.030454232622161e-05, "loss": 0.9558, "step": 112700 }, { "epoch": 19.39951823812801, "grad_norm": 6.985152244567871, "learning_rate": 4.0300240880936e-05, "loss": 0.9826, "step": 112750 }, { "epoch": 19.408121128699243, "grad_norm": 6.793111801147461, "learning_rate": 4.0295939435650385e-05, "loss": 1.0143, "step": 112800 }, { "epoch": 19.416724019270475, "grad_norm": 10.690536499023438, "learning_rate": 4.0291637990364766e-05, "loss": 0.9349, "step": 112850 }, { "epoch": 19.425326909841708, "grad_norm": 13.86011791229248, "learning_rate": 4.028733654507915e-05, "loss": 1.0493, "step": 112900 }, { "epoch": 19.43392980041294, "grad_norm": 6.022976875305176, "learning_rate": 4.028303509979353e-05, "loss": 0.9102, "step": 112950 }, { "epoch": 19.44253269098417, "grad_norm": 9.672783851623535, "learning_rate": 4.0278733654507916e-05, "loss": 0.9764, "step": 113000 }, { "epoch": 19.451135581555402, "grad_norm": 7.88127326965332, "learning_rate": 4.0274432209222304e-05, "loss": 0.9639, "step": 113050 }, { "epoch": 19.459738472126634, "grad_norm": 7.388211250305176, "learning_rate": 4.0270130763936685e-05, "loss": 0.967, "step": 113100 }, { "epoch": 19.468341362697867, "grad_norm": 16.663427352905273, "learning_rate": 4.0265829318651067e-05, "loss": 0.9406, "step": 113150 }, { "epoch": 19.4769442532691, "grad_norm": 7.168600082397461, "learning_rate": 4.026152787336545e-05, "loss": 0.9929, "step": 113200 }, { "epoch": 19.485547143840332, "grad_norm": 8.632010459899902, "learning_rate": 4.0257226428079836e-05, "loss": 0.9672, "step": 113250 }, { "epoch": 19.49415003441156, "grad_norm": 9.71005916595459, "learning_rate": 4.0252924982794224e-05, "loss": 0.9714, "step": 113300 }, { "epoch": 19.502752924982794, "grad_norm": 10.824752807617188, "learning_rate": 4.0248623537508605e-05, "loss": 1.0076, "step": 113350 }, { "epoch": 19.511355815554026, "grad_norm": 12.411925315856934, "learning_rate": 4.024432209222299e-05, "loss": 0.9255, "step": 113400 }, { "epoch": 19.51995870612526, "grad_norm": 11.123435020446777, "learning_rate": 4.0240020646937374e-05, "loss": 0.9189, "step": 113450 }, { "epoch": 19.52856159669649, "grad_norm": 10.420899391174316, "learning_rate": 4.0235719201651755e-05, "loss": 0.9279, "step": 113500 }, { "epoch": 19.537164487267724, "grad_norm": 9.807923316955566, "learning_rate": 4.023141775636614e-05, "loss": 0.9059, "step": 113550 }, { "epoch": 19.545767377838953, "grad_norm": 10.88342571258545, "learning_rate": 4.0227116311080524e-05, "loss": 0.9882, "step": 113600 }, { "epoch": 19.554370268410185, "grad_norm": 9.138763427734375, "learning_rate": 4.022281486579491e-05, "loss": 0.9758, "step": 113650 }, { "epoch": 19.562973158981418, "grad_norm": 9.724268913269043, "learning_rate": 4.021851342050929e-05, "loss": 0.9116, "step": 113700 }, { "epoch": 19.57157604955265, "grad_norm": 9.160836219787598, "learning_rate": 4.0214211975223675e-05, "loss": 0.9012, "step": 113750 }, { "epoch": 19.580178940123883, "grad_norm": 8.83164119720459, "learning_rate": 4.020991052993806e-05, "loss": 0.997, "step": 113800 }, { "epoch": 19.588781830695112, "grad_norm": 10.049444198608398, "learning_rate": 4.0205609084652444e-05, "loss": 0.9301, "step": 113850 }, { "epoch": 19.597384721266344, "grad_norm": 9.867188453674316, "learning_rate": 4.020130763936683e-05, "loss": 0.9917, "step": 113900 }, { "epoch": 19.605987611837577, "grad_norm": 10.073433876037598, "learning_rate": 4.019700619408121e-05, "loss": 0.9391, "step": 113950 }, { "epoch": 19.61459050240881, "grad_norm": 7.288029193878174, "learning_rate": 4.0192704748795594e-05, "loss": 0.9674, "step": 114000 }, { "epoch": 19.623193392980042, "grad_norm": 8.892648696899414, "learning_rate": 4.018840330350998e-05, "loss": 0.9846, "step": 114050 }, { "epoch": 19.631796283551274, "grad_norm": 7.45117712020874, "learning_rate": 4.018410185822436e-05, "loss": 0.9947, "step": 114100 }, { "epoch": 19.640399174122503, "grad_norm": 8.92116928100586, "learning_rate": 4.017980041293875e-05, "loss": 0.9331, "step": 114150 }, { "epoch": 19.649002064693736, "grad_norm": 17.30290985107422, "learning_rate": 4.017549896765314e-05, "loss": 0.9704, "step": 114200 }, { "epoch": 19.65760495526497, "grad_norm": 12.260952949523926, "learning_rate": 4.017119752236751e-05, "loss": 1.022, "step": 114250 }, { "epoch": 19.6662078458362, "grad_norm": 9.202252388000488, "learning_rate": 4.01668960770819e-05, "loss": 0.9383, "step": 114300 }, { "epoch": 19.674810736407434, "grad_norm": 7.65976095199585, "learning_rate": 4.016259463179628e-05, "loss": 0.9169, "step": 114350 }, { "epoch": 19.683413626978666, "grad_norm": 8.008522987365723, "learning_rate": 4.015829318651067e-05, "loss": 0.9497, "step": 114400 }, { "epoch": 19.692016517549895, "grad_norm": 9.782032012939453, "learning_rate": 4.015399174122506e-05, "loss": 0.935, "step": 114450 }, { "epoch": 19.700619408121128, "grad_norm": 8.153741836547852, "learning_rate": 4.014969029593943e-05, "loss": 0.9279, "step": 114500 }, { "epoch": 19.70922229869236, "grad_norm": 8.403447151184082, "learning_rate": 4.014538885065382e-05, "loss": 0.9276, "step": 114550 }, { "epoch": 19.717825189263593, "grad_norm": 10.842999458312988, "learning_rate": 4.01410874053682e-05, "loss": 0.9383, "step": 114600 }, { "epoch": 19.726428079834825, "grad_norm": 9.867798805236816, "learning_rate": 4.013678596008259e-05, "loss": 0.9801, "step": 114650 }, { "epoch": 19.735030970406058, "grad_norm": 9.77640438079834, "learning_rate": 4.013248451479698e-05, "loss": 0.9718, "step": 114700 }, { "epoch": 19.743633860977287, "grad_norm": 7.802949905395508, "learning_rate": 4.012818306951136e-05, "loss": 0.9273, "step": 114750 }, { "epoch": 19.75223675154852, "grad_norm": 13.641395568847656, "learning_rate": 4.012388162422574e-05, "loss": 0.9568, "step": 114800 }, { "epoch": 19.760839642119752, "grad_norm": 9.124244689941406, "learning_rate": 4.011958017894013e-05, "loss": 0.8992, "step": 114850 }, { "epoch": 19.769442532690984, "grad_norm": 11.872655868530273, "learning_rate": 4.011527873365451e-05, "loss": 0.9711, "step": 114900 }, { "epoch": 19.778045423262217, "grad_norm": 5.2566962242126465, "learning_rate": 4.01109772883689e-05, "loss": 0.9137, "step": 114950 }, { "epoch": 19.78664831383345, "grad_norm": 7.7365899085998535, "learning_rate": 4.010667584308328e-05, "loss": 0.9183, "step": 115000 }, { "epoch": 19.79525120440468, "grad_norm": 8.650155067443848, "learning_rate": 4.010237439779766e-05, "loss": 0.944, "step": 115050 }, { "epoch": 19.80385409497591, "grad_norm": 10.278815269470215, "learning_rate": 4.009807295251205e-05, "loss": 0.9785, "step": 115100 }, { "epoch": 19.812456985547144, "grad_norm": 16.295860290527344, "learning_rate": 4.009377150722643e-05, "loss": 0.9806, "step": 115150 }, { "epoch": 19.821059876118376, "grad_norm": 14.654317855834961, "learning_rate": 4.008947006194082e-05, "loss": 0.9638, "step": 115200 }, { "epoch": 19.82966276668961, "grad_norm": 13.133405685424805, "learning_rate": 4.00851686166552e-05, "loss": 0.9332, "step": 115250 }, { "epoch": 19.83826565726084, "grad_norm": 6.459465980529785, "learning_rate": 4.008086717136958e-05, "loss": 0.9616, "step": 115300 }, { "epoch": 19.84686854783207, "grad_norm": 13.102578163146973, "learning_rate": 4.007656572608397e-05, "loss": 0.9858, "step": 115350 }, { "epoch": 19.855471438403303, "grad_norm": 9.977508544921875, "learning_rate": 4.007226428079835e-05, "loss": 0.9622, "step": 115400 }, { "epoch": 19.864074328974535, "grad_norm": 7.547395706176758, "learning_rate": 4.0067962835512736e-05, "loss": 0.9543, "step": 115450 }, { "epoch": 19.872677219545768, "grad_norm": 11.823989868164062, "learning_rate": 4.006366139022712e-05, "loss": 0.9785, "step": 115500 }, { "epoch": 19.881280110117, "grad_norm": 9.290002822875977, "learning_rate": 4.0059359944941505e-05, "loss": 0.9632, "step": 115550 }, { "epoch": 19.889883000688233, "grad_norm": 14.280745506286621, "learning_rate": 4.0055058499655886e-05, "loss": 0.9905, "step": 115600 }, { "epoch": 19.89848589125946, "grad_norm": 11.402361869812012, "learning_rate": 4.005075705437027e-05, "loss": 0.9915, "step": 115650 }, { "epoch": 19.907088781830694, "grad_norm": 11.112833023071289, "learning_rate": 4.0046455609084655e-05, "loss": 0.945, "step": 115700 }, { "epoch": 19.915691672401927, "grad_norm": 9.060050010681152, "learning_rate": 4.004215416379904e-05, "loss": 0.912, "step": 115750 }, { "epoch": 19.92429456297316, "grad_norm": 10.933808326721191, "learning_rate": 4.0037852718513425e-05, "loss": 0.9654, "step": 115800 }, { "epoch": 19.932897453544392, "grad_norm": 11.865239143371582, "learning_rate": 4.0033551273227806e-05, "loss": 0.9405, "step": 115850 }, { "epoch": 19.941500344115624, "grad_norm": 9.439332962036133, "learning_rate": 4.002924982794219e-05, "loss": 0.9796, "step": 115900 }, { "epoch": 19.950103234686853, "grad_norm": 9.455728530883789, "learning_rate": 4.0024948382656575e-05, "loss": 0.9754, "step": 115950 }, { "epoch": 19.958706125258086, "grad_norm": 10.43875503540039, "learning_rate": 4.0020646937370956e-05, "loss": 0.9516, "step": 116000 }, { "epoch": 19.96730901582932, "grad_norm": 6.554933547973633, "learning_rate": 4.0016345492085344e-05, "loss": 0.956, "step": 116050 }, { "epoch": 19.97591190640055, "grad_norm": 8.55671501159668, "learning_rate": 4.001204404679973e-05, "loss": 0.9259, "step": 116100 }, { "epoch": 19.984514796971784, "grad_norm": 11.617072105407715, "learning_rate": 4.0007742601514106e-05, "loss": 0.9182, "step": 116150 }, { "epoch": 19.993117687543016, "grad_norm": 14.037917137145996, "learning_rate": 4.0003441156228494e-05, "loss": 1.0278, "step": 116200 }, { "epoch": 20.0, "eval_accuracy": 0.49492342109791776, "eval_f1": 0.4756551055847811, "eval_f1_DuraRiadoRio_16x16": 0.5028094979155338, "eval_f1_Mole_16x16": 0.2373134328358209, "eval_f1_Quebrado_16x16": 0.6511245200219419, "eval_f1_RiadoRio_16x16": 0.45937211449676824, "eval_f1_RioFechado_16x16": 0.5276559626538407, "eval_loss": 1.8705594539642334, "eval_precision": 0.6149769943966572, "eval_precision_DuraRiadoRio_16x16": 0.4316837846249611, "eval_precision_Mole_16x16": 0.8457446808510638, "eval_precision_Quebrado_16x16": 0.8845007451564829, "eval_precision_RiadoRio_16x16": 0.5236842105263158, "eval_precision_RioFechado_16x16": 0.3892715508244625, "eval_recall": 0.49660744747624114, "eval_recall_DuraRiadoRio_16x16": 0.6019965277777778, "eval_recall_Mole_16x16": 0.13802083333333334, "eval_recall_Quebrado_16x16": 0.5151909722222222, "eval_recall_RiadoRio_16x16": 0.40912828947368424, "eval_recall_RioFechado_16x16": 0.8187006145741879, "eval_runtime": 46.2123, "eval_samples_per_second": 251.492, "eval_steps_per_second": 15.732, "step": 116240 }, { "epoch": 20.001720578114245, "grad_norm": 12.118302345275879, "learning_rate": 3.999913971094288e-05, "loss": 0.9486, "step": 116250 }, { "epoch": 20.010323468685478, "grad_norm": 9.989070892333984, "learning_rate": 3.9994838265657263e-05, "loss": 0.9754, "step": 116300 }, { "epoch": 20.01892635925671, "grad_norm": 11.197558403015137, "learning_rate": 3.999053682037165e-05, "loss": 0.8661, "step": 116350 }, { "epoch": 20.027529249827943, "grad_norm": 9.291105270385742, "learning_rate": 3.9986235375086026e-05, "loss": 0.9758, "step": 116400 }, { "epoch": 20.036132140399175, "grad_norm": 11.185808181762695, "learning_rate": 3.9981933929800414e-05, "loss": 0.9913, "step": 116450 }, { "epoch": 20.044735030970408, "grad_norm": 8.925411224365234, "learning_rate": 3.99776324845148e-05, "loss": 0.9553, "step": 116500 }, { "epoch": 20.053337921541637, "grad_norm": 10.630066871643066, "learning_rate": 3.997333103922918e-05, "loss": 0.9368, "step": 116550 }, { "epoch": 20.06194081211287, "grad_norm": 14.013883590698242, "learning_rate": 3.996902959394357e-05, "loss": 0.971, "step": 116600 }, { "epoch": 20.070543702684102, "grad_norm": 8.223773002624512, "learning_rate": 3.9964728148657945e-05, "loss": 0.9684, "step": 116650 }, { "epoch": 20.079146593255334, "grad_norm": 12.887332916259766, "learning_rate": 3.996042670337233e-05, "loss": 1.0017, "step": 116700 }, { "epoch": 20.087749483826567, "grad_norm": 12.604436874389648, "learning_rate": 3.995612525808672e-05, "loss": 0.9367, "step": 116750 }, { "epoch": 20.0963523743978, "grad_norm": 12.499285697937012, "learning_rate": 3.99518238128011e-05, "loss": 0.969, "step": 116800 }, { "epoch": 20.10495526496903, "grad_norm": 10.539002418518066, "learning_rate": 3.994752236751549e-05, "loss": 0.8977, "step": 116850 }, { "epoch": 20.11355815554026, "grad_norm": 9.93300724029541, "learning_rate": 3.994322092222987e-05, "loss": 0.9289, "step": 116900 }, { "epoch": 20.122161046111493, "grad_norm": 12.653271675109863, "learning_rate": 3.993891947694425e-05, "loss": 0.9573, "step": 116950 }, { "epoch": 20.130763936682726, "grad_norm": 10.66557502746582, "learning_rate": 3.993461803165864e-05, "loss": 0.9346, "step": 117000 }, { "epoch": 20.13936682725396, "grad_norm": 15.55444622039795, "learning_rate": 3.993031658637302e-05, "loss": 0.9498, "step": 117050 }, { "epoch": 20.147969717825188, "grad_norm": 13.73047924041748, "learning_rate": 3.992601514108741e-05, "loss": 0.9078, "step": 117100 }, { "epoch": 20.15657260839642, "grad_norm": 9.701229095458984, "learning_rate": 3.992171369580179e-05, "loss": 0.945, "step": 117150 }, { "epoch": 20.165175498967653, "grad_norm": 10.596701622009277, "learning_rate": 3.991741225051617e-05, "loss": 0.908, "step": 117200 }, { "epoch": 20.173778389538885, "grad_norm": 10.27924919128418, "learning_rate": 3.991311080523056e-05, "loss": 0.9916, "step": 117250 }, { "epoch": 20.182381280110118, "grad_norm": 11.613537788391113, "learning_rate": 3.990880935994494e-05, "loss": 0.9506, "step": 117300 }, { "epoch": 20.19098417068135, "grad_norm": 7.729433059692383, "learning_rate": 3.990450791465933e-05, "loss": 1.0336, "step": 117350 }, { "epoch": 20.19958706125258, "grad_norm": 11.023784637451172, "learning_rate": 3.990020646937371e-05, "loss": 1.0257, "step": 117400 }, { "epoch": 20.20818995182381, "grad_norm": 11.470181465148926, "learning_rate": 3.98959050240881e-05, "loss": 0.9247, "step": 117450 }, { "epoch": 20.216792842395044, "grad_norm": 10.870711326599121, "learning_rate": 3.989160357880248e-05, "loss": 0.9361, "step": 117500 }, { "epoch": 20.225395732966277, "grad_norm": 11.935465812683105, "learning_rate": 3.988730213351686e-05, "loss": 0.9435, "step": 117550 }, { "epoch": 20.23399862353751, "grad_norm": 9.565665245056152, "learning_rate": 3.988300068823125e-05, "loss": 0.9385, "step": 117600 }, { "epoch": 20.242601514108742, "grad_norm": 9.729279518127441, "learning_rate": 3.9878699242945636e-05, "loss": 0.9396, "step": 117650 }, { "epoch": 20.25120440467997, "grad_norm": 11.03827953338623, "learning_rate": 3.987439779766002e-05, "loss": 0.9281, "step": 117700 }, { "epoch": 20.259807295251203, "grad_norm": 7.229173183441162, "learning_rate": 3.98700963523744e-05, "loss": 0.8583, "step": 117750 }, { "epoch": 20.268410185822436, "grad_norm": 10.150164604187012, "learning_rate": 3.986579490708878e-05, "loss": 0.9394, "step": 117800 }, { "epoch": 20.27701307639367, "grad_norm": 11.78131103515625, "learning_rate": 3.986149346180317e-05, "loss": 0.9848, "step": 117850 }, { "epoch": 20.2856159669649, "grad_norm": 11.139106750488281, "learning_rate": 3.9857192016517556e-05, "loss": 0.9439, "step": 117900 }, { "epoch": 20.294218857536134, "grad_norm": 8.082491874694824, "learning_rate": 3.985289057123194e-05, "loss": 0.9241, "step": 117950 }, { "epoch": 20.302821748107363, "grad_norm": 9.877345085144043, "learning_rate": 3.984858912594632e-05, "loss": 0.913, "step": 118000 }, { "epoch": 20.311424638678595, "grad_norm": 10.460057258605957, "learning_rate": 3.98442876806607e-05, "loss": 0.9815, "step": 118050 }, { "epoch": 20.320027529249828, "grad_norm": 9.88249683380127, "learning_rate": 3.983998623537509e-05, "loss": 0.9431, "step": 118100 }, { "epoch": 20.32863041982106, "grad_norm": 6.896435260772705, "learning_rate": 3.9835684790089475e-05, "loss": 0.9532, "step": 118150 }, { "epoch": 20.337233310392293, "grad_norm": 9.385590553283691, "learning_rate": 3.9831383344803856e-05, "loss": 0.9529, "step": 118200 }, { "epoch": 20.345836200963525, "grad_norm": 8.484158515930176, "learning_rate": 3.9827081899518244e-05, "loss": 0.9252, "step": 118250 }, { "epoch": 20.354439091534754, "grad_norm": 6.935211181640625, "learning_rate": 3.9822780454232625e-05, "loss": 0.9549, "step": 118300 }, { "epoch": 20.363041982105987, "grad_norm": 9.474265098571777, "learning_rate": 3.981847900894701e-05, "loss": 0.9622, "step": 118350 }, { "epoch": 20.37164487267722, "grad_norm": 8.84957218170166, "learning_rate": 3.9814177563661395e-05, "loss": 0.9652, "step": 118400 }, { "epoch": 20.38024776324845, "grad_norm": 8.125936508178711, "learning_rate": 3.9809876118375776e-05, "loss": 0.9289, "step": 118450 }, { "epoch": 20.388850653819684, "grad_norm": 8.3505277633667, "learning_rate": 3.9805574673090164e-05, "loss": 0.8952, "step": 118500 }, { "epoch": 20.397453544390917, "grad_norm": 7.967491626739502, "learning_rate": 3.9801273227804545e-05, "loss": 0.9211, "step": 118550 }, { "epoch": 20.406056434962146, "grad_norm": 13.488630294799805, "learning_rate": 3.9796971782518926e-05, "loss": 0.9417, "step": 118600 }, { "epoch": 20.41465932553338, "grad_norm": 9.122420310974121, "learning_rate": 3.9792670337233314e-05, "loss": 0.9776, "step": 118650 }, { "epoch": 20.42326221610461, "grad_norm": 8.593908309936523, "learning_rate": 3.9788368891947695e-05, "loss": 0.923, "step": 118700 }, { "epoch": 20.431865106675843, "grad_norm": 11.618017196655273, "learning_rate": 3.978406744666208e-05, "loss": 1.0099, "step": 118750 }, { "epoch": 20.440467997247076, "grad_norm": 11.712356567382812, "learning_rate": 3.9779766001376464e-05, "loss": 0.9342, "step": 118800 }, { "epoch": 20.44907088781831, "grad_norm": 8.41947078704834, "learning_rate": 3.9775464556090845e-05, "loss": 0.938, "step": 118850 }, { "epoch": 20.457673778389537, "grad_norm": 8.99519157409668, "learning_rate": 3.9771163110805233e-05, "loss": 0.996, "step": 118900 }, { "epoch": 20.46627666896077, "grad_norm": 9.879176139831543, "learning_rate": 3.9766861665519615e-05, "loss": 0.9901, "step": 118950 }, { "epoch": 20.474879559532003, "grad_norm": 7.515507698059082, "learning_rate": 3.9762560220234e-05, "loss": 0.9397, "step": 119000 }, { "epoch": 20.483482450103235, "grad_norm": 8.87443733215332, "learning_rate": 3.9758258774948384e-05, "loss": 1.002, "step": 119050 }, { "epoch": 20.492085340674468, "grad_norm": 7.584824562072754, "learning_rate": 3.9753957329662765e-05, "loss": 0.965, "step": 119100 }, { "epoch": 20.5006882312457, "grad_norm": 7.133138179779053, "learning_rate": 3.974965588437715e-05, "loss": 0.989, "step": 119150 }, { "epoch": 20.50929112181693, "grad_norm": 9.050799369812012, "learning_rate": 3.9745354439091534e-05, "loss": 0.8879, "step": 119200 }, { "epoch": 20.51789401238816, "grad_norm": 10.090590476989746, "learning_rate": 3.974105299380592e-05, "loss": 0.943, "step": 119250 }, { "epoch": 20.526496902959394, "grad_norm": 9.474884986877441, "learning_rate": 3.973675154852031e-05, "loss": 0.8722, "step": 119300 }, { "epoch": 20.535099793530627, "grad_norm": 8.669706344604492, "learning_rate": 3.9732450103234684e-05, "loss": 0.9184, "step": 119350 }, { "epoch": 20.54370268410186, "grad_norm": 11.066435813903809, "learning_rate": 3.972814865794907e-05, "loss": 0.9653, "step": 119400 }, { "epoch": 20.552305574673092, "grad_norm": 9.983501434326172, "learning_rate": 3.9723847212663453e-05, "loss": 0.9274, "step": 119450 }, { "epoch": 20.56090846524432, "grad_norm": 10.545122146606445, "learning_rate": 3.971954576737784e-05, "loss": 0.9426, "step": 119500 }, { "epoch": 20.569511355815553, "grad_norm": 13.092851638793945, "learning_rate": 3.971524432209223e-05, "loss": 0.9619, "step": 119550 }, { "epoch": 20.578114246386786, "grad_norm": 8.905010223388672, "learning_rate": 3.971094287680661e-05, "loss": 0.9343, "step": 119600 }, { "epoch": 20.58671713695802, "grad_norm": 9.455330848693848, "learning_rate": 3.970664143152099e-05, "loss": 0.9365, "step": 119650 }, { "epoch": 20.59532002752925, "grad_norm": 9.361190795898438, "learning_rate": 3.970233998623537e-05, "loss": 0.9245, "step": 119700 }, { "epoch": 20.60392291810048, "grad_norm": 11.736649513244629, "learning_rate": 3.969803854094976e-05, "loss": 0.9059, "step": 119750 }, { "epoch": 20.612525808671712, "grad_norm": 9.166767120361328, "learning_rate": 3.969373709566415e-05, "loss": 0.973, "step": 119800 }, { "epoch": 20.621128699242945, "grad_norm": 8.481391906738281, "learning_rate": 3.968943565037853e-05, "loss": 0.9537, "step": 119850 }, { "epoch": 20.629731589814178, "grad_norm": 13.109936714172363, "learning_rate": 3.968513420509291e-05, "loss": 0.9622, "step": 119900 }, { "epoch": 20.63833448038541, "grad_norm": 6.845538139343262, "learning_rate": 3.96808327598073e-05, "loss": 0.9676, "step": 119950 }, { "epoch": 20.646937370956643, "grad_norm": 12.486204147338867, "learning_rate": 3.967653131452168e-05, "loss": 0.9569, "step": 120000 }, { "epoch": 20.655540261527875, "grad_norm": 10.167862892150879, "learning_rate": 3.967222986923607e-05, "loss": 0.9763, "step": 120050 }, { "epoch": 20.664143152099104, "grad_norm": 9.699884414672852, "learning_rate": 3.966792842395045e-05, "loss": 0.9344, "step": 120100 }, { "epoch": 20.672746042670337, "grad_norm": 11.607645034790039, "learning_rate": 3.966362697866483e-05, "loss": 0.9601, "step": 120150 }, { "epoch": 20.68134893324157, "grad_norm": 12.083114624023438, "learning_rate": 3.965932553337922e-05, "loss": 0.9426, "step": 120200 }, { "epoch": 20.6899518238128, "grad_norm": 16.054641723632812, "learning_rate": 3.96550240880936e-05, "loss": 0.9482, "step": 120250 }, { "epoch": 20.698554714384034, "grad_norm": 13.599163055419922, "learning_rate": 3.965072264280799e-05, "loss": 1.0092, "step": 120300 }, { "epoch": 20.707157604955263, "grad_norm": 10.359939575195312, "learning_rate": 3.964642119752237e-05, "loss": 0.9237, "step": 120350 }, { "epoch": 20.715760495526496, "grad_norm": 10.208233833312988, "learning_rate": 3.964211975223676e-05, "loss": 0.9566, "step": 120400 }, { "epoch": 20.72436338609773, "grad_norm": 17.25006866455078, "learning_rate": 3.963781830695114e-05, "loss": 0.9961, "step": 120450 }, { "epoch": 20.73296627666896, "grad_norm": 7.178842544555664, "learning_rate": 3.963351686166552e-05, "loss": 0.9502, "step": 120500 }, { "epoch": 20.741569167240193, "grad_norm": 12.86752700805664, "learning_rate": 3.962921541637991e-05, "loss": 0.9349, "step": 120550 }, { "epoch": 20.750172057811426, "grad_norm": 11.541356086730957, "learning_rate": 3.962491397109429e-05, "loss": 0.9992, "step": 120600 }, { "epoch": 20.758774948382655, "grad_norm": 13.724189758300781, "learning_rate": 3.9620612525808676e-05, "loss": 0.943, "step": 120650 }, { "epoch": 20.767377838953887, "grad_norm": 7.481085777282715, "learning_rate": 3.961631108052306e-05, "loss": 0.9202, "step": 120700 }, { "epoch": 20.77598072952512, "grad_norm": 12.143173217773438, "learning_rate": 3.961200963523744e-05, "loss": 1.0043, "step": 120750 }, { "epoch": 20.784583620096353, "grad_norm": 12.833910942077637, "learning_rate": 3.9607708189951826e-05, "loss": 0.8907, "step": 120800 }, { "epoch": 20.793186510667585, "grad_norm": 6.5598602294921875, "learning_rate": 3.960340674466621e-05, "loss": 0.917, "step": 120850 }, { "epoch": 20.801789401238818, "grad_norm": 10.850086212158203, "learning_rate": 3.9599105299380596e-05, "loss": 0.9827, "step": 120900 }, { "epoch": 20.810392291810047, "grad_norm": 8.68698787689209, "learning_rate": 3.959480385409498e-05, "loss": 0.9593, "step": 120950 }, { "epoch": 20.81899518238128, "grad_norm": 11.038634300231934, "learning_rate": 3.959050240880936e-05, "loss": 0.9318, "step": 121000 }, { "epoch": 20.82759807295251, "grad_norm": 11.422515869140625, "learning_rate": 3.9586200963523746e-05, "loss": 0.9814, "step": 121050 }, { "epoch": 20.836200963523744, "grad_norm": 9.453520774841309, "learning_rate": 3.958189951823813e-05, "loss": 0.9702, "step": 121100 }, { "epoch": 20.844803854094977, "grad_norm": 7.8362627029418945, "learning_rate": 3.9577598072952515e-05, "loss": 0.9335, "step": 121150 }, { "epoch": 20.85340674466621, "grad_norm": 10.54257583618164, "learning_rate": 3.95732966276669e-05, "loss": 0.9596, "step": 121200 }, { "epoch": 20.86200963523744, "grad_norm": 14.348710060119629, "learning_rate": 3.956899518238128e-05, "loss": 0.9678, "step": 121250 }, { "epoch": 20.87061252580867, "grad_norm": 7.104825496673584, "learning_rate": 3.9564693737095665e-05, "loss": 0.9734, "step": 121300 }, { "epoch": 20.879215416379903, "grad_norm": 13.072660446166992, "learning_rate": 3.956039229181005e-05, "loss": 0.9725, "step": 121350 }, { "epoch": 20.887818306951136, "grad_norm": 7.718062877655029, "learning_rate": 3.9556090846524434e-05, "loss": 0.9401, "step": 121400 }, { "epoch": 20.89642119752237, "grad_norm": 8.554508209228516, "learning_rate": 3.955178940123882e-05, "loss": 0.9834, "step": 121450 }, { "epoch": 20.9050240880936, "grad_norm": 17.747159957885742, "learning_rate": 3.95474879559532e-05, "loss": 0.9793, "step": 121500 }, { "epoch": 20.91362697866483, "grad_norm": 11.682493209838867, "learning_rate": 3.9543186510667585e-05, "loss": 0.9755, "step": 121550 }, { "epoch": 20.922229869236062, "grad_norm": 11.279479026794434, "learning_rate": 3.953888506538197e-05, "loss": 0.9439, "step": 121600 }, { "epoch": 20.930832759807295, "grad_norm": 12.442818641662598, "learning_rate": 3.9534583620096354e-05, "loss": 0.9077, "step": 121650 }, { "epoch": 20.939435650378527, "grad_norm": 10.011394500732422, "learning_rate": 3.953028217481074e-05, "loss": 0.9798, "step": 121700 }, { "epoch": 20.94803854094976, "grad_norm": 9.92039680480957, "learning_rate": 3.952598072952512e-05, "loss": 0.9383, "step": 121750 }, { "epoch": 20.956641431520993, "grad_norm": 8.138964653015137, "learning_rate": 3.9521679284239504e-05, "loss": 0.9839, "step": 121800 }, { "epoch": 20.96524432209222, "grad_norm": 9.626153945922852, "learning_rate": 3.951737783895389e-05, "loss": 0.9825, "step": 121850 }, { "epoch": 20.973847212663454, "grad_norm": 9.543601989746094, "learning_rate": 3.951307639366827e-05, "loss": 0.9696, "step": 121900 }, { "epoch": 20.982450103234687, "grad_norm": 11.632718086242676, "learning_rate": 3.950877494838266e-05, "loss": 0.916, "step": 121950 }, { "epoch": 20.99105299380592, "grad_norm": 5.477049827575684, "learning_rate": 3.950447350309704e-05, "loss": 0.9648, "step": 122000 }, { "epoch": 20.99965588437715, "grad_norm": 12.999841690063477, "learning_rate": 3.9500172057811423e-05, "loss": 0.9218, "step": 122050 }, { "epoch": 21.0, "eval_accuracy": 0.43890896575460336, "eval_f1": 0.37470397563181246, "eval_f1_DuraRiadoRio_16x16": 0.3262124711316397, "eval_f1_Mole_16x16": 0.5110181311018132, "eval_f1_Quebrado_16x16": 0.610204081632653, "eval_f1_RiadoRio_16x16": 0.1381692573402418, "eval_f1_RioFechado_16x16": 0.28791593695271456, "eval_loss": 2.4841432571411133, "eval_precision": 0.49344471396982803, "eval_precision_DuraRiadoRio_16x16": 0.4870689655172414, "eval_precision_Mole_16x16": 0.3764899301274147, "eval_precision_Quebrado_16x16": 0.4593942054433714, "eval_precision_RiadoRio_16x16": 0.4319654427645788, "eval_precision_RioFechado_16x16": 0.7123050259965338, "eval_recall": 0.44228859732554976, "eval_recall_DuraRiadoRio_16x16": 0.24522569444444445, "eval_recall_Mole_16x16": 0.7951388888888888, "eval_recall_Quebrado_16x16": 0.9084201388888888, "eval_recall_RiadoRio_16x16": 0.08223684210526316, "eval_recall_RioFechado_16x16": 0.1804214223002634, "eval_runtime": 46.2271, "eval_samples_per_second": 251.411, "eval_steps_per_second": 15.727, "step": 122052 }, { "epoch": 21.008258774948384, "grad_norm": 6.312785625457764, "learning_rate": 3.949587061252581e-05, "loss": 0.96, "step": 122100 }, { "epoch": 21.016861665519613, "grad_norm": 7.104005813598633, "learning_rate": 3.949156916724019e-05, "loss": 0.8705, "step": 122150 }, { "epoch": 21.025464556090846, "grad_norm": 8.187152862548828, "learning_rate": 3.948726772195458e-05, "loss": 0.9606, "step": 122200 }, { "epoch": 21.03406744666208, "grad_norm": 14.745664596557617, "learning_rate": 3.948296627666896e-05, "loss": 0.9336, "step": 122250 }, { "epoch": 21.04267033723331, "grad_norm": 7.0594000816345215, "learning_rate": 3.947866483138334e-05, "loss": 0.9147, "step": 122300 }, { "epoch": 21.051273227804543, "grad_norm": 8.758723258972168, "learning_rate": 3.947436338609773e-05, "loss": 0.9648, "step": 122350 }, { "epoch": 21.059876118375776, "grad_norm": 8.492600440979004, "learning_rate": 3.947006194081211e-05, "loss": 0.9747, "step": 122400 }, { "epoch": 21.068479008947005, "grad_norm": 7.069737911224365, "learning_rate": 3.94657604955265e-05, "loss": 0.9329, "step": 122450 }, { "epoch": 21.077081899518237, "grad_norm": 6.679958343505859, "learning_rate": 3.946145905024088e-05, "loss": 0.9076, "step": 122500 }, { "epoch": 21.08568479008947, "grad_norm": 13.6814603805542, "learning_rate": 3.945715760495527e-05, "loss": 0.9369, "step": 122550 }, { "epoch": 21.094287680660702, "grad_norm": 10.259180068969727, "learning_rate": 3.945285615966965e-05, "loss": 0.8967, "step": 122600 }, { "epoch": 21.102890571231935, "grad_norm": 8.479695320129395, "learning_rate": 3.944855471438403e-05, "loss": 0.9101, "step": 122650 }, { "epoch": 21.111493461803168, "grad_norm": 5.997395038604736, "learning_rate": 3.944425326909842e-05, "loss": 0.9403, "step": 122700 }, { "epoch": 21.120096352374397, "grad_norm": 9.167291641235352, "learning_rate": 3.943995182381281e-05, "loss": 0.9116, "step": 122750 }, { "epoch": 21.12869924294563, "grad_norm": 10.980735778808594, "learning_rate": 3.943565037852719e-05, "loss": 0.8894, "step": 122800 }, { "epoch": 21.13730213351686, "grad_norm": 8.224651336669922, "learning_rate": 3.943134893324157e-05, "loss": 0.9777, "step": 122850 }, { "epoch": 21.145905024088094, "grad_norm": 8.854667663574219, "learning_rate": 3.942704748795595e-05, "loss": 0.9315, "step": 122900 }, { "epoch": 21.154507914659327, "grad_norm": 9.383535385131836, "learning_rate": 3.942274604267034e-05, "loss": 0.9102, "step": 122950 }, { "epoch": 21.16311080523056, "grad_norm": 12.862723350524902, "learning_rate": 3.941844459738473e-05, "loss": 0.9446, "step": 123000 }, { "epoch": 21.171713695801788, "grad_norm": 12.297831535339355, "learning_rate": 3.941414315209911e-05, "loss": 0.9507, "step": 123050 }, { "epoch": 21.18031658637302, "grad_norm": 11.549473762512207, "learning_rate": 3.9409841706813496e-05, "loss": 0.9353, "step": 123100 }, { "epoch": 21.188919476944253, "grad_norm": 7.663021564483643, "learning_rate": 3.940554026152787e-05, "loss": 0.9573, "step": 123150 }, { "epoch": 21.197522367515486, "grad_norm": 10.46956729888916, "learning_rate": 3.940123881624226e-05, "loss": 0.958, "step": 123200 }, { "epoch": 21.20612525808672, "grad_norm": 12.557709693908691, "learning_rate": 3.9396937370956646e-05, "loss": 0.94, "step": 123250 }, { "epoch": 21.214728148657947, "grad_norm": 11.26285171508789, "learning_rate": 3.939263592567103e-05, "loss": 0.9621, "step": 123300 }, { "epoch": 21.22333103922918, "grad_norm": 14.288579940795898, "learning_rate": 3.9388334480385415e-05, "loss": 0.9183, "step": 123350 }, { "epoch": 21.231933929800412, "grad_norm": 9.668583869934082, "learning_rate": 3.9384033035099796e-05, "loss": 0.9782, "step": 123400 }, { "epoch": 21.240536820371645, "grad_norm": 6.540658950805664, "learning_rate": 3.937973158981418e-05, "loss": 0.9793, "step": 123450 }, { "epoch": 21.249139710942877, "grad_norm": 10.462762832641602, "learning_rate": 3.9375430144528566e-05, "loss": 0.9405, "step": 123500 }, { "epoch": 21.25774260151411, "grad_norm": 14.55323600769043, "learning_rate": 3.937112869924295e-05, "loss": 0.9411, "step": 123550 }, { "epoch": 21.26634549208534, "grad_norm": 8.05602741241455, "learning_rate": 3.9366827253957335e-05, "loss": 0.9331, "step": 123600 }, { "epoch": 21.27494838265657, "grad_norm": 8.788267135620117, "learning_rate": 3.9362525808671716e-05, "loss": 0.9288, "step": 123650 }, { "epoch": 21.283551273227804, "grad_norm": 10.53455924987793, "learning_rate": 3.93582243633861e-05, "loss": 0.9153, "step": 123700 }, { "epoch": 21.292154163799037, "grad_norm": 9.101053237915039, "learning_rate": 3.9353922918100485e-05, "loss": 0.8986, "step": 123750 }, { "epoch": 21.30075705437027, "grad_norm": 10.010592460632324, "learning_rate": 3.9349621472814866e-05, "loss": 0.9329, "step": 123800 }, { "epoch": 21.3093599449415, "grad_norm": 10.2041654586792, "learning_rate": 3.9345320027529254e-05, "loss": 0.9731, "step": 123850 }, { "epoch": 21.31796283551273, "grad_norm": 8.192954063415527, "learning_rate": 3.9341018582243635e-05, "loss": 0.9838, "step": 123900 }, { "epoch": 21.326565726083963, "grad_norm": 9.894124031066895, "learning_rate": 3.9336717136958016e-05, "loss": 0.9782, "step": 123950 }, { "epoch": 21.335168616655196, "grad_norm": 9.366199493408203, "learning_rate": 3.9332415691672404e-05, "loss": 0.9302, "step": 124000 }, { "epoch": 21.34377150722643, "grad_norm": 9.029410362243652, "learning_rate": 3.9328114246386786e-05, "loss": 0.9483, "step": 124050 }, { "epoch": 21.35237439779766, "grad_norm": 8.378783226013184, "learning_rate": 3.9323812801101173e-05, "loss": 0.9717, "step": 124100 }, { "epoch": 21.360977288368893, "grad_norm": 8.97499942779541, "learning_rate": 3.931951135581556e-05, "loss": 0.9469, "step": 124150 }, { "epoch": 21.369580178940122, "grad_norm": 10.061330795288086, "learning_rate": 3.9315209910529936e-05, "loss": 0.9284, "step": 124200 }, { "epoch": 21.378183069511355, "grad_norm": 10.71595573425293, "learning_rate": 3.9310908465244324e-05, "loss": 0.985, "step": 124250 }, { "epoch": 21.386785960082587, "grad_norm": 10.386908531188965, "learning_rate": 3.9306607019958705e-05, "loss": 0.9341, "step": 124300 }, { "epoch": 21.39538885065382, "grad_norm": 9.648521423339844, "learning_rate": 3.930230557467309e-05, "loss": 0.9201, "step": 124350 }, { "epoch": 21.403991741225052, "grad_norm": 10.432940483093262, "learning_rate": 3.929800412938748e-05, "loss": 0.9812, "step": 124400 }, { "epoch": 21.412594631796285, "grad_norm": 12.09512996673584, "learning_rate": 3.9293702684101855e-05, "loss": 1.0037, "step": 124450 }, { "epoch": 21.421197522367514, "grad_norm": 6.75709342956543, "learning_rate": 3.928940123881624e-05, "loss": 0.9191, "step": 124500 }, { "epoch": 21.429800412938746, "grad_norm": 11.071005821228027, "learning_rate": 3.9285099793530624e-05, "loss": 0.9815, "step": 124550 }, { "epoch": 21.43840330350998, "grad_norm": 9.420500755310059, "learning_rate": 3.928079834824501e-05, "loss": 0.9122, "step": 124600 }, { "epoch": 21.44700619408121, "grad_norm": 16.520954132080078, "learning_rate": 3.92764969029594e-05, "loss": 0.9066, "step": 124650 }, { "epoch": 21.455609084652444, "grad_norm": 6.972529411315918, "learning_rate": 3.927219545767378e-05, "loss": 0.9111, "step": 124700 }, { "epoch": 21.464211975223677, "grad_norm": 7.42445707321167, "learning_rate": 3.926789401238816e-05, "loss": 0.9353, "step": 124750 }, { "epoch": 21.472814865794906, "grad_norm": 11.594476699829102, "learning_rate": 3.926359256710255e-05, "loss": 0.9046, "step": 124800 }, { "epoch": 21.481417756366138, "grad_norm": 9.218186378479004, "learning_rate": 3.925929112181693e-05, "loss": 0.9442, "step": 124850 }, { "epoch": 21.49002064693737, "grad_norm": 9.33782958984375, "learning_rate": 3.925498967653132e-05, "loss": 0.9114, "step": 124900 }, { "epoch": 21.498623537508603, "grad_norm": 8.553351402282715, "learning_rate": 3.92506882312457e-05, "loss": 0.9271, "step": 124950 }, { "epoch": 21.507226428079836, "grad_norm": 12.19158935546875, "learning_rate": 3.924638678596008e-05, "loss": 0.9145, "step": 125000 }, { "epoch": 21.51582931865107, "grad_norm": 9.198257446289062, "learning_rate": 3.924208534067447e-05, "loss": 0.9276, "step": 125050 }, { "epoch": 21.524432209222297, "grad_norm": 9.646537780761719, "learning_rate": 3.923778389538885e-05, "loss": 0.9639, "step": 125100 }, { "epoch": 21.53303509979353, "grad_norm": 8.859037399291992, "learning_rate": 3.923348245010324e-05, "loss": 0.9589, "step": 125150 }, { "epoch": 21.541637990364762, "grad_norm": 11.657605171203613, "learning_rate": 3.922918100481762e-05, "loss": 0.8935, "step": 125200 }, { "epoch": 21.550240880935995, "grad_norm": 10.800978660583496, "learning_rate": 3.922487955953201e-05, "loss": 0.9555, "step": 125250 }, { "epoch": 21.558843771507227, "grad_norm": 10.547904968261719, "learning_rate": 3.922057811424639e-05, "loss": 0.9456, "step": 125300 }, { "epoch": 21.56744666207846, "grad_norm": 7.84328031539917, "learning_rate": 3.921627666896077e-05, "loss": 0.9378, "step": 125350 }, { "epoch": 21.57604955264969, "grad_norm": 8.267740249633789, "learning_rate": 3.921197522367516e-05, "loss": 0.8962, "step": 125400 }, { "epoch": 21.58465244322092, "grad_norm": 14.94235610961914, "learning_rate": 3.920767377838954e-05, "loss": 0.9302, "step": 125450 }, { "epoch": 21.593255333792154, "grad_norm": 12.276426315307617, "learning_rate": 3.920337233310393e-05, "loss": 0.9489, "step": 125500 }, { "epoch": 21.601858224363387, "grad_norm": 11.84604263305664, "learning_rate": 3.919907088781831e-05, "loss": 0.9191, "step": 125550 }, { "epoch": 21.61046111493462, "grad_norm": 8.484162330627441, "learning_rate": 3.919476944253269e-05, "loss": 0.9322, "step": 125600 }, { "epoch": 21.61906400550585, "grad_norm": 6.6797590255737305, "learning_rate": 3.919046799724708e-05, "loss": 1.0033, "step": 125650 }, { "epoch": 21.62766689607708, "grad_norm": 9.759799003601074, "learning_rate": 3.918616655196146e-05, "loss": 0.9564, "step": 125700 }, { "epoch": 21.636269786648313, "grad_norm": 8.045982360839844, "learning_rate": 3.918186510667585e-05, "loss": 0.9337, "step": 125750 }, { "epoch": 21.644872677219546, "grad_norm": 8.354508399963379, "learning_rate": 3.917756366139023e-05, "loss": 0.9726, "step": 125800 }, { "epoch": 21.653475567790778, "grad_norm": 8.472122192382812, "learning_rate": 3.917326221610461e-05, "loss": 0.9661, "step": 125850 }, { "epoch": 21.66207845836201, "grad_norm": 15.123336791992188, "learning_rate": 3.9168960770819e-05, "loss": 0.9623, "step": 125900 }, { "epoch": 21.670681348933243, "grad_norm": 9.275153160095215, "learning_rate": 3.916465932553338e-05, "loss": 0.8659, "step": 125950 }, { "epoch": 21.679284239504472, "grad_norm": 10.875959396362305, "learning_rate": 3.9160357880247766e-05, "loss": 0.8859, "step": 126000 }, { "epoch": 21.687887130075705, "grad_norm": 12.186909675598145, "learning_rate": 3.9156056434962154e-05, "loss": 0.9285, "step": 126050 }, { "epoch": 21.696490020646937, "grad_norm": 12.704384803771973, "learning_rate": 3.915175498967653e-05, "loss": 0.9351, "step": 126100 }, { "epoch": 21.70509291121817, "grad_norm": 10.862345695495605, "learning_rate": 3.914745354439092e-05, "loss": 0.9444, "step": 126150 }, { "epoch": 21.713695801789402, "grad_norm": 7.211763858795166, "learning_rate": 3.9143152099105305e-05, "loss": 0.9468, "step": 126200 }, { "epoch": 21.72229869236063, "grad_norm": 9.53628921508789, "learning_rate": 3.9138850653819686e-05, "loss": 1.0037, "step": 126250 }, { "epoch": 21.730901582931864, "grad_norm": 13.79422378540039, "learning_rate": 3.9134549208534074e-05, "loss": 0.9736, "step": 126300 }, { "epoch": 21.739504473503096, "grad_norm": 15.156098365783691, "learning_rate": 3.913024776324845e-05, "loss": 0.9186, "step": 126350 }, { "epoch": 21.74810736407433, "grad_norm": 10.134681701660156, "learning_rate": 3.9125946317962836e-05, "loss": 0.9791, "step": 126400 }, { "epoch": 21.75671025464556, "grad_norm": 7.182354927062988, "learning_rate": 3.9121644872677224e-05, "loss": 0.9327, "step": 126450 }, { "epoch": 21.765313145216794, "grad_norm": 9.696478843688965, "learning_rate": 3.9117343427391605e-05, "loss": 0.9089, "step": 126500 }, { "epoch": 21.773916035788027, "grad_norm": 10.094278335571289, "learning_rate": 3.911304198210599e-05, "loss": 0.9482, "step": 126550 }, { "epoch": 21.782518926359256, "grad_norm": 11.311054229736328, "learning_rate": 3.910874053682037e-05, "loss": 0.9016, "step": 126600 }, { "epoch": 21.791121816930488, "grad_norm": 11.920093536376953, "learning_rate": 3.9104439091534756e-05, "loss": 0.9082, "step": 126650 }, { "epoch": 21.79972470750172, "grad_norm": 5.503902912139893, "learning_rate": 3.9100137646249144e-05, "loss": 0.9287, "step": 126700 }, { "epoch": 21.808327598072953, "grad_norm": 8.48508358001709, "learning_rate": 3.9095836200963525e-05, "loss": 0.8987, "step": 126750 }, { "epoch": 21.816930488644186, "grad_norm": 14.298482894897461, "learning_rate": 3.909153475567791e-05, "loss": 0.979, "step": 126800 }, { "epoch": 21.825533379215415, "grad_norm": 8.54508113861084, "learning_rate": 3.9087233310392294e-05, "loss": 0.8942, "step": 126850 }, { "epoch": 21.834136269786647, "grad_norm": 10.48590087890625, "learning_rate": 3.9082931865106675e-05, "loss": 0.9491, "step": 126900 }, { "epoch": 21.84273916035788, "grad_norm": 7.562083721160889, "learning_rate": 3.907863041982106e-05, "loss": 0.9652, "step": 126950 }, { "epoch": 21.851342050929112, "grad_norm": 8.31627082824707, "learning_rate": 3.9074328974535444e-05, "loss": 0.9527, "step": 127000 }, { "epoch": 21.859944941500345, "grad_norm": 10.768111228942871, "learning_rate": 3.907002752924983e-05, "loss": 0.9018, "step": 127050 }, { "epoch": 21.868547832071577, "grad_norm": 14.35283088684082, "learning_rate": 3.906572608396421e-05, "loss": 0.8544, "step": 127100 }, { "epoch": 21.877150722642806, "grad_norm": 7.546442985534668, "learning_rate": 3.9061424638678594e-05, "loss": 0.988, "step": 127150 }, { "epoch": 21.88575361321404, "grad_norm": 7.882843017578125, "learning_rate": 3.905712319339298e-05, "loss": 0.9322, "step": 127200 }, { "epoch": 21.89435650378527, "grad_norm": 10.823969841003418, "learning_rate": 3.9052821748107364e-05, "loss": 0.9486, "step": 127250 }, { "epoch": 21.902959394356504, "grad_norm": 8.272994041442871, "learning_rate": 3.904852030282175e-05, "loss": 0.9714, "step": 127300 }, { "epoch": 21.911562284927736, "grad_norm": 8.233833312988281, "learning_rate": 3.904421885753613e-05, "loss": 0.9344, "step": 127350 }, { "epoch": 21.92016517549897, "grad_norm": 12.915020942687988, "learning_rate": 3.903991741225052e-05, "loss": 0.955, "step": 127400 }, { "epoch": 21.928768066070198, "grad_norm": 7.8154168128967285, "learning_rate": 3.90356159669649e-05, "loss": 0.9064, "step": 127450 }, { "epoch": 21.93737095664143, "grad_norm": 14.198649406433105, "learning_rate": 3.903131452167928e-05, "loss": 0.8971, "step": 127500 }, { "epoch": 21.945973847212663, "grad_norm": 12.418644905090332, "learning_rate": 3.902701307639367e-05, "loss": 1.0106, "step": 127550 }, { "epoch": 21.954576737783896, "grad_norm": 10.103485107421875, "learning_rate": 3.902271163110805e-05, "loss": 0.9969, "step": 127600 }, { "epoch": 21.963179628355128, "grad_norm": 7.385933876037598, "learning_rate": 3.901841018582244e-05, "loss": 0.9608, "step": 127650 }, { "epoch": 21.97178251892636, "grad_norm": 10.676424026489258, "learning_rate": 3.901410874053682e-05, "loss": 0.9688, "step": 127700 }, { "epoch": 21.98038540949759, "grad_norm": 11.419791221618652, "learning_rate": 3.90098072952512e-05, "loss": 0.9344, "step": 127750 }, { "epoch": 21.988988300068822, "grad_norm": 10.221179008483887, "learning_rate": 3.900550584996559e-05, "loss": 0.9122, "step": 127800 }, { "epoch": 21.997591190640055, "grad_norm": 8.520549774169922, "learning_rate": 3.900120440467998e-05, "loss": 0.9428, "step": 127850 }, { "epoch": 22.0, "eval_accuracy": 0.5750301152985716, "eval_f1": 0.5714878822603764, "eval_f1_DuraRiadoRio_16x16": 0.5734896302975654, "eval_f1_Mole_16x16": 0.6020222971221156, "eval_f1_Quebrado_16x16": 0.7367167024064808, "eval_f1_RiadoRio_16x16": 0.37800875273522977, "eval_f1_RioFechado_16x16": 0.5672020287404903, "eval_loss": 1.37923002243042, "eval_precision": 0.6286157252486444, "eval_precision_DuraRiadoRio_16x16": 0.5966228893058161, "eval_precision_Mole_16x16": 0.7475853187379266, "eval_precision_Quebrado_16x16": 0.8166930797675647, "eval_precision_RiadoRio_16x16": 0.5645424836601307, "eval_precision_RioFechado_16x16": 0.41763485477178425, "eval_recall": 0.5789589406232512, "eval_recall_DuraRiadoRio_16x16": 0.5520833333333334, "eval_recall_Mole_16x16": 0.50390625, "eval_recall_Quebrado_16x16": 0.6710069444444444, "eval_recall_RiadoRio_16x16": 0.28412828947368424, "eval_recall_RioFechado_16x16": 0.8836698858647937, "eval_runtime": 45.9159, "eval_samples_per_second": 253.115, "eval_steps_per_second": 15.833, "step": 127864 }, { "epoch": 22.006194081211287, "grad_norm": 10.461928367614746, "learning_rate": 3.899690295939436e-05, "loss": 0.9127, "step": 127900 }, { "epoch": 22.01479697178252, "grad_norm": 8.490116119384766, "learning_rate": 3.899260151410874e-05, "loss": 0.9185, "step": 127950 }, { "epoch": 22.023399862353752, "grad_norm": 10.252252578735352, "learning_rate": 3.898830006882312e-05, "loss": 0.9608, "step": 128000 }, { "epoch": 22.03200275292498, "grad_norm": 9.294808387756348, "learning_rate": 3.898399862353751e-05, "loss": 0.9396, "step": 128050 }, { "epoch": 22.040605643496214, "grad_norm": 7.001893997192383, "learning_rate": 3.89796971782519e-05, "loss": 0.9674, "step": 128100 }, { "epoch": 22.049208534067446, "grad_norm": 12.708827018737793, "learning_rate": 3.897539573296628e-05, "loss": 0.9067, "step": 128150 }, { "epoch": 22.05781142463868, "grad_norm": 15.853740692138672, "learning_rate": 3.897109428768067e-05, "loss": 0.9252, "step": 128200 }, { "epoch": 22.06641431520991, "grad_norm": 15.707086563110352, "learning_rate": 3.896679284239504e-05, "loss": 0.983, "step": 128250 }, { "epoch": 22.075017205781144, "grad_norm": 10.15360164642334, "learning_rate": 3.896249139710943e-05, "loss": 0.9309, "step": 128300 }, { "epoch": 22.083620096352373, "grad_norm": 9.125489234924316, "learning_rate": 3.895818995182382e-05, "loss": 0.9577, "step": 128350 }, { "epoch": 22.092222986923606, "grad_norm": 7.435138702392578, "learning_rate": 3.89538885065382e-05, "loss": 0.9309, "step": 128400 }, { "epoch": 22.100825877494838, "grad_norm": 14.459527015686035, "learning_rate": 3.8949587061252586e-05, "loss": 0.9116, "step": 128450 }, { "epoch": 22.10942876806607, "grad_norm": 9.293680191040039, "learning_rate": 3.894528561596697e-05, "loss": 0.9499, "step": 128500 }, { "epoch": 22.118031658637303, "grad_norm": 9.821526527404785, "learning_rate": 3.894098417068135e-05, "loss": 0.9003, "step": 128550 }, { "epoch": 22.126634549208536, "grad_norm": 13.191445350646973, "learning_rate": 3.8936682725395736e-05, "loss": 0.9644, "step": 128600 }, { "epoch": 22.135237439779765, "grad_norm": 9.68014907836914, "learning_rate": 3.893238128011012e-05, "loss": 0.8734, "step": 128650 }, { "epoch": 22.143840330350997, "grad_norm": 8.480592727661133, "learning_rate": 3.8928079834824506e-05, "loss": 0.9217, "step": 128700 }, { "epoch": 22.15244322092223, "grad_norm": 12.240468978881836, "learning_rate": 3.892377838953889e-05, "loss": 0.9097, "step": 128750 }, { "epoch": 22.161046111493462, "grad_norm": 8.29737377166748, "learning_rate": 3.891947694425327e-05, "loss": 0.9146, "step": 128800 }, { "epoch": 22.169649002064695, "grad_norm": 11.43956470489502, "learning_rate": 3.8915175498967656e-05, "loss": 0.9257, "step": 128850 }, { "epoch": 22.178251892635927, "grad_norm": 9.498772621154785, "learning_rate": 3.891087405368204e-05, "loss": 0.9237, "step": 128900 }, { "epoch": 22.186854783207156, "grad_norm": 7.940093994140625, "learning_rate": 3.8906572608396425e-05, "loss": 0.9832, "step": 128950 }, { "epoch": 22.19545767377839, "grad_norm": 16.963834762573242, "learning_rate": 3.8902271163110806e-05, "loss": 0.9605, "step": 129000 }, { "epoch": 22.20406056434962, "grad_norm": 9.034221649169922, "learning_rate": 3.889796971782519e-05, "loss": 0.9479, "step": 129050 }, { "epoch": 22.212663454920854, "grad_norm": 12.421195030212402, "learning_rate": 3.8893668272539575e-05, "loss": 0.9601, "step": 129100 }, { "epoch": 22.221266345492086, "grad_norm": 10.092775344848633, "learning_rate": 3.8889366827253956e-05, "loss": 0.9212, "step": 129150 }, { "epoch": 22.22986923606332, "grad_norm": 8.615799903869629, "learning_rate": 3.8885065381968344e-05, "loss": 0.9517, "step": 129200 }, { "epoch": 22.238472126634548, "grad_norm": 9.576804161071777, "learning_rate": 3.888076393668273e-05, "loss": 0.9983, "step": 129250 }, { "epoch": 22.24707501720578, "grad_norm": 9.789751052856445, "learning_rate": 3.887646249139711e-05, "loss": 0.9074, "step": 129300 }, { "epoch": 22.255677907777013, "grad_norm": 12.599346160888672, "learning_rate": 3.8872161046111495e-05, "loss": 0.9736, "step": 129350 }, { "epoch": 22.264280798348246, "grad_norm": 8.955522537231445, "learning_rate": 3.8867859600825876e-05, "loss": 0.92, "step": 129400 }, { "epoch": 22.272883688919478, "grad_norm": 9.819647789001465, "learning_rate": 3.8863558155540264e-05, "loss": 0.9573, "step": 129450 }, { "epoch": 22.28148657949071, "grad_norm": 11.076264381408691, "learning_rate": 3.885925671025465e-05, "loss": 0.9568, "step": 129500 }, { "epoch": 22.29008947006194, "grad_norm": 6.2475175857543945, "learning_rate": 3.885495526496903e-05, "loss": 0.955, "step": 129550 }, { "epoch": 22.298692360633172, "grad_norm": 10.482626914978027, "learning_rate": 3.8850653819683414e-05, "loss": 0.9646, "step": 129600 }, { "epoch": 22.307295251204405, "grad_norm": 15.69408893585205, "learning_rate": 3.8846352374397795e-05, "loss": 0.9137, "step": 129650 }, { "epoch": 22.315898141775637, "grad_norm": 7.598124027252197, "learning_rate": 3.884205092911218e-05, "loss": 0.9318, "step": 129700 }, { "epoch": 22.32450103234687, "grad_norm": 9.550304412841797, "learning_rate": 3.883774948382657e-05, "loss": 0.9181, "step": 129750 }, { "epoch": 22.3331039229181, "grad_norm": 10.702160835266113, "learning_rate": 3.883344803854095e-05, "loss": 0.9306, "step": 129800 }, { "epoch": 22.34170681348933, "grad_norm": 10.865273475646973, "learning_rate": 3.8829146593255334e-05, "loss": 0.9336, "step": 129850 }, { "epoch": 22.350309704060564, "grad_norm": 10.087119102478027, "learning_rate": 3.882484514796972e-05, "loss": 0.8798, "step": 129900 }, { "epoch": 22.358912594631796, "grad_norm": 7.768732070922852, "learning_rate": 3.88205437026841e-05, "loss": 0.9282, "step": 129950 }, { "epoch": 22.36751548520303, "grad_norm": 11.708263397216797, "learning_rate": 3.881624225739849e-05, "loss": 0.9553, "step": 130000 }, { "epoch": 22.37611837577426, "grad_norm": 11.077170372009277, "learning_rate": 3.881194081211287e-05, "loss": 0.915, "step": 130050 }, { "epoch": 22.38472126634549, "grad_norm": 13.086820602416992, "learning_rate": 3.880763936682725e-05, "loss": 0.8744, "step": 130100 }, { "epoch": 22.393324156916723, "grad_norm": 6.496758937835693, "learning_rate": 3.880333792154164e-05, "loss": 0.9234, "step": 130150 }, { "epoch": 22.401927047487955, "grad_norm": 10.193819999694824, "learning_rate": 3.879903647625602e-05, "loss": 0.9456, "step": 130200 }, { "epoch": 22.410529938059188, "grad_norm": 10.168835639953613, "learning_rate": 3.879473503097041e-05, "loss": 1.0064, "step": 130250 }, { "epoch": 22.41913282863042, "grad_norm": 8.882741928100586, "learning_rate": 3.879043358568479e-05, "loss": 0.9123, "step": 130300 }, { "epoch": 22.427735719201653, "grad_norm": 11.182729721069336, "learning_rate": 3.878613214039918e-05, "loss": 0.9282, "step": 130350 }, { "epoch": 22.436338609772882, "grad_norm": 7.320985794067383, "learning_rate": 3.878183069511356e-05, "loss": 0.9486, "step": 130400 }, { "epoch": 22.444941500344115, "grad_norm": 10.239843368530273, "learning_rate": 3.877752924982794e-05, "loss": 0.8475, "step": 130450 }, { "epoch": 22.453544390915347, "grad_norm": 12.961734771728516, "learning_rate": 3.877322780454233e-05, "loss": 0.8622, "step": 130500 }, { "epoch": 22.46214728148658, "grad_norm": 10.288259506225586, "learning_rate": 3.876892635925671e-05, "loss": 0.9501, "step": 130550 }, { "epoch": 22.470750172057812, "grad_norm": 5.97032356262207, "learning_rate": 3.87646249139711e-05, "loss": 0.8856, "step": 130600 }, { "epoch": 22.479353062629045, "grad_norm": 10.568169593811035, "learning_rate": 3.876032346868548e-05, "loss": 0.9407, "step": 130650 }, { "epoch": 22.487955953200274, "grad_norm": 6.060198783874512, "learning_rate": 3.875602202339986e-05, "loss": 0.9375, "step": 130700 }, { "epoch": 22.496558843771506, "grad_norm": 15.960383415222168, "learning_rate": 3.875172057811425e-05, "loss": 0.9724, "step": 130750 }, { "epoch": 22.50516173434274, "grad_norm": 11.378029823303223, "learning_rate": 3.874741913282863e-05, "loss": 0.9148, "step": 130800 }, { "epoch": 22.51376462491397, "grad_norm": 9.297844886779785, "learning_rate": 3.874311768754302e-05, "loss": 0.923, "step": 130850 }, { "epoch": 22.522367515485204, "grad_norm": 11.493600845336914, "learning_rate": 3.8738816242257406e-05, "loss": 0.9706, "step": 130900 }, { "epoch": 22.530970406056436, "grad_norm": 11.484955787658691, "learning_rate": 3.873451479697178e-05, "loss": 0.9277, "step": 130950 }, { "epoch": 22.539573296627665, "grad_norm": 8.592428207397461, "learning_rate": 3.873021335168617e-05, "loss": 0.9392, "step": 131000 }, { "epoch": 22.548176187198898, "grad_norm": 11.977800369262695, "learning_rate": 3.872591190640055e-05, "loss": 0.9894, "step": 131050 }, { "epoch": 22.55677907777013, "grad_norm": 11.150737762451172, "learning_rate": 3.872161046111494e-05, "loss": 0.9642, "step": 131100 }, { "epoch": 22.565381968341363, "grad_norm": 9.378612518310547, "learning_rate": 3.8717309015829325e-05, "loss": 0.9523, "step": 131150 }, { "epoch": 22.573984858912596, "grad_norm": 15.662482261657715, "learning_rate": 3.87130075705437e-05, "loss": 0.9299, "step": 131200 }, { "epoch": 22.582587749483828, "grad_norm": 9.106439590454102, "learning_rate": 3.870870612525809e-05, "loss": 0.8551, "step": 131250 }, { "epoch": 22.591190640055057, "grad_norm": 10.014338493347168, "learning_rate": 3.8704404679972476e-05, "loss": 0.8991, "step": 131300 }, { "epoch": 22.59979353062629, "grad_norm": 13.016349792480469, "learning_rate": 3.870010323468686e-05, "loss": 0.9428, "step": 131350 }, { "epoch": 22.608396421197522, "grad_norm": 12.380034446716309, "learning_rate": 3.8695801789401245e-05, "loss": 0.8924, "step": 131400 }, { "epoch": 22.616999311768755, "grad_norm": 10.746407508850098, "learning_rate": 3.869150034411562e-05, "loss": 0.9275, "step": 131450 }, { "epoch": 22.625602202339987, "grad_norm": 12.287127494812012, "learning_rate": 3.868719889883001e-05, "loss": 0.8977, "step": 131500 }, { "epoch": 22.63420509291122, "grad_norm": 11.84957218170166, "learning_rate": 3.8682897453544395e-05, "loss": 0.9176, "step": 131550 }, { "epoch": 22.64280798348245, "grad_norm": 10.82559871673584, "learning_rate": 3.8678596008258776e-05, "loss": 0.9983, "step": 131600 }, { "epoch": 22.65141087405368, "grad_norm": 8.551328659057617, "learning_rate": 3.8674294562973164e-05, "loss": 0.9752, "step": 131650 }, { "epoch": 22.660013764624914, "grad_norm": 9.830801963806152, "learning_rate": 3.8669993117687545e-05, "loss": 0.9441, "step": 131700 }, { "epoch": 22.668616655196146, "grad_norm": 11.936351776123047, "learning_rate": 3.8665691672401927e-05, "loss": 0.9243, "step": 131750 }, { "epoch": 22.67721954576738, "grad_norm": 8.018609046936035, "learning_rate": 3.8661390227116314e-05, "loss": 0.9439, "step": 131800 }, { "epoch": 22.68582243633861, "grad_norm": 13.282700538635254, "learning_rate": 3.8657088781830696e-05, "loss": 0.9377, "step": 131850 }, { "epoch": 22.69442532690984, "grad_norm": 10.308577537536621, "learning_rate": 3.8652787336545084e-05, "loss": 0.9341, "step": 131900 }, { "epoch": 22.703028217481073, "grad_norm": 10.264933586120605, "learning_rate": 3.8648485891259465e-05, "loss": 0.9192, "step": 131950 }, { "epoch": 22.711631108052305, "grad_norm": 12.706130981445312, "learning_rate": 3.8644184445973846e-05, "loss": 0.9646, "step": 132000 }, { "epoch": 22.720233998623538, "grad_norm": 9.376574516296387, "learning_rate": 3.8639883000688234e-05, "loss": 0.8994, "step": 132050 }, { "epoch": 22.72883688919477, "grad_norm": 8.906094551086426, "learning_rate": 3.8635581555402615e-05, "loss": 0.9384, "step": 132100 }, { "epoch": 22.737439779766003, "grad_norm": 6.351970672607422, "learning_rate": 3.8631280110117e-05, "loss": 0.9571, "step": 132150 }, { "epoch": 22.746042670337232, "grad_norm": 7.85660457611084, "learning_rate": 3.8626978664831384e-05, "loss": 0.9182, "step": 132200 }, { "epoch": 22.754645560908465, "grad_norm": 18.12624168395996, "learning_rate": 3.8622677219545765e-05, "loss": 0.9678, "step": 132250 }, { "epoch": 22.763248451479697, "grad_norm": 8.639195442199707, "learning_rate": 3.861837577426015e-05, "loss": 0.937, "step": 132300 }, { "epoch": 22.77185134205093, "grad_norm": 8.722452163696289, "learning_rate": 3.8614074328974534e-05, "loss": 0.8557, "step": 132350 }, { "epoch": 22.780454232622162, "grad_norm": 9.711675643920898, "learning_rate": 3.860977288368892e-05, "loss": 1.002, "step": 132400 }, { "epoch": 22.789057123193395, "grad_norm": 7.344510078430176, "learning_rate": 3.8605471438403304e-05, "loss": 0.9603, "step": 132450 }, { "epoch": 22.797660013764624, "grad_norm": 12.728010177612305, "learning_rate": 3.860116999311769e-05, "loss": 0.9104, "step": 132500 }, { "epoch": 22.806262904335856, "grad_norm": 10.096725463867188, "learning_rate": 3.859686854783207e-05, "loss": 0.9291, "step": 132550 }, { "epoch": 22.81486579490709, "grad_norm": 9.064233779907227, "learning_rate": 3.8592567102546454e-05, "loss": 0.9588, "step": 132600 }, { "epoch": 22.82346868547832, "grad_norm": 7.449126720428467, "learning_rate": 3.858826565726084e-05, "loss": 0.9032, "step": 132650 }, { "epoch": 22.832071576049554, "grad_norm": 7.6983137130737305, "learning_rate": 3.858396421197523e-05, "loss": 0.9161, "step": 132700 }, { "epoch": 22.840674466620783, "grad_norm": 12.433198928833008, "learning_rate": 3.857966276668961e-05, "loss": 0.9127, "step": 132750 }, { "epoch": 22.849277357192015, "grad_norm": 12.284041404724121, "learning_rate": 3.857536132140399e-05, "loss": 0.9526, "step": 132800 }, { "epoch": 22.857880247763248, "grad_norm": 11.895554542541504, "learning_rate": 3.857105987611837e-05, "loss": 0.9802, "step": 132850 }, { "epoch": 22.86648313833448, "grad_norm": 12.369098663330078, "learning_rate": 3.856675843083276e-05, "loss": 0.9503, "step": 132900 }, { "epoch": 22.875086028905713, "grad_norm": 10.282855987548828, "learning_rate": 3.856245698554715e-05, "loss": 0.9509, "step": 132950 }, { "epoch": 22.883688919476946, "grad_norm": 13.339397430419922, "learning_rate": 3.855815554026153e-05, "loss": 0.9081, "step": 133000 }, { "epoch": 22.892291810048174, "grad_norm": 9.644953727722168, "learning_rate": 3.855385409497592e-05, "loss": 0.9388, "step": 133050 }, { "epoch": 22.900894700619407, "grad_norm": 13.044384002685547, "learning_rate": 3.854955264969029e-05, "loss": 0.9479, "step": 133100 }, { "epoch": 22.90949759119064, "grad_norm": 7.6776509284973145, "learning_rate": 3.854525120440468e-05, "loss": 0.9715, "step": 133150 }, { "epoch": 22.918100481761872, "grad_norm": 10.02032470703125, "learning_rate": 3.854094975911907e-05, "loss": 0.9604, "step": 133200 }, { "epoch": 22.926703372333105, "grad_norm": 9.623347282409668, "learning_rate": 3.853664831383345e-05, "loss": 0.9718, "step": 133250 }, { "epoch": 22.935306262904337, "grad_norm": 9.034906387329102, "learning_rate": 3.853234686854784e-05, "loss": 0.953, "step": 133300 }, { "epoch": 22.943909153475566, "grad_norm": 7.347438335418701, "learning_rate": 3.852804542326222e-05, "loss": 0.9163, "step": 133350 }, { "epoch": 22.9525120440468, "grad_norm": 6.912693500518799, "learning_rate": 3.85237439779766e-05, "loss": 0.9008, "step": 133400 }, { "epoch": 22.96111493461803, "grad_norm": 18.0524959564209, "learning_rate": 3.851944253269099e-05, "loss": 0.8904, "step": 133450 }, { "epoch": 22.969717825189264, "grad_norm": 8.974712371826172, "learning_rate": 3.851514108740537e-05, "loss": 0.9158, "step": 133500 }, { "epoch": 22.978320715760496, "grad_norm": 6.812099933624268, "learning_rate": 3.851083964211976e-05, "loss": 0.903, "step": 133550 }, { "epoch": 22.98692360633173, "grad_norm": 10.875104904174805, "learning_rate": 3.850653819683414e-05, "loss": 0.951, "step": 133600 }, { "epoch": 22.995526496902958, "grad_norm": 10.592730522155762, "learning_rate": 3.850223675154852e-05, "loss": 0.9655, "step": 133650 }, { "epoch": 23.0, "eval_accuracy": 0.6582343830665979, "eval_f1": 0.6479564484261263, "eval_f1_DuraRiadoRio_16x16": 0.6181423139598045, "eval_f1_Mole_16x16": 0.6908529525279815, "eval_f1_Quebrado_16x16": 0.7867279632220667, "eval_f1_RiadoRio_16x16": 0.4651962188484675, "eval_f1_RioFechado_16x16": 0.6788627935723115, "eval_loss": 1.016800045967102, "eval_precision": 0.6850383700263991, "eval_precision_DuraRiadoRio_16x16": 0.5302889095992545, "eval_precision_Mole_16x16": 0.6219596942321056, "eval_precision_Quebrado_16x16": 0.729158947758429, "eval_precision_RiadoRio_16x16": 0.7667610953729934, "eval_precision_RioFechado_16x16": 0.7770232031692134, "eval_recall": 0.661713014038425, "eval_recall_DuraRiadoRio_16x16": 0.7408854166666666, "eval_recall_Mole_16x16": 0.7769097222222222, "eval_recall_Quebrado_16x16": 0.8541666666666666, "eval_recall_RiadoRio_16x16": 0.3338815789473684, "eval_recall_RioFechado_16x16": 0.602721685689201, "eval_runtime": 46.0279, "eval_samples_per_second": 252.499, "eval_steps_per_second": 15.795, "step": 133676 }, { "epoch": 23.00412938747419, "grad_norm": 14.895306587219238, "learning_rate": 3.849793530626291e-05, "loss": 0.9024, "step": 133700 }, { "epoch": 23.012732278045423, "grad_norm": 8.155150413513184, "learning_rate": 3.849363386097729e-05, "loss": 0.9267, "step": 133750 }, { "epoch": 23.021335168616655, "grad_norm": 8.232610702514648, "learning_rate": 3.8489332415691677e-05, "loss": 0.9597, "step": 133800 }, { "epoch": 23.029938059187888, "grad_norm": 9.232503890991211, "learning_rate": 3.848503097040606e-05, "loss": 0.948, "step": 133850 }, { "epoch": 23.03854094975912, "grad_norm": 13.48752212524414, "learning_rate": 3.848072952512044e-05, "loss": 0.9227, "step": 133900 }, { "epoch": 23.04714384033035, "grad_norm": 9.571722984313965, "learning_rate": 3.847642807983483e-05, "loss": 0.9121, "step": 133950 }, { "epoch": 23.055746730901582, "grad_norm": 9.100560188293457, "learning_rate": 3.847212663454921e-05, "loss": 0.9431, "step": 134000 }, { "epoch": 23.064349621472815, "grad_norm": 10.167379379272461, "learning_rate": 3.8467825189263596e-05, "loss": 0.9068, "step": 134050 }, { "epoch": 23.072952512044047, "grad_norm": 9.734572410583496, "learning_rate": 3.846352374397798e-05, "loss": 0.9162, "step": 134100 }, { "epoch": 23.08155540261528, "grad_norm": 11.49625015258789, "learning_rate": 3.845922229869236e-05, "loss": 0.9518, "step": 134150 }, { "epoch": 23.090158293186512, "grad_norm": 8.816681861877441, "learning_rate": 3.8454920853406746e-05, "loss": 0.9094, "step": 134200 }, { "epoch": 23.09876118375774, "grad_norm": 6.748520374298096, "learning_rate": 3.845061940812113e-05, "loss": 0.8925, "step": 134250 }, { "epoch": 23.107364074328974, "grad_norm": 10.287398338317871, "learning_rate": 3.8446317962835515e-05, "loss": 0.9286, "step": 134300 }, { "epoch": 23.115966964900206, "grad_norm": 8.67827033996582, "learning_rate": 3.84420165175499e-05, "loss": 0.8535, "step": 134350 }, { "epoch": 23.12456985547144, "grad_norm": 12.788808822631836, "learning_rate": 3.8437715072264285e-05, "loss": 0.8915, "step": 134400 }, { "epoch": 23.13317274604267, "grad_norm": 9.494373321533203, "learning_rate": 3.8433413626978666e-05, "loss": 0.8873, "step": 134450 }, { "epoch": 23.141775636613904, "grad_norm": 8.737595558166504, "learning_rate": 3.842911218169305e-05, "loss": 0.9255, "step": 134500 }, { "epoch": 23.150378527185133, "grad_norm": 14.482037544250488, "learning_rate": 3.8424810736407435e-05, "loss": 0.919, "step": 134550 }, { "epoch": 23.158981417756365, "grad_norm": 10.224617004394531, "learning_rate": 3.842050929112182e-05, "loss": 0.9348, "step": 134600 }, { "epoch": 23.167584308327598, "grad_norm": 11.524622917175293, "learning_rate": 3.8416207845836204e-05, "loss": 0.9556, "step": 134650 }, { "epoch": 23.17618719889883, "grad_norm": 10.096221923828125, "learning_rate": 3.8411906400550585e-05, "loss": 0.9228, "step": 134700 }, { "epoch": 23.184790089470063, "grad_norm": 10.244658470153809, "learning_rate": 3.840760495526497e-05, "loss": 0.9212, "step": 134750 }, { "epoch": 23.193392980041295, "grad_norm": 8.127251625061035, "learning_rate": 3.8403303509979354e-05, "loss": 0.9104, "step": 134800 }, { "epoch": 23.201995870612524, "grad_norm": 12.333660125732422, "learning_rate": 3.839900206469374e-05, "loss": 0.9006, "step": 134850 }, { "epoch": 23.210598761183757, "grad_norm": 9.94881820678711, "learning_rate": 3.839470061940812e-05, "loss": 0.9187, "step": 134900 }, { "epoch": 23.21920165175499, "grad_norm": 10.462562561035156, "learning_rate": 3.8390399174122505e-05, "loss": 0.9446, "step": 134950 }, { "epoch": 23.227804542326222, "grad_norm": 6.602960586547852, "learning_rate": 3.838609772883689e-05, "loss": 0.9425, "step": 135000 }, { "epoch": 23.236407432897455, "grad_norm": 7.699405193328857, "learning_rate": 3.8381796283551274e-05, "loss": 0.9422, "step": 135050 }, { "epoch": 23.245010323468687, "grad_norm": 6.477211952209473, "learning_rate": 3.837749483826566e-05, "loss": 0.9318, "step": 135100 }, { "epoch": 23.253613214039916, "grad_norm": 12.36583423614502, "learning_rate": 3.837319339298004e-05, "loss": 0.9195, "step": 135150 }, { "epoch": 23.26221610461115, "grad_norm": 8.348941802978516, "learning_rate": 3.836889194769443e-05, "loss": 0.912, "step": 135200 }, { "epoch": 23.27081899518238, "grad_norm": 7.254903316497803, "learning_rate": 3.836459050240881e-05, "loss": 0.9223, "step": 135250 }, { "epoch": 23.279421885753614, "grad_norm": 9.76523494720459, "learning_rate": 3.836028905712319e-05, "loss": 0.9089, "step": 135300 }, { "epoch": 23.288024776324846, "grad_norm": 8.894597053527832, "learning_rate": 3.835598761183758e-05, "loss": 0.9174, "step": 135350 }, { "epoch": 23.29662766689608, "grad_norm": 8.18835735321045, "learning_rate": 3.835168616655196e-05, "loss": 0.93, "step": 135400 }, { "epoch": 23.305230557467308, "grad_norm": 12.229753494262695, "learning_rate": 3.834738472126635e-05, "loss": 0.9083, "step": 135450 }, { "epoch": 23.31383344803854, "grad_norm": 10.266366958618164, "learning_rate": 3.834308327598073e-05, "loss": 0.9962, "step": 135500 }, { "epoch": 23.322436338609773, "grad_norm": 7.549715995788574, "learning_rate": 3.833878183069511e-05, "loss": 0.9423, "step": 135550 }, { "epoch": 23.331039229181005, "grad_norm": 13.027995109558105, "learning_rate": 3.83344803854095e-05, "loss": 0.8719, "step": 135600 }, { "epoch": 23.339642119752238, "grad_norm": 9.548795700073242, "learning_rate": 3.833017894012388e-05, "loss": 0.8689, "step": 135650 }, { "epoch": 23.348245010323467, "grad_norm": 10.246988296508789, "learning_rate": 3.832587749483827e-05, "loss": 0.9622, "step": 135700 }, { "epoch": 23.3568479008947, "grad_norm": 11.722721099853516, "learning_rate": 3.832157604955265e-05, "loss": 0.9257, "step": 135750 }, { "epoch": 23.365450791465932, "grad_norm": 11.12967300415039, "learning_rate": 3.831727460426703e-05, "loss": 0.9591, "step": 135800 }, { "epoch": 23.374053682037164, "grad_norm": 11.0017671585083, "learning_rate": 3.831297315898142e-05, "loss": 0.9469, "step": 135850 }, { "epoch": 23.382656572608397, "grad_norm": 8.450535774230957, "learning_rate": 3.83086717136958e-05, "loss": 0.9262, "step": 135900 }, { "epoch": 23.39125946317963, "grad_norm": 14.036128044128418, "learning_rate": 3.830437026841019e-05, "loss": 0.9717, "step": 135950 }, { "epoch": 23.39986235375086, "grad_norm": 7.460312843322754, "learning_rate": 3.830006882312458e-05, "loss": 0.8911, "step": 136000 }, { "epoch": 23.40846524432209, "grad_norm": 11.926862716674805, "learning_rate": 3.829576737783895e-05, "loss": 0.9934, "step": 136050 }, { "epoch": 23.417068134893324, "grad_norm": 8.09646224975586, "learning_rate": 3.829146593255334e-05, "loss": 0.9277, "step": 136100 }, { "epoch": 23.425671025464556, "grad_norm": 9.074581146240234, "learning_rate": 3.828716448726772e-05, "loss": 0.9636, "step": 136150 }, { "epoch": 23.43427391603579, "grad_norm": 10.36101245880127, "learning_rate": 3.828286304198211e-05, "loss": 0.9161, "step": 136200 }, { "epoch": 23.44287680660702, "grad_norm": 5.837637424468994, "learning_rate": 3.8278561596696496e-05, "loss": 0.9551, "step": 136250 }, { "epoch": 23.45147969717825, "grad_norm": 12.712016105651855, "learning_rate": 3.827426015141087e-05, "loss": 0.9748, "step": 136300 }, { "epoch": 23.460082587749483, "grad_norm": 9.861677169799805, "learning_rate": 3.826995870612526e-05, "loss": 0.9082, "step": 136350 }, { "epoch": 23.468685478320715, "grad_norm": 11.202341079711914, "learning_rate": 3.8265657260839647e-05, "loss": 0.9397, "step": 136400 }, { "epoch": 23.477288368891948, "grad_norm": 11.622251510620117, "learning_rate": 3.826135581555403e-05, "loss": 0.926, "step": 136450 }, { "epoch": 23.48589125946318, "grad_norm": 12.289475440979004, "learning_rate": 3.8257054370268416e-05, "loss": 0.8885, "step": 136500 }, { "epoch": 23.494494150034413, "grad_norm": 10.288081169128418, "learning_rate": 3.82527529249828e-05, "loss": 0.9532, "step": 136550 }, { "epoch": 23.503097040605642, "grad_norm": 9.302433013916016, "learning_rate": 3.824845147969718e-05, "loss": 0.9135, "step": 136600 }, { "epoch": 23.511699931176874, "grad_norm": 8.012349128723145, "learning_rate": 3.8244150034411566e-05, "loss": 0.8661, "step": 136650 }, { "epoch": 23.520302821748107, "grad_norm": 9.659966468811035, "learning_rate": 3.823984858912595e-05, "loss": 0.8627, "step": 136700 }, { "epoch": 23.52890571231934, "grad_norm": 14.830811500549316, "learning_rate": 3.8235547143840335e-05, "loss": 1.0032, "step": 136750 }, { "epoch": 23.537508602890572, "grad_norm": 7.502387046813965, "learning_rate": 3.8231245698554716e-05, "loss": 0.9637, "step": 136800 }, { "epoch": 23.546111493461805, "grad_norm": 8.461153984069824, "learning_rate": 3.82269442532691e-05, "loss": 0.8644, "step": 136850 }, { "epoch": 23.554714384033034, "grad_norm": 10.116989135742188, "learning_rate": 3.8222642807983485e-05, "loss": 0.928, "step": 136900 }, { "epoch": 23.563317274604266, "grad_norm": 9.224021911621094, "learning_rate": 3.8218341362697867e-05, "loss": 0.9296, "step": 136950 }, { "epoch": 23.5719201651755, "grad_norm": 7.695443153381348, "learning_rate": 3.8214039917412255e-05, "loss": 0.9027, "step": 137000 }, { "epoch": 23.58052305574673, "grad_norm": 11.763172149658203, "learning_rate": 3.8209738472126636e-05, "loss": 0.9553, "step": 137050 }, { "epoch": 23.589125946317964, "grad_norm": 8.543238639831543, "learning_rate": 3.820543702684102e-05, "loss": 0.9488, "step": 137100 }, { "epoch": 23.597728836889196, "grad_norm": 8.5350341796875, "learning_rate": 3.8201135581555405e-05, "loss": 0.9673, "step": 137150 }, { "epoch": 23.606331727460425, "grad_norm": 12.366891860961914, "learning_rate": 3.8196834136269786e-05, "loss": 0.955, "step": 137200 }, { "epoch": 23.614934618031658, "grad_norm": 11.986250877380371, "learning_rate": 3.8192532690984174e-05, "loss": 0.8672, "step": 137250 }, { "epoch": 23.62353750860289, "grad_norm": 8.80652904510498, "learning_rate": 3.8188231245698555e-05, "loss": 0.9245, "step": 137300 }, { "epoch": 23.632140399174123, "grad_norm": 9.740911483764648, "learning_rate": 3.818392980041294e-05, "loss": 0.9731, "step": 137350 }, { "epoch": 23.640743289745355, "grad_norm": 11.168083190917969, "learning_rate": 3.8179628355127324e-05, "loss": 0.887, "step": 137400 }, { "epoch": 23.649346180316588, "grad_norm": 7.387316703796387, "learning_rate": 3.8175326909841705e-05, "loss": 0.8563, "step": 137450 }, { "epoch": 23.657949070887817, "grad_norm": 11.214729309082031, "learning_rate": 3.817102546455609e-05, "loss": 0.9133, "step": 137500 }, { "epoch": 23.66655196145905, "grad_norm": 13.835205078125, "learning_rate": 3.8166724019270475e-05, "loss": 0.9817, "step": 137550 }, { "epoch": 23.675154852030282, "grad_norm": 8.699366569519043, "learning_rate": 3.816242257398486e-05, "loss": 0.9119, "step": 137600 }, { "epoch": 23.683757742601514, "grad_norm": 9.165700912475586, "learning_rate": 3.8158121128699244e-05, "loss": 0.9364, "step": 137650 }, { "epoch": 23.692360633172747, "grad_norm": 6.82261323928833, "learning_rate": 3.8153819683413625e-05, "loss": 0.9139, "step": 137700 }, { "epoch": 23.70096352374398, "grad_norm": 8.067182540893555, "learning_rate": 3.814951823812801e-05, "loss": 0.932, "step": 137750 }, { "epoch": 23.70956641431521, "grad_norm": 8.449584007263184, "learning_rate": 3.81452167928424e-05, "loss": 0.9359, "step": 137800 }, { "epoch": 23.71816930488644, "grad_norm": 10.884464263916016, "learning_rate": 3.814091534755678e-05, "loss": 0.9233, "step": 137850 }, { "epoch": 23.726772195457674, "grad_norm": 10.3787841796875, "learning_rate": 3.813661390227117e-05, "loss": 0.9357, "step": 137900 }, { "epoch": 23.735375086028906, "grad_norm": 14.175910949707031, "learning_rate": 3.8132312456985544e-05, "loss": 0.8842, "step": 137950 }, { "epoch": 23.74397797660014, "grad_norm": 14.676584243774414, "learning_rate": 3.812801101169993e-05, "loss": 0.9578, "step": 138000 }, { "epoch": 23.75258086717137, "grad_norm": 9.637797355651855, "learning_rate": 3.812370956641432e-05, "loss": 0.9267, "step": 138050 }, { "epoch": 23.7611837577426, "grad_norm": 8.018956184387207, "learning_rate": 3.81194081211287e-05, "loss": 0.977, "step": 138100 }, { "epoch": 23.769786648313833, "grad_norm": 9.62653923034668, "learning_rate": 3.811510667584309e-05, "loss": 0.9816, "step": 138150 }, { "epoch": 23.778389538885065, "grad_norm": 10.324944496154785, "learning_rate": 3.8110805230557464e-05, "loss": 0.8955, "step": 138200 }, { "epoch": 23.786992429456298, "grad_norm": 9.100518226623535, "learning_rate": 3.810650378527185e-05, "loss": 0.9575, "step": 138250 }, { "epoch": 23.79559532002753, "grad_norm": 9.752341270446777, "learning_rate": 3.810220233998624e-05, "loss": 0.9329, "step": 138300 }, { "epoch": 23.804198210598763, "grad_norm": 11.364791870117188, "learning_rate": 3.809790089470062e-05, "loss": 0.9488, "step": 138350 }, { "epoch": 23.812801101169992, "grad_norm": 8.073541641235352, "learning_rate": 3.809359944941501e-05, "loss": 0.9638, "step": 138400 }, { "epoch": 23.821403991741224, "grad_norm": 11.974804878234863, "learning_rate": 3.808929800412939e-05, "loss": 0.9437, "step": 138450 }, { "epoch": 23.830006882312457, "grad_norm": 14.356040000915527, "learning_rate": 3.808499655884377e-05, "loss": 0.8708, "step": 138500 }, { "epoch": 23.83860977288369, "grad_norm": 16.075292587280273, "learning_rate": 3.808069511355816e-05, "loss": 0.9433, "step": 138550 }, { "epoch": 23.847212663454922, "grad_norm": 10.37588119506836, "learning_rate": 3.807639366827254e-05, "loss": 0.8987, "step": 138600 }, { "epoch": 23.855815554026155, "grad_norm": 8.777589797973633, "learning_rate": 3.807209222298693e-05, "loss": 0.9516, "step": 138650 }, { "epoch": 23.864418444597383, "grad_norm": 10.456183433532715, "learning_rate": 3.806779077770131e-05, "loss": 0.8864, "step": 138700 }, { "epoch": 23.873021335168616, "grad_norm": 5.042362213134766, "learning_rate": 3.806348933241569e-05, "loss": 0.8857, "step": 138750 }, { "epoch": 23.88162422573985, "grad_norm": 11.314642906188965, "learning_rate": 3.805918788713008e-05, "loss": 0.9097, "step": 138800 }, { "epoch": 23.89022711631108, "grad_norm": 9.700013160705566, "learning_rate": 3.805488644184446e-05, "loss": 0.9188, "step": 138850 }, { "epoch": 23.898830006882314, "grad_norm": 10.57367992401123, "learning_rate": 3.805058499655885e-05, "loss": 0.8996, "step": 138900 }, { "epoch": 23.907432897453546, "grad_norm": 10.423626899719238, "learning_rate": 3.804628355127323e-05, "loss": 0.9083, "step": 138950 }, { "epoch": 23.916035788024775, "grad_norm": 10.90226936340332, "learning_rate": 3.804198210598761e-05, "loss": 0.9028, "step": 139000 }, { "epoch": 23.924638678596008, "grad_norm": 11.211145401000977, "learning_rate": 3.8037680660702e-05, "loss": 0.9386, "step": 139050 }, { "epoch": 23.93324156916724, "grad_norm": 14.998205184936523, "learning_rate": 3.803337921541638e-05, "loss": 0.9091, "step": 139100 }, { "epoch": 23.941844459738473, "grad_norm": 8.953941345214844, "learning_rate": 3.802907777013077e-05, "loss": 0.9086, "step": 139150 }, { "epoch": 23.950447350309705, "grad_norm": 8.615900039672852, "learning_rate": 3.8024776324845155e-05, "loss": 0.951, "step": 139200 }, { "epoch": 23.959050240880934, "grad_norm": 10.494682312011719, "learning_rate": 3.802047487955953e-05, "loss": 0.921, "step": 139250 }, { "epoch": 23.967653131452167, "grad_norm": 10.957459449768066, "learning_rate": 3.801617343427392e-05, "loss": 0.9609, "step": 139300 }, { "epoch": 23.9762560220234, "grad_norm": 12.052117347717285, "learning_rate": 3.80118719889883e-05, "loss": 0.9144, "step": 139350 }, { "epoch": 23.984858912594632, "grad_norm": 14.63453197479248, "learning_rate": 3.8007570543702686e-05, "loss": 0.9227, "step": 139400 }, { "epoch": 23.993461803165864, "grad_norm": 13.122603416442871, "learning_rate": 3.8003269098417074e-05, "loss": 0.9555, "step": 139450 }, { "epoch": 24.0, "eval_accuracy": 0.5775253828945104, "eval_f1": 0.54693692413263, "eval_f1_DuraRiadoRio_16x16": 0.3595206391478029, "eval_f1_Mole_16x16": 0.6426123765369885, "eval_f1_Quebrado_16x16": 0.7149387456573414, "eval_f1_RiadoRio_16x16": 0.41669032074936135, "eval_f1_RioFechado_16x16": 0.6009225385716558, "eval_loss": 1.3300529718399048, "eval_precision": 0.6266027825861328, "eval_precision_DuraRiadoRio_16x16": 0.7714285714285715, "eval_precision_Mole_16x16": 0.5999247271358675, "eval_precision_Quebrado_16x16": 0.617693522906793, "eval_precision_RiadoRio_16x16": 0.6727772685609532, "eval_precision_RioFechado_16x16": 0.47118982289847844, "eval_recall": 0.5811569931880844, "eval_recall_DuraRiadoRio_16x16": 0.234375, "eval_recall_Mole_16x16": 0.6918402777777778, "eval_recall_Quebrado_16x16": 0.8485243055555556, "eval_recall_RiadoRio_16x16": 0.30180921052631576, "eval_recall_RioFechado_16x16": 0.8292361720807726, "eval_runtime": 46.3547, "eval_samples_per_second": 250.719, "eval_steps_per_second": 15.683, "step": 139488 }, { "epoch": 24.002064693737097, "grad_norm": 8.340795516967773, "learning_rate": 3.7998967653131455e-05, "loss": 0.9547, "step": 139500 }, { "epoch": 24.010667584308326, "grad_norm": 10.860660552978516, "learning_rate": 3.799466620784584e-05, "loss": 0.8931, "step": 139550 }, { "epoch": 24.01927047487956, "grad_norm": 9.272323608398438, "learning_rate": 3.799036476256022e-05, "loss": 0.9586, "step": 139600 }, { "epoch": 24.02787336545079, "grad_norm": 8.084892272949219, "learning_rate": 3.7986063317274606e-05, "loss": 0.9175, "step": 139650 }, { "epoch": 24.036476256022024, "grad_norm": 8.117597579956055, "learning_rate": 3.7981761871988994e-05, "loss": 0.9594, "step": 139700 }, { "epoch": 24.045079146593256, "grad_norm": 15.724629402160645, "learning_rate": 3.7977460426703375e-05, "loss": 0.8868, "step": 139750 }, { "epoch": 24.05368203716449, "grad_norm": 12.188080787658691, "learning_rate": 3.7973158981417756e-05, "loss": 0.9279, "step": 139800 }, { "epoch": 24.062284927735718, "grad_norm": 9.223344802856445, "learning_rate": 3.7968857536132144e-05, "loss": 0.8824, "step": 139850 }, { "epoch": 24.07088781830695, "grad_norm": 13.997721672058105, "learning_rate": 3.7964556090846525e-05, "loss": 0.9192, "step": 139900 }, { "epoch": 24.079490708878183, "grad_norm": 9.624273300170898, "learning_rate": 3.796025464556091e-05, "loss": 0.8846, "step": 139950 }, { "epoch": 24.088093599449415, "grad_norm": 7.635201454162598, "learning_rate": 3.7955953200275294e-05, "loss": 0.8988, "step": 140000 }, { "epoch": 24.096696490020648, "grad_norm": 10.257547378540039, "learning_rate": 3.795165175498968e-05, "loss": 0.8888, "step": 140050 }, { "epoch": 24.10529938059188, "grad_norm": 14.43117904663086, "learning_rate": 3.7947350309704063e-05, "loss": 0.902, "step": 140100 }, { "epoch": 24.11390227116311, "grad_norm": 7.624350547790527, "learning_rate": 3.7943048864418445e-05, "loss": 0.8935, "step": 140150 }, { "epoch": 24.122505161734342, "grad_norm": 7.781637668609619, "learning_rate": 3.793874741913283e-05, "loss": 0.9315, "step": 140200 }, { "epoch": 24.131108052305574, "grad_norm": 9.290365219116211, "learning_rate": 3.7934445973847214e-05, "loss": 0.9472, "step": 140250 }, { "epoch": 24.139710942876807, "grad_norm": 17.702199935913086, "learning_rate": 3.79301445285616e-05, "loss": 0.9754, "step": 140300 }, { "epoch": 24.14831383344804, "grad_norm": 11.779065132141113, "learning_rate": 3.792584308327598e-05, "loss": 0.9061, "step": 140350 }, { "epoch": 24.156916724019272, "grad_norm": 9.94943904876709, "learning_rate": 3.7921541637990364e-05, "loss": 0.8715, "step": 140400 }, { "epoch": 24.1655196145905, "grad_norm": 12.426412582397461, "learning_rate": 3.791724019270475e-05, "loss": 0.8735, "step": 140450 }, { "epoch": 24.174122505161733, "grad_norm": 10.793440818786621, "learning_rate": 3.791293874741913e-05, "loss": 0.9442, "step": 140500 }, { "epoch": 24.182725395732966, "grad_norm": 9.697668075561523, "learning_rate": 3.790863730213352e-05, "loss": 0.9082, "step": 140550 }, { "epoch": 24.1913282863042, "grad_norm": 15.752410888671875, "learning_rate": 3.79043358568479e-05, "loss": 0.8967, "step": 140600 }, { "epoch": 24.19993117687543, "grad_norm": 8.891989707946777, "learning_rate": 3.7900034411562283e-05, "loss": 0.8434, "step": 140650 }, { "epoch": 24.208534067446664, "grad_norm": 11.3290433883667, "learning_rate": 3.789573296627667e-05, "loss": 0.9228, "step": 140700 }, { "epoch": 24.217136958017893, "grad_norm": 13.787455558776855, "learning_rate": 3.789143152099105e-05, "loss": 0.8706, "step": 140750 }, { "epoch": 24.225739848589125, "grad_norm": 8.554631233215332, "learning_rate": 3.788713007570544e-05, "loss": 0.8643, "step": 140800 }, { "epoch": 24.234342739160358, "grad_norm": 8.921427726745605, "learning_rate": 3.788282863041983e-05, "loss": 0.8515, "step": 140850 }, { "epoch": 24.24294562973159, "grad_norm": 8.27197551727295, "learning_rate": 3.78785271851342e-05, "loss": 0.9112, "step": 140900 }, { "epoch": 24.251548520302823, "grad_norm": 9.141945838928223, "learning_rate": 3.787422573984859e-05, "loss": 0.9711, "step": 140950 }, { "epoch": 24.260151410874055, "grad_norm": 8.464612007141113, "learning_rate": 3.786992429456297e-05, "loss": 0.9571, "step": 141000 }, { "epoch": 24.268754301445284, "grad_norm": 11.67750358581543, "learning_rate": 3.786562284927736e-05, "loss": 0.9171, "step": 141050 }, { "epoch": 24.277357192016517, "grad_norm": 9.675850868225098, "learning_rate": 3.786132140399175e-05, "loss": 0.9317, "step": 141100 }, { "epoch": 24.28596008258775, "grad_norm": 12.122220993041992, "learning_rate": 3.785701995870612e-05, "loss": 0.8874, "step": 141150 }, { "epoch": 24.294562973158982, "grad_norm": 12.603663444519043, "learning_rate": 3.785271851342051e-05, "loss": 0.9083, "step": 141200 }, { "epoch": 24.303165863730214, "grad_norm": 8.356821060180664, "learning_rate": 3.78484170681349e-05, "loss": 0.9337, "step": 141250 }, { "epoch": 24.311768754301447, "grad_norm": 8.173026084899902, "learning_rate": 3.784411562284928e-05, "loss": 0.8894, "step": 141300 }, { "epoch": 24.320371644872676, "grad_norm": 8.665340423583984, "learning_rate": 3.783981417756367e-05, "loss": 0.9038, "step": 141350 }, { "epoch": 24.32897453544391, "grad_norm": 10.357563018798828, "learning_rate": 3.783551273227804e-05, "loss": 0.9272, "step": 141400 }, { "epoch": 24.33757742601514, "grad_norm": 9.425708770751953, "learning_rate": 3.783121128699243e-05, "loss": 0.8713, "step": 141450 }, { "epoch": 24.346180316586373, "grad_norm": 8.697186470031738, "learning_rate": 3.782690984170682e-05, "loss": 0.9439, "step": 141500 }, { "epoch": 24.354783207157606, "grad_norm": 10.255887985229492, "learning_rate": 3.78226083964212e-05, "loss": 0.8633, "step": 141550 }, { "epoch": 24.36338609772884, "grad_norm": 12.739855766296387, "learning_rate": 3.781830695113559e-05, "loss": 0.9271, "step": 141600 }, { "epoch": 24.371988988300068, "grad_norm": 9.49609375, "learning_rate": 3.781400550584997e-05, "loss": 0.9152, "step": 141650 }, { "epoch": 24.3805918788713, "grad_norm": 9.771905899047852, "learning_rate": 3.780970406056435e-05, "loss": 0.9472, "step": 141700 }, { "epoch": 24.389194769442533, "grad_norm": 9.009293556213379, "learning_rate": 3.780540261527874e-05, "loss": 0.9098, "step": 141750 }, { "epoch": 24.397797660013765, "grad_norm": 10.386629104614258, "learning_rate": 3.780110116999312e-05, "loss": 0.9303, "step": 141800 }, { "epoch": 24.406400550584998, "grad_norm": 11.944914817810059, "learning_rate": 3.7796799724707506e-05, "loss": 0.8961, "step": 141850 }, { "epoch": 24.41500344115623, "grad_norm": 11.29998779296875, "learning_rate": 3.779249827942189e-05, "loss": 0.9157, "step": 141900 }, { "epoch": 24.42360633172746, "grad_norm": 7.472931861877441, "learning_rate": 3.778819683413627e-05, "loss": 0.8872, "step": 141950 }, { "epoch": 24.43220922229869, "grad_norm": 7.031731605529785, "learning_rate": 3.7783895388850656e-05, "loss": 0.9093, "step": 142000 }, { "epoch": 24.440812112869924, "grad_norm": 9.085681915283203, "learning_rate": 3.777959394356504e-05, "loss": 1.0102, "step": 142050 }, { "epoch": 24.449415003441157, "grad_norm": 7.760104656219482, "learning_rate": 3.7775292498279426e-05, "loss": 0.9706, "step": 142100 }, { "epoch": 24.45801789401239, "grad_norm": 17.422828674316406, "learning_rate": 3.777099105299381e-05, "loss": 0.9186, "step": 142150 }, { "epoch": 24.46662078458362, "grad_norm": 7.091302394866943, "learning_rate": 3.7766689607708195e-05, "loss": 0.9137, "step": 142200 }, { "epoch": 24.47522367515485, "grad_norm": 12.595724105834961, "learning_rate": 3.7762388162422576e-05, "loss": 0.8846, "step": 142250 }, { "epoch": 24.483826565726083, "grad_norm": 10.492859840393066, "learning_rate": 3.775808671713696e-05, "loss": 0.9814, "step": 142300 }, { "epoch": 24.492429456297316, "grad_norm": 9.367167472839355, "learning_rate": 3.7753785271851345e-05, "loss": 0.8597, "step": 142350 }, { "epoch": 24.50103234686855, "grad_norm": 9.357158660888672, "learning_rate": 3.7749483826565726e-05, "loss": 0.9558, "step": 142400 }, { "epoch": 24.50963523743978, "grad_norm": 12.066889762878418, "learning_rate": 3.7745182381280114e-05, "loss": 0.9553, "step": 142450 }, { "epoch": 24.51823812801101, "grad_norm": 7.649158477783203, "learning_rate": 3.7740880935994495e-05, "loss": 0.8703, "step": 142500 }, { "epoch": 24.526841018582243, "grad_norm": 11.34870719909668, "learning_rate": 3.7736579490708876e-05, "loss": 0.8741, "step": 142550 }, { "epoch": 24.535443909153475, "grad_norm": 13.663084030151367, "learning_rate": 3.7732278045423264e-05, "loss": 0.8732, "step": 142600 }, { "epoch": 24.544046799724708, "grad_norm": 9.40065860748291, "learning_rate": 3.772797660013765e-05, "loss": 0.9709, "step": 142650 }, { "epoch": 24.55264969029594, "grad_norm": 11.207592964172363, "learning_rate": 3.7723675154852033e-05, "loss": 0.9323, "step": 142700 }, { "epoch": 24.561252580867173, "grad_norm": 15.729707717895508, "learning_rate": 3.7719373709566415e-05, "loss": 0.9498, "step": 142750 }, { "epoch": 24.5698554714384, "grad_norm": 9.64367961883545, "learning_rate": 3.7715072264280796e-05, "loss": 0.976, "step": 142800 }, { "epoch": 24.578458362009634, "grad_norm": 7.452882289886475, "learning_rate": 3.7710770818995184e-05, "loss": 0.9305, "step": 142850 }, { "epoch": 24.587061252580867, "grad_norm": 14.15317440032959, "learning_rate": 3.770646937370957e-05, "loss": 0.8963, "step": 142900 }, { "epoch": 24.5956641431521, "grad_norm": 11.624592781066895, "learning_rate": 3.770216792842395e-05, "loss": 0.9052, "step": 142950 }, { "epoch": 24.604267033723332, "grad_norm": 11.002046585083008, "learning_rate": 3.769786648313834e-05, "loss": 0.8697, "step": 143000 }, { "epoch": 24.612869924294564, "grad_norm": 6.9384002685546875, "learning_rate": 3.7693565037852715e-05, "loss": 0.8985, "step": 143050 }, { "epoch": 24.621472814865793, "grad_norm": 8.897109031677246, "learning_rate": 3.76892635925671e-05, "loss": 0.959, "step": 143100 }, { "epoch": 24.630075705437026, "grad_norm": 11.747153282165527, "learning_rate": 3.768496214728149e-05, "loss": 0.899, "step": 143150 }, { "epoch": 24.63867859600826, "grad_norm": 11.745481491088867, "learning_rate": 3.768066070199587e-05, "loss": 0.9245, "step": 143200 }, { "epoch": 24.64728148657949, "grad_norm": 11.661211967468262, "learning_rate": 3.767635925671026e-05, "loss": 0.9321, "step": 143250 }, { "epoch": 24.655884377150723, "grad_norm": 9.963964462280273, "learning_rate": 3.767205781142464e-05, "loss": 0.9049, "step": 143300 }, { "epoch": 24.664487267721956, "grad_norm": 9.32172679901123, "learning_rate": 3.766775636613902e-05, "loss": 0.9274, "step": 143350 }, { "epoch": 24.673090158293185, "grad_norm": 7.1860785484313965, "learning_rate": 3.766345492085341e-05, "loss": 0.9428, "step": 143400 }, { "epoch": 24.681693048864418, "grad_norm": 14.743624687194824, "learning_rate": 3.765915347556779e-05, "loss": 0.8769, "step": 143450 }, { "epoch": 24.69029593943565, "grad_norm": 11.926392555236816, "learning_rate": 3.765485203028218e-05, "loss": 0.8919, "step": 143500 }, { "epoch": 24.698898830006883, "grad_norm": 7.04101037979126, "learning_rate": 3.765055058499656e-05, "loss": 0.9235, "step": 143550 }, { "epoch": 24.707501720578115, "grad_norm": 15.299920082092285, "learning_rate": 3.764624913971094e-05, "loss": 0.9225, "step": 143600 }, { "epoch": 24.716104611149348, "grad_norm": 26.10154914855957, "learning_rate": 3.764194769442533e-05, "loss": 0.9384, "step": 143650 }, { "epoch": 24.724707501720577, "grad_norm": 10.54284954071045, "learning_rate": 3.763764624913971e-05, "loss": 0.9398, "step": 143700 }, { "epoch": 24.73331039229181, "grad_norm": 13.635038375854492, "learning_rate": 3.76333448038541e-05, "loss": 0.9102, "step": 143750 }, { "epoch": 24.74191328286304, "grad_norm": 9.202703475952148, "learning_rate": 3.762904335856848e-05, "loss": 0.9477, "step": 143800 }, { "epoch": 24.750516173434274, "grad_norm": 13.46757984161377, "learning_rate": 3.762474191328286e-05, "loss": 0.9164, "step": 143850 }, { "epoch": 24.759119064005507, "grad_norm": 11.796211242675781, "learning_rate": 3.762044046799725e-05, "loss": 0.9248, "step": 143900 }, { "epoch": 24.76772195457674, "grad_norm": 8.8829345703125, "learning_rate": 3.761613902271163e-05, "loss": 0.9293, "step": 143950 }, { "epoch": 24.77632484514797, "grad_norm": 8.812234878540039, "learning_rate": 3.761183757742602e-05, "loss": 0.9392, "step": 144000 }, { "epoch": 24.7849277357192, "grad_norm": 8.036493301391602, "learning_rate": 3.76075361321404e-05, "loss": 0.9, "step": 144050 }, { "epoch": 24.793530626290433, "grad_norm": 10.532805442810059, "learning_rate": 3.760323468685478e-05, "loss": 0.9289, "step": 144100 }, { "epoch": 24.802133516861666, "grad_norm": 13.541930198669434, "learning_rate": 3.759893324156917e-05, "loss": 0.9304, "step": 144150 }, { "epoch": 24.8107364074329, "grad_norm": 6.406330585479736, "learning_rate": 3.759463179628355e-05, "loss": 0.9319, "step": 144200 }, { "epoch": 24.81933929800413, "grad_norm": 10.32689380645752, "learning_rate": 3.759033035099794e-05, "loss": 0.8823, "step": 144250 }, { "epoch": 24.82794218857536, "grad_norm": 10.687844276428223, "learning_rate": 3.7586028905712326e-05, "loss": 0.953, "step": 144300 }, { "epoch": 24.836545079146592, "grad_norm": 18.702486038208008, "learning_rate": 3.758172746042671e-05, "loss": 0.9467, "step": 144350 }, { "epoch": 24.845147969717825, "grad_norm": 14.088067054748535, "learning_rate": 3.757742601514109e-05, "loss": 0.9231, "step": 144400 }, { "epoch": 24.853750860289058, "grad_norm": 9.328703880310059, "learning_rate": 3.757312456985547e-05, "loss": 0.9131, "step": 144450 }, { "epoch": 24.86235375086029, "grad_norm": 5.744831562042236, "learning_rate": 3.756882312456986e-05, "loss": 0.888, "step": 144500 }, { "epoch": 24.870956641431523, "grad_norm": 8.849129676818848, "learning_rate": 3.7564521679284245e-05, "loss": 0.9531, "step": 144550 }, { "epoch": 24.87955953200275, "grad_norm": 9.945921897888184, "learning_rate": 3.7560220233998626e-05, "loss": 0.8947, "step": 144600 }, { "epoch": 24.888162422573984, "grad_norm": 12.873255729675293, "learning_rate": 3.755591878871301e-05, "loss": 0.9574, "step": 144650 }, { "epoch": 24.896765313145217, "grad_norm": 9.286680221557617, "learning_rate": 3.755161734342739e-05, "loss": 0.8709, "step": 144700 }, { "epoch": 24.90536820371645, "grad_norm": 13.795872688293457, "learning_rate": 3.754731589814178e-05, "loss": 0.8958, "step": 144750 }, { "epoch": 24.91397109428768, "grad_norm": 12.989541053771973, "learning_rate": 3.7543014452856165e-05, "loss": 0.9841, "step": 144800 }, { "epoch": 24.922573984858914, "grad_norm": 8.717784881591797, "learning_rate": 3.7538713007570546e-05, "loss": 0.9038, "step": 144850 }, { "epoch": 24.931176875430143, "grad_norm": 13.537964820861816, "learning_rate": 3.753441156228493e-05, "loss": 0.9873, "step": 144900 }, { "epoch": 24.939779766001376, "grad_norm": 10.116386413574219, "learning_rate": 3.7530110116999315e-05, "loss": 0.951, "step": 144950 }, { "epoch": 24.94838265657261, "grad_norm": 8.86430549621582, "learning_rate": 3.7525808671713696e-05, "loss": 0.8996, "step": 145000 }, { "epoch": 24.95698554714384, "grad_norm": 12.672709465026855, "learning_rate": 3.7521507226428084e-05, "loss": 0.8292, "step": 145050 }, { "epoch": 24.965588437715073, "grad_norm": 11.787375450134277, "learning_rate": 3.7517205781142465e-05, "loss": 0.8967, "step": 145100 }, { "epoch": 24.974191328286302, "grad_norm": 11.519380569458008, "learning_rate": 3.751290433585685e-05, "loss": 0.872, "step": 145150 }, { "epoch": 24.982794218857535, "grad_norm": 15.951248168945312, "learning_rate": 3.7508602890571234e-05, "loss": 0.9581, "step": 145200 }, { "epoch": 24.991397109428767, "grad_norm": 13.818696022033691, "learning_rate": 3.7504301445285616e-05, "loss": 0.919, "step": 145250 }, { "epoch": 25.0, "grad_norm": 27.183576583862305, "learning_rate": 3.7500000000000003e-05, "loss": 0.9223, "step": 145300 }, { "epoch": 25.0, "eval_accuracy": 0.6484254001032524, "eval_f1": 0.6480692080859816, "eval_f1_DuraRiadoRio_16x16": 0.6291376765760067, "eval_f1_Mole_16x16": 0.6905721192586624, "eval_f1_Quebrado_16x16": 0.739386296763346, "eval_f1_RiadoRio_16x16": 0.5221088435374149, "eval_f1_RioFechado_16x16": 0.6591411042944785, "eval_loss": 0.9976791739463806, "eval_precision": 0.6521451197878377, "eval_precision_DuraRiadoRio_16x16": 0.6117261172611727, "eval_precision_Mole_16x16": 0.6443609022556391, "eval_precision_Quebrado_16x16": 0.7167889160554197, "eval_precision_RiadoRio_16x16": 0.5404929577464789, "eval_precision_RioFechado_16x16": 0.7473567056204786, "eval_recall": 0.6498868731997905, "eval_recall_DuraRiadoRio_16x16": 0.6475694444444444, "eval_recall_Mole_16x16": 0.7439236111111112, "eval_recall_Quebrado_16x16": 0.7634548611111112, "eval_recall_RiadoRio_16x16": 0.5049342105263158, "eval_recall_RioFechado_16x16": 0.5895522388059702, "eval_runtime": 46.7744, "eval_samples_per_second": 248.469, "eval_steps_per_second": 15.543, "step": 145300 }, { "epoch": 25.008602890571233, "grad_norm": 13.38866138458252, "learning_rate": 3.7495698554714385e-05, "loss": 0.9337, "step": 145350 }, { "epoch": 25.017205781142465, "grad_norm": 10.061627388000488, "learning_rate": 3.749139710942877e-05, "loss": 0.9044, "step": 145400 }, { "epoch": 25.025808671713694, "grad_norm": 9.122878074645996, "learning_rate": 3.7487095664143154e-05, "loss": 0.8768, "step": 145450 }, { "epoch": 25.034411562284927, "grad_norm": 9.449451446533203, "learning_rate": 3.7482794218857535e-05, "loss": 0.9321, "step": 145500 }, { "epoch": 25.04301445285616, "grad_norm": 8.051238059997559, "learning_rate": 3.747849277357192e-05, "loss": 0.9522, "step": 145550 }, { "epoch": 25.05161734342739, "grad_norm": 14.002348899841309, "learning_rate": 3.7474191328286304e-05, "loss": 0.9199, "step": 145600 }, { "epoch": 25.060220233998624, "grad_norm": 8.674985885620117, "learning_rate": 3.746988988300069e-05, "loss": 0.9066, "step": 145650 }, { "epoch": 25.068823124569857, "grad_norm": 7.8529791831970215, "learning_rate": 3.746558843771508e-05, "loss": 0.8377, "step": 145700 }, { "epoch": 25.077426015141086, "grad_norm": 7.174069881439209, "learning_rate": 3.7461286992429454e-05, "loss": 0.9087, "step": 145750 }, { "epoch": 25.08602890571232, "grad_norm": 9.702253341674805, "learning_rate": 3.745698554714384e-05, "loss": 0.9664, "step": 145800 }, { "epoch": 25.09463179628355, "grad_norm": 10.831591606140137, "learning_rate": 3.7452684101858223e-05, "loss": 0.872, "step": 145850 }, { "epoch": 25.103234686854783, "grad_norm": 9.077827453613281, "learning_rate": 3.744838265657261e-05, "loss": 0.8972, "step": 145900 }, { "epoch": 25.111837577426016, "grad_norm": 7.602989196777344, "learning_rate": 3.7444081211287e-05, "loss": 0.9133, "step": 145950 }, { "epoch": 25.12044046799725, "grad_norm": 9.815960884094238, "learning_rate": 3.7439779766001374e-05, "loss": 0.914, "step": 146000 }, { "epoch": 25.129043358568477, "grad_norm": 8.829757690429688, "learning_rate": 3.743547832071576e-05, "loss": 0.8988, "step": 146050 }, { "epoch": 25.13764624913971, "grad_norm": 8.054925918579102, "learning_rate": 3.743117687543014e-05, "loss": 0.8907, "step": 146100 }, { "epoch": 25.146249139710942, "grad_norm": 7.4178690910339355, "learning_rate": 3.742687543014453e-05, "loss": 0.9024, "step": 146150 }, { "epoch": 25.154852030282175, "grad_norm": 15.080707550048828, "learning_rate": 3.742257398485892e-05, "loss": 0.9305, "step": 146200 }, { "epoch": 25.163454920853408, "grad_norm": 9.612289428710938, "learning_rate": 3.741827253957329e-05, "loss": 0.9497, "step": 146250 }, { "epoch": 25.17205781142464, "grad_norm": 8.960163116455078, "learning_rate": 3.741397109428768e-05, "loss": 0.9416, "step": 146300 }, { "epoch": 25.18066070199587, "grad_norm": 10.878570556640625, "learning_rate": 3.740966964900207e-05, "loss": 0.9232, "step": 146350 }, { "epoch": 25.1892635925671, "grad_norm": 7.559167861938477, "learning_rate": 3.740536820371645e-05, "loss": 0.8955, "step": 146400 }, { "epoch": 25.197866483138334, "grad_norm": 10.139278411865234, "learning_rate": 3.740106675843084e-05, "loss": 0.9076, "step": 146450 }, { "epoch": 25.206469373709567, "grad_norm": 11.40548324584961, "learning_rate": 3.739676531314522e-05, "loss": 0.9485, "step": 146500 }, { "epoch": 25.2150722642808, "grad_norm": 12.426542282104492, "learning_rate": 3.73924638678596e-05, "loss": 0.9339, "step": 146550 }, { "epoch": 25.22367515485203, "grad_norm": 10.384832382202148, "learning_rate": 3.738816242257399e-05, "loss": 0.9099, "step": 146600 }, { "epoch": 25.23227804542326, "grad_norm": 10.781933784484863, "learning_rate": 3.738386097728837e-05, "loss": 0.8756, "step": 146650 }, { "epoch": 25.240880935994493, "grad_norm": 12.382231712341309, "learning_rate": 3.737955953200276e-05, "loss": 0.8836, "step": 146700 }, { "epoch": 25.249483826565726, "grad_norm": 11.761810302734375, "learning_rate": 3.737525808671714e-05, "loss": 0.9571, "step": 146750 }, { "epoch": 25.25808671713696, "grad_norm": 8.153233528137207, "learning_rate": 3.737095664143152e-05, "loss": 0.947, "step": 146800 }, { "epoch": 25.26668960770819, "grad_norm": 12.1353120803833, "learning_rate": 3.736665519614591e-05, "loss": 0.8942, "step": 146850 }, { "epoch": 25.275292498279423, "grad_norm": 8.295010566711426, "learning_rate": 3.736235375086029e-05, "loss": 0.8903, "step": 146900 }, { "epoch": 25.283895388850652, "grad_norm": 10.891989707946777, "learning_rate": 3.735805230557468e-05, "loss": 0.9323, "step": 146950 }, { "epoch": 25.292498279421885, "grad_norm": 13.445262908935547, "learning_rate": 3.735375086028906e-05, "loss": 0.8773, "step": 147000 }, { "epoch": 25.301101169993117, "grad_norm": 6.344038486480713, "learning_rate": 3.734944941500344e-05, "loss": 0.9424, "step": 147050 }, { "epoch": 25.30970406056435, "grad_norm": 8.760103225708008, "learning_rate": 3.734514796971783e-05, "loss": 0.8887, "step": 147100 }, { "epoch": 25.318306951135582, "grad_norm": 7.32329797744751, "learning_rate": 3.734084652443221e-05, "loss": 0.8883, "step": 147150 }, { "epoch": 25.326909841706815, "grad_norm": 8.409423828125, "learning_rate": 3.7336545079146596e-05, "loss": 0.8787, "step": 147200 }, { "epoch": 25.335512732278044, "grad_norm": 7.220482349395752, "learning_rate": 3.733224363386098e-05, "loss": 0.8939, "step": 147250 }, { "epoch": 25.344115622849277, "grad_norm": 9.54015064239502, "learning_rate": 3.7327942188575366e-05, "loss": 0.9756, "step": 147300 }, { "epoch": 25.35271851342051, "grad_norm": 9.637151718139648, "learning_rate": 3.732364074328975e-05, "loss": 0.915, "step": 147350 }, { "epoch": 25.36132140399174, "grad_norm": 9.65040111541748, "learning_rate": 3.731933929800413e-05, "loss": 0.8868, "step": 147400 }, { "epoch": 25.369924294562974, "grad_norm": 7.0348944664001465, "learning_rate": 3.7315037852718516e-05, "loss": 0.846, "step": 147450 }, { "epoch": 25.378527185134207, "grad_norm": 14.153687477111816, "learning_rate": 3.73107364074329e-05, "loss": 0.9115, "step": 147500 }, { "epoch": 25.387130075705436, "grad_norm": 15.891119003295898, "learning_rate": 3.7306434962147285e-05, "loss": 0.9254, "step": 147550 }, { "epoch": 25.395732966276668, "grad_norm": 9.778393745422363, "learning_rate": 3.7302133516861666e-05, "loss": 0.9045, "step": 147600 }, { "epoch": 25.4043358568479, "grad_norm": 6.2017741203308105, "learning_rate": 3.729783207157605e-05, "loss": 0.8843, "step": 147650 }, { "epoch": 25.412938747419133, "grad_norm": 12.440858840942383, "learning_rate": 3.7293530626290435e-05, "loss": 0.8699, "step": 147700 }, { "epoch": 25.421541637990366, "grad_norm": 7.209077835083008, "learning_rate": 3.728922918100482e-05, "loss": 0.9562, "step": 147750 }, { "epoch": 25.4301445285616, "grad_norm": 11.642709732055664, "learning_rate": 3.7284927735719204e-05, "loss": 0.927, "step": 147800 }, { "epoch": 25.438747419132827, "grad_norm": 9.97142219543457, "learning_rate": 3.728062629043359e-05, "loss": 0.924, "step": 147850 }, { "epoch": 25.44735030970406, "grad_norm": 7.572685241699219, "learning_rate": 3.727632484514797e-05, "loss": 0.8703, "step": 147900 }, { "epoch": 25.455953200275292, "grad_norm": 9.818136215209961, "learning_rate": 3.7272023399862355e-05, "loss": 0.8659, "step": 147950 }, { "epoch": 25.464556090846525, "grad_norm": 10.649150848388672, "learning_rate": 3.726772195457674e-05, "loss": 0.8379, "step": 148000 }, { "epoch": 25.473158981417757, "grad_norm": 10.903157234191895, "learning_rate": 3.7263420509291124e-05, "loss": 0.8573, "step": 148050 }, { "epoch": 25.48176187198899, "grad_norm": 12.040077209472656, "learning_rate": 3.725911906400551e-05, "loss": 0.9124, "step": 148100 }, { "epoch": 25.49036476256022, "grad_norm": 13.389047622680664, "learning_rate": 3.7254817618719886e-05, "loss": 0.9222, "step": 148150 }, { "epoch": 25.49896765313145, "grad_norm": 12.148234367370605, "learning_rate": 3.7250516173434274e-05, "loss": 0.8337, "step": 148200 }, { "epoch": 25.507570543702684, "grad_norm": 10.940281867980957, "learning_rate": 3.724621472814866e-05, "loss": 0.9492, "step": 148250 }, { "epoch": 25.516173434273917, "grad_norm": 10.226743698120117, "learning_rate": 3.724191328286304e-05, "loss": 0.9206, "step": 148300 }, { "epoch": 25.52477632484515, "grad_norm": 9.229767799377441, "learning_rate": 3.723761183757743e-05, "loss": 0.9225, "step": 148350 }, { "epoch": 25.53337921541638, "grad_norm": 8.7374267578125, "learning_rate": 3.723331039229181e-05, "loss": 0.9082, "step": 148400 }, { "epoch": 25.54198210598761, "grad_norm": 12.559805870056152, "learning_rate": 3.7229008947006194e-05, "loss": 0.9063, "step": 148450 }, { "epoch": 25.550584996558843, "grad_norm": 10.638786315917969, "learning_rate": 3.722470750172058e-05, "loss": 0.9527, "step": 148500 }, { "epoch": 25.559187887130076, "grad_norm": 12.312384605407715, "learning_rate": 3.722040605643496e-05, "loss": 0.9465, "step": 148550 }, { "epoch": 25.56779077770131, "grad_norm": 13.155719757080078, "learning_rate": 3.721610461114935e-05, "loss": 0.8837, "step": 148600 }, { "epoch": 25.57639366827254, "grad_norm": 13.403854370117188, "learning_rate": 3.721180316586373e-05, "loss": 0.8963, "step": 148650 }, { "epoch": 25.58499655884377, "grad_norm": 10.854264259338379, "learning_rate": 3.720750172057811e-05, "loss": 0.9347, "step": 148700 }, { "epoch": 25.593599449415002, "grad_norm": 13.782352447509766, "learning_rate": 3.72032002752925e-05, "loss": 0.9359, "step": 148750 }, { "epoch": 25.602202339986235, "grad_norm": 6.851571083068848, "learning_rate": 3.719889883000688e-05, "loss": 0.8657, "step": 148800 }, { "epoch": 25.610805230557467, "grad_norm": 7.1159539222717285, "learning_rate": 3.719459738472127e-05, "loss": 0.8738, "step": 148850 }, { "epoch": 25.6194081211287, "grad_norm": 6.396842956542969, "learning_rate": 3.719029593943565e-05, "loss": 0.8893, "step": 148900 }, { "epoch": 25.628011011699932, "grad_norm": 9.560259819030762, "learning_rate": 3.718599449415003e-05, "loss": 0.9681, "step": 148950 }, { "epoch": 25.63661390227116, "grad_norm": 7.272044658660889, "learning_rate": 3.718169304886442e-05, "loss": 0.9198, "step": 149000 }, { "epoch": 25.645216792842394, "grad_norm": 11.352949142456055, "learning_rate": 3.71773916035788e-05, "loss": 0.9561, "step": 149050 }, { "epoch": 25.653819683413627, "grad_norm": 9.513941764831543, "learning_rate": 3.717309015829319e-05, "loss": 0.8767, "step": 149100 }, { "epoch": 25.66242257398486, "grad_norm": 8.280470848083496, "learning_rate": 3.716878871300758e-05, "loss": 0.9237, "step": 149150 }, { "epoch": 25.67102546455609, "grad_norm": 9.154404640197754, "learning_rate": 3.716448726772196e-05, "loss": 0.9195, "step": 149200 }, { "epoch": 25.679628355127324, "grad_norm": 9.486564636230469, "learning_rate": 3.716018582243634e-05, "loss": 0.9614, "step": 149250 }, { "epoch": 25.688231245698553, "grad_norm": 10.951675415039062, "learning_rate": 3.715588437715072e-05, "loss": 0.8865, "step": 149300 }, { "epoch": 25.696834136269786, "grad_norm": 14.078728675842285, "learning_rate": 3.715158293186511e-05, "loss": 0.9109, "step": 149350 }, { "epoch": 25.705437026841018, "grad_norm": 9.551496505737305, "learning_rate": 3.71472814865795e-05, "loss": 0.8829, "step": 149400 }, { "epoch": 25.71403991741225, "grad_norm": 10.811774253845215, "learning_rate": 3.714298004129388e-05, "loss": 0.9062, "step": 149450 }, { "epoch": 25.722642807983483, "grad_norm": 11.25102710723877, "learning_rate": 3.713867859600826e-05, "loss": 0.9097, "step": 149500 }, { "epoch": 25.731245698554716, "grad_norm": 10.145870208740234, "learning_rate": 3.713437715072264e-05, "loss": 0.9118, "step": 149550 }, { "epoch": 25.739848589125945, "grad_norm": 13.593633651733398, "learning_rate": 3.713007570543703e-05, "loss": 0.9193, "step": 149600 }, { "epoch": 25.748451479697177, "grad_norm": 9.45044231414795, "learning_rate": 3.7125774260151416e-05, "loss": 0.9215, "step": 149650 }, { "epoch": 25.75705437026841, "grad_norm": 11.37908935546875, "learning_rate": 3.71214728148658e-05, "loss": 0.9096, "step": 149700 }, { "epoch": 25.765657260839642, "grad_norm": 9.452600479125977, "learning_rate": 3.711717136958018e-05, "loss": 0.9004, "step": 149750 }, { "epoch": 25.774260151410875, "grad_norm": 9.162858009338379, "learning_rate": 3.7112869924294566e-05, "loss": 0.8589, "step": 149800 }, { "epoch": 25.782863041982107, "grad_norm": 13.534106254577637, "learning_rate": 3.710856847900895e-05, "loss": 0.9169, "step": 149850 }, { "epoch": 25.791465932553336, "grad_norm": 9.969512939453125, "learning_rate": 3.7104267033723336e-05, "loss": 0.9422, "step": 149900 }, { "epoch": 25.80006882312457, "grad_norm": 12.672919273376465, "learning_rate": 3.709996558843772e-05, "loss": 0.8753, "step": 149950 }, { "epoch": 25.8086717136958, "grad_norm": 14.76850700378418, "learning_rate": 3.7095664143152105e-05, "loss": 0.9105, "step": 150000 }, { "epoch": 25.817274604267034, "grad_norm": 8.424470901489258, "learning_rate": 3.7091362697866486e-05, "loss": 0.9264, "step": 150050 }, { "epoch": 25.825877494838267, "grad_norm": 11.622244834899902, "learning_rate": 3.708706125258087e-05, "loss": 0.952, "step": 150100 }, { "epoch": 25.8344803854095, "grad_norm": 14.809576034545898, "learning_rate": 3.7082759807295255e-05, "loss": 0.9102, "step": 150150 }, { "epoch": 25.843083275980728, "grad_norm": 9.117555618286133, "learning_rate": 3.7078458362009636e-05, "loss": 0.906, "step": 150200 }, { "epoch": 25.85168616655196, "grad_norm": 11.7745361328125, "learning_rate": 3.7074156916724024e-05, "loss": 0.9315, "step": 150250 }, { "epoch": 25.860289057123193, "grad_norm": 10.720389366149902, "learning_rate": 3.7069855471438405e-05, "loss": 0.8784, "step": 150300 }, { "epoch": 25.868891947694426, "grad_norm": 8.434577941894531, "learning_rate": 3.7065554026152786e-05, "loss": 0.9709, "step": 150350 }, { "epoch": 25.877494838265658, "grad_norm": 5.848171710968018, "learning_rate": 3.7061252580867174e-05, "loss": 0.9422, "step": 150400 }, { "epoch": 25.88609772883689, "grad_norm": 12.245061874389648, "learning_rate": 3.7056951135581556e-05, "loss": 0.8874, "step": 150450 }, { "epoch": 25.89470061940812, "grad_norm": 10.500883102416992, "learning_rate": 3.7052649690295944e-05, "loss": 0.9206, "step": 150500 }, { "epoch": 25.903303509979352, "grad_norm": 10.666751861572266, "learning_rate": 3.7048348245010325e-05, "loss": 0.8738, "step": 150550 }, { "epoch": 25.911906400550585, "grad_norm": 8.557937622070312, "learning_rate": 3.7044046799724706e-05, "loss": 0.8944, "step": 150600 }, { "epoch": 25.920509291121817, "grad_norm": 9.67363452911377, "learning_rate": 3.7039745354439094e-05, "loss": 0.9654, "step": 150650 }, { "epoch": 25.92911218169305, "grad_norm": 11.063363075256348, "learning_rate": 3.7035443909153475e-05, "loss": 0.9334, "step": 150700 }, { "epoch": 25.937715072264282, "grad_norm": 9.158751487731934, "learning_rate": 3.703114246386786e-05, "loss": 0.913, "step": 150750 }, { "epoch": 25.94631796283551, "grad_norm": 12.214499473571777, "learning_rate": 3.702684101858225e-05, "loss": 0.883, "step": 150800 }, { "epoch": 25.954920853406744, "grad_norm": 9.296566009521484, "learning_rate": 3.7022539573296625e-05, "loss": 0.932, "step": 150850 }, { "epoch": 25.963523743977976, "grad_norm": 8.056695938110352, "learning_rate": 3.701823812801101e-05, "loss": 0.901, "step": 150900 }, { "epoch": 25.97212663454921, "grad_norm": 13.331932067871094, "learning_rate": 3.7013936682725394e-05, "loss": 0.9583, "step": 150950 }, { "epoch": 25.98072952512044, "grad_norm": 11.876636505126953, "learning_rate": 3.700963523743978e-05, "loss": 0.9352, "step": 151000 }, { "epoch": 25.989332415691674, "grad_norm": 14.361923217773438, "learning_rate": 3.700533379215417e-05, "loss": 0.9324, "step": 151050 }, { "epoch": 25.997935306262903, "grad_norm": 10.564800262451172, "learning_rate": 3.7001032346868545e-05, "loss": 0.9173, "step": 151100 }, { "epoch": 26.0, "eval_accuracy": 0.6089313371192566, "eval_f1": 0.6063579480078742, "eval_f1_DuraRiadoRio_16x16": 0.5210727969348659, "eval_f1_Mole_16x16": 0.6736259054111632, "eval_f1_Quebrado_16x16": 0.6957878315132605, "eval_f1_RiadoRio_16x16": 0.5077740602243653, "eval_f1_RioFechado_16x16": 0.6335291459557162, "eval_loss": 1.1395978927612305, "eval_precision": 0.6132115677461462, "eval_precision_DuraRiadoRio_16x16": 0.633147113594041, "eval_precision_Mole_16x16": 0.6615062761506276, "eval_precision_Quebrado_16x16": 0.6317280453257791, "eval_precision_RiadoRio_16x16": 0.4869762174405436, "eval_precision_RioFechado_16x16": 0.6527001862197392, "eval_recall": 0.6098183176288321, "eval_recall_DuraRiadoRio_16x16": 0.4427083333333333, "eval_recall_Mole_16x16": 0.6861979166666666, "eval_recall_Quebrado_16x16": 0.7743055555555556, "eval_recall_RiadoRio_16x16": 0.5304276315789473, "eval_recall_RioFechado_16x16": 0.6154521510096576, "eval_runtime": 46.1535, "eval_samples_per_second": 251.812, "eval_steps_per_second": 15.752, "step": 151112 }, { "epoch": 26.006538196834136, "grad_norm": 7.775354862213135, "learning_rate": 3.699673090158293e-05, "loss": 0.9014, "step": 151150 }, { "epoch": 26.015141087405368, "grad_norm": 9.015837669372559, "learning_rate": 3.699242945629732e-05, "loss": 0.9124, "step": 151200 }, { "epoch": 26.0237439779766, "grad_norm": 10.395739555358887, "learning_rate": 3.69881280110117e-05, "loss": 0.8757, "step": 151250 }, { "epoch": 26.032346868547833, "grad_norm": 8.623872756958008, "learning_rate": 3.698382656572609e-05, "loss": 0.8431, "step": 151300 }, { "epoch": 26.040949759119066, "grad_norm": 15.613020896911621, "learning_rate": 3.697952512044047e-05, "loss": 0.877, "step": 151350 }, { "epoch": 26.049552649690295, "grad_norm": 10.29453182220459, "learning_rate": 3.697522367515485e-05, "loss": 0.8874, "step": 151400 }, { "epoch": 26.058155540261527, "grad_norm": 7.572412014007568, "learning_rate": 3.697092222986924e-05, "loss": 0.8387, "step": 151450 }, { "epoch": 26.06675843083276, "grad_norm": 15.688015937805176, "learning_rate": 3.696662078458362e-05, "loss": 0.9403, "step": 151500 }, { "epoch": 26.075361321403992, "grad_norm": 9.723471641540527, "learning_rate": 3.696231933929801e-05, "loss": 0.9045, "step": 151550 }, { "epoch": 26.083964211975225, "grad_norm": 13.034832000732422, "learning_rate": 3.695801789401239e-05, "loss": 0.8976, "step": 151600 }, { "epoch": 26.092567102546454, "grad_norm": 8.994726181030273, "learning_rate": 3.695371644872677e-05, "loss": 0.9184, "step": 151650 }, { "epoch": 26.101169993117686, "grad_norm": 8.773737907409668, "learning_rate": 3.694941500344116e-05, "loss": 0.9027, "step": 151700 }, { "epoch": 26.10977288368892, "grad_norm": 10.642932891845703, "learning_rate": 3.694511355815554e-05, "loss": 0.9115, "step": 151750 }, { "epoch": 26.11837577426015, "grad_norm": 10.535770416259766, "learning_rate": 3.694081211286993e-05, "loss": 0.8923, "step": 151800 }, { "epoch": 26.126978664831384, "grad_norm": 8.373555183410645, "learning_rate": 3.693651066758431e-05, "loss": 0.9323, "step": 151850 }, { "epoch": 26.135581555402617, "grad_norm": 11.308286666870117, "learning_rate": 3.693220922229869e-05, "loss": 0.908, "step": 151900 }, { "epoch": 26.144184445973845, "grad_norm": 10.794553756713867, "learning_rate": 3.692790777701308e-05, "loss": 0.8877, "step": 151950 }, { "epoch": 26.152787336545078, "grad_norm": 7.324456691741943, "learning_rate": 3.692360633172746e-05, "loss": 0.9042, "step": 152000 }, { "epoch": 26.16139022711631, "grad_norm": 8.430316925048828, "learning_rate": 3.691930488644185e-05, "loss": 0.9262, "step": 152050 }, { "epoch": 26.169993117687543, "grad_norm": 10.305166244506836, "learning_rate": 3.691500344115623e-05, "loss": 0.9129, "step": 152100 }, { "epoch": 26.178596008258776, "grad_norm": 8.778353691101074, "learning_rate": 3.691070199587062e-05, "loss": 0.8974, "step": 152150 }, { "epoch": 26.187198898830008, "grad_norm": 10.56392765045166, "learning_rate": 3.6906400550585e-05, "loss": 0.9856, "step": 152200 }, { "epoch": 26.195801789401237, "grad_norm": 11.934979438781738, "learning_rate": 3.690209910529938e-05, "loss": 0.9197, "step": 152250 }, { "epoch": 26.20440467997247, "grad_norm": 16.750957489013672, "learning_rate": 3.689779766001377e-05, "loss": 0.916, "step": 152300 }, { "epoch": 26.213007570543702, "grad_norm": 8.737133026123047, "learning_rate": 3.689349621472815e-05, "loss": 0.8907, "step": 152350 }, { "epoch": 26.221610461114935, "grad_norm": 10.56751823425293, "learning_rate": 3.6889194769442537e-05, "loss": 0.9315, "step": 152400 }, { "epoch": 26.230213351686167, "grad_norm": 10.943368911743164, "learning_rate": 3.688489332415692e-05, "loss": 0.8979, "step": 152450 }, { "epoch": 26.2388162422574, "grad_norm": 17.707345962524414, "learning_rate": 3.68805918788713e-05, "loss": 0.8832, "step": 152500 }, { "epoch": 26.24741913282863, "grad_norm": 8.913859367370605, "learning_rate": 3.687629043358569e-05, "loss": 0.8905, "step": 152550 }, { "epoch": 26.25602202339986, "grad_norm": 12.466801643371582, "learning_rate": 3.687198898830007e-05, "loss": 0.8787, "step": 152600 }, { "epoch": 26.264624913971094, "grad_norm": 8.501591682434082, "learning_rate": 3.6867687543014456e-05, "loss": 0.8819, "step": 152650 }, { "epoch": 26.273227804542326, "grad_norm": 11.302335739135742, "learning_rate": 3.686338609772884e-05, "loss": 0.9394, "step": 152700 }, { "epoch": 26.28183069511356, "grad_norm": 7.216952323913574, "learning_rate": 3.685908465244322e-05, "loss": 0.8523, "step": 152750 }, { "epoch": 26.29043358568479, "grad_norm": 9.291361808776855, "learning_rate": 3.6854783207157606e-05, "loss": 0.9309, "step": 152800 }, { "epoch": 26.29903647625602, "grad_norm": 11.12073802947998, "learning_rate": 3.6850481761871994e-05, "loss": 0.8728, "step": 152850 }, { "epoch": 26.307639366827253, "grad_norm": 11.319964408874512, "learning_rate": 3.6846180316586375e-05, "loss": 0.9033, "step": 152900 }, { "epoch": 26.316242257398486, "grad_norm": 10.755248069763184, "learning_rate": 3.684187887130076e-05, "loss": 0.881, "step": 152950 }, { "epoch": 26.324845147969718, "grad_norm": 10.287301063537598, "learning_rate": 3.683757742601514e-05, "loss": 0.9418, "step": 153000 }, { "epoch": 26.33344803854095, "grad_norm": 8.744412422180176, "learning_rate": 3.6833275980729526e-05, "loss": 0.8985, "step": 153050 }, { "epoch": 26.342050929112183, "grad_norm": 13.150025367736816, "learning_rate": 3.6828974535443914e-05, "loss": 0.9336, "step": 153100 }, { "epoch": 26.350653819683412, "grad_norm": 8.21713638305664, "learning_rate": 3.6824673090158295e-05, "loss": 0.9149, "step": 153150 }, { "epoch": 26.359256710254645, "grad_norm": 13.585381507873535, "learning_rate": 3.682037164487268e-05, "loss": 0.9247, "step": 153200 }, { "epoch": 26.367859600825877, "grad_norm": 11.015069961547852, "learning_rate": 3.681607019958706e-05, "loss": 0.902, "step": 153250 }, { "epoch": 26.37646249139711, "grad_norm": 7.827136039733887, "learning_rate": 3.6811768754301445e-05, "loss": 0.9167, "step": 153300 }, { "epoch": 26.385065381968342, "grad_norm": 11.21259880065918, "learning_rate": 3.680746730901583e-05, "loss": 0.9003, "step": 153350 }, { "epoch": 26.393668272539575, "grad_norm": 11.420376777648926, "learning_rate": 3.6803165863730214e-05, "loss": 0.935, "step": 153400 }, { "epoch": 26.402271163110804, "grad_norm": 11.573356628417969, "learning_rate": 3.67988644184446e-05, "loss": 0.9348, "step": 153450 }, { "epoch": 26.410874053682036, "grad_norm": 8.265669822692871, "learning_rate": 3.679456297315898e-05, "loss": 0.9239, "step": 153500 }, { "epoch": 26.41947694425327, "grad_norm": 9.955547332763672, "learning_rate": 3.6790261527873364e-05, "loss": 0.8986, "step": 153550 }, { "epoch": 26.4280798348245, "grad_norm": 7.077062129974365, "learning_rate": 3.678596008258775e-05, "loss": 0.8686, "step": 153600 }, { "epoch": 26.436682725395734, "grad_norm": 9.153634071350098, "learning_rate": 3.6781658637302134e-05, "loss": 0.8875, "step": 153650 }, { "epoch": 26.445285615966966, "grad_norm": 8.992431640625, "learning_rate": 3.677735719201652e-05, "loss": 0.9122, "step": 153700 }, { "epoch": 26.453888506538195, "grad_norm": 10.168208122253418, "learning_rate": 3.67730557467309e-05, "loss": 0.9586, "step": 153750 }, { "epoch": 26.462491397109428, "grad_norm": 7.83193302154541, "learning_rate": 3.6768754301445284e-05, "loss": 0.8832, "step": 153800 }, { "epoch": 26.47109428768066, "grad_norm": 10.111612319946289, "learning_rate": 3.676445285615967e-05, "loss": 0.8861, "step": 153850 }, { "epoch": 26.479697178251893, "grad_norm": 11.167220115661621, "learning_rate": 3.676015141087405e-05, "loss": 0.9216, "step": 153900 }, { "epoch": 26.488300068823126, "grad_norm": 10.11123275756836, "learning_rate": 3.675584996558844e-05, "loss": 0.8824, "step": 153950 }, { "epoch": 26.496902959394358, "grad_norm": 9.491327285766602, "learning_rate": 3.675154852030282e-05, "loss": 0.9313, "step": 154000 }, { "epoch": 26.505505849965587, "grad_norm": 15.761955261230469, "learning_rate": 3.67472470750172e-05, "loss": 0.901, "step": 154050 }, { "epoch": 26.51410874053682, "grad_norm": 10.628678321838379, "learning_rate": 3.674294562973159e-05, "loss": 0.9054, "step": 154100 }, { "epoch": 26.522711631108052, "grad_norm": 15.604118347167969, "learning_rate": 3.673864418444597e-05, "loss": 0.961, "step": 154150 }, { "epoch": 26.531314521679285, "grad_norm": 8.404451370239258, "learning_rate": 3.673434273916036e-05, "loss": 0.9126, "step": 154200 }, { "epoch": 26.539917412250517, "grad_norm": 9.997146606445312, "learning_rate": 3.673004129387475e-05, "loss": 0.9189, "step": 154250 }, { "epoch": 26.54852030282175, "grad_norm": 12.741114616394043, "learning_rate": 3.672573984858913e-05, "loss": 0.9578, "step": 154300 }, { "epoch": 26.55712319339298, "grad_norm": 16.49055290222168, "learning_rate": 3.672143840330351e-05, "loss": 0.8868, "step": 154350 }, { "epoch": 26.56572608396421, "grad_norm": 11.098302841186523, "learning_rate": 3.671713695801789e-05, "loss": 0.8724, "step": 154400 }, { "epoch": 26.574328974535444, "grad_norm": 9.841108322143555, "learning_rate": 3.671283551273228e-05, "loss": 0.854, "step": 154450 }, { "epoch": 26.582931865106676, "grad_norm": 10.03341007232666, "learning_rate": 3.670853406744667e-05, "loss": 0.8716, "step": 154500 }, { "epoch": 26.59153475567791, "grad_norm": 14.063738822937012, "learning_rate": 3.670423262216105e-05, "loss": 0.8845, "step": 154550 }, { "epoch": 26.600137646249138, "grad_norm": 8.049010276794434, "learning_rate": 3.669993117687543e-05, "loss": 0.8745, "step": 154600 }, { "epoch": 26.60874053682037, "grad_norm": 6.430475234985352, "learning_rate": 3.669562973158981e-05, "loss": 0.876, "step": 154650 }, { "epoch": 26.617343427391603, "grad_norm": 12.147594451904297, "learning_rate": 3.66913282863042e-05, "loss": 0.9248, "step": 154700 }, { "epoch": 26.625946317962836, "grad_norm": 8.088692665100098, "learning_rate": 3.668702684101859e-05, "loss": 0.8341, "step": 154750 }, { "epoch": 26.634549208534068, "grad_norm": 14.021026611328125, "learning_rate": 3.668272539573297e-05, "loss": 0.8758, "step": 154800 }, { "epoch": 26.6431520991053, "grad_norm": 11.571836471557617, "learning_rate": 3.6678423950447356e-05, "loss": 0.843, "step": 154850 }, { "epoch": 26.651754989676533, "grad_norm": 7.917642116546631, "learning_rate": 3.667412250516174e-05, "loss": 0.8738, "step": 154900 }, { "epoch": 26.660357880247762, "grad_norm": 8.614834785461426, "learning_rate": 3.666982105987612e-05, "loss": 0.9638, "step": 154950 }, { "epoch": 26.668960770818995, "grad_norm": 7.179855823516846, "learning_rate": 3.6665519614590507e-05, "loss": 0.891, "step": 155000 }, { "epoch": 26.677563661390227, "grad_norm": 11.354161262512207, "learning_rate": 3.666121816930489e-05, "loss": 0.8649, "step": 155050 }, { "epoch": 26.68616655196146, "grad_norm": 10.664604187011719, "learning_rate": 3.6656916724019276e-05, "loss": 0.9361, "step": 155100 }, { "epoch": 26.694769442532692, "grad_norm": 5.728518009185791, "learning_rate": 3.665261527873366e-05, "loss": 0.9127, "step": 155150 }, { "epoch": 26.70337233310392, "grad_norm": 13.307659149169922, "learning_rate": 3.664831383344804e-05, "loss": 0.9454, "step": 155200 }, { "epoch": 26.711975223675154, "grad_norm": 7.817727088928223, "learning_rate": 3.6644012388162426e-05, "loss": 0.8705, "step": 155250 }, { "epoch": 26.720578114246386, "grad_norm": 7.867726802825928, "learning_rate": 3.663971094287681e-05, "loss": 0.8694, "step": 155300 }, { "epoch": 26.72918100481762, "grad_norm": 14.517541885375977, "learning_rate": 3.6635409497591195e-05, "loss": 0.9045, "step": 155350 }, { "epoch": 26.73778389538885, "grad_norm": 7.654894828796387, "learning_rate": 3.6631108052305576e-05, "loss": 0.846, "step": 155400 }, { "epoch": 26.746386785960084, "grad_norm": 8.84024715423584, "learning_rate": 3.662680660701996e-05, "loss": 0.9599, "step": 155450 }, { "epoch": 26.754989676531313, "grad_norm": 11.220619201660156, "learning_rate": 3.6622505161734345e-05, "loss": 0.9065, "step": 155500 }, { "epoch": 26.763592567102545, "grad_norm": 11.025317192077637, "learning_rate": 3.6618203716448727e-05, "loss": 0.9518, "step": 155550 }, { "epoch": 26.772195457673778, "grad_norm": 16.474876403808594, "learning_rate": 3.6613902271163115e-05, "loss": 0.9195, "step": 155600 }, { "epoch": 26.78079834824501, "grad_norm": 14.19460391998291, "learning_rate": 3.66096008258775e-05, "loss": 0.929, "step": 155650 }, { "epoch": 26.789401238816243, "grad_norm": 12.288394927978516, "learning_rate": 3.660529938059188e-05, "loss": 0.8707, "step": 155700 }, { "epoch": 26.798004129387476, "grad_norm": 12.309657096862793, "learning_rate": 3.6600997935306265e-05, "loss": 0.9192, "step": 155750 }, { "epoch": 26.806607019958705, "grad_norm": 16.709671020507812, "learning_rate": 3.6596696490020646e-05, "loss": 0.9111, "step": 155800 }, { "epoch": 26.815209910529937, "grad_norm": 8.89453125, "learning_rate": 3.6592395044735034e-05, "loss": 0.8671, "step": 155850 }, { "epoch": 26.82381280110117, "grad_norm": 11.762892723083496, "learning_rate": 3.658809359944942e-05, "loss": 0.9338, "step": 155900 }, { "epoch": 26.832415691672402, "grad_norm": 8.471491813659668, "learning_rate": 3.6583792154163796e-05, "loss": 0.8986, "step": 155950 }, { "epoch": 26.841018582243635, "grad_norm": 9.988759994506836, "learning_rate": 3.6579490708878184e-05, "loss": 0.9285, "step": 156000 }, { "epoch": 26.849621472814867, "grad_norm": 12.624263763427734, "learning_rate": 3.6575189263592565e-05, "loss": 0.9359, "step": 156050 }, { "epoch": 26.858224363386096, "grad_norm": 13.240352630615234, "learning_rate": 3.657088781830695e-05, "loss": 0.8525, "step": 156100 }, { "epoch": 26.86682725395733, "grad_norm": 9.114359855651855, "learning_rate": 3.656658637302134e-05, "loss": 0.9159, "step": 156150 }, { "epoch": 26.87543014452856, "grad_norm": 11.557374954223633, "learning_rate": 3.6562284927735716e-05, "loss": 0.8994, "step": 156200 }, { "epoch": 26.884033035099794, "grad_norm": 11.399320602416992, "learning_rate": 3.6557983482450104e-05, "loss": 0.9144, "step": 156250 }, { "epoch": 26.892635925671026, "grad_norm": 8.348628044128418, "learning_rate": 3.655368203716449e-05, "loss": 0.9065, "step": 156300 }, { "epoch": 26.90123881624226, "grad_norm": 20.327556610107422, "learning_rate": 3.654938059187887e-05, "loss": 0.9243, "step": 156350 }, { "epoch": 26.909841706813488, "grad_norm": 9.924915313720703, "learning_rate": 3.654507914659326e-05, "loss": 0.8838, "step": 156400 }, { "epoch": 26.91844459738472, "grad_norm": 12.393834114074707, "learning_rate": 3.654077770130764e-05, "loss": 0.9027, "step": 156450 }, { "epoch": 26.927047487955953, "grad_norm": 10.720602989196777, "learning_rate": 3.653647625602202e-05, "loss": 0.8886, "step": 156500 }, { "epoch": 26.935650378527185, "grad_norm": 12.27884578704834, "learning_rate": 3.653217481073641e-05, "loss": 0.8987, "step": 156550 }, { "epoch": 26.944253269098418, "grad_norm": 14.19636058807373, "learning_rate": 3.652787336545079e-05, "loss": 0.9161, "step": 156600 }, { "epoch": 26.95285615966965, "grad_norm": 14.1400728225708, "learning_rate": 3.652357192016518e-05, "loss": 0.9227, "step": 156650 }, { "epoch": 26.96145905024088, "grad_norm": 12.369368553161621, "learning_rate": 3.651927047487956e-05, "loss": 0.9109, "step": 156700 }, { "epoch": 26.970061940812112, "grad_norm": 8.903060913085938, "learning_rate": 3.651496902959394e-05, "loss": 0.8877, "step": 156750 }, { "epoch": 26.978664831383345, "grad_norm": 16.616863250732422, "learning_rate": 3.651066758430833e-05, "loss": 0.8871, "step": 156800 }, { "epoch": 26.987267721954577, "grad_norm": 7.903749465942383, "learning_rate": 3.650636613902271e-05, "loss": 0.8751, "step": 156850 }, { "epoch": 26.99587061252581, "grad_norm": 12.44982624053955, "learning_rate": 3.65020646937371e-05, "loss": 0.9415, "step": 156900 }, { "epoch": 27.0, "eval_accuracy": 0.4913956289795216, "eval_f1": 0.4543688749038379, "eval_f1_DuraRiadoRio_16x16": 0.48223860589812334, "eval_f1_Mole_16x16": 0.1332263242375602, "eval_f1_Quebrado_16x16": 0.6864746627779803, "eval_f1_RiadoRio_16x16": 0.42370255615801705, "eval_f1_RioFechado_16x16": 0.5462022254475085, "eval_loss": 2.376081705093384, "eval_precision": 0.5752443470624581, "eval_precision_DuraRiadoRio_16x16": 0.39274017467248906, "eval_precision_Mole_16x16": 0.8829787234042553, "eval_precision_Quebrado_16x16": 0.5917661847894407, "eval_precision_RiadoRio_16x16": 0.40043923865300146, "eval_precision_RioFechado_16x16": 0.6082974137931034, "eval_recall": 0.4918669199153869, "eval_recall_DuraRiadoRio_16x16": 0.6245659722222222, "eval_recall_Mole_16x16": 0.0720486111111111, "eval_recall_Quebrado_16x16": 0.8172743055555556, "eval_recall_RiadoRio_16x16": 0.4498355263157895, "eval_recall_RioFechado_16x16": 0.49561018437225635, "eval_runtime": 45.7931, "eval_samples_per_second": 253.794, "eval_steps_per_second": 15.876, "step": 156924 }, { "epoch": 27.004473503097042, "grad_norm": 8.865287780761719, "learning_rate": 3.649776324845148e-05, "loss": 0.8693, "step": 156950 }, { "epoch": 27.01307639366827, "grad_norm": 12.612858772277832, "learning_rate": 3.649346180316587e-05, "loss": 0.8529, "step": 157000 }, { "epoch": 27.021679284239504, "grad_norm": 11.393730163574219, "learning_rate": 3.648916035788025e-05, "loss": 0.8766, "step": 157050 }, { "epoch": 27.030282174810736, "grad_norm": 13.067209243774414, "learning_rate": 3.648485891259463e-05, "loss": 0.9695, "step": 157100 }, { "epoch": 27.03888506538197, "grad_norm": 12.501884460449219, "learning_rate": 3.648055746730902e-05, "loss": 0.9081, "step": 157150 }, { "epoch": 27.0474879559532, "grad_norm": 7.396037578582764, "learning_rate": 3.64762560220234e-05, "loss": 0.8869, "step": 157200 }, { "epoch": 27.056090846524434, "grad_norm": 9.626848220825195, "learning_rate": 3.647195457673779e-05, "loss": 0.8814, "step": 157250 }, { "epoch": 27.064693737095663, "grad_norm": 11.145471572875977, "learning_rate": 3.646765313145217e-05, "loss": 0.8904, "step": 157300 }, { "epoch": 27.073296627666895, "grad_norm": 8.560053825378418, "learning_rate": 3.646335168616655e-05, "loss": 0.8807, "step": 157350 }, { "epoch": 27.081899518238128, "grad_norm": 9.514331817626953, "learning_rate": 3.645905024088094e-05, "loss": 0.9138, "step": 157400 }, { "epoch": 27.09050240880936, "grad_norm": 9.31457233428955, "learning_rate": 3.645474879559532e-05, "loss": 0.9271, "step": 157450 }, { "epoch": 27.099105299380593, "grad_norm": 11.257427215576172, "learning_rate": 3.645044735030971e-05, "loss": 0.8659, "step": 157500 }, { "epoch": 27.107708189951826, "grad_norm": 8.244112014770508, "learning_rate": 3.644614590502409e-05, "loss": 0.9126, "step": 157550 }, { "epoch": 27.116311080523054, "grad_norm": 10.12486743927002, "learning_rate": 3.644184445973847e-05, "loss": 0.8628, "step": 157600 }, { "epoch": 27.124913971094287, "grad_norm": 11.22184944152832, "learning_rate": 3.643754301445286e-05, "loss": 0.8437, "step": 157650 }, { "epoch": 27.13351686166552, "grad_norm": 14.551734924316406, "learning_rate": 3.6433241569167246e-05, "loss": 0.8695, "step": 157700 }, { "epoch": 27.142119752236752, "grad_norm": 8.143622398376465, "learning_rate": 3.642894012388163e-05, "loss": 0.8649, "step": 157750 }, { "epoch": 27.150722642807985, "grad_norm": 8.405390739440918, "learning_rate": 3.6424638678596015e-05, "loss": 0.8533, "step": 157800 }, { "epoch": 27.159325533379217, "grad_norm": 13.426033973693848, "learning_rate": 3.642033723331039e-05, "loss": 0.8925, "step": 157850 }, { "epoch": 27.167928423950446, "grad_norm": 9.135150909423828, "learning_rate": 3.641603578802478e-05, "loss": 0.8238, "step": 157900 }, { "epoch": 27.17653131452168, "grad_norm": 8.918893814086914, "learning_rate": 3.6411734342739165e-05, "loss": 0.8891, "step": 157950 }, { "epoch": 27.18513420509291, "grad_norm": 13.150062561035156, "learning_rate": 3.6407432897453546e-05, "loss": 0.9006, "step": 158000 }, { "epoch": 27.193737095664144, "grad_norm": 13.528446197509766, "learning_rate": 3.6403131452167934e-05, "loss": 0.9071, "step": 158050 }, { "epoch": 27.202339986235376, "grad_norm": 13.268595695495605, "learning_rate": 3.639883000688231e-05, "loss": 0.8933, "step": 158100 }, { "epoch": 27.210942876806605, "grad_norm": 9.235493659973145, "learning_rate": 3.6394528561596697e-05, "loss": 0.8923, "step": 158150 }, { "epoch": 27.219545767377838, "grad_norm": 8.753786087036133, "learning_rate": 3.6390227116311085e-05, "loss": 0.8955, "step": 158200 }, { "epoch": 27.22814865794907, "grad_norm": 11.56154727935791, "learning_rate": 3.6385925671025466e-05, "loss": 0.9703, "step": 158250 }, { "epoch": 27.236751548520303, "grad_norm": 8.779301643371582, "learning_rate": 3.6381624225739854e-05, "loss": 0.87, "step": 158300 }, { "epoch": 27.245354439091535, "grad_norm": 9.978301048278809, "learning_rate": 3.6377322780454235e-05, "loss": 0.8672, "step": 158350 }, { "epoch": 27.253957329662768, "grad_norm": 10.276266098022461, "learning_rate": 3.6373021335168616e-05, "loss": 0.889, "step": 158400 }, { "epoch": 27.262560220233997, "grad_norm": 15.03113079071045, "learning_rate": 3.6368719889883004e-05, "loss": 0.8634, "step": 158450 }, { "epoch": 27.27116311080523, "grad_norm": 11.556618690490723, "learning_rate": 3.6364418444597385e-05, "loss": 0.8967, "step": 158500 }, { "epoch": 27.279766001376462, "grad_norm": 8.905078887939453, "learning_rate": 3.636011699931177e-05, "loss": 0.8283, "step": 158550 }, { "epoch": 27.288368891947695, "grad_norm": 17.822914123535156, "learning_rate": 3.6355815554026154e-05, "loss": 0.9282, "step": 158600 }, { "epoch": 27.296971782518927, "grad_norm": 11.30942440032959, "learning_rate": 3.6351514108740535e-05, "loss": 0.9325, "step": 158650 }, { "epoch": 27.30557467309016, "grad_norm": 13.879469871520996, "learning_rate": 3.634721266345492e-05, "loss": 0.8728, "step": 158700 }, { "epoch": 27.31417756366139, "grad_norm": 8.824873924255371, "learning_rate": 3.6342911218169305e-05, "loss": 0.8357, "step": 158750 }, { "epoch": 27.32278045423262, "grad_norm": 14.419418334960938, "learning_rate": 3.633860977288369e-05, "loss": 0.932, "step": 158800 }, { "epoch": 27.331383344803854, "grad_norm": 6.053401947021484, "learning_rate": 3.6334308327598074e-05, "loss": 0.9196, "step": 158850 }, { "epoch": 27.339986235375086, "grad_norm": 15.099733352661133, "learning_rate": 3.6330006882312455e-05, "loss": 0.8414, "step": 158900 }, { "epoch": 27.34858912594632, "grad_norm": 8.027508735656738, "learning_rate": 3.632570543702684e-05, "loss": 0.9017, "step": 158950 }, { "epoch": 27.35719201651755, "grad_norm": 16.046005249023438, "learning_rate": 3.6321403991741224e-05, "loss": 0.9425, "step": 159000 }, { "epoch": 27.36579490708878, "grad_norm": 22.017793655395508, "learning_rate": 3.631710254645561e-05, "loss": 0.8787, "step": 159050 }, { "epoch": 27.374397797660013, "grad_norm": 7.373079299926758, "learning_rate": 3.631280110116999e-05, "loss": 0.8756, "step": 159100 }, { "epoch": 27.383000688231245, "grad_norm": 11.269092559814453, "learning_rate": 3.630849965588438e-05, "loss": 0.909, "step": 159150 }, { "epoch": 27.391603578802478, "grad_norm": 10.260244369506836, "learning_rate": 3.630419821059876e-05, "loss": 0.9507, "step": 159200 }, { "epoch": 27.40020646937371, "grad_norm": 10.70195198059082, "learning_rate": 3.629989676531314e-05, "loss": 0.9173, "step": 159250 }, { "epoch": 27.408809359944943, "grad_norm": 10.1634521484375, "learning_rate": 3.629559532002753e-05, "loss": 0.8773, "step": 159300 }, { "epoch": 27.417412250516172, "grad_norm": 9.219460487365723, "learning_rate": 3.629129387474192e-05, "loss": 0.8464, "step": 159350 }, { "epoch": 27.426015141087404, "grad_norm": 10.567719459533691, "learning_rate": 3.62869924294563e-05, "loss": 0.8954, "step": 159400 }, { "epoch": 27.434618031658637, "grad_norm": 10.759013175964355, "learning_rate": 3.628269098417068e-05, "loss": 0.8717, "step": 159450 }, { "epoch": 27.44322092222987, "grad_norm": 7.195890426635742, "learning_rate": 3.627838953888506e-05, "loss": 0.9841, "step": 159500 }, { "epoch": 27.451823812801102, "grad_norm": 11.878470420837402, "learning_rate": 3.627408809359945e-05, "loss": 0.8873, "step": 159550 }, { "epoch": 27.460426703372335, "grad_norm": 14.695958137512207, "learning_rate": 3.626978664831384e-05, "loss": 0.8738, "step": 159600 }, { "epoch": 27.469029593943564, "grad_norm": 6.9437360763549805, "learning_rate": 3.626548520302822e-05, "loss": 0.8647, "step": 159650 }, { "epoch": 27.477632484514796, "grad_norm": 10.293545722961426, "learning_rate": 3.62611837577426e-05, "loss": 0.8897, "step": 159700 }, { "epoch": 27.48623537508603, "grad_norm": 10.008435249328613, "learning_rate": 3.625688231245699e-05, "loss": 0.9054, "step": 159750 }, { "epoch": 27.49483826565726, "grad_norm": 14.944574356079102, "learning_rate": 3.625258086717137e-05, "loss": 0.9415, "step": 159800 }, { "epoch": 27.503441156228494, "grad_norm": 8.598651885986328, "learning_rate": 3.624827942188576e-05, "loss": 0.9734, "step": 159850 }, { "epoch": 27.512044046799726, "grad_norm": 7.228833198547363, "learning_rate": 3.624397797660014e-05, "loss": 0.8932, "step": 159900 }, { "epoch": 27.520646937370955, "grad_norm": 9.829763412475586, "learning_rate": 3.623967653131453e-05, "loss": 0.8888, "step": 159950 }, { "epoch": 27.529249827942188, "grad_norm": 14.084701538085938, "learning_rate": 3.623537508602891e-05, "loss": 0.9261, "step": 160000 }, { "epoch": 27.53785271851342, "grad_norm": 15.761123657226562, "learning_rate": 3.623107364074329e-05, "loss": 0.9165, "step": 160050 }, { "epoch": 27.546455609084653, "grad_norm": 10.938589096069336, "learning_rate": 3.622677219545768e-05, "loss": 0.9104, "step": 160100 }, { "epoch": 27.555058499655885, "grad_norm": 10.297532081604004, "learning_rate": 3.622247075017206e-05, "loss": 0.8998, "step": 160150 }, { "epoch": 27.563661390227118, "grad_norm": 10.901227951049805, "learning_rate": 3.621816930488645e-05, "loss": 0.9235, "step": 160200 }, { "epoch": 27.572264280798347, "grad_norm": 12.199424743652344, "learning_rate": 3.621386785960083e-05, "loss": 0.9585, "step": 160250 }, { "epoch": 27.58086717136958, "grad_norm": 11.968239784240723, "learning_rate": 3.620956641431521e-05, "loss": 0.9434, "step": 160300 }, { "epoch": 27.589470061940812, "grad_norm": 12.585365295410156, "learning_rate": 3.62052649690296e-05, "loss": 0.8964, "step": 160350 }, { "epoch": 27.598072952512045, "grad_norm": 11.995943069458008, "learning_rate": 3.620096352374398e-05, "loss": 0.9036, "step": 160400 }, { "epoch": 27.606675843083277, "grad_norm": 11.047300338745117, "learning_rate": 3.6196662078458366e-05, "loss": 0.8919, "step": 160450 }, { "epoch": 27.61527873365451, "grad_norm": 8.879595756530762, "learning_rate": 3.619236063317275e-05, "loss": 0.8998, "step": 160500 }, { "epoch": 27.62388162422574, "grad_norm": 9.668859481811523, "learning_rate": 3.618805918788713e-05, "loss": 0.9134, "step": 160550 }, { "epoch": 27.63248451479697, "grad_norm": 24.779741287231445, "learning_rate": 3.6183757742601516e-05, "loss": 0.898, "step": 160600 }, { "epoch": 27.641087405368204, "grad_norm": 6.211843013763428, "learning_rate": 3.61794562973159e-05, "loss": 0.9095, "step": 160650 }, { "epoch": 27.649690295939436, "grad_norm": 13.849813461303711, "learning_rate": 3.6175154852030285e-05, "loss": 0.8861, "step": 160700 }, { "epoch": 27.65829318651067, "grad_norm": 11.586739540100098, "learning_rate": 3.6170853406744673e-05, "loss": 0.9115, "step": 160750 }, { "epoch": 27.6668960770819, "grad_norm": 12.389143943786621, "learning_rate": 3.616655196145905e-05, "loss": 0.8947, "step": 160800 }, { "epoch": 27.67549896765313, "grad_norm": 13.945556640625, "learning_rate": 3.6162250516173436e-05, "loss": 0.8542, "step": 160850 }, { "epoch": 27.684101858224363, "grad_norm": 15.351140975952148, "learning_rate": 3.615794907088782e-05, "loss": 0.8652, "step": 160900 }, { "epoch": 27.692704748795595, "grad_norm": 8.066584587097168, "learning_rate": 3.6153647625602205e-05, "loss": 0.9056, "step": 160950 }, { "epoch": 27.701307639366828, "grad_norm": 10.37514591217041, "learning_rate": 3.614934618031659e-05, "loss": 0.9031, "step": 161000 }, { "epoch": 27.70991052993806, "grad_norm": 9.094402313232422, "learning_rate": 3.614504473503097e-05, "loss": 0.9322, "step": 161050 }, { "epoch": 27.71851342050929, "grad_norm": 6.294773101806641, "learning_rate": 3.6140743289745355e-05, "loss": 0.9057, "step": 161100 }, { "epoch": 27.727116311080522, "grad_norm": 16.59613037109375, "learning_rate": 3.6136441844459736e-05, "loss": 0.8787, "step": 161150 }, { "epoch": 27.735719201651754, "grad_norm": 11.388501167297363, "learning_rate": 3.6132140399174124e-05, "loss": 0.893, "step": 161200 }, { "epoch": 27.744322092222987, "grad_norm": 12.760761260986328, "learning_rate": 3.612783895388851e-05, "loss": 0.9037, "step": 161250 }, { "epoch": 27.75292498279422, "grad_norm": 10.854598045349121, "learning_rate": 3.6123537508602893e-05, "loss": 0.9695, "step": 161300 }, { "epoch": 27.761527873365452, "grad_norm": 7.867920398712158, "learning_rate": 3.6119236063317275e-05, "loss": 0.8478, "step": 161350 }, { "epoch": 27.77013076393668, "grad_norm": 7.520150661468506, "learning_rate": 3.611493461803166e-05, "loss": 0.9303, "step": 161400 }, { "epoch": 27.778733654507914, "grad_norm": 8.487213134765625, "learning_rate": 3.6110633172746044e-05, "loss": 0.9183, "step": 161450 }, { "epoch": 27.787336545079146, "grad_norm": 8.782894134521484, "learning_rate": 3.610633172746043e-05, "loss": 0.9023, "step": 161500 }, { "epoch": 27.79593943565038, "grad_norm": 11.873407363891602, "learning_rate": 3.610203028217481e-05, "loss": 0.9124, "step": 161550 }, { "epoch": 27.80454232622161, "grad_norm": 13.072087287902832, "learning_rate": 3.6097728836889194e-05, "loss": 0.916, "step": 161600 }, { "epoch": 27.813145216792844, "grad_norm": 13.63879680633545, "learning_rate": 3.609342739160358e-05, "loss": 0.8907, "step": 161650 }, { "epoch": 27.821748107364073, "grad_norm": 11.307513236999512, "learning_rate": 3.608912594631796e-05, "loss": 0.863, "step": 161700 }, { "epoch": 27.830350997935305, "grad_norm": 10.395702362060547, "learning_rate": 3.608482450103235e-05, "loss": 0.9197, "step": 161750 }, { "epoch": 27.838953888506538, "grad_norm": 11.415350914001465, "learning_rate": 3.608052305574673e-05, "loss": 0.9292, "step": 161800 }, { "epoch": 27.84755677907777, "grad_norm": 12.75048828125, "learning_rate": 3.6076221610461113e-05, "loss": 0.9004, "step": 161850 }, { "epoch": 27.856159669649003, "grad_norm": 14.466719627380371, "learning_rate": 3.60719201651755e-05, "loss": 0.8757, "step": 161900 }, { "epoch": 27.864762560220235, "grad_norm": 14.342036247253418, "learning_rate": 3.606761871988988e-05, "loss": 0.8751, "step": 161950 }, { "epoch": 27.873365450791464, "grad_norm": 10.613945960998535, "learning_rate": 3.606331727460427e-05, "loss": 0.9126, "step": 162000 }, { "epoch": 27.881968341362697, "grad_norm": 15.982879638671875, "learning_rate": 3.605901582931865e-05, "loss": 0.913, "step": 162050 }, { "epoch": 27.89057123193393, "grad_norm": 21.44317626953125, "learning_rate": 3.605471438403304e-05, "loss": 0.8782, "step": 162100 }, { "epoch": 27.899174122505162, "grad_norm": 12.822269439697266, "learning_rate": 3.605041293874742e-05, "loss": 0.9218, "step": 162150 }, { "epoch": 27.907777013076394, "grad_norm": 6.378741264343262, "learning_rate": 3.60461114934618e-05, "loss": 0.9077, "step": 162200 }, { "epoch": 27.916379903647627, "grad_norm": 9.86790657043457, "learning_rate": 3.604181004817619e-05, "loss": 0.9081, "step": 162250 }, { "epoch": 27.924982794218856, "grad_norm": 11.631595611572266, "learning_rate": 3.603750860289057e-05, "loss": 0.9239, "step": 162300 }, { "epoch": 27.93358568479009, "grad_norm": 12.178318977355957, "learning_rate": 3.603320715760496e-05, "loss": 0.9039, "step": 162350 }, { "epoch": 27.94218857536132, "grad_norm": 8.336461067199707, "learning_rate": 3.602890571231934e-05, "loss": 0.9155, "step": 162400 }, { "epoch": 27.950791465932554, "grad_norm": 8.422204971313477, "learning_rate": 3.602460426703372e-05, "loss": 0.9532, "step": 162450 }, { "epoch": 27.959394356503786, "grad_norm": 9.665349006652832, "learning_rate": 3.602030282174811e-05, "loss": 0.9044, "step": 162500 }, { "epoch": 27.96799724707502, "grad_norm": 9.239334106445312, "learning_rate": 3.601600137646249e-05, "loss": 0.9252, "step": 162550 }, { "epoch": 27.976600137646248, "grad_norm": 8.091375350952148, "learning_rate": 3.601169993117688e-05, "loss": 0.8748, "step": 162600 }, { "epoch": 27.98520302821748, "grad_norm": 11.754035949707031, "learning_rate": 3.6007398485891266e-05, "loss": 0.9372, "step": 162650 }, { "epoch": 27.993805918788713, "grad_norm": 12.574356079101562, "learning_rate": 3.600309704060564e-05, "loss": 0.8774, "step": 162700 }, { "epoch": 28.0, "eval_accuracy": 0.45560144553433146, "eval_f1": 0.4249789205548461, "eval_f1_DuraRiadoRio_16x16": 0.342365317182021, "eval_f1_Mole_16x16": 0.4770847306058574, "eval_f1_Quebrado_16x16": 0.5554721513286294, "eval_f1_RiadoRio_16x16": 0.22526991537788152, "eval_f1_RioFechado_16x16": 0.5247024882798413, "eval_loss": 2.6207549571990967, "eval_precision": 0.47887197019935, "eval_precision_DuraRiadoRio_16x16": 0.6446578631452581, "eval_precision_Mole_16x16": 0.4919317657906869, "eval_precision_Quebrado_16x16": 0.4246040853798485, "eval_precision_RiadoRio_16x16": 0.3879396984924623, "eval_precision_RioFechado_16x16": 0.4452264381884945, "eval_recall": 0.45931344470886026, "eval_recall_DuraRiadoRio_16x16": 0.23307291666666666, "eval_recall_Mole_16x16": 0.4631076388888889, "eval_recall_Quebrado_16x16": 0.8029513888888888, "eval_recall_RiadoRio_16x16": 0.15871710526315788, "eval_recall_RioFechado_16x16": 0.6387181738366988, "eval_runtime": 45.7588, "eval_samples_per_second": 253.984, "eval_steps_per_second": 15.888, "step": 162736 }, { "epoch": 28.002408809359945, "grad_norm": 14.768881797790527, "learning_rate": 3.599879559532003e-05, "loss": 0.9469, "step": 162750 }, { "epoch": 28.011011699931178, "grad_norm": 11.763396263122559, "learning_rate": 3.599449415003442e-05, "loss": 0.8439, "step": 162800 }, { "epoch": 28.01961459050241, "grad_norm": 8.629043579101562, "learning_rate": 3.59901927047488e-05, "loss": 0.9065, "step": 162850 }, { "epoch": 28.02821748107364, "grad_norm": 10.915056228637695, "learning_rate": 3.5985891259463186e-05, "loss": 0.8822, "step": 162900 }, { "epoch": 28.036820371644872, "grad_norm": 13.30303955078125, "learning_rate": 3.598158981417756e-05, "loss": 0.8705, "step": 162950 }, { "epoch": 28.045423262216104, "grad_norm": 13.142707824707031, "learning_rate": 3.597728836889195e-05, "loss": 0.8865, "step": 163000 }, { "epoch": 28.054026152787337, "grad_norm": 9.181596755981445, "learning_rate": 3.5972986923606336e-05, "loss": 0.8517, "step": 163050 }, { "epoch": 28.06262904335857, "grad_norm": 6.573636054992676, "learning_rate": 3.596868547832072e-05, "loss": 0.881, "step": 163100 }, { "epoch": 28.071231933929802, "grad_norm": 16.13492202758789, "learning_rate": 3.5964384033035105e-05, "loss": 0.8986, "step": 163150 }, { "epoch": 28.07983482450103, "grad_norm": 12.628347396850586, "learning_rate": 3.596008258774948e-05, "loss": 0.8354, "step": 163200 }, { "epoch": 28.088437715072264, "grad_norm": 10.476160049438477, "learning_rate": 3.595578114246387e-05, "loss": 0.8022, "step": 163250 }, { "epoch": 28.097040605643496, "grad_norm": 22.767383575439453, "learning_rate": 3.5951479697178256e-05, "loss": 0.8959, "step": 163300 }, { "epoch": 28.10564349621473, "grad_norm": 9.599798202514648, "learning_rate": 3.594717825189264e-05, "loss": 0.8982, "step": 163350 }, { "epoch": 28.11424638678596, "grad_norm": 12.43053913116455, "learning_rate": 3.5942876806607025e-05, "loss": 0.8324, "step": 163400 }, { "epoch": 28.122849277357194, "grad_norm": 12.93720531463623, "learning_rate": 3.5938575361321406e-05, "loss": 0.8511, "step": 163450 }, { "epoch": 28.131452167928423, "grad_norm": 14.220437049865723, "learning_rate": 3.593427391603579e-05, "loss": 0.8861, "step": 163500 }, { "epoch": 28.140055058499655, "grad_norm": 12.862784385681152, "learning_rate": 3.5929972470750175e-05, "loss": 0.8913, "step": 163550 }, { "epoch": 28.148657949070888, "grad_norm": 9.989619255065918, "learning_rate": 3.5925671025464556e-05, "loss": 0.9005, "step": 163600 }, { "epoch": 28.15726083964212, "grad_norm": 13.383058547973633, "learning_rate": 3.5921369580178944e-05, "loss": 0.8945, "step": 163650 }, { "epoch": 28.165863730213353, "grad_norm": 9.561820983886719, "learning_rate": 3.5917068134893325e-05, "loss": 0.9182, "step": 163700 }, { "epoch": 28.174466620784585, "grad_norm": 12.399988174438477, "learning_rate": 3.5912766689607706e-05, "loss": 0.9497, "step": 163750 }, { "epoch": 28.183069511355814, "grad_norm": 8.568437576293945, "learning_rate": 3.5908465244322094e-05, "loss": 0.9296, "step": 163800 }, { "epoch": 28.191672401927047, "grad_norm": 11.089445114135742, "learning_rate": 3.5904163799036476e-05, "loss": 0.8929, "step": 163850 }, { "epoch": 28.20027529249828, "grad_norm": 12.688273429870605, "learning_rate": 3.5899862353750863e-05, "loss": 0.8852, "step": 163900 }, { "epoch": 28.208878183069512, "grad_norm": 11.890348434448242, "learning_rate": 3.5895560908465245e-05, "loss": 0.9365, "step": 163950 }, { "epoch": 28.217481073640744, "grad_norm": 13.517451286315918, "learning_rate": 3.589125946317963e-05, "loss": 0.8476, "step": 164000 }, { "epoch": 28.226083964211977, "grad_norm": 7.629504680633545, "learning_rate": 3.5886958017894014e-05, "loss": 0.9327, "step": 164050 }, { "epoch": 28.234686854783206, "grad_norm": 13.92194938659668, "learning_rate": 3.5882656572608395e-05, "loss": 0.8939, "step": 164100 }, { "epoch": 28.24328974535444, "grad_norm": 9.769216537475586, "learning_rate": 3.587835512732278e-05, "loss": 0.8454, "step": 164150 }, { "epoch": 28.25189263592567, "grad_norm": 8.49891185760498, "learning_rate": 3.587405368203717e-05, "loss": 0.9222, "step": 164200 }, { "epoch": 28.260495526496904, "grad_norm": 11.948705673217773, "learning_rate": 3.586975223675155e-05, "loss": 0.9058, "step": 164250 }, { "epoch": 28.269098417068136, "grad_norm": 7.919321537017822, "learning_rate": 3.586545079146593e-05, "loss": 0.8835, "step": 164300 }, { "epoch": 28.27770130763937, "grad_norm": 10.843772888183594, "learning_rate": 3.5861149346180314e-05, "loss": 0.8636, "step": 164350 }, { "epoch": 28.286304198210598, "grad_norm": 9.005457878112793, "learning_rate": 3.58568479008947e-05, "loss": 0.8616, "step": 164400 }, { "epoch": 28.29490708878183, "grad_norm": 16.002159118652344, "learning_rate": 3.585254645560909e-05, "loss": 0.8857, "step": 164450 }, { "epoch": 28.303509979353063, "grad_norm": 8.533577919006348, "learning_rate": 3.584824501032347e-05, "loss": 0.8876, "step": 164500 }, { "epoch": 28.312112869924295, "grad_norm": 11.030136108398438, "learning_rate": 3.584394356503785e-05, "loss": 0.9408, "step": 164550 }, { "epoch": 28.320715760495528, "grad_norm": 12.19456958770752, "learning_rate": 3.5839642119752234e-05, "loss": 0.87, "step": 164600 }, { "epoch": 28.329318651066757, "grad_norm": 10.80595588684082, "learning_rate": 3.583534067446662e-05, "loss": 0.8383, "step": 164650 }, { "epoch": 28.33792154163799, "grad_norm": 9.42219352722168, "learning_rate": 3.583103922918101e-05, "loss": 0.9739, "step": 164700 }, { "epoch": 28.346524432209222, "grad_norm": 7.272032260894775, "learning_rate": 3.582673778389539e-05, "loss": 0.9292, "step": 164750 }, { "epoch": 28.355127322780454, "grad_norm": 9.790452003479004, "learning_rate": 3.582243633860978e-05, "loss": 0.8751, "step": 164800 }, { "epoch": 28.363730213351687, "grad_norm": 9.514883041381836, "learning_rate": 3.581813489332416e-05, "loss": 0.8984, "step": 164850 }, { "epoch": 28.37233310392292, "grad_norm": 15.846576690673828, "learning_rate": 3.581383344803854e-05, "loss": 0.9462, "step": 164900 }, { "epoch": 28.38093599449415, "grad_norm": 13.019845008850098, "learning_rate": 3.580953200275293e-05, "loss": 0.9284, "step": 164950 }, { "epoch": 28.38953888506538, "grad_norm": 8.808255195617676, "learning_rate": 3.580523055746731e-05, "loss": 0.8947, "step": 165000 }, { "epoch": 28.398141775636613, "grad_norm": 13.216740608215332, "learning_rate": 3.58009291121817e-05, "loss": 0.9333, "step": 165050 }, { "epoch": 28.406744666207846, "grad_norm": 8.217879295349121, "learning_rate": 3.579662766689608e-05, "loss": 0.8823, "step": 165100 }, { "epoch": 28.41534755677908, "grad_norm": 11.580968856811523, "learning_rate": 3.579232622161046e-05, "loss": 0.8454, "step": 165150 }, { "epoch": 28.42395044735031, "grad_norm": 9.976579666137695, "learning_rate": 3.578802477632485e-05, "loss": 0.9083, "step": 165200 }, { "epoch": 28.43255333792154, "grad_norm": 9.2488431930542, "learning_rate": 3.578372333103923e-05, "loss": 0.9095, "step": 165250 }, { "epoch": 28.441156228492773, "grad_norm": 10.56016731262207, "learning_rate": 3.577942188575362e-05, "loss": 0.9091, "step": 165300 }, { "epoch": 28.449759119064005, "grad_norm": 14.093789100646973, "learning_rate": 3.5775120440468e-05, "loss": 0.8544, "step": 165350 }, { "epoch": 28.458362009635238, "grad_norm": 12.151350975036621, "learning_rate": 3.577081899518238e-05, "loss": 0.935, "step": 165400 }, { "epoch": 28.46696490020647, "grad_norm": 13.038127899169922, "learning_rate": 3.576651754989677e-05, "loss": 0.8901, "step": 165450 }, { "epoch": 28.475567790777703, "grad_norm": 9.624930381774902, "learning_rate": 3.576221610461115e-05, "loss": 0.8855, "step": 165500 }, { "epoch": 28.48417068134893, "grad_norm": 8.309195518493652, "learning_rate": 3.575791465932554e-05, "loss": 0.8828, "step": 165550 }, { "epoch": 28.492773571920164, "grad_norm": 8.777364730834961, "learning_rate": 3.5753613214039925e-05, "loss": 0.8686, "step": 165600 }, { "epoch": 28.501376462491397, "grad_norm": 8.715258598327637, "learning_rate": 3.57493117687543e-05, "loss": 0.9267, "step": 165650 }, { "epoch": 28.50997935306263, "grad_norm": 15.225371360778809, "learning_rate": 3.574501032346869e-05, "loss": 0.8893, "step": 165700 }, { "epoch": 28.518582243633862, "grad_norm": 8.26577091217041, "learning_rate": 3.574070887818307e-05, "loss": 0.8761, "step": 165750 }, { "epoch": 28.527185134205094, "grad_norm": 9.896106719970703, "learning_rate": 3.5736407432897456e-05, "loss": 0.8902, "step": 165800 }, { "epoch": 28.535788024776323, "grad_norm": 10.827815055847168, "learning_rate": 3.5732105987611844e-05, "loss": 0.9207, "step": 165850 }, { "epoch": 28.544390915347556, "grad_norm": 12.975659370422363, "learning_rate": 3.572780454232622e-05, "loss": 0.8553, "step": 165900 }, { "epoch": 28.55299380591879, "grad_norm": 8.808277130126953, "learning_rate": 3.572350309704061e-05, "loss": 0.8995, "step": 165950 }, { "epoch": 28.56159669649002, "grad_norm": 9.75830078125, "learning_rate": 3.571920165175499e-05, "loss": 0.9048, "step": 166000 }, { "epoch": 28.570199587061254, "grad_norm": 10.196927070617676, "learning_rate": 3.5714900206469376e-05, "loss": 0.8408, "step": 166050 }, { "epoch": 28.578802477632486, "grad_norm": 11.32848834991455, "learning_rate": 3.5710598761183764e-05, "loss": 0.8904, "step": 166100 }, { "epoch": 28.587405368203715, "grad_norm": 13.869812965393066, "learning_rate": 3.5706297315898145e-05, "loss": 0.9401, "step": 166150 }, { "epoch": 28.596008258774948, "grad_norm": 12.692739486694336, "learning_rate": 3.5701995870612526e-05, "loss": 0.9024, "step": 166200 }, { "epoch": 28.60461114934618, "grad_norm": 9.639309883117676, "learning_rate": 3.5697694425326914e-05, "loss": 0.9431, "step": 166250 }, { "epoch": 28.613214039917413, "grad_norm": 5.82691764831543, "learning_rate": 3.5693392980041295e-05, "loss": 0.8893, "step": 166300 }, { "epoch": 28.621816930488645, "grad_norm": 14.48844051361084, "learning_rate": 3.568909153475568e-05, "loss": 0.9378, "step": 166350 }, { "epoch": 28.630419821059878, "grad_norm": 10.1719388961792, "learning_rate": 3.5684790089470064e-05, "loss": 0.8994, "step": 166400 }, { "epoch": 28.639022711631107, "grad_norm": 5.662003040313721, "learning_rate": 3.5680488644184446e-05, "loss": 0.8862, "step": 166450 }, { "epoch": 28.64762560220234, "grad_norm": 14.81537914276123, "learning_rate": 3.5676187198898833e-05, "loss": 0.8967, "step": 166500 }, { "epoch": 28.65622849277357, "grad_norm": 6.945750713348389, "learning_rate": 3.5671885753613215e-05, "loss": 0.8893, "step": 166550 }, { "epoch": 28.664831383344804, "grad_norm": 7.174513816833496, "learning_rate": 3.56675843083276e-05, "loss": 0.8895, "step": 166600 }, { "epoch": 28.673434273916037, "grad_norm": 12.401618003845215, "learning_rate": 3.5663282863041984e-05, "loss": 0.9201, "step": 166650 }, { "epoch": 28.68203716448727, "grad_norm": 14.432713508605957, "learning_rate": 3.5658981417756365e-05, "loss": 0.925, "step": 166700 }, { "epoch": 28.6906400550585, "grad_norm": 8.164896011352539, "learning_rate": 3.565467997247075e-05, "loss": 0.9297, "step": 166750 }, { "epoch": 28.69924294562973, "grad_norm": 7.483365058898926, "learning_rate": 3.5650378527185134e-05, "loss": 0.9042, "step": 166800 }, { "epoch": 28.707845836200963, "grad_norm": 8.09028148651123, "learning_rate": 3.564607708189952e-05, "loss": 0.8979, "step": 166850 }, { "epoch": 28.716448726772196, "grad_norm": 9.713075637817383, "learning_rate": 3.56417756366139e-05, "loss": 0.8709, "step": 166900 }, { "epoch": 28.72505161734343, "grad_norm": 10.576570510864258, "learning_rate": 3.563747419132829e-05, "loss": 0.854, "step": 166950 }, { "epoch": 28.73365450791466, "grad_norm": 9.617383003234863, "learning_rate": 3.563317274604267e-05, "loss": 0.8734, "step": 167000 }, { "epoch": 28.74225739848589, "grad_norm": 13.668933868408203, "learning_rate": 3.5628871300757053e-05, "loss": 0.876, "step": 167050 }, { "epoch": 28.750860289057123, "grad_norm": 10.475780487060547, "learning_rate": 3.562456985547144e-05, "loss": 0.9053, "step": 167100 }, { "epoch": 28.759463179628355, "grad_norm": 11.017013549804688, "learning_rate": 3.562026841018582e-05, "loss": 0.9009, "step": 167150 }, { "epoch": 28.768066070199588, "grad_norm": 11.232638359069824, "learning_rate": 3.561596696490021e-05, "loss": 0.9252, "step": 167200 }, { "epoch": 28.77666896077082, "grad_norm": 11.539307594299316, "learning_rate": 3.561166551961459e-05, "loss": 0.8972, "step": 167250 }, { "epoch": 28.785271851342053, "grad_norm": 10.612844467163086, "learning_rate": 3.560736407432897e-05, "loss": 0.9146, "step": 167300 }, { "epoch": 28.79387474191328, "grad_norm": 9.014254570007324, "learning_rate": 3.560306262904336e-05, "loss": 0.8943, "step": 167350 }, { "epoch": 28.802477632484514, "grad_norm": 11.453691482543945, "learning_rate": 3.559876118375774e-05, "loss": 0.8938, "step": 167400 }, { "epoch": 28.811080523055747, "grad_norm": 12.569795608520508, "learning_rate": 3.559445973847213e-05, "loss": 0.907, "step": 167450 }, { "epoch": 28.81968341362698, "grad_norm": 9.520381927490234, "learning_rate": 3.559015829318651e-05, "loss": 0.9247, "step": 167500 }, { "epoch": 28.828286304198212, "grad_norm": 10.340448379516602, "learning_rate": 3.558585684790089e-05, "loss": 0.8986, "step": 167550 }, { "epoch": 28.83688919476944, "grad_norm": 9.472855567932129, "learning_rate": 3.558155540261528e-05, "loss": 0.8887, "step": 167600 }, { "epoch": 28.845492085340673, "grad_norm": 9.76060962677002, "learning_rate": 3.557725395732966e-05, "loss": 0.8987, "step": 167650 }, { "epoch": 28.854094975911906, "grad_norm": 9.679862022399902, "learning_rate": 3.557295251204405e-05, "loss": 0.8748, "step": 167700 }, { "epoch": 28.86269786648314, "grad_norm": 10.139730453491211, "learning_rate": 3.556865106675844e-05, "loss": 0.9309, "step": 167750 }, { "epoch": 28.87130075705437, "grad_norm": 8.147704124450684, "learning_rate": 3.556434962147281e-05, "loss": 0.86, "step": 167800 }, { "epoch": 28.879903647625603, "grad_norm": 10.217731475830078, "learning_rate": 3.55600481761872e-05, "loss": 0.9386, "step": 167850 }, { "epoch": 28.888506538196832, "grad_norm": 11.757140159606934, "learning_rate": 3.555574673090159e-05, "loss": 0.8678, "step": 167900 }, { "epoch": 28.897109428768065, "grad_norm": 16.408475875854492, "learning_rate": 3.555144528561597e-05, "loss": 0.8882, "step": 167950 }, { "epoch": 28.905712319339298, "grad_norm": 12.710854530334473, "learning_rate": 3.554714384033036e-05, "loss": 0.8442, "step": 168000 }, { "epoch": 28.91431520991053, "grad_norm": 12.580416679382324, "learning_rate": 3.554284239504473e-05, "loss": 0.9477, "step": 168050 }, { "epoch": 28.922918100481763, "grad_norm": 14.231019973754883, "learning_rate": 3.553854094975912e-05, "loss": 0.9244, "step": 168100 }, { "epoch": 28.931520991052995, "grad_norm": 13.350014686584473, "learning_rate": 3.553423950447351e-05, "loss": 0.8878, "step": 168150 }, { "epoch": 28.940123881624224, "grad_norm": 20.142030715942383, "learning_rate": 3.552993805918789e-05, "loss": 0.8937, "step": 168200 }, { "epoch": 28.948726772195457, "grad_norm": 9.37252426147461, "learning_rate": 3.5525636613902276e-05, "loss": 0.9331, "step": 168250 }, { "epoch": 28.95732966276669, "grad_norm": 17.48586082458496, "learning_rate": 3.552133516861666e-05, "loss": 0.9189, "step": 168300 }, { "epoch": 28.96593255333792, "grad_norm": 10.70217514038086, "learning_rate": 3.551703372333104e-05, "loss": 0.9183, "step": 168350 }, { "epoch": 28.974535443909154, "grad_norm": 9.276798248291016, "learning_rate": 3.5512732278045426e-05, "loss": 0.828, "step": 168400 }, { "epoch": 28.983138334480387, "grad_norm": 9.048457145690918, "learning_rate": 3.550843083275981e-05, "loss": 0.9198, "step": 168450 }, { "epoch": 28.991741225051616, "grad_norm": 12.530298233032227, "learning_rate": 3.5504129387474196e-05, "loss": 0.9192, "step": 168500 }, { "epoch": 29.0, "eval_accuracy": 0.558165548098434, "eval_f1": 0.5623321064678286, "eval_f1_DuraRiadoRio_16x16": 0.5639940765813413, "eval_f1_Mole_16x16": 0.5104602510460251, "eval_f1_Quebrado_16x16": 0.6887109274089382, "eval_f1_RiadoRio_16x16": 0.5272539395505037, "eval_f1_RioFechado_16x16": 0.521241337752335, "eval_loss": 1.4114956855773926, "eval_precision": 0.6660802289481851, "eval_precision_DuraRiadoRio_16x16": 0.5501444490301279, "eval_precision_Mole_16x16": 0.7142857142857143, "eval_precision_Quebrado_16x16": 0.8506700701978303, "eval_precision_RiadoRio_16x16": 0.38436911487758946, "eval_precision_RioFechado_16x16": 0.8309317963496637, "eval_recall": 0.5546398995412515, "eval_recall_DuraRiadoRio_16x16": 0.5785590277777778, "eval_recall_Mole_16x16": 0.3971354166666667, "eval_recall_Quebrado_16x16": 0.5785590277777778, "eval_recall_RiadoRio_16x16": 0.8392269736842105, "eval_recall_RioFechado_16x16": 0.3797190517998244, "eval_runtime": 45.9264, "eval_samples_per_second": 253.057, "eval_steps_per_second": 15.83, "step": 168548 }, { "epoch": 29.00034411562285, "grad_norm": 12.308828353881836, "learning_rate": 3.549982794218858e-05, "loss": 0.8935, "step": 168550 }, { "epoch": 29.00894700619408, "grad_norm": 11.792147636413574, "learning_rate": 3.549552649690296e-05, "loss": 0.8843, "step": 168600 }, { "epoch": 29.017549896765313, "grad_norm": 9.818890571594238, "learning_rate": 3.5491225051617346e-05, "loss": 0.8308, "step": 168650 }, { "epoch": 29.026152787336546, "grad_norm": 12.29063892364502, "learning_rate": 3.548692360633173e-05, "loss": 0.8789, "step": 168700 }, { "epoch": 29.03475567790778, "grad_norm": 8.137463569641113, "learning_rate": 3.5482622161046115e-05, "loss": 0.8763, "step": 168750 }, { "epoch": 29.043358568479007, "grad_norm": 17.189464569091797, "learning_rate": 3.5478320715760496e-05, "loss": 0.9864, "step": 168800 }, { "epoch": 29.05196145905024, "grad_norm": 11.926457405090332, "learning_rate": 3.547401927047488e-05, "loss": 0.8916, "step": 168850 }, { "epoch": 29.060564349621473, "grad_norm": 10.093523025512695, "learning_rate": 3.5469717825189265e-05, "loss": 0.849, "step": 168900 }, { "epoch": 29.069167240192705, "grad_norm": 7.572750568389893, "learning_rate": 3.5465416379903646e-05, "loss": 0.8587, "step": 168950 }, { "epoch": 29.077770130763938, "grad_norm": 10.877213478088379, "learning_rate": 3.5461114934618034e-05, "loss": 0.8902, "step": 169000 }, { "epoch": 29.08637302133517, "grad_norm": 12.154471397399902, "learning_rate": 3.5456813489332416e-05, "loss": 0.862, "step": 169050 }, { "epoch": 29.0949759119064, "grad_norm": 14.565794944763184, "learning_rate": 3.5452512044046804e-05, "loss": 0.9051, "step": 169100 }, { "epoch": 29.10357880247763, "grad_norm": 10.94548225402832, "learning_rate": 3.5448210598761185e-05, "loss": 0.8777, "step": 169150 }, { "epoch": 29.112181693048864, "grad_norm": 9.953696250915527, "learning_rate": 3.5443909153475566e-05, "loss": 0.8506, "step": 169200 }, { "epoch": 29.120784583620097, "grad_norm": 8.217525482177734, "learning_rate": 3.5439607708189954e-05, "loss": 0.9099, "step": 169250 }, { "epoch": 29.12938747419133, "grad_norm": 7.318860054016113, "learning_rate": 3.543530626290434e-05, "loss": 0.8253, "step": 169300 }, { "epoch": 29.13799036476256, "grad_norm": 14.135316848754883, "learning_rate": 3.543100481761872e-05, "loss": 0.8962, "step": 169350 }, { "epoch": 29.14659325533379, "grad_norm": 10.082715034484863, "learning_rate": 3.5426703372333104e-05, "loss": 0.9037, "step": 169400 }, { "epoch": 29.155196145905023, "grad_norm": 11.452386856079102, "learning_rate": 3.5422401927047485e-05, "loss": 0.923, "step": 169450 }, { "epoch": 29.163799036476256, "grad_norm": 9.704338073730469, "learning_rate": 3.541810048176187e-05, "loss": 0.8502, "step": 169500 }, { "epoch": 29.17240192704749, "grad_norm": 8.841708183288574, "learning_rate": 3.541379903647626e-05, "loss": 0.896, "step": 169550 }, { "epoch": 29.18100481761872, "grad_norm": 10.550212860107422, "learning_rate": 3.540949759119064e-05, "loss": 0.8873, "step": 169600 }, { "epoch": 29.189607708189953, "grad_norm": 12.12903118133545, "learning_rate": 3.540519614590503e-05, "loss": 0.8799, "step": 169650 }, { "epoch": 29.198210598761182, "grad_norm": 10.53392219543457, "learning_rate": 3.5400894700619405e-05, "loss": 0.8899, "step": 169700 }, { "epoch": 29.206813489332415, "grad_norm": 10.806631088256836, "learning_rate": 3.539659325533379e-05, "loss": 0.8935, "step": 169750 }, { "epoch": 29.215416379903647, "grad_norm": 8.589448928833008, "learning_rate": 3.539229181004818e-05, "loss": 0.857, "step": 169800 }, { "epoch": 29.22401927047488, "grad_norm": 12.869205474853516, "learning_rate": 3.538799036476256e-05, "loss": 0.9538, "step": 169850 }, { "epoch": 29.232622161046113, "grad_norm": 18.149141311645508, "learning_rate": 3.538368891947695e-05, "loss": 0.8541, "step": 169900 }, { "epoch": 29.241225051617345, "grad_norm": 10.929856300354004, "learning_rate": 3.537938747419133e-05, "loss": 0.9265, "step": 169950 }, { "epoch": 29.249827942188574, "grad_norm": 17.026283264160156, "learning_rate": 3.537508602890571e-05, "loss": 0.8543, "step": 170000 }, { "epoch": 29.258430832759807, "grad_norm": 11.215804100036621, "learning_rate": 3.53707845836201e-05, "loss": 0.9063, "step": 170050 }, { "epoch": 29.26703372333104, "grad_norm": 7.651368141174316, "learning_rate": 3.536648313833448e-05, "loss": 0.8783, "step": 170100 }, { "epoch": 29.27563661390227, "grad_norm": 10.658320426940918, "learning_rate": 3.536218169304887e-05, "loss": 0.8878, "step": 170150 }, { "epoch": 29.284239504473504, "grad_norm": 10.997809410095215, "learning_rate": 3.535788024776325e-05, "loss": 0.8152, "step": 170200 }, { "epoch": 29.292842395044737, "grad_norm": 10.796038627624512, "learning_rate": 3.535357880247763e-05, "loss": 0.8749, "step": 170250 }, { "epoch": 29.301445285615966, "grad_norm": 10.161111831665039, "learning_rate": 3.534927735719202e-05, "loss": 0.9064, "step": 170300 }, { "epoch": 29.3100481761872, "grad_norm": 11.132206916809082, "learning_rate": 3.53449759119064e-05, "loss": 0.9338, "step": 170350 }, { "epoch": 29.31865106675843, "grad_norm": 12.313220977783203, "learning_rate": 3.534067446662079e-05, "loss": 0.8582, "step": 170400 }, { "epoch": 29.327253957329663, "grad_norm": 6.837219715118408, "learning_rate": 3.533637302133517e-05, "loss": 0.8785, "step": 170450 }, { "epoch": 29.335856847900896, "grad_norm": 14.84729290008545, "learning_rate": 3.533207157604955e-05, "loss": 0.9161, "step": 170500 }, { "epoch": 29.344459738472125, "grad_norm": 11.822949409484863, "learning_rate": 3.532777013076394e-05, "loss": 0.966, "step": 170550 }, { "epoch": 29.353062629043357, "grad_norm": 10.928886413574219, "learning_rate": 3.532346868547832e-05, "loss": 0.8391, "step": 170600 }, { "epoch": 29.36166551961459, "grad_norm": 9.065164566040039, "learning_rate": 3.531916724019271e-05, "loss": 0.8596, "step": 170650 }, { "epoch": 29.370268410185822, "grad_norm": 10.89782428741455, "learning_rate": 3.5314865794907096e-05, "loss": 0.9021, "step": 170700 }, { "epoch": 29.378871300757055, "grad_norm": 9.379724502563477, "learning_rate": 3.531056434962147e-05, "loss": 0.8953, "step": 170750 }, { "epoch": 29.387474191328288, "grad_norm": 8.664748191833496, "learning_rate": 3.530626290433586e-05, "loss": 0.9284, "step": 170800 }, { "epoch": 29.396077081899517, "grad_norm": 8.32276439666748, "learning_rate": 3.530196145905024e-05, "loss": 0.8902, "step": 170850 }, { "epoch": 29.40467997247075, "grad_norm": 12.017376899719238, "learning_rate": 3.529766001376463e-05, "loss": 0.8974, "step": 170900 }, { "epoch": 29.41328286304198, "grad_norm": 10.85587215423584, "learning_rate": 3.5293358568479015e-05, "loss": 0.8752, "step": 170950 }, { "epoch": 29.421885753613214, "grad_norm": 16.593303680419922, "learning_rate": 3.528905712319339e-05, "loss": 0.8804, "step": 171000 }, { "epoch": 29.430488644184447, "grad_norm": 9.513830184936523, "learning_rate": 3.528475567790778e-05, "loss": 0.8602, "step": 171050 }, { "epoch": 29.43909153475568, "grad_norm": 11.959114074707031, "learning_rate": 3.528045423262216e-05, "loss": 0.9095, "step": 171100 }, { "epoch": 29.447694425326908, "grad_norm": 6.051203727722168, "learning_rate": 3.527615278733655e-05, "loss": 0.8223, "step": 171150 }, { "epoch": 29.45629731589814, "grad_norm": 16.718095779418945, "learning_rate": 3.5271851342050935e-05, "loss": 0.8875, "step": 171200 }, { "epoch": 29.464900206469373, "grad_norm": 7.588572025299072, "learning_rate": 3.5267549896765316e-05, "loss": 0.8877, "step": 171250 }, { "epoch": 29.473503097040606, "grad_norm": 5.419987678527832, "learning_rate": 3.52632484514797e-05, "loss": 0.8414, "step": 171300 }, { "epoch": 29.48210598761184, "grad_norm": 13.88894271850586, "learning_rate": 3.5258947006194085e-05, "loss": 0.8831, "step": 171350 }, { "epoch": 29.49070887818307, "grad_norm": 15.74412727355957, "learning_rate": 3.5254645560908466e-05, "loss": 0.8633, "step": 171400 }, { "epoch": 29.4993117687543, "grad_norm": 10.889122009277344, "learning_rate": 3.5250344115622854e-05, "loss": 0.9086, "step": 171450 }, { "epoch": 29.507914659325532, "grad_norm": 9.462313652038574, "learning_rate": 3.5246042670337235e-05, "loss": 0.8538, "step": 171500 }, { "epoch": 29.516517549896765, "grad_norm": 13.816802024841309, "learning_rate": 3.5241741225051616e-05, "loss": 0.9064, "step": 171550 }, { "epoch": 29.525120440467997, "grad_norm": 13.754746437072754, "learning_rate": 3.5237439779766004e-05, "loss": 0.8597, "step": 171600 }, { "epoch": 29.53372333103923, "grad_norm": 10.469192504882812, "learning_rate": 3.5233138334480386e-05, "loss": 0.9693, "step": 171650 }, { "epoch": 29.542326221610463, "grad_norm": 10.825498580932617, "learning_rate": 3.5228836889194774e-05, "loss": 0.8948, "step": 171700 }, { "epoch": 29.55092911218169, "grad_norm": 10.083127975463867, "learning_rate": 3.5224535443909155e-05, "loss": 0.8909, "step": 171750 }, { "epoch": 29.559532002752924, "grad_norm": 16.572912216186523, "learning_rate": 3.522023399862354e-05, "loss": 0.9625, "step": 171800 }, { "epoch": 29.568134893324157, "grad_norm": 8.276968002319336, "learning_rate": 3.5215932553337924e-05, "loss": 0.8806, "step": 171850 }, { "epoch": 29.57673778389539, "grad_norm": 14.428668975830078, "learning_rate": 3.5211631108052305e-05, "loss": 0.9068, "step": 171900 }, { "epoch": 29.58534067446662, "grad_norm": 9.375755310058594, "learning_rate": 3.520732966276669e-05, "loss": 0.8841, "step": 171950 }, { "epoch": 29.593943565037854, "grad_norm": 11.360371589660645, "learning_rate": 3.5203028217481074e-05, "loss": 0.8743, "step": 172000 }, { "epoch": 29.602546455609083, "grad_norm": 5.869993209838867, "learning_rate": 3.519872677219546e-05, "loss": 0.8719, "step": 172050 }, { "epoch": 29.611149346180316, "grad_norm": 8.538166999816895, "learning_rate": 3.519442532690984e-05, "loss": 0.9235, "step": 172100 }, { "epoch": 29.61975223675155, "grad_norm": 11.62211799621582, "learning_rate": 3.5190123881624224e-05, "loss": 0.8862, "step": 172150 }, { "epoch": 29.62835512732278, "grad_norm": 12.64539623260498, "learning_rate": 3.518582243633861e-05, "loss": 0.909, "step": 172200 }, { "epoch": 29.636958017894013, "grad_norm": 11.540177345275879, "learning_rate": 3.5181520991052994e-05, "loss": 0.8807, "step": 172250 }, { "epoch": 29.645560908465246, "grad_norm": 7.885228633880615, "learning_rate": 3.517721954576738e-05, "loss": 0.8612, "step": 172300 }, { "epoch": 29.654163799036475, "grad_norm": 12.164545059204102, "learning_rate": 3.517291810048176e-05, "loss": 0.8394, "step": 172350 }, { "epoch": 29.662766689607707, "grad_norm": 7.821190357208252, "learning_rate": 3.5168616655196144e-05, "loss": 0.9027, "step": 172400 }, { "epoch": 29.67136958017894, "grad_norm": 9.679380416870117, "learning_rate": 3.516431520991053e-05, "loss": 0.8375, "step": 172450 }, { "epoch": 29.679972470750172, "grad_norm": 12.641105651855469, "learning_rate": 3.516001376462491e-05, "loss": 0.8914, "step": 172500 }, { "epoch": 29.688575361321405, "grad_norm": 10.458820343017578, "learning_rate": 3.51557123193393e-05, "loss": 0.9258, "step": 172550 }, { "epoch": 29.697178251892637, "grad_norm": 11.627311706542969, "learning_rate": 3.515141087405369e-05, "loss": 0.8713, "step": 172600 }, { "epoch": 29.705781142463866, "grad_norm": 11.586438179016113, "learning_rate": 3.514710942876806e-05, "loss": 0.8581, "step": 172650 }, { "epoch": 29.7143840330351, "grad_norm": 10.648429870605469, "learning_rate": 3.514280798348245e-05, "loss": 0.9328, "step": 172700 }, { "epoch": 29.72298692360633, "grad_norm": 8.778736114501953, "learning_rate": 3.513850653819684e-05, "loss": 0.8548, "step": 172750 }, { "epoch": 29.731589814177564, "grad_norm": 15.45086669921875, "learning_rate": 3.513420509291122e-05, "loss": 0.9071, "step": 172800 }, { "epoch": 29.740192704748797, "grad_norm": 9.165240287780762, "learning_rate": 3.512990364762561e-05, "loss": 0.8395, "step": 172850 }, { "epoch": 29.74879559532003, "grad_norm": 15.65156364440918, "learning_rate": 3.512560220233998e-05, "loss": 0.8945, "step": 172900 }, { "epoch": 29.757398485891258, "grad_norm": 12.963729858398438, "learning_rate": 3.512130075705437e-05, "loss": 0.8994, "step": 172950 }, { "epoch": 29.76600137646249, "grad_norm": 12.526384353637695, "learning_rate": 3.511699931176876e-05, "loss": 0.8803, "step": 173000 }, { "epoch": 29.774604267033723, "grad_norm": 12.331477165222168, "learning_rate": 3.511269786648314e-05, "loss": 0.8801, "step": 173050 }, { "epoch": 29.783207157604956, "grad_norm": 9.256677627563477, "learning_rate": 3.510839642119753e-05, "loss": 0.826, "step": 173100 }, { "epoch": 29.79181004817619, "grad_norm": 12.421409606933594, "learning_rate": 3.51040949759119e-05, "loss": 0.8424, "step": 173150 }, { "epoch": 29.80041293874742, "grad_norm": 12.929179191589355, "learning_rate": 3.509979353062629e-05, "loss": 0.8663, "step": 173200 }, { "epoch": 29.80901582931865, "grad_norm": 12.072053909301758, "learning_rate": 3.509549208534068e-05, "loss": 0.8664, "step": 173250 }, { "epoch": 29.817618719889882, "grad_norm": 14.216740608215332, "learning_rate": 3.509119064005506e-05, "loss": 0.9654, "step": 173300 }, { "epoch": 29.826221610461115, "grad_norm": 7.001824378967285, "learning_rate": 3.508688919476945e-05, "loss": 0.8892, "step": 173350 }, { "epoch": 29.834824501032347, "grad_norm": 6.124659538269043, "learning_rate": 3.508258774948383e-05, "loss": 0.8958, "step": 173400 }, { "epoch": 29.84342739160358, "grad_norm": 11.637874603271484, "learning_rate": 3.507828630419821e-05, "loss": 0.9053, "step": 173450 }, { "epoch": 29.85203028217481, "grad_norm": 11.124427795410156, "learning_rate": 3.50739848589126e-05, "loss": 0.8775, "step": 173500 }, { "epoch": 29.86063317274604, "grad_norm": 11.891438484191895, "learning_rate": 3.506968341362698e-05, "loss": 0.9165, "step": 173550 }, { "epoch": 29.869236063317274, "grad_norm": 10.45441722869873, "learning_rate": 3.5065381968341367e-05, "loss": 0.8829, "step": 173600 }, { "epoch": 29.877838953888507, "grad_norm": 9.658273696899414, "learning_rate": 3.506108052305575e-05, "loss": 0.9146, "step": 173650 }, { "epoch": 29.88644184445974, "grad_norm": 11.707548141479492, "learning_rate": 3.505677907777013e-05, "loss": 0.8876, "step": 173700 }, { "epoch": 29.89504473503097, "grad_norm": 10.02408218383789, "learning_rate": 3.505247763248452e-05, "loss": 0.8866, "step": 173750 }, { "epoch": 29.903647625602204, "grad_norm": 8.361163139343262, "learning_rate": 3.50481761871989e-05, "loss": 0.8695, "step": 173800 }, { "epoch": 29.912250516173433, "grad_norm": 12.514907836914062, "learning_rate": 3.5043874741913286e-05, "loss": 0.8895, "step": 173850 }, { "epoch": 29.920853406744666, "grad_norm": 9.360941886901855, "learning_rate": 3.503957329662767e-05, "loss": 0.8172, "step": 173900 }, { "epoch": 29.929456297315898, "grad_norm": 12.720256805419922, "learning_rate": 3.5035271851342055e-05, "loss": 0.8381, "step": 173950 }, { "epoch": 29.93805918788713, "grad_norm": 12.987907409667969, "learning_rate": 3.5030970406056436e-05, "loss": 0.8939, "step": 174000 }, { "epoch": 29.946662078458363, "grad_norm": 8.3461332321167, "learning_rate": 3.502666896077082e-05, "loss": 0.9049, "step": 174050 }, { "epoch": 29.955264969029592, "grad_norm": 11.318819999694824, "learning_rate": 3.5022367515485205e-05, "loss": 0.8547, "step": 174100 }, { "epoch": 29.963867859600825, "grad_norm": 13.111542701721191, "learning_rate": 3.501806607019959e-05, "loss": 0.9035, "step": 174150 }, { "epoch": 29.972470750172057, "grad_norm": 12.381363868713379, "learning_rate": 3.5013764624913974e-05, "loss": 0.8768, "step": 174200 }, { "epoch": 29.98107364074329, "grad_norm": 13.837462425231934, "learning_rate": 3.5009463179628356e-05, "loss": 0.9149, "step": 174250 }, { "epoch": 29.989676531314522, "grad_norm": 13.367965698242188, "learning_rate": 3.500516173434274e-05, "loss": 0.8849, "step": 174300 }, { "epoch": 29.998279421885755, "grad_norm": 12.58559513092041, "learning_rate": 3.5000860289057125e-05, "loss": 0.9385, "step": 174350 }, { "epoch": 30.0, "eval_accuracy": 0.5241782825675443, "eval_f1": 0.4950923306793813, "eval_f1_DuraRiadoRio_16x16": 0.29834651329978434, "eval_f1_Mole_16x16": 0.6229371213703782, "eval_f1_Quebrado_16x16": 0.5957100190062449, "eval_f1_RiadoRio_16x16": 0.4357298474945534, "eval_f1_RioFechado_16x16": 0.5227381522259454, "eval_loss": 1.7991423606872559, "eval_precision": 0.6013139710934576, "eval_precision_DuraRiadoRio_16x16": 0.8682008368200836, "eval_precision_Mole_16x16": 0.6004832863471606, "eval_precision_Quebrado_16x16": 0.43342552350849467, "eval_precision_RiadoRio_16x16": 0.5297233666862861, "eval_precision_RioFechado_16x16": 0.5747368421052632, "eval_recall": 0.5257895089824356, "eval_recall_DuraRiadoRio_16x16": 0.1801215277777778, "eval_recall_Mole_16x16": 0.6471354166666666, "eval_recall_Quebrado_16x16": 0.9522569444444444, "eval_recall_RiadoRio_16x16": 0.37006578947368424, "eval_recall_RioFechado_16x16": 0.4793678665496049, "eval_runtime": 46.5736, "eval_samples_per_second": 249.541, "eval_steps_per_second": 15.61, "step": 174360 }, { "epoch": 30.006882312456984, "grad_norm": 11.890076637268066, "learning_rate": 3.499655884377151e-05, "loss": 0.8909, "step": 174400 }, { "epoch": 30.015485203028216, "grad_norm": 16.707992553710938, "learning_rate": 3.4992257398485894e-05, "loss": 0.9473, "step": 174450 }, { "epoch": 30.02408809359945, "grad_norm": 7.032901763916016, "learning_rate": 3.4987955953200275e-05, "loss": 0.8488, "step": 174500 }, { "epoch": 30.03269098417068, "grad_norm": 9.894580841064453, "learning_rate": 3.4983654507914656e-05, "loss": 0.8971, "step": 174550 }, { "epoch": 30.041293874741914, "grad_norm": 13.180243492126465, "learning_rate": 3.4979353062629044e-05, "loss": 0.8666, "step": 174600 }, { "epoch": 30.049896765313147, "grad_norm": 8.820587158203125, "learning_rate": 3.497505161734343e-05, "loss": 0.8327, "step": 174650 }, { "epoch": 30.058499655884376, "grad_norm": 12.81863784790039, "learning_rate": 3.497075017205781e-05, "loss": 0.8598, "step": 174700 }, { "epoch": 30.067102546455608, "grad_norm": 11.37162971496582, "learning_rate": 3.49664487267722e-05, "loss": 0.9001, "step": 174750 }, { "epoch": 30.07570543702684, "grad_norm": 10.967655181884766, "learning_rate": 3.496214728148658e-05, "loss": 0.8952, "step": 174800 }, { "epoch": 30.084308327598073, "grad_norm": 9.460411071777344, "learning_rate": 3.4957845836200964e-05, "loss": 0.8909, "step": 174850 }, { "epoch": 30.092911218169306, "grad_norm": 9.471195220947266, "learning_rate": 3.495354439091535e-05, "loss": 0.8503, "step": 174900 }, { "epoch": 30.10151410874054, "grad_norm": 10.539765357971191, "learning_rate": 3.494924294562973e-05, "loss": 0.8633, "step": 174950 }, { "epoch": 30.110116999311767, "grad_norm": 14.677083015441895, "learning_rate": 3.494494150034412e-05, "loss": 0.9138, "step": 175000 }, { "epoch": 30.118719889883, "grad_norm": 7.499022960662842, "learning_rate": 3.49406400550585e-05, "loss": 0.8537, "step": 175050 }, { "epoch": 30.127322780454232, "grad_norm": 9.905425071716309, "learning_rate": 3.493633860977288e-05, "loss": 0.8903, "step": 175100 }, { "epoch": 30.135925671025465, "grad_norm": 9.763962745666504, "learning_rate": 3.493203716448727e-05, "loss": 0.8552, "step": 175150 }, { "epoch": 30.144528561596697, "grad_norm": 16.258657455444336, "learning_rate": 3.492773571920165e-05, "loss": 0.8616, "step": 175200 }, { "epoch": 30.15313145216793, "grad_norm": 8.779834747314453, "learning_rate": 3.492343427391604e-05, "loss": 0.8989, "step": 175250 }, { "epoch": 30.16173434273916, "grad_norm": 13.481534004211426, "learning_rate": 3.491913282863042e-05, "loss": 0.8843, "step": 175300 }, { "epoch": 30.17033723331039, "grad_norm": 9.5375337600708, "learning_rate": 3.49148313833448e-05, "loss": 0.845, "step": 175350 }, { "epoch": 30.178940123881624, "grad_norm": 10.270456314086914, "learning_rate": 3.491052993805919e-05, "loss": 0.937, "step": 175400 }, { "epoch": 30.187543014452856, "grad_norm": 9.922199249267578, "learning_rate": 3.490622849277357e-05, "loss": 0.9156, "step": 175450 }, { "epoch": 30.19614590502409, "grad_norm": 10.338312149047852, "learning_rate": 3.490192704748796e-05, "loss": 0.8834, "step": 175500 }, { "epoch": 30.20474879559532, "grad_norm": 13.315771102905273, "learning_rate": 3.489762560220234e-05, "loss": 0.8524, "step": 175550 }, { "epoch": 30.21335168616655, "grad_norm": 14.746391296386719, "learning_rate": 3.489332415691672e-05, "loss": 0.8477, "step": 175600 }, { "epoch": 30.221954576737783, "grad_norm": 9.102505683898926, "learning_rate": 3.488902271163111e-05, "loss": 0.8908, "step": 175650 }, { "epoch": 30.230557467309016, "grad_norm": 11.054033279418945, "learning_rate": 3.488472126634549e-05, "loss": 0.8565, "step": 175700 }, { "epoch": 30.239160357880248, "grad_norm": 11.91186809539795, "learning_rate": 3.488041982105988e-05, "loss": 0.8998, "step": 175750 }, { "epoch": 30.24776324845148, "grad_norm": 7.80615758895874, "learning_rate": 3.487611837577427e-05, "loss": 0.8684, "step": 175800 }, { "epoch": 30.256366139022713, "grad_norm": 9.988863945007324, "learning_rate": 3.487181693048864e-05, "loss": 0.9242, "step": 175850 }, { "epoch": 30.264969029593942, "grad_norm": 9.690640449523926, "learning_rate": 3.486751548520303e-05, "loss": 0.9177, "step": 175900 }, { "epoch": 30.273571920165175, "grad_norm": 8.970932006835938, "learning_rate": 3.486321403991741e-05, "loss": 0.8505, "step": 175950 }, { "epoch": 30.282174810736407, "grad_norm": 11.558496475219727, "learning_rate": 3.48589125946318e-05, "loss": 0.8541, "step": 176000 }, { "epoch": 30.29077770130764, "grad_norm": 9.797102928161621, "learning_rate": 3.4854611149346186e-05, "loss": 0.8955, "step": 176050 }, { "epoch": 30.299380591878872, "grad_norm": 6.630373954772949, "learning_rate": 3.485030970406057e-05, "loss": 0.8679, "step": 176100 }, { "epoch": 30.307983482450105, "grad_norm": 12.067896842956543, "learning_rate": 3.484600825877495e-05, "loss": 0.8958, "step": 176150 }, { "epoch": 30.316586373021334, "grad_norm": 10.259530067443848, "learning_rate": 3.4841706813489337e-05, "loss": 0.9393, "step": 176200 }, { "epoch": 30.325189263592566, "grad_norm": 11.977352142333984, "learning_rate": 3.483740536820372e-05, "loss": 0.8773, "step": 176250 }, { "epoch": 30.3337921541638, "grad_norm": 9.486679077148438, "learning_rate": 3.4833103922918106e-05, "loss": 0.909, "step": 176300 }, { "epoch": 30.34239504473503, "grad_norm": 12.71806812286377, "learning_rate": 3.482880247763249e-05, "loss": 0.8817, "step": 176350 }, { "epoch": 30.350997935306264, "grad_norm": 11.308537483215332, "learning_rate": 3.482450103234687e-05, "loss": 0.9061, "step": 176400 }, { "epoch": 30.359600825877497, "grad_norm": 13.78849983215332, "learning_rate": 3.4820199587061256e-05, "loss": 0.8891, "step": 176450 }, { "epoch": 30.368203716448726, "grad_norm": 12.59046745300293, "learning_rate": 3.481589814177564e-05, "loss": 0.9377, "step": 176500 }, { "epoch": 30.376806607019958, "grad_norm": 12.875066757202148, "learning_rate": 3.4811596696490025e-05, "loss": 0.823, "step": 176550 }, { "epoch": 30.38540949759119, "grad_norm": 12.248563766479492, "learning_rate": 3.4807295251204406e-05, "loss": 0.7938, "step": 176600 }, { "epoch": 30.394012388162423, "grad_norm": 9.840903282165527, "learning_rate": 3.480299380591879e-05, "loss": 0.8823, "step": 176650 }, { "epoch": 30.402615278733656, "grad_norm": 16.578508377075195, "learning_rate": 3.4798692360633175e-05, "loss": 0.8806, "step": 176700 }, { "epoch": 30.411218169304888, "grad_norm": 13.521478652954102, "learning_rate": 3.4794390915347557e-05, "loss": 0.8864, "step": 176750 }, { "epoch": 30.419821059876117, "grad_norm": 9.055959701538086, "learning_rate": 3.4790089470061945e-05, "loss": 0.883, "step": 176800 }, { "epoch": 30.42842395044735, "grad_norm": 7.282461643218994, "learning_rate": 3.4785788024776326e-05, "loss": 0.8904, "step": 176850 }, { "epoch": 30.437026841018582, "grad_norm": 11.535749435424805, "learning_rate": 3.4781486579490714e-05, "loss": 0.8962, "step": 176900 }, { "epoch": 30.445629731589815, "grad_norm": 8.567361831665039, "learning_rate": 3.4777185134205095e-05, "loss": 0.9256, "step": 176950 }, { "epoch": 30.454232622161047, "grad_norm": 10.453493118286133, "learning_rate": 3.4772883688919476e-05, "loss": 0.8758, "step": 177000 }, { "epoch": 30.462835512732276, "grad_norm": 8.817400932312012, "learning_rate": 3.4768582243633864e-05, "loss": 0.8763, "step": 177050 }, { "epoch": 30.47143840330351, "grad_norm": 10.58356761932373, "learning_rate": 3.4764280798348245e-05, "loss": 0.8799, "step": 177100 }, { "epoch": 30.48004129387474, "grad_norm": 10.533042907714844, "learning_rate": 3.475997935306263e-05, "loss": 0.849, "step": 177150 }, { "epoch": 30.488644184445974, "grad_norm": 16.766727447509766, "learning_rate": 3.4755677907777014e-05, "loss": 0.8984, "step": 177200 }, { "epoch": 30.497247075017206, "grad_norm": 8.498181343078613, "learning_rate": 3.4751376462491395e-05, "loss": 0.8183, "step": 177250 }, { "epoch": 30.50584996558844, "grad_norm": 12.801713943481445, "learning_rate": 3.474707501720578e-05, "loss": 0.8656, "step": 177300 }, { "epoch": 30.514452856159668, "grad_norm": 11.576750755310059, "learning_rate": 3.4742773571920165e-05, "loss": 0.8352, "step": 177350 }, { "epoch": 30.5230557467309, "grad_norm": 9.983803749084473, "learning_rate": 3.473847212663455e-05, "loss": 0.9167, "step": 177400 }, { "epoch": 30.531658637302133, "grad_norm": 15.404607772827148, "learning_rate": 3.473417068134894e-05, "loss": 0.8723, "step": 177450 }, { "epoch": 30.540261527873366, "grad_norm": 7.813915252685547, "learning_rate": 3.4729869236063315e-05, "loss": 0.886, "step": 177500 }, { "epoch": 30.548864418444598, "grad_norm": 12.990458488464355, "learning_rate": 3.47255677907777e-05, "loss": 0.8955, "step": 177550 }, { "epoch": 30.55746730901583, "grad_norm": 10.960675239562988, "learning_rate": 3.4721266345492084e-05, "loss": 0.9132, "step": 177600 }, { "epoch": 30.56607019958706, "grad_norm": 9.430967330932617, "learning_rate": 3.471696490020647e-05, "loss": 0.8702, "step": 177650 }, { "epoch": 30.574673090158292, "grad_norm": 15.221870422363281, "learning_rate": 3.471266345492086e-05, "loss": 0.9068, "step": 177700 }, { "epoch": 30.583275980729525, "grad_norm": 8.335898399353027, "learning_rate": 3.4708362009635234e-05, "loss": 0.8422, "step": 177750 }, { "epoch": 30.591878871300757, "grad_norm": 15.346182823181152, "learning_rate": 3.470406056434962e-05, "loss": 0.9087, "step": 177800 }, { "epoch": 30.60048176187199, "grad_norm": 10.7843656539917, "learning_rate": 3.469975911906401e-05, "loss": 0.8401, "step": 177850 }, { "epoch": 30.609084652443222, "grad_norm": 7.866519451141357, "learning_rate": 3.469545767377839e-05, "loss": 0.8843, "step": 177900 }, { "epoch": 30.61768754301445, "grad_norm": 8.035837173461914, "learning_rate": 3.469115622849278e-05, "loss": 0.8254, "step": 177950 }, { "epoch": 30.626290433585684, "grad_norm": 10.450389862060547, "learning_rate": 3.4686854783207154e-05, "loss": 0.9025, "step": 178000 }, { "epoch": 30.634893324156916, "grad_norm": 10.5713472366333, "learning_rate": 3.468255333792154e-05, "loss": 0.916, "step": 178050 }, { "epoch": 30.64349621472815, "grad_norm": 9.751781463623047, "learning_rate": 3.467825189263593e-05, "loss": 0.8848, "step": 178100 }, { "epoch": 30.65209910529938, "grad_norm": 11.11854362487793, "learning_rate": 3.467395044735031e-05, "loss": 0.9032, "step": 178150 }, { "epoch": 30.660701995870614, "grad_norm": 8.956361770629883, "learning_rate": 3.46696490020647e-05, "loss": 0.865, "step": 178200 }, { "epoch": 30.669304886441843, "grad_norm": 19.155094146728516, "learning_rate": 3.466534755677908e-05, "loss": 0.8096, "step": 178250 }, { "epoch": 30.677907777013075, "grad_norm": 10.964194297790527, "learning_rate": 3.466104611149346e-05, "loss": 0.8773, "step": 178300 }, { "epoch": 30.686510667584308, "grad_norm": 11.244915008544922, "learning_rate": 3.465674466620785e-05, "loss": 0.9212, "step": 178350 }, { "epoch": 30.69511355815554, "grad_norm": 11.017247200012207, "learning_rate": 3.465244322092223e-05, "loss": 0.8591, "step": 178400 }, { "epoch": 30.703716448726773, "grad_norm": 8.421277046203613, "learning_rate": 3.464814177563662e-05, "loss": 0.8525, "step": 178450 }, { "epoch": 30.712319339298006, "grad_norm": 23.024417877197266, "learning_rate": 3.4643840330351e-05, "loss": 0.8485, "step": 178500 }, { "epoch": 30.720922229869235, "grad_norm": 20.762685775756836, "learning_rate": 3.463953888506538e-05, "loss": 0.8787, "step": 178550 }, { "epoch": 30.729525120440467, "grad_norm": 11.432365417480469, "learning_rate": 3.463523743977977e-05, "loss": 0.8534, "step": 178600 }, { "epoch": 30.7381280110117, "grad_norm": 14.926942825317383, "learning_rate": 3.463093599449415e-05, "loss": 0.8893, "step": 178650 }, { "epoch": 30.746730901582932, "grad_norm": 10.498910903930664, "learning_rate": 3.462663454920854e-05, "loss": 0.9167, "step": 178700 }, { "epoch": 30.755333792154165, "grad_norm": 12.741588592529297, "learning_rate": 3.462233310392292e-05, "loss": 0.92, "step": 178750 }, { "epoch": 30.763936682725397, "grad_norm": 7.046477794647217, "learning_rate": 3.46180316586373e-05, "loss": 0.9186, "step": 178800 }, { "epoch": 30.772539573296626, "grad_norm": 9.612916946411133, "learning_rate": 3.461373021335169e-05, "loss": 0.9081, "step": 178850 }, { "epoch": 30.78114246386786, "grad_norm": 8.24023723602295, "learning_rate": 3.460942876806607e-05, "loss": 0.8886, "step": 178900 }, { "epoch": 30.78974535443909, "grad_norm": 11.124068260192871, "learning_rate": 3.460512732278046e-05, "loss": 0.8693, "step": 178950 }, { "epoch": 30.798348245010324, "grad_norm": 12.62869644165039, "learning_rate": 3.460082587749484e-05, "loss": 0.8223, "step": 179000 }, { "epoch": 30.806951135581556, "grad_norm": 10.604713439941406, "learning_rate": 3.4596524432209226e-05, "loss": 0.9072, "step": 179050 }, { "epoch": 30.81555402615279, "grad_norm": 7.648578643798828, "learning_rate": 3.459222298692361e-05, "loss": 0.849, "step": 179100 }, { "epoch": 30.824156916724018, "grad_norm": 12.38643741607666, "learning_rate": 3.458792154163799e-05, "loss": 0.8938, "step": 179150 }, { "epoch": 30.83275980729525, "grad_norm": 9.481351852416992, "learning_rate": 3.4583620096352376e-05, "loss": 0.9124, "step": 179200 }, { "epoch": 30.841362697866483, "grad_norm": 10.522784233093262, "learning_rate": 3.4579318651066764e-05, "loss": 0.8956, "step": 179250 }, { "epoch": 30.849965588437716, "grad_norm": 9.174946784973145, "learning_rate": 3.4575017205781145e-05, "loss": 0.8266, "step": 179300 }, { "epoch": 30.858568479008948, "grad_norm": 13.58225154876709, "learning_rate": 3.4570715760495527e-05, "loss": 0.8857, "step": 179350 }, { "epoch": 30.86717136958018, "grad_norm": 11.843667030334473, "learning_rate": 3.456641431520991e-05, "loss": 0.844, "step": 179400 }, { "epoch": 30.87577426015141, "grad_norm": 12.02992057800293, "learning_rate": 3.4562112869924296e-05, "loss": 0.9332, "step": 179450 }, { "epoch": 30.884377150722642, "grad_norm": 11.653255462646484, "learning_rate": 3.4557811424638684e-05, "loss": 0.8944, "step": 179500 }, { "epoch": 30.892980041293875, "grad_norm": 13.023357391357422, "learning_rate": 3.4553509979353065e-05, "loss": 0.914, "step": 179550 }, { "epoch": 30.901582931865107, "grad_norm": 12.230290412902832, "learning_rate": 3.454920853406745e-05, "loss": 0.8982, "step": 179600 }, { "epoch": 30.91018582243634, "grad_norm": 11.07699203491211, "learning_rate": 3.454490708878183e-05, "loss": 0.878, "step": 179650 }, { "epoch": 30.918788713007572, "grad_norm": 10.325149536132812, "learning_rate": 3.4540605643496215e-05, "loss": 0.8374, "step": 179700 }, { "epoch": 30.9273916035788, "grad_norm": 9.404661178588867, "learning_rate": 3.45363041982106e-05, "loss": 0.9092, "step": 179750 }, { "epoch": 30.935994494150034, "grad_norm": 10.659913063049316, "learning_rate": 3.4532002752924984e-05, "loss": 0.8549, "step": 179800 }, { "epoch": 30.944597384721266, "grad_norm": 13.091554641723633, "learning_rate": 3.452770130763937e-05, "loss": 0.873, "step": 179850 }, { "epoch": 30.9532002752925, "grad_norm": 9.922138214111328, "learning_rate": 3.452339986235375e-05, "loss": 0.8718, "step": 179900 }, { "epoch": 30.96180316586373, "grad_norm": 12.632086753845215, "learning_rate": 3.4519098417068135e-05, "loss": 0.8427, "step": 179950 }, { "epoch": 30.97040605643496, "grad_norm": 12.107108116149902, "learning_rate": 3.451479697178252e-05, "loss": 0.8835, "step": 180000 }, { "epoch": 30.979008947006193, "grad_norm": 9.630020141601562, "learning_rate": 3.4510495526496904e-05, "loss": 0.9087, "step": 180050 }, { "epoch": 30.987611837577425, "grad_norm": 8.743014335632324, "learning_rate": 3.450619408121129e-05, "loss": 0.8915, "step": 180100 }, { "epoch": 30.996214728148658, "grad_norm": 10.99495792388916, "learning_rate": 3.450189263592567e-05, "loss": 0.8657, "step": 180150 }, { "epoch": 31.0, "eval_accuracy": 0.5604026845637584, "eval_f1": 0.5486806087811826, "eval_f1_DuraRiadoRio_16x16": 0.30428863172226006, "eval_f1_Mole_16x16": 0.6059870550161812, "eval_f1_Quebrado_16x16": 0.7313534345953299, "eval_f1_RiadoRio_16x16": 0.47543329623151537, "eval_f1_RioFechado_16x16": 0.6263406263406264, "eval_loss": 1.5598013401031494, "eval_precision": 0.6076076086994544, "eval_precision_DuraRiadoRio_16x16": 0.7050473186119873, "eval_precision_Mole_16x16": 0.5674242424242424, "eval_precision_Quebrado_16x16": 0.7655434266729948, "eval_precision_RiadoRio_16x16": 0.3876069484054965, "eval_precision_RioFechado_16x16": 0.6124161073825504, "eval_recall": 0.5599808619441492, "eval_recall_DuraRiadoRio_16x16": 0.19401041666666666, "eval_recall_Mole_16x16": 0.6501736111111112, "eval_recall_Quebrado_16x16": 0.7000868055555556, "eval_recall_RiadoRio_16x16": 0.6147203947368421, "eval_recall_RioFechado_16x16": 0.6409130816505707, "eval_runtime": 46.3471, "eval_samples_per_second": 250.76, "eval_steps_per_second": 15.686, "step": 180172 }, { "epoch": 31.00481761871989, "grad_norm": 8.51795482635498, "learning_rate": 3.4497591190640054e-05, "loss": 0.887, "step": 180200 }, { "epoch": 31.013420509291123, "grad_norm": 7.459472179412842, "learning_rate": 3.449328974535444e-05, "loss": 0.8936, "step": 180250 }, { "epoch": 31.022023399862352, "grad_norm": 10.595137596130371, "learning_rate": 3.448898830006882e-05, "loss": 0.8321, "step": 180300 }, { "epoch": 31.030626290433585, "grad_norm": 12.209539413452148, "learning_rate": 3.448468685478321e-05, "loss": 0.8944, "step": 180350 }, { "epoch": 31.039229181004817, "grad_norm": 10.020302772521973, "learning_rate": 3.448038540949759e-05, "loss": 0.8863, "step": 180400 }, { "epoch": 31.04783207157605, "grad_norm": 12.370579719543457, "learning_rate": 3.447608396421197e-05, "loss": 0.8938, "step": 180450 }, { "epoch": 31.056434962147282, "grad_norm": 9.406523704528809, "learning_rate": 3.447178251892636e-05, "loss": 0.9148, "step": 180500 }, { "epoch": 31.065037852718515, "grad_norm": 7.83812141418457, "learning_rate": 3.446748107364074e-05, "loss": 0.8847, "step": 180550 }, { "epoch": 31.073640743289744, "grad_norm": 8.736783981323242, "learning_rate": 3.446317962835513e-05, "loss": 0.8244, "step": 180600 }, { "epoch": 31.082243633860976, "grad_norm": 11.557326316833496, "learning_rate": 3.445887818306952e-05, "loss": 0.8977, "step": 180650 }, { "epoch": 31.09084652443221, "grad_norm": 13.117768287658691, "learning_rate": 3.445457673778389e-05, "loss": 0.8841, "step": 180700 }, { "epoch": 31.09944941500344, "grad_norm": 7.975518703460693, "learning_rate": 3.445027529249828e-05, "loss": 0.8308, "step": 180750 }, { "epoch": 31.108052305574674, "grad_norm": 9.087517738342285, "learning_rate": 3.444597384721266e-05, "loss": 0.8356, "step": 180800 }, { "epoch": 31.116655196145906, "grad_norm": 14.561482429504395, "learning_rate": 3.444167240192705e-05, "loss": 0.8622, "step": 180850 }, { "epoch": 31.125258086717135, "grad_norm": 8.425077438354492, "learning_rate": 3.443737095664144e-05, "loss": 0.802, "step": 180900 }, { "epoch": 31.133860977288368, "grad_norm": 12.845963478088379, "learning_rate": 3.443306951135582e-05, "loss": 0.8361, "step": 180950 }, { "epoch": 31.1424638678596, "grad_norm": 15.879399299621582, "learning_rate": 3.44287680660702e-05, "loss": 0.8549, "step": 181000 }, { "epoch": 31.151066758430833, "grad_norm": 11.29841423034668, "learning_rate": 3.442446662078458e-05, "loss": 0.8845, "step": 181050 }, { "epoch": 31.159669649002065, "grad_norm": 9.184696197509766, "learning_rate": 3.442016517549897e-05, "loss": 0.8798, "step": 181100 }, { "epoch": 31.168272539573298, "grad_norm": 9.433289527893066, "learning_rate": 3.441586373021336e-05, "loss": 0.8505, "step": 181150 }, { "epoch": 31.176875430144527, "grad_norm": 8.492671012878418, "learning_rate": 3.441156228492774e-05, "loss": 0.8085, "step": 181200 }, { "epoch": 31.18547832071576, "grad_norm": 9.358036041259766, "learning_rate": 3.440726083964212e-05, "loss": 0.8733, "step": 181250 }, { "epoch": 31.194081211286992, "grad_norm": 11.389634132385254, "learning_rate": 3.440295939435651e-05, "loss": 0.8338, "step": 181300 }, { "epoch": 31.202684101858225, "grad_norm": 10.842451095581055, "learning_rate": 3.439865794907089e-05, "loss": 0.8225, "step": 181350 }, { "epoch": 31.211286992429457, "grad_norm": 11.28657054901123, "learning_rate": 3.439435650378528e-05, "loss": 0.8218, "step": 181400 }, { "epoch": 31.21988988300069, "grad_norm": 13.833444595336914, "learning_rate": 3.439005505849966e-05, "loss": 0.8781, "step": 181450 }, { "epoch": 31.22849277357192, "grad_norm": 12.008960723876953, "learning_rate": 3.438575361321404e-05, "loss": 0.8368, "step": 181500 }, { "epoch": 31.23709566414315, "grad_norm": 8.585112571716309, "learning_rate": 3.438145216792843e-05, "loss": 0.8849, "step": 181550 }, { "epoch": 31.245698554714384, "grad_norm": 11.187397956848145, "learning_rate": 3.437715072264281e-05, "loss": 0.8705, "step": 181600 }, { "epoch": 31.254301445285616, "grad_norm": 12.844679832458496, "learning_rate": 3.4372849277357196e-05, "loss": 0.8948, "step": 181650 }, { "epoch": 31.26290433585685, "grad_norm": 9.057875633239746, "learning_rate": 3.436854783207158e-05, "loss": 0.9203, "step": 181700 }, { "epoch": 31.27150722642808, "grad_norm": 12.383111953735352, "learning_rate": 3.4364246386785965e-05, "loss": 0.9184, "step": 181750 }, { "epoch": 31.28011011699931, "grad_norm": 18.525346755981445, "learning_rate": 3.4359944941500346e-05, "loss": 0.8497, "step": 181800 }, { "epoch": 31.288713007570543, "grad_norm": 9.537247657775879, "learning_rate": 3.435564349621473e-05, "loss": 0.863, "step": 181850 }, { "epoch": 31.297315898141775, "grad_norm": 6.9219489097595215, "learning_rate": 3.4351342050929115e-05, "loss": 0.814, "step": 181900 }, { "epoch": 31.305918788713008, "grad_norm": 8.996870994567871, "learning_rate": 3.43470406056435e-05, "loss": 0.8875, "step": 181950 }, { "epoch": 31.31452167928424, "grad_norm": 8.666257858276367, "learning_rate": 3.4342739160357885e-05, "loss": 0.8513, "step": 182000 }, { "epoch": 31.323124569855473, "grad_norm": 9.52623462677002, "learning_rate": 3.4338437715072266e-05, "loss": 0.8545, "step": 182050 }, { "epoch": 31.331727460426702, "grad_norm": 8.942296028137207, "learning_rate": 3.433413626978665e-05, "loss": 0.8762, "step": 182100 }, { "epoch": 31.340330350997935, "grad_norm": 10.144529342651367, "learning_rate": 3.4329834824501035e-05, "loss": 0.8805, "step": 182150 }, { "epoch": 31.348933241569167, "grad_norm": 13.878555297851562, "learning_rate": 3.4325533379215416e-05, "loss": 0.9049, "step": 182200 }, { "epoch": 31.3575361321404, "grad_norm": 13.898591041564941, "learning_rate": 3.4321231933929804e-05, "loss": 0.907, "step": 182250 }, { "epoch": 31.366139022711632, "grad_norm": 12.248201370239258, "learning_rate": 3.4316930488644185e-05, "loss": 0.9179, "step": 182300 }, { "epoch": 31.374741913282865, "grad_norm": 10.030795097351074, "learning_rate": 3.4312629043358566e-05, "loss": 0.8786, "step": 182350 }, { "epoch": 31.383344803854094, "grad_norm": 12.654565811157227, "learning_rate": 3.4308327598072954e-05, "loss": 0.8671, "step": 182400 }, { "epoch": 31.391947694425326, "grad_norm": 12.052231788635254, "learning_rate": 3.4304026152787335e-05, "loss": 0.8167, "step": 182450 }, { "epoch": 31.40055058499656, "grad_norm": 9.62735366821289, "learning_rate": 3.4299724707501723e-05, "loss": 0.8144, "step": 182500 }, { "epoch": 31.40915347556779, "grad_norm": 11.00317668914795, "learning_rate": 3.429542326221611e-05, "loss": 0.9246, "step": 182550 }, { "epoch": 31.417756366139024, "grad_norm": 12.362236022949219, "learning_rate": 3.4291121816930486e-05, "loss": 0.849, "step": 182600 }, { "epoch": 31.426359256710256, "grad_norm": 9.107636451721191, "learning_rate": 3.4286820371644874e-05, "loss": 0.8576, "step": 182650 }, { "epoch": 31.434962147281485, "grad_norm": 14.06654167175293, "learning_rate": 3.428251892635926e-05, "loss": 0.8919, "step": 182700 }, { "epoch": 31.443565037852718, "grad_norm": 13.124178886413574, "learning_rate": 3.427821748107364e-05, "loss": 0.9063, "step": 182750 }, { "epoch": 31.45216792842395, "grad_norm": 11.174102783203125, "learning_rate": 3.427391603578803e-05, "loss": 0.861, "step": 182800 }, { "epoch": 31.460770818995183, "grad_norm": 9.072160720825195, "learning_rate": 3.4269614590502405e-05, "loss": 0.8843, "step": 182850 }, { "epoch": 31.469373709566415, "grad_norm": 12.708187103271484, "learning_rate": 3.426531314521679e-05, "loss": 0.8393, "step": 182900 }, { "epoch": 31.477976600137648, "grad_norm": 13.106014251708984, "learning_rate": 3.426101169993118e-05, "loss": 0.8326, "step": 182950 }, { "epoch": 31.486579490708877, "grad_norm": 12.417990684509277, "learning_rate": 3.425671025464556e-05, "loss": 0.895, "step": 183000 }, { "epoch": 31.49518238128011, "grad_norm": 11.091959953308105, "learning_rate": 3.425240880935995e-05, "loss": 0.8821, "step": 183050 }, { "epoch": 31.503785271851342, "grad_norm": 14.368752479553223, "learning_rate": 3.424810736407433e-05, "loss": 0.8427, "step": 183100 }, { "epoch": 31.512388162422575, "grad_norm": 8.028721809387207, "learning_rate": 3.424380591878871e-05, "loss": 0.8649, "step": 183150 }, { "epoch": 31.520991052993807, "grad_norm": 7.962997913360596, "learning_rate": 3.42395044735031e-05, "loss": 0.834, "step": 183200 }, { "epoch": 31.52959394356504, "grad_norm": 13.993117332458496, "learning_rate": 3.423520302821748e-05, "loss": 0.8863, "step": 183250 }, { "epoch": 31.53819683413627, "grad_norm": 6.866849422454834, "learning_rate": 3.423090158293187e-05, "loss": 0.874, "step": 183300 }, { "epoch": 31.5467997247075, "grad_norm": 9.718472480773926, "learning_rate": 3.422660013764625e-05, "loss": 0.895, "step": 183350 }, { "epoch": 31.555402615278734, "grad_norm": 12.189906120300293, "learning_rate": 3.422229869236063e-05, "loss": 0.8494, "step": 183400 }, { "epoch": 31.564005505849966, "grad_norm": 6.3189873695373535, "learning_rate": 3.421799724707502e-05, "loss": 0.8134, "step": 183450 }, { "epoch": 31.5726083964212, "grad_norm": 14.317989349365234, "learning_rate": 3.42136958017894e-05, "loss": 0.9025, "step": 183500 }, { "epoch": 31.581211286992428, "grad_norm": 9.150619506835938, "learning_rate": 3.420939435650379e-05, "loss": 0.8513, "step": 183550 }, { "epoch": 31.58981417756366, "grad_norm": 10.079627990722656, "learning_rate": 3.420509291121817e-05, "loss": 0.8789, "step": 183600 }, { "epoch": 31.598417068134893, "grad_norm": 11.987564086914062, "learning_rate": 3.420079146593255e-05, "loss": 0.8465, "step": 183650 }, { "epoch": 31.607019958706125, "grad_norm": 8.814218521118164, "learning_rate": 3.419649002064694e-05, "loss": 0.8934, "step": 183700 }, { "epoch": 31.615622849277358, "grad_norm": 9.422844886779785, "learning_rate": 3.419218857536132e-05, "loss": 0.8789, "step": 183750 }, { "epoch": 31.62422573984859, "grad_norm": 9.19926643371582, "learning_rate": 3.418788713007571e-05, "loss": 0.8679, "step": 183800 }, { "epoch": 31.63282863041982, "grad_norm": 9.416884422302246, "learning_rate": 3.418358568479009e-05, "loss": 0.893, "step": 183850 }, { "epoch": 31.641431520991052, "grad_norm": 15.531486511230469, "learning_rate": 3.417928423950448e-05, "loss": 0.9125, "step": 183900 }, { "epoch": 31.650034411562284, "grad_norm": 11.072798728942871, "learning_rate": 3.417498279421886e-05, "loss": 0.8673, "step": 183950 }, { "epoch": 31.658637302133517, "grad_norm": 9.392868995666504, "learning_rate": 3.417068134893324e-05, "loss": 0.898, "step": 184000 }, { "epoch": 31.66724019270475, "grad_norm": 12.575972557067871, "learning_rate": 3.416637990364763e-05, "loss": 0.9039, "step": 184050 }, { "epoch": 31.675843083275982, "grad_norm": 12.036766052246094, "learning_rate": 3.416207845836201e-05, "loss": 0.8544, "step": 184100 }, { "epoch": 31.68444597384721, "grad_norm": 11.104174613952637, "learning_rate": 3.41577770130764e-05, "loss": 0.9781, "step": 184150 }, { "epoch": 31.693048864418444, "grad_norm": 11.652667045593262, "learning_rate": 3.415347556779078e-05, "loss": 0.8375, "step": 184200 }, { "epoch": 31.701651754989676, "grad_norm": 10.907254219055176, "learning_rate": 3.414917412250516e-05, "loss": 0.8281, "step": 184250 }, { "epoch": 31.71025464556091, "grad_norm": 14.30970573425293, "learning_rate": 3.414487267721955e-05, "loss": 0.8936, "step": 184300 }, { "epoch": 31.71885753613214, "grad_norm": 7.531245231628418, "learning_rate": 3.4140571231933935e-05, "loss": 0.8761, "step": 184350 }, { "epoch": 31.727460426703374, "grad_norm": 11.012463569641113, "learning_rate": 3.4136269786648316e-05, "loss": 0.8943, "step": 184400 }, { "epoch": 31.736063317274603, "grad_norm": 8.841084480285645, "learning_rate": 3.4131968341362704e-05, "loss": 0.8744, "step": 184450 }, { "epoch": 31.744666207845835, "grad_norm": 13.298781394958496, "learning_rate": 3.412766689607708e-05, "loss": 0.8592, "step": 184500 }, { "epoch": 31.753269098417068, "grad_norm": 8.305105209350586, "learning_rate": 3.412336545079147e-05, "loss": 0.8603, "step": 184550 }, { "epoch": 31.7618719889883, "grad_norm": 9.2412748336792, "learning_rate": 3.4119064005505855e-05, "loss": 0.9236, "step": 184600 }, { "epoch": 31.770474879559533, "grad_norm": 9.875741004943848, "learning_rate": 3.4114762560220236e-05, "loss": 0.8048, "step": 184650 }, { "epoch": 31.779077770130765, "grad_norm": 21.59575843811035, "learning_rate": 3.4110461114934624e-05, "loss": 0.8942, "step": 184700 }, { "epoch": 31.787680660701994, "grad_norm": 10.407164573669434, "learning_rate": 3.4106159669649005e-05, "loss": 0.9386, "step": 184750 }, { "epoch": 31.796283551273227, "grad_norm": 17.40414047241211, "learning_rate": 3.4101858224363386e-05, "loss": 0.9271, "step": 184800 }, { "epoch": 31.80488644184446, "grad_norm": 8.305562019348145, "learning_rate": 3.4097556779077774e-05, "loss": 0.8464, "step": 184850 }, { "epoch": 31.813489332415692, "grad_norm": 14.609291076660156, "learning_rate": 3.4093255333792155e-05, "loss": 0.8713, "step": 184900 }, { "epoch": 31.822092222986925, "grad_norm": 11.87563705444336, "learning_rate": 3.408895388850654e-05, "loss": 0.8925, "step": 184950 }, { "epoch": 31.830695113558157, "grad_norm": 11.745320320129395, "learning_rate": 3.4084652443220924e-05, "loss": 0.8758, "step": 185000 }, { "epoch": 31.839298004129386, "grad_norm": 10.236390113830566, "learning_rate": 3.4080350997935305e-05, "loss": 0.8869, "step": 185050 }, { "epoch": 31.84790089470062, "grad_norm": 11.038844108581543, "learning_rate": 3.4076049552649693e-05, "loss": 0.9843, "step": 185100 }, { "epoch": 31.85650378527185, "grad_norm": 8.567197799682617, "learning_rate": 3.4071748107364075e-05, "loss": 0.858, "step": 185150 }, { "epoch": 31.865106675843084, "grad_norm": 12.069087982177734, "learning_rate": 3.406744666207846e-05, "loss": 0.8779, "step": 185200 }, { "epoch": 31.873709566414316, "grad_norm": 13.92436408996582, "learning_rate": 3.4063145216792844e-05, "loss": 0.9144, "step": 185250 }, { "epoch": 31.88231245698555, "grad_norm": 8.071537017822266, "learning_rate": 3.4058843771507225e-05, "loss": 0.8082, "step": 185300 }, { "epoch": 31.890915347556778, "grad_norm": 12.524648666381836, "learning_rate": 3.405454232622161e-05, "loss": 0.8509, "step": 185350 }, { "epoch": 31.89951823812801, "grad_norm": 10.308311462402344, "learning_rate": 3.4050240880935994e-05, "loss": 0.8296, "step": 185400 }, { "epoch": 31.908121128699243, "grad_norm": 13.551347732543945, "learning_rate": 3.404593943565038e-05, "loss": 0.8735, "step": 185450 }, { "epoch": 31.916724019270475, "grad_norm": 14.347978591918945, "learning_rate": 3.404163799036476e-05, "loss": 0.885, "step": 185500 }, { "epoch": 31.925326909841708, "grad_norm": 8.502416610717773, "learning_rate": 3.4037336545079144e-05, "loss": 0.9028, "step": 185550 }, { "epoch": 31.93392980041294, "grad_norm": 13.302851676940918, "learning_rate": 3.403303509979353e-05, "loss": 0.8837, "step": 185600 }, { "epoch": 31.94253269098417, "grad_norm": 11.731345176696777, "learning_rate": 3.4028733654507913e-05, "loss": 0.8473, "step": 185650 }, { "epoch": 31.951135581555402, "grad_norm": 11.68016529083252, "learning_rate": 3.40244322092223e-05, "loss": 0.8954, "step": 185700 }, { "epoch": 31.959738472126634, "grad_norm": 13.042068481445312, "learning_rate": 3.402013076393669e-05, "loss": 0.8671, "step": 185750 }, { "epoch": 31.968341362697867, "grad_norm": 12.504968643188477, "learning_rate": 3.4015829318651064e-05, "loss": 0.8469, "step": 185800 }, { "epoch": 31.9769442532691, "grad_norm": 15.3025541305542, "learning_rate": 3.401152787336545e-05, "loss": 0.8798, "step": 185850 }, { "epoch": 31.985547143840332, "grad_norm": 13.445629119873047, "learning_rate": 3.400722642807983e-05, "loss": 0.8845, "step": 185900 }, { "epoch": 31.99415003441156, "grad_norm": 12.099645614624023, "learning_rate": 3.400292498279422e-05, "loss": 0.9113, "step": 185950 }, { "epoch": 32.0, "eval_accuracy": 0.5470659094820168, "eval_f1": 0.5393414732099125, "eval_f1_DuraRiadoRio_16x16": 0.4763563962170234, "eval_f1_Mole_16x16": 0.5841184387617766, "eval_f1_Quebrado_16x16": 0.6857635149687702, "eval_f1_RiadoRio_16x16": 0.3844984802431611, "eval_f1_RioFechado_16x16": 0.5659705358588312, "eval_loss": 1.5645190477371216, "eval_precision": 0.5584824686811936, "eval_precision_DuraRiadoRio_16x16": 0.558343057176196, "eval_precision_Mole_16x16": 0.6044568245125348, "eval_precision_Quebrado_16x16": 0.6806327490380505, "eval_precision_RiadoRio_16x16": 0.5006596306068601, "eval_precision_RioFechado_16x16": 0.44832008207232626, "eval_recall": 0.5501739119482567, "eval_recall_DuraRiadoRio_16x16": 0.4153645833333333, "eval_recall_Mole_16x16": 0.5651041666666666, "eval_recall_Quebrado_16x16": 0.6909722222222222, "eval_recall_RiadoRio_16x16": 0.3120888157894737, "eval_recall_RioFechado_16x16": 0.7673397717295873, "eval_runtime": 46.1129, "eval_samples_per_second": 252.034, "eval_steps_per_second": 15.766, "step": 185984 }, { "epoch": 32.002752924982794, "grad_norm": 10.435075759887695, "learning_rate": 3.399862353750861e-05, "loss": 0.8334, "step": 186000 }, { "epoch": 32.011355815554026, "grad_norm": 17.234420776367188, "learning_rate": 3.399432209222299e-05, "loss": 0.8539, "step": 186050 }, { "epoch": 32.01995870612526, "grad_norm": 10.052386283874512, "learning_rate": 3.399002064693737e-05, "loss": 0.8686, "step": 186100 }, { "epoch": 32.02856159669649, "grad_norm": 15.339449882507324, "learning_rate": 3.398571920165175e-05, "loss": 0.8516, "step": 186150 }, { "epoch": 32.037164487267724, "grad_norm": 13.94561767578125, "learning_rate": 3.398141775636614e-05, "loss": 0.8437, "step": 186200 }, { "epoch": 32.045767377838956, "grad_norm": 14.484305381774902, "learning_rate": 3.397711631108053e-05, "loss": 0.8746, "step": 186250 }, { "epoch": 32.05437026841019, "grad_norm": 9.901473045349121, "learning_rate": 3.397281486579491e-05, "loss": 0.8613, "step": 186300 }, { "epoch": 32.06297315898142, "grad_norm": 7.9248456954956055, "learning_rate": 3.396851342050929e-05, "loss": 0.8383, "step": 186350 }, { "epoch": 32.07157604955265, "grad_norm": 10.483495712280273, "learning_rate": 3.396421197522368e-05, "loss": 0.888, "step": 186400 }, { "epoch": 32.08017894012388, "grad_norm": 10.151693344116211, "learning_rate": 3.395991052993806e-05, "loss": 0.8556, "step": 186450 }, { "epoch": 32.08878183069511, "grad_norm": 8.06038761138916, "learning_rate": 3.395560908465245e-05, "loss": 0.8809, "step": 186500 }, { "epoch": 32.097384721266344, "grad_norm": 7.418259620666504, "learning_rate": 3.395130763936683e-05, "loss": 0.8799, "step": 186550 }, { "epoch": 32.10598761183758, "grad_norm": 12.313608169555664, "learning_rate": 3.394700619408122e-05, "loss": 0.8796, "step": 186600 }, { "epoch": 32.11459050240881, "grad_norm": 13.476503372192383, "learning_rate": 3.39427047487956e-05, "loss": 0.8168, "step": 186650 }, { "epoch": 32.12319339298004, "grad_norm": 9.883244514465332, "learning_rate": 3.393840330350998e-05, "loss": 0.8781, "step": 186700 }, { "epoch": 32.131796283551274, "grad_norm": 10.70998764038086, "learning_rate": 3.393410185822437e-05, "loss": 0.864, "step": 186750 }, { "epoch": 32.14039917412251, "grad_norm": 9.888232231140137, "learning_rate": 3.392980041293875e-05, "loss": 0.8509, "step": 186800 }, { "epoch": 32.14900206469374, "grad_norm": 13.30727767944336, "learning_rate": 3.3925498967653136e-05, "loss": 0.8525, "step": 186850 }, { "epoch": 32.15760495526497, "grad_norm": 10.06889820098877, "learning_rate": 3.392119752236752e-05, "loss": 0.8731, "step": 186900 }, { "epoch": 32.1662078458362, "grad_norm": 9.501011848449707, "learning_rate": 3.39168960770819e-05, "loss": 0.8597, "step": 186950 }, { "epoch": 32.17481073640743, "grad_norm": 10.629328727722168, "learning_rate": 3.3912594631796286e-05, "loss": 0.8876, "step": 187000 }, { "epoch": 32.18341362697866, "grad_norm": 9.334572792053223, "learning_rate": 3.390829318651067e-05, "loss": 0.8449, "step": 187050 }, { "epoch": 32.192016517549895, "grad_norm": 9.340981483459473, "learning_rate": 3.3903991741225056e-05, "loss": 0.8696, "step": 187100 }, { "epoch": 32.20061940812113, "grad_norm": 10.24375057220459, "learning_rate": 3.389969029593944e-05, "loss": 0.8559, "step": 187150 }, { "epoch": 32.20922229869236, "grad_norm": 9.866820335388184, "learning_rate": 3.389538885065382e-05, "loss": 0.8919, "step": 187200 }, { "epoch": 32.21782518926359, "grad_norm": 16.001134872436523, "learning_rate": 3.3891087405368206e-05, "loss": 0.8596, "step": 187250 }, { "epoch": 32.226428079834825, "grad_norm": 6.659049987792969, "learning_rate": 3.388678596008259e-05, "loss": 0.874, "step": 187300 }, { "epoch": 32.23503097040606, "grad_norm": 16.52001953125, "learning_rate": 3.3882484514796975e-05, "loss": 0.8746, "step": 187350 }, { "epoch": 32.24363386097729, "grad_norm": 9.901860237121582, "learning_rate": 3.387818306951136e-05, "loss": 0.8734, "step": 187400 }, { "epoch": 32.25223675154852, "grad_norm": 12.053766250610352, "learning_rate": 3.387388162422574e-05, "loss": 0.8651, "step": 187450 }, { "epoch": 32.260839642119755, "grad_norm": 8.940925598144531, "learning_rate": 3.3869580178940125e-05, "loss": 0.8763, "step": 187500 }, { "epoch": 32.26944253269098, "grad_norm": 8.308158874511719, "learning_rate": 3.3865278733654506e-05, "loss": 0.7919, "step": 187550 }, { "epoch": 32.27804542326221, "grad_norm": 8.011115074157715, "learning_rate": 3.3860977288368894e-05, "loss": 0.8715, "step": 187600 }, { "epoch": 32.286648313833446, "grad_norm": 12.528432846069336, "learning_rate": 3.385667584308328e-05, "loss": 0.8784, "step": 187650 }, { "epoch": 32.29525120440468, "grad_norm": 7.175879001617432, "learning_rate": 3.385237439779766e-05, "loss": 0.8411, "step": 187700 }, { "epoch": 32.30385409497591, "grad_norm": 12.066157341003418, "learning_rate": 3.3848072952512045e-05, "loss": 0.8938, "step": 187750 }, { "epoch": 32.31245698554714, "grad_norm": 14.7401123046875, "learning_rate": 3.384377150722643e-05, "loss": 0.8462, "step": 187800 }, { "epoch": 32.321059876118376, "grad_norm": 9.202737808227539, "learning_rate": 3.3839470061940814e-05, "loss": 0.8309, "step": 187850 }, { "epoch": 32.32966276668961, "grad_norm": 13.13135814666748, "learning_rate": 3.38351686166552e-05, "loss": 0.8922, "step": 187900 }, { "epoch": 32.33826565726084, "grad_norm": 7.544773101806641, "learning_rate": 3.3830867171369576e-05, "loss": 0.8643, "step": 187950 }, { "epoch": 32.346868547832074, "grad_norm": 9.951685905456543, "learning_rate": 3.3826565726083964e-05, "loss": 0.8337, "step": 188000 }, { "epoch": 32.355471438403306, "grad_norm": 13.979231834411621, "learning_rate": 3.382226428079835e-05, "loss": 0.8747, "step": 188050 }, { "epoch": 32.36407432897454, "grad_norm": 8.443633079528809, "learning_rate": 3.381796283551273e-05, "loss": 0.8567, "step": 188100 }, { "epoch": 32.372677219545764, "grad_norm": 12.806077003479004, "learning_rate": 3.381366139022712e-05, "loss": 0.8532, "step": 188150 }, { "epoch": 32.381280110117, "grad_norm": 11.161148071289062, "learning_rate": 3.38093599449415e-05, "loss": 0.8421, "step": 188200 }, { "epoch": 32.38988300068823, "grad_norm": 8.560043334960938, "learning_rate": 3.3805058499655883e-05, "loss": 0.8772, "step": 188250 }, { "epoch": 32.39848589125946, "grad_norm": 8.069417953491211, "learning_rate": 3.380075705437027e-05, "loss": 0.8756, "step": 188300 }, { "epoch": 32.407088781830694, "grad_norm": 11.629646301269531, "learning_rate": 3.379645560908465e-05, "loss": 0.881, "step": 188350 }, { "epoch": 32.41569167240193, "grad_norm": 8.257584571838379, "learning_rate": 3.379215416379904e-05, "loss": 0.8686, "step": 188400 }, { "epoch": 32.42429456297316, "grad_norm": 8.291685104370117, "learning_rate": 3.378785271851342e-05, "loss": 0.8835, "step": 188450 }, { "epoch": 32.43289745354439, "grad_norm": 12.612107276916504, "learning_rate": 3.37835512732278e-05, "loss": 0.8897, "step": 188500 }, { "epoch": 32.441500344115624, "grad_norm": 12.954550743103027, "learning_rate": 3.377924982794219e-05, "loss": 0.8463, "step": 188550 }, { "epoch": 32.45010323468686, "grad_norm": 7.533907890319824, "learning_rate": 3.377494838265657e-05, "loss": 0.8894, "step": 188600 }, { "epoch": 32.45870612525809, "grad_norm": 13.819576263427734, "learning_rate": 3.377064693737096e-05, "loss": 0.8377, "step": 188650 }, { "epoch": 32.46730901582932, "grad_norm": 11.84616756439209, "learning_rate": 3.376634549208534e-05, "loss": 0.8872, "step": 188700 }, { "epoch": 32.47591190640055, "grad_norm": 10.2267427444458, "learning_rate": 3.376204404679973e-05, "loss": 0.8565, "step": 188750 }, { "epoch": 32.48451479697178, "grad_norm": 10.66508960723877, "learning_rate": 3.375774260151411e-05, "loss": 0.8928, "step": 188800 }, { "epoch": 32.49311768754301, "grad_norm": 11.722112655639648, "learning_rate": 3.375344115622849e-05, "loss": 0.8911, "step": 188850 }, { "epoch": 32.501720578114245, "grad_norm": 15.544063568115234, "learning_rate": 3.374913971094288e-05, "loss": 0.8558, "step": 188900 }, { "epoch": 32.51032346868548, "grad_norm": 10.279569625854492, "learning_rate": 3.374483826565726e-05, "loss": 0.84, "step": 188950 }, { "epoch": 32.51892635925671, "grad_norm": 10.521358489990234, "learning_rate": 3.374053682037165e-05, "loss": 0.9437, "step": 189000 }, { "epoch": 32.52752924982794, "grad_norm": 11.703659057617188, "learning_rate": 3.373623537508603e-05, "loss": 0.8561, "step": 189050 }, { "epoch": 32.536132140399175, "grad_norm": 10.834888458251953, "learning_rate": 3.373193392980041e-05, "loss": 0.8174, "step": 189100 }, { "epoch": 32.54473503097041, "grad_norm": 12.905241012573242, "learning_rate": 3.37276324845148e-05, "loss": 0.9003, "step": 189150 }, { "epoch": 32.55333792154164, "grad_norm": 11.236388206481934, "learning_rate": 3.372333103922919e-05, "loss": 0.8965, "step": 189200 }, { "epoch": 32.56194081211287, "grad_norm": 10.16446304321289, "learning_rate": 3.371902959394357e-05, "loss": 0.8747, "step": 189250 }, { "epoch": 32.570543702684105, "grad_norm": 11.064828872680664, "learning_rate": 3.371472814865795e-05, "loss": 0.885, "step": 189300 }, { "epoch": 32.57914659325533, "grad_norm": 11.44881534576416, "learning_rate": 3.371042670337233e-05, "loss": 0.8906, "step": 189350 }, { "epoch": 32.58774948382656, "grad_norm": 9.482641220092773, "learning_rate": 3.370612525808672e-05, "loss": 0.8414, "step": 189400 }, { "epoch": 32.596352374397796, "grad_norm": 8.697303771972656, "learning_rate": 3.3701823812801106e-05, "loss": 0.8357, "step": 189450 }, { "epoch": 32.60495526496903, "grad_norm": 9.710620880126953, "learning_rate": 3.369752236751549e-05, "loss": 0.8589, "step": 189500 }, { "epoch": 32.61355815554026, "grad_norm": 8.503951072692871, "learning_rate": 3.3693220922229875e-05, "loss": 0.8529, "step": 189550 }, { "epoch": 32.62216104611149, "grad_norm": 15.183716773986816, "learning_rate": 3.368891947694425e-05, "loss": 0.8321, "step": 189600 }, { "epoch": 32.630763936682726, "grad_norm": 9.213330268859863, "learning_rate": 3.368461803165864e-05, "loss": 0.8339, "step": 189650 }, { "epoch": 32.63936682725396, "grad_norm": 15.639291763305664, "learning_rate": 3.3680316586373026e-05, "loss": 0.9015, "step": 189700 }, { "epoch": 32.64796971782519, "grad_norm": 10.625306129455566, "learning_rate": 3.367601514108741e-05, "loss": 0.8774, "step": 189750 }, { "epoch": 32.656572608396424, "grad_norm": 8.455765724182129, "learning_rate": 3.3671713695801795e-05, "loss": 0.8532, "step": 189800 }, { "epoch": 32.665175498967656, "grad_norm": 9.207921981811523, "learning_rate": 3.3667412250516176e-05, "loss": 0.8632, "step": 189850 }, { "epoch": 32.67377838953888, "grad_norm": 12.323335647583008, "learning_rate": 3.366311080523056e-05, "loss": 0.8555, "step": 189900 }, { "epoch": 32.682381280110114, "grad_norm": 7.628798961639404, "learning_rate": 3.3658809359944945e-05, "loss": 0.8544, "step": 189950 }, { "epoch": 32.69098417068135, "grad_norm": 6.630911350250244, "learning_rate": 3.3654507914659326e-05, "loss": 0.8551, "step": 190000 }, { "epoch": 32.69958706125258, "grad_norm": 13.304728507995605, "learning_rate": 3.3650206469373714e-05, "loss": 0.8967, "step": 190050 }, { "epoch": 32.70818995182381, "grad_norm": 11.635164260864258, "learning_rate": 3.3645905024088095e-05, "loss": 0.8754, "step": 190100 }, { "epoch": 32.716792842395044, "grad_norm": 9.786354064941406, "learning_rate": 3.3641603578802476e-05, "loss": 0.8776, "step": 190150 }, { "epoch": 32.72539573296628, "grad_norm": 9.038719177246094, "learning_rate": 3.3637302133516864e-05, "loss": 0.846, "step": 190200 }, { "epoch": 32.73399862353751, "grad_norm": 10.038090705871582, "learning_rate": 3.3633000688231246e-05, "loss": 0.8934, "step": 190250 }, { "epoch": 32.74260151410874, "grad_norm": 11.641119003295898, "learning_rate": 3.3628699242945634e-05, "loss": 0.8917, "step": 190300 }, { "epoch": 32.751204404679974, "grad_norm": 21.56242561340332, "learning_rate": 3.3624397797660015e-05, "loss": 0.9062, "step": 190350 }, { "epoch": 32.75980729525121, "grad_norm": 12.563766479492188, "learning_rate": 3.3620096352374396e-05, "loss": 0.889, "step": 190400 }, { "epoch": 32.76841018582244, "grad_norm": 6.548441410064697, "learning_rate": 3.3615794907088784e-05, "loss": 0.8415, "step": 190450 }, { "epoch": 32.777013076393665, "grad_norm": 14.331366539001465, "learning_rate": 3.3611493461803165e-05, "loss": 0.8845, "step": 190500 }, { "epoch": 32.7856159669649, "grad_norm": 10.145983695983887, "learning_rate": 3.360719201651755e-05, "loss": 0.8668, "step": 190550 }, { "epoch": 32.79421885753613, "grad_norm": 12.357900619506836, "learning_rate": 3.360289057123194e-05, "loss": 0.8253, "step": 190600 }, { "epoch": 32.80282174810736, "grad_norm": 12.31905746459961, "learning_rate": 3.3598589125946315e-05, "loss": 0.896, "step": 190650 }, { "epoch": 32.811424638678595, "grad_norm": 10.334495544433594, "learning_rate": 3.35942876806607e-05, "loss": 0.8283, "step": 190700 }, { "epoch": 32.82002752924983, "grad_norm": 10.523308753967285, "learning_rate": 3.3589986235375084e-05, "loss": 0.9139, "step": 190750 }, { "epoch": 32.82863041982106, "grad_norm": 12.194304466247559, "learning_rate": 3.358568479008947e-05, "loss": 0.9007, "step": 190800 }, { "epoch": 32.83723331039229, "grad_norm": 10.363595962524414, "learning_rate": 3.358138334480386e-05, "loss": 0.938, "step": 190850 }, { "epoch": 32.845836200963525, "grad_norm": 7.254922389984131, "learning_rate": 3.357708189951824e-05, "loss": 0.845, "step": 190900 }, { "epoch": 32.85443909153476, "grad_norm": 16.320560455322266, "learning_rate": 3.357278045423262e-05, "loss": 0.8659, "step": 190950 }, { "epoch": 32.86304198210599, "grad_norm": 16.62286949157715, "learning_rate": 3.3568479008947004e-05, "loss": 0.8318, "step": 191000 }, { "epoch": 32.87164487267722, "grad_norm": 12.134632110595703, "learning_rate": 3.356417756366139e-05, "loss": 0.9338, "step": 191050 }, { "epoch": 32.88024776324845, "grad_norm": 11.425152778625488, "learning_rate": 3.355987611837578e-05, "loss": 0.8967, "step": 191100 }, { "epoch": 32.88885065381968, "grad_norm": 13.639885902404785, "learning_rate": 3.355557467309016e-05, "loss": 0.89, "step": 191150 }, { "epoch": 32.89745354439091, "grad_norm": 6.716434001922607, "learning_rate": 3.355127322780454e-05, "loss": 0.887, "step": 191200 }, { "epoch": 32.906056434962146, "grad_norm": 10.206987380981445, "learning_rate": 3.354697178251893e-05, "loss": 0.8681, "step": 191250 }, { "epoch": 32.91465932553338, "grad_norm": 10.807086944580078, "learning_rate": 3.354267033723331e-05, "loss": 0.8173, "step": 191300 }, { "epoch": 32.92326221610461, "grad_norm": 9.778143882751465, "learning_rate": 3.35383688919477e-05, "loss": 0.8512, "step": 191350 }, { "epoch": 32.93186510667584, "grad_norm": 9.861985206604004, "learning_rate": 3.353406744666208e-05, "loss": 0.8728, "step": 191400 }, { "epoch": 32.940467997247076, "grad_norm": 17.93320655822754, "learning_rate": 3.352976600137646e-05, "loss": 0.8863, "step": 191450 }, { "epoch": 32.94907088781831, "grad_norm": 12.988165855407715, "learning_rate": 3.352546455609085e-05, "loss": 0.8468, "step": 191500 }, { "epoch": 32.95767377838954, "grad_norm": 9.446928977966309, "learning_rate": 3.352116311080523e-05, "loss": 0.8718, "step": 191550 }, { "epoch": 32.966276668960774, "grad_norm": 7.750868320465088, "learning_rate": 3.351686166551962e-05, "loss": 0.8341, "step": 191600 }, { "epoch": 32.974879559532006, "grad_norm": 14.95274829864502, "learning_rate": 3.3512560220234e-05, "loss": 0.9394, "step": 191650 }, { "epoch": 32.98348245010323, "grad_norm": 8.635055541992188, "learning_rate": 3.350825877494839e-05, "loss": 0.8887, "step": 191700 }, { "epoch": 32.992085340674464, "grad_norm": 10.792703628540039, "learning_rate": 3.350395732966277e-05, "loss": 0.7984, "step": 191750 }, { "epoch": 33.0, "eval_accuracy": 0.5234038891757012, "eval_f1": 0.4854448652631179, "eval_f1_DuraRiadoRio_16x16": 0.2416058394160584, "eval_f1_Mole_16x16": 0.5712804561949196, "eval_f1_Quebrado_16x16": 0.6429319371727749, "eval_f1_RiadoRio_16x16": 0.41275602800103706, "eval_f1_RioFechado_16x16": 0.5586500655307994, "eval_loss": 1.9435709714889526, "eval_precision": 0.5926141834481806, "eval_precision_DuraRiadoRio_16x16": 0.7591743119266054, "eval_precision_Mole_16x16": 0.7091377091377091, "eval_precision_Quebrado_16x16": 0.49052727687742526, "eval_precision_RiadoRio_16x16": 0.5585964912280702, "eval_precision_RioFechado_16x16": 0.4456351280710925, "eval_recall": 0.5260907392218475, "eval_recall_DuraRiadoRio_16x16": 0.14366319444444445, "eval_recall_Mole_16x16": 0.4782986111111111, "eval_recall_Quebrado_16x16": 0.9327256944444444, "eval_recall_RiadoRio_16x16": 0.32730263157894735, "eval_recall_RioFechado_16x16": 0.7484635645302897, "eval_runtime": 46.4487, "eval_samples_per_second": 250.211, "eval_steps_per_second": 15.652, "step": 191796 }, { "epoch": 33.0006882312457, "grad_norm": 9.699004173278809, "learning_rate": 3.349965588437715e-05, "loss": 0.8832, "step": 191800 }, { "epoch": 33.00929112181693, "grad_norm": 18.64847183227539, "learning_rate": 3.349535443909154e-05, "loss": 0.8037, "step": 191850 }, { "epoch": 33.01789401238816, "grad_norm": 13.728870391845703, "learning_rate": 3.349105299380592e-05, "loss": 0.8883, "step": 191900 }, { "epoch": 33.026496902959394, "grad_norm": 11.986940383911133, "learning_rate": 3.348675154852031e-05, "loss": 0.8904, "step": 191950 }, { "epoch": 33.03509979353063, "grad_norm": 15.343351364135742, "learning_rate": 3.348245010323469e-05, "loss": 0.8687, "step": 192000 }, { "epoch": 33.04370268410186, "grad_norm": 10.15105152130127, "learning_rate": 3.347814865794907e-05, "loss": 0.9186, "step": 192050 }, { "epoch": 33.05230557467309, "grad_norm": 9.447779655456543, "learning_rate": 3.347384721266346e-05, "loss": 0.8446, "step": 192100 }, { "epoch": 33.060908465244324, "grad_norm": 9.645816802978516, "learning_rate": 3.346954576737784e-05, "loss": 0.8125, "step": 192150 }, { "epoch": 33.06951135581556, "grad_norm": 11.190258979797363, "learning_rate": 3.3465244322092226e-05, "loss": 0.8391, "step": 192200 }, { "epoch": 33.07811424638679, "grad_norm": 10.748268127441406, "learning_rate": 3.3460942876806614e-05, "loss": 0.8221, "step": 192250 }, { "epoch": 33.086717136958015, "grad_norm": 8.323376655578613, "learning_rate": 3.345664143152099e-05, "loss": 0.8874, "step": 192300 }, { "epoch": 33.09532002752925, "grad_norm": 25.91495132446289, "learning_rate": 3.345233998623538e-05, "loss": 0.8698, "step": 192350 }, { "epoch": 33.10392291810048, "grad_norm": 11.19621467590332, "learning_rate": 3.344803854094976e-05, "loss": 0.8279, "step": 192400 }, { "epoch": 33.11252580867171, "grad_norm": 14.981266021728516, "learning_rate": 3.3443737095664146e-05, "loss": 0.8432, "step": 192450 }, { "epoch": 33.121128699242945, "grad_norm": 10.64856243133545, "learning_rate": 3.3439435650378534e-05, "loss": 0.8954, "step": 192500 }, { "epoch": 33.12973158981418, "grad_norm": 12.402923583984375, "learning_rate": 3.343513420509291e-05, "loss": 0.8496, "step": 192550 }, { "epoch": 33.13833448038541, "grad_norm": 11.432872772216797, "learning_rate": 3.3430832759807296e-05, "loss": 0.8687, "step": 192600 }, { "epoch": 33.14693737095664, "grad_norm": 8.962390899658203, "learning_rate": 3.342653131452168e-05, "loss": 0.8516, "step": 192650 }, { "epoch": 33.155540261527875, "grad_norm": 9.379138946533203, "learning_rate": 3.3422229869236065e-05, "loss": 0.8039, "step": 192700 }, { "epoch": 33.16414315209911, "grad_norm": 11.94077205657959, "learning_rate": 3.341792842395045e-05, "loss": 0.8154, "step": 192750 }, { "epoch": 33.17274604267034, "grad_norm": 14.519234657287598, "learning_rate": 3.341362697866483e-05, "loss": 0.8887, "step": 192800 }, { "epoch": 33.181348933241566, "grad_norm": 12.769603729248047, "learning_rate": 3.3409325533379216e-05, "loss": 0.8698, "step": 192850 }, { "epoch": 33.1899518238128, "grad_norm": 11.412074089050293, "learning_rate": 3.3405024088093604e-05, "loss": 0.8341, "step": 192900 }, { "epoch": 33.19855471438403, "grad_norm": 10.651707649230957, "learning_rate": 3.3400722642807985e-05, "loss": 0.8001, "step": 192950 }, { "epoch": 33.20715760495526, "grad_norm": 9.09465503692627, "learning_rate": 3.339642119752237e-05, "loss": 0.8293, "step": 193000 }, { "epoch": 33.215760495526496, "grad_norm": 9.850098609924316, "learning_rate": 3.3392119752236754e-05, "loss": 0.8249, "step": 193050 }, { "epoch": 33.22436338609773, "grad_norm": 12.091716766357422, "learning_rate": 3.3387818306951135e-05, "loss": 0.829, "step": 193100 }, { "epoch": 33.23296627666896, "grad_norm": 13.473051071166992, "learning_rate": 3.338351686166552e-05, "loss": 0.8653, "step": 193150 }, { "epoch": 33.24156916724019, "grad_norm": 8.75447940826416, "learning_rate": 3.3379215416379904e-05, "loss": 0.8908, "step": 193200 }, { "epoch": 33.250172057811426, "grad_norm": 6.843907356262207, "learning_rate": 3.337491397109429e-05, "loss": 0.8381, "step": 193250 }, { "epoch": 33.25877494838266, "grad_norm": 7.381378650665283, "learning_rate": 3.337061252580867e-05, "loss": 0.8633, "step": 193300 }, { "epoch": 33.26737783895389, "grad_norm": 6.620952129364014, "learning_rate": 3.3366311080523054e-05, "loss": 0.8827, "step": 193350 }, { "epoch": 33.27598072952512, "grad_norm": 9.853689193725586, "learning_rate": 3.336200963523744e-05, "loss": 0.845, "step": 193400 }, { "epoch": 33.28458362009635, "grad_norm": 12.771637916564941, "learning_rate": 3.3357708189951824e-05, "loss": 0.868, "step": 193450 }, { "epoch": 33.29318651066758, "grad_norm": 10.63200569152832, "learning_rate": 3.335340674466621e-05, "loss": 0.9179, "step": 193500 }, { "epoch": 33.301789401238814, "grad_norm": 8.105987548828125, "learning_rate": 3.334910529938059e-05, "loss": 0.8507, "step": 193550 }, { "epoch": 33.31039229181005, "grad_norm": 10.570005416870117, "learning_rate": 3.3344803854094974e-05, "loss": 0.8593, "step": 193600 }, { "epoch": 33.31899518238128, "grad_norm": 11.999038696289062, "learning_rate": 3.334050240880936e-05, "loss": 0.8951, "step": 193650 }, { "epoch": 33.32759807295251, "grad_norm": 9.962215423583984, "learning_rate": 3.333620096352374e-05, "loss": 0.8147, "step": 193700 }, { "epoch": 33.336200963523744, "grad_norm": 12.575662612915039, "learning_rate": 3.333189951823813e-05, "loss": 0.908, "step": 193750 }, { "epoch": 33.34480385409498, "grad_norm": 10.704075813293457, "learning_rate": 3.332759807295251e-05, "loss": 0.8175, "step": 193800 }, { "epoch": 33.35340674466621, "grad_norm": 9.703815460205078, "learning_rate": 3.33232966276669e-05, "loss": 0.8767, "step": 193850 }, { "epoch": 33.36200963523744, "grad_norm": 12.228446006774902, "learning_rate": 3.331899518238128e-05, "loss": 0.892, "step": 193900 }, { "epoch": 33.370612525808674, "grad_norm": 8.775775909423828, "learning_rate": 3.331469373709566e-05, "loss": 0.8631, "step": 193950 }, { "epoch": 33.37921541637991, "grad_norm": 5.258145332336426, "learning_rate": 3.331039229181005e-05, "loss": 0.8887, "step": 194000 }, { "epoch": 33.38781830695113, "grad_norm": 12.532306671142578, "learning_rate": 3.330609084652443e-05, "loss": 0.8889, "step": 194050 }, { "epoch": 33.396421197522365, "grad_norm": 9.383208274841309, "learning_rate": 3.330178940123882e-05, "loss": 0.8832, "step": 194100 }, { "epoch": 33.4050240880936, "grad_norm": 14.34406852722168, "learning_rate": 3.32974879559532e-05, "loss": 0.8592, "step": 194150 }, { "epoch": 33.41362697866483, "grad_norm": 15.159299850463867, "learning_rate": 3.329318651066758e-05, "loss": 0.9168, "step": 194200 }, { "epoch": 33.42222986923606, "grad_norm": 14.340958595275879, "learning_rate": 3.328888506538197e-05, "loss": 0.8828, "step": 194250 }, { "epoch": 33.430832759807295, "grad_norm": 12.878695487976074, "learning_rate": 3.328458362009636e-05, "loss": 0.8916, "step": 194300 }, { "epoch": 33.43943565037853, "grad_norm": 10.822718620300293, "learning_rate": 3.328028217481074e-05, "loss": 0.8405, "step": 194350 }, { "epoch": 33.44803854094976, "grad_norm": 10.323404312133789, "learning_rate": 3.327598072952513e-05, "loss": 0.9, "step": 194400 }, { "epoch": 33.45664143152099, "grad_norm": 12.852296829223633, "learning_rate": 3.32716792842395e-05, "loss": 0.8629, "step": 194450 }, { "epoch": 33.465244322092225, "grad_norm": 15.907588958740234, "learning_rate": 3.326737783895389e-05, "loss": 0.8852, "step": 194500 }, { "epoch": 33.47384721266346, "grad_norm": 9.066055297851562, "learning_rate": 3.326307639366828e-05, "loss": 0.8664, "step": 194550 }, { "epoch": 33.48245010323469, "grad_norm": 12.775718688964844, "learning_rate": 3.325877494838266e-05, "loss": 0.8476, "step": 194600 }, { "epoch": 33.491052993805916, "grad_norm": 10.260148048400879, "learning_rate": 3.3254473503097046e-05, "loss": 0.84, "step": 194650 }, { "epoch": 33.49965588437715, "grad_norm": 11.926032066345215, "learning_rate": 3.325017205781142e-05, "loss": 0.8733, "step": 194700 }, { "epoch": 33.50825877494838, "grad_norm": 10.106258392333984, "learning_rate": 3.324587061252581e-05, "loss": 0.9413, "step": 194750 }, { "epoch": 33.51686166551961, "grad_norm": 10.910012245178223, "learning_rate": 3.3241569167240197e-05, "loss": 0.873, "step": 194800 }, { "epoch": 33.525464556090846, "grad_norm": 8.669344902038574, "learning_rate": 3.323726772195458e-05, "loss": 0.8764, "step": 194850 }, { "epoch": 33.53406744666208, "grad_norm": 13.077116012573242, "learning_rate": 3.3232966276668966e-05, "loss": 0.887, "step": 194900 }, { "epoch": 33.54267033723331, "grad_norm": 9.952001571655273, "learning_rate": 3.322866483138335e-05, "loss": 0.8657, "step": 194950 }, { "epoch": 33.55127322780454, "grad_norm": 5.781915664672852, "learning_rate": 3.322436338609773e-05, "loss": 0.8702, "step": 195000 }, { "epoch": 33.559876118375776, "grad_norm": 13.807308197021484, "learning_rate": 3.3220061940812116e-05, "loss": 0.8436, "step": 195050 }, { "epoch": 33.56847900894701, "grad_norm": 8.36359691619873, "learning_rate": 3.32157604955265e-05, "loss": 0.906, "step": 195100 }, { "epoch": 33.57708189951824, "grad_norm": 5.728250026702881, "learning_rate": 3.3211459050240885e-05, "loss": 0.8064, "step": 195150 }, { "epoch": 33.58568479008947, "grad_norm": 10.418073654174805, "learning_rate": 3.3207157604955266e-05, "loss": 0.8756, "step": 195200 }, { "epoch": 33.5942876806607, "grad_norm": 8.57862663269043, "learning_rate": 3.320285615966965e-05, "loss": 0.8654, "step": 195250 }, { "epoch": 33.60289057123193, "grad_norm": 13.738333702087402, "learning_rate": 3.3198554714384035e-05, "loss": 0.847, "step": 195300 }, { "epoch": 33.611493461803164, "grad_norm": 11.397505760192871, "learning_rate": 3.3194253269098417e-05, "loss": 0.8276, "step": 195350 }, { "epoch": 33.6200963523744, "grad_norm": 10.822166442871094, "learning_rate": 3.3189951823812804e-05, "loss": 0.8811, "step": 195400 }, { "epoch": 33.62869924294563, "grad_norm": 11.750923156738281, "learning_rate": 3.3185650378527186e-05, "loss": 0.8773, "step": 195450 }, { "epoch": 33.63730213351686, "grad_norm": 9.79946517944336, "learning_rate": 3.318134893324157e-05, "loss": 0.8747, "step": 195500 }, { "epoch": 33.645905024088094, "grad_norm": 10.176294326782227, "learning_rate": 3.3177047487955955e-05, "loss": 0.8205, "step": 195550 }, { "epoch": 33.65450791465933, "grad_norm": 11.73713493347168, "learning_rate": 3.3172746042670336e-05, "loss": 0.8211, "step": 195600 }, { "epoch": 33.66311080523056, "grad_norm": 9.382670402526855, "learning_rate": 3.3168444597384724e-05, "loss": 0.8607, "step": 195650 }, { "epoch": 33.67171369580179, "grad_norm": 12.333935737609863, "learning_rate": 3.316414315209911e-05, "loss": 0.8892, "step": 195700 }, { "epoch": 33.680316586373024, "grad_norm": 13.933906555175781, "learning_rate": 3.315984170681349e-05, "loss": 0.8571, "step": 195750 }, { "epoch": 33.68891947694425, "grad_norm": 11.068760871887207, "learning_rate": 3.3155540261527874e-05, "loss": 0.8584, "step": 195800 }, { "epoch": 33.69752236751548, "grad_norm": 11.068602561950684, "learning_rate": 3.3151238816242255e-05, "loss": 0.8607, "step": 195850 }, { "epoch": 33.706125258086715, "grad_norm": 14.479357719421387, "learning_rate": 3.314693737095664e-05, "loss": 0.8717, "step": 195900 }, { "epoch": 33.71472814865795, "grad_norm": 11.105730056762695, "learning_rate": 3.314263592567103e-05, "loss": 0.8608, "step": 195950 }, { "epoch": 33.72333103922918, "grad_norm": 10.497045516967773, "learning_rate": 3.313833448038541e-05, "loss": 0.8734, "step": 196000 }, { "epoch": 33.73193392980041, "grad_norm": 9.10722541809082, "learning_rate": 3.3134033035099794e-05, "loss": 0.8489, "step": 196050 }, { "epoch": 33.740536820371645, "grad_norm": 11.328387260437012, "learning_rate": 3.3129731589814175e-05, "loss": 0.8533, "step": 196100 }, { "epoch": 33.74913971094288, "grad_norm": 7.430356025695801, "learning_rate": 3.312543014452856e-05, "loss": 0.8544, "step": 196150 }, { "epoch": 33.75774260151411, "grad_norm": 9.897027969360352, "learning_rate": 3.312112869924295e-05, "loss": 0.844, "step": 196200 }, { "epoch": 33.76634549208534, "grad_norm": 14.270951271057129, "learning_rate": 3.311682725395733e-05, "loss": 0.9018, "step": 196250 }, { "epoch": 33.774948382656575, "grad_norm": 11.42479133605957, "learning_rate": 3.311252580867171e-05, "loss": 0.8839, "step": 196300 }, { "epoch": 33.78355127322781, "grad_norm": 13.008315086364746, "learning_rate": 3.31082243633861e-05, "loss": 0.8323, "step": 196350 }, { "epoch": 33.79215416379903, "grad_norm": 12.204937934875488, "learning_rate": 3.310392291810048e-05, "loss": 0.8222, "step": 196400 }, { "epoch": 33.800757054370266, "grad_norm": 12.357587814331055, "learning_rate": 3.309962147281487e-05, "loss": 0.892, "step": 196450 }, { "epoch": 33.8093599449415, "grad_norm": 19.930360794067383, "learning_rate": 3.309532002752925e-05, "loss": 0.8777, "step": 196500 }, { "epoch": 33.81796283551273, "grad_norm": 11.878382682800293, "learning_rate": 3.309101858224364e-05, "loss": 0.8216, "step": 196550 }, { "epoch": 33.82656572608396, "grad_norm": 8.474522590637207, "learning_rate": 3.308671713695802e-05, "loss": 0.9008, "step": 196600 }, { "epoch": 33.835168616655196, "grad_norm": 4.859290599822998, "learning_rate": 3.30824156916724e-05, "loss": 0.8792, "step": 196650 }, { "epoch": 33.84377150722643, "grad_norm": 7.4809722900390625, "learning_rate": 3.307811424638679e-05, "loss": 0.8448, "step": 196700 }, { "epoch": 33.85237439779766, "grad_norm": 7.1181159019470215, "learning_rate": 3.307381280110117e-05, "loss": 0.8136, "step": 196750 }, { "epoch": 33.86097728836889, "grad_norm": 8.849923133850098, "learning_rate": 3.306951135581556e-05, "loss": 0.8506, "step": 196800 }, { "epoch": 33.869580178940126, "grad_norm": 12.217486381530762, "learning_rate": 3.306520991052994e-05, "loss": 0.8437, "step": 196850 }, { "epoch": 33.87818306951136, "grad_norm": 12.96523380279541, "learning_rate": 3.306090846524432e-05, "loss": 0.8417, "step": 196900 }, { "epoch": 33.88678596008259, "grad_norm": 9.116828918457031, "learning_rate": 3.305660701995871e-05, "loss": 0.8964, "step": 196950 }, { "epoch": 33.895388850653816, "grad_norm": 11.96809196472168, "learning_rate": 3.305230557467309e-05, "loss": 0.8861, "step": 197000 }, { "epoch": 33.90399174122505, "grad_norm": 7.593081474304199, "learning_rate": 3.304800412938748e-05, "loss": 0.8583, "step": 197050 }, { "epoch": 33.91259463179628, "grad_norm": 14.326255798339844, "learning_rate": 3.304370268410186e-05, "loss": 0.8662, "step": 197100 }, { "epoch": 33.921197522367514, "grad_norm": 14.618711471557617, "learning_rate": 3.303940123881624e-05, "loss": 0.9066, "step": 197150 }, { "epoch": 33.92980041293875, "grad_norm": 13.191388130187988, "learning_rate": 3.303509979353063e-05, "loss": 0.8544, "step": 197200 }, { "epoch": 33.93840330350998, "grad_norm": 14.89144229888916, "learning_rate": 3.303079834824501e-05, "loss": 0.8512, "step": 197250 }, { "epoch": 33.94700619408121, "grad_norm": 8.70503044128418, "learning_rate": 3.30264969029594e-05, "loss": 0.9259, "step": 197300 }, { "epoch": 33.955609084652444, "grad_norm": 11.301817893981934, "learning_rate": 3.3022195457673785e-05, "loss": 0.8644, "step": 197350 }, { "epoch": 33.96421197522368, "grad_norm": 8.822723388671875, "learning_rate": 3.301789401238816e-05, "loss": 0.9266, "step": 197400 }, { "epoch": 33.97281486579491, "grad_norm": 19.081026077270508, "learning_rate": 3.301359256710255e-05, "loss": 0.8579, "step": 197450 }, { "epoch": 33.98141775636614, "grad_norm": 9.327852249145508, "learning_rate": 3.300929112181693e-05, "loss": 0.8979, "step": 197500 }, { "epoch": 33.990020646937374, "grad_norm": 11.58326530456543, "learning_rate": 3.300498967653132e-05, "loss": 0.8879, "step": 197550 }, { "epoch": 33.9986235375086, "grad_norm": 8.314979553222656, "learning_rate": 3.3000688231245705e-05, "loss": 0.8127, "step": 197600 }, { "epoch": 34.0, "eval_accuracy": 0.5472379969024265, "eval_f1": 0.5343809010219418, "eval_f1_DuraRiadoRio_16x16": 0.5173845355474831, "eval_f1_Mole_16x16": 0.6051179126944305, "eval_f1_Quebrado_16x16": 0.4376237623762376, "eval_f1_RiadoRio_16x16": 0.49196393571148567, "eval_f1_RioFechado_16x16": 0.619814358780072, "eval_loss": 1.594391942024231, "eval_precision": 0.6127765828436285, "eval_precision_DuraRiadoRio_16x16": 0.6432258064516129, "eval_precision_Mole_16x16": 0.4922448979591837, "eval_precision_Quebrado_16x16": 0.9132231404958677, "eval_precision_RiadoRio_16x16": 0.4700374531835206, "eval_precision_RioFechado_16x16": 0.5451516161279574, "eval_recall": 0.5479704764040992, "eval_recall_DuraRiadoRio_16x16": 0.4327256944444444, "eval_recall_Mole_16x16": 0.78515625, "eval_recall_Quebrado_16x16": 0.2877604166666667, "eval_recall_RiadoRio_16x16": 0.5160361842105263, "eval_recall_RioFechado_16x16": 0.7181738366988587, "eval_runtime": 46.4367, "eval_samples_per_second": 250.276, "eval_steps_per_second": 15.656, "step": 197608 }, { "epoch": 34.00722642807983, "grad_norm": 7.723941326141357, "learning_rate": 3.299638678596008e-05, "loss": 0.809, "step": 197650 }, { "epoch": 34.015829318651065, "grad_norm": 14.416890144348145, "learning_rate": 3.299208534067447e-05, "loss": 0.8433, "step": 197700 }, { "epoch": 34.0244322092223, "grad_norm": 10.013671875, "learning_rate": 3.2987783895388855e-05, "loss": 0.8522, "step": 197750 }, { "epoch": 34.03303509979353, "grad_norm": 10.805571556091309, "learning_rate": 3.2983482450103236e-05, "loss": 0.8178, "step": 197800 }, { "epoch": 34.04163799036476, "grad_norm": 10.244780540466309, "learning_rate": 3.2979181004817624e-05, "loss": 0.932, "step": 197850 }, { "epoch": 34.050240880935995, "grad_norm": 12.19654369354248, "learning_rate": 3.2974879559532005e-05, "loss": 0.8682, "step": 197900 }, { "epoch": 34.05884377150723, "grad_norm": 13.01709270477295, "learning_rate": 3.2970578114246387e-05, "loss": 0.8023, "step": 197950 }, { "epoch": 34.06744666207846, "grad_norm": 16.12141990661621, "learning_rate": 3.2966276668960775e-05, "loss": 0.827, "step": 198000 }, { "epoch": 34.07604955264969, "grad_norm": 11.566180229187012, "learning_rate": 3.2961975223675156e-05, "loss": 0.8239, "step": 198050 }, { "epoch": 34.084652443220925, "grad_norm": 9.825575828552246, "learning_rate": 3.2957673778389544e-05, "loss": 0.9169, "step": 198100 }, { "epoch": 34.09325533379216, "grad_norm": 9.080620765686035, "learning_rate": 3.2953372333103925e-05, "loss": 0.8886, "step": 198150 }, { "epoch": 34.10185822436338, "grad_norm": 11.807164192199707, "learning_rate": 3.2949070887818306e-05, "loss": 0.8585, "step": 198200 }, { "epoch": 34.110461114934616, "grad_norm": 8.559258460998535, "learning_rate": 3.2944769442532694e-05, "loss": 0.8462, "step": 198250 }, { "epoch": 34.11906400550585, "grad_norm": 11.563594818115234, "learning_rate": 3.2940467997247075e-05, "loss": 0.8645, "step": 198300 }, { "epoch": 34.12766689607708, "grad_norm": 9.319289207458496, "learning_rate": 3.293616655196146e-05, "loss": 0.8559, "step": 198350 }, { "epoch": 34.13626978664831, "grad_norm": 10.927430152893066, "learning_rate": 3.2931865106675844e-05, "loss": 0.9131, "step": 198400 }, { "epoch": 34.144872677219546, "grad_norm": 10.257214546203613, "learning_rate": 3.2927563661390225e-05, "loss": 0.9004, "step": 198450 }, { "epoch": 34.15347556779078, "grad_norm": 12.505215644836426, "learning_rate": 3.292326221610461e-05, "loss": 0.8239, "step": 198500 }, { "epoch": 34.16207845836201, "grad_norm": 12.887667655944824, "learning_rate": 3.2918960770818995e-05, "loss": 0.8451, "step": 198550 }, { "epoch": 34.17068134893324, "grad_norm": 13.04271125793457, "learning_rate": 3.291465932553338e-05, "loss": 0.8683, "step": 198600 }, { "epoch": 34.179284239504476, "grad_norm": 8.453643798828125, "learning_rate": 3.2910357880247764e-05, "loss": 0.8582, "step": 198650 }, { "epoch": 34.18788713007571, "grad_norm": 9.793142318725586, "learning_rate": 3.290605643496215e-05, "loss": 0.8404, "step": 198700 }, { "epoch": 34.19649002064694, "grad_norm": 14.654833793640137, "learning_rate": 3.290175498967653e-05, "loss": 0.9128, "step": 198750 }, { "epoch": 34.205092911218166, "grad_norm": 10.625388145446777, "learning_rate": 3.2897453544390914e-05, "loss": 0.8002, "step": 198800 }, { "epoch": 34.2136958017894, "grad_norm": 13.571146011352539, "learning_rate": 3.28931520991053e-05, "loss": 0.8445, "step": 198850 }, { "epoch": 34.22229869236063, "grad_norm": 12.116853713989258, "learning_rate": 3.288885065381968e-05, "loss": 0.8328, "step": 198900 }, { "epoch": 34.230901582931864, "grad_norm": 11.576618194580078, "learning_rate": 3.288454920853407e-05, "loss": 0.8841, "step": 198950 }, { "epoch": 34.2395044735031, "grad_norm": 16.99598503112793, "learning_rate": 3.288024776324845e-05, "loss": 0.7813, "step": 199000 }, { "epoch": 34.24810736407433, "grad_norm": 10.626957893371582, "learning_rate": 3.287594631796283e-05, "loss": 0.8648, "step": 199050 }, { "epoch": 34.25671025464556, "grad_norm": 9.408656120300293, "learning_rate": 3.287164487267722e-05, "loss": 0.7967, "step": 199100 }, { "epoch": 34.265313145216794, "grad_norm": 13.486054420471191, "learning_rate": 3.286734342739161e-05, "loss": 0.8331, "step": 199150 }, { "epoch": 34.27391603578803, "grad_norm": 15.685680389404297, "learning_rate": 3.286304198210599e-05, "loss": 0.7667, "step": 199200 }, { "epoch": 34.28251892635926, "grad_norm": 16.838695526123047, "learning_rate": 3.285874053682037e-05, "loss": 0.8273, "step": 199250 }, { "epoch": 34.29112181693049, "grad_norm": 11.013328552246094, "learning_rate": 3.285443909153475e-05, "loss": 0.8755, "step": 199300 }, { "epoch": 34.29972470750172, "grad_norm": 14.851521492004395, "learning_rate": 3.285013764624914e-05, "loss": 0.855, "step": 199350 }, { "epoch": 34.30832759807295, "grad_norm": 12.133782386779785, "learning_rate": 3.284583620096353e-05, "loss": 0.8518, "step": 199400 }, { "epoch": 34.31693048864418, "grad_norm": 10.07300090789795, "learning_rate": 3.284153475567791e-05, "loss": 0.8848, "step": 199450 }, { "epoch": 34.325533379215415, "grad_norm": 9.922211647033691, "learning_rate": 3.28372333103923e-05, "loss": 0.872, "step": 199500 }, { "epoch": 34.33413626978665, "grad_norm": 14.482342720031738, "learning_rate": 3.283293186510667e-05, "loss": 0.8686, "step": 199550 }, { "epoch": 34.34273916035788, "grad_norm": 8.868735313415527, "learning_rate": 3.282863041982106e-05, "loss": 0.8579, "step": 199600 }, { "epoch": 34.35134205092911, "grad_norm": 12.80249309539795, "learning_rate": 3.282432897453545e-05, "loss": 0.8064, "step": 199650 }, { "epoch": 34.359944941500345, "grad_norm": 11.95151138305664, "learning_rate": 3.282002752924983e-05, "loss": 0.8524, "step": 199700 }, { "epoch": 34.36854783207158, "grad_norm": 15.771743774414062, "learning_rate": 3.281572608396422e-05, "loss": 0.8349, "step": 199750 }, { "epoch": 34.37715072264281, "grad_norm": 17.295560836791992, "learning_rate": 3.28114246386786e-05, "loss": 0.807, "step": 199800 }, { "epoch": 34.38575361321404, "grad_norm": 18.44343376159668, "learning_rate": 3.280712319339298e-05, "loss": 0.8918, "step": 199850 }, { "epoch": 34.394356503785275, "grad_norm": 28.4755802154541, "learning_rate": 3.280282174810737e-05, "loss": 0.8517, "step": 199900 }, { "epoch": 34.4029593943565, "grad_norm": 12.384471893310547, "learning_rate": 3.279852030282175e-05, "loss": 0.8334, "step": 199950 }, { "epoch": 34.41156228492773, "grad_norm": 10.190892219543457, "learning_rate": 3.2794218857536137e-05, "loss": 0.8704, "step": 200000 }, { "epoch": 34.420165175498965, "grad_norm": 11.947005271911621, "learning_rate": 3.278991741225052e-05, "loss": 0.8372, "step": 200050 }, { "epoch": 34.4287680660702, "grad_norm": 13.69653034210205, "learning_rate": 3.27856159669649e-05, "loss": 0.8931, "step": 200100 }, { "epoch": 34.43737095664143, "grad_norm": 17.037906646728516, "learning_rate": 3.278131452167929e-05, "loss": 0.8047, "step": 200150 }, { "epoch": 34.44597384721266, "grad_norm": 8.69920539855957, "learning_rate": 3.277701307639367e-05, "loss": 0.8341, "step": 200200 }, { "epoch": 34.454576737783896, "grad_norm": 7.63215970993042, "learning_rate": 3.2772711631108056e-05, "loss": 0.873, "step": 200250 }, { "epoch": 34.46317962835513, "grad_norm": 11.903081893920898, "learning_rate": 3.276841018582244e-05, "loss": 0.8901, "step": 200300 }, { "epoch": 34.47178251892636, "grad_norm": 16.121551513671875, "learning_rate": 3.276410874053682e-05, "loss": 0.9143, "step": 200350 }, { "epoch": 34.48038540949759, "grad_norm": 12.798028945922852, "learning_rate": 3.2759807295251206e-05, "loss": 0.8812, "step": 200400 }, { "epoch": 34.488988300068826, "grad_norm": 9.837823867797852, "learning_rate": 3.275550584996559e-05, "loss": 0.8838, "step": 200450 }, { "epoch": 34.49759119064006, "grad_norm": 9.326920509338379, "learning_rate": 3.2751204404679975e-05, "loss": 0.9076, "step": 200500 }, { "epoch": 34.506194081211284, "grad_norm": 9.476529121398926, "learning_rate": 3.2746902959394357e-05, "loss": 0.8727, "step": 200550 }, { "epoch": 34.514796971782516, "grad_norm": 11.420526504516602, "learning_rate": 3.274260151410874e-05, "loss": 0.881, "step": 200600 }, { "epoch": 34.52339986235375, "grad_norm": 10.934907913208008, "learning_rate": 3.2738300068823126e-05, "loss": 0.8561, "step": 200650 }, { "epoch": 34.53200275292498, "grad_norm": 8.996284484863281, "learning_rate": 3.273399862353751e-05, "loss": 0.8312, "step": 200700 }, { "epoch": 34.540605643496214, "grad_norm": 13.84523868560791, "learning_rate": 3.2729697178251895e-05, "loss": 0.8224, "step": 200750 }, { "epoch": 34.549208534067446, "grad_norm": 18.270729064941406, "learning_rate": 3.272539573296628e-05, "loss": 0.8261, "step": 200800 }, { "epoch": 34.55781142463868, "grad_norm": 11.23880672454834, "learning_rate": 3.2721094287680664e-05, "loss": 0.8145, "step": 200850 }, { "epoch": 34.56641431520991, "grad_norm": 9.64745044708252, "learning_rate": 3.2716792842395045e-05, "loss": 0.8315, "step": 200900 }, { "epoch": 34.575017205781144, "grad_norm": 13.657811164855957, "learning_rate": 3.2712491397109426e-05, "loss": 0.8843, "step": 200950 }, { "epoch": 34.58362009635238, "grad_norm": 13.083840370178223, "learning_rate": 3.2708189951823814e-05, "loss": 0.882, "step": 201000 }, { "epoch": 34.59222298692361, "grad_norm": 13.917457580566406, "learning_rate": 3.27038885065382e-05, "loss": 0.8294, "step": 201050 }, { "epoch": 34.60082587749484, "grad_norm": 10.44485092163086, "learning_rate": 3.269958706125258e-05, "loss": 0.8615, "step": 201100 }, { "epoch": 34.60942876806607, "grad_norm": 9.959376335144043, "learning_rate": 3.2695285615966965e-05, "loss": 0.8195, "step": 201150 }, { "epoch": 34.6180316586373, "grad_norm": 9.238410949707031, "learning_rate": 3.269098417068135e-05, "loss": 0.8238, "step": 201200 }, { "epoch": 34.62663454920853, "grad_norm": 7.912295818328857, "learning_rate": 3.2686682725395734e-05, "loss": 0.8458, "step": 201250 }, { "epoch": 34.635237439779765, "grad_norm": 9.038687705993652, "learning_rate": 3.268238128011012e-05, "loss": 0.8676, "step": 201300 }, { "epoch": 34.643840330351, "grad_norm": 10.211465835571289, "learning_rate": 3.26780798348245e-05, "loss": 0.8155, "step": 201350 }, { "epoch": 34.65244322092223, "grad_norm": 11.582569122314453, "learning_rate": 3.267377838953889e-05, "loss": 0.8662, "step": 201400 }, { "epoch": 34.66104611149346, "grad_norm": 13.041566848754883, "learning_rate": 3.266947694425327e-05, "loss": 0.8426, "step": 201450 }, { "epoch": 34.669649002064695, "grad_norm": 13.754462242126465, "learning_rate": 3.266517549896765e-05, "loss": 0.8836, "step": 201500 }, { "epoch": 34.67825189263593, "grad_norm": 12.6000394821167, "learning_rate": 3.266087405368204e-05, "loss": 0.8258, "step": 201550 }, { "epoch": 34.68685478320716, "grad_norm": 11.624961853027344, "learning_rate": 3.265657260839642e-05, "loss": 0.9458, "step": 201600 }, { "epoch": 34.69545767377839, "grad_norm": 11.61853313446045, "learning_rate": 3.265227116311081e-05, "loss": 0.8073, "step": 201650 }, { "epoch": 34.704060564349625, "grad_norm": 12.148681640625, "learning_rate": 3.264796971782519e-05, "loss": 0.8564, "step": 201700 }, { "epoch": 34.71266345492085, "grad_norm": 6.24528169631958, "learning_rate": 3.264366827253957e-05, "loss": 0.9144, "step": 201750 }, { "epoch": 34.72126634549208, "grad_norm": 13.118940353393555, "learning_rate": 3.263936682725396e-05, "loss": 0.8471, "step": 201800 }, { "epoch": 34.729869236063315, "grad_norm": 11.430907249450684, "learning_rate": 3.263506538196834e-05, "loss": 0.9226, "step": 201850 }, { "epoch": 34.73847212663455, "grad_norm": 16.389917373657227, "learning_rate": 3.263076393668273e-05, "loss": 0.8104, "step": 201900 }, { "epoch": 34.74707501720578, "grad_norm": 6.924263954162598, "learning_rate": 3.262646249139711e-05, "loss": 0.8536, "step": 201950 }, { "epoch": 34.75567790777701, "grad_norm": 12.243240356445312, "learning_rate": 3.262216104611149e-05, "loss": 0.9415, "step": 202000 }, { "epoch": 34.764280798348246, "grad_norm": 8.640816688537598, "learning_rate": 3.261785960082588e-05, "loss": 0.8602, "step": 202050 }, { "epoch": 34.77288368891948, "grad_norm": 10.558335304260254, "learning_rate": 3.261355815554026e-05, "loss": 0.8969, "step": 202100 }, { "epoch": 34.78148657949071, "grad_norm": 11.468293190002441, "learning_rate": 3.260925671025465e-05, "loss": 0.8314, "step": 202150 }, { "epoch": 34.79008947006194, "grad_norm": 14.859932899475098, "learning_rate": 3.260495526496904e-05, "loss": 0.866, "step": 202200 }, { "epoch": 34.798692360633176, "grad_norm": 11.885013580322266, "learning_rate": 3.260065381968341e-05, "loss": 0.8867, "step": 202250 }, { "epoch": 34.80729525120441, "grad_norm": 9.604657173156738, "learning_rate": 3.25963523743978e-05, "loss": 0.8432, "step": 202300 }, { "epoch": 34.815898141775634, "grad_norm": 9.546141624450684, "learning_rate": 3.259205092911218e-05, "loss": 0.9122, "step": 202350 }, { "epoch": 34.824501032346866, "grad_norm": 7.891623497009277, "learning_rate": 3.258774948382657e-05, "loss": 0.8736, "step": 202400 }, { "epoch": 34.8331039229181, "grad_norm": 9.382537841796875, "learning_rate": 3.2583448038540956e-05, "loss": 0.847, "step": 202450 }, { "epoch": 34.84170681348933, "grad_norm": 8.035417556762695, "learning_rate": 3.257914659325533e-05, "loss": 0.8544, "step": 202500 }, { "epoch": 34.850309704060564, "grad_norm": 16.956335067749023, "learning_rate": 3.257484514796972e-05, "loss": 0.862, "step": 202550 }, { "epoch": 34.858912594631796, "grad_norm": 9.938112258911133, "learning_rate": 3.25705437026841e-05, "loss": 0.8372, "step": 202600 }, { "epoch": 34.86751548520303, "grad_norm": 7.819533824920654, "learning_rate": 3.256624225739849e-05, "loss": 0.8797, "step": 202650 }, { "epoch": 34.87611837577426, "grad_norm": 12.400238037109375, "learning_rate": 3.2561940812112876e-05, "loss": 0.853, "step": 202700 }, { "epoch": 34.884721266345494, "grad_norm": 10.989931106567383, "learning_rate": 3.255763936682725e-05, "loss": 0.8467, "step": 202750 }, { "epoch": 34.89332415691673, "grad_norm": 14.052227020263672, "learning_rate": 3.255333792154164e-05, "loss": 0.8297, "step": 202800 }, { "epoch": 34.90192704748796, "grad_norm": 10.35075855255127, "learning_rate": 3.2549036476256026e-05, "loss": 0.8843, "step": 202850 }, { "epoch": 34.910529938059184, "grad_norm": 12.46611213684082, "learning_rate": 3.254473503097041e-05, "loss": 0.8197, "step": 202900 }, { "epoch": 34.91913282863042, "grad_norm": 19.872556686401367, "learning_rate": 3.2540433585684795e-05, "loss": 0.8208, "step": 202950 }, { "epoch": 34.92773571920165, "grad_norm": 12.612303733825684, "learning_rate": 3.2536132140399176e-05, "loss": 0.8735, "step": 203000 }, { "epoch": 34.93633860977288, "grad_norm": 11.7134370803833, "learning_rate": 3.253183069511356e-05, "loss": 0.9014, "step": 203050 }, { "epoch": 34.944941500344115, "grad_norm": 11.015841484069824, "learning_rate": 3.2527529249827945e-05, "loss": 0.9053, "step": 203100 }, { "epoch": 34.95354439091535, "grad_norm": 12.252843856811523, "learning_rate": 3.252322780454233e-05, "loss": 0.8578, "step": 203150 }, { "epoch": 34.96214728148658, "grad_norm": 12.952484130859375, "learning_rate": 3.2518926359256715e-05, "loss": 0.8737, "step": 203200 }, { "epoch": 34.97075017205781, "grad_norm": 13.39712905883789, "learning_rate": 3.2514624913971096e-05, "loss": 0.8547, "step": 203250 }, { "epoch": 34.979353062629045, "grad_norm": 19.47435188293457, "learning_rate": 3.251032346868548e-05, "loss": 0.8004, "step": 203300 }, { "epoch": 34.98795595320028, "grad_norm": 13.374895095825195, "learning_rate": 3.2506022023399865e-05, "loss": 0.8283, "step": 203350 }, { "epoch": 34.99655884377151, "grad_norm": 10.70610237121582, "learning_rate": 3.2501720578114246e-05, "loss": 0.8669, "step": 203400 }, { "epoch": 35.0, "eval_accuracy": 0.6148683531233867, "eval_f1": 0.6036365565168297, "eval_f1_DuraRiadoRio_16x16": 0.5822596630327056, "eval_f1_Mole_16x16": 0.4958233890214797, "eval_f1_Quebrado_16x16": 0.7082976903336184, "eval_f1_RiadoRio_16x16": 0.5496153846153846, "eval_f1_RioFechado_16x16": 0.6821866555809603, "eval_loss": 1.3819918632507324, "eval_precision": 0.6440063302581777, "eval_precision_DuraRiadoRio_16x16": 0.6784064665127021, "eval_precision_Mole_16x16": 0.7929389312977099, "eval_precision_Quebrado_16x16": 0.5845806269415419, "eval_precision_RiadoRio_16x16": 0.516257225433526, "eval_precision_RioFechado_16x16": 0.6478484011054086, "eval_recall": 0.6154096407154116, "eval_recall_DuraRiadoRio_16x16": 0.5099826388888888, "eval_recall_Mole_16x16": 0.3606770833333333, "eval_recall_Quebrado_16x16": 0.8984375, "eval_recall_RiadoRio_16x16": 0.5875822368421053, "eval_recall_RioFechado_16x16": 0.7203687445127305, "eval_runtime": 46.7644, "eval_samples_per_second": 248.522, "eval_steps_per_second": 15.546, "step": 203420 }, { "epoch": 35.00516173434274, "grad_norm": 15.23050594329834, "learning_rate": 3.2497419132828634e-05, "loss": 0.8363, "step": 203450 }, { "epoch": 35.01376462491397, "grad_norm": 14.519754409790039, "learning_rate": 3.2493117687543015e-05, "loss": 0.8353, "step": 203500 }, { "epoch": 35.0223675154852, "grad_norm": 8.585149765014648, "learning_rate": 3.24888162422574e-05, "loss": 0.8247, "step": 203550 }, { "epoch": 35.03097040605643, "grad_norm": 12.665912628173828, "learning_rate": 3.2484514796971784e-05, "loss": 0.8732, "step": 203600 }, { "epoch": 35.039573296627665, "grad_norm": 11.531346321105957, "learning_rate": 3.2480213351686165e-05, "loss": 0.8796, "step": 203650 }, { "epoch": 35.0481761871989, "grad_norm": 8.061259269714355, "learning_rate": 3.2475911906400553e-05, "loss": 0.8155, "step": 203700 }, { "epoch": 35.05677907777013, "grad_norm": 11.775321960449219, "learning_rate": 3.2471610461114935e-05, "loss": 0.8548, "step": 203750 }, { "epoch": 35.06538196834136, "grad_norm": 9.494305610656738, "learning_rate": 3.246730901582932e-05, "loss": 0.933, "step": 203800 }, { "epoch": 35.073984858912596, "grad_norm": 10.234720230102539, "learning_rate": 3.2463007570543704e-05, "loss": 0.804, "step": 203850 }, { "epoch": 35.08258774948383, "grad_norm": 8.548134803771973, "learning_rate": 3.2458706125258085e-05, "loss": 0.8446, "step": 203900 }, { "epoch": 35.09119064005506, "grad_norm": 9.06960391998291, "learning_rate": 3.245440467997247e-05, "loss": 0.8346, "step": 203950 }, { "epoch": 35.09979353062629, "grad_norm": 11.947551727294922, "learning_rate": 3.2450103234686854e-05, "loss": 0.7738, "step": 204000 }, { "epoch": 35.108396421197526, "grad_norm": 8.404254913330078, "learning_rate": 3.244580178940124e-05, "loss": 0.8509, "step": 204050 }, { "epoch": 35.11699931176875, "grad_norm": 8.9825439453125, "learning_rate": 3.244150034411562e-05, "loss": 0.8484, "step": 204100 }, { "epoch": 35.125602202339984, "grad_norm": 10.275079727172852, "learning_rate": 3.2437198898830004e-05, "loss": 0.8595, "step": 204150 }, { "epoch": 35.134205092911216, "grad_norm": 14.068306922912598, "learning_rate": 3.243289745354439e-05, "loss": 0.821, "step": 204200 }, { "epoch": 35.14280798348245, "grad_norm": 8.772857666015625, "learning_rate": 3.242859600825878e-05, "loss": 0.8452, "step": 204250 }, { "epoch": 35.15141087405368, "grad_norm": 13.983622550964355, "learning_rate": 3.242429456297316e-05, "loss": 0.8299, "step": 204300 }, { "epoch": 35.160013764624914, "grad_norm": 16.922348022460938, "learning_rate": 3.241999311768755e-05, "loss": 0.8875, "step": 204350 }, { "epoch": 35.168616655196146, "grad_norm": 10.875700950622559, "learning_rate": 3.2415691672401924e-05, "loss": 0.8544, "step": 204400 }, { "epoch": 35.17721954576738, "grad_norm": 13.057642936706543, "learning_rate": 3.241139022711631e-05, "loss": 0.8336, "step": 204450 }, { "epoch": 35.18582243633861, "grad_norm": 11.049738883972168, "learning_rate": 3.24070887818307e-05, "loss": 0.8447, "step": 204500 }, { "epoch": 35.194425326909844, "grad_norm": 7.253073215484619, "learning_rate": 3.240278733654508e-05, "loss": 0.8912, "step": 204550 }, { "epoch": 35.20302821748108, "grad_norm": 15.423830032348633, "learning_rate": 3.239848589125947e-05, "loss": 0.8701, "step": 204600 }, { "epoch": 35.21163110805231, "grad_norm": 11.391202926635742, "learning_rate": 3.239418444597384e-05, "loss": 0.8212, "step": 204650 }, { "epoch": 35.220233998623534, "grad_norm": 8.787407875061035, "learning_rate": 3.238988300068823e-05, "loss": 0.855, "step": 204700 }, { "epoch": 35.22883688919477, "grad_norm": 12.965995788574219, "learning_rate": 3.238558155540262e-05, "loss": 0.8617, "step": 204750 }, { "epoch": 35.237439779766, "grad_norm": 12.243977546691895, "learning_rate": 3.2381280110117e-05, "loss": 0.834, "step": 204800 }, { "epoch": 35.24604267033723, "grad_norm": 19.16217803955078, "learning_rate": 3.237697866483139e-05, "loss": 0.828, "step": 204850 }, { "epoch": 35.254645560908465, "grad_norm": 14.03455638885498, "learning_rate": 3.237267721954577e-05, "loss": 0.8834, "step": 204900 }, { "epoch": 35.2632484514797, "grad_norm": 14.087629318237305, "learning_rate": 3.236837577426015e-05, "loss": 0.8431, "step": 204950 }, { "epoch": 35.27185134205093, "grad_norm": 14.416988372802734, "learning_rate": 3.236407432897454e-05, "loss": 0.8292, "step": 205000 }, { "epoch": 35.28045423262216, "grad_norm": 14.396838188171387, "learning_rate": 3.235977288368892e-05, "loss": 0.8268, "step": 205050 }, { "epoch": 35.289057123193395, "grad_norm": 10.09204387664795, "learning_rate": 3.235547143840331e-05, "loss": 0.8654, "step": 205100 }, { "epoch": 35.29766001376463, "grad_norm": 9.268141746520996, "learning_rate": 3.235116999311769e-05, "loss": 0.8234, "step": 205150 }, { "epoch": 35.30626290433586, "grad_norm": 11.624702453613281, "learning_rate": 3.234686854783207e-05, "loss": 0.8107, "step": 205200 }, { "epoch": 35.31486579490709, "grad_norm": 13.02381706237793, "learning_rate": 3.234256710254646e-05, "loss": 0.9107, "step": 205250 }, { "epoch": 35.32346868547832, "grad_norm": 14.241048812866211, "learning_rate": 3.233826565726084e-05, "loss": 0.8075, "step": 205300 }, { "epoch": 35.33207157604955, "grad_norm": 10.67221736907959, "learning_rate": 3.233396421197523e-05, "loss": 0.7904, "step": 205350 }, { "epoch": 35.34067446662078, "grad_norm": 10.5592679977417, "learning_rate": 3.232966276668961e-05, "loss": 0.9018, "step": 205400 }, { "epoch": 35.349277357192015, "grad_norm": 19.252527236938477, "learning_rate": 3.232536132140399e-05, "loss": 0.901, "step": 205450 }, { "epoch": 35.35788024776325, "grad_norm": 10.334161758422852, "learning_rate": 3.232105987611838e-05, "loss": 0.801, "step": 205500 }, { "epoch": 35.36648313833448, "grad_norm": 12.247603416442871, "learning_rate": 3.231675843083276e-05, "loss": 0.8019, "step": 205550 }, { "epoch": 35.37508602890571, "grad_norm": 10.126092910766602, "learning_rate": 3.2312456985547146e-05, "loss": 0.8484, "step": 205600 }, { "epoch": 35.383688919476946, "grad_norm": 12.603821754455566, "learning_rate": 3.2308155540261534e-05, "loss": 0.8692, "step": 205650 }, { "epoch": 35.39229181004818, "grad_norm": 10.566073417663574, "learning_rate": 3.2303854094975915e-05, "loss": 0.8509, "step": 205700 }, { "epoch": 35.40089470061941, "grad_norm": 12.726466178894043, "learning_rate": 3.22995526496903e-05, "loss": 0.8563, "step": 205750 }, { "epoch": 35.40949759119064, "grad_norm": 15.036700248718262, "learning_rate": 3.229525120440468e-05, "loss": 0.8158, "step": 205800 }, { "epoch": 35.41810048176187, "grad_norm": 12.446409225463867, "learning_rate": 3.2290949759119066e-05, "loss": 0.8999, "step": 205850 }, { "epoch": 35.4267033723331, "grad_norm": 12.096962928771973, "learning_rate": 3.2286648313833454e-05, "loss": 0.8115, "step": 205900 }, { "epoch": 35.435306262904334, "grad_norm": 9.862338066101074, "learning_rate": 3.2282346868547835e-05, "loss": 0.8713, "step": 205950 }, { "epoch": 35.443909153475566, "grad_norm": 11.034460067749023, "learning_rate": 3.2278045423262216e-05, "loss": 0.86, "step": 206000 }, { "epoch": 35.4525120440468, "grad_norm": 13.573153495788574, "learning_rate": 3.22737439779766e-05, "loss": 0.8461, "step": 206050 }, { "epoch": 35.46111493461803, "grad_norm": 12.564433097839355, "learning_rate": 3.2269442532690985e-05, "loss": 0.8358, "step": 206100 }, { "epoch": 35.469717825189264, "grad_norm": 10.990483283996582, "learning_rate": 3.226514108740537e-05, "loss": 0.8707, "step": 206150 }, { "epoch": 35.478320715760496, "grad_norm": 11.448390007019043, "learning_rate": 3.2260839642119754e-05, "loss": 0.8996, "step": 206200 }, { "epoch": 35.48692360633173, "grad_norm": 9.41377067565918, "learning_rate": 3.2256538196834135e-05, "loss": 0.8662, "step": 206250 }, { "epoch": 35.49552649690296, "grad_norm": 7.602402210235596, "learning_rate": 3.2252236751548523e-05, "loss": 0.8212, "step": 206300 }, { "epoch": 35.504129387474194, "grad_norm": 13.971085548400879, "learning_rate": 3.2247935306262905e-05, "loss": 0.8386, "step": 206350 }, { "epoch": 35.512732278045426, "grad_norm": 9.849902153015137, "learning_rate": 3.224363386097729e-05, "loss": 0.8938, "step": 206400 }, { "epoch": 35.52133516861665, "grad_norm": 12.407340049743652, "learning_rate": 3.2239332415691674e-05, "loss": 0.9431, "step": 206450 }, { "epoch": 35.529938059187884, "grad_norm": 8.196109771728516, "learning_rate": 3.223503097040606e-05, "loss": 0.8682, "step": 206500 }, { "epoch": 35.53854094975912, "grad_norm": 13.807161331176758, "learning_rate": 3.223072952512044e-05, "loss": 0.8397, "step": 206550 }, { "epoch": 35.54714384033035, "grad_norm": 11.38847541809082, "learning_rate": 3.2226428079834824e-05, "loss": 0.883, "step": 206600 }, { "epoch": 35.55574673090158, "grad_norm": 12.884778022766113, "learning_rate": 3.222212663454921e-05, "loss": 0.8697, "step": 206650 }, { "epoch": 35.564349621472815, "grad_norm": 8.718083381652832, "learning_rate": 3.221782518926359e-05, "loss": 0.7755, "step": 206700 }, { "epoch": 35.57295251204405, "grad_norm": 8.099611282348633, "learning_rate": 3.221352374397798e-05, "loss": 0.8857, "step": 206750 }, { "epoch": 35.58155540261528, "grad_norm": 18.99228858947754, "learning_rate": 3.220922229869236e-05, "loss": 0.8203, "step": 206800 }, { "epoch": 35.59015829318651, "grad_norm": 9.95068073272705, "learning_rate": 3.2204920853406743e-05, "loss": 0.8852, "step": 206850 }, { "epoch": 35.598761183757745, "grad_norm": 12.226740837097168, "learning_rate": 3.220061940812113e-05, "loss": 0.8219, "step": 206900 }, { "epoch": 35.60736407432898, "grad_norm": 18.516803741455078, "learning_rate": 3.219631796283551e-05, "loss": 0.8189, "step": 206950 }, { "epoch": 35.61596696490021, "grad_norm": 8.626640319824219, "learning_rate": 3.21920165175499e-05, "loss": 0.8465, "step": 207000 }, { "epoch": 35.624569855471435, "grad_norm": 10.158951759338379, "learning_rate": 3.218771507226428e-05, "loss": 0.8588, "step": 207050 }, { "epoch": 35.63317274604267, "grad_norm": 8.776076316833496, "learning_rate": 3.218341362697866e-05, "loss": 0.909, "step": 207100 }, { "epoch": 35.6417756366139, "grad_norm": 11.987627029418945, "learning_rate": 3.217911218169305e-05, "loss": 0.8717, "step": 207150 }, { "epoch": 35.65037852718513, "grad_norm": 13.41974925994873, "learning_rate": 3.217481073640743e-05, "loss": 0.8638, "step": 207200 }, { "epoch": 35.658981417756365, "grad_norm": 10.466955184936523, "learning_rate": 3.217050929112182e-05, "loss": 0.8786, "step": 207250 }, { "epoch": 35.6675843083276, "grad_norm": 15.833537101745605, "learning_rate": 3.216620784583621e-05, "loss": 0.8433, "step": 207300 }, { "epoch": 35.67618719889883, "grad_norm": 14.117100715637207, "learning_rate": 3.216190640055058e-05, "loss": 0.8393, "step": 207350 }, { "epoch": 35.68479008947006, "grad_norm": 14.032395362854004, "learning_rate": 3.215760495526497e-05, "loss": 0.8797, "step": 207400 }, { "epoch": 35.693392980041295, "grad_norm": 13.010339736938477, "learning_rate": 3.215330350997935e-05, "loss": 0.8681, "step": 207450 }, { "epoch": 35.70199587061253, "grad_norm": 12.067227363586426, "learning_rate": 3.214900206469374e-05, "loss": 0.8541, "step": 207500 }, { "epoch": 35.71059876118376, "grad_norm": 12.792498588562012, "learning_rate": 3.214470061940813e-05, "loss": 0.8561, "step": 207550 }, { "epoch": 35.71920165175499, "grad_norm": 10.234371185302734, "learning_rate": 3.21403991741225e-05, "loss": 0.8154, "step": 207600 }, { "epoch": 35.72780454232622, "grad_norm": 13.87519359588623, "learning_rate": 3.213609772883689e-05, "loss": 0.8859, "step": 207650 }, { "epoch": 35.73640743289745, "grad_norm": 10.709038734436035, "learning_rate": 3.213179628355128e-05, "loss": 0.8524, "step": 207700 }, { "epoch": 35.745010323468684, "grad_norm": 9.880308151245117, "learning_rate": 3.212749483826566e-05, "loss": 0.8329, "step": 207750 }, { "epoch": 35.753613214039916, "grad_norm": 12.550567626953125, "learning_rate": 3.212319339298005e-05, "loss": 0.8339, "step": 207800 }, { "epoch": 35.76221610461115, "grad_norm": 14.629541397094727, "learning_rate": 3.211889194769443e-05, "loss": 0.8599, "step": 207850 }, { "epoch": 35.77081899518238, "grad_norm": 11.355667114257812, "learning_rate": 3.211459050240881e-05, "loss": 0.895, "step": 207900 }, { "epoch": 35.779421885753614, "grad_norm": 14.560222625732422, "learning_rate": 3.21102890571232e-05, "loss": 0.8613, "step": 207950 }, { "epoch": 35.788024776324846, "grad_norm": 10.550712585449219, "learning_rate": 3.210598761183758e-05, "loss": 0.8676, "step": 208000 }, { "epoch": 35.79662766689608, "grad_norm": 8.045562744140625, "learning_rate": 3.2101686166551966e-05, "loss": 0.8166, "step": 208050 }, { "epoch": 35.80523055746731, "grad_norm": 8.904256820678711, "learning_rate": 3.209738472126635e-05, "loss": 0.7977, "step": 208100 }, { "epoch": 35.813833448038544, "grad_norm": 11.790861129760742, "learning_rate": 3.209308327598073e-05, "loss": 0.8729, "step": 208150 }, { "epoch": 35.822436338609776, "grad_norm": 9.401518821716309, "learning_rate": 3.2088781830695116e-05, "loss": 0.8375, "step": 208200 }, { "epoch": 35.831039229181, "grad_norm": 9.367497444152832, "learning_rate": 3.20844803854095e-05, "loss": 0.8611, "step": 208250 }, { "epoch": 35.839642119752234, "grad_norm": 11.82066535949707, "learning_rate": 3.2080178940123886e-05, "loss": 0.9057, "step": 208300 }, { "epoch": 35.84824501032347, "grad_norm": 9.52621078491211, "learning_rate": 3.207587749483827e-05, "loss": 0.8423, "step": 208350 }, { "epoch": 35.8568479008947, "grad_norm": 11.33305549621582, "learning_rate": 3.207157604955265e-05, "loss": 0.8951, "step": 208400 }, { "epoch": 35.86545079146593, "grad_norm": 11.892388343811035, "learning_rate": 3.2067274604267036e-05, "loss": 0.8553, "step": 208450 }, { "epoch": 35.874053682037164, "grad_norm": 12.401168823242188, "learning_rate": 3.206297315898142e-05, "loss": 0.831, "step": 208500 }, { "epoch": 35.8826565726084, "grad_norm": 18.541221618652344, "learning_rate": 3.2058671713695805e-05, "loss": 0.8821, "step": 208550 }, { "epoch": 35.89125946317963, "grad_norm": 15.034602165222168, "learning_rate": 3.2054370268410186e-05, "loss": 0.8393, "step": 208600 }, { "epoch": 35.89986235375086, "grad_norm": 12.854809761047363, "learning_rate": 3.2050068823124574e-05, "loss": 0.8697, "step": 208650 }, { "epoch": 35.908465244322095, "grad_norm": 15.75217342376709, "learning_rate": 3.2045767377838955e-05, "loss": 0.832, "step": 208700 }, { "epoch": 35.91706813489333, "grad_norm": 15.229450225830078, "learning_rate": 3.2041465932553336e-05, "loss": 0.8854, "step": 208750 }, { "epoch": 35.92567102546455, "grad_norm": 8.521600723266602, "learning_rate": 3.2037164487267724e-05, "loss": 0.8384, "step": 208800 }, { "epoch": 35.934273916035785, "grad_norm": 6.792489528656006, "learning_rate": 3.2032863041982106e-05, "loss": 0.8506, "step": 208850 }, { "epoch": 35.94287680660702, "grad_norm": 15.492838859558105, "learning_rate": 3.2028561596696493e-05, "loss": 0.89, "step": 208900 }, { "epoch": 35.95147969717825, "grad_norm": 9.828019142150879, "learning_rate": 3.2024260151410875e-05, "loss": 0.8319, "step": 208950 }, { "epoch": 35.96008258774948, "grad_norm": 12.11175537109375, "learning_rate": 3.2019958706125256e-05, "loss": 0.8593, "step": 209000 }, { "epoch": 35.968685478320715, "grad_norm": 12.252793312072754, "learning_rate": 3.2015657260839644e-05, "loss": 0.7989, "step": 209050 }, { "epoch": 35.97728836889195, "grad_norm": 13.75327205657959, "learning_rate": 3.2011355815554025e-05, "loss": 0.7732, "step": 209100 }, { "epoch": 35.98589125946318, "grad_norm": 11.379630088806152, "learning_rate": 3.200705437026841e-05, "loss": 0.7844, "step": 209150 }, { "epoch": 35.99449415003441, "grad_norm": 12.197700500488281, "learning_rate": 3.20027529249828e-05, "loss": 0.8443, "step": 209200 }, { "epoch": 36.0, "eval_accuracy": 0.5782137325761487, "eval_f1": 0.5678459911196408, "eval_f1_DuraRiadoRio_16x16": 0.5203376822716808, "eval_f1_Mole_16x16": 0.6319675156884459, "eval_f1_Quebrado_16x16": 0.719443908090365, "eval_f1_RiadoRio_16x16": 0.4990597969161339, "eval_f1_RioFechado_16x16": 0.46842105263157896, "eval_loss": 1.5134761333465576, "eval_precision": 0.5985254158283764, "eval_precision_DuraRiadoRio_16x16": 0.6336448598130842, "eval_precision_Mole_16x16": 0.5497752087347463, "eval_precision_Quebrado_16x16": 0.648, "eval_precision_RiadoRio_16x16": 0.4598059598059598, "eval_precision_RioFechado_16x16": 0.7014010507880911, "eval_recall": 0.5780642469412484, "eval_recall_DuraRiadoRio_16x16": 0.44140625, "eval_recall_Mole_16x16": 0.7430555555555556, "eval_recall_Quebrado_16x16": 0.80859375, "eval_recall_RiadoRio_16x16": 0.545641447368421, "eval_recall_RioFechado_16x16": 0.35162423178226515, "eval_runtime": 46.3209, "eval_samples_per_second": 250.902, "eval_steps_per_second": 15.695, "step": 209232 }, { "epoch": 36.003097040605645, "grad_norm": 15.663206100463867, "learning_rate": 3.1998451479697175e-05, "loss": 0.8572, "step": 209250 }, { "epoch": 36.01169993117688, "grad_norm": 12.988362312316895, "learning_rate": 3.199415003441156e-05, "loss": 0.7917, "step": 209300 }, { "epoch": 36.02030282174811, "grad_norm": 11.747440338134766, "learning_rate": 3.198984858912595e-05, "loss": 0.823, "step": 209350 }, { "epoch": 36.028905712319336, "grad_norm": 13.042610168457031, "learning_rate": 3.198554714384033e-05, "loss": 0.8379, "step": 209400 }, { "epoch": 36.03750860289057, "grad_norm": 8.895848274230957, "learning_rate": 3.198124569855472e-05, "loss": 0.8392, "step": 209450 }, { "epoch": 36.0461114934618, "grad_norm": 7.967841625213623, "learning_rate": 3.1976944253269095e-05, "loss": 0.833, "step": 209500 }, { "epoch": 36.05471438403303, "grad_norm": 6.570825576782227, "learning_rate": 3.197264280798348e-05, "loss": 0.7972, "step": 209550 }, { "epoch": 36.063317274604266, "grad_norm": 13.465058326721191, "learning_rate": 3.196834136269787e-05, "loss": 0.8798, "step": 209600 }, { "epoch": 36.0719201651755, "grad_norm": 14.53465461730957, "learning_rate": 3.196403991741225e-05, "loss": 0.8156, "step": 209650 }, { "epoch": 36.08052305574673, "grad_norm": 14.391077041625977, "learning_rate": 3.195973847212664e-05, "loss": 0.7983, "step": 209700 }, { "epoch": 36.089125946317964, "grad_norm": 17.481096267700195, "learning_rate": 3.195543702684102e-05, "loss": 0.9225, "step": 209750 }, { "epoch": 36.097728836889196, "grad_norm": 10.639642715454102, "learning_rate": 3.19511355815554e-05, "loss": 0.8104, "step": 209800 }, { "epoch": 36.10633172746043, "grad_norm": 8.35386848449707, "learning_rate": 3.194683413626979e-05, "loss": 0.856, "step": 209850 }, { "epoch": 36.11493461803166, "grad_norm": 9.009659767150879, "learning_rate": 3.194253269098417e-05, "loss": 0.8442, "step": 209900 }, { "epoch": 36.123537508602894, "grad_norm": 14.107779502868652, "learning_rate": 3.193823124569856e-05, "loss": 0.8678, "step": 209950 }, { "epoch": 36.13214039917412, "grad_norm": 11.719018936157227, "learning_rate": 3.193392980041294e-05, "loss": 0.8251, "step": 210000 }, { "epoch": 36.14074328974535, "grad_norm": 14.797170639038086, "learning_rate": 3.192962835512732e-05, "loss": 0.8203, "step": 210050 }, { "epoch": 36.149346180316584, "grad_norm": 9.622892379760742, "learning_rate": 3.192532690984171e-05, "loss": 0.8411, "step": 210100 }, { "epoch": 36.15794907088782, "grad_norm": 12.519696235656738, "learning_rate": 3.192102546455609e-05, "loss": 0.8514, "step": 210150 }, { "epoch": 36.16655196145905, "grad_norm": 9.374659538269043, "learning_rate": 3.191672401927048e-05, "loss": 0.8733, "step": 210200 }, { "epoch": 36.17515485203028, "grad_norm": 11.401491165161133, "learning_rate": 3.191242257398486e-05, "loss": 0.8233, "step": 210250 }, { "epoch": 36.183757742601514, "grad_norm": 11.999564170837402, "learning_rate": 3.190812112869924e-05, "loss": 0.7964, "step": 210300 }, { "epoch": 36.19236063317275, "grad_norm": 25.933177947998047, "learning_rate": 3.190381968341363e-05, "loss": 0.8474, "step": 210350 }, { "epoch": 36.20096352374398, "grad_norm": 12.276843070983887, "learning_rate": 3.189951823812801e-05, "loss": 0.8258, "step": 210400 }, { "epoch": 36.20956641431521, "grad_norm": 15.32671070098877, "learning_rate": 3.18952167928424e-05, "loss": 0.7837, "step": 210450 }, { "epoch": 36.218169304886445, "grad_norm": 15.58676528930664, "learning_rate": 3.189091534755678e-05, "loss": 0.8225, "step": 210500 }, { "epoch": 36.22677219545768, "grad_norm": 13.005181312561035, "learning_rate": 3.188661390227117e-05, "loss": 0.8505, "step": 210550 }, { "epoch": 36.2353750860289, "grad_norm": 12.780974388122559, "learning_rate": 3.188231245698555e-05, "loss": 0.8247, "step": 210600 }, { "epoch": 36.243977976600135, "grad_norm": 7.058206081390381, "learning_rate": 3.187801101169993e-05, "loss": 0.8621, "step": 210650 }, { "epoch": 36.25258086717137, "grad_norm": 12.810040473937988, "learning_rate": 3.187370956641432e-05, "loss": 0.8347, "step": 210700 }, { "epoch": 36.2611837577426, "grad_norm": 10.615395545959473, "learning_rate": 3.1869408121128705e-05, "loss": 0.8741, "step": 210750 }, { "epoch": 36.26978664831383, "grad_norm": 7.698345184326172, "learning_rate": 3.1865106675843086e-05, "loss": 0.8238, "step": 210800 }, { "epoch": 36.278389538885065, "grad_norm": 18.25507354736328, "learning_rate": 3.186080523055747e-05, "loss": 0.8733, "step": 210850 }, { "epoch": 36.2869924294563, "grad_norm": 15.912452697753906, "learning_rate": 3.185650378527185e-05, "loss": 0.839, "step": 210900 }, { "epoch": 36.29559532002753, "grad_norm": 7.755764484405518, "learning_rate": 3.185220233998624e-05, "loss": 0.8512, "step": 210950 }, { "epoch": 36.30419821059876, "grad_norm": 9.400238990783691, "learning_rate": 3.1847900894700625e-05, "loss": 0.8161, "step": 211000 }, { "epoch": 36.312801101169995, "grad_norm": 10.08273983001709, "learning_rate": 3.1843599449415006e-05, "loss": 0.8797, "step": 211050 }, { "epoch": 36.32140399174123, "grad_norm": 9.96353530883789, "learning_rate": 3.183929800412939e-05, "loss": 0.838, "step": 211100 }, { "epoch": 36.33000688231246, "grad_norm": 8.443548202514648, "learning_rate": 3.183499655884377e-05, "loss": 0.855, "step": 211150 }, { "epoch": 36.338609772883686, "grad_norm": 13.769987106323242, "learning_rate": 3.1830695113558156e-05, "loss": 0.8074, "step": 211200 }, { "epoch": 36.34721266345492, "grad_norm": 15.650532722473145, "learning_rate": 3.1826393668272544e-05, "loss": 0.8321, "step": 211250 }, { "epoch": 36.35581555402615, "grad_norm": 10.502561569213867, "learning_rate": 3.1822092222986925e-05, "loss": 0.8093, "step": 211300 }, { "epoch": 36.36441844459738, "grad_norm": 15.344661712646484, "learning_rate": 3.181779077770131e-05, "loss": 0.7908, "step": 211350 }, { "epoch": 36.373021335168616, "grad_norm": 8.438776016235352, "learning_rate": 3.1813489332415694e-05, "loss": 0.9009, "step": 211400 }, { "epoch": 36.38162422573985, "grad_norm": 8.130911827087402, "learning_rate": 3.1809187887130076e-05, "loss": 0.8809, "step": 211450 }, { "epoch": 36.39022711631108, "grad_norm": 9.197090148925781, "learning_rate": 3.1804886441844464e-05, "loss": 0.7677, "step": 211500 }, { "epoch": 36.398830006882314, "grad_norm": 13.910935401916504, "learning_rate": 3.1800584996558845e-05, "loss": 0.8025, "step": 211550 }, { "epoch": 36.407432897453546, "grad_norm": 10.185547828674316, "learning_rate": 3.179628355127323e-05, "loss": 0.8112, "step": 211600 }, { "epoch": 36.41603578802478, "grad_norm": 14.685128211975098, "learning_rate": 3.1791982105987614e-05, "loss": 0.8696, "step": 211650 }, { "epoch": 36.42463867859601, "grad_norm": 12.073147773742676, "learning_rate": 3.1787680660701995e-05, "loss": 0.8539, "step": 211700 }, { "epoch": 36.43324156916724, "grad_norm": 11.14186954498291, "learning_rate": 3.178337921541638e-05, "loss": 0.8355, "step": 211750 }, { "epoch": 36.44184445973847, "grad_norm": 7.1596784591674805, "learning_rate": 3.1779077770130764e-05, "loss": 0.8774, "step": 211800 }, { "epoch": 36.4504473503097, "grad_norm": 12.679871559143066, "learning_rate": 3.177477632484515e-05, "loss": 0.8768, "step": 211850 }, { "epoch": 36.459050240880934, "grad_norm": 12.975618362426758, "learning_rate": 3.177047487955953e-05, "loss": 0.8312, "step": 211900 }, { "epoch": 36.46765313145217, "grad_norm": 10.177264213562012, "learning_rate": 3.1766173434273914e-05, "loss": 0.833, "step": 211950 }, { "epoch": 36.4762560220234, "grad_norm": 11.034743309020996, "learning_rate": 3.17618719889883e-05, "loss": 0.8406, "step": 212000 }, { "epoch": 36.48485891259463, "grad_norm": 13.694477081298828, "learning_rate": 3.1757570543702684e-05, "loss": 0.8389, "step": 212050 }, { "epoch": 36.493461803165864, "grad_norm": 9.627153396606445, "learning_rate": 3.175326909841707e-05, "loss": 0.8444, "step": 212100 }, { "epoch": 36.5020646937371, "grad_norm": 12.221199989318848, "learning_rate": 3.174896765313146e-05, "loss": 0.8384, "step": 212150 }, { "epoch": 36.51066758430833, "grad_norm": 9.016278266906738, "learning_rate": 3.1744666207845834e-05, "loss": 0.8753, "step": 212200 }, { "epoch": 36.51927047487956, "grad_norm": 15.83446216583252, "learning_rate": 3.174036476256022e-05, "loss": 0.8323, "step": 212250 }, { "epoch": 36.527873365450795, "grad_norm": 10.576913833618164, "learning_rate": 3.17360633172746e-05, "loss": 0.8207, "step": 212300 }, { "epoch": 36.53647625602202, "grad_norm": 11.40130615234375, "learning_rate": 3.173176187198899e-05, "loss": 0.8542, "step": 212350 }, { "epoch": 36.54507914659325, "grad_norm": 13.53122615814209, "learning_rate": 3.172746042670338e-05, "loss": 0.7925, "step": 212400 }, { "epoch": 36.553682037164485, "grad_norm": 7.746920108795166, "learning_rate": 3.172315898141775e-05, "loss": 0.8654, "step": 212450 }, { "epoch": 36.56228492773572, "grad_norm": 10.182924270629883, "learning_rate": 3.171885753613214e-05, "loss": 0.8723, "step": 212500 }, { "epoch": 36.57088781830695, "grad_norm": 13.898863792419434, "learning_rate": 3.171455609084652e-05, "loss": 0.778, "step": 212550 }, { "epoch": 36.57949070887818, "grad_norm": 8.909965515136719, "learning_rate": 3.171025464556091e-05, "loss": 0.7899, "step": 212600 }, { "epoch": 36.588093599449415, "grad_norm": 15.220949172973633, "learning_rate": 3.17059532002753e-05, "loss": 0.8664, "step": 212650 }, { "epoch": 36.59669649002065, "grad_norm": 8.621286392211914, "learning_rate": 3.170165175498968e-05, "loss": 0.8149, "step": 212700 }, { "epoch": 36.60529938059188, "grad_norm": 10.734785079956055, "learning_rate": 3.169735030970406e-05, "loss": 0.8366, "step": 212750 }, { "epoch": 36.61390227116311, "grad_norm": 9.296852111816406, "learning_rate": 3.169304886441845e-05, "loss": 0.8373, "step": 212800 }, { "epoch": 36.622505161734345, "grad_norm": 13.35773754119873, "learning_rate": 3.168874741913283e-05, "loss": 0.8536, "step": 212850 }, { "epoch": 36.63110805230558, "grad_norm": 10.948402404785156, "learning_rate": 3.168444597384722e-05, "loss": 0.8908, "step": 212900 }, { "epoch": 36.6397109428768, "grad_norm": 10.119270324707031, "learning_rate": 3.16801445285616e-05, "loss": 0.825, "step": 212950 }, { "epoch": 36.648313833448036, "grad_norm": 15.038161277770996, "learning_rate": 3.167584308327598e-05, "loss": 0.8788, "step": 213000 }, { "epoch": 36.65691672401927, "grad_norm": 12.429109573364258, "learning_rate": 3.167154163799037e-05, "loss": 0.8683, "step": 213050 }, { "epoch": 36.6655196145905, "grad_norm": 10.629107475280762, "learning_rate": 3.166724019270475e-05, "loss": 0.821, "step": 213100 }, { "epoch": 36.67412250516173, "grad_norm": 12.088299751281738, "learning_rate": 3.166293874741914e-05, "loss": 0.7901, "step": 213150 }, { "epoch": 36.682725395732966, "grad_norm": 17.220006942749023, "learning_rate": 3.165863730213352e-05, "loss": 0.8549, "step": 213200 }, { "epoch": 36.6913282863042, "grad_norm": 11.632436752319336, "learning_rate": 3.16543358568479e-05, "loss": 0.7866, "step": 213250 }, { "epoch": 36.69993117687543, "grad_norm": 6.916300296783447, "learning_rate": 3.165003441156229e-05, "loss": 0.8389, "step": 213300 }, { "epoch": 36.708534067446664, "grad_norm": 11.289132118225098, "learning_rate": 3.164573296627667e-05, "loss": 0.8316, "step": 213350 }, { "epoch": 36.717136958017896, "grad_norm": 8.92952823638916, "learning_rate": 3.1641431520991056e-05, "loss": 0.8716, "step": 213400 }, { "epoch": 36.72573984858913, "grad_norm": 14.417888641357422, "learning_rate": 3.163713007570544e-05, "loss": 0.8901, "step": 213450 }, { "epoch": 36.73434273916036, "grad_norm": 14.249300003051758, "learning_rate": 3.1632828630419826e-05, "loss": 0.8168, "step": 213500 }, { "epoch": 36.74294562973159, "grad_norm": 10.685072898864746, "learning_rate": 3.162852718513421e-05, "loss": 0.7624, "step": 213550 }, { "epoch": 36.75154852030282, "grad_norm": 13.069330215454102, "learning_rate": 3.162422573984859e-05, "loss": 0.8504, "step": 213600 }, { "epoch": 36.76015141087405, "grad_norm": 9.494500160217285, "learning_rate": 3.1619924294562976e-05, "loss": 0.8256, "step": 213650 }, { "epoch": 36.768754301445284, "grad_norm": 9.769332885742188, "learning_rate": 3.161562284927736e-05, "loss": 0.8822, "step": 213700 }, { "epoch": 36.77735719201652, "grad_norm": 10.202199935913086, "learning_rate": 3.1611321403991745e-05, "loss": 0.9007, "step": 213750 }, { "epoch": 36.78596008258775, "grad_norm": 8.966304779052734, "learning_rate": 3.1607019958706126e-05, "loss": 0.8385, "step": 213800 }, { "epoch": 36.79456297315898, "grad_norm": 15.700901985168457, "learning_rate": 3.160271851342051e-05, "loss": 0.866, "step": 213850 }, { "epoch": 36.803165863730214, "grad_norm": 11.860697746276855, "learning_rate": 3.1598417068134895e-05, "loss": 0.8363, "step": 213900 }, { "epoch": 36.81176875430145, "grad_norm": 13.565535545349121, "learning_rate": 3.1594115622849276e-05, "loss": 0.8431, "step": 213950 }, { "epoch": 36.82037164487268, "grad_norm": 12.287016868591309, "learning_rate": 3.1589814177563664e-05, "loss": 0.8608, "step": 214000 }, { "epoch": 36.82897453544391, "grad_norm": 12.788101196289062, "learning_rate": 3.1585512732278046e-05, "loss": 0.8176, "step": 214050 }, { "epoch": 36.837577426015145, "grad_norm": 15.223047256469727, "learning_rate": 3.158121128699243e-05, "loss": 0.8385, "step": 214100 }, { "epoch": 36.84618031658637, "grad_norm": 13.88369369506836, "learning_rate": 3.1576909841706815e-05, "loss": 0.8776, "step": 214150 }, { "epoch": 36.8547832071576, "grad_norm": 9.162918090820312, "learning_rate": 3.15726083964212e-05, "loss": 0.9151, "step": 214200 }, { "epoch": 36.863386097728835, "grad_norm": 9.12954330444336, "learning_rate": 3.1568306951135584e-05, "loss": 0.8816, "step": 214250 }, { "epoch": 36.87198898830007, "grad_norm": 20.75003433227539, "learning_rate": 3.156400550584997e-05, "loss": 0.8239, "step": 214300 }, { "epoch": 36.8805918788713, "grad_norm": 19.187055587768555, "learning_rate": 3.1559704060564346e-05, "loss": 0.8386, "step": 214350 }, { "epoch": 36.88919476944253, "grad_norm": 12.214380264282227, "learning_rate": 3.1555402615278734e-05, "loss": 0.8657, "step": 214400 }, { "epoch": 36.897797660013765, "grad_norm": 7.761457920074463, "learning_rate": 3.155110116999312e-05, "loss": 0.8359, "step": 214450 }, { "epoch": 36.906400550585, "grad_norm": 12.956029891967773, "learning_rate": 3.15467997247075e-05, "loss": 0.8779, "step": 214500 }, { "epoch": 36.91500344115623, "grad_norm": 13.250981330871582, "learning_rate": 3.154249827942189e-05, "loss": 0.828, "step": 214550 }, { "epoch": 36.92360633172746, "grad_norm": 14.911140441894531, "learning_rate": 3.1538196834136266e-05, "loss": 0.8298, "step": 214600 }, { "epoch": 36.932209222298695, "grad_norm": 8.965885162353516, "learning_rate": 3.1533895388850654e-05, "loss": 0.8614, "step": 214650 }, { "epoch": 36.94081211286992, "grad_norm": 7.661518573760986, "learning_rate": 3.152959394356504e-05, "loss": 0.8328, "step": 214700 }, { "epoch": 36.94941500344115, "grad_norm": 11.21900749206543, "learning_rate": 3.152529249827942e-05, "loss": 0.8362, "step": 214750 }, { "epoch": 36.958017894012386, "grad_norm": 15.49968433380127, "learning_rate": 3.152099105299381e-05, "loss": 0.8209, "step": 214800 }, { "epoch": 36.96662078458362, "grad_norm": 12.219710350036621, "learning_rate": 3.151668960770819e-05, "loss": 0.852, "step": 214850 }, { "epoch": 36.97522367515485, "grad_norm": 12.7200345993042, "learning_rate": 3.151238816242257e-05, "loss": 0.8735, "step": 214900 }, { "epoch": 36.98382656572608, "grad_norm": 11.538724899291992, "learning_rate": 3.150808671713696e-05, "loss": 0.867, "step": 214950 }, { "epoch": 36.992429456297316, "grad_norm": 11.255666732788086, "learning_rate": 3.150378527185134e-05, "loss": 0.8108, "step": 215000 }, { "epoch": 37.0, "eval_accuracy": 0.6030803648253312, "eval_f1": 0.6011022980842771, "eval_f1_DuraRiadoRio_16x16": 0.6077687443541102, "eval_f1_Mole_16x16": 0.6357492426007924, "eval_f1_Quebrado_16x16": 0.6369919823057782, "eval_f1_RiadoRio_16x16": 0.503138337936229, "eval_f1_RioFechado_16x16": 0.6218631832244758, "eval_loss": 1.4273808002471924, "eval_precision": 0.6482951354633755, "eval_precision_DuraRiadoRio_16x16": 0.5205818632002476, "eval_precision_Mole_16x16": 0.6864620030196276, "eval_precision_Quebrado_16x16": 0.8773800456968773, "eval_precision_RiadoRio_16x16": 0.6460348162475822, "eval_precision_RioFechado_16x16": 0.5110169491525424, "eval_recall": 0.6056345674234607, "eval_recall_DuraRiadoRio_16x16": 0.7300347222222222, "eval_recall_Mole_16x16": 0.5920138888888888, "eval_recall_Quebrado_16x16": 0.5, "eval_recall_RiadoRio_16x16": 0.4120065789473684, "eval_recall_RioFechado_16x16": 0.7941176470588235, "eval_runtime": 45.2819, "eval_samples_per_second": 256.659, "eval_steps_per_second": 16.055, "step": 215044 }, { "epoch": 37.00103234686855, "grad_norm": 11.648045539855957, "learning_rate": 3.149948382656573e-05, "loss": 0.8778, "step": 215050 }, { "epoch": 37.00963523743978, "grad_norm": 11.781638145446777, "learning_rate": 3.149518238128011e-05, "loss": 0.8352, "step": 215100 }, { "epoch": 37.01823812801101, "grad_norm": 9.80825424194336, "learning_rate": 3.149088093599449e-05, "loss": 0.8517, "step": 215150 }, { "epoch": 37.026841018582246, "grad_norm": 10.323509216308594, "learning_rate": 3.148657949070888e-05, "loss": 0.8316, "step": 215200 }, { "epoch": 37.03544390915348, "grad_norm": 8.568190574645996, "learning_rate": 3.148227804542326e-05, "loss": 0.9008, "step": 215250 }, { "epoch": 37.044046799724704, "grad_norm": 8.059196472167969, "learning_rate": 3.147797660013765e-05, "loss": 0.8588, "step": 215300 }, { "epoch": 37.05264969029594, "grad_norm": 8.097976684570312, "learning_rate": 3.147367515485203e-05, "loss": 0.836, "step": 215350 }, { "epoch": 37.06125258086717, "grad_norm": 8.022001266479492, "learning_rate": 3.146937370956641e-05, "loss": 0.8163, "step": 215400 }, { "epoch": 37.0698554714384, "grad_norm": 13.187414169311523, "learning_rate": 3.14650722642808e-05, "loss": 0.8196, "step": 215450 }, { "epoch": 37.078458362009634, "grad_norm": 13.757959365844727, "learning_rate": 3.146077081899518e-05, "loss": 0.8029, "step": 215500 }, { "epoch": 37.08706125258087, "grad_norm": 13.255040168762207, "learning_rate": 3.145646937370957e-05, "loss": 0.8627, "step": 215550 }, { "epoch": 37.0956641431521, "grad_norm": 11.285919189453125, "learning_rate": 3.145216792842396e-05, "loss": 0.7898, "step": 215600 }, { "epoch": 37.10426703372333, "grad_norm": 10.390264511108398, "learning_rate": 3.144786648313834e-05, "loss": 0.8529, "step": 215650 }, { "epoch": 37.112869924294564, "grad_norm": 13.803706169128418, "learning_rate": 3.144356503785272e-05, "loss": 0.8187, "step": 215700 }, { "epoch": 37.1214728148658, "grad_norm": 10.15179443359375, "learning_rate": 3.14392635925671e-05, "loss": 0.8064, "step": 215750 }, { "epoch": 37.13007570543703, "grad_norm": 9.411649703979492, "learning_rate": 3.143496214728149e-05, "loss": 0.8157, "step": 215800 }, { "epoch": 37.13867859600826, "grad_norm": 7.0661749839782715, "learning_rate": 3.1430660701995876e-05, "loss": 0.7799, "step": 215850 }, { "epoch": 37.14728148657949, "grad_norm": 9.21900749206543, "learning_rate": 3.142635925671026e-05, "loss": 0.8229, "step": 215900 }, { "epoch": 37.15588437715072, "grad_norm": 12.955976486206055, "learning_rate": 3.142205781142464e-05, "loss": 0.815, "step": 215950 }, { "epoch": 37.16448726772195, "grad_norm": 13.09589958190918, "learning_rate": 3.141775636613902e-05, "loss": 0.8574, "step": 216000 }, { "epoch": 37.173090158293185, "grad_norm": 9.9086275100708, "learning_rate": 3.141345492085341e-05, "loss": 0.7707, "step": 216050 }, { "epoch": 37.18169304886442, "grad_norm": 13.493806838989258, "learning_rate": 3.1409153475567796e-05, "loss": 0.8562, "step": 216100 }, { "epoch": 37.19029593943565, "grad_norm": 11.721061706542969, "learning_rate": 3.140485203028218e-05, "loss": 0.8798, "step": 216150 }, { "epoch": 37.19889883000688, "grad_norm": 9.863381385803223, "learning_rate": 3.1400550584996565e-05, "loss": 0.8707, "step": 216200 }, { "epoch": 37.207501720578115, "grad_norm": 9.119588851928711, "learning_rate": 3.1396249139710946e-05, "loss": 0.9179, "step": 216250 }, { "epoch": 37.21610461114935, "grad_norm": 16.232250213623047, "learning_rate": 3.139194769442533e-05, "loss": 0.8509, "step": 216300 }, { "epoch": 37.22470750172058, "grad_norm": 10.192798614501953, "learning_rate": 3.1387646249139715e-05, "loss": 0.837, "step": 216350 }, { "epoch": 37.23331039229181, "grad_norm": 7.919525623321533, "learning_rate": 3.1383344803854096e-05, "loss": 0.8216, "step": 216400 }, { "epoch": 37.241913282863045, "grad_norm": 9.884795188903809, "learning_rate": 3.1379043358568484e-05, "loss": 0.8394, "step": 216450 }, { "epoch": 37.25051617343427, "grad_norm": 10.56735610961914, "learning_rate": 3.1374741913282865e-05, "loss": 0.8678, "step": 216500 }, { "epoch": 37.2591190640055, "grad_norm": 14.768850326538086, "learning_rate": 3.1370440467997247e-05, "loss": 0.841, "step": 216550 }, { "epoch": 37.267721954576736, "grad_norm": 8.617777824401855, "learning_rate": 3.1366139022711634e-05, "loss": 0.7646, "step": 216600 }, { "epoch": 37.27632484514797, "grad_norm": 13.067809104919434, "learning_rate": 3.1361837577426016e-05, "loss": 0.8239, "step": 216650 }, { "epoch": 37.2849277357192, "grad_norm": 9.353818893432617, "learning_rate": 3.1357536132140404e-05, "loss": 0.7596, "step": 216700 }, { "epoch": 37.29353062629043, "grad_norm": 13.119492530822754, "learning_rate": 3.1353234686854785e-05, "loss": 0.8566, "step": 216750 }, { "epoch": 37.302133516861666, "grad_norm": 12.483606338500977, "learning_rate": 3.1348933241569166e-05, "loss": 0.8582, "step": 216800 }, { "epoch": 37.3107364074329, "grad_norm": 13.717815399169922, "learning_rate": 3.1344631796283554e-05, "loss": 0.8566, "step": 216850 }, { "epoch": 37.31933929800413, "grad_norm": 14.465103149414062, "learning_rate": 3.1340330350997935e-05, "loss": 0.818, "step": 216900 }, { "epoch": 37.32794218857536, "grad_norm": 13.79445743560791, "learning_rate": 3.133602890571232e-05, "loss": 0.8495, "step": 216950 }, { "epoch": 37.336545079146596, "grad_norm": 12.804020881652832, "learning_rate": 3.1331727460426704e-05, "loss": 0.8175, "step": 217000 }, { "epoch": 37.34514796971783, "grad_norm": 20.399627685546875, "learning_rate": 3.1327426015141085e-05, "loss": 0.8038, "step": 217050 }, { "epoch": 37.353750860289054, "grad_norm": 16.706613540649414, "learning_rate": 3.132312456985547e-05, "loss": 0.8679, "step": 217100 }, { "epoch": 37.36235375086029, "grad_norm": 12.519326210021973, "learning_rate": 3.1318823124569854e-05, "loss": 0.8347, "step": 217150 }, { "epoch": 37.37095664143152, "grad_norm": 15.160606384277344, "learning_rate": 3.131452167928424e-05, "loss": 0.8258, "step": 217200 }, { "epoch": 37.37955953200275, "grad_norm": 21.666847229003906, "learning_rate": 3.131022023399863e-05, "loss": 0.8585, "step": 217250 }, { "epoch": 37.388162422573984, "grad_norm": 9.60906982421875, "learning_rate": 3.1305918788713005e-05, "loss": 0.8326, "step": 217300 }, { "epoch": 37.39676531314522, "grad_norm": 8.391597747802734, "learning_rate": 3.130161734342739e-05, "loss": 0.8881, "step": 217350 }, { "epoch": 37.40536820371645, "grad_norm": 8.070333480834961, "learning_rate": 3.1297315898141774e-05, "loss": 0.8353, "step": 217400 }, { "epoch": 37.41397109428768, "grad_norm": 15.130317687988281, "learning_rate": 3.129301445285616e-05, "loss": 0.8474, "step": 217450 }, { "epoch": 37.422573984858914, "grad_norm": 9.785294532775879, "learning_rate": 3.128871300757055e-05, "loss": 0.8643, "step": 217500 }, { "epoch": 37.43117687543015, "grad_norm": 11.46236515045166, "learning_rate": 3.1284411562284924e-05, "loss": 0.868, "step": 217550 }, { "epoch": 37.43977976600138, "grad_norm": 8.768597602844238, "learning_rate": 3.128011011699931e-05, "loss": 0.7848, "step": 217600 }, { "epoch": 37.44838265657261, "grad_norm": 12.7876615524292, "learning_rate": 3.127580867171369e-05, "loss": 0.8921, "step": 217650 }, { "epoch": 37.45698554714384, "grad_norm": 17.081323623657227, "learning_rate": 3.127150722642808e-05, "loss": 0.8723, "step": 217700 }, { "epoch": 37.46558843771507, "grad_norm": 16.133724212646484, "learning_rate": 3.126720578114247e-05, "loss": 0.8202, "step": 217750 }, { "epoch": 37.4741913282863, "grad_norm": 12.028861999511719, "learning_rate": 3.126290433585685e-05, "loss": 0.852, "step": 217800 }, { "epoch": 37.482794218857535, "grad_norm": 7.93231201171875, "learning_rate": 3.125860289057123e-05, "loss": 0.8301, "step": 217850 }, { "epoch": 37.49139710942877, "grad_norm": 19.489643096923828, "learning_rate": 3.125430144528562e-05, "loss": 0.8668, "step": 217900 }, { "epoch": 37.5, "grad_norm": 16.72347640991211, "learning_rate": 3.125e-05, "loss": 0.8647, "step": 217950 }, { "epoch": 37.50860289057123, "grad_norm": 15.92646312713623, "learning_rate": 3.124569855471439e-05, "loss": 0.8911, "step": 218000 }, { "epoch": 37.517205781142465, "grad_norm": 13.877013206481934, "learning_rate": 3.124139710942877e-05, "loss": 0.8578, "step": 218050 }, { "epoch": 37.5258086717137, "grad_norm": 9.733473777770996, "learning_rate": 3.123709566414315e-05, "loss": 0.822, "step": 218100 }, { "epoch": 37.53441156228493, "grad_norm": 14.212911605834961, "learning_rate": 3.123279421885754e-05, "loss": 0.8224, "step": 218150 }, { "epoch": 37.54301445285616, "grad_norm": 10.331561088562012, "learning_rate": 3.122849277357192e-05, "loss": 0.8887, "step": 218200 }, { "epoch": 37.55161734342739, "grad_norm": 14.534672737121582, "learning_rate": 3.122419132828631e-05, "loss": 0.8874, "step": 218250 }, { "epoch": 37.56022023399862, "grad_norm": 11.678380966186523, "learning_rate": 3.121988988300069e-05, "loss": 0.8363, "step": 218300 }, { "epoch": 37.56882312456985, "grad_norm": 15.745450019836426, "learning_rate": 3.121558843771508e-05, "loss": 0.8403, "step": 218350 }, { "epoch": 37.577426015141086, "grad_norm": 13.254032135009766, "learning_rate": 3.121128699242946e-05, "loss": 0.8418, "step": 218400 }, { "epoch": 37.58602890571232, "grad_norm": 10.24193286895752, "learning_rate": 3.120698554714384e-05, "loss": 0.889, "step": 218450 }, { "epoch": 37.59463179628355, "grad_norm": 12.145990371704102, "learning_rate": 3.120268410185823e-05, "loss": 0.882, "step": 218500 }, { "epoch": 37.60323468685478, "grad_norm": 11.320862770080566, "learning_rate": 3.119838265657261e-05, "loss": 0.8168, "step": 218550 }, { "epoch": 37.611837577426016, "grad_norm": 14.369271278381348, "learning_rate": 3.1194081211286997e-05, "loss": 0.8349, "step": 218600 }, { "epoch": 37.62044046799725, "grad_norm": 9.293390274047852, "learning_rate": 3.118977976600138e-05, "loss": 0.7985, "step": 218650 }, { "epoch": 37.62904335856848, "grad_norm": 12.415495872497559, "learning_rate": 3.118547832071576e-05, "loss": 0.777, "step": 218700 }, { "epoch": 37.63764624913971, "grad_norm": 7.257807731628418, "learning_rate": 3.118117687543015e-05, "loss": 0.8389, "step": 218750 }, { "epoch": 37.646249139710946, "grad_norm": 8.756677627563477, "learning_rate": 3.117687543014453e-05, "loss": 0.8393, "step": 218800 }, { "epoch": 37.65485203028217, "grad_norm": 9.351263999938965, "learning_rate": 3.1172573984858916e-05, "loss": 0.8291, "step": 218850 }, { "epoch": 37.663454920853404, "grad_norm": 13.980607986450195, "learning_rate": 3.11682725395733e-05, "loss": 0.7945, "step": 218900 }, { "epoch": 37.67205781142464, "grad_norm": 10.676091194152832, "learning_rate": 3.116397109428768e-05, "loss": 0.8689, "step": 218950 }, { "epoch": 37.68066070199587, "grad_norm": 17.025287628173828, "learning_rate": 3.1159669649002066e-05, "loss": 0.8061, "step": 219000 }, { "epoch": 37.6892635925671, "grad_norm": 15.366522789001465, "learning_rate": 3.115536820371645e-05, "loss": 0.8632, "step": 219050 }, { "epoch": 37.697866483138334, "grad_norm": 8.967570304870605, "learning_rate": 3.1151066758430835e-05, "loss": 0.8559, "step": 219100 }, { "epoch": 37.70646937370957, "grad_norm": 9.633513450622559, "learning_rate": 3.114676531314522e-05, "loss": 0.8852, "step": 219150 }, { "epoch": 37.7150722642808, "grad_norm": 9.013849258422852, "learning_rate": 3.11424638678596e-05, "loss": 0.8175, "step": 219200 }, { "epoch": 37.72367515485203, "grad_norm": 13.81950569152832, "learning_rate": 3.1138162422573986e-05, "loss": 0.8379, "step": 219250 }, { "epoch": 37.732278045423264, "grad_norm": 9.807594299316406, "learning_rate": 3.1133860977288374e-05, "loss": 0.8254, "step": 219300 }, { "epoch": 37.7408809359945, "grad_norm": 13.579500198364258, "learning_rate": 3.1129559532002755e-05, "loss": 0.859, "step": 219350 }, { "epoch": 37.74948382656573, "grad_norm": 16.11495018005371, "learning_rate": 3.112525808671714e-05, "loss": 0.8304, "step": 219400 }, { "epoch": 37.758086717136955, "grad_norm": 10.784406661987305, "learning_rate": 3.112095664143152e-05, "loss": 0.8887, "step": 219450 }, { "epoch": 37.76668960770819, "grad_norm": 18.662675857543945, "learning_rate": 3.1116655196145905e-05, "loss": 0.8831, "step": 219500 }, { "epoch": 37.77529249827942, "grad_norm": 15.991307258605957, "learning_rate": 3.111235375086029e-05, "loss": 0.8858, "step": 219550 }, { "epoch": 37.78389538885065, "grad_norm": 15.302413940429688, "learning_rate": 3.1108052305574674e-05, "loss": 0.8138, "step": 219600 }, { "epoch": 37.792498279421885, "grad_norm": 9.09868335723877, "learning_rate": 3.110375086028906e-05, "loss": 0.828, "step": 219650 }, { "epoch": 37.80110116999312, "grad_norm": 20.333744049072266, "learning_rate": 3.1099449415003437e-05, "loss": 0.853, "step": 219700 }, { "epoch": 37.80970406056435, "grad_norm": 10.877531051635742, "learning_rate": 3.1095147969717824e-05, "loss": 0.8043, "step": 219750 }, { "epoch": 37.81830695113558, "grad_norm": 8.216778755187988, "learning_rate": 3.109084652443221e-05, "loss": 0.8445, "step": 219800 }, { "epoch": 37.826909841706815, "grad_norm": 8.740213394165039, "learning_rate": 3.1086545079146594e-05, "loss": 0.8417, "step": 219850 }, { "epoch": 37.83551273227805, "grad_norm": 9.853479385375977, "learning_rate": 3.108224363386098e-05, "loss": 0.8503, "step": 219900 }, { "epoch": 37.84411562284928, "grad_norm": 15.297186851501465, "learning_rate": 3.107794218857536e-05, "loss": 0.8187, "step": 219950 }, { "epoch": 37.85271851342051, "grad_norm": 16.220014572143555, "learning_rate": 3.1073640743289744e-05, "loss": 0.8695, "step": 220000 }, { "epoch": 37.86132140399174, "grad_norm": 9.988815307617188, "learning_rate": 3.106933929800413e-05, "loss": 0.8166, "step": 220050 }, { "epoch": 37.86992429456297, "grad_norm": 11.127697944641113, "learning_rate": 3.106503785271851e-05, "loss": 0.8131, "step": 220100 }, { "epoch": 37.8785271851342, "grad_norm": 10.753122329711914, "learning_rate": 3.10607364074329e-05, "loss": 0.8145, "step": 220150 }, { "epoch": 37.887130075705436, "grad_norm": 14.345871925354004, "learning_rate": 3.105643496214728e-05, "loss": 0.8574, "step": 220200 }, { "epoch": 37.89573296627667, "grad_norm": 13.05732536315918, "learning_rate": 3.105213351686166e-05, "loss": 0.8319, "step": 220250 }, { "epoch": 37.9043358568479, "grad_norm": 9.9769926071167, "learning_rate": 3.104783207157605e-05, "loss": 0.7967, "step": 220300 }, { "epoch": 37.91293874741913, "grad_norm": 15.037422180175781, "learning_rate": 3.104353062629043e-05, "loss": 0.8822, "step": 220350 }, { "epoch": 37.921541637990366, "grad_norm": 11.529967308044434, "learning_rate": 3.103922918100482e-05, "loss": 0.8436, "step": 220400 }, { "epoch": 37.9301445285616, "grad_norm": 10.409283638000488, "learning_rate": 3.10349277357192e-05, "loss": 0.8293, "step": 220450 }, { "epoch": 37.93874741913283, "grad_norm": 8.534031867980957, "learning_rate": 3.103062629043359e-05, "loss": 0.8579, "step": 220500 }, { "epoch": 37.94735030970406, "grad_norm": 13.278609275817871, "learning_rate": 3.102632484514797e-05, "loss": 0.8579, "step": 220550 }, { "epoch": 37.955953200275296, "grad_norm": 8.066826820373535, "learning_rate": 3.102202339986235e-05, "loss": 0.8888, "step": 220600 }, { "epoch": 37.96455609084652, "grad_norm": 9.460746765136719, "learning_rate": 3.101772195457674e-05, "loss": 0.8837, "step": 220650 }, { "epoch": 37.973158981417754, "grad_norm": 9.436470031738281, "learning_rate": 3.101342050929113e-05, "loss": 0.8269, "step": 220700 }, { "epoch": 37.98176187198899, "grad_norm": 8.228453636169434, "learning_rate": 3.100911906400551e-05, "loss": 0.8297, "step": 220750 }, { "epoch": 37.99036476256022, "grad_norm": 14.060323715209961, "learning_rate": 3.100481761871989e-05, "loss": 0.8559, "step": 220800 }, { "epoch": 37.99896765313145, "grad_norm": 10.818705558776855, "learning_rate": 3.100051617343427e-05, "loss": 0.8155, "step": 220850 }, { "epoch": 38.0, "eval_accuracy": 0.5718464980209946, "eval_f1": 0.561116984953236, "eval_f1_DuraRiadoRio_16x16": 0.5526785714285715, "eval_f1_Mole_16x16": 0.44062688366485836, "eval_f1_Quebrado_16x16": 0.6944696282864914, "eval_f1_RiadoRio_16x16": 0.485131195335277, "eval_f1_RioFechado_16x16": 0.632678646050982, "eval_loss": 2.1081631183624268, "eval_precision": 0.5899808346171126, "eval_precision_DuraRiadoRio_16x16": 0.5689338235294118, "eval_precision_Mole_16x16": 0.7209072978303748, "eval_precision_Quebrado_16x16": 0.5963874182497664, "eval_precision_RiadoRio_16x16": 0.46000737191301144, "eval_precision_RioFechado_16x16": 0.6036682615629984, "eval_recall": 0.5727079739332235, "eval_recall_DuraRiadoRio_16x16": 0.5373263888888888, "eval_recall_Mole_16x16": 0.3172743055555556, "eval_recall_Quebrado_16x16": 0.8311631944444444, "eval_recall_RiadoRio_16x16": 0.5131578947368421, "eval_recall_RioFechado_16x16": 0.6646180860403863, "eval_runtime": 45.856, "eval_samples_per_second": 253.446, "eval_steps_per_second": 15.854, "step": 220856 }, { "epoch": 38.007570543702684, "grad_norm": 6.313615322113037, "learning_rate": 3.099621472814866e-05, "loss": 0.8434, "step": 220900 }, { "epoch": 38.01617343427392, "grad_norm": 13.431744575500488, "learning_rate": 3.099191328286305e-05, "loss": 0.7919, "step": 220950 }, { "epoch": 38.02477632484515, "grad_norm": 9.242823600769043, "learning_rate": 3.098761183757743e-05, "loss": 0.7561, "step": 221000 }, { "epoch": 38.03337921541638, "grad_norm": 5.692231178283691, "learning_rate": 3.098331039229181e-05, "loss": 0.8638, "step": 221050 }, { "epoch": 38.041982105987614, "grad_norm": 10.526327133178711, "learning_rate": 3.097900894700619e-05, "loss": 0.8685, "step": 221100 }, { "epoch": 38.05058499655885, "grad_norm": 7.239251136779785, "learning_rate": 3.097470750172058e-05, "loss": 0.8722, "step": 221150 }, { "epoch": 38.05918788713007, "grad_norm": 12.368369102478027, "learning_rate": 3.0970406056434967e-05, "loss": 0.8273, "step": 221200 }, { "epoch": 38.067790777701305, "grad_norm": 17.53847885131836, "learning_rate": 3.096610461114935e-05, "loss": 0.9077, "step": 221250 }, { "epoch": 38.07639366827254, "grad_norm": 12.469592094421387, "learning_rate": 3.0961803165863736e-05, "loss": 0.86, "step": 221300 }, { "epoch": 38.08499655884377, "grad_norm": 12.633320808410645, "learning_rate": 3.095750172057812e-05, "loss": 0.7985, "step": 221350 }, { "epoch": 38.093599449415, "grad_norm": 11.26282787322998, "learning_rate": 3.09532002752925e-05, "loss": 0.8721, "step": 221400 }, { "epoch": 38.102202339986235, "grad_norm": 8.540878295898438, "learning_rate": 3.0948898830006886e-05, "loss": 0.872, "step": 221450 }, { "epoch": 38.11080523055747, "grad_norm": 14.092565536499023, "learning_rate": 3.094459738472127e-05, "loss": 0.854, "step": 221500 }, { "epoch": 38.1194081211287, "grad_norm": 7.167665004730225, "learning_rate": 3.0940295939435655e-05, "loss": 0.8919, "step": 221550 }, { "epoch": 38.12801101169993, "grad_norm": 10.959187507629395, "learning_rate": 3.0935994494150036e-05, "loss": 0.7794, "step": 221600 }, { "epoch": 38.136613902271165, "grad_norm": 13.90500259399414, "learning_rate": 3.093169304886442e-05, "loss": 0.7834, "step": 221650 }, { "epoch": 38.1452167928424, "grad_norm": 14.182353973388672, "learning_rate": 3.0927391603578805e-05, "loss": 0.8497, "step": 221700 }, { "epoch": 38.15381968341363, "grad_norm": 19.11916160583496, "learning_rate": 3.0923090158293187e-05, "loss": 0.875, "step": 221750 }, { "epoch": 38.162422573984855, "grad_norm": 10.889452934265137, "learning_rate": 3.0918788713007575e-05, "loss": 0.7704, "step": 221800 }, { "epoch": 38.17102546455609, "grad_norm": 11.562237739562988, "learning_rate": 3.0914487267721956e-05, "loss": 0.8314, "step": 221850 }, { "epoch": 38.17962835512732, "grad_norm": 13.061066627502441, "learning_rate": 3.091018582243634e-05, "loss": 0.82, "step": 221900 }, { "epoch": 38.18823124569855, "grad_norm": 13.227030754089355, "learning_rate": 3.0905884377150725e-05, "loss": 0.8411, "step": 221950 }, { "epoch": 38.196834136269786, "grad_norm": 13.091363906860352, "learning_rate": 3.0901582931865106e-05, "loss": 0.8151, "step": 222000 }, { "epoch": 38.20543702684102, "grad_norm": 11.223121643066406, "learning_rate": 3.0897281486579494e-05, "loss": 0.8788, "step": 222050 }, { "epoch": 38.21403991741225, "grad_norm": 11.616930961608887, "learning_rate": 3.089298004129388e-05, "loss": 0.8796, "step": 222100 }, { "epoch": 38.22264280798348, "grad_norm": 11.129182815551758, "learning_rate": 3.0888678596008256e-05, "loss": 0.81, "step": 222150 }, { "epoch": 38.231245698554716, "grad_norm": 8.903253555297852, "learning_rate": 3.0884377150722644e-05, "loss": 0.8479, "step": 222200 }, { "epoch": 38.23984858912595, "grad_norm": 11.169637680053711, "learning_rate": 3.0880075705437025e-05, "loss": 0.8541, "step": 222250 }, { "epoch": 38.24845147969718, "grad_norm": 14.711309432983398, "learning_rate": 3.087577426015141e-05, "loss": 0.7943, "step": 222300 }, { "epoch": 38.25705437026841, "grad_norm": 9.840728759765625, "learning_rate": 3.08714728148658e-05, "loss": 0.8606, "step": 222350 }, { "epoch": 38.26565726083964, "grad_norm": 9.896696090698242, "learning_rate": 3.0867171369580176e-05, "loss": 0.7695, "step": 222400 }, { "epoch": 38.27426015141087, "grad_norm": 12.823501586914062, "learning_rate": 3.0862869924294564e-05, "loss": 0.8725, "step": 222450 }, { "epoch": 38.282863041982104, "grad_norm": 13.091451644897461, "learning_rate": 3.0858568479008945e-05, "loss": 0.9214, "step": 222500 }, { "epoch": 38.291465932553336, "grad_norm": 13.96158218383789, "learning_rate": 3.085426703372333e-05, "loss": 0.8755, "step": 222550 }, { "epoch": 38.30006882312457, "grad_norm": 6.914041519165039, "learning_rate": 3.084996558843772e-05, "loss": 0.8632, "step": 222600 }, { "epoch": 38.3086717136958, "grad_norm": 16.598268508911133, "learning_rate": 3.08456641431521e-05, "loss": 0.8589, "step": 222650 }, { "epoch": 38.317274604267034, "grad_norm": 8.038679122924805, "learning_rate": 3.084136269786648e-05, "loss": 0.7884, "step": 222700 }, { "epoch": 38.32587749483827, "grad_norm": 10.656070709228516, "learning_rate": 3.083706125258087e-05, "loss": 0.8278, "step": 222750 }, { "epoch": 38.3344803854095, "grad_norm": 13.108458518981934, "learning_rate": 3.083275980729525e-05, "loss": 0.8616, "step": 222800 }, { "epoch": 38.34308327598073, "grad_norm": 24.692485809326172, "learning_rate": 3.082845836200964e-05, "loss": 0.8538, "step": 222850 }, { "epoch": 38.351686166551964, "grad_norm": 10.5029296875, "learning_rate": 3.082415691672402e-05, "loss": 0.7972, "step": 222900 }, { "epoch": 38.3602890571232, "grad_norm": 12.13703727722168, "learning_rate": 3.08198554714384e-05, "loss": 0.7807, "step": 222950 }, { "epoch": 38.36889194769442, "grad_norm": 10.453768730163574, "learning_rate": 3.081555402615279e-05, "loss": 0.8711, "step": 223000 }, { "epoch": 38.377494838265655, "grad_norm": 13.59195613861084, "learning_rate": 3.081125258086717e-05, "loss": 0.8566, "step": 223050 }, { "epoch": 38.38609772883689, "grad_norm": 12.778953552246094, "learning_rate": 3.080695113558156e-05, "loss": 0.819, "step": 223100 }, { "epoch": 38.39470061940812, "grad_norm": 14.787617683410645, "learning_rate": 3.080264969029594e-05, "loss": 0.8143, "step": 223150 }, { "epoch": 38.40330350997935, "grad_norm": 18.432621002197266, "learning_rate": 3.079834824501032e-05, "loss": 0.7806, "step": 223200 }, { "epoch": 38.411906400550585, "grad_norm": 10.887399673461914, "learning_rate": 3.079404679972471e-05, "loss": 0.8379, "step": 223250 }, { "epoch": 38.42050929112182, "grad_norm": 10.102716445922852, "learning_rate": 3.078974535443909e-05, "loss": 0.8384, "step": 223300 }, { "epoch": 38.42911218169305, "grad_norm": 11.034936904907227, "learning_rate": 3.078544390915348e-05, "loss": 0.7763, "step": 223350 }, { "epoch": 38.43771507226428, "grad_norm": 8.516402244567871, "learning_rate": 3.078114246386786e-05, "loss": 0.8332, "step": 223400 }, { "epoch": 38.446317962835515, "grad_norm": 12.99522876739502, "learning_rate": 3.077684101858225e-05, "loss": 0.8402, "step": 223450 }, { "epoch": 38.45492085340675, "grad_norm": 17.361448287963867, "learning_rate": 3.077253957329663e-05, "loss": 0.7904, "step": 223500 }, { "epoch": 38.46352374397798, "grad_norm": 9.736222267150879, "learning_rate": 3.076823812801101e-05, "loss": 0.7936, "step": 223550 }, { "epoch": 38.472126634549205, "grad_norm": 10.301980018615723, "learning_rate": 3.07639366827254e-05, "loss": 0.8334, "step": 223600 }, { "epoch": 38.48072952512044, "grad_norm": 15.77480411529541, "learning_rate": 3.075963523743978e-05, "loss": 0.8418, "step": 223650 }, { "epoch": 38.48933241569167, "grad_norm": 11.10028076171875, "learning_rate": 3.075533379215417e-05, "loss": 0.8243, "step": 223700 }, { "epoch": 38.4979353062629, "grad_norm": 11.654536247253418, "learning_rate": 3.075103234686855e-05, "loss": 0.8273, "step": 223750 }, { "epoch": 38.506538196834136, "grad_norm": 13.105803489685059, "learning_rate": 3.074673090158293e-05, "loss": 0.8501, "step": 223800 }, { "epoch": 38.51514108740537, "grad_norm": 12.814179420471191, "learning_rate": 3.074242945629732e-05, "loss": 0.8193, "step": 223850 }, { "epoch": 38.5237439779766, "grad_norm": 13.412431716918945, "learning_rate": 3.07381280110117e-05, "loss": 0.9179, "step": 223900 }, { "epoch": 38.53234686854783, "grad_norm": 10.652851104736328, "learning_rate": 3.073382656572609e-05, "loss": 0.8314, "step": 223950 }, { "epoch": 38.540949759119066, "grad_norm": 9.0834321975708, "learning_rate": 3.0729525120440475e-05, "loss": 0.7971, "step": 224000 }, { "epoch": 38.5495526496903, "grad_norm": 15.180404663085938, "learning_rate": 3.072522367515485e-05, "loss": 0.8709, "step": 224050 }, { "epoch": 38.55815554026153, "grad_norm": 18.10888671875, "learning_rate": 3.072092222986924e-05, "loss": 0.7875, "step": 224100 }, { "epoch": 38.56675843083276, "grad_norm": 12.444089889526367, "learning_rate": 3.0716620784583625e-05, "loss": 0.9232, "step": 224150 }, { "epoch": 38.57536132140399, "grad_norm": 11.506142616271973, "learning_rate": 3.0712319339298006e-05, "loss": 0.8655, "step": 224200 }, { "epoch": 38.58396421197522, "grad_norm": 19.037099838256836, "learning_rate": 3.0708017894012394e-05, "loss": 0.81, "step": 224250 }, { "epoch": 38.592567102546454, "grad_norm": 10.260311126708984, "learning_rate": 3.070371644872677e-05, "loss": 0.8713, "step": 224300 }, { "epoch": 38.601169993117686, "grad_norm": 9.203017234802246, "learning_rate": 3.069941500344116e-05, "loss": 0.8272, "step": 224350 }, { "epoch": 38.60977288368892, "grad_norm": 9.76535415649414, "learning_rate": 3.0695113558155545e-05, "loss": 0.818, "step": 224400 }, { "epoch": 38.61837577426015, "grad_norm": 8.359705924987793, "learning_rate": 3.0690812112869926e-05, "loss": 0.8214, "step": 224450 }, { "epoch": 38.626978664831384, "grad_norm": 12.801188468933105, "learning_rate": 3.0686510667584314e-05, "loss": 0.8287, "step": 224500 }, { "epoch": 38.63558155540262, "grad_norm": 15.298574447631836, "learning_rate": 3.068220922229869e-05, "loss": 0.8388, "step": 224550 }, { "epoch": 38.64418444597385, "grad_norm": 16.953529357910156, "learning_rate": 3.0677907777013076e-05, "loss": 0.8839, "step": 224600 }, { "epoch": 38.65278733654508, "grad_norm": 16.55194091796875, "learning_rate": 3.0673606331727464e-05, "loss": 0.8394, "step": 224650 }, { "epoch": 38.661390227116314, "grad_norm": 13.140300750732422, "learning_rate": 3.0669304886441845e-05, "loss": 0.8074, "step": 224700 }, { "epoch": 38.66999311768754, "grad_norm": 13.82580280303955, "learning_rate": 3.066500344115623e-05, "loss": 0.8266, "step": 224750 }, { "epoch": 38.67859600825877, "grad_norm": 12.801486015319824, "learning_rate": 3.0660701995870614e-05, "loss": 0.8752, "step": 224800 }, { "epoch": 38.687198898830005, "grad_norm": 14.460489273071289, "learning_rate": 3.0656400550584995e-05, "loss": 0.8359, "step": 224850 }, { "epoch": 38.69580178940124, "grad_norm": 17.528152465820312, "learning_rate": 3.0652099105299383e-05, "loss": 0.8376, "step": 224900 }, { "epoch": 38.70440467997247, "grad_norm": 13.116251945495605, "learning_rate": 3.0647797660013765e-05, "loss": 0.8302, "step": 224950 }, { "epoch": 38.7130075705437, "grad_norm": 11.041553497314453, "learning_rate": 3.064349621472815e-05, "loss": 0.8894, "step": 225000 }, { "epoch": 38.721610461114935, "grad_norm": 12.072100639343262, "learning_rate": 3.0639194769442534e-05, "loss": 0.8553, "step": 225050 }, { "epoch": 38.73021335168617, "grad_norm": 12.661510467529297, "learning_rate": 3.0634893324156915e-05, "loss": 0.809, "step": 225100 }, { "epoch": 38.7388162422574, "grad_norm": 8.987753868103027, "learning_rate": 3.06305918788713e-05, "loss": 0.857, "step": 225150 }, { "epoch": 38.74741913282863, "grad_norm": 12.951506614685059, "learning_rate": 3.0626290433585684e-05, "loss": 0.8034, "step": 225200 }, { "epoch": 38.756022023399865, "grad_norm": 13.372241973876953, "learning_rate": 3.062198898830007e-05, "loss": 0.8195, "step": 225250 }, { "epoch": 38.7646249139711, "grad_norm": 8.355611801147461, "learning_rate": 3.061768754301445e-05, "loss": 0.8086, "step": 225300 }, { "epoch": 38.77322780454232, "grad_norm": 7.68317985534668, "learning_rate": 3.0613386097728834e-05, "loss": 0.7635, "step": 225350 }, { "epoch": 38.781830695113555, "grad_norm": 14.048848152160645, "learning_rate": 3.060908465244322e-05, "loss": 0.8247, "step": 225400 }, { "epoch": 38.79043358568479, "grad_norm": 11.125616073608398, "learning_rate": 3.0604783207157603e-05, "loss": 0.7931, "step": 225450 }, { "epoch": 38.79903647625602, "grad_norm": 12.657208442687988, "learning_rate": 3.060048176187199e-05, "loss": 0.8752, "step": 225500 }, { "epoch": 38.80763936682725, "grad_norm": 8.58807373046875, "learning_rate": 3.059618031658637e-05, "loss": 0.7781, "step": 225550 }, { "epoch": 38.816242257398486, "grad_norm": 11.324459075927734, "learning_rate": 3.059187887130076e-05, "loss": 0.8949, "step": 225600 }, { "epoch": 38.82484514796972, "grad_norm": 11.03003978729248, "learning_rate": 3.058757742601514e-05, "loss": 0.8352, "step": 225650 }, { "epoch": 38.83344803854095, "grad_norm": 13.6701078414917, "learning_rate": 3.058327598072952e-05, "loss": 0.8362, "step": 225700 }, { "epoch": 38.84205092911218, "grad_norm": 9.975738525390625, "learning_rate": 3.057897453544391e-05, "loss": 0.8821, "step": 225750 }, { "epoch": 38.850653819683416, "grad_norm": 9.564926147460938, "learning_rate": 3.05746730901583e-05, "loss": 0.8513, "step": 225800 }, { "epoch": 38.85925671025465, "grad_norm": 14.574480056762695, "learning_rate": 3.057037164487268e-05, "loss": 0.8414, "step": 225850 }, { "epoch": 38.86785960082588, "grad_norm": 13.458272933959961, "learning_rate": 3.056607019958706e-05, "loss": 0.9311, "step": 225900 }, { "epoch": 38.876462491397106, "grad_norm": 12.326568603515625, "learning_rate": 3.056176875430144e-05, "loss": 0.8436, "step": 225950 }, { "epoch": 38.88506538196834, "grad_norm": 14.677090644836426, "learning_rate": 3.055746730901583e-05, "loss": 0.813, "step": 226000 }, { "epoch": 38.89366827253957, "grad_norm": 13.506479263305664, "learning_rate": 3.055316586373022e-05, "loss": 0.797, "step": 226050 }, { "epoch": 38.902271163110804, "grad_norm": 9.511887550354004, "learning_rate": 3.05488644184446e-05, "loss": 0.862, "step": 226100 }, { "epoch": 38.910874053682036, "grad_norm": 7.246212959289551, "learning_rate": 3.054456297315899e-05, "loss": 0.8311, "step": 226150 }, { "epoch": 38.91947694425327, "grad_norm": 14.512063980102539, "learning_rate": 3.054026152787337e-05, "loss": 0.8655, "step": 226200 }, { "epoch": 38.9280798348245, "grad_norm": 11.661053657531738, "learning_rate": 3.053596008258775e-05, "loss": 0.8684, "step": 226250 }, { "epoch": 38.936682725395734, "grad_norm": 12.401809692382812, "learning_rate": 3.053165863730214e-05, "loss": 0.8214, "step": 226300 }, { "epoch": 38.94528561596697, "grad_norm": 14.105721473693848, "learning_rate": 3.052735719201652e-05, "loss": 0.8318, "step": 226350 }, { "epoch": 38.9538885065382, "grad_norm": 13.418106079101562, "learning_rate": 3.052305574673091e-05, "loss": 0.8313, "step": 226400 }, { "epoch": 38.96249139710943, "grad_norm": 12.303299903869629, "learning_rate": 3.051875430144529e-05, "loss": 0.822, "step": 226450 }, { "epoch": 38.971094287680664, "grad_norm": 8.416863441467285, "learning_rate": 3.051445285615967e-05, "loss": 0.8604, "step": 226500 }, { "epoch": 38.97969717825189, "grad_norm": 13.648849487304688, "learning_rate": 3.0510151410874054e-05, "loss": 0.7971, "step": 226550 }, { "epoch": 38.98830006882312, "grad_norm": 13.000356674194336, "learning_rate": 3.050584996558844e-05, "loss": 0.8154, "step": 226600 }, { "epoch": 38.996902959394355, "grad_norm": 15.074055671691895, "learning_rate": 3.0501548520302826e-05, "loss": 0.8587, "step": 226650 }, { "epoch": 39.0, "eval_accuracy": 0.6412837721562554, "eval_f1": 0.6350699028097229, "eval_f1_DuraRiadoRio_16x16": 0.6022749950109758, "eval_f1_Mole_16x16": 0.6574447646493756, "eval_f1_Quebrado_16x16": 0.7713267658771746, "eval_f1_RiadoRio_16x16": 0.49730893433799783, "eval_f1_RioFechado_16x16": 0.6469940541730896, "eval_loss": 1.2546249628067017, "eval_precision": 0.6523699845374813, "eval_precision_DuraRiadoRio_16x16": 0.5574436645733284, "eval_precision_Mole_16x16": 0.589796621854533, "eval_precision_Quebrado_16x16": 0.7458451560599919, "eval_precision_RiadoRio_16x16": 0.719626168224299, "eval_precision_RioFechado_16x16": 0.6491383119752541, "eval_recall": 0.6441957363594822, "eval_recall_DuraRiadoRio_16x16": 0.6549479166666666, "eval_recall_Mole_16x16": 0.7426215277777778, "eval_recall_Quebrado_16x16": 0.7986111111111112, "eval_recall_RiadoRio_16x16": 0.37993421052631576, "eval_recall_RioFechado_16x16": 0.6448639157155399, "eval_runtime": 45.8442, "eval_samples_per_second": 253.511, "eval_steps_per_second": 15.858, "step": 226668 }, { "epoch": 39.00550584996559, "grad_norm": 9.929780960083008, "learning_rate": 3.0497247075017204e-05, "loss": 0.8156, "step": 226700 }, { "epoch": 39.01410874053682, "grad_norm": 12.01915454864502, "learning_rate": 3.049294562973159e-05, "loss": 0.8374, "step": 226750 }, { "epoch": 39.02271163110805, "grad_norm": 5.850263595581055, "learning_rate": 3.0488644184445976e-05, "loss": 0.8539, "step": 226800 }, { "epoch": 39.031314521679285, "grad_norm": 8.79038143157959, "learning_rate": 3.048434273916036e-05, "loss": 0.8231, "step": 226850 }, { "epoch": 39.03991741225052, "grad_norm": 12.024097442626953, "learning_rate": 3.0480041293874745e-05, "loss": 0.8181, "step": 226900 }, { "epoch": 39.04852030282175, "grad_norm": 11.644914627075195, "learning_rate": 3.047573984858913e-05, "loss": 0.7602, "step": 226950 }, { "epoch": 39.05712319339298, "grad_norm": 14.128507614135742, "learning_rate": 3.0471438403303508e-05, "loss": 0.8444, "step": 227000 }, { "epoch": 39.065726083964215, "grad_norm": 13.259644508361816, "learning_rate": 3.0467136958017896e-05, "loss": 0.8468, "step": 227050 }, { "epoch": 39.07432897453545, "grad_norm": 11.05765151977539, "learning_rate": 3.046283551273228e-05, "loss": 0.801, "step": 227100 }, { "epoch": 39.08293186510667, "grad_norm": 9.016724586486816, "learning_rate": 3.0458534067446665e-05, "loss": 0.8124, "step": 227150 }, { "epoch": 39.091534755677905, "grad_norm": 22.00237274169922, "learning_rate": 3.045423262216105e-05, "loss": 0.8268, "step": 227200 }, { "epoch": 39.10013764624914, "grad_norm": 13.834192276000977, "learning_rate": 3.044993117687543e-05, "loss": 0.8189, "step": 227250 }, { "epoch": 39.10874053682037, "grad_norm": 11.937128067016602, "learning_rate": 3.0445629731589815e-05, "loss": 0.7884, "step": 227300 }, { "epoch": 39.1173434273916, "grad_norm": 9.681209564208984, "learning_rate": 3.04413282863042e-05, "loss": 0.8206, "step": 227350 }, { "epoch": 39.125946317962836, "grad_norm": 10.059327125549316, "learning_rate": 3.0437026841018584e-05, "loss": 0.8256, "step": 227400 }, { "epoch": 39.13454920853407, "grad_norm": 14.95361042022705, "learning_rate": 3.043272539573297e-05, "loss": 0.7916, "step": 227450 }, { "epoch": 39.1431520991053, "grad_norm": 13.537753105163574, "learning_rate": 3.0428423950447353e-05, "loss": 0.8625, "step": 227500 }, { "epoch": 39.15175498967653, "grad_norm": 11.668670654296875, "learning_rate": 3.0424122505161735e-05, "loss": 0.887, "step": 227550 }, { "epoch": 39.160357880247766, "grad_norm": 15.715250968933105, "learning_rate": 3.041982105987612e-05, "loss": 0.8551, "step": 227600 }, { "epoch": 39.168960770819, "grad_norm": 10.937134742736816, "learning_rate": 3.0415519614590504e-05, "loss": 0.8449, "step": 227650 }, { "epoch": 39.177563661390224, "grad_norm": 15.671525001525879, "learning_rate": 3.0411218169304888e-05, "loss": 0.8681, "step": 227700 }, { "epoch": 39.186166551961456, "grad_norm": 13.183513641357422, "learning_rate": 3.0406916724019273e-05, "loss": 0.8166, "step": 227750 }, { "epoch": 39.19476944253269, "grad_norm": 9.698931694030762, "learning_rate": 3.0402615278733654e-05, "loss": 0.8134, "step": 227800 }, { "epoch": 39.20337233310392, "grad_norm": 11.329048156738281, "learning_rate": 3.039831383344804e-05, "loss": 0.8363, "step": 227850 }, { "epoch": 39.211975223675154, "grad_norm": 13.487167358398438, "learning_rate": 3.0394012388162423e-05, "loss": 0.855, "step": 227900 }, { "epoch": 39.220578114246386, "grad_norm": 11.60549545288086, "learning_rate": 3.0389710942876808e-05, "loss": 0.8484, "step": 227950 }, { "epoch": 39.22918100481762, "grad_norm": 9.677910804748535, "learning_rate": 3.0385409497591196e-05, "loss": 0.8489, "step": 228000 }, { "epoch": 39.23778389538885, "grad_norm": 19.07427978515625, "learning_rate": 3.0381108052305573e-05, "loss": 0.8304, "step": 228050 }, { "epoch": 39.246386785960084, "grad_norm": 9.337539672851562, "learning_rate": 3.0376806607019958e-05, "loss": 0.801, "step": 228100 }, { "epoch": 39.254989676531316, "grad_norm": 15.9208402633667, "learning_rate": 3.0372505161734343e-05, "loss": 0.832, "step": 228150 }, { "epoch": 39.26359256710255, "grad_norm": 12.241842269897461, "learning_rate": 3.0368203716448727e-05, "loss": 0.8364, "step": 228200 }, { "epoch": 39.27219545767378, "grad_norm": 15.473211288452148, "learning_rate": 3.0363902271163115e-05, "loss": 0.8291, "step": 228250 }, { "epoch": 39.28079834824501, "grad_norm": 13.98387622833252, "learning_rate": 3.03596008258775e-05, "loss": 0.8088, "step": 228300 }, { "epoch": 39.28940123881624, "grad_norm": 12.87624454498291, "learning_rate": 3.0355299380591877e-05, "loss": 0.8097, "step": 228350 }, { "epoch": 39.29800412938747, "grad_norm": 11.089701652526855, "learning_rate": 3.0350997935306262e-05, "loss": 0.7713, "step": 228400 }, { "epoch": 39.306607019958705, "grad_norm": 9.194135665893555, "learning_rate": 3.034669649002065e-05, "loss": 0.8436, "step": 228450 }, { "epoch": 39.31520991052994, "grad_norm": 9.336625099182129, "learning_rate": 3.0342395044735034e-05, "loss": 0.8138, "step": 228500 }, { "epoch": 39.32381280110117, "grad_norm": 8.389690399169922, "learning_rate": 3.033809359944942e-05, "loss": 0.8044, "step": 228550 }, { "epoch": 39.3324156916724, "grad_norm": 7.4670634269714355, "learning_rate": 3.0333792154163797e-05, "loss": 0.7989, "step": 228600 }, { "epoch": 39.341018582243635, "grad_norm": 11.857172966003418, "learning_rate": 3.0329490708878185e-05, "loss": 0.8549, "step": 228650 }, { "epoch": 39.34962147281487, "grad_norm": 17.310792922973633, "learning_rate": 3.032518926359257e-05, "loss": 0.8534, "step": 228700 }, { "epoch": 39.3582243633861, "grad_norm": 11.100146293640137, "learning_rate": 3.0320887818306954e-05, "loss": 0.8807, "step": 228750 }, { "epoch": 39.36682725395733, "grad_norm": 17.071022033691406, "learning_rate": 3.031658637302134e-05, "loss": 0.839, "step": 228800 }, { "epoch": 39.375430144528565, "grad_norm": 11.228771209716797, "learning_rate": 3.031228492773572e-05, "loss": 0.8435, "step": 228850 }, { "epoch": 39.38403303509979, "grad_norm": 13.199332237243652, "learning_rate": 3.0307983482450104e-05, "loss": 0.7978, "step": 228900 }, { "epoch": 39.39263592567102, "grad_norm": 11.991995811462402, "learning_rate": 3.030368203716449e-05, "loss": 0.8249, "step": 228950 }, { "epoch": 39.401238816242255, "grad_norm": 13.931419372558594, "learning_rate": 3.0299380591878873e-05, "loss": 0.8129, "step": 229000 }, { "epoch": 39.40984170681349, "grad_norm": 18.419677734375, "learning_rate": 3.0295079146593258e-05, "loss": 0.8005, "step": 229050 }, { "epoch": 39.41844459738472, "grad_norm": 18.3383731842041, "learning_rate": 3.0290777701307642e-05, "loss": 0.8374, "step": 229100 }, { "epoch": 39.42704748795595, "grad_norm": 15.237173080444336, "learning_rate": 3.0286476256022024e-05, "loss": 0.8541, "step": 229150 }, { "epoch": 39.435650378527185, "grad_norm": 6.38608455657959, "learning_rate": 3.0282174810736408e-05, "loss": 0.7735, "step": 229200 }, { "epoch": 39.44425326909842, "grad_norm": 13.275871276855469, "learning_rate": 3.0277873365450793e-05, "loss": 0.86, "step": 229250 }, { "epoch": 39.45285615966965, "grad_norm": 10.964957237243652, "learning_rate": 3.0273571920165177e-05, "loss": 0.788, "step": 229300 }, { "epoch": 39.46145905024088, "grad_norm": 12.741104125976562, "learning_rate": 3.0269270474879562e-05, "loss": 0.8194, "step": 229350 }, { "epoch": 39.470061940812116, "grad_norm": 16.114675521850586, "learning_rate": 3.0264969029593943e-05, "loss": 0.7941, "step": 229400 }, { "epoch": 39.47866483138335, "grad_norm": 14.485494613647461, "learning_rate": 3.0260667584308328e-05, "loss": 0.8547, "step": 229450 }, { "epoch": 39.487267721954574, "grad_norm": 12.355427742004395, "learning_rate": 3.0256366139022712e-05, "loss": 0.8511, "step": 229500 }, { "epoch": 39.495870612525806, "grad_norm": 13.766806602478027, "learning_rate": 3.0252064693737097e-05, "loss": 0.8217, "step": 229550 }, { "epoch": 39.50447350309704, "grad_norm": 9.397790908813477, "learning_rate": 3.024776324845148e-05, "loss": 0.8635, "step": 229600 }, { "epoch": 39.51307639366827, "grad_norm": 15.863462448120117, "learning_rate": 3.024346180316587e-05, "loss": 0.8254, "step": 229650 }, { "epoch": 39.521679284239504, "grad_norm": 16.51403045654297, "learning_rate": 3.0239160357880247e-05, "loss": 0.7822, "step": 229700 }, { "epoch": 39.530282174810736, "grad_norm": 12.505497932434082, "learning_rate": 3.023485891259463e-05, "loss": 0.8172, "step": 229750 }, { "epoch": 39.53888506538197, "grad_norm": 8.90392780303955, "learning_rate": 3.0230557467309016e-05, "loss": 0.7767, "step": 229800 }, { "epoch": 39.5474879559532, "grad_norm": 11.128409385681152, "learning_rate": 3.0226256022023404e-05, "loss": 0.8413, "step": 229850 }, { "epoch": 39.556090846524434, "grad_norm": 9.225093841552734, "learning_rate": 3.022195457673779e-05, "loss": 0.8321, "step": 229900 }, { "epoch": 39.564693737095666, "grad_norm": 9.666722297668457, "learning_rate": 3.0217653131452166e-05, "loss": 0.8347, "step": 229950 }, { "epoch": 39.5732966276669, "grad_norm": 17.87047576904297, "learning_rate": 3.021335168616655e-05, "loss": 0.8389, "step": 230000 }, { "epoch": 39.58189951823813, "grad_norm": 12.746573448181152, "learning_rate": 3.020905024088094e-05, "loss": 0.8166, "step": 230050 }, { "epoch": 39.59050240880936, "grad_norm": 15.611976623535156, "learning_rate": 3.0204748795595323e-05, "loss": 0.8415, "step": 230100 }, { "epoch": 39.59910529938059, "grad_norm": 15.364056587219238, "learning_rate": 3.0200447350309708e-05, "loss": 0.867, "step": 230150 }, { "epoch": 39.60770818995182, "grad_norm": 15.823247909545898, "learning_rate": 3.0196145905024086e-05, "loss": 0.8808, "step": 230200 }, { "epoch": 39.616311080523054, "grad_norm": 13.054478645324707, "learning_rate": 3.019184445973847e-05, "loss": 0.8183, "step": 230250 }, { "epoch": 39.62491397109429, "grad_norm": 14.908544540405273, "learning_rate": 3.018754301445286e-05, "loss": 0.8134, "step": 230300 }, { "epoch": 39.63351686166552, "grad_norm": 9.50129222869873, "learning_rate": 3.0183241569167243e-05, "loss": 0.8454, "step": 230350 }, { "epoch": 39.64211975223675, "grad_norm": 12.149151802062988, "learning_rate": 3.0178940123881627e-05, "loss": 0.8518, "step": 230400 }, { "epoch": 39.650722642807985, "grad_norm": 13.119181632995605, "learning_rate": 3.0174638678596012e-05, "loss": 0.8357, "step": 230450 }, { "epoch": 39.65932553337922, "grad_norm": 11.790984153747559, "learning_rate": 3.0170337233310393e-05, "loss": 0.7998, "step": 230500 }, { "epoch": 39.66792842395045, "grad_norm": 11.532367706298828, "learning_rate": 3.0166035788024778e-05, "loss": 0.805, "step": 230550 }, { "epoch": 39.67653131452168, "grad_norm": 16.624267578125, "learning_rate": 3.0161734342739162e-05, "loss": 0.803, "step": 230600 }, { "epoch": 39.68513420509291, "grad_norm": 12.141376495361328, "learning_rate": 3.0157432897453547e-05, "loss": 0.7628, "step": 230650 }, { "epoch": 39.69373709566414, "grad_norm": 7.329588890075684, "learning_rate": 3.015313145216793e-05, "loss": 0.7962, "step": 230700 }, { "epoch": 39.70233998623537, "grad_norm": 15.111682891845703, "learning_rate": 3.0148830006882313e-05, "loss": 0.8315, "step": 230750 }, { "epoch": 39.710942876806605, "grad_norm": 13.07227611541748, "learning_rate": 3.0144528561596697e-05, "loss": 0.8305, "step": 230800 }, { "epoch": 39.71954576737784, "grad_norm": 11.57765007019043, "learning_rate": 3.0140227116311082e-05, "loss": 0.7958, "step": 230850 }, { "epoch": 39.72814865794907, "grad_norm": 8.280070304870605, "learning_rate": 3.0135925671025466e-05, "loss": 0.8577, "step": 230900 }, { "epoch": 39.7367515485203, "grad_norm": 11.898401260375977, "learning_rate": 3.013162422573985e-05, "loss": 0.866, "step": 230950 }, { "epoch": 39.745354439091535, "grad_norm": 19.14908790588379, "learning_rate": 3.0127322780454235e-05, "loss": 0.9041, "step": 231000 }, { "epoch": 39.75395732966277, "grad_norm": 9.723931312561035, "learning_rate": 3.0123021335168617e-05, "loss": 0.7759, "step": 231050 }, { "epoch": 39.762560220234, "grad_norm": 15.632049560546875, "learning_rate": 3.0118719889883e-05, "loss": 0.8651, "step": 231100 }, { "epoch": 39.77116311080523, "grad_norm": 11.398330688476562, "learning_rate": 3.0114418444597386e-05, "loss": 0.8363, "step": 231150 }, { "epoch": 39.779766001376466, "grad_norm": 14.212312698364258, "learning_rate": 3.011011699931177e-05, "loss": 0.8056, "step": 231200 }, { "epoch": 39.78836889194769, "grad_norm": 9.447726249694824, "learning_rate": 3.0105815554026158e-05, "loss": 0.7773, "step": 231250 }, { "epoch": 39.79697178251892, "grad_norm": 11.09247875213623, "learning_rate": 3.0101514108740536e-05, "loss": 0.8314, "step": 231300 }, { "epoch": 39.805574673090156, "grad_norm": 13.929061889648438, "learning_rate": 3.009721266345492e-05, "loss": 0.845, "step": 231350 }, { "epoch": 39.81417756366139, "grad_norm": 17.735431671142578, "learning_rate": 3.0092911218169305e-05, "loss": 0.8323, "step": 231400 }, { "epoch": 39.82278045423262, "grad_norm": 19.31398582458496, "learning_rate": 3.008860977288369e-05, "loss": 0.8275, "step": 231450 }, { "epoch": 39.831383344803854, "grad_norm": 14.280865669250488, "learning_rate": 3.0084308327598078e-05, "loss": 0.8359, "step": 231500 }, { "epoch": 39.839986235375086, "grad_norm": 9.362622261047363, "learning_rate": 3.0080006882312455e-05, "loss": 0.9071, "step": 231550 }, { "epoch": 39.84858912594632, "grad_norm": 12.926568984985352, "learning_rate": 3.007570543702684e-05, "loss": 0.8621, "step": 231600 }, { "epoch": 39.85719201651755, "grad_norm": 14.660911560058594, "learning_rate": 3.0071403991741225e-05, "loss": 0.8164, "step": 231650 }, { "epoch": 39.865794907088784, "grad_norm": 11.949248313903809, "learning_rate": 3.0067102546455612e-05, "loss": 0.8027, "step": 231700 }, { "epoch": 39.874397797660016, "grad_norm": 11.513373374938965, "learning_rate": 3.0062801101169997e-05, "loss": 0.8316, "step": 231750 }, { "epoch": 39.88300068823125, "grad_norm": 12.047130584716797, "learning_rate": 3.005849965588438e-05, "loss": 0.8138, "step": 231800 }, { "epoch": 39.891603578802474, "grad_norm": 14.076888084411621, "learning_rate": 3.005419821059876e-05, "loss": 0.8693, "step": 231850 }, { "epoch": 39.90020646937371, "grad_norm": 23.53899574279785, "learning_rate": 3.0049896765313147e-05, "loss": 0.818, "step": 231900 }, { "epoch": 39.90880935994494, "grad_norm": 10.038372039794922, "learning_rate": 3.0045595320027532e-05, "loss": 0.7804, "step": 231950 }, { "epoch": 39.91741225051617, "grad_norm": 15.943347930908203, "learning_rate": 3.0041293874741916e-05, "loss": 0.8201, "step": 232000 }, { "epoch": 39.926015141087404, "grad_norm": 11.659899711608887, "learning_rate": 3.00369924294563e-05, "loss": 0.8186, "step": 232050 }, { "epoch": 39.93461803165864, "grad_norm": 11.149589538574219, "learning_rate": 3.0032690984170682e-05, "loss": 0.8174, "step": 232100 }, { "epoch": 39.94322092222987, "grad_norm": 13.017484664916992, "learning_rate": 3.0028389538885067e-05, "loss": 0.8617, "step": 232150 }, { "epoch": 39.9518238128011, "grad_norm": 8.724315643310547, "learning_rate": 3.002408809359945e-05, "loss": 0.8748, "step": 232200 }, { "epoch": 39.960426703372335, "grad_norm": 14.261041641235352, "learning_rate": 3.0019786648313836e-05, "loss": 0.8146, "step": 232250 }, { "epoch": 39.96902959394357, "grad_norm": 11.040193557739258, "learning_rate": 3.001548520302822e-05, "loss": 0.7835, "step": 232300 }, { "epoch": 39.9776324845148, "grad_norm": 11.029266357421875, "learning_rate": 3.00111837577426e-05, "loss": 0.862, "step": 232350 }, { "epoch": 39.98623537508603, "grad_norm": 12.291226387023926, "learning_rate": 3.0006882312456986e-05, "loss": 0.8177, "step": 232400 }, { "epoch": 39.99483826565726, "grad_norm": 17.93740463256836, "learning_rate": 3.000258086717137e-05, "loss": 0.8124, "step": 232450 }, { "epoch": 40.0, "eval_accuracy": 0.5579074169678196, "eval_f1": 0.53732775120247, "eval_f1_DuraRiadoRio_16x16": 0.5446857582164136, "eval_f1_Mole_16x16": 0.5789345657990757, "eval_f1_Quebrado_16x16": 0.7253093363329584, "eval_f1_RiadoRio_16x16": 0.2483113541331618, "eval_f1_RioFechado_16x16": 0.5893977415307403, "eval_loss": 1.831228256225586, "eval_precision": 0.5857874696581403, "eval_precision_DuraRiadoRio_16x16": 0.48878923766816146, "eval_precision_Mole_16x16": 0.6585500830105147, "eval_precision_Quebrado_16x16": 0.7529191966370855, "eval_precision_RiadoRio_16x16": 0.5701624815361891, "eval_precision_RioFechado_16x16": 0.45851634943875064, "eval_recall": 0.5629453312321263, "eval_recall_DuraRiadoRio_16x16": 0.6150173611111112, "eval_recall_Mole_16x16": 0.5164930555555556, "eval_recall_Quebrado_16x16": 0.6996527777777778, "eval_recall_RiadoRio_16x16": 0.15871710526315788, "eval_recall_RioFechado_16x16": 0.824846356453029, "eval_runtime": 46.263, "eval_samples_per_second": 251.216, "eval_steps_per_second": 15.715, "step": 232480 }, { "epoch": 40.00344115622849, "grad_norm": 11.612451553344727, "learning_rate": 2.9998279421885755e-05, "loss": 0.804, "step": 232500 }, { "epoch": 40.01204404679972, "grad_norm": 11.678960800170898, "learning_rate": 2.999397797660014e-05, "loss": 0.7779, "step": 232550 }, { "epoch": 40.020646937370955, "grad_norm": 12.962372779846191, "learning_rate": 2.9989676531314524e-05, "loss": 0.8578, "step": 232600 }, { "epoch": 40.02924982794219, "grad_norm": 9.021584510803223, "learning_rate": 2.9985375086028906e-05, "loss": 0.8531, "step": 232650 }, { "epoch": 40.03785271851342, "grad_norm": 13.26291275024414, "learning_rate": 2.998107364074329e-05, "loss": 0.8519, "step": 232700 }, { "epoch": 40.04645560908465, "grad_norm": 9.442376136779785, "learning_rate": 2.9976772195457675e-05, "loss": 0.8384, "step": 232750 }, { "epoch": 40.055058499655885, "grad_norm": 10.713613510131836, "learning_rate": 2.997247075017206e-05, "loss": 0.8071, "step": 232800 }, { "epoch": 40.06366139022712, "grad_norm": 15.669389724731445, "learning_rate": 2.9968169304886444e-05, "loss": 0.8007, "step": 232850 }, { "epoch": 40.07226428079835, "grad_norm": 21.873937606811523, "learning_rate": 2.9963867859600825e-05, "loss": 0.843, "step": 232900 }, { "epoch": 40.08086717136958, "grad_norm": 11.961814880371094, "learning_rate": 2.995956641431521e-05, "loss": 0.8673, "step": 232950 }, { "epoch": 40.089470061940816, "grad_norm": 14.48397159576416, "learning_rate": 2.9955264969029594e-05, "loss": 0.809, "step": 233000 }, { "epoch": 40.09807295251204, "grad_norm": 9.6854829788208, "learning_rate": 2.995096352374398e-05, "loss": 0.7993, "step": 233050 }, { "epoch": 40.10667584308327, "grad_norm": 13.739323616027832, "learning_rate": 2.9946662078458367e-05, "loss": 0.8235, "step": 233100 }, { "epoch": 40.115278733654506, "grad_norm": 12.601007461547852, "learning_rate": 2.994236063317275e-05, "loss": 0.789, "step": 233150 }, { "epoch": 40.12388162422574, "grad_norm": 11.715237617492676, "learning_rate": 2.993805918788713e-05, "loss": 0.8161, "step": 233200 }, { "epoch": 40.13248451479697, "grad_norm": 11.54177474975586, "learning_rate": 2.9933757742601514e-05, "loss": 0.7785, "step": 233250 }, { "epoch": 40.141087405368204, "grad_norm": 11.808496475219727, "learning_rate": 2.99294562973159e-05, "loss": 0.8579, "step": 233300 }, { "epoch": 40.149690295939436, "grad_norm": 10.181523323059082, "learning_rate": 2.9925154852030286e-05, "loss": 0.8257, "step": 233350 }, { "epoch": 40.15829318651067, "grad_norm": 15.260072708129883, "learning_rate": 2.992085340674467e-05, "loss": 0.7782, "step": 233400 }, { "epoch": 40.1668960770819, "grad_norm": 16.15797996520996, "learning_rate": 2.991655196145905e-05, "loss": 0.7998, "step": 233450 }, { "epoch": 40.175498967653134, "grad_norm": 14.563396453857422, "learning_rate": 2.9912250516173433e-05, "loss": 0.8528, "step": 233500 }, { "epoch": 40.184101858224366, "grad_norm": 11.413307189941406, "learning_rate": 2.990794907088782e-05, "loss": 0.7878, "step": 233550 }, { "epoch": 40.1927047487956, "grad_norm": 11.24429988861084, "learning_rate": 2.9903647625602205e-05, "loss": 0.8312, "step": 233600 }, { "epoch": 40.201307639366824, "grad_norm": 11.949051856994629, "learning_rate": 2.989934618031659e-05, "loss": 0.8318, "step": 233650 }, { "epoch": 40.20991052993806, "grad_norm": 5.215050220489502, "learning_rate": 2.9895044735030968e-05, "loss": 0.8211, "step": 233700 }, { "epoch": 40.21851342050929, "grad_norm": 10.046882629394531, "learning_rate": 2.9890743289745356e-05, "loss": 0.8153, "step": 233750 }, { "epoch": 40.22711631108052, "grad_norm": 14.303393363952637, "learning_rate": 2.988644184445974e-05, "loss": 0.8021, "step": 233800 }, { "epoch": 40.235719201651754, "grad_norm": 17.105989456176758, "learning_rate": 2.9882140399174125e-05, "loss": 0.7548, "step": 233850 }, { "epoch": 40.24432209222299, "grad_norm": 12.5493803024292, "learning_rate": 2.987783895388851e-05, "loss": 0.7983, "step": 233900 }, { "epoch": 40.25292498279422, "grad_norm": 11.594947814941406, "learning_rate": 2.9873537508602894e-05, "loss": 0.8464, "step": 233950 }, { "epoch": 40.26152787336545, "grad_norm": 22.019929885864258, "learning_rate": 2.9869236063317275e-05, "loss": 0.9222, "step": 234000 }, { "epoch": 40.270130763936685, "grad_norm": 16.151779174804688, "learning_rate": 2.986493461803166e-05, "loss": 0.8311, "step": 234050 }, { "epoch": 40.27873365450792, "grad_norm": 9.717228889465332, "learning_rate": 2.9860633172746044e-05, "loss": 0.8725, "step": 234100 }, { "epoch": 40.28733654507915, "grad_norm": 13.724078178405762, "learning_rate": 2.985633172746043e-05, "loss": 0.7823, "step": 234150 }, { "epoch": 40.295939435650375, "grad_norm": 10.474478721618652, "learning_rate": 2.9852030282174813e-05, "loss": 0.8255, "step": 234200 }, { "epoch": 40.30454232622161, "grad_norm": 8.32099723815918, "learning_rate": 2.9847728836889195e-05, "loss": 0.8208, "step": 234250 }, { "epoch": 40.31314521679284, "grad_norm": 10.488977432250977, "learning_rate": 2.984342739160358e-05, "loss": 0.805, "step": 234300 }, { "epoch": 40.32174810736407, "grad_norm": 9.203238487243652, "learning_rate": 2.9839125946317964e-05, "loss": 0.8065, "step": 234350 }, { "epoch": 40.330350997935305, "grad_norm": 14.118271827697754, "learning_rate": 2.9834824501032348e-05, "loss": 0.8275, "step": 234400 }, { "epoch": 40.33895388850654, "grad_norm": 14.719852447509766, "learning_rate": 2.9830523055746733e-05, "loss": 0.8337, "step": 234450 }, { "epoch": 40.34755677907777, "grad_norm": 5.246501445770264, "learning_rate": 2.9826221610461114e-05, "loss": 0.8016, "step": 234500 }, { "epoch": 40.356159669649, "grad_norm": 12.348990440368652, "learning_rate": 2.98219201651755e-05, "loss": 0.7856, "step": 234550 }, { "epoch": 40.364762560220235, "grad_norm": 16.188379287719727, "learning_rate": 2.9817618719889883e-05, "loss": 0.7985, "step": 234600 }, { "epoch": 40.37336545079147, "grad_norm": 13.460387229919434, "learning_rate": 2.9813317274604268e-05, "loss": 0.8768, "step": 234650 }, { "epoch": 40.3819683413627, "grad_norm": 10.690510749816895, "learning_rate": 2.9809015829318652e-05, "loss": 0.8393, "step": 234700 }, { "epoch": 40.39057123193393, "grad_norm": 11.652181625366211, "learning_rate": 2.980471438403304e-05, "loss": 0.8113, "step": 234750 }, { "epoch": 40.39917412250516, "grad_norm": 8.865718841552734, "learning_rate": 2.9800412938747418e-05, "loss": 0.8049, "step": 234800 }, { "epoch": 40.40777701307639, "grad_norm": 7.705862522125244, "learning_rate": 2.9796111493461802e-05, "loss": 0.8301, "step": 234850 }, { "epoch": 40.41637990364762, "grad_norm": 14.75493335723877, "learning_rate": 2.9791810048176187e-05, "loss": 0.8371, "step": 234900 }, { "epoch": 40.424982794218856, "grad_norm": 8.159524917602539, "learning_rate": 2.9787508602890575e-05, "loss": 0.8314, "step": 234950 }, { "epoch": 40.43358568479009, "grad_norm": 13.01003360748291, "learning_rate": 2.978320715760496e-05, "loss": 0.8486, "step": 235000 }, { "epoch": 40.44218857536132, "grad_norm": 16.232181549072266, "learning_rate": 2.9778905712319337e-05, "loss": 0.8665, "step": 235050 }, { "epoch": 40.450791465932554, "grad_norm": 16.303571701049805, "learning_rate": 2.9774604267033722e-05, "loss": 0.8265, "step": 235100 }, { "epoch": 40.459394356503786, "grad_norm": 8.593423843383789, "learning_rate": 2.977030282174811e-05, "loss": 0.8443, "step": 235150 }, { "epoch": 40.46799724707502, "grad_norm": 13.479256629943848, "learning_rate": 2.9766001376462494e-05, "loss": 0.8497, "step": 235200 }, { "epoch": 40.47660013764625, "grad_norm": 4.742778301239014, "learning_rate": 2.976169993117688e-05, "loss": 0.8364, "step": 235250 }, { "epoch": 40.485203028217484, "grad_norm": 7.966705799102783, "learning_rate": 2.9757398485891264e-05, "loss": 0.8449, "step": 235300 }, { "epoch": 40.493805918788716, "grad_norm": 13.308069229125977, "learning_rate": 2.9753097040605645e-05, "loss": 0.8181, "step": 235350 }, { "epoch": 40.50240880935994, "grad_norm": 17.81526756286621, "learning_rate": 2.974879559532003e-05, "loss": 0.8267, "step": 235400 }, { "epoch": 40.511011699931174, "grad_norm": 10.182741165161133, "learning_rate": 2.9744494150034414e-05, "loss": 0.8101, "step": 235450 }, { "epoch": 40.51961459050241, "grad_norm": 12.420433044433594, "learning_rate": 2.97401927047488e-05, "loss": 0.8319, "step": 235500 }, { "epoch": 40.52821748107364, "grad_norm": 7.353178024291992, "learning_rate": 2.9735891259463183e-05, "loss": 0.8372, "step": 235550 }, { "epoch": 40.53682037164487, "grad_norm": 11.24759292602539, "learning_rate": 2.9731589814177564e-05, "loss": 0.7873, "step": 235600 }, { "epoch": 40.545423262216104, "grad_norm": 12.406911849975586, "learning_rate": 2.972728836889195e-05, "loss": 0.7932, "step": 235650 }, { "epoch": 40.55402615278734, "grad_norm": 12.272226333618164, "learning_rate": 2.9722986923606333e-05, "loss": 0.8808, "step": 235700 }, { "epoch": 40.56262904335857, "grad_norm": 13.828827857971191, "learning_rate": 2.9718685478320718e-05, "loss": 0.8131, "step": 235750 }, { "epoch": 40.5712319339298, "grad_norm": 12.42386245727539, "learning_rate": 2.9714384033035102e-05, "loss": 0.8109, "step": 235800 }, { "epoch": 40.579834824501035, "grad_norm": 9.907657623291016, "learning_rate": 2.9710082587749484e-05, "loss": 0.822, "step": 235850 }, { "epoch": 40.58843771507227, "grad_norm": 10.260066986083984, "learning_rate": 2.9705781142463868e-05, "loss": 0.8882, "step": 235900 }, { "epoch": 40.5970406056435, "grad_norm": 14.395687103271484, "learning_rate": 2.9701479697178253e-05, "loss": 0.8334, "step": 235950 }, { "epoch": 40.605643496214725, "grad_norm": 11.287680625915527, "learning_rate": 2.9697178251892637e-05, "loss": 0.831, "step": 236000 }, { "epoch": 40.61424638678596, "grad_norm": 19.701868057250977, "learning_rate": 2.9692876806607022e-05, "loss": 0.852, "step": 236050 }, { "epoch": 40.62284927735719, "grad_norm": 12.41551685333252, "learning_rate": 2.9688575361321406e-05, "loss": 0.7862, "step": 236100 }, { "epoch": 40.63145216792842, "grad_norm": 9.521626472473145, "learning_rate": 2.9684273916035788e-05, "loss": 0.873, "step": 236150 }, { "epoch": 40.640055058499655, "grad_norm": 8.292929649353027, "learning_rate": 2.9679972470750172e-05, "loss": 0.8173, "step": 236200 }, { "epoch": 40.64865794907089, "grad_norm": 10.113822937011719, "learning_rate": 2.9675671025464557e-05, "loss": 0.8431, "step": 236250 }, { "epoch": 40.65726083964212, "grad_norm": 11.885089874267578, "learning_rate": 2.967136958017894e-05, "loss": 0.8296, "step": 236300 }, { "epoch": 40.66586373021335, "grad_norm": 14.85073184967041, "learning_rate": 2.966706813489333e-05, "loss": 0.823, "step": 236350 }, { "epoch": 40.674466620784585, "grad_norm": 9.497675895690918, "learning_rate": 2.9662766689607707e-05, "loss": 0.811, "step": 236400 }, { "epoch": 40.68306951135582, "grad_norm": 7.655157566070557, "learning_rate": 2.965846524432209e-05, "loss": 0.8205, "step": 236450 }, { "epoch": 40.69167240192705, "grad_norm": 11.342281341552734, "learning_rate": 2.9654163799036476e-05, "loss": 0.8235, "step": 236500 }, { "epoch": 40.70027529249828, "grad_norm": 9.343525886535645, "learning_rate": 2.9649862353750864e-05, "loss": 0.8045, "step": 236550 }, { "epoch": 40.70887818306951, "grad_norm": 10.503178596496582, "learning_rate": 2.964556090846525e-05, "loss": 0.7949, "step": 236600 }, { "epoch": 40.71748107364074, "grad_norm": 8.300965309143066, "learning_rate": 2.9641259463179633e-05, "loss": 0.8356, "step": 236650 }, { "epoch": 40.72608396421197, "grad_norm": 16.707019805908203, "learning_rate": 2.963695801789401e-05, "loss": 0.8245, "step": 236700 }, { "epoch": 40.734686854783206, "grad_norm": 9.520484924316406, "learning_rate": 2.9632656572608395e-05, "loss": 0.8649, "step": 236750 }, { "epoch": 40.74328974535444, "grad_norm": 16.446346282958984, "learning_rate": 2.9628355127322783e-05, "loss": 0.8283, "step": 236800 }, { "epoch": 40.75189263592567, "grad_norm": 6.520540237426758, "learning_rate": 2.9624053682037168e-05, "loss": 0.8133, "step": 236850 }, { "epoch": 40.7604955264969, "grad_norm": 9.151076316833496, "learning_rate": 2.9619752236751553e-05, "loss": 0.8299, "step": 236900 }, { "epoch": 40.769098417068136, "grad_norm": 9.314895629882812, "learning_rate": 2.961545079146593e-05, "loss": 0.8461, "step": 236950 }, { "epoch": 40.77770130763937, "grad_norm": 8.635174751281738, "learning_rate": 2.9611149346180318e-05, "loss": 0.8018, "step": 237000 }, { "epoch": 40.7863041982106, "grad_norm": 15.60713005065918, "learning_rate": 2.9606847900894703e-05, "loss": 0.8568, "step": 237050 }, { "epoch": 40.794907088781834, "grad_norm": 8.770522117614746, "learning_rate": 2.9602546455609087e-05, "loss": 0.8432, "step": 237100 }, { "epoch": 40.803509979353066, "grad_norm": 14.140973091125488, "learning_rate": 2.9598245010323472e-05, "loss": 0.8196, "step": 237150 }, { "epoch": 40.81211286992429, "grad_norm": 11.614298820495605, "learning_rate": 2.9593943565037853e-05, "loss": 0.8273, "step": 237200 }, { "epoch": 40.820715760495524, "grad_norm": 13.511590003967285, "learning_rate": 2.9589642119752238e-05, "loss": 0.8245, "step": 237250 }, { "epoch": 40.82931865106676, "grad_norm": 10.10208797454834, "learning_rate": 2.9585340674466622e-05, "loss": 0.8358, "step": 237300 }, { "epoch": 40.83792154163799, "grad_norm": 7.1326904296875, "learning_rate": 2.9581039229181007e-05, "loss": 0.7807, "step": 237350 }, { "epoch": 40.84652443220922, "grad_norm": 10.960712432861328, "learning_rate": 2.957673778389539e-05, "loss": 0.8547, "step": 237400 }, { "epoch": 40.855127322780454, "grad_norm": 8.978744506835938, "learning_rate": 2.9572436338609776e-05, "loss": 0.7867, "step": 237450 }, { "epoch": 40.86373021335169, "grad_norm": 14.500258445739746, "learning_rate": 2.9568134893324157e-05, "loss": 0.8269, "step": 237500 }, { "epoch": 40.87233310392292, "grad_norm": 14.024763107299805, "learning_rate": 2.956383344803854e-05, "loss": 0.8716, "step": 237550 }, { "epoch": 40.88093599449415, "grad_norm": 11.733505249023438, "learning_rate": 2.9559532002752926e-05, "loss": 0.826, "step": 237600 }, { "epoch": 40.889538885065384, "grad_norm": 10.890397071838379, "learning_rate": 2.955523055746731e-05, "loss": 0.8088, "step": 237650 }, { "epoch": 40.89814177563662, "grad_norm": 16.12793731689453, "learning_rate": 2.9550929112181695e-05, "loss": 0.8596, "step": 237700 }, { "epoch": 40.90674466620784, "grad_norm": 15.321951866149902, "learning_rate": 2.9546627666896077e-05, "loss": 0.8011, "step": 237750 }, { "epoch": 40.915347556779075, "grad_norm": 13.486268997192383, "learning_rate": 2.954232622161046e-05, "loss": 0.837, "step": 237800 }, { "epoch": 40.92395044735031, "grad_norm": 11.580890655517578, "learning_rate": 2.9538024776324846e-05, "loss": 0.8017, "step": 237850 }, { "epoch": 40.93255333792154, "grad_norm": 11.597046852111816, "learning_rate": 2.953372333103923e-05, "loss": 0.8329, "step": 237900 }, { "epoch": 40.94115622849277, "grad_norm": 6.622074127197266, "learning_rate": 2.9529421885753618e-05, "loss": 0.8462, "step": 237950 }, { "epoch": 40.949759119064005, "grad_norm": 14.926924705505371, "learning_rate": 2.9525120440467996e-05, "loss": 0.835, "step": 238000 }, { "epoch": 40.95836200963524, "grad_norm": 12.909444808959961, "learning_rate": 2.952081899518238e-05, "loss": 0.7964, "step": 238050 }, { "epoch": 40.96696490020647, "grad_norm": 8.795742988586426, "learning_rate": 2.9516517549896765e-05, "loss": 0.8323, "step": 238100 }, { "epoch": 40.9755677907777, "grad_norm": 14.687792778015137, "learning_rate": 2.951221610461115e-05, "loss": 0.7856, "step": 238150 }, { "epoch": 40.984170681348935, "grad_norm": 10.331775665283203, "learning_rate": 2.9507914659325538e-05, "loss": 0.8911, "step": 238200 }, { "epoch": 40.99277357192017, "grad_norm": 12.304204940795898, "learning_rate": 2.9503613214039922e-05, "loss": 0.8602, "step": 238250 }, { "epoch": 41.0, "eval_accuracy": 0.5134228187919463, "eval_f1": 0.4999978433024895, "eval_f1_DuraRiadoRio_16x16": 0.4720399054870045, "eval_f1_Mole_16x16": 0.3342140026420079, "eval_f1_Quebrado_16x16": 0.6623826000988631, "eval_f1_RiadoRio_16x16": 0.4896694214876033, "eval_f1_RioFechado_16x16": 0.5416832867969685, "eval_loss": 1.9507423639297485, "eval_precision": 0.5892191296259293, "eval_precision_DuraRiadoRio_16x16": 0.59734219269103, "eval_precision_Mole_16x16": 0.6988950276243094, "eval_precision_Quebrado_16x16": 0.7692307692307693, "eval_precision_RiadoRio_16x16": 0.49210963455149503, "eval_precision_RioFechado_16x16": 0.3885180240320427, "eval_recall": 0.5145729965690126, "eval_recall_DuraRiadoRio_16x16": 0.3901909722222222, "eval_recall_Mole_16x16": 0.21961805555555555, "eval_recall_Quebrado_16x16": 0.5815972222222222, "eval_recall_RiadoRio_16x16": 0.48725328947368424, "eval_recall_RioFechado_16x16": 0.8942054433713784, "eval_runtime": 46.3744, "eval_samples_per_second": 250.612, "eval_steps_per_second": 15.677, "step": 238292 }, { "epoch": 41.0013764624914, "grad_norm": 9.256872177124023, "learning_rate": 2.94993117687543e-05, "loss": 0.838, "step": 238300 }, { "epoch": 41.009979353062626, "grad_norm": 12.084928512573242, "learning_rate": 2.9495010323468684e-05, "loss": 0.7662, "step": 238350 }, { "epoch": 41.01858224363386, "grad_norm": 12.367027282714844, "learning_rate": 2.9490708878183072e-05, "loss": 0.8139, "step": 238400 }, { "epoch": 41.02718513420509, "grad_norm": 7.044132709503174, "learning_rate": 2.9486407432897457e-05, "loss": 0.8088, "step": 238450 }, { "epoch": 41.03578802477632, "grad_norm": 10.766160011291504, "learning_rate": 2.948210598761184e-05, "loss": 0.8308, "step": 238500 }, { "epoch": 41.044390915347556, "grad_norm": 8.357415199279785, "learning_rate": 2.947780454232622e-05, "loss": 0.7822, "step": 238550 }, { "epoch": 41.05299380591879, "grad_norm": 12.480719566345215, "learning_rate": 2.9473503097040607e-05, "loss": 0.8356, "step": 238600 }, { "epoch": 41.06159669649002, "grad_norm": 20.236997604370117, "learning_rate": 2.9469201651754992e-05, "loss": 0.848, "step": 238650 }, { "epoch": 41.07019958706125, "grad_norm": 13.827505111694336, "learning_rate": 2.9464900206469376e-05, "loss": 0.7928, "step": 238700 }, { "epoch": 41.078802477632486, "grad_norm": 11.439004898071289, "learning_rate": 2.946059876118376e-05, "loss": 0.7548, "step": 238750 }, { "epoch": 41.08740536820372, "grad_norm": 14.980766296386719, "learning_rate": 2.9456297315898145e-05, "loss": 0.8151, "step": 238800 }, { "epoch": 41.09600825877495, "grad_norm": 18.738901138305664, "learning_rate": 2.9451995870612527e-05, "loss": 0.8446, "step": 238850 }, { "epoch": 41.104611149346184, "grad_norm": 8.0914945602417, "learning_rate": 2.944769442532691e-05, "loss": 0.7685, "step": 238900 }, { "epoch": 41.11321403991741, "grad_norm": 11.798049926757812, "learning_rate": 2.9443392980041296e-05, "loss": 0.8201, "step": 238950 }, { "epoch": 41.12181693048864, "grad_norm": 11.076517105102539, "learning_rate": 2.943909153475568e-05, "loss": 0.804, "step": 239000 }, { "epoch": 41.130419821059874, "grad_norm": 16.61061668395996, "learning_rate": 2.9434790089470065e-05, "loss": 0.8038, "step": 239050 }, { "epoch": 41.13902271163111, "grad_norm": 16.12880516052246, "learning_rate": 2.9430488644184446e-05, "loss": 0.8112, "step": 239100 }, { "epoch": 41.14762560220234, "grad_norm": 12.771553993225098, "learning_rate": 2.942618719889883e-05, "loss": 0.8135, "step": 239150 }, { "epoch": 41.15622849277357, "grad_norm": 12.3074951171875, "learning_rate": 2.9421885753613215e-05, "loss": 0.8121, "step": 239200 }, { "epoch": 41.164831383344804, "grad_norm": 9.71275806427002, "learning_rate": 2.94175843083276e-05, "loss": 0.863, "step": 239250 }, { "epoch": 41.17343427391604, "grad_norm": 14.673721313476562, "learning_rate": 2.9413282863041984e-05, "loss": 0.8345, "step": 239300 }, { "epoch": 41.18203716448727, "grad_norm": 18.194034576416016, "learning_rate": 2.9408981417756365e-05, "loss": 0.8423, "step": 239350 }, { "epoch": 41.1906400550585, "grad_norm": 14.847515106201172, "learning_rate": 2.940467997247075e-05, "loss": 0.837, "step": 239400 }, { "epoch": 41.199242945629734, "grad_norm": 13.294322967529297, "learning_rate": 2.9400378527185135e-05, "loss": 0.8349, "step": 239450 }, { "epoch": 41.20784583620097, "grad_norm": 9.280374526977539, "learning_rate": 2.939607708189952e-05, "loss": 0.8156, "step": 239500 }, { "epoch": 41.21644872677219, "grad_norm": 19.699445724487305, "learning_rate": 2.9391775636613904e-05, "loss": 0.8357, "step": 239550 }, { "epoch": 41.225051617343425, "grad_norm": 9.11159610748291, "learning_rate": 2.938747419132829e-05, "loss": 0.811, "step": 239600 }, { "epoch": 41.23365450791466, "grad_norm": 13.324408531188965, "learning_rate": 2.938317274604267e-05, "loss": 0.8157, "step": 239650 }, { "epoch": 41.24225739848589, "grad_norm": 12.780108451843262, "learning_rate": 2.9378871300757054e-05, "loss": 0.9055, "step": 239700 }, { "epoch": 41.25086028905712, "grad_norm": 8.5637788772583, "learning_rate": 2.937456985547144e-05, "loss": 0.8131, "step": 239750 }, { "epoch": 41.259463179628355, "grad_norm": 10.821245193481445, "learning_rate": 2.9370268410185827e-05, "loss": 0.7718, "step": 239800 }, { "epoch": 41.26806607019959, "grad_norm": 14.756125450134277, "learning_rate": 2.936596696490021e-05, "loss": 0.869, "step": 239850 }, { "epoch": 41.27666896077082, "grad_norm": 15.090322494506836, "learning_rate": 2.936166551961459e-05, "loss": 0.8354, "step": 239900 }, { "epoch": 41.28527185134205, "grad_norm": 11.784309387207031, "learning_rate": 2.9357364074328973e-05, "loss": 0.8401, "step": 239950 }, { "epoch": 41.293874741913285, "grad_norm": 10.645408630371094, "learning_rate": 2.9353062629043358e-05, "loss": 0.8067, "step": 240000 }, { "epoch": 41.30247763248452, "grad_norm": 13.073332786560059, "learning_rate": 2.9348761183757746e-05, "loss": 0.8234, "step": 240050 }, { "epoch": 41.31108052305575, "grad_norm": 15.906852722167969, "learning_rate": 2.934445973847213e-05, "loss": 0.8153, "step": 240100 }, { "epoch": 41.319683413626976, "grad_norm": 10.712103843688965, "learning_rate": 2.9340158293186508e-05, "loss": 0.7734, "step": 240150 }, { "epoch": 41.32828630419821, "grad_norm": 15.567662239074707, "learning_rate": 2.9335856847900893e-05, "loss": 0.8413, "step": 240200 }, { "epoch": 41.33688919476944, "grad_norm": 15.18772029876709, "learning_rate": 2.933155540261528e-05, "loss": 0.8684, "step": 240250 }, { "epoch": 41.34549208534067, "grad_norm": 12.103958129882812, "learning_rate": 2.9327253957329665e-05, "loss": 0.8206, "step": 240300 }, { "epoch": 41.354094975911906, "grad_norm": 13.086195945739746, "learning_rate": 2.932295251204405e-05, "loss": 0.8402, "step": 240350 }, { "epoch": 41.36269786648314, "grad_norm": 11.10002326965332, "learning_rate": 2.9318651066758434e-05, "loss": 0.8584, "step": 240400 }, { "epoch": 41.37130075705437, "grad_norm": 12.874799728393555, "learning_rate": 2.9314349621472816e-05, "loss": 0.8369, "step": 240450 }, { "epoch": 41.3799036476256, "grad_norm": 15.304957389831543, "learning_rate": 2.93100481761872e-05, "loss": 0.78, "step": 240500 }, { "epoch": 41.388506538196836, "grad_norm": 13.65037727355957, "learning_rate": 2.9305746730901585e-05, "loss": 0.7778, "step": 240550 }, { "epoch": 41.39710942876807, "grad_norm": 11.144920349121094, "learning_rate": 2.930144528561597e-05, "loss": 0.8709, "step": 240600 }, { "epoch": 41.4057123193393, "grad_norm": 5.8132710456848145, "learning_rate": 2.9297143840330354e-05, "loss": 0.8659, "step": 240650 }, { "epoch": 41.41431520991053, "grad_norm": 15.361316680908203, "learning_rate": 2.9292842395044735e-05, "loss": 0.8211, "step": 240700 }, { "epoch": 41.42291810048176, "grad_norm": 10.807408332824707, "learning_rate": 2.928854094975912e-05, "loss": 0.8149, "step": 240750 }, { "epoch": 41.43152099105299, "grad_norm": 9.568943977355957, "learning_rate": 2.9284239504473504e-05, "loss": 0.8395, "step": 240800 }, { "epoch": 41.440123881624224, "grad_norm": 11.706027030944824, "learning_rate": 2.927993805918789e-05, "loss": 0.7854, "step": 240850 }, { "epoch": 41.44872677219546, "grad_norm": 10.087101936340332, "learning_rate": 2.9275636613902273e-05, "loss": 0.8846, "step": 240900 }, { "epoch": 41.45732966276669, "grad_norm": 12.44240951538086, "learning_rate": 2.9271335168616658e-05, "loss": 0.8653, "step": 240950 }, { "epoch": 41.46593255333792, "grad_norm": 8.573914527893066, "learning_rate": 2.926703372333104e-05, "loss": 0.752, "step": 241000 }, { "epoch": 41.474535443909154, "grad_norm": 12.496503829956055, "learning_rate": 2.9262732278045424e-05, "loss": 0.8791, "step": 241050 }, { "epoch": 41.48313833448039, "grad_norm": 9.883359909057617, "learning_rate": 2.9258430832759808e-05, "loss": 0.841, "step": 241100 }, { "epoch": 41.49174122505162, "grad_norm": 9.416950225830078, "learning_rate": 2.9254129387474193e-05, "loss": 0.825, "step": 241150 }, { "epoch": 41.50034411562285, "grad_norm": 11.282757759094238, "learning_rate": 2.924982794218858e-05, "loss": 0.7918, "step": 241200 }, { "epoch": 41.508947006194084, "grad_norm": 12.020237922668457, "learning_rate": 2.924552649690296e-05, "loss": 0.8348, "step": 241250 }, { "epoch": 41.51754989676531, "grad_norm": 14.400588035583496, "learning_rate": 2.9241225051617343e-05, "loss": 0.841, "step": 241300 }, { "epoch": 41.52615278733654, "grad_norm": 9.644591331481934, "learning_rate": 2.9236923606331728e-05, "loss": 0.8602, "step": 241350 }, { "epoch": 41.534755677907775, "grad_norm": 11.778895378112793, "learning_rate": 2.9232622161046112e-05, "loss": 0.818, "step": 241400 }, { "epoch": 41.54335856847901, "grad_norm": 12.422785758972168, "learning_rate": 2.92283207157605e-05, "loss": 0.8362, "step": 241450 }, { "epoch": 41.55196145905024, "grad_norm": 18.059837341308594, "learning_rate": 2.9224019270474878e-05, "loss": 0.8451, "step": 241500 }, { "epoch": 41.56056434962147, "grad_norm": 17.335363388061523, "learning_rate": 2.9219717825189262e-05, "loss": 0.869, "step": 241550 }, { "epoch": 41.569167240192705, "grad_norm": 16.02141761779785, "learning_rate": 2.9215416379903647e-05, "loss": 0.8431, "step": 241600 }, { "epoch": 41.57777013076394, "grad_norm": 12.644451141357422, "learning_rate": 2.9211114934618035e-05, "loss": 0.7942, "step": 241650 }, { "epoch": 41.58637302133517, "grad_norm": 19.65835189819336, "learning_rate": 2.920681348933242e-05, "loss": 0.825, "step": 241700 }, { "epoch": 41.5949759119064, "grad_norm": 12.673420906066895, "learning_rate": 2.9202512044046804e-05, "loss": 0.86, "step": 241750 }, { "epoch": 41.603578802477635, "grad_norm": 11.896744728088379, "learning_rate": 2.9198210598761182e-05, "loss": 0.7685, "step": 241800 }, { "epoch": 41.61218169304887, "grad_norm": 15.436060905456543, "learning_rate": 2.919390915347557e-05, "loss": 0.8609, "step": 241850 }, { "epoch": 41.62078458362009, "grad_norm": 9.031325340270996, "learning_rate": 2.9189607708189954e-05, "loss": 0.8189, "step": 241900 }, { "epoch": 41.629387474191326, "grad_norm": 9.990264892578125, "learning_rate": 2.918530626290434e-05, "loss": 0.7836, "step": 241950 }, { "epoch": 41.63799036476256, "grad_norm": 10.635930061340332, "learning_rate": 2.9181004817618723e-05, "loss": 0.8495, "step": 242000 }, { "epoch": 41.64659325533379, "grad_norm": 10.97263240814209, "learning_rate": 2.91767033723331e-05, "loss": 0.8, "step": 242050 }, { "epoch": 41.65519614590502, "grad_norm": 8.615398406982422, "learning_rate": 2.917240192704749e-05, "loss": 0.8442, "step": 242100 }, { "epoch": 41.663799036476256, "grad_norm": 7.26699686050415, "learning_rate": 2.9168100481761874e-05, "loss": 0.8399, "step": 242150 }, { "epoch": 41.67240192704749, "grad_norm": 12.199705123901367, "learning_rate": 2.916379903647626e-05, "loss": 0.7824, "step": 242200 }, { "epoch": 41.68100481761872, "grad_norm": 13.320418357849121, "learning_rate": 2.9159497591190643e-05, "loss": 0.8806, "step": 242250 }, { "epoch": 41.68960770818995, "grad_norm": 15.591451644897461, "learning_rate": 2.9155196145905027e-05, "loss": 0.8182, "step": 242300 }, { "epoch": 41.698210598761186, "grad_norm": 14.86335277557373, "learning_rate": 2.915089470061941e-05, "loss": 0.821, "step": 242350 }, { "epoch": 41.70681348933242, "grad_norm": 12.388840675354004, "learning_rate": 2.9146593255333793e-05, "loss": 0.8575, "step": 242400 }, { "epoch": 41.71541637990365, "grad_norm": 15.908899307250977, "learning_rate": 2.9142291810048178e-05, "loss": 0.8608, "step": 242450 }, { "epoch": 41.72401927047488, "grad_norm": 15.799535751342773, "learning_rate": 2.9137990364762562e-05, "loss": 0.8729, "step": 242500 }, { "epoch": 41.73262216104611, "grad_norm": 10.699694633483887, "learning_rate": 2.9133688919476947e-05, "loss": 0.832, "step": 242550 }, { "epoch": 41.74122505161734, "grad_norm": 11.459444046020508, "learning_rate": 2.9129387474191328e-05, "loss": 0.7942, "step": 242600 }, { "epoch": 41.749827942188574, "grad_norm": 13.654376983642578, "learning_rate": 2.9125086028905713e-05, "loss": 0.8014, "step": 242650 }, { "epoch": 41.75843083275981, "grad_norm": 12.79162883758545, "learning_rate": 2.9120784583620097e-05, "loss": 0.8216, "step": 242700 }, { "epoch": 41.76703372333104, "grad_norm": 9.114250183105469, "learning_rate": 2.9116483138334482e-05, "loss": 0.794, "step": 242750 }, { "epoch": 41.77563661390227, "grad_norm": 11.039403915405273, "learning_rate": 2.9112181693048866e-05, "loss": 0.791, "step": 242800 }, { "epoch": 41.784239504473504, "grad_norm": 6.821073532104492, "learning_rate": 2.9107880247763247e-05, "loss": 0.8327, "step": 242850 }, { "epoch": 41.79284239504474, "grad_norm": 14.500955581665039, "learning_rate": 2.9103578802477632e-05, "loss": 0.7868, "step": 242900 }, { "epoch": 41.80144528561597, "grad_norm": 9.586188316345215, "learning_rate": 2.9099277357192017e-05, "loss": 0.8542, "step": 242950 }, { "epoch": 41.8100481761872, "grad_norm": 10.385194778442383, "learning_rate": 2.90949759119064e-05, "loss": 0.799, "step": 243000 }, { "epoch": 41.818651066758434, "grad_norm": 16.173053741455078, "learning_rate": 2.909067446662079e-05, "loss": 0.8421, "step": 243050 }, { "epoch": 41.82725395732966, "grad_norm": 15.409577369689941, "learning_rate": 2.9086373021335174e-05, "loss": 0.8446, "step": 243100 }, { "epoch": 41.83585684790089, "grad_norm": 24.829578399658203, "learning_rate": 2.908207157604955e-05, "loss": 0.8031, "step": 243150 }, { "epoch": 41.844459738472125, "grad_norm": 17.127038955688477, "learning_rate": 2.9077770130763936e-05, "loss": 0.8453, "step": 243200 }, { "epoch": 41.85306262904336, "grad_norm": 12.952582359313965, "learning_rate": 2.9073468685478324e-05, "loss": 0.828, "step": 243250 }, { "epoch": 41.86166551961459, "grad_norm": 10.63305377960205, "learning_rate": 2.906916724019271e-05, "loss": 0.8094, "step": 243300 }, { "epoch": 41.87026841018582, "grad_norm": 11.149521827697754, "learning_rate": 2.9064865794907093e-05, "loss": 0.7846, "step": 243350 }, { "epoch": 41.878871300757055, "grad_norm": 10.605463027954102, "learning_rate": 2.906056434962147e-05, "loss": 0.8457, "step": 243400 }, { "epoch": 41.88747419132829, "grad_norm": 14.449097633361816, "learning_rate": 2.9056262904335855e-05, "loss": 0.8224, "step": 243450 }, { "epoch": 41.89607708189952, "grad_norm": 14.065131187438965, "learning_rate": 2.9051961459050243e-05, "loss": 0.8507, "step": 243500 }, { "epoch": 41.90467997247075, "grad_norm": 15.586149215698242, "learning_rate": 2.9047660013764628e-05, "loss": 0.8721, "step": 243550 }, { "epoch": 41.913282863041985, "grad_norm": 8.788996696472168, "learning_rate": 2.9043358568479012e-05, "loss": 0.8702, "step": 243600 }, { "epoch": 41.92188575361321, "grad_norm": 12.191900253295898, "learning_rate": 2.903905712319339e-05, "loss": 0.8177, "step": 243650 }, { "epoch": 41.93048864418444, "grad_norm": 17.25728988647461, "learning_rate": 2.9034755677907778e-05, "loss": 0.7918, "step": 243700 }, { "epoch": 41.939091534755676, "grad_norm": 13.738006591796875, "learning_rate": 2.9030454232622163e-05, "loss": 0.8425, "step": 243750 }, { "epoch": 41.94769442532691, "grad_norm": 10.030852317810059, "learning_rate": 2.9026152787336547e-05, "loss": 0.8391, "step": 243800 }, { "epoch": 41.95629731589814, "grad_norm": 11.028874397277832, "learning_rate": 2.9021851342050932e-05, "loss": 0.841, "step": 243850 }, { "epoch": 41.96490020646937, "grad_norm": 12.35028076171875, "learning_rate": 2.9017549896765316e-05, "loss": 0.8606, "step": 243900 }, { "epoch": 41.973503097040606, "grad_norm": 8.74411678314209, "learning_rate": 2.9013248451479698e-05, "loss": 0.8417, "step": 243950 }, { "epoch": 41.98210598761184, "grad_norm": 11.727701187133789, "learning_rate": 2.9008947006194082e-05, "loss": 0.806, "step": 244000 }, { "epoch": 41.99070887818307, "grad_norm": 8.772538185119629, "learning_rate": 2.9004645560908467e-05, "loss": 0.75, "step": 244050 }, { "epoch": 41.9993117687543, "grad_norm": 15.333983421325684, "learning_rate": 2.900034411562285e-05, "loss": 0.8136, "step": 244100 }, { "epoch": 42.0, "eval_accuracy": 0.6209774565479264, "eval_f1": 0.6174393798242791, "eval_f1_DuraRiadoRio_16x16": 0.5978083469340173, "eval_f1_Mole_16x16": 0.6700043535045712, "eval_f1_Quebrado_16x16": 0.6948148148148148, "eval_f1_RiadoRio_16x16": 0.5076096687555953, "eval_f1_RioFechado_16x16": 0.6169597151123971, "eval_loss": 1.29599928855896, "eval_precision": 0.6213094953906667, "eval_precision_DuraRiadoRio_16x16": 0.6458438287153653, "eval_precision_Mole_16x16": 0.6720524017467249, "eval_precision_Quebrado_16x16": 0.6059431524547804, "eval_precision_RiadoRio_16x16": 0.5569744597249509, "eval_precision_RioFechado_16x16": 0.6257336343115124, "eval_recall": 0.6226679625928664, "eval_recall_DuraRiadoRio_16x16": 0.5564236111111112, "eval_recall_Mole_16x16": 0.66796875, "eval_recall_Quebrado_16x16": 0.8142361111111112, "eval_recall_RiadoRio_16x16": 0.4662828947368421, "eval_recall_RioFechado_16x16": 0.6084284460052678, "eval_runtime": 46.0147, "eval_samples_per_second": 252.572, "eval_steps_per_second": 15.799, "step": 244104 }, { "epoch": 42.007914659325536, "grad_norm": 12.405303001403809, "learning_rate": 2.8996042670337236e-05, "loss": 0.8359, "step": 244150 }, { "epoch": 42.01651754989677, "grad_norm": 15.5551176071167, "learning_rate": 2.8991741225051617e-05, "loss": 0.8672, "step": 244200 }, { "epoch": 42.025120440467994, "grad_norm": 11.71169662475586, "learning_rate": 2.8987439779766e-05, "loss": 0.7734, "step": 244250 }, { "epoch": 42.033723331039226, "grad_norm": 11.010003089904785, "learning_rate": 2.8983138334480386e-05, "loss": 0.8497, "step": 244300 }, { "epoch": 42.04232622161046, "grad_norm": 12.918739318847656, "learning_rate": 2.897883688919477e-05, "loss": 0.8298, "step": 244350 }, { "epoch": 42.05092911218169, "grad_norm": 9.055424690246582, "learning_rate": 2.8974535443909155e-05, "loss": 0.7285, "step": 244400 }, { "epoch": 42.059532002752924, "grad_norm": 8.90597915649414, "learning_rate": 2.8970233998623543e-05, "loss": 0.7543, "step": 244450 }, { "epoch": 42.06813489332416, "grad_norm": 13.00718879699707, "learning_rate": 2.896593255333792e-05, "loss": 0.8302, "step": 244500 }, { "epoch": 42.07673778389539, "grad_norm": 11.988373756408691, "learning_rate": 2.8961631108052306e-05, "loss": 0.7935, "step": 244550 }, { "epoch": 42.08534067446662, "grad_norm": 12.517027854919434, "learning_rate": 2.895732966276669e-05, "loss": 0.8157, "step": 244600 }, { "epoch": 42.093943565037854, "grad_norm": 11.139464378356934, "learning_rate": 2.8953028217481075e-05, "loss": 0.7448, "step": 244650 }, { "epoch": 42.10254645560909, "grad_norm": 16.462129592895508, "learning_rate": 2.8948726772195463e-05, "loss": 0.8398, "step": 244700 }, { "epoch": 42.11114934618032, "grad_norm": 13.256123542785645, "learning_rate": 2.894442532690984e-05, "loss": 0.8356, "step": 244750 }, { "epoch": 42.11975223675155, "grad_norm": 16.854116439819336, "learning_rate": 2.8940123881624225e-05, "loss": 0.8104, "step": 244800 }, { "epoch": 42.12835512732278, "grad_norm": 13.282781600952148, "learning_rate": 2.893582243633861e-05, "loss": 0.8263, "step": 244850 }, { "epoch": 42.13695801789401, "grad_norm": 10.077037811279297, "learning_rate": 2.8931520991052997e-05, "loss": 0.8422, "step": 244900 }, { "epoch": 42.14556090846524, "grad_norm": 12.444628715515137, "learning_rate": 2.8927219545767382e-05, "loss": 0.8026, "step": 244950 }, { "epoch": 42.154163799036475, "grad_norm": 14.589683532714844, "learning_rate": 2.892291810048176e-05, "loss": 0.8627, "step": 245000 }, { "epoch": 42.16276668960771, "grad_norm": 11.46100902557373, "learning_rate": 2.8918616655196144e-05, "loss": 0.7674, "step": 245050 }, { "epoch": 42.17136958017894, "grad_norm": 12.694865226745605, "learning_rate": 2.8914315209910532e-05, "loss": 0.7874, "step": 245100 }, { "epoch": 42.17997247075017, "grad_norm": 16.3592586517334, "learning_rate": 2.8910013764624917e-05, "loss": 0.8322, "step": 245150 }, { "epoch": 42.188575361321405, "grad_norm": 15.941341400146484, "learning_rate": 2.89057123193393e-05, "loss": 0.8492, "step": 245200 }, { "epoch": 42.19717825189264, "grad_norm": 11.980931282043457, "learning_rate": 2.8901410874053686e-05, "loss": 0.8317, "step": 245250 }, { "epoch": 42.20578114246387, "grad_norm": 16.67333221435547, "learning_rate": 2.8897109428768064e-05, "loss": 0.8349, "step": 245300 }, { "epoch": 42.2143840330351, "grad_norm": 14.911019325256348, "learning_rate": 2.8892807983482452e-05, "loss": 0.7868, "step": 245350 }, { "epoch": 42.222986923606335, "grad_norm": 11.145984649658203, "learning_rate": 2.8888506538196836e-05, "loss": 0.7539, "step": 245400 }, { "epoch": 42.23158981417756, "grad_norm": 13.174148559570312, "learning_rate": 2.888420509291122e-05, "loss": 0.8115, "step": 245450 }, { "epoch": 42.24019270474879, "grad_norm": 15.336188316345215, "learning_rate": 2.8879903647625605e-05, "loss": 0.8308, "step": 245500 }, { "epoch": 42.248795595320026, "grad_norm": 12.341429710388184, "learning_rate": 2.8875602202339987e-05, "loss": 0.7911, "step": 245550 }, { "epoch": 42.25739848589126, "grad_norm": 13.053993225097656, "learning_rate": 2.887130075705437e-05, "loss": 0.8308, "step": 245600 }, { "epoch": 42.26600137646249, "grad_norm": 7.296080589294434, "learning_rate": 2.8866999311768756e-05, "loss": 0.7877, "step": 245650 }, { "epoch": 42.27460426703372, "grad_norm": 12.040671348571777, "learning_rate": 2.886269786648314e-05, "loss": 0.7662, "step": 245700 }, { "epoch": 42.283207157604956, "grad_norm": 23.18377685546875, "learning_rate": 2.8858396421197525e-05, "loss": 0.8478, "step": 245750 }, { "epoch": 42.29181004817619, "grad_norm": 15.246589660644531, "learning_rate": 2.8854094975911906e-05, "loss": 0.8265, "step": 245800 }, { "epoch": 42.30041293874742, "grad_norm": 22.625085830688477, "learning_rate": 2.884979353062629e-05, "loss": 0.7411, "step": 245850 }, { "epoch": 42.30901582931865, "grad_norm": 9.30186653137207, "learning_rate": 2.8845492085340675e-05, "loss": 0.8105, "step": 245900 }, { "epoch": 42.317618719889886, "grad_norm": 10.70002269744873, "learning_rate": 2.884119064005506e-05, "loss": 0.7904, "step": 245950 }, { "epoch": 42.32622161046112, "grad_norm": 11.053893089294434, "learning_rate": 2.8836889194769444e-05, "loss": 0.8354, "step": 246000 }, { "epoch": 42.334824501032344, "grad_norm": 13.18179702758789, "learning_rate": 2.883258774948383e-05, "loss": 0.8092, "step": 246050 }, { "epoch": 42.343427391603576, "grad_norm": 17.62956428527832, "learning_rate": 2.882828630419821e-05, "loss": 0.7773, "step": 246100 }, { "epoch": 42.35203028217481, "grad_norm": 13.388457298278809, "learning_rate": 2.8823984858912595e-05, "loss": 0.7362, "step": 246150 }, { "epoch": 42.36063317274604, "grad_norm": 10.796355247497559, "learning_rate": 2.881968341362698e-05, "loss": 0.7371, "step": 246200 }, { "epoch": 42.369236063317274, "grad_norm": 12.475397109985352, "learning_rate": 2.8815381968341364e-05, "loss": 0.8116, "step": 246250 }, { "epoch": 42.37783895388851, "grad_norm": 15.76215934753418, "learning_rate": 2.881108052305575e-05, "loss": 0.781, "step": 246300 }, { "epoch": 42.38644184445974, "grad_norm": 11.963041305541992, "learning_rate": 2.880677907777013e-05, "loss": 0.7748, "step": 246350 }, { "epoch": 42.39504473503097, "grad_norm": 14.899174690246582, "learning_rate": 2.8802477632484514e-05, "loss": 0.7818, "step": 246400 }, { "epoch": 42.403647625602204, "grad_norm": 16.522701263427734, "learning_rate": 2.87981761871989e-05, "loss": 0.8106, "step": 246450 }, { "epoch": 42.41225051617344, "grad_norm": 12.372855186462402, "learning_rate": 2.8793874741913286e-05, "loss": 0.7911, "step": 246500 }, { "epoch": 42.42085340674467, "grad_norm": 16.911354064941406, "learning_rate": 2.878957329662767e-05, "loss": 0.8712, "step": 246550 }, { "epoch": 42.429456297315895, "grad_norm": 15.787970542907715, "learning_rate": 2.8785271851342056e-05, "loss": 0.7834, "step": 246600 }, { "epoch": 42.43805918788713, "grad_norm": 9.231060981750488, "learning_rate": 2.8780970406056433e-05, "loss": 0.8347, "step": 246650 }, { "epoch": 42.44666207845836, "grad_norm": 12.25950813293457, "learning_rate": 2.8776668960770818e-05, "loss": 0.7702, "step": 246700 }, { "epoch": 42.45526496902959, "grad_norm": 13.37641716003418, "learning_rate": 2.8772367515485206e-05, "loss": 0.8326, "step": 246750 }, { "epoch": 42.463867859600825, "grad_norm": 11.521696090698242, "learning_rate": 2.876806607019959e-05, "loss": 0.8159, "step": 246800 }, { "epoch": 42.47247075017206, "grad_norm": 14.11585807800293, "learning_rate": 2.8763764624913975e-05, "loss": 0.8389, "step": 246850 }, { "epoch": 42.48107364074329, "grad_norm": 6.414308547973633, "learning_rate": 2.8759463179628353e-05, "loss": 0.7342, "step": 246900 }, { "epoch": 42.48967653131452, "grad_norm": 14.323902130126953, "learning_rate": 2.875516173434274e-05, "loss": 0.8592, "step": 246950 }, { "epoch": 42.498279421885755, "grad_norm": 11.981422424316406, "learning_rate": 2.8750860289057125e-05, "loss": 0.7952, "step": 247000 }, { "epoch": 42.50688231245699, "grad_norm": 18.365270614624023, "learning_rate": 2.874655884377151e-05, "loss": 0.8812, "step": 247050 }, { "epoch": 42.51548520302822, "grad_norm": 14.23944091796875, "learning_rate": 2.8742257398485894e-05, "loss": 0.7692, "step": 247100 }, { "epoch": 42.52408809359945, "grad_norm": 17.516273498535156, "learning_rate": 2.8737955953200276e-05, "loss": 0.7834, "step": 247150 }, { "epoch": 42.53269098417068, "grad_norm": 6.375917434692383, "learning_rate": 2.873365450791466e-05, "loss": 0.845, "step": 247200 }, { "epoch": 42.54129387474191, "grad_norm": 14.74416446685791, "learning_rate": 2.8729353062629045e-05, "loss": 0.8121, "step": 247250 }, { "epoch": 42.54989676531314, "grad_norm": 15.271775245666504, "learning_rate": 2.872505161734343e-05, "loss": 0.8128, "step": 247300 }, { "epoch": 42.558499655884376, "grad_norm": 7.069077014923096, "learning_rate": 2.8720750172057814e-05, "loss": 0.8148, "step": 247350 }, { "epoch": 42.56710254645561, "grad_norm": 16.839385986328125, "learning_rate": 2.87164487267722e-05, "loss": 0.8263, "step": 247400 }, { "epoch": 42.57570543702684, "grad_norm": 18.780080795288086, "learning_rate": 2.871214728148658e-05, "loss": 0.8058, "step": 247450 }, { "epoch": 42.58430832759807, "grad_norm": 17.772809982299805, "learning_rate": 2.8707845836200964e-05, "loss": 0.7465, "step": 247500 }, { "epoch": 42.592911218169306, "grad_norm": 13.072630882263184, "learning_rate": 2.870354439091535e-05, "loss": 0.7944, "step": 247550 }, { "epoch": 42.60151410874054, "grad_norm": 7.415162563323975, "learning_rate": 2.8699242945629733e-05, "loss": 0.8346, "step": 247600 }, { "epoch": 42.61011699931177, "grad_norm": 8.293112754821777, "learning_rate": 2.8694941500344118e-05, "loss": 0.8142, "step": 247650 }, { "epoch": 42.618719889883, "grad_norm": 12.290423393249512, "learning_rate": 2.86906400550585e-05, "loss": 0.8166, "step": 247700 }, { "epoch": 42.627322780454236, "grad_norm": 14.491817474365234, "learning_rate": 2.8686338609772884e-05, "loss": 0.8424, "step": 247750 }, { "epoch": 42.63592567102546, "grad_norm": 15.577738761901855, "learning_rate": 2.8682037164487268e-05, "loss": 0.7509, "step": 247800 }, { "epoch": 42.644528561596694, "grad_norm": 13.080686569213867, "learning_rate": 2.8677735719201653e-05, "loss": 0.8833, "step": 247850 }, { "epoch": 42.653131452167926, "grad_norm": 17.152935028076172, "learning_rate": 2.8673434273916037e-05, "loss": 0.8319, "step": 247900 }, { "epoch": 42.66173434273916, "grad_norm": 11.436297416687012, "learning_rate": 2.8669132828630425e-05, "loss": 0.837, "step": 247950 }, { "epoch": 42.67033723331039, "grad_norm": 16.810489654541016, "learning_rate": 2.8664831383344803e-05, "loss": 0.8715, "step": 248000 }, { "epoch": 42.678940123881624, "grad_norm": 13.854514122009277, "learning_rate": 2.8660529938059188e-05, "loss": 0.9016, "step": 248050 }, { "epoch": 42.68754301445286, "grad_norm": 12.908648490905762, "learning_rate": 2.8656228492773572e-05, "loss": 0.8397, "step": 248100 }, { "epoch": 42.69614590502409, "grad_norm": 9.85252857208252, "learning_rate": 2.865192704748796e-05, "loss": 0.7991, "step": 248150 }, { "epoch": 42.70474879559532, "grad_norm": 14.318450927734375, "learning_rate": 2.8647625602202345e-05, "loss": 0.7861, "step": 248200 }, { "epoch": 42.713351686166554, "grad_norm": 11.757149696350098, "learning_rate": 2.8643324156916722e-05, "loss": 0.821, "step": 248250 }, { "epoch": 42.72195457673779, "grad_norm": 12.90434741973877, "learning_rate": 2.8639022711631107e-05, "loss": 0.7899, "step": 248300 }, { "epoch": 42.73055746730902, "grad_norm": 9.40472412109375, "learning_rate": 2.8634721266345495e-05, "loss": 0.7938, "step": 248350 }, { "epoch": 42.739160357880245, "grad_norm": 12.55527114868164, "learning_rate": 2.863041982105988e-05, "loss": 0.8043, "step": 248400 }, { "epoch": 42.74776324845148, "grad_norm": 14.668388366699219, "learning_rate": 2.8626118375774264e-05, "loss": 0.7898, "step": 248450 }, { "epoch": 42.75636613902271, "grad_norm": 16.612852096557617, "learning_rate": 2.8621816930488642e-05, "loss": 0.8429, "step": 248500 }, { "epoch": 42.76496902959394, "grad_norm": 21.359542846679688, "learning_rate": 2.861751548520303e-05, "loss": 0.8149, "step": 248550 }, { "epoch": 42.773571920165175, "grad_norm": 10.848679542541504, "learning_rate": 2.8613214039917414e-05, "loss": 0.8361, "step": 248600 }, { "epoch": 42.78217481073641, "grad_norm": 11.37220573425293, "learning_rate": 2.86089125946318e-05, "loss": 0.8203, "step": 248650 }, { "epoch": 42.79077770130764, "grad_norm": 10.268576622009277, "learning_rate": 2.8604611149346183e-05, "loss": 0.8091, "step": 248700 }, { "epoch": 42.79938059187887, "grad_norm": 12.618490219116211, "learning_rate": 2.8600309704060568e-05, "loss": 0.8414, "step": 248750 }, { "epoch": 42.807983482450105, "grad_norm": 12.281126976013184, "learning_rate": 2.859600825877495e-05, "loss": 0.808, "step": 248800 }, { "epoch": 42.81658637302134, "grad_norm": 15.84571361541748, "learning_rate": 2.8591706813489334e-05, "loss": 0.8621, "step": 248850 }, { "epoch": 42.82518926359257, "grad_norm": 13.014327049255371, "learning_rate": 2.8587405368203718e-05, "loss": 0.7439, "step": 248900 }, { "epoch": 42.8337921541638, "grad_norm": 8.32402515411377, "learning_rate": 2.8583103922918103e-05, "loss": 0.8958, "step": 248950 }, { "epoch": 42.84239504473503, "grad_norm": 10.310498237609863, "learning_rate": 2.8578802477632487e-05, "loss": 0.8641, "step": 249000 }, { "epoch": 42.85099793530626, "grad_norm": 6.86198616027832, "learning_rate": 2.857450103234687e-05, "loss": 0.7689, "step": 249050 }, { "epoch": 42.85960082587749, "grad_norm": 15.18387222290039, "learning_rate": 2.8570199587061253e-05, "loss": 0.8588, "step": 249100 }, { "epoch": 42.868203716448726, "grad_norm": 18.804094314575195, "learning_rate": 2.8565898141775638e-05, "loss": 0.8286, "step": 249150 }, { "epoch": 42.87680660701996, "grad_norm": 8.214497566223145, "learning_rate": 2.8561596696490022e-05, "loss": 0.7419, "step": 249200 }, { "epoch": 42.88540949759119, "grad_norm": 15.550756454467773, "learning_rate": 2.8557295251204407e-05, "loss": 0.8157, "step": 249250 }, { "epoch": 42.89401238816242, "grad_norm": 13.976816177368164, "learning_rate": 2.8552993805918788e-05, "loss": 0.8288, "step": 249300 }, { "epoch": 42.902615278733656, "grad_norm": 15.836051940917969, "learning_rate": 2.8548692360633173e-05, "loss": 0.7918, "step": 249350 }, { "epoch": 42.91121816930489, "grad_norm": 9.228628158569336, "learning_rate": 2.8544390915347557e-05, "loss": 0.7906, "step": 249400 }, { "epoch": 42.91982105987612, "grad_norm": 14.41982364654541, "learning_rate": 2.854008947006194e-05, "loss": 0.8688, "step": 249450 }, { "epoch": 42.92842395044735, "grad_norm": 18.1798038482666, "learning_rate": 2.8535788024776326e-05, "loss": 0.8177, "step": 249500 }, { "epoch": 42.93702684101858, "grad_norm": 14.237772941589355, "learning_rate": 2.8531486579490714e-05, "loss": 0.7975, "step": 249550 }, { "epoch": 42.94562973158981, "grad_norm": 8.653955459594727, "learning_rate": 2.8527185134205092e-05, "loss": 0.8348, "step": 249600 }, { "epoch": 42.954232622161044, "grad_norm": 14.19395923614502, "learning_rate": 2.8522883688919477e-05, "loss": 0.8137, "step": 249650 }, { "epoch": 42.962835512732276, "grad_norm": 10.018625259399414, "learning_rate": 2.851858224363386e-05, "loss": 0.7734, "step": 249700 }, { "epoch": 42.97143840330351, "grad_norm": 10.240616798400879, "learning_rate": 2.851428079834825e-05, "loss": 0.8667, "step": 249750 }, { "epoch": 42.98004129387474, "grad_norm": 12.3809814453125, "learning_rate": 2.8509979353062634e-05, "loss": 0.8362, "step": 249800 }, { "epoch": 42.988644184445974, "grad_norm": 12.625358581542969, "learning_rate": 2.850567790777701e-05, "loss": 0.8115, "step": 249850 }, { "epoch": 42.997247075017206, "grad_norm": 6.730236530303955, "learning_rate": 2.8501376462491396e-05, "loss": 0.8146, "step": 249900 }, { "epoch": 43.0, "eval_accuracy": 0.5647048700739976, "eval_f1": 0.5431624785368493, "eval_f1_DuraRiadoRio_16x16": 0.48477284073889165, "eval_f1_Mole_16x16": 0.6093313265964309, "eval_f1_Quebrado_16x16": 0.6934027777777778, "eval_f1_RiadoRio_16x16": 0.3484066767830046, "eval_f1_RioFechado_16x16": 0.5798987707881417, "eval_loss": 1.59076988697052, "eval_precision": 0.5820287696646412, "eval_precision_DuraRiadoRio_16x16": 0.5705052878965923, "eval_precision_Mole_16x16": 0.6037494674051981, "eval_precision_Quebrado_16x16": 0.5778356481481481, "eval_precision_RiadoRio_16x16": 0.6651216685979142, "eval_precision_RioFechado_16x16": 0.4929317762753534, "eval_recall": 0.568671593817548, "eval_recall_DuraRiadoRio_16x16": 0.4214409722222222, "eval_recall_Mole_16x16": 0.6150173611111112, "eval_recall_Quebrado_16x16": 0.8667534722222222, "eval_recall_RiadoRio_16x16": 0.23601973684210525, "eval_recall_RioFechado_16x16": 0.704126426690079, "eval_runtime": 47.0444, "eval_samples_per_second": 247.043, "eval_steps_per_second": 15.454, "step": 249916 }, { "epoch": 43.00584996558844, "grad_norm": 12.441060066223145, "learning_rate": 2.849707501720578e-05, "loss": 0.8174, "step": 249950 }, { "epoch": 43.01445285615967, "grad_norm": 19.348827362060547, "learning_rate": 2.849277357192017e-05, "loss": 0.8577, "step": 250000 }, { "epoch": 43.023055746730904, "grad_norm": 12.276566505432129, "learning_rate": 2.8488472126634553e-05, "loss": 0.7885, "step": 250050 }, { "epoch": 43.03165863730214, "grad_norm": 10.822646141052246, "learning_rate": 2.8484170681348938e-05, "loss": 0.7984, "step": 250100 }, { "epoch": 43.04026152787336, "grad_norm": 8.743988037109375, "learning_rate": 2.8479869236063315e-05, "loss": 0.8004, "step": 250150 }, { "epoch": 43.048864418444595, "grad_norm": 9.365748405456543, "learning_rate": 2.8475567790777703e-05, "loss": 0.856, "step": 250200 }, { "epoch": 43.05746730901583, "grad_norm": 8.009637832641602, "learning_rate": 2.8471266345492088e-05, "loss": 0.7432, "step": 250250 }, { "epoch": 43.06607019958706, "grad_norm": 11.18679141998291, "learning_rate": 2.8466964900206472e-05, "loss": 0.7924, "step": 250300 }, { "epoch": 43.07467309015829, "grad_norm": 13.890905380249023, "learning_rate": 2.8462663454920857e-05, "loss": 0.7732, "step": 250350 }, { "epoch": 43.083275980729525, "grad_norm": 13.688037872314453, "learning_rate": 2.8458362009635238e-05, "loss": 0.7234, "step": 250400 }, { "epoch": 43.09187887130076, "grad_norm": 13.293739318847656, "learning_rate": 2.8454060564349623e-05, "loss": 0.727, "step": 250450 }, { "epoch": 43.10048176187199, "grad_norm": 17.387861251831055, "learning_rate": 2.8449759119064007e-05, "loss": 0.8467, "step": 250500 }, { "epoch": 43.10908465244322, "grad_norm": 16.467947006225586, "learning_rate": 2.8445457673778392e-05, "loss": 0.8424, "step": 250550 }, { "epoch": 43.117687543014455, "grad_norm": 12.250519752502441, "learning_rate": 2.8441156228492776e-05, "loss": 0.777, "step": 250600 }, { "epoch": 43.12629043358569, "grad_norm": 13.637967109680176, "learning_rate": 2.8436854783207158e-05, "loss": 0.7975, "step": 250650 }, { "epoch": 43.13489332415692, "grad_norm": 14.31413459777832, "learning_rate": 2.8432553337921542e-05, "loss": 0.8101, "step": 250700 }, { "epoch": 43.143496214728145, "grad_norm": 11.120884895324707, "learning_rate": 2.8428251892635927e-05, "loss": 0.7836, "step": 250750 }, { "epoch": 43.15209910529938, "grad_norm": 15.506579399108887, "learning_rate": 2.842395044735031e-05, "loss": 0.7821, "step": 250800 }, { "epoch": 43.16070199587061, "grad_norm": 14.76003646850586, "learning_rate": 2.8419649002064696e-05, "loss": 0.8538, "step": 250850 }, { "epoch": 43.16930488644184, "grad_norm": 12.033001899719238, "learning_rate": 2.841534755677908e-05, "loss": 0.8104, "step": 250900 }, { "epoch": 43.177907777013075, "grad_norm": 9.425711631774902, "learning_rate": 2.841104611149346e-05, "loss": 0.8297, "step": 250950 }, { "epoch": 43.18651066758431, "grad_norm": 8.827445030212402, "learning_rate": 2.8406744666207846e-05, "loss": 0.8198, "step": 251000 }, { "epoch": 43.19511355815554, "grad_norm": 11.827754974365234, "learning_rate": 2.840244322092223e-05, "loss": 0.7893, "step": 251050 }, { "epoch": 43.20371644872677, "grad_norm": 12.493127822875977, "learning_rate": 2.8398141775636615e-05, "loss": 0.8232, "step": 251100 }, { "epoch": 43.212319339298006, "grad_norm": 10.925695419311523, "learning_rate": 2.8393840330351e-05, "loss": 0.7653, "step": 251150 }, { "epoch": 43.22092222986924, "grad_norm": 12.584985733032227, "learning_rate": 2.838953888506538e-05, "loss": 0.8029, "step": 251200 }, { "epoch": 43.22952512044047, "grad_norm": 14.806260108947754, "learning_rate": 2.8385237439779766e-05, "loss": 0.8572, "step": 251250 }, { "epoch": 43.2381280110117, "grad_norm": 13.408487319946289, "learning_rate": 2.838093599449415e-05, "loss": 0.8141, "step": 251300 }, { "epoch": 43.24673090158293, "grad_norm": 14.102158546447754, "learning_rate": 2.8376634549208535e-05, "loss": 0.7472, "step": 251350 }, { "epoch": 43.25533379215416, "grad_norm": 14.053845405578613, "learning_rate": 2.8372333103922923e-05, "loss": 0.8316, "step": 251400 }, { "epoch": 43.263936682725394, "grad_norm": 13.342300415039062, "learning_rate": 2.83680316586373e-05, "loss": 0.7886, "step": 251450 }, { "epoch": 43.272539573296626, "grad_norm": 14.746055603027344, "learning_rate": 2.8363730213351685e-05, "loss": 0.8189, "step": 251500 }, { "epoch": 43.28114246386786, "grad_norm": 14.517843246459961, "learning_rate": 2.835942876806607e-05, "loss": 0.8177, "step": 251550 }, { "epoch": 43.28974535443909, "grad_norm": 10.432485580444336, "learning_rate": 2.8355127322780457e-05, "loss": 0.7398, "step": 251600 }, { "epoch": 43.298348245010324, "grad_norm": 11.342174530029297, "learning_rate": 2.8350825877494842e-05, "loss": 0.7911, "step": 251650 }, { "epoch": 43.306951135581556, "grad_norm": 12.323138236999512, "learning_rate": 2.8346524432209227e-05, "loss": 0.8328, "step": 251700 }, { "epoch": 43.31555402615279, "grad_norm": 6.683281898498535, "learning_rate": 2.8342222986923604e-05, "loss": 0.7638, "step": 251750 }, { "epoch": 43.32415691672402, "grad_norm": 13.43424129486084, "learning_rate": 2.8337921541637992e-05, "loss": 0.8686, "step": 251800 }, { "epoch": 43.332759807295254, "grad_norm": 17.391210556030273, "learning_rate": 2.8333620096352377e-05, "loss": 0.745, "step": 251850 }, { "epoch": 43.34136269786649, "grad_norm": 12.526958465576172, "learning_rate": 2.832931865106676e-05, "loss": 0.7825, "step": 251900 }, { "epoch": 43.34996558843771, "grad_norm": 10.131121635437012, "learning_rate": 2.8325017205781146e-05, "loss": 0.8673, "step": 251950 }, { "epoch": 43.358568479008945, "grad_norm": 9.361234664916992, "learning_rate": 2.8320715760495524e-05, "loss": 0.793, "step": 252000 }, { "epoch": 43.36717136958018, "grad_norm": 14.401649475097656, "learning_rate": 2.8316414315209912e-05, "loss": 0.7723, "step": 252050 }, { "epoch": 43.37577426015141, "grad_norm": 9.568771362304688, "learning_rate": 2.8312112869924296e-05, "loss": 0.801, "step": 252100 }, { "epoch": 43.38437715072264, "grad_norm": 12.75228500366211, "learning_rate": 2.830781142463868e-05, "loss": 0.8158, "step": 252150 }, { "epoch": 43.392980041293875, "grad_norm": 11.567577362060547, "learning_rate": 2.8303509979353065e-05, "loss": 0.8404, "step": 252200 }, { "epoch": 43.40158293186511, "grad_norm": 9.115324974060059, "learning_rate": 2.829920853406745e-05, "loss": 0.8521, "step": 252250 }, { "epoch": 43.41018582243634, "grad_norm": 7.899247169494629, "learning_rate": 2.829490708878183e-05, "loss": 0.8542, "step": 252300 }, { "epoch": 43.41878871300757, "grad_norm": 10.808917045593262, "learning_rate": 2.8290605643496216e-05, "loss": 0.8428, "step": 252350 }, { "epoch": 43.427391603578805, "grad_norm": 10.26830768585205, "learning_rate": 2.82863041982106e-05, "loss": 0.769, "step": 252400 }, { "epoch": 43.43599449415004, "grad_norm": 11.034826278686523, "learning_rate": 2.8282002752924985e-05, "loss": 0.854, "step": 252450 }, { "epoch": 43.44459738472127, "grad_norm": 12.762770652770996, "learning_rate": 2.827770130763937e-05, "loss": 0.8307, "step": 252500 }, { "epoch": 43.453200275292495, "grad_norm": 20.09299659729004, "learning_rate": 2.827339986235375e-05, "loss": 0.7786, "step": 252550 }, { "epoch": 43.46180316586373, "grad_norm": 6.755995273590088, "learning_rate": 2.8269098417068135e-05, "loss": 0.8006, "step": 252600 }, { "epoch": 43.47040605643496, "grad_norm": 14.030315399169922, "learning_rate": 2.826479697178252e-05, "loss": 0.7989, "step": 252650 }, { "epoch": 43.47900894700619, "grad_norm": 16.482744216918945, "learning_rate": 2.8260495526496904e-05, "loss": 0.854, "step": 252700 }, { "epoch": 43.487611837577425, "grad_norm": 11.982182502746582, "learning_rate": 2.825619408121129e-05, "loss": 0.7902, "step": 252750 }, { "epoch": 43.49621472814866, "grad_norm": 14.757675170898438, "learning_rate": 2.825189263592567e-05, "loss": 0.7823, "step": 252800 }, { "epoch": 43.50481761871989, "grad_norm": 11.018411636352539, "learning_rate": 2.8247591190640055e-05, "loss": 0.8123, "step": 252850 }, { "epoch": 43.51342050929112, "grad_norm": 14.224705696105957, "learning_rate": 2.824328974535444e-05, "loss": 0.8016, "step": 252900 }, { "epoch": 43.522023399862356, "grad_norm": 14.507723808288574, "learning_rate": 2.8238988300068824e-05, "loss": 0.769, "step": 252950 }, { "epoch": 43.53062629043359, "grad_norm": 15.463092803955078, "learning_rate": 2.823468685478321e-05, "loss": 0.8334, "step": 253000 }, { "epoch": 43.53922918100482, "grad_norm": 13.656926155090332, "learning_rate": 2.8230385409497596e-05, "loss": 0.7828, "step": 253050 }, { "epoch": 43.547832071576046, "grad_norm": 12.06508731842041, "learning_rate": 2.8226083964211974e-05, "loss": 0.778, "step": 253100 }, { "epoch": 43.55643496214728, "grad_norm": 15.26755142211914, "learning_rate": 2.822178251892636e-05, "loss": 0.7759, "step": 253150 }, { "epoch": 43.56503785271851, "grad_norm": 15.002814292907715, "learning_rate": 2.8217481073640743e-05, "loss": 0.8119, "step": 253200 }, { "epoch": 43.573640743289744, "grad_norm": 15.392219543457031, "learning_rate": 2.821317962835513e-05, "loss": 0.7939, "step": 253250 }, { "epoch": 43.582243633860976, "grad_norm": 14.52626895904541, "learning_rate": 2.8208878183069516e-05, "loss": 0.7944, "step": 253300 }, { "epoch": 43.59084652443221, "grad_norm": 5.8132805824279785, "learning_rate": 2.8204576737783893e-05, "loss": 0.7745, "step": 253350 }, { "epoch": 43.59944941500344, "grad_norm": 11.387760162353516, "learning_rate": 2.8200275292498278e-05, "loss": 0.8032, "step": 253400 }, { "epoch": 43.608052305574674, "grad_norm": 20.84513282775879, "learning_rate": 2.8195973847212666e-05, "loss": 0.8661, "step": 253450 }, { "epoch": 43.616655196145906, "grad_norm": 11.441704750061035, "learning_rate": 2.819167240192705e-05, "loss": 0.7824, "step": 253500 }, { "epoch": 43.62525808671714, "grad_norm": 9.441070556640625, "learning_rate": 2.8187370956641435e-05, "loss": 0.7447, "step": 253550 }, { "epoch": 43.63386097728837, "grad_norm": 9.980401992797852, "learning_rate": 2.818306951135582e-05, "loss": 0.8197, "step": 253600 }, { "epoch": 43.642463867859604, "grad_norm": 15.075821876525879, "learning_rate": 2.81787680660702e-05, "loss": 0.8226, "step": 253650 }, { "epoch": 43.65106675843083, "grad_norm": 14.459430694580078, "learning_rate": 2.8174466620784585e-05, "loss": 0.8291, "step": 253700 }, { "epoch": 43.65966964900206, "grad_norm": 14.023740768432617, "learning_rate": 2.817016517549897e-05, "loss": 0.7969, "step": 253750 }, { "epoch": 43.668272539573294, "grad_norm": 7.805643558502197, "learning_rate": 2.8165863730213354e-05, "loss": 0.7652, "step": 253800 }, { "epoch": 43.67687543014453, "grad_norm": 19.482864379882812, "learning_rate": 2.816156228492774e-05, "loss": 0.8223, "step": 253850 }, { "epoch": 43.68547832071576, "grad_norm": 16.532480239868164, "learning_rate": 2.815726083964212e-05, "loss": 0.8431, "step": 253900 }, { "epoch": 43.69408121128699, "grad_norm": 17.920696258544922, "learning_rate": 2.8152959394356505e-05, "loss": 0.8029, "step": 253950 }, { "epoch": 43.702684101858225, "grad_norm": 20.144207000732422, "learning_rate": 2.814865794907089e-05, "loss": 0.8148, "step": 254000 }, { "epoch": 43.71128699242946, "grad_norm": 15.093043327331543, "learning_rate": 2.8144356503785274e-05, "loss": 0.8746, "step": 254050 }, { "epoch": 43.71988988300069, "grad_norm": 10.123041152954102, "learning_rate": 2.814005505849966e-05, "loss": 0.8272, "step": 254100 }, { "epoch": 43.72849277357192, "grad_norm": 19.861963272094727, "learning_rate": 2.813575361321404e-05, "loss": 0.8162, "step": 254150 }, { "epoch": 43.737095664143155, "grad_norm": 16.44536781311035, "learning_rate": 2.8131452167928424e-05, "loss": 0.8376, "step": 254200 }, { "epoch": 43.74569855471439, "grad_norm": 10.942607879638672, "learning_rate": 2.812715072264281e-05, "loss": 0.7759, "step": 254250 }, { "epoch": 43.75430144528561, "grad_norm": 13.184904098510742, "learning_rate": 2.8122849277357193e-05, "loss": 0.8055, "step": 254300 }, { "epoch": 43.762904335856845, "grad_norm": 14.432320594787598, "learning_rate": 2.8118547832071578e-05, "loss": 0.8703, "step": 254350 }, { "epoch": 43.77150722642808, "grad_norm": 11.459864616394043, "learning_rate": 2.8114246386785962e-05, "loss": 0.8115, "step": 254400 }, { "epoch": 43.78011011699931, "grad_norm": 19.650100708007812, "learning_rate": 2.8109944941500343e-05, "loss": 0.801, "step": 254450 }, { "epoch": 43.78871300757054, "grad_norm": 15.598544120788574, "learning_rate": 2.8105643496214728e-05, "loss": 0.8287, "step": 254500 }, { "epoch": 43.797315898141775, "grad_norm": 11.473956108093262, "learning_rate": 2.8101342050929113e-05, "loss": 0.7706, "step": 254550 }, { "epoch": 43.80591878871301, "grad_norm": 10.646279335021973, "learning_rate": 2.8097040605643497e-05, "loss": 0.8506, "step": 254600 }, { "epoch": 43.81452167928424, "grad_norm": 5.812572002410889, "learning_rate": 2.8092739160357885e-05, "loss": 0.8229, "step": 254650 }, { "epoch": 43.82312456985547, "grad_norm": 14.091140747070312, "learning_rate": 2.8088437715072263e-05, "loss": 0.8367, "step": 254700 }, { "epoch": 43.831727460426706, "grad_norm": 12.500253677368164, "learning_rate": 2.8084136269786647e-05, "loss": 0.7728, "step": 254750 }, { "epoch": 43.84033035099794, "grad_norm": 13.954549789428711, "learning_rate": 2.8079834824501032e-05, "loss": 0.8222, "step": 254800 }, { "epoch": 43.84893324156917, "grad_norm": 17.542512893676758, "learning_rate": 2.807553337921542e-05, "loss": 0.823, "step": 254850 }, { "epoch": 43.857536132140396, "grad_norm": 11.172863006591797, "learning_rate": 2.8071231933929805e-05, "loss": 0.8318, "step": 254900 }, { "epoch": 43.86613902271163, "grad_norm": 12.088754653930664, "learning_rate": 2.8066930488644182e-05, "loss": 0.7917, "step": 254950 }, { "epoch": 43.87474191328286, "grad_norm": 9.418240547180176, "learning_rate": 2.8062629043358567e-05, "loss": 0.8226, "step": 255000 }, { "epoch": 43.883344803854094, "grad_norm": 13.522285461425781, "learning_rate": 2.8058327598072955e-05, "loss": 0.8673, "step": 255050 }, { "epoch": 43.891947694425326, "grad_norm": 13.25660514831543, "learning_rate": 2.805402615278734e-05, "loss": 0.8356, "step": 255100 }, { "epoch": 43.90055058499656, "grad_norm": 12.26223373413086, "learning_rate": 2.8049724707501724e-05, "loss": 0.8835, "step": 255150 }, { "epoch": 43.90915347556779, "grad_norm": 9.391382217407227, "learning_rate": 2.804542326221611e-05, "loss": 0.796, "step": 255200 }, { "epoch": 43.917756366139024, "grad_norm": 7.106705665588379, "learning_rate": 2.8041121816930486e-05, "loss": 0.8119, "step": 255250 }, { "epoch": 43.926359256710256, "grad_norm": 19.113075256347656, "learning_rate": 2.8036820371644874e-05, "loss": 0.8607, "step": 255300 }, { "epoch": 43.93496214728149, "grad_norm": 11.448079109191895, "learning_rate": 2.803251892635926e-05, "loss": 0.7922, "step": 255350 }, { "epoch": 43.94356503785272, "grad_norm": 12.818455696105957, "learning_rate": 2.8028217481073643e-05, "loss": 0.7889, "step": 255400 }, { "epoch": 43.952167928423954, "grad_norm": 12.305718421936035, "learning_rate": 2.8023916035788028e-05, "loss": 0.8138, "step": 255450 }, { "epoch": 43.96077081899518, "grad_norm": 10.830150604248047, "learning_rate": 2.801961459050241e-05, "loss": 0.7872, "step": 255500 }, { "epoch": 43.96937370956641, "grad_norm": 19.01058006286621, "learning_rate": 2.8015313145216794e-05, "loss": 0.7905, "step": 255550 }, { "epoch": 43.977976600137644, "grad_norm": 10.845855712890625, "learning_rate": 2.8011011699931178e-05, "loss": 0.891, "step": 255600 }, { "epoch": 43.98657949070888, "grad_norm": 17.329322814941406, "learning_rate": 2.8006710254645563e-05, "loss": 0.7766, "step": 255650 }, { "epoch": 43.99518238128011, "grad_norm": 9.265697479248047, "learning_rate": 2.8002408809359947e-05, "loss": 0.8045, "step": 255700 }, { "epoch": 44.0, "eval_accuracy": 0.543452073653416, "eval_f1": 0.5249983063322747, "eval_f1_DuraRiadoRio_16x16": 0.48758465011286684, "eval_f1_Mole_16x16": 0.41631866366848697, "eval_f1_Quebrado_16x16": 0.6391034378312888, "eval_f1_RiadoRio_16x16": 0.5089440276976341, "eval_f1_RioFechado_16x16": 0.5730407523510972, "eval_loss": 2.11653995513916, "eval_precision": 0.6027157634587227, "eval_precision_DuraRiadoRio_16x16": 0.6967741935483871, "eval_precision_Mole_16x16": 0.8009888751545118, "eval_precision_Quebrado_16x16": 0.4908118167015585, "eval_precision_RiadoRio_16x16": 0.4781351644380195, "eval_precision_RioFechado_16x16": 0.5468687674511368, "eval_recall": 0.5435778088402159, "eval_recall_DuraRiadoRio_16x16": 0.375, "eval_recall_Mole_16x16": 0.28125, "eval_recall_Quebrado_16x16": 0.9157986111111112, "eval_recall_RiadoRio_16x16": 0.5439967105263158, "eval_recall_RioFechado_16x16": 0.6018437225636524, "eval_runtime": 46.5407, "eval_samples_per_second": 249.717, "eval_steps_per_second": 15.621, "step": 255728 }, { "epoch": 44.00378527185134, "grad_norm": 12.085381507873535, "learning_rate": 2.7998107364074332e-05, "loss": 0.8119, "step": 255750 }, { "epoch": 44.012388162422575, "grad_norm": 10.212953567504883, "learning_rate": 2.7993805918788713e-05, "loss": 0.7967, "step": 255800 }, { "epoch": 44.02099105299381, "grad_norm": 15.47095775604248, "learning_rate": 2.7989504473503098e-05, "loss": 0.8111, "step": 255850 }, { "epoch": 44.02959394356504, "grad_norm": 9.976460456848145, "learning_rate": 2.7985203028217482e-05, "loss": 0.8246, "step": 255900 }, { "epoch": 44.03819683413627, "grad_norm": 16.91411018371582, "learning_rate": 2.7980901582931867e-05, "loss": 0.7878, "step": 255950 }, { "epoch": 44.046799724707505, "grad_norm": 11.249085426330566, "learning_rate": 2.797660013764625e-05, "loss": 0.7901, "step": 256000 }, { "epoch": 44.05540261527873, "grad_norm": 9.102032661437988, "learning_rate": 2.7972298692360632e-05, "loss": 0.8204, "step": 256050 }, { "epoch": 44.06400550584996, "grad_norm": 10.349048614501953, "learning_rate": 2.7967997247075017e-05, "loss": 0.8338, "step": 256100 }, { "epoch": 44.072608396421195, "grad_norm": 12.793670654296875, "learning_rate": 2.79636958017894e-05, "loss": 0.7914, "step": 256150 }, { "epoch": 44.08121128699243, "grad_norm": 11.02894401550293, "learning_rate": 2.7959394356503786e-05, "loss": 0.8349, "step": 256200 }, { "epoch": 44.08981417756366, "grad_norm": 12.970077514648438, "learning_rate": 2.7955092911218174e-05, "loss": 0.8109, "step": 256250 }, { "epoch": 44.09841706813489, "grad_norm": 7.896594524383545, "learning_rate": 2.7950791465932552e-05, "loss": 0.7659, "step": 256300 }, { "epoch": 44.107019958706125, "grad_norm": 8.891749382019043, "learning_rate": 2.7946490020646936e-05, "loss": 0.7978, "step": 256350 }, { "epoch": 44.11562284927736, "grad_norm": 20.075485229492188, "learning_rate": 2.794218857536132e-05, "loss": 0.7959, "step": 256400 }, { "epoch": 44.12422573984859, "grad_norm": 5.365163803100586, "learning_rate": 2.7937887130075706e-05, "loss": 0.8842, "step": 256450 }, { "epoch": 44.13282863041982, "grad_norm": 14.37308406829834, "learning_rate": 2.7933585684790094e-05, "loss": 0.8279, "step": 256500 }, { "epoch": 44.141431520991055, "grad_norm": 13.063761711120605, "learning_rate": 2.7929284239504478e-05, "loss": 0.7832, "step": 256550 }, { "epoch": 44.15003441156229, "grad_norm": 10.86627197265625, "learning_rate": 2.7924982794218856e-05, "loss": 0.8214, "step": 256600 }, { "epoch": 44.15863730213351, "grad_norm": 15.947823524475098, "learning_rate": 2.792068134893324e-05, "loss": 0.7865, "step": 256650 }, { "epoch": 44.167240192704746, "grad_norm": 9.672406196594238, "learning_rate": 2.791637990364763e-05, "loss": 0.7907, "step": 256700 }, { "epoch": 44.17584308327598, "grad_norm": 13.900660514831543, "learning_rate": 2.7912078458362013e-05, "loss": 0.8278, "step": 256750 }, { "epoch": 44.18444597384721, "grad_norm": 10.973573684692383, "learning_rate": 2.7907777013076398e-05, "loss": 0.8245, "step": 256800 }, { "epoch": 44.193048864418444, "grad_norm": 7.580682754516602, "learning_rate": 2.7903475567790775e-05, "loss": 0.8507, "step": 256850 }, { "epoch": 44.201651754989676, "grad_norm": 19.238571166992188, "learning_rate": 2.7899174122505163e-05, "loss": 0.7345, "step": 256900 }, { "epoch": 44.21025464556091, "grad_norm": 8.128317832946777, "learning_rate": 2.7894872677219548e-05, "loss": 0.8292, "step": 256950 }, { "epoch": 44.21885753613214, "grad_norm": 17.81446647644043, "learning_rate": 2.7890571231933932e-05, "loss": 0.7701, "step": 257000 }, { "epoch": 44.227460426703374, "grad_norm": 11.543777465820312, "learning_rate": 2.7886269786648317e-05, "loss": 0.745, "step": 257050 }, { "epoch": 44.236063317274606, "grad_norm": 13.831305503845215, "learning_rate": 2.78819683413627e-05, "loss": 0.7496, "step": 257100 }, { "epoch": 44.24466620784584, "grad_norm": 11.003203392028809, "learning_rate": 2.7877666896077083e-05, "loss": 0.8262, "step": 257150 }, { "epoch": 44.25326909841707, "grad_norm": 12.462993621826172, "learning_rate": 2.7873365450791467e-05, "loss": 0.7851, "step": 257200 }, { "epoch": 44.2618719889883, "grad_norm": 14.19619369506836, "learning_rate": 2.7869064005505852e-05, "loss": 0.8382, "step": 257250 }, { "epoch": 44.27047487955953, "grad_norm": 13.965798377990723, "learning_rate": 2.7864762560220236e-05, "loss": 0.758, "step": 257300 }, { "epoch": 44.27907777013076, "grad_norm": 12.834152221679688, "learning_rate": 2.786046111493462e-05, "loss": 0.8002, "step": 257350 }, { "epoch": 44.287680660701994, "grad_norm": 10.308147430419922, "learning_rate": 2.7856159669649002e-05, "loss": 0.8424, "step": 257400 }, { "epoch": 44.29628355127323, "grad_norm": 9.723011016845703, "learning_rate": 2.7851858224363387e-05, "loss": 0.8217, "step": 257450 }, { "epoch": 44.30488644184446, "grad_norm": 15.81091022491455, "learning_rate": 2.784755677907777e-05, "loss": 0.8117, "step": 257500 }, { "epoch": 44.31348933241569, "grad_norm": 17.848575592041016, "learning_rate": 2.7843255333792156e-05, "loss": 0.8606, "step": 257550 }, { "epoch": 44.322092222986925, "grad_norm": 14.904017448425293, "learning_rate": 2.783895388850654e-05, "loss": 0.8064, "step": 257600 }, { "epoch": 44.33069511355816, "grad_norm": 11.345322608947754, "learning_rate": 2.783465244322092e-05, "loss": 0.8121, "step": 257650 }, { "epoch": 44.33929800412939, "grad_norm": 14.31788158416748, "learning_rate": 2.7830350997935306e-05, "loss": 0.8021, "step": 257700 }, { "epoch": 44.34790089470062, "grad_norm": 18.190261840820312, "learning_rate": 2.782604955264969e-05, "loss": 0.8209, "step": 257750 }, { "epoch": 44.356503785271855, "grad_norm": 12.422480583190918, "learning_rate": 2.7821748107364075e-05, "loss": 0.7867, "step": 257800 }, { "epoch": 44.36510667584308, "grad_norm": 9.878108024597168, "learning_rate": 2.781744666207846e-05, "loss": 0.8167, "step": 257850 }, { "epoch": 44.37370956641431, "grad_norm": 14.627795219421387, "learning_rate": 2.7813145216792848e-05, "loss": 0.7916, "step": 257900 }, { "epoch": 44.382312456985545, "grad_norm": 11.443700790405273, "learning_rate": 2.7808843771507225e-05, "loss": 0.8225, "step": 257950 }, { "epoch": 44.39091534755678, "grad_norm": 12.913394927978516, "learning_rate": 2.780454232622161e-05, "loss": 0.7387, "step": 258000 }, { "epoch": 44.39951823812801, "grad_norm": 16.87530517578125, "learning_rate": 2.7800240880935995e-05, "loss": 0.7909, "step": 258050 }, { "epoch": 44.40812112869924, "grad_norm": 22.117504119873047, "learning_rate": 2.7795939435650383e-05, "loss": 0.8524, "step": 258100 }, { "epoch": 44.416724019270475, "grad_norm": 10.8303861618042, "learning_rate": 2.7791637990364767e-05, "loss": 0.8028, "step": 258150 }, { "epoch": 44.42532690984171, "grad_norm": 14.847999572753906, "learning_rate": 2.7787336545079145e-05, "loss": 0.8266, "step": 258200 }, { "epoch": 44.43392980041294, "grad_norm": 9.76375961303711, "learning_rate": 2.778303509979353e-05, "loss": 0.8469, "step": 258250 }, { "epoch": 44.44253269098417, "grad_norm": 25.04828453063965, "learning_rate": 2.7778733654507917e-05, "loss": 0.8066, "step": 258300 }, { "epoch": 44.451135581555405, "grad_norm": 24.766525268554688, "learning_rate": 2.7774432209222302e-05, "loss": 0.7886, "step": 258350 }, { "epoch": 44.45973847212664, "grad_norm": 10.973868370056152, "learning_rate": 2.7770130763936686e-05, "loss": 0.756, "step": 258400 }, { "epoch": 44.46834136269786, "grad_norm": 12.032751083374023, "learning_rate": 2.7765829318651064e-05, "loss": 0.8066, "step": 258450 }, { "epoch": 44.476944253269096, "grad_norm": 10.552319526672363, "learning_rate": 2.776152787336545e-05, "loss": 0.7827, "step": 258500 }, { "epoch": 44.48554714384033, "grad_norm": 9.626855850219727, "learning_rate": 2.7757226428079837e-05, "loss": 0.7769, "step": 258550 }, { "epoch": 44.49415003441156, "grad_norm": 17.414167404174805, "learning_rate": 2.775292498279422e-05, "loss": 0.7972, "step": 258600 }, { "epoch": 44.502752924982794, "grad_norm": 12.002792358398438, "learning_rate": 2.7748623537508606e-05, "loss": 0.8221, "step": 258650 }, { "epoch": 44.511355815554026, "grad_norm": 12.92994499206543, "learning_rate": 2.774432209222299e-05, "loss": 0.7503, "step": 258700 }, { "epoch": 44.51995870612526, "grad_norm": 14.225043296813965, "learning_rate": 2.774002064693737e-05, "loss": 0.8489, "step": 258750 }, { "epoch": 44.52856159669649, "grad_norm": 16.947477340698242, "learning_rate": 2.7735719201651756e-05, "loss": 0.7974, "step": 258800 }, { "epoch": 44.537164487267724, "grad_norm": 13.888713836669922, "learning_rate": 2.773141775636614e-05, "loss": 0.8019, "step": 258850 }, { "epoch": 44.545767377838956, "grad_norm": 11.535835266113281, "learning_rate": 2.7727116311080525e-05, "loss": 0.8064, "step": 258900 }, { "epoch": 44.55437026841019, "grad_norm": 6.5084099769592285, "learning_rate": 2.772281486579491e-05, "loss": 0.8358, "step": 258950 }, { "epoch": 44.56297315898142, "grad_norm": 14.43639850616455, "learning_rate": 2.771851342050929e-05, "loss": 0.8323, "step": 259000 }, { "epoch": 44.57157604955265, "grad_norm": 20.45232582092285, "learning_rate": 2.7714211975223676e-05, "loss": 0.8033, "step": 259050 }, { "epoch": 44.58017894012388, "grad_norm": 13.17978286743164, "learning_rate": 2.770991052993806e-05, "loss": 0.8012, "step": 259100 }, { "epoch": 44.58878183069511, "grad_norm": 11.624163627624512, "learning_rate": 2.7705609084652445e-05, "loss": 0.8078, "step": 259150 }, { "epoch": 44.597384721266344, "grad_norm": 10.039764404296875, "learning_rate": 2.770130763936683e-05, "loss": 0.8557, "step": 259200 }, { "epoch": 44.60598761183758, "grad_norm": 17.248140335083008, "learning_rate": 2.7697006194081214e-05, "loss": 0.7694, "step": 259250 }, { "epoch": 44.61459050240881, "grad_norm": 17.450496673583984, "learning_rate": 2.7692704748795595e-05, "loss": 0.8075, "step": 259300 }, { "epoch": 44.62319339298004, "grad_norm": 9.468414306640625, "learning_rate": 2.768840330350998e-05, "loss": 0.7809, "step": 259350 }, { "epoch": 44.631796283551274, "grad_norm": 8.625378608703613, "learning_rate": 2.7684101858224364e-05, "loss": 0.7485, "step": 259400 }, { "epoch": 44.64039917412251, "grad_norm": 11.803553581237793, "learning_rate": 2.767980041293875e-05, "loss": 0.7984, "step": 259450 }, { "epoch": 44.64900206469374, "grad_norm": 11.637051582336426, "learning_rate": 2.7675498967653137e-05, "loss": 0.7946, "step": 259500 }, { "epoch": 44.65760495526497, "grad_norm": 10.588994026184082, "learning_rate": 2.7671197522367514e-05, "loss": 0.8201, "step": 259550 }, { "epoch": 44.6662078458362, "grad_norm": 12.679435729980469, "learning_rate": 2.76668960770819e-05, "loss": 0.8051, "step": 259600 }, { "epoch": 44.67481073640743, "grad_norm": 15.506795883178711, "learning_rate": 2.7662594631796284e-05, "loss": 0.8068, "step": 259650 }, { "epoch": 44.68341362697866, "grad_norm": 17.240476608276367, "learning_rate": 2.7658293186510668e-05, "loss": 0.8017, "step": 259700 }, { "epoch": 44.692016517549895, "grad_norm": 11.776106834411621, "learning_rate": 2.7653991741225056e-05, "loss": 0.7362, "step": 259750 }, { "epoch": 44.70061940812113, "grad_norm": 11.234804153442383, "learning_rate": 2.7649690295939434e-05, "loss": 0.7977, "step": 259800 }, { "epoch": 44.70922229869236, "grad_norm": 10.450583457946777, "learning_rate": 2.764538885065382e-05, "loss": 0.8154, "step": 259850 }, { "epoch": 44.71782518926359, "grad_norm": 13.47439193725586, "learning_rate": 2.7641087405368203e-05, "loss": 0.8323, "step": 259900 }, { "epoch": 44.726428079834825, "grad_norm": 13.245407104492188, "learning_rate": 2.763678596008259e-05, "loss": 0.7596, "step": 259950 }, { "epoch": 44.73503097040606, "grad_norm": 9.632479667663574, "learning_rate": 2.7632484514796975e-05, "loss": 0.7928, "step": 260000 }, { "epoch": 44.74363386097729, "grad_norm": 15.03664779663086, "learning_rate": 2.762818306951136e-05, "loss": 0.7945, "step": 260050 }, { "epoch": 44.75223675154852, "grad_norm": 14.854204177856445, "learning_rate": 2.7623881624225738e-05, "loss": 0.7893, "step": 260100 }, { "epoch": 44.760839642119755, "grad_norm": 11.479286193847656, "learning_rate": 2.7619580178940126e-05, "loss": 0.7928, "step": 260150 }, { "epoch": 44.76944253269098, "grad_norm": 10.24255657196045, "learning_rate": 2.761527873365451e-05, "loss": 0.8522, "step": 260200 }, { "epoch": 44.77804542326221, "grad_norm": 10.954974174499512, "learning_rate": 2.7610977288368895e-05, "loss": 0.8284, "step": 260250 }, { "epoch": 44.786648313833446, "grad_norm": 15.391246795654297, "learning_rate": 2.760667584308328e-05, "loss": 0.8404, "step": 260300 }, { "epoch": 44.79525120440468, "grad_norm": 12.641975402832031, "learning_rate": 2.760237439779766e-05, "loss": 0.8099, "step": 260350 }, { "epoch": 44.80385409497591, "grad_norm": 15.023813247680664, "learning_rate": 2.7598072952512045e-05, "loss": 0.8024, "step": 260400 }, { "epoch": 44.81245698554714, "grad_norm": 14.565075874328613, "learning_rate": 2.759377150722643e-05, "loss": 0.8269, "step": 260450 }, { "epoch": 44.821059876118376, "grad_norm": 14.406018257141113, "learning_rate": 2.7589470061940814e-05, "loss": 0.7998, "step": 260500 }, { "epoch": 44.82966276668961, "grad_norm": 14.130977630615234, "learning_rate": 2.75851686166552e-05, "loss": 0.8108, "step": 260550 }, { "epoch": 44.83826565726084, "grad_norm": 15.683358192443848, "learning_rate": 2.758086717136958e-05, "loss": 0.7699, "step": 260600 }, { "epoch": 44.846868547832074, "grad_norm": 9.746614456176758, "learning_rate": 2.7576565726083965e-05, "loss": 0.7547, "step": 260650 }, { "epoch": 44.855471438403306, "grad_norm": 15.207605361938477, "learning_rate": 2.757226428079835e-05, "loss": 0.8176, "step": 260700 }, { "epoch": 44.86407432897454, "grad_norm": 36.2233772277832, "learning_rate": 2.7567962835512734e-05, "loss": 0.7967, "step": 260750 }, { "epoch": 44.872677219545764, "grad_norm": 16.6459903717041, "learning_rate": 2.7563661390227118e-05, "loss": 0.8118, "step": 260800 }, { "epoch": 44.881280110117, "grad_norm": 9.180767059326172, "learning_rate": 2.7559359944941503e-05, "loss": 0.7858, "step": 260850 }, { "epoch": 44.88988300068823, "grad_norm": 17.16820526123047, "learning_rate": 2.7555058499655884e-05, "loss": 0.7725, "step": 260900 }, { "epoch": 44.89848589125946, "grad_norm": 17.9899845123291, "learning_rate": 2.755075705437027e-05, "loss": 0.8889, "step": 260950 }, { "epoch": 44.907088781830694, "grad_norm": 25.459726333618164, "learning_rate": 2.7546455609084653e-05, "loss": 0.7977, "step": 261000 }, { "epoch": 44.91569167240193, "grad_norm": 8.369806289672852, "learning_rate": 2.7542154163799038e-05, "loss": 0.8096, "step": 261050 }, { "epoch": 44.92429456297316, "grad_norm": 14.081635475158691, "learning_rate": 2.7537852718513422e-05, "loss": 0.7796, "step": 261100 }, { "epoch": 44.93289745354439, "grad_norm": 11.146088600158691, "learning_rate": 2.7533551273227803e-05, "loss": 0.8425, "step": 261150 }, { "epoch": 44.941500344115624, "grad_norm": 18.07229995727539, "learning_rate": 2.7529249827942188e-05, "loss": 0.7411, "step": 261200 }, { "epoch": 44.95010323468686, "grad_norm": 13.901552200317383, "learning_rate": 2.7524948382656573e-05, "loss": 0.814, "step": 261250 }, { "epoch": 44.95870612525809, "grad_norm": 5.994570732116699, "learning_rate": 2.7520646937370957e-05, "loss": 0.8793, "step": 261300 }, { "epoch": 44.96730901582932, "grad_norm": 9.566688537597656, "learning_rate": 2.7516345492085345e-05, "loss": 0.8569, "step": 261350 }, { "epoch": 44.97591190640055, "grad_norm": 8.33790397644043, "learning_rate": 2.751204404679973e-05, "loss": 0.8264, "step": 261400 }, { "epoch": 44.98451479697178, "grad_norm": 9.187797546386719, "learning_rate": 2.7507742601514107e-05, "loss": 0.7958, "step": 261450 }, { "epoch": 44.99311768754301, "grad_norm": 7.39698600769043, "learning_rate": 2.7503441156228492e-05, "loss": 0.8172, "step": 261500 }, { "epoch": 45.0, "eval_accuracy": 0.5096368955429358, "eval_f1": 0.5039144979514093, "eval_f1_DuraRiadoRio_16x16": 0.3718073068218558, "eval_f1_Mole_16x16": 0.5684261974584556, "eval_f1_Quebrado_16x16": 0.6721649484536083, "eval_f1_RiadoRio_16x16": 0.40266311584553927, "eval_f1_RioFechado_16x16": 0.5045109211775879, "eval_loss": 2.459604263305664, "eval_precision": 0.6247622194855682, "eval_precision_DuraRiadoRio_16x16": 0.7287705956907478, "eval_precision_Mole_16x16": 0.6504474272930649, "eval_precision_Quebrado_16x16": 0.8274111675126904, "eval_precision_RiadoRio_16x16": 0.5714285714285714, "eval_precision_RioFechado_16x16": 0.34575333550276605, "eval_recall": 0.5128007168106834, "eval_recall_DuraRiadoRio_16x16": 0.2495659722222222, "eval_recall_Mole_16x16": 0.5047743055555556, "eval_recall_Quebrado_16x16": 0.5659722222222222, "eval_recall_RiadoRio_16x16": 0.31085526315789475, "eval_recall_RioFechado_16x16": 0.9328358208955224, "eval_runtime": 46.3002, "eval_samples_per_second": 251.014, "eval_steps_per_second": 15.702, "step": 261540 }, { "epoch": 45.001720578114245, "grad_norm": 16.977060317993164, "learning_rate": 2.749913971094288e-05, "loss": 0.8088, "step": 261550 }, { "epoch": 45.01032346868548, "grad_norm": 5.879305839538574, "learning_rate": 2.7494838265657264e-05, "loss": 0.7925, "step": 261600 }, { "epoch": 45.01892635925671, "grad_norm": 7.5443196296691895, "learning_rate": 2.749053682037165e-05, "loss": 0.8105, "step": 261650 }, { "epoch": 45.02752924982794, "grad_norm": 15.130269050598145, "learning_rate": 2.7486235375086027e-05, "loss": 0.8393, "step": 261700 }, { "epoch": 45.036132140399175, "grad_norm": 17.010271072387695, "learning_rate": 2.748193392980041e-05, "loss": 0.7626, "step": 261750 }, { "epoch": 45.04473503097041, "grad_norm": 10.153508186340332, "learning_rate": 2.74776324845148e-05, "loss": 0.7733, "step": 261800 }, { "epoch": 45.05333792154164, "grad_norm": 16.41318702697754, "learning_rate": 2.7473331039229184e-05, "loss": 0.83, "step": 261850 }, { "epoch": 45.06194081211287, "grad_norm": 13.84674072265625, "learning_rate": 2.746902959394357e-05, "loss": 0.7822, "step": 261900 }, { "epoch": 45.070543702684105, "grad_norm": 15.560593605041504, "learning_rate": 2.7464728148657946e-05, "loss": 0.8282, "step": 261950 }, { "epoch": 45.07914659325533, "grad_norm": 12.96896743774414, "learning_rate": 2.7460426703372334e-05, "loss": 0.8288, "step": 262000 }, { "epoch": 45.08774948382656, "grad_norm": 14.698625564575195, "learning_rate": 2.745612525808672e-05, "loss": 0.8139, "step": 262050 }, { "epoch": 45.096352374397796, "grad_norm": 10.719035148620605, "learning_rate": 2.7451823812801103e-05, "loss": 0.8138, "step": 262100 }, { "epoch": 45.10495526496903, "grad_norm": 12.203104019165039, "learning_rate": 2.7447522367515488e-05, "loss": 0.7987, "step": 262150 }, { "epoch": 45.11355815554026, "grad_norm": 18.374061584472656, "learning_rate": 2.7443220922229872e-05, "loss": 0.8219, "step": 262200 }, { "epoch": 45.12216104611149, "grad_norm": 13.925897598266602, "learning_rate": 2.7438919476944254e-05, "loss": 0.8464, "step": 262250 }, { "epoch": 45.130763936682726, "grad_norm": 13.558019638061523, "learning_rate": 2.7434618031658638e-05, "loss": 0.8324, "step": 262300 }, { "epoch": 45.13936682725396, "grad_norm": 13.528894424438477, "learning_rate": 2.7430316586373023e-05, "loss": 0.7912, "step": 262350 }, { "epoch": 45.14796971782519, "grad_norm": 16.181827545166016, "learning_rate": 2.7426015141087407e-05, "loss": 0.8448, "step": 262400 }, { "epoch": 45.156572608396424, "grad_norm": 16.694812774658203, "learning_rate": 2.7421713695801792e-05, "loss": 0.7933, "step": 262450 }, { "epoch": 45.165175498967656, "grad_norm": 14.695483207702637, "learning_rate": 2.7417412250516173e-05, "loss": 0.8231, "step": 262500 }, { "epoch": 45.17377838953888, "grad_norm": 11.522321701049805, "learning_rate": 2.7413110805230558e-05, "loss": 0.7299, "step": 262550 }, { "epoch": 45.182381280110114, "grad_norm": 10.51090145111084, "learning_rate": 2.7408809359944942e-05, "loss": 0.7555, "step": 262600 }, { "epoch": 45.19098417068135, "grad_norm": 13.615876197814941, "learning_rate": 2.7404507914659327e-05, "loss": 0.7977, "step": 262650 }, { "epoch": 45.19958706125258, "grad_norm": 11.07104778289795, "learning_rate": 2.740020646937371e-05, "loss": 0.8103, "step": 262700 }, { "epoch": 45.20818995182381, "grad_norm": 9.400008201599121, "learning_rate": 2.73959050240881e-05, "loss": 0.7454, "step": 262750 }, { "epoch": 45.216792842395044, "grad_norm": 18.13142204284668, "learning_rate": 2.7391603578802477e-05, "loss": 0.7927, "step": 262800 }, { "epoch": 45.22539573296628, "grad_norm": 8.216788291931152, "learning_rate": 2.738730213351686e-05, "loss": 0.8451, "step": 262850 }, { "epoch": 45.23399862353751, "grad_norm": 13.800544738769531, "learning_rate": 2.7383000688231246e-05, "loss": 0.7714, "step": 262900 }, { "epoch": 45.24260151410874, "grad_norm": 19.238632202148438, "learning_rate": 2.7378699242945634e-05, "loss": 0.7965, "step": 262950 }, { "epoch": 45.251204404679974, "grad_norm": 11.059441566467285, "learning_rate": 2.737439779766002e-05, "loss": 0.7758, "step": 263000 }, { "epoch": 45.25980729525121, "grad_norm": 18.435331344604492, "learning_rate": 2.7370096352374396e-05, "loss": 0.8542, "step": 263050 }, { "epoch": 45.26841018582244, "grad_norm": 13.4454984664917, "learning_rate": 2.736579490708878e-05, "loss": 0.769, "step": 263100 }, { "epoch": 45.277013076393665, "grad_norm": 9.064343452453613, "learning_rate": 2.7361493461803166e-05, "loss": 0.8442, "step": 263150 }, { "epoch": 45.2856159669649, "grad_norm": 14.679962158203125, "learning_rate": 2.7357192016517553e-05, "loss": 0.8021, "step": 263200 }, { "epoch": 45.29421885753613, "grad_norm": 16.221223831176758, "learning_rate": 2.7352890571231938e-05, "loss": 0.7911, "step": 263250 }, { "epoch": 45.30282174810736, "grad_norm": 15.079080581665039, "learning_rate": 2.7348589125946316e-05, "loss": 0.774, "step": 263300 }, { "epoch": 45.311424638678595, "grad_norm": 8.728806495666504, "learning_rate": 2.73442876806607e-05, "loss": 0.8272, "step": 263350 }, { "epoch": 45.32002752924983, "grad_norm": 6.589792251586914, "learning_rate": 2.733998623537509e-05, "loss": 0.8527, "step": 263400 }, { "epoch": 45.32863041982106, "grad_norm": 9.896239280700684, "learning_rate": 2.7335684790089473e-05, "loss": 0.8008, "step": 263450 }, { "epoch": 45.33723331039229, "grad_norm": 15.350398063659668, "learning_rate": 2.7331383344803857e-05, "loss": 0.7886, "step": 263500 }, { "epoch": 45.345836200963525, "grad_norm": 11.568831443786621, "learning_rate": 2.7327081899518242e-05, "loss": 0.7781, "step": 263550 }, { "epoch": 45.35443909153476, "grad_norm": 14.970113754272461, "learning_rate": 2.7322780454232623e-05, "loss": 0.7905, "step": 263600 }, { "epoch": 45.36304198210599, "grad_norm": 6.305039405822754, "learning_rate": 2.7318479008947008e-05, "loss": 0.8218, "step": 263650 }, { "epoch": 45.37164487267722, "grad_norm": 12.415478706359863, "learning_rate": 2.7314177563661392e-05, "loss": 0.8306, "step": 263700 }, { "epoch": 45.38024776324845, "grad_norm": 10.723408699035645, "learning_rate": 2.7309876118375777e-05, "loss": 0.782, "step": 263750 }, { "epoch": 45.38885065381968, "grad_norm": 9.64030647277832, "learning_rate": 2.730557467309016e-05, "loss": 0.8241, "step": 263800 }, { "epoch": 45.39745354439091, "grad_norm": 14.421387672424316, "learning_rate": 2.7301273227804543e-05, "loss": 0.8008, "step": 263850 }, { "epoch": 45.406056434962146, "grad_norm": 14.95654296875, "learning_rate": 2.7296971782518927e-05, "loss": 0.7836, "step": 263900 }, { "epoch": 45.41465932553338, "grad_norm": 12.098333358764648, "learning_rate": 2.7292670337233312e-05, "loss": 0.7678, "step": 263950 }, { "epoch": 45.42326221610461, "grad_norm": 15.926487922668457, "learning_rate": 2.7288368891947696e-05, "loss": 0.7604, "step": 264000 }, { "epoch": 45.43186510667584, "grad_norm": 14.918033599853516, "learning_rate": 2.728406744666208e-05, "loss": 0.8493, "step": 264050 }, { "epoch": 45.440467997247076, "grad_norm": 13.921390533447266, "learning_rate": 2.7279766001376462e-05, "loss": 0.8065, "step": 264100 }, { "epoch": 45.44907088781831, "grad_norm": 10.108048439025879, "learning_rate": 2.7275464556090847e-05, "loss": 0.8129, "step": 264150 }, { "epoch": 45.45767377838954, "grad_norm": 14.158368110656738, "learning_rate": 2.727116311080523e-05, "loss": 0.7744, "step": 264200 }, { "epoch": 45.466276668960774, "grad_norm": 13.188648223876953, "learning_rate": 2.7266861665519616e-05, "loss": 0.7987, "step": 264250 }, { "epoch": 45.474879559532006, "grad_norm": 6.623232841491699, "learning_rate": 2.7262560220234e-05, "loss": 0.7871, "step": 264300 }, { "epoch": 45.48348245010323, "grad_norm": 9.24501895904541, "learning_rate": 2.7258258774948385e-05, "loss": 0.8472, "step": 264350 }, { "epoch": 45.492085340674464, "grad_norm": 6.605987548828125, "learning_rate": 2.7253957329662766e-05, "loss": 0.8014, "step": 264400 }, { "epoch": 45.5006882312457, "grad_norm": 6.750746250152588, "learning_rate": 2.724965588437715e-05, "loss": 0.893, "step": 264450 }, { "epoch": 45.50929112181693, "grad_norm": 9.790690422058105, "learning_rate": 2.7245354439091535e-05, "loss": 0.7962, "step": 264500 }, { "epoch": 45.51789401238816, "grad_norm": 13.630228996276855, "learning_rate": 2.724105299380592e-05, "loss": 0.7582, "step": 264550 }, { "epoch": 45.526496902959394, "grad_norm": 14.108724594116211, "learning_rate": 2.7236751548520308e-05, "loss": 0.7368, "step": 264600 }, { "epoch": 45.53509979353063, "grad_norm": 11.37462043762207, "learning_rate": 2.7232450103234685e-05, "loss": 0.7886, "step": 264650 }, { "epoch": 45.54370268410186, "grad_norm": 14.141410827636719, "learning_rate": 2.722814865794907e-05, "loss": 0.7975, "step": 264700 }, { "epoch": 45.55230557467309, "grad_norm": 11.474081039428711, "learning_rate": 2.7223847212663455e-05, "loss": 0.7803, "step": 264750 }, { "epoch": 45.560908465244324, "grad_norm": 11.858009338378906, "learning_rate": 2.7219545767377842e-05, "loss": 0.8047, "step": 264800 }, { "epoch": 45.56951135581556, "grad_norm": 7.280492305755615, "learning_rate": 2.7215244322092227e-05, "loss": 0.7687, "step": 264850 }, { "epoch": 45.57811424638679, "grad_norm": 12.19548511505127, "learning_rate": 2.721094287680661e-05, "loss": 0.8128, "step": 264900 }, { "epoch": 45.586717136958015, "grad_norm": 10.286602020263672, "learning_rate": 2.720664143152099e-05, "loss": 0.7654, "step": 264950 }, { "epoch": 45.59532002752925, "grad_norm": 14.160049438476562, "learning_rate": 2.7202339986235374e-05, "loss": 0.7354, "step": 265000 }, { "epoch": 45.60392291810048, "grad_norm": 11.410614013671875, "learning_rate": 2.7198038540949762e-05, "loss": 0.8055, "step": 265050 }, { "epoch": 45.61252580867171, "grad_norm": 13.123403549194336, "learning_rate": 2.7193737095664146e-05, "loss": 0.7738, "step": 265100 }, { "epoch": 45.621128699242945, "grad_norm": 18.473812103271484, "learning_rate": 2.718943565037853e-05, "loss": 0.7505, "step": 265150 }, { "epoch": 45.62973158981418, "grad_norm": 23.577762603759766, "learning_rate": 2.718513420509291e-05, "loss": 0.7688, "step": 265200 }, { "epoch": 45.63833448038541, "grad_norm": 12.706047058105469, "learning_rate": 2.7180832759807297e-05, "loss": 0.7874, "step": 265250 }, { "epoch": 45.64693737095664, "grad_norm": 15.127314567565918, "learning_rate": 2.717653131452168e-05, "loss": 0.8288, "step": 265300 }, { "epoch": 45.655540261527875, "grad_norm": 18.09365463256836, "learning_rate": 2.7172229869236066e-05, "loss": 0.8444, "step": 265350 }, { "epoch": 45.66414315209911, "grad_norm": 12.514717102050781, "learning_rate": 2.716792842395045e-05, "loss": 0.7756, "step": 265400 }, { "epoch": 45.67274604267034, "grad_norm": 9.755219459533691, "learning_rate": 2.716362697866483e-05, "loss": 0.8006, "step": 265450 }, { "epoch": 45.681348933241566, "grad_norm": 16.951171875, "learning_rate": 2.7159325533379216e-05, "loss": 0.8312, "step": 265500 }, { "epoch": 45.6899518238128, "grad_norm": 12.058581352233887, "learning_rate": 2.71550240880936e-05, "loss": 0.7712, "step": 265550 }, { "epoch": 45.69855471438403, "grad_norm": 20.24626350402832, "learning_rate": 2.7150722642807985e-05, "loss": 0.8122, "step": 265600 }, { "epoch": 45.70715760495526, "grad_norm": 17.254226684570312, "learning_rate": 2.714642119752237e-05, "loss": 0.8048, "step": 265650 }, { "epoch": 45.715760495526496, "grad_norm": 12.626811981201172, "learning_rate": 2.7142119752236754e-05, "loss": 0.7877, "step": 265700 }, { "epoch": 45.72436338609773, "grad_norm": 13.627229690551758, "learning_rate": 2.7137818306951136e-05, "loss": 0.8095, "step": 265750 }, { "epoch": 45.73296627666896, "grad_norm": 8.40965747833252, "learning_rate": 2.713351686166552e-05, "loss": 0.8172, "step": 265800 }, { "epoch": 45.74156916724019, "grad_norm": 8.178217887878418, "learning_rate": 2.7129215416379905e-05, "loss": 0.826, "step": 265850 }, { "epoch": 45.750172057811426, "grad_norm": 15.586624145507812, "learning_rate": 2.712491397109429e-05, "loss": 0.8427, "step": 265900 }, { "epoch": 45.75877494838266, "grad_norm": 9.394388198852539, "learning_rate": 2.7120612525808674e-05, "loss": 0.8065, "step": 265950 }, { "epoch": 45.76737783895389, "grad_norm": 8.810588836669922, "learning_rate": 2.7116311080523055e-05, "loss": 0.8395, "step": 266000 }, { "epoch": 45.77598072952512, "grad_norm": 12.46452522277832, "learning_rate": 2.711200963523744e-05, "loss": 0.8219, "step": 266050 }, { "epoch": 45.78458362009635, "grad_norm": 12.874374389648438, "learning_rate": 2.7107708189951824e-05, "loss": 0.7485, "step": 266100 }, { "epoch": 45.79318651066758, "grad_norm": 12.150903701782227, "learning_rate": 2.710340674466621e-05, "loss": 0.8213, "step": 266150 }, { "epoch": 45.801789401238814, "grad_norm": 11.843587875366211, "learning_rate": 2.7099105299380597e-05, "loss": 0.7545, "step": 266200 }, { "epoch": 45.81039229181005, "grad_norm": 12.510703086853027, "learning_rate": 2.7094803854094974e-05, "loss": 0.8358, "step": 266250 }, { "epoch": 45.81899518238128, "grad_norm": 22.222078323364258, "learning_rate": 2.709050240880936e-05, "loss": 0.746, "step": 266300 }, { "epoch": 45.82759807295251, "grad_norm": 11.072065353393555, "learning_rate": 2.7086200963523744e-05, "loss": 0.7566, "step": 266350 }, { "epoch": 45.836200963523744, "grad_norm": 13.033350944519043, "learning_rate": 2.7081899518238128e-05, "loss": 0.8159, "step": 266400 }, { "epoch": 45.84480385409498, "grad_norm": 13.12066650390625, "learning_rate": 2.7077598072952516e-05, "loss": 0.7593, "step": 266450 }, { "epoch": 45.85340674466621, "grad_norm": 12.088726997375488, "learning_rate": 2.70732966276669e-05, "loss": 0.8795, "step": 266500 }, { "epoch": 45.86200963523744, "grad_norm": 14.036001205444336, "learning_rate": 2.706899518238128e-05, "loss": 0.8228, "step": 266550 }, { "epoch": 45.870612525808674, "grad_norm": 16.964744567871094, "learning_rate": 2.7064693737095663e-05, "loss": 0.7845, "step": 266600 }, { "epoch": 45.87921541637991, "grad_norm": 11.934040069580078, "learning_rate": 2.706039229181005e-05, "loss": 0.815, "step": 266650 }, { "epoch": 45.88781830695113, "grad_norm": 14.48957347869873, "learning_rate": 2.7056090846524435e-05, "loss": 0.8221, "step": 266700 }, { "epoch": 45.896421197522365, "grad_norm": 13.456685066223145, "learning_rate": 2.705178940123882e-05, "loss": 0.8157, "step": 266750 }, { "epoch": 45.9050240880936, "grad_norm": 14.469048500061035, "learning_rate": 2.7047487955953198e-05, "loss": 0.7998, "step": 266800 }, { "epoch": 45.91362697866483, "grad_norm": 20.982789993286133, "learning_rate": 2.7043186510667586e-05, "loss": 0.7722, "step": 266850 }, { "epoch": 45.92222986923606, "grad_norm": 14.446940422058105, "learning_rate": 2.703888506538197e-05, "loss": 0.8212, "step": 266900 }, { "epoch": 45.930832759807295, "grad_norm": 16.413650512695312, "learning_rate": 2.7034583620096355e-05, "loss": 0.7624, "step": 266950 }, { "epoch": 45.93943565037853, "grad_norm": 9.234062194824219, "learning_rate": 2.703028217481074e-05, "loss": 0.8274, "step": 267000 }, { "epoch": 45.94803854094976, "grad_norm": 14.33342456817627, "learning_rate": 2.7025980729525124e-05, "loss": 0.7845, "step": 267050 }, { "epoch": 45.95664143152099, "grad_norm": 7.276498794555664, "learning_rate": 2.7021679284239505e-05, "loss": 0.7361, "step": 267100 }, { "epoch": 45.965244322092225, "grad_norm": 11.649659156799316, "learning_rate": 2.701737783895389e-05, "loss": 0.7944, "step": 267150 }, { "epoch": 45.97384721266346, "grad_norm": 8.807027816772461, "learning_rate": 2.7013076393668274e-05, "loss": 0.8034, "step": 267200 }, { "epoch": 45.98245010323469, "grad_norm": 12.348913192749023, "learning_rate": 2.700877494838266e-05, "loss": 0.8175, "step": 267250 }, { "epoch": 45.991052993805916, "grad_norm": 12.125904083251953, "learning_rate": 2.7004473503097043e-05, "loss": 0.8191, "step": 267300 }, { "epoch": 45.99965588437715, "grad_norm": 16.128494262695312, "learning_rate": 2.7000172057811425e-05, "loss": 0.7635, "step": 267350 }, { "epoch": 46.0, "eval_accuracy": 0.5456031664085356, "eval_f1": 0.5529200636639849, "eval_f1_DuraRiadoRio_16x16": 0.44585663446873414, "eval_f1_Mole_16x16": 0.5436507936507936, "eval_f1_Quebrado_16x16": 0.7154592979835698, "eval_f1_RiadoRio_16x16": 0.5012739707657234, "eval_f1_RioFechado_16x16": 0.5583596214511041, "eval_loss": 1.6403169631958008, "eval_precision": 0.6158105234769407, "eval_precision_DuraRiadoRio_16x16": 0.5380368098159509, "eval_precision_Mole_16x16": 0.6342592592592593, "eval_precision_Quebrado_16x16": 0.8388791593695272, "eval_precision_RiadoRio_16x16": 0.3719402985074627, "eval_precision_RioFechado_16x16": 0.6959370904325033, "eval_recall": 0.542947286272456, "eval_recall_DuraRiadoRio_16x16": 0.3806423611111111, "eval_recall_Mole_16x16": 0.4756944444444444, "eval_recall_Quebrado_16x16": 0.6236979166666666, "eval_recall_RiadoRio_16x16": 0.7685032894736842, "eval_recall_RioFechado_16x16": 0.466198419666374, "eval_runtime": 46.2394, "eval_samples_per_second": 251.344, "eval_steps_per_second": 15.723, "step": 267352 }, { "epoch": 46.00825877494838, "grad_norm": 11.591780662536621, "learning_rate": 2.699587061252581e-05, "loss": 0.795, "step": 267400 }, { "epoch": 46.01686166551961, "grad_norm": 10.63853931427002, "learning_rate": 2.6991569167240194e-05, "loss": 0.7892, "step": 267450 }, { "epoch": 46.025464556090846, "grad_norm": 14.908363342285156, "learning_rate": 2.6987267721954578e-05, "loss": 0.7891, "step": 267500 }, { "epoch": 46.03406744666208, "grad_norm": 11.66747760772705, "learning_rate": 2.6982966276668963e-05, "loss": 0.7784, "step": 267550 }, { "epoch": 46.04267033723331, "grad_norm": 17.485300064086914, "learning_rate": 2.6978664831383344e-05, "loss": 0.8141, "step": 267600 }, { "epoch": 46.05127322780454, "grad_norm": 7.227141857147217, "learning_rate": 2.697436338609773e-05, "loss": 0.7703, "step": 267650 }, { "epoch": 46.059876118375776, "grad_norm": 10.654823303222656, "learning_rate": 2.6970061940812113e-05, "loss": 0.7463, "step": 267700 }, { "epoch": 46.06847900894701, "grad_norm": 10.186059951782227, "learning_rate": 2.6965760495526498e-05, "loss": 0.7555, "step": 267750 }, { "epoch": 46.07708189951824, "grad_norm": 12.649618148803711, "learning_rate": 2.6961459050240882e-05, "loss": 0.7894, "step": 267800 }, { "epoch": 46.08568479008947, "grad_norm": 12.468655586242676, "learning_rate": 2.695715760495527e-05, "loss": 0.749, "step": 267850 }, { "epoch": 46.0942876806607, "grad_norm": 13.139361381530762, "learning_rate": 2.6952856159669648e-05, "loss": 0.7377, "step": 267900 }, { "epoch": 46.10289057123193, "grad_norm": 11.959507942199707, "learning_rate": 2.6948554714384033e-05, "loss": 0.766, "step": 267950 }, { "epoch": 46.111493461803164, "grad_norm": 17.312915802001953, "learning_rate": 2.6944253269098417e-05, "loss": 0.7893, "step": 268000 }, { "epoch": 46.1200963523744, "grad_norm": 11.890108108520508, "learning_rate": 2.6939951823812805e-05, "loss": 0.8015, "step": 268050 }, { "epoch": 46.12869924294563, "grad_norm": 10.838847160339355, "learning_rate": 2.693565037852719e-05, "loss": 0.7809, "step": 268100 }, { "epoch": 46.13730213351686, "grad_norm": 22.641082763671875, "learning_rate": 2.6931348933241567e-05, "loss": 0.7696, "step": 268150 }, { "epoch": 46.145905024088094, "grad_norm": 12.152515411376953, "learning_rate": 2.6927047487955952e-05, "loss": 0.8712, "step": 268200 }, { "epoch": 46.15450791465933, "grad_norm": 20.192670822143555, "learning_rate": 2.692274604267034e-05, "loss": 0.8276, "step": 268250 }, { "epoch": 46.16311080523056, "grad_norm": 16.90357780456543, "learning_rate": 2.6918444597384724e-05, "loss": 0.7759, "step": 268300 }, { "epoch": 46.17171369580179, "grad_norm": 9.060422897338867, "learning_rate": 2.691414315209911e-05, "loss": 0.7584, "step": 268350 }, { "epoch": 46.180316586373024, "grad_norm": 14.12893009185791, "learning_rate": 2.6909841706813494e-05, "loss": 0.8072, "step": 268400 }, { "epoch": 46.18891947694425, "grad_norm": 16.016984939575195, "learning_rate": 2.690554026152787e-05, "loss": 0.8027, "step": 268450 }, { "epoch": 46.19752236751548, "grad_norm": 19.17196273803711, "learning_rate": 2.690123881624226e-05, "loss": 0.8367, "step": 268500 }, { "epoch": 46.206125258086715, "grad_norm": 9.568649291992188, "learning_rate": 2.6896937370956644e-05, "loss": 0.7358, "step": 268550 }, { "epoch": 46.21472814865795, "grad_norm": 17.3360538482666, "learning_rate": 2.689263592567103e-05, "loss": 0.8056, "step": 268600 }, { "epoch": 46.22333103922918, "grad_norm": 10.042970657348633, "learning_rate": 2.6888334480385413e-05, "loss": 0.7041, "step": 268650 }, { "epoch": 46.23193392980041, "grad_norm": 12.214495658874512, "learning_rate": 2.6884033035099794e-05, "loss": 0.8204, "step": 268700 }, { "epoch": 46.240536820371645, "grad_norm": 12.924118995666504, "learning_rate": 2.687973158981418e-05, "loss": 0.8054, "step": 268750 }, { "epoch": 46.24913971094288, "grad_norm": 13.028773307800293, "learning_rate": 2.6875430144528563e-05, "loss": 0.7797, "step": 268800 }, { "epoch": 46.25774260151411, "grad_norm": 14.840530395507812, "learning_rate": 2.6871128699242948e-05, "loss": 0.7706, "step": 268850 }, { "epoch": 46.26634549208534, "grad_norm": 15.284623146057129, "learning_rate": 2.6866827253957332e-05, "loss": 0.7819, "step": 268900 }, { "epoch": 46.274948382656575, "grad_norm": 12.406807899475098, "learning_rate": 2.6862525808671714e-05, "loss": 0.7479, "step": 268950 }, { "epoch": 46.28355127322781, "grad_norm": 12.338337898254395, "learning_rate": 2.6858224363386098e-05, "loss": 0.7912, "step": 269000 }, { "epoch": 46.29215416379903, "grad_norm": 11.251130104064941, "learning_rate": 2.6853922918100483e-05, "loss": 0.804, "step": 269050 }, { "epoch": 46.300757054370266, "grad_norm": 17.137929916381836, "learning_rate": 2.6849621472814867e-05, "loss": 0.8432, "step": 269100 }, { "epoch": 46.3093599449415, "grad_norm": 12.257840156555176, "learning_rate": 2.6845320027529252e-05, "loss": 0.7628, "step": 269150 }, { "epoch": 46.31796283551273, "grad_norm": 11.691289901733398, "learning_rate": 2.6841018582243636e-05, "loss": 0.8004, "step": 269200 }, { "epoch": 46.32656572608396, "grad_norm": 13.621804237365723, "learning_rate": 2.6836717136958018e-05, "loss": 0.8281, "step": 269250 }, { "epoch": 46.335168616655196, "grad_norm": 11.119935989379883, "learning_rate": 2.6832415691672402e-05, "loss": 0.7711, "step": 269300 }, { "epoch": 46.34377150722643, "grad_norm": 13.119930267333984, "learning_rate": 2.6828114246386787e-05, "loss": 0.8326, "step": 269350 }, { "epoch": 46.35237439779766, "grad_norm": 7.272193431854248, "learning_rate": 2.682381280110117e-05, "loss": 0.7547, "step": 269400 }, { "epoch": 46.36097728836889, "grad_norm": 11.971617698669434, "learning_rate": 2.681951135581556e-05, "loss": 0.7621, "step": 269450 }, { "epoch": 46.369580178940126, "grad_norm": 17.888813018798828, "learning_rate": 2.6815209910529937e-05, "loss": 0.7852, "step": 269500 }, { "epoch": 46.37818306951136, "grad_norm": 18.71697998046875, "learning_rate": 2.681090846524432e-05, "loss": 0.7688, "step": 269550 }, { "epoch": 46.38678596008259, "grad_norm": 16.286357879638672, "learning_rate": 2.6806607019958706e-05, "loss": 0.7822, "step": 269600 }, { "epoch": 46.395388850653816, "grad_norm": 20.772804260253906, "learning_rate": 2.680230557467309e-05, "loss": 0.7816, "step": 269650 }, { "epoch": 46.40399174122505, "grad_norm": 10.472904205322266, "learning_rate": 2.679800412938748e-05, "loss": 0.7549, "step": 269700 }, { "epoch": 46.41259463179628, "grad_norm": 15.288223266601562, "learning_rate": 2.6793702684101856e-05, "loss": 0.7645, "step": 269750 }, { "epoch": 46.421197522367514, "grad_norm": 10.849883079528809, "learning_rate": 2.678940123881624e-05, "loss": 0.8136, "step": 269800 }, { "epoch": 46.42980041293875, "grad_norm": 14.905678749084473, "learning_rate": 2.6785099793530625e-05, "loss": 0.7703, "step": 269850 }, { "epoch": 46.43840330350998, "grad_norm": 16.749174118041992, "learning_rate": 2.6780798348245013e-05, "loss": 0.8237, "step": 269900 }, { "epoch": 46.44700619408121, "grad_norm": 15.175206184387207, "learning_rate": 2.6776496902959398e-05, "loss": 0.7724, "step": 269950 }, { "epoch": 46.455609084652444, "grad_norm": 9.39867115020752, "learning_rate": 2.6772195457673783e-05, "loss": 0.7603, "step": 270000 }, { "epoch": 46.46421197522368, "grad_norm": 16.441497802734375, "learning_rate": 2.676789401238816e-05, "loss": 0.8167, "step": 270050 }, { "epoch": 46.47281486579491, "grad_norm": 20.430316925048828, "learning_rate": 2.6763592567102548e-05, "loss": 0.8207, "step": 270100 }, { "epoch": 46.48141775636614, "grad_norm": 10.81107234954834, "learning_rate": 2.6759291121816933e-05, "loss": 0.821, "step": 270150 }, { "epoch": 46.490020646937374, "grad_norm": 8.503812789916992, "learning_rate": 2.6754989676531317e-05, "loss": 0.7979, "step": 270200 }, { "epoch": 46.4986235375086, "grad_norm": 17.889150619506836, "learning_rate": 2.6750688231245702e-05, "loss": 0.8579, "step": 270250 }, { "epoch": 46.50722642807983, "grad_norm": 8.535149574279785, "learning_rate": 2.674638678596008e-05, "loss": 0.8204, "step": 270300 }, { "epoch": 46.515829318651065, "grad_norm": 18.93460464477539, "learning_rate": 2.6742085340674468e-05, "loss": 0.7393, "step": 270350 }, { "epoch": 46.5244322092223, "grad_norm": 12.396068572998047, "learning_rate": 2.6737783895388852e-05, "loss": 0.8515, "step": 270400 }, { "epoch": 46.53303509979353, "grad_norm": 13.911520957946777, "learning_rate": 2.6733482450103237e-05, "loss": 0.7527, "step": 270450 }, { "epoch": 46.54163799036476, "grad_norm": 11.207242965698242, "learning_rate": 2.672918100481762e-05, "loss": 0.7829, "step": 270500 }, { "epoch": 46.550240880935995, "grad_norm": 17.82866668701172, "learning_rate": 2.6724879559532006e-05, "loss": 0.7728, "step": 270550 }, { "epoch": 46.55884377150723, "grad_norm": 17.791580200195312, "learning_rate": 2.6720578114246387e-05, "loss": 0.7703, "step": 270600 }, { "epoch": 46.56744666207846, "grad_norm": 14.990754127502441, "learning_rate": 2.671627666896077e-05, "loss": 0.8217, "step": 270650 }, { "epoch": 46.57604955264969, "grad_norm": 10.075067520141602, "learning_rate": 2.6711975223675156e-05, "loss": 0.8011, "step": 270700 }, { "epoch": 46.584652443220925, "grad_norm": 9.590933799743652, "learning_rate": 2.670767377838954e-05, "loss": 0.7928, "step": 270750 }, { "epoch": 46.59325533379216, "grad_norm": 13.380291938781738, "learning_rate": 2.6703372333103925e-05, "loss": 0.7929, "step": 270800 }, { "epoch": 46.60185822436338, "grad_norm": 8.992301940917969, "learning_rate": 2.6699070887818307e-05, "loss": 0.7825, "step": 270850 }, { "epoch": 46.610461114934616, "grad_norm": 9.309147834777832, "learning_rate": 2.669476944253269e-05, "loss": 0.7437, "step": 270900 }, { "epoch": 46.61906400550585, "grad_norm": 11.262320518493652, "learning_rate": 2.6690467997247076e-05, "loss": 0.7498, "step": 270950 }, { "epoch": 46.62766689607708, "grad_norm": 15.092366218566895, "learning_rate": 2.668616655196146e-05, "loss": 0.7431, "step": 271000 }, { "epoch": 46.63626978664831, "grad_norm": 15.826822280883789, "learning_rate": 2.6681865106675845e-05, "loss": 0.7392, "step": 271050 }, { "epoch": 46.644872677219546, "grad_norm": 10.36652660369873, "learning_rate": 2.6677563661390226e-05, "loss": 0.7334, "step": 271100 }, { "epoch": 46.65347556779078, "grad_norm": 9.910470008850098, "learning_rate": 2.667326221610461e-05, "loss": 0.789, "step": 271150 }, { "epoch": 46.66207845836201, "grad_norm": 12.385234832763672, "learning_rate": 2.6668960770818995e-05, "loss": 0.8249, "step": 271200 }, { "epoch": 46.67068134893324, "grad_norm": 11.709662437438965, "learning_rate": 2.666465932553338e-05, "loss": 0.7954, "step": 271250 }, { "epoch": 46.679284239504476, "grad_norm": 14.219642639160156, "learning_rate": 2.6660357880247768e-05, "loss": 0.8264, "step": 271300 }, { "epoch": 46.68788713007571, "grad_norm": 16.222335815429688, "learning_rate": 2.6656056434962152e-05, "loss": 0.8145, "step": 271350 }, { "epoch": 46.696490020646934, "grad_norm": 10.61889934539795, "learning_rate": 2.665175498967653e-05, "loss": 0.8683, "step": 271400 }, { "epoch": 46.705092911218166, "grad_norm": 13.015026092529297, "learning_rate": 2.6647453544390914e-05, "loss": 0.8003, "step": 271450 }, { "epoch": 46.7136958017894, "grad_norm": 5.916774272918701, "learning_rate": 2.6643152099105302e-05, "loss": 0.7673, "step": 271500 }, { "epoch": 46.72229869236063, "grad_norm": 9.331177711486816, "learning_rate": 2.6638850653819687e-05, "loss": 0.8084, "step": 271550 }, { "epoch": 46.730901582931864, "grad_norm": 11.224925994873047, "learning_rate": 2.663454920853407e-05, "loss": 0.7818, "step": 271600 }, { "epoch": 46.7395044735031, "grad_norm": 17.302288055419922, "learning_rate": 2.663024776324845e-05, "loss": 0.8005, "step": 271650 }, { "epoch": 46.74810736407433, "grad_norm": 14.933327674865723, "learning_rate": 2.6625946317962834e-05, "loss": 0.7692, "step": 271700 }, { "epoch": 46.75671025464556, "grad_norm": 12.553208351135254, "learning_rate": 2.6621644872677222e-05, "loss": 0.8059, "step": 271750 }, { "epoch": 46.765313145216794, "grad_norm": 16.76412582397461, "learning_rate": 2.6617343427391606e-05, "loss": 0.8303, "step": 271800 }, { "epoch": 46.77391603578803, "grad_norm": 15.937189102172852, "learning_rate": 2.661304198210599e-05, "loss": 0.8396, "step": 271850 }, { "epoch": 46.78251892635926, "grad_norm": 14.358351707458496, "learning_rate": 2.660874053682037e-05, "loss": 0.8094, "step": 271900 }, { "epoch": 46.79112181693049, "grad_norm": 12.345145225524902, "learning_rate": 2.6604439091534757e-05, "loss": 0.7204, "step": 271950 }, { "epoch": 46.79972470750172, "grad_norm": 13.119680404663086, "learning_rate": 2.660013764624914e-05, "loss": 0.8036, "step": 272000 }, { "epoch": 46.80832759807295, "grad_norm": 14.042450904846191, "learning_rate": 2.6595836200963526e-05, "loss": 0.8111, "step": 272050 }, { "epoch": 46.81693048864418, "grad_norm": 12.498499870300293, "learning_rate": 2.659153475567791e-05, "loss": 0.8328, "step": 272100 }, { "epoch": 46.825533379215415, "grad_norm": 17.381874084472656, "learning_rate": 2.6587233310392295e-05, "loss": 0.7609, "step": 272150 }, { "epoch": 46.83413626978665, "grad_norm": 15.65877914428711, "learning_rate": 2.6582931865106676e-05, "loss": 0.8012, "step": 272200 }, { "epoch": 46.84273916035788, "grad_norm": 13.98975944519043, "learning_rate": 2.657863041982106e-05, "loss": 0.7827, "step": 272250 }, { "epoch": 46.85134205092911, "grad_norm": 16.20604705810547, "learning_rate": 2.6574328974535445e-05, "loss": 0.7558, "step": 272300 }, { "epoch": 46.859944941500345, "grad_norm": 18.77338981628418, "learning_rate": 2.657002752924983e-05, "loss": 0.8405, "step": 272350 }, { "epoch": 46.86854783207158, "grad_norm": 10.95426082611084, "learning_rate": 2.6565726083964214e-05, "loss": 0.8244, "step": 272400 }, { "epoch": 46.87715072264281, "grad_norm": 14.892943382263184, "learning_rate": 2.6561424638678596e-05, "loss": 0.8271, "step": 272450 }, { "epoch": 46.88575361321404, "grad_norm": 16.19541358947754, "learning_rate": 2.655712319339298e-05, "loss": 0.8101, "step": 272500 }, { "epoch": 46.894356503785275, "grad_norm": 13.700130462646484, "learning_rate": 2.6552821748107365e-05, "loss": 0.7671, "step": 272550 }, { "epoch": 46.9029593943565, "grad_norm": 10.835649490356445, "learning_rate": 2.654852030282175e-05, "loss": 0.8235, "step": 272600 }, { "epoch": 46.91156228492773, "grad_norm": 13.084644317626953, "learning_rate": 2.6544218857536134e-05, "loss": 0.8122, "step": 272650 }, { "epoch": 46.920165175498965, "grad_norm": 14.81629467010498, "learning_rate": 2.6539917412250522e-05, "loss": 0.8259, "step": 272700 }, { "epoch": 46.9287680660702, "grad_norm": 17.057802200317383, "learning_rate": 2.65356159669649e-05, "loss": 0.8025, "step": 272750 }, { "epoch": 46.93737095664143, "grad_norm": 14.353205680847168, "learning_rate": 2.6531314521679284e-05, "loss": 0.8248, "step": 272800 }, { "epoch": 46.94597384721266, "grad_norm": 14.447163581848145, "learning_rate": 2.652701307639367e-05, "loss": 0.7782, "step": 272850 }, { "epoch": 46.954576737783896, "grad_norm": 11.539849281311035, "learning_rate": 2.6522711631108053e-05, "loss": 0.7867, "step": 272900 }, { "epoch": 46.96317962835513, "grad_norm": 13.750905990600586, "learning_rate": 2.651841018582244e-05, "loss": 0.8481, "step": 272950 }, { "epoch": 46.97178251892636, "grad_norm": 15.07282829284668, "learning_rate": 2.651410874053682e-05, "loss": 0.7926, "step": 273000 }, { "epoch": 46.98038540949759, "grad_norm": 11.328878402709961, "learning_rate": 2.6509807295251203e-05, "loss": 0.7977, "step": 273050 }, { "epoch": 46.988988300068826, "grad_norm": 12.39560317993164, "learning_rate": 2.6505505849965588e-05, "loss": 0.7615, "step": 273100 }, { "epoch": 46.99759119064006, "grad_norm": 16.126182556152344, "learning_rate": 2.6501204404679976e-05, "loss": 0.7885, "step": 273150 }, { "epoch": 47.0, "eval_accuracy": 0.5753742901393908, "eval_f1": 0.5723695212725405, "eval_f1_DuraRiadoRio_16x16": 0.5323865604433669, "eval_f1_Mole_16x16": 0.5036517674554485, "eval_f1_Quebrado_16x16": 0.7289848415250344, "eval_f1_RiadoRio_16x16": 0.47309312119794106, "eval_f1_RioFechado_16x16": 0.6237313157409116, "eval_loss": 1.7048436403274536, "eval_precision": 0.6148643236970693, "eval_precision_DuraRiadoRio_16x16": 0.4429394812680115, "eval_precision_Mole_16x16": 0.7703306523681859, "eval_precision_Quebrado_16x16": 0.7741463414634147, "eval_precision_RiadoRio_16x16": 0.5488599348534202, "eval_precision_RioFechado_16x16": 0.5380452085323145, "eval_recall": 0.5775241600306004, "eval_recall_DuraRiadoRio_16x16": 0.6671006944444444, "eval_recall_Mole_16x16": 0.3741319444444444, "eval_recall_Quebrado_16x16": 0.6888020833333334, "eval_recall_RiadoRio_16x16": 0.41570723684210525, "eval_recall_RioFechado_16x16": 0.7418788410886743, "eval_runtime": 45.809, "eval_samples_per_second": 253.706, "eval_steps_per_second": 15.87, "step": 273164 }, { "epoch": 47.006194081211284, "grad_norm": 8.714055061340332, "learning_rate": 2.649690295939436e-05, "loss": 0.8133, "step": 273200 }, { "epoch": 47.014796971782516, "grad_norm": 11.685221672058105, "learning_rate": 2.649260151410874e-05, "loss": 0.8147, "step": 273250 }, { "epoch": 47.02339986235375, "grad_norm": 25.44700813293457, "learning_rate": 2.6488300068823123e-05, "loss": 0.8089, "step": 273300 }, { "epoch": 47.03200275292498, "grad_norm": 21.205772399902344, "learning_rate": 2.648399862353751e-05, "loss": 0.7255, "step": 273350 }, { "epoch": 47.040605643496214, "grad_norm": 14.598151206970215, "learning_rate": 2.6479697178251895e-05, "loss": 0.7659, "step": 273400 }, { "epoch": 47.049208534067446, "grad_norm": 13.553678512573242, "learning_rate": 2.647539573296628e-05, "loss": 0.7799, "step": 273450 }, { "epoch": 47.05781142463868, "grad_norm": 7.745999813079834, "learning_rate": 2.6471094287680664e-05, "loss": 0.7843, "step": 273500 }, { "epoch": 47.06641431520991, "grad_norm": 17.394447326660156, "learning_rate": 2.6466792842395042e-05, "loss": 0.8075, "step": 273550 }, { "epoch": 47.075017205781144, "grad_norm": 13.67025089263916, "learning_rate": 2.646249139710943e-05, "loss": 0.7382, "step": 273600 }, { "epoch": 47.08362009635238, "grad_norm": 13.002673149108887, "learning_rate": 2.6458189951823815e-05, "loss": 0.7512, "step": 273650 }, { "epoch": 47.09222298692361, "grad_norm": 14.43137264251709, "learning_rate": 2.64538885065382e-05, "loss": 0.7798, "step": 273700 }, { "epoch": 47.10082587749484, "grad_norm": 10.500000953674316, "learning_rate": 2.6449587061252584e-05, "loss": 0.7943, "step": 273750 }, { "epoch": 47.10942876806607, "grad_norm": 15.98896598815918, "learning_rate": 2.6445285615966965e-05, "loss": 0.7998, "step": 273800 }, { "epoch": 47.1180316586373, "grad_norm": 14.805818557739258, "learning_rate": 2.644098417068135e-05, "loss": 0.7942, "step": 273850 }, { "epoch": 47.12663454920853, "grad_norm": 13.665179252624512, "learning_rate": 2.6436682725395734e-05, "loss": 0.8088, "step": 273900 }, { "epoch": 47.135237439779765, "grad_norm": 11.717373847961426, "learning_rate": 2.643238128011012e-05, "loss": 0.7895, "step": 273950 }, { "epoch": 47.143840330351, "grad_norm": 10.853753089904785, "learning_rate": 2.6428079834824503e-05, "loss": 0.787, "step": 274000 }, { "epoch": 47.15244322092223, "grad_norm": 10.194761276245117, "learning_rate": 2.6423778389538888e-05, "loss": 0.7692, "step": 274050 }, { "epoch": 47.16104611149346, "grad_norm": 15.102686882019043, "learning_rate": 2.641947694425327e-05, "loss": 0.7772, "step": 274100 }, { "epoch": 47.169649002064695, "grad_norm": 10.086029052734375, "learning_rate": 2.6415175498967654e-05, "loss": 0.7883, "step": 274150 }, { "epoch": 47.17825189263593, "grad_norm": 14.797948837280273, "learning_rate": 2.6410874053682038e-05, "loss": 0.7499, "step": 274200 }, { "epoch": 47.18685478320716, "grad_norm": 11.72014045715332, "learning_rate": 2.6406572608396423e-05, "loss": 0.7719, "step": 274250 }, { "epoch": 47.19545767377839, "grad_norm": 12.916595458984375, "learning_rate": 2.6402271163110807e-05, "loss": 0.7995, "step": 274300 }, { "epoch": 47.204060564349625, "grad_norm": 12.562392234802246, "learning_rate": 2.639796971782519e-05, "loss": 0.7473, "step": 274350 }, { "epoch": 47.21266345492085, "grad_norm": 16.007936477661133, "learning_rate": 2.6393668272539573e-05, "loss": 0.7861, "step": 274400 }, { "epoch": 47.22126634549208, "grad_norm": 11.142617225646973, "learning_rate": 2.6389366827253958e-05, "loss": 0.8364, "step": 274450 }, { "epoch": 47.229869236063315, "grad_norm": 12.94870662689209, "learning_rate": 2.6385065381968342e-05, "loss": 0.8071, "step": 274500 }, { "epoch": 47.23847212663455, "grad_norm": 15.241424560546875, "learning_rate": 2.638076393668273e-05, "loss": 0.7995, "step": 274550 }, { "epoch": 47.24707501720578, "grad_norm": 17.401166915893555, "learning_rate": 2.6376462491397108e-05, "loss": 0.7777, "step": 274600 }, { "epoch": 47.25567790777701, "grad_norm": 15.386534690856934, "learning_rate": 2.6372161046111492e-05, "loss": 0.7251, "step": 274650 }, { "epoch": 47.264280798348246, "grad_norm": 10.443163871765137, "learning_rate": 2.6367859600825877e-05, "loss": 0.7512, "step": 274700 }, { "epoch": 47.27288368891948, "grad_norm": 10.670332908630371, "learning_rate": 2.6363558155540265e-05, "loss": 0.8046, "step": 274750 }, { "epoch": 47.28148657949071, "grad_norm": 18.69503402709961, "learning_rate": 2.635925671025465e-05, "loss": 0.8508, "step": 274800 }, { "epoch": 47.29008947006194, "grad_norm": 10.615769386291504, "learning_rate": 2.6354955264969034e-05, "loss": 0.7576, "step": 274850 }, { "epoch": 47.298692360633176, "grad_norm": 9.125410079956055, "learning_rate": 2.6350653819683412e-05, "loss": 0.8132, "step": 274900 }, { "epoch": 47.30729525120441, "grad_norm": 10.511174201965332, "learning_rate": 2.6346352374397796e-05, "loss": 0.7856, "step": 274950 }, { "epoch": 47.315898141775634, "grad_norm": 14.420333862304688, "learning_rate": 2.6342050929112184e-05, "loss": 0.8003, "step": 275000 }, { "epoch": 47.324501032346866, "grad_norm": 11.438892364501953, "learning_rate": 2.633774948382657e-05, "loss": 0.7927, "step": 275050 }, { "epoch": 47.3331039229181, "grad_norm": 7.093317985534668, "learning_rate": 2.6333448038540953e-05, "loss": 0.7668, "step": 275100 }, { "epoch": 47.34170681348933, "grad_norm": 15.746354103088379, "learning_rate": 2.632914659325533e-05, "loss": 0.7762, "step": 275150 }, { "epoch": 47.350309704060564, "grad_norm": 10.997113227844238, "learning_rate": 2.632484514796972e-05, "loss": 0.835, "step": 275200 }, { "epoch": 47.358912594631796, "grad_norm": 16.428760528564453, "learning_rate": 2.6320543702684104e-05, "loss": 0.7964, "step": 275250 }, { "epoch": 47.36751548520303, "grad_norm": 12.952919006347656, "learning_rate": 2.631624225739849e-05, "loss": 0.8016, "step": 275300 }, { "epoch": 47.37611837577426, "grad_norm": 12.750831604003906, "learning_rate": 2.6311940812112873e-05, "loss": 0.8098, "step": 275350 }, { "epoch": 47.384721266345494, "grad_norm": 11.831620216369629, "learning_rate": 2.6307639366827254e-05, "loss": 0.7306, "step": 275400 }, { "epoch": 47.39332415691673, "grad_norm": 12.15782356262207, "learning_rate": 2.630333792154164e-05, "loss": 0.8401, "step": 275450 }, { "epoch": 47.40192704748796, "grad_norm": 14.839345932006836, "learning_rate": 2.6299036476256023e-05, "loss": 0.8253, "step": 275500 }, { "epoch": 47.410529938059184, "grad_norm": 11.433370590209961, "learning_rate": 2.6294735030970408e-05, "loss": 0.8039, "step": 275550 }, { "epoch": 47.41913282863042, "grad_norm": 9.924315452575684, "learning_rate": 2.6290433585684792e-05, "loss": 0.8196, "step": 275600 }, { "epoch": 47.42773571920165, "grad_norm": 13.87786865234375, "learning_rate": 2.6286132140399177e-05, "loss": 0.8171, "step": 275650 }, { "epoch": 47.43633860977288, "grad_norm": 11.051419258117676, "learning_rate": 2.6281830695113558e-05, "loss": 0.7718, "step": 275700 }, { "epoch": 47.444941500344115, "grad_norm": 12.311654090881348, "learning_rate": 2.6277529249827943e-05, "loss": 0.7969, "step": 275750 }, { "epoch": 47.45354439091535, "grad_norm": 10.019000053405762, "learning_rate": 2.6273227804542327e-05, "loss": 0.7378, "step": 275800 }, { "epoch": 47.46214728148658, "grad_norm": 11.456536293029785, "learning_rate": 2.6268926359256712e-05, "loss": 0.842, "step": 275850 }, { "epoch": 47.47075017205781, "grad_norm": 14.089086532592773, "learning_rate": 2.6264624913971096e-05, "loss": 0.7919, "step": 275900 }, { "epoch": 47.479353062629045, "grad_norm": 10.680676460266113, "learning_rate": 2.6260323468685477e-05, "loss": 0.787, "step": 275950 }, { "epoch": 47.48795595320028, "grad_norm": 10.06486701965332, "learning_rate": 2.6256022023399862e-05, "loss": 0.7657, "step": 276000 }, { "epoch": 47.49655884377151, "grad_norm": 13.848481178283691, "learning_rate": 2.6251720578114247e-05, "loss": 0.7921, "step": 276050 }, { "epoch": 47.50516173434274, "grad_norm": 7.741215705871582, "learning_rate": 2.624741913282863e-05, "loss": 0.8004, "step": 276100 }, { "epoch": 47.51376462491397, "grad_norm": 14.975469589233398, "learning_rate": 2.6243117687543016e-05, "loss": 0.8005, "step": 276150 }, { "epoch": 47.5223675154852, "grad_norm": 19.776792526245117, "learning_rate": 2.6238816242257404e-05, "loss": 0.8152, "step": 276200 }, { "epoch": 47.53097040605643, "grad_norm": 7.861680030822754, "learning_rate": 2.623451479697178e-05, "loss": 0.7866, "step": 276250 }, { "epoch": 47.539573296627665, "grad_norm": 8.55832576751709, "learning_rate": 2.6230213351686166e-05, "loss": 0.8273, "step": 276300 }, { "epoch": 47.5481761871989, "grad_norm": 11.441006660461426, "learning_rate": 2.622591190640055e-05, "loss": 0.8113, "step": 276350 }, { "epoch": 47.55677907777013, "grad_norm": 15.94068717956543, "learning_rate": 2.622161046111494e-05, "loss": 0.8147, "step": 276400 }, { "epoch": 47.56538196834136, "grad_norm": 12.273500442504883, "learning_rate": 2.6217309015829323e-05, "loss": 0.8147, "step": 276450 }, { "epoch": 47.573984858912596, "grad_norm": 13.214219093322754, "learning_rate": 2.62130075705437e-05, "loss": 0.7487, "step": 276500 }, { "epoch": 47.58258774948383, "grad_norm": 9.834537506103516, "learning_rate": 2.6208706125258085e-05, "loss": 0.7818, "step": 276550 }, { "epoch": 47.59119064005506, "grad_norm": 6.0050835609436035, "learning_rate": 2.6204404679972473e-05, "loss": 0.7175, "step": 276600 }, { "epoch": 47.59979353062629, "grad_norm": 10.644506454467773, "learning_rate": 2.6200103234686858e-05, "loss": 0.7953, "step": 276650 }, { "epoch": 47.608396421197526, "grad_norm": 17.703798294067383, "learning_rate": 2.6195801789401242e-05, "loss": 0.7963, "step": 276700 }, { "epoch": 47.61699931176875, "grad_norm": 8.126569747924805, "learning_rate": 2.619150034411562e-05, "loss": 0.7717, "step": 276750 }, { "epoch": 47.625602202339984, "grad_norm": 10.974896430969238, "learning_rate": 2.6187198898830008e-05, "loss": 0.8407, "step": 276800 }, { "epoch": 47.634205092911216, "grad_norm": 12.771106719970703, "learning_rate": 2.6182897453544393e-05, "loss": 0.786, "step": 276850 }, { "epoch": 47.64280798348245, "grad_norm": 14.266757011413574, "learning_rate": 2.6178596008258777e-05, "loss": 0.7746, "step": 276900 }, { "epoch": 47.65141087405368, "grad_norm": 8.820364952087402, "learning_rate": 2.6174294562973162e-05, "loss": 0.763, "step": 276950 }, { "epoch": 47.660013764624914, "grad_norm": 11.669449806213379, "learning_rate": 2.6169993117687546e-05, "loss": 0.7925, "step": 277000 }, { "epoch": 47.668616655196146, "grad_norm": 21.134084701538086, "learning_rate": 2.6165691672401928e-05, "loss": 0.7677, "step": 277050 }, { "epoch": 47.67721954576738, "grad_norm": 14.10937213897705, "learning_rate": 2.6161390227116312e-05, "loss": 0.8264, "step": 277100 }, { "epoch": 47.68582243633861, "grad_norm": 13.367321014404297, "learning_rate": 2.6157088781830697e-05, "loss": 0.7565, "step": 277150 }, { "epoch": 47.694425326909844, "grad_norm": 14.692923545837402, "learning_rate": 2.615278733654508e-05, "loss": 0.8046, "step": 277200 }, { "epoch": 47.70302821748108, "grad_norm": 12.211639404296875, "learning_rate": 2.6148485891259466e-05, "loss": 0.7829, "step": 277250 }, { "epoch": 47.71163110805231, "grad_norm": 14.144275665283203, "learning_rate": 2.6144184445973847e-05, "loss": 0.7827, "step": 277300 }, { "epoch": 47.720233998623534, "grad_norm": 10.009529113769531, "learning_rate": 2.613988300068823e-05, "loss": 0.8227, "step": 277350 }, { "epoch": 47.72883688919477, "grad_norm": 5.611095905303955, "learning_rate": 2.6135581555402616e-05, "loss": 0.7375, "step": 277400 }, { "epoch": 47.737439779766, "grad_norm": 14.001260757446289, "learning_rate": 2.6131280110117e-05, "loss": 0.7716, "step": 277450 }, { "epoch": 47.74604267033723, "grad_norm": 11.688395500183105, "learning_rate": 2.6126978664831385e-05, "loss": 0.7939, "step": 277500 }, { "epoch": 47.754645560908465, "grad_norm": 10.393745422363281, "learning_rate": 2.6122677219545766e-05, "loss": 0.7823, "step": 277550 }, { "epoch": 47.7632484514797, "grad_norm": 13.944498062133789, "learning_rate": 2.611837577426015e-05, "loss": 0.7712, "step": 277600 }, { "epoch": 47.77185134205093, "grad_norm": 17.090251922607422, "learning_rate": 2.6114074328974536e-05, "loss": 0.8425, "step": 277650 }, { "epoch": 47.78045423262216, "grad_norm": 14.876072883605957, "learning_rate": 2.610977288368892e-05, "loss": 0.8102, "step": 277700 }, { "epoch": 47.789057123193395, "grad_norm": 11.830301284790039, "learning_rate": 2.6105471438403305e-05, "loss": 0.7559, "step": 277750 }, { "epoch": 47.79766001376463, "grad_norm": 11.95892333984375, "learning_rate": 2.6101169993117693e-05, "loss": 0.7496, "step": 277800 }, { "epoch": 47.80626290433586, "grad_norm": 11.575028419494629, "learning_rate": 2.609686854783207e-05, "loss": 0.7537, "step": 277850 }, { "epoch": 47.81486579490709, "grad_norm": 10.760931015014648, "learning_rate": 2.6092567102546455e-05, "loss": 0.8378, "step": 277900 }, { "epoch": 47.82346868547832, "grad_norm": 8.586880683898926, "learning_rate": 2.608826565726084e-05, "loss": 0.7744, "step": 277950 }, { "epoch": 47.83207157604955, "grad_norm": 10.107904434204102, "learning_rate": 2.6083964211975228e-05, "loss": 0.8229, "step": 278000 }, { "epoch": 47.84067446662078, "grad_norm": 11.6121187210083, "learning_rate": 2.6079662766689612e-05, "loss": 0.7926, "step": 278050 }, { "epoch": 47.849277357192015, "grad_norm": 6.663557052612305, "learning_rate": 2.607536132140399e-05, "loss": 0.7897, "step": 278100 }, { "epoch": 47.85788024776325, "grad_norm": 12.204401969909668, "learning_rate": 2.6071059876118374e-05, "loss": 0.7995, "step": 278150 }, { "epoch": 47.86648313833448, "grad_norm": 12.369773864746094, "learning_rate": 2.606675843083276e-05, "loss": 0.813, "step": 278200 }, { "epoch": 47.87508602890571, "grad_norm": 14.605649948120117, "learning_rate": 2.6062456985547147e-05, "loss": 0.7392, "step": 278250 }, { "epoch": 47.883688919476946, "grad_norm": 13.652969360351562, "learning_rate": 2.605815554026153e-05, "loss": 0.7538, "step": 278300 }, { "epoch": 47.89229181004818, "grad_norm": 10.323820114135742, "learning_rate": 2.6053854094975916e-05, "loss": 0.8405, "step": 278350 }, { "epoch": 47.90089470061941, "grad_norm": 8.888130187988281, "learning_rate": 2.6049552649690294e-05, "loss": 0.7924, "step": 278400 }, { "epoch": 47.90949759119064, "grad_norm": 17.058727264404297, "learning_rate": 2.6045251204404682e-05, "loss": 0.7282, "step": 278450 }, { "epoch": 47.91810048176187, "grad_norm": 12.425224304199219, "learning_rate": 2.6040949759119066e-05, "loss": 0.7689, "step": 278500 }, { "epoch": 47.9267033723331, "grad_norm": 16.89198112487793, "learning_rate": 2.603664831383345e-05, "loss": 0.7855, "step": 278550 }, { "epoch": 47.935306262904334, "grad_norm": 9.782870292663574, "learning_rate": 2.6032346868547835e-05, "loss": 0.8403, "step": 278600 }, { "epoch": 47.943909153475566, "grad_norm": 6.136988639831543, "learning_rate": 2.6028045423262217e-05, "loss": 0.8086, "step": 278650 }, { "epoch": 47.9525120440468, "grad_norm": 9.506657600402832, "learning_rate": 2.60237439779766e-05, "loss": 0.7722, "step": 278700 }, { "epoch": 47.96111493461803, "grad_norm": 11.6210298538208, "learning_rate": 2.6019442532690986e-05, "loss": 0.7582, "step": 278750 }, { "epoch": 47.969717825189264, "grad_norm": 14.280275344848633, "learning_rate": 2.601514108740537e-05, "loss": 0.8466, "step": 278800 }, { "epoch": 47.978320715760496, "grad_norm": 18.89058494567871, "learning_rate": 2.6010839642119755e-05, "loss": 0.8222, "step": 278850 }, { "epoch": 47.98692360633173, "grad_norm": 16.924758911132812, "learning_rate": 2.6006538196834136e-05, "loss": 0.8248, "step": 278900 }, { "epoch": 47.99552649690296, "grad_norm": 15.620051383972168, "learning_rate": 2.600223675154852e-05, "loss": 0.8135, "step": 278950 }, { "epoch": 48.0, "eval_accuracy": 0.6211495439683359, "eval_f1": 0.6217548801996642, "eval_f1_DuraRiadoRio_16x16": 0.6151079136690647, "eval_f1_Mole_16x16": 0.6287339971550497, "eval_f1_Quebrado_16x16": 0.6936558874352012, "eval_f1_RiadoRio_16x16": 0.5200341005967605, "eval_f1_RioFechado_16x16": 0.6512425021422451, "eval_loss": 1.3005459308624268, "eval_precision": 0.6418133020327286, "eval_precision_DuraRiadoRio_16x16": 0.6380597014925373, "eval_precision_Mole_16x16": 0.6927899686520376, "eval_precision_Quebrado_16x16": 0.8042358328563252, "eval_precision_RiadoRio_16x16": 0.5398230088495575, "eval_precision_RioFechado_16x16": 0.5341579983131852, "eval_recall": 0.6229579134449014, "eval_recall_DuraRiadoRio_16x16": 0.59375, "eval_recall_Mole_16x16": 0.5755208333333334, "eval_recall_Quebrado_16x16": 0.6098090277777778, "eval_recall_RiadoRio_16x16": 0.5016447368421053, "eval_recall_RioFechado_16x16": 0.8340649692712906, "eval_runtime": 47.3002, "eval_samples_per_second": 245.707, "eval_steps_per_second": 15.37, "step": 278976 }, { "epoch": 48.004129387474194, "grad_norm": 15.859496116638184, "learning_rate": 2.5997935306262905e-05, "loss": 0.7321, "step": 279000 }, { "epoch": 48.012732278045426, "grad_norm": 17.652360916137695, "learning_rate": 2.599363386097729e-05, "loss": 0.8112, "step": 279050 }, { "epoch": 48.02133516861665, "grad_norm": 14.06562328338623, "learning_rate": 2.5989332415691674e-05, "loss": 0.7281, "step": 279100 }, { "epoch": 48.029938059187884, "grad_norm": 12.134862899780273, "learning_rate": 2.598503097040606e-05, "loss": 0.7785, "step": 279150 }, { "epoch": 48.03854094975912, "grad_norm": 11.814508438110352, "learning_rate": 2.598072952512044e-05, "loss": 0.762, "step": 279200 }, { "epoch": 48.04714384033035, "grad_norm": 9.341639518737793, "learning_rate": 2.5976428079834825e-05, "loss": 0.8572, "step": 279250 }, { "epoch": 48.05574673090158, "grad_norm": 8.86536979675293, "learning_rate": 2.597212663454921e-05, "loss": 0.7346, "step": 279300 }, { "epoch": 48.064349621472815, "grad_norm": 7.27018928527832, "learning_rate": 2.5967825189263594e-05, "loss": 0.7611, "step": 279350 }, { "epoch": 48.07295251204405, "grad_norm": 10.81411075592041, "learning_rate": 2.5963523743977978e-05, "loss": 0.742, "step": 279400 }, { "epoch": 48.08155540261528, "grad_norm": 18.806873321533203, "learning_rate": 2.595922229869236e-05, "loss": 0.8036, "step": 279450 }, { "epoch": 48.09015829318651, "grad_norm": 13.275239944458008, "learning_rate": 2.5954920853406744e-05, "loss": 0.7542, "step": 279500 }, { "epoch": 48.098761183757745, "grad_norm": 12.75310230255127, "learning_rate": 2.595061940812113e-05, "loss": 0.8277, "step": 279550 }, { "epoch": 48.10736407432898, "grad_norm": 14.795042991638184, "learning_rate": 2.5946317962835513e-05, "loss": 0.7503, "step": 279600 }, { "epoch": 48.11596696490021, "grad_norm": 15.923666000366211, "learning_rate": 2.59420165175499e-05, "loss": 0.8186, "step": 279650 }, { "epoch": 48.124569855471435, "grad_norm": 8.768579483032227, "learning_rate": 2.5937715072264286e-05, "loss": 0.7815, "step": 279700 }, { "epoch": 48.13317274604267, "grad_norm": 24.577373504638672, "learning_rate": 2.5933413626978663e-05, "loss": 0.8281, "step": 279750 }, { "epoch": 48.1417756366139, "grad_norm": 11.783133506774902, "learning_rate": 2.5929112181693048e-05, "loss": 0.7617, "step": 279800 }, { "epoch": 48.15037852718513, "grad_norm": 14.139538764953613, "learning_rate": 2.5924810736407436e-05, "loss": 0.8012, "step": 279850 }, { "epoch": 48.158981417756365, "grad_norm": 11.672821044921875, "learning_rate": 2.592050929112182e-05, "loss": 0.7539, "step": 279900 }, { "epoch": 48.1675843083276, "grad_norm": 15.971484184265137, "learning_rate": 2.5916207845836205e-05, "loss": 0.7795, "step": 279950 }, { "epoch": 48.17618719889883, "grad_norm": 14.96009635925293, "learning_rate": 2.5911906400550583e-05, "loss": 0.7865, "step": 280000 }, { "epoch": 48.18479008947006, "grad_norm": 17.78426170349121, "learning_rate": 2.590760495526497e-05, "loss": 0.8094, "step": 280050 }, { "epoch": 48.193392980041295, "grad_norm": 16.16896629333496, "learning_rate": 2.5903303509979355e-05, "loss": 0.7537, "step": 280100 }, { "epoch": 48.20199587061253, "grad_norm": 19.274850845336914, "learning_rate": 2.589900206469374e-05, "loss": 0.7893, "step": 280150 }, { "epoch": 48.21059876118376, "grad_norm": 9.839242935180664, "learning_rate": 2.5894700619408124e-05, "loss": 0.8191, "step": 280200 }, { "epoch": 48.21920165175499, "grad_norm": 9.352812767028809, "learning_rate": 2.5890399174122502e-05, "loss": 0.8125, "step": 280250 }, { "epoch": 48.22780454232622, "grad_norm": 7.955507755279541, "learning_rate": 2.588609772883689e-05, "loss": 0.7373, "step": 280300 }, { "epoch": 48.23640743289745, "grad_norm": 14.136201858520508, "learning_rate": 2.5881796283551275e-05, "loss": 0.8041, "step": 280350 }, { "epoch": 48.245010323468684, "grad_norm": 12.330214500427246, "learning_rate": 2.587749483826566e-05, "loss": 0.7617, "step": 280400 }, { "epoch": 48.253613214039916, "grad_norm": 14.53570556640625, "learning_rate": 2.5873193392980044e-05, "loss": 0.815, "step": 280450 }, { "epoch": 48.26221610461115, "grad_norm": 10.199761390686035, "learning_rate": 2.586889194769443e-05, "loss": 0.7952, "step": 280500 }, { "epoch": 48.27081899518238, "grad_norm": 15.318307876586914, "learning_rate": 2.586459050240881e-05, "loss": 0.7941, "step": 280550 }, { "epoch": 48.279421885753614, "grad_norm": 16.782747268676758, "learning_rate": 2.5860289057123194e-05, "loss": 0.7139, "step": 280600 }, { "epoch": 48.288024776324846, "grad_norm": 12.493013381958008, "learning_rate": 2.585598761183758e-05, "loss": 0.8011, "step": 280650 }, { "epoch": 48.29662766689608, "grad_norm": 10.476158142089844, "learning_rate": 2.5851686166551963e-05, "loss": 0.7708, "step": 280700 }, { "epoch": 48.30523055746731, "grad_norm": 15.946463584899902, "learning_rate": 2.5847384721266348e-05, "loss": 0.7904, "step": 280750 }, { "epoch": 48.313833448038544, "grad_norm": 18.805158615112305, "learning_rate": 2.584308327598073e-05, "loss": 0.7735, "step": 280800 }, { "epoch": 48.322436338609776, "grad_norm": 7.892700672149658, "learning_rate": 2.5838781830695114e-05, "loss": 0.7813, "step": 280850 }, { "epoch": 48.331039229181, "grad_norm": 12.353172302246094, "learning_rate": 2.5834480385409498e-05, "loss": 0.7878, "step": 280900 }, { "epoch": 48.339642119752234, "grad_norm": 10.50256061553955, "learning_rate": 2.5830178940123883e-05, "loss": 0.7778, "step": 280950 }, { "epoch": 48.34824501032347, "grad_norm": 15.532363891601562, "learning_rate": 2.5825877494838267e-05, "loss": 0.7961, "step": 281000 }, { "epoch": 48.3568479008947, "grad_norm": 21.840749740600586, "learning_rate": 2.582157604955265e-05, "loss": 0.8397, "step": 281050 }, { "epoch": 48.36545079146593, "grad_norm": 17.098674774169922, "learning_rate": 2.5817274604267033e-05, "loss": 0.8444, "step": 281100 }, { "epoch": 48.374053682037164, "grad_norm": 11.964585304260254, "learning_rate": 2.5812973158981418e-05, "loss": 0.8523, "step": 281150 }, { "epoch": 48.3826565726084, "grad_norm": 10.833194732666016, "learning_rate": 2.5808671713695802e-05, "loss": 0.7562, "step": 281200 }, { "epoch": 48.39125946317963, "grad_norm": 10.972952842712402, "learning_rate": 2.580437026841019e-05, "loss": 0.8243, "step": 281250 }, { "epoch": 48.39986235375086, "grad_norm": 17.635820388793945, "learning_rate": 2.5800068823124575e-05, "loss": 0.7503, "step": 281300 }, { "epoch": 48.408465244322095, "grad_norm": 10.125372886657715, "learning_rate": 2.5795767377838952e-05, "loss": 0.835, "step": 281350 }, { "epoch": 48.41706813489333, "grad_norm": 18.002683639526367, "learning_rate": 2.5791465932553337e-05, "loss": 0.7901, "step": 281400 }, { "epoch": 48.42567102546455, "grad_norm": 11.827004432678223, "learning_rate": 2.578716448726772e-05, "loss": 0.7784, "step": 281450 }, { "epoch": 48.434273916035785, "grad_norm": 12.885754585266113, "learning_rate": 2.578286304198211e-05, "loss": 0.7135, "step": 281500 }, { "epoch": 48.44287680660702, "grad_norm": 16.484928131103516, "learning_rate": 2.5778561596696494e-05, "loss": 0.7863, "step": 281550 }, { "epoch": 48.45147969717825, "grad_norm": 12.273377418518066, "learning_rate": 2.5774260151410872e-05, "loss": 0.7561, "step": 281600 }, { "epoch": 48.46008258774948, "grad_norm": 16.514244079589844, "learning_rate": 2.5769958706125256e-05, "loss": 0.7931, "step": 281650 }, { "epoch": 48.468685478320715, "grad_norm": 13.649559020996094, "learning_rate": 2.5765657260839644e-05, "loss": 0.7954, "step": 281700 }, { "epoch": 48.47728836889195, "grad_norm": 13.849393844604492, "learning_rate": 2.576135581555403e-05, "loss": 0.7663, "step": 281750 }, { "epoch": 48.48589125946318, "grad_norm": 13.775201797485352, "learning_rate": 2.5757054370268413e-05, "loss": 0.8066, "step": 281800 }, { "epoch": 48.49449415003441, "grad_norm": 10.844870567321777, "learning_rate": 2.5752752924982798e-05, "loss": 0.8052, "step": 281850 }, { "epoch": 48.503097040605645, "grad_norm": 12.114198684692383, "learning_rate": 2.574845147969718e-05, "loss": 0.7698, "step": 281900 }, { "epoch": 48.51169993117688, "grad_norm": 17.852170944213867, "learning_rate": 2.5744150034411564e-05, "loss": 0.7498, "step": 281950 }, { "epoch": 48.52030282174811, "grad_norm": 9.183165550231934, "learning_rate": 2.5739848589125948e-05, "loss": 0.773, "step": 282000 }, { "epoch": 48.528905712319336, "grad_norm": 11.974312782287598, "learning_rate": 2.5735547143840333e-05, "loss": 0.7921, "step": 282050 }, { "epoch": 48.53750860289057, "grad_norm": 9.755135536193848, "learning_rate": 2.5731245698554717e-05, "loss": 0.8056, "step": 282100 }, { "epoch": 48.5461114934618, "grad_norm": 19.980899810791016, "learning_rate": 2.57269442532691e-05, "loss": 0.7927, "step": 282150 }, { "epoch": 48.55471438403303, "grad_norm": 13.843938827514648, "learning_rate": 2.5722642807983483e-05, "loss": 0.7652, "step": 282200 }, { "epoch": 48.563317274604266, "grad_norm": 11.01641845703125, "learning_rate": 2.5718341362697868e-05, "loss": 0.7642, "step": 282250 }, { "epoch": 48.5719201651755, "grad_norm": 9.695611953735352, "learning_rate": 2.5714039917412252e-05, "loss": 0.7707, "step": 282300 }, { "epoch": 48.58052305574673, "grad_norm": 13.037498474121094, "learning_rate": 2.5709738472126637e-05, "loss": 0.7729, "step": 282350 }, { "epoch": 48.589125946317964, "grad_norm": 11.606212615966797, "learning_rate": 2.5705437026841018e-05, "loss": 0.7987, "step": 282400 }, { "epoch": 48.597728836889196, "grad_norm": 14.711237907409668, "learning_rate": 2.5701135581555403e-05, "loss": 0.8044, "step": 282450 }, { "epoch": 48.60633172746043, "grad_norm": 15.461472511291504, "learning_rate": 2.5696834136269787e-05, "loss": 0.767, "step": 282500 }, { "epoch": 48.61493461803166, "grad_norm": 6.139580726623535, "learning_rate": 2.569253269098417e-05, "loss": 0.8064, "step": 282550 }, { "epoch": 48.623537508602894, "grad_norm": 16.517642974853516, "learning_rate": 2.5688231245698556e-05, "loss": 0.747, "step": 282600 }, { "epoch": 48.63214039917412, "grad_norm": 11.136242866516113, "learning_rate": 2.5683929800412944e-05, "loss": 0.775, "step": 282650 }, { "epoch": 48.64074328974535, "grad_norm": 12.555342674255371, "learning_rate": 2.5679628355127322e-05, "loss": 0.7465, "step": 282700 }, { "epoch": 48.649346180316584, "grad_norm": 17.693132400512695, "learning_rate": 2.5675326909841707e-05, "loss": 0.793, "step": 282750 }, { "epoch": 48.65794907088782, "grad_norm": 8.825231552124023, "learning_rate": 2.567102546455609e-05, "loss": 0.7236, "step": 282800 }, { "epoch": 48.66655196145905, "grad_norm": 13.427656173706055, "learning_rate": 2.5666724019270476e-05, "loss": 0.8193, "step": 282850 }, { "epoch": 48.67515485203028, "grad_norm": 12.359375, "learning_rate": 2.5662422573984864e-05, "loss": 0.834, "step": 282900 }, { "epoch": 48.683757742601514, "grad_norm": 14.847198486328125, "learning_rate": 2.565812112869924e-05, "loss": 0.7323, "step": 282950 }, { "epoch": 48.69236063317275, "grad_norm": 14.481402397155762, "learning_rate": 2.5653819683413626e-05, "loss": 0.7873, "step": 283000 }, { "epoch": 48.70096352374398, "grad_norm": 11.889378547668457, "learning_rate": 2.564951823812801e-05, "loss": 0.7836, "step": 283050 }, { "epoch": 48.70956641431521, "grad_norm": 16.92194938659668, "learning_rate": 2.56452167928424e-05, "loss": 0.7742, "step": 283100 }, { "epoch": 48.718169304886445, "grad_norm": 9.49715805053711, "learning_rate": 2.5640915347556783e-05, "loss": 0.7733, "step": 283150 }, { "epoch": 48.72677219545768, "grad_norm": 14.684920310974121, "learning_rate": 2.5636613902271168e-05, "loss": 0.7965, "step": 283200 }, { "epoch": 48.7353750860289, "grad_norm": 15.320023536682129, "learning_rate": 2.5632312456985545e-05, "loss": 0.7974, "step": 283250 }, { "epoch": 48.743977976600135, "grad_norm": 12.57377815246582, "learning_rate": 2.5628011011699933e-05, "loss": 0.778, "step": 283300 }, { "epoch": 48.75258086717137, "grad_norm": 13.33337688446045, "learning_rate": 2.5623709566414318e-05, "loss": 0.7553, "step": 283350 }, { "epoch": 48.7611837577426, "grad_norm": 18.933055877685547, "learning_rate": 2.5619408121128702e-05, "loss": 0.7771, "step": 283400 }, { "epoch": 48.76978664831383, "grad_norm": 14.488242149353027, "learning_rate": 2.5615106675843087e-05, "loss": 0.7712, "step": 283450 }, { "epoch": 48.778389538885065, "grad_norm": 11.040003776550293, "learning_rate": 2.5610805230557465e-05, "loss": 0.7918, "step": 283500 }, { "epoch": 48.7869924294563, "grad_norm": 11.702054977416992, "learning_rate": 2.5606503785271853e-05, "loss": 0.8084, "step": 283550 }, { "epoch": 48.79559532002753, "grad_norm": 8.777935028076172, "learning_rate": 2.5602202339986237e-05, "loss": 0.8067, "step": 283600 }, { "epoch": 48.80419821059876, "grad_norm": 9.75833511352539, "learning_rate": 2.5597900894700622e-05, "loss": 0.7215, "step": 283650 }, { "epoch": 48.812801101169995, "grad_norm": 20.63547706604004, "learning_rate": 2.5593599449415006e-05, "loss": 0.8173, "step": 283700 }, { "epoch": 48.82140399174123, "grad_norm": 19.477996826171875, "learning_rate": 2.5589298004129388e-05, "loss": 0.7741, "step": 283750 }, { "epoch": 48.83000688231246, "grad_norm": 12.933279037475586, "learning_rate": 2.5584996558843772e-05, "loss": 0.7969, "step": 283800 }, { "epoch": 48.838609772883686, "grad_norm": 13.86420726776123, "learning_rate": 2.5580695113558157e-05, "loss": 0.7432, "step": 283850 }, { "epoch": 48.84721266345492, "grad_norm": 12.977071762084961, "learning_rate": 2.557639366827254e-05, "loss": 0.8209, "step": 283900 }, { "epoch": 48.85581555402615, "grad_norm": 5.724295616149902, "learning_rate": 2.5572092222986926e-05, "loss": 0.781, "step": 283950 }, { "epoch": 48.86441844459738, "grad_norm": 22.880334854125977, "learning_rate": 2.556779077770131e-05, "loss": 0.8306, "step": 284000 }, { "epoch": 48.873021335168616, "grad_norm": 11.989106178283691, "learning_rate": 2.556348933241569e-05, "loss": 0.7949, "step": 284050 }, { "epoch": 48.88162422573985, "grad_norm": 12.024293899536133, "learning_rate": 2.5559187887130076e-05, "loss": 0.7871, "step": 284100 }, { "epoch": 48.89022711631108, "grad_norm": 12.427447319030762, "learning_rate": 2.555488644184446e-05, "loss": 0.8057, "step": 284150 }, { "epoch": 48.898830006882314, "grad_norm": 10.657662391662598, "learning_rate": 2.5550584996558845e-05, "loss": 0.7727, "step": 284200 }, { "epoch": 48.907432897453546, "grad_norm": 15.205107688903809, "learning_rate": 2.554628355127323e-05, "loss": 0.7816, "step": 284250 }, { "epoch": 48.91603578802478, "grad_norm": 15.362290382385254, "learning_rate": 2.554198210598761e-05, "loss": 0.7803, "step": 284300 }, { "epoch": 48.92463867859601, "grad_norm": 17.960390090942383, "learning_rate": 2.5537680660701996e-05, "loss": 0.7936, "step": 284350 }, { "epoch": 48.93324156916724, "grad_norm": 8.857782363891602, "learning_rate": 2.553337921541638e-05, "loss": 0.8027, "step": 284400 }, { "epoch": 48.94184445973847, "grad_norm": 14.515323638916016, "learning_rate": 2.5529077770130765e-05, "loss": 0.7477, "step": 284450 }, { "epoch": 48.9504473503097, "grad_norm": 13.344963073730469, "learning_rate": 2.5524776324845153e-05, "loss": 0.8103, "step": 284500 }, { "epoch": 48.959050240880934, "grad_norm": 9.09996509552002, "learning_rate": 2.552047487955953e-05, "loss": 0.7802, "step": 284550 }, { "epoch": 48.96765313145217, "grad_norm": 14.302042007446289, "learning_rate": 2.5516173434273915e-05, "loss": 0.7731, "step": 284600 }, { "epoch": 48.9762560220234, "grad_norm": 10.482579231262207, "learning_rate": 2.55118719889883e-05, "loss": 0.7288, "step": 284650 }, { "epoch": 48.98485891259463, "grad_norm": 15.001465797424316, "learning_rate": 2.5507570543702684e-05, "loss": 0.7668, "step": 284700 }, { "epoch": 48.993461803165864, "grad_norm": 9.136077880859375, "learning_rate": 2.5503269098417072e-05, "loss": 0.8136, "step": 284750 }, { "epoch": 49.0, "eval_accuracy": 0.5690070555842368, "eval_f1": 0.5596441655892674, "eval_f1_DuraRiadoRio_16x16": 0.5154130702836005, "eval_f1_Mole_16x16": 0.507103825136612, "eval_f1_Quebrado_16x16": 0.6782734856728327, "eval_f1_RiadoRio_16x16": 0.4866609294320138, "eval_f1_RioFechado_16x16": 0.6107695174212782, "eval_loss": 1.716134786605835, "eval_precision": 0.5809387224953525, "eval_precision_DuraRiadoRio_16x16": 0.5968018275271274, "eval_precision_Mole_16x16": 0.6843657817109144, "eval_precision_Quebrado_16x16": 0.5825545171339563, "eval_precision_RiadoRio_16x16": 0.5103790613718412, "eval_precision_RioFechado_16x16": 0.5305924247329232, "eval_recall": 0.570501774698489, "eval_recall_DuraRiadoRio_16x16": 0.4535590277777778, "eval_recall_Mole_16x16": 0.4027777777777778, "eval_recall_Quebrado_16x16": 0.8116319444444444, "eval_recall_RiadoRio_16x16": 0.46504934210526316, "eval_recall_RioFechado_16x16": 0.7194907813871817, "eval_runtime": 46.3099, "eval_samples_per_second": 250.962, "eval_steps_per_second": 15.699, "step": 284788 }, { "epoch": 49.0020646937371, "grad_norm": 14.458956718444824, "learning_rate": 2.5498967653131457e-05, "loss": 0.7852, "step": 284800 }, { "epoch": 49.01066758430833, "grad_norm": 10.98720645904541, "learning_rate": 2.5494666207845834e-05, "loss": 0.7407, "step": 284850 }, { "epoch": 49.01927047487956, "grad_norm": 14.917850494384766, "learning_rate": 2.549036476256022e-05, "loss": 0.802, "step": 284900 }, { "epoch": 49.027873365450795, "grad_norm": 12.459712982177734, "learning_rate": 2.5486063317274607e-05, "loss": 0.806, "step": 284950 }, { "epoch": 49.03647625602202, "grad_norm": 15.182781219482422, "learning_rate": 2.548176187198899e-05, "loss": 0.8168, "step": 285000 }, { "epoch": 49.04507914659325, "grad_norm": 9.734964370727539, "learning_rate": 2.5477460426703376e-05, "loss": 0.791, "step": 285050 }, { "epoch": 49.053682037164485, "grad_norm": 10.266756057739258, "learning_rate": 2.5473158981417754e-05, "loss": 0.7773, "step": 285100 }, { "epoch": 49.06228492773572, "grad_norm": 14.090635299682617, "learning_rate": 2.5468857536132142e-05, "loss": 0.7961, "step": 285150 }, { "epoch": 49.07088781830695, "grad_norm": 16.612091064453125, "learning_rate": 2.5464556090846526e-05, "loss": 0.7722, "step": 285200 }, { "epoch": 49.07949070887818, "grad_norm": 12.42093563079834, "learning_rate": 2.546025464556091e-05, "loss": 0.7311, "step": 285250 }, { "epoch": 49.088093599449415, "grad_norm": 7.263678073883057, "learning_rate": 2.5455953200275295e-05, "loss": 0.8036, "step": 285300 }, { "epoch": 49.09669649002065, "grad_norm": 13.831400871276855, "learning_rate": 2.545165175498968e-05, "loss": 0.7616, "step": 285350 }, { "epoch": 49.10529938059188, "grad_norm": 15.562992095947266, "learning_rate": 2.544735030970406e-05, "loss": 0.8142, "step": 285400 }, { "epoch": 49.11390227116311, "grad_norm": 13.766745567321777, "learning_rate": 2.5443048864418446e-05, "loss": 0.8143, "step": 285450 }, { "epoch": 49.122505161734345, "grad_norm": 14.697474479675293, "learning_rate": 2.543874741913283e-05, "loss": 0.7873, "step": 285500 }, { "epoch": 49.13110805230558, "grad_norm": 13.782334327697754, "learning_rate": 2.5434445973847215e-05, "loss": 0.7755, "step": 285550 }, { "epoch": 49.1397109428768, "grad_norm": 7.897702693939209, "learning_rate": 2.54301445285616e-05, "loss": 0.8209, "step": 285600 }, { "epoch": 49.148313833448036, "grad_norm": 15.744916915893555, "learning_rate": 2.542584308327598e-05, "loss": 0.8354, "step": 285650 }, { "epoch": 49.15691672401927, "grad_norm": 22.52106285095215, "learning_rate": 2.5421541637990365e-05, "loss": 0.7435, "step": 285700 }, { "epoch": 49.1655196145905, "grad_norm": 7.791121959686279, "learning_rate": 2.541724019270475e-05, "loss": 0.8097, "step": 285750 }, { "epoch": 49.17412250516173, "grad_norm": 10.938132286071777, "learning_rate": 2.5412938747419134e-05, "loss": 0.7716, "step": 285800 }, { "epoch": 49.182725395732966, "grad_norm": 13.488805770874023, "learning_rate": 2.540863730213352e-05, "loss": 0.7825, "step": 285850 }, { "epoch": 49.1913282863042, "grad_norm": 5.355918884277344, "learning_rate": 2.54043358568479e-05, "loss": 0.7315, "step": 285900 }, { "epoch": 49.19993117687543, "grad_norm": 11.294045448303223, "learning_rate": 2.5400034411562285e-05, "loss": 0.8129, "step": 285950 }, { "epoch": 49.208534067446664, "grad_norm": 30.750673294067383, "learning_rate": 2.539573296627667e-05, "loss": 0.7633, "step": 286000 }, { "epoch": 49.217136958017896, "grad_norm": 9.314048767089844, "learning_rate": 2.5391431520991054e-05, "loss": 0.8343, "step": 286050 }, { "epoch": 49.22573984858913, "grad_norm": 10.61589527130127, "learning_rate": 2.5387130075705438e-05, "loss": 0.8237, "step": 286100 }, { "epoch": 49.23434273916036, "grad_norm": 12.923331260681152, "learning_rate": 2.5382828630419826e-05, "loss": 0.7706, "step": 286150 }, { "epoch": 49.24294562973159, "grad_norm": 13.035685539245605, "learning_rate": 2.5378527185134204e-05, "loss": 0.7582, "step": 286200 }, { "epoch": 49.25154852030282, "grad_norm": 10.183847427368164, "learning_rate": 2.537422573984859e-05, "loss": 0.8237, "step": 286250 }, { "epoch": 49.26015141087405, "grad_norm": 23.588298797607422, "learning_rate": 2.5369924294562973e-05, "loss": 0.7677, "step": 286300 }, { "epoch": 49.268754301445284, "grad_norm": 15.567853927612305, "learning_rate": 2.536562284927736e-05, "loss": 0.7922, "step": 286350 }, { "epoch": 49.27735719201652, "grad_norm": 9.045783042907715, "learning_rate": 2.5361321403991746e-05, "loss": 0.772, "step": 286400 }, { "epoch": 49.28596008258775, "grad_norm": 12.343469619750977, "learning_rate": 2.5357019958706123e-05, "loss": 0.8083, "step": 286450 }, { "epoch": 49.29456297315898, "grad_norm": 17.658174514770508, "learning_rate": 2.5352718513420508e-05, "loss": 0.7187, "step": 286500 }, { "epoch": 49.303165863730214, "grad_norm": 18.790790557861328, "learning_rate": 2.5348417068134896e-05, "loss": 0.7508, "step": 286550 }, { "epoch": 49.31176875430145, "grad_norm": 10.485434532165527, "learning_rate": 2.534411562284928e-05, "loss": 0.7838, "step": 286600 }, { "epoch": 49.32037164487268, "grad_norm": 11.754314422607422, "learning_rate": 2.5339814177563665e-05, "loss": 0.67, "step": 286650 }, { "epoch": 49.32897453544391, "grad_norm": 11.602588653564453, "learning_rate": 2.5335512732278043e-05, "loss": 0.782, "step": 286700 }, { "epoch": 49.337577426015145, "grad_norm": 16.280818939208984, "learning_rate": 2.5331211286992427e-05, "loss": 0.7773, "step": 286750 }, { "epoch": 49.34618031658637, "grad_norm": 13.30538558959961, "learning_rate": 2.5326909841706815e-05, "loss": 0.8292, "step": 286800 }, { "epoch": 49.3547832071576, "grad_norm": 18.33893585205078, "learning_rate": 2.53226083964212e-05, "loss": 0.8034, "step": 286850 }, { "epoch": 49.363386097728835, "grad_norm": 11.83631706237793, "learning_rate": 2.5318306951135584e-05, "loss": 0.819, "step": 286900 }, { "epoch": 49.37198898830007, "grad_norm": 13.525436401367188, "learning_rate": 2.531400550584997e-05, "loss": 0.7484, "step": 286950 }, { "epoch": 49.3805918788713, "grad_norm": 12.451898574829102, "learning_rate": 2.530970406056435e-05, "loss": 0.8367, "step": 287000 }, { "epoch": 49.38919476944253, "grad_norm": 15.888591766357422, "learning_rate": 2.5305402615278735e-05, "loss": 0.784, "step": 287050 }, { "epoch": 49.397797660013765, "grad_norm": 18.30401039123535, "learning_rate": 2.530110116999312e-05, "loss": 0.7454, "step": 287100 }, { "epoch": 49.406400550585, "grad_norm": 14.560074806213379, "learning_rate": 2.5296799724707504e-05, "loss": 0.7916, "step": 287150 }, { "epoch": 49.41500344115623, "grad_norm": 10.862629890441895, "learning_rate": 2.529249827942189e-05, "loss": 0.7386, "step": 287200 }, { "epoch": 49.42360633172746, "grad_norm": 10.25634765625, "learning_rate": 2.528819683413627e-05, "loss": 0.8088, "step": 287250 }, { "epoch": 49.432209222298695, "grad_norm": 11.074009895324707, "learning_rate": 2.5283895388850654e-05, "loss": 0.7808, "step": 287300 }, { "epoch": 49.44081211286992, "grad_norm": 17.033456802368164, "learning_rate": 2.527959394356504e-05, "loss": 0.7953, "step": 287350 }, { "epoch": 49.44941500344115, "grad_norm": 12.158028602600098, "learning_rate": 2.5275292498279423e-05, "loss": 0.7646, "step": 287400 }, { "epoch": 49.458017894012386, "grad_norm": 12.736845970153809, "learning_rate": 2.5270991052993808e-05, "loss": 0.7795, "step": 287450 }, { "epoch": 49.46662078458362, "grad_norm": 15.93651008605957, "learning_rate": 2.5266689607708192e-05, "loss": 0.8123, "step": 287500 }, { "epoch": 49.47522367515485, "grad_norm": 13.094764709472656, "learning_rate": 2.5262388162422574e-05, "loss": 0.8746, "step": 287550 }, { "epoch": 49.48382656572608, "grad_norm": 13.52327823638916, "learning_rate": 2.5258086717136958e-05, "loss": 0.7751, "step": 287600 }, { "epoch": 49.492429456297316, "grad_norm": 16.38433265686035, "learning_rate": 2.5253785271851343e-05, "loss": 0.7462, "step": 287650 }, { "epoch": 49.50103234686855, "grad_norm": 9.447290420532227, "learning_rate": 2.5249483826565727e-05, "loss": 0.8024, "step": 287700 }, { "epoch": 49.50963523743978, "grad_norm": 14.87409496307373, "learning_rate": 2.5245182381280115e-05, "loss": 0.7911, "step": 287750 }, { "epoch": 49.51823812801101, "grad_norm": 9.113334655761719, "learning_rate": 2.5240880935994493e-05, "loss": 0.8099, "step": 287800 }, { "epoch": 49.526841018582246, "grad_norm": 14.598387718200684, "learning_rate": 2.5236579490708877e-05, "loss": 0.7487, "step": 287850 }, { "epoch": 49.53544390915348, "grad_norm": 14.077424049377441, "learning_rate": 2.5232278045423262e-05, "loss": 0.723, "step": 287900 }, { "epoch": 49.544046799724704, "grad_norm": 8.970601081848145, "learning_rate": 2.522797660013765e-05, "loss": 0.7432, "step": 287950 }, { "epoch": 49.55264969029594, "grad_norm": 10.844277381896973, "learning_rate": 2.5223675154852035e-05, "loss": 0.7762, "step": 288000 }, { "epoch": 49.56125258086717, "grad_norm": 13.110955238342285, "learning_rate": 2.5219373709566412e-05, "loss": 0.7931, "step": 288050 }, { "epoch": 49.5698554714384, "grad_norm": 15.16531753540039, "learning_rate": 2.5215072264280797e-05, "loss": 0.8138, "step": 288100 }, { "epoch": 49.578458362009634, "grad_norm": 11.381049156188965, "learning_rate": 2.521077081899518e-05, "loss": 0.8198, "step": 288150 }, { "epoch": 49.58706125258087, "grad_norm": 15.170456886291504, "learning_rate": 2.520646937370957e-05, "loss": 0.7959, "step": 288200 }, { "epoch": 49.5956641431521, "grad_norm": 14.224203109741211, "learning_rate": 2.5202167928423954e-05, "loss": 0.7719, "step": 288250 }, { "epoch": 49.60426703372333, "grad_norm": 9.411438941955566, "learning_rate": 2.519786648313834e-05, "loss": 0.7595, "step": 288300 }, { "epoch": 49.612869924294564, "grad_norm": 12.056334495544434, "learning_rate": 2.5193565037852716e-05, "loss": 0.784, "step": 288350 }, { "epoch": 49.6214728148658, "grad_norm": 9.17379093170166, "learning_rate": 2.5189263592567104e-05, "loss": 0.8512, "step": 288400 }, { "epoch": 49.63007570543703, "grad_norm": 13.369263648986816, "learning_rate": 2.518496214728149e-05, "loss": 0.8187, "step": 288450 }, { "epoch": 49.63867859600826, "grad_norm": 9.71693229675293, "learning_rate": 2.5180660701995873e-05, "loss": 0.764, "step": 288500 }, { "epoch": 49.64728148657949, "grad_norm": 11.930364608764648, "learning_rate": 2.5176359256710258e-05, "loss": 0.856, "step": 288550 }, { "epoch": 49.65588437715072, "grad_norm": 12.864754676818848, "learning_rate": 2.517205781142464e-05, "loss": 0.7583, "step": 288600 }, { "epoch": 49.66448726772195, "grad_norm": 12.883216857910156, "learning_rate": 2.5167756366139024e-05, "loss": 0.747, "step": 288650 }, { "epoch": 49.673090158293185, "grad_norm": 10.580684661865234, "learning_rate": 2.5163454920853408e-05, "loss": 0.7695, "step": 288700 }, { "epoch": 49.68169304886442, "grad_norm": 13.02924633026123, "learning_rate": 2.5159153475567793e-05, "loss": 0.7998, "step": 288750 }, { "epoch": 49.69029593943565, "grad_norm": 15.290536880493164, "learning_rate": 2.5154852030282177e-05, "loss": 0.7517, "step": 288800 }, { "epoch": 49.69889883000688, "grad_norm": 10.767369270324707, "learning_rate": 2.5150550584996562e-05, "loss": 0.8364, "step": 288850 }, { "epoch": 49.707501720578115, "grad_norm": 11.273589134216309, "learning_rate": 2.5146249139710943e-05, "loss": 0.8098, "step": 288900 }, { "epoch": 49.71610461114935, "grad_norm": 13.717353820800781, "learning_rate": 2.5141947694425328e-05, "loss": 0.7324, "step": 288950 }, { "epoch": 49.72470750172058, "grad_norm": 14.22813892364502, "learning_rate": 2.5137646249139712e-05, "loss": 0.7278, "step": 289000 }, { "epoch": 49.73331039229181, "grad_norm": 8.962784767150879, "learning_rate": 2.5133344803854097e-05, "loss": 0.7313, "step": 289050 }, { "epoch": 49.741913282863045, "grad_norm": 12.882214546203613, "learning_rate": 2.512904335856848e-05, "loss": 0.7412, "step": 289100 }, { "epoch": 49.75051617343427, "grad_norm": 17.84402847290039, "learning_rate": 2.5124741913282862e-05, "loss": 0.7878, "step": 289150 }, { "epoch": 49.7591190640055, "grad_norm": 12.562623023986816, "learning_rate": 2.5120440467997247e-05, "loss": 0.7931, "step": 289200 }, { "epoch": 49.767721954576736, "grad_norm": 16.538253784179688, "learning_rate": 2.511613902271163e-05, "loss": 0.8181, "step": 289250 }, { "epoch": 49.77632484514797, "grad_norm": 10.784673690795898, "learning_rate": 2.5111837577426016e-05, "loss": 0.8325, "step": 289300 }, { "epoch": 49.7849277357192, "grad_norm": 16.759227752685547, "learning_rate": 2.51075361321404e-05, "loss": 0.7283, "step": 289350 }, { "epoch": 49.79353062629043, "grad_norm": 15.76153564453125, "learning_rate": 2.5103234686854782e-05, "loss": 0.7869, "step": 289400 }, { "epoch": 49.802133516861666, "grad_norm": 11.015127182006836, "learning_rate": 2.5098933241569166e-05, "loss": 0.7539, "step": 289450 }, { "epoch": 49.8107364074329, "grad_norm": 9.004149436950684, "learning_rate": 2.509463179628355e-05, "loss": 0.7511, "step": 289500 }, { "epoch": 49.81933929800413, "grad_norm": 10.449600219726562, "learning_rate": 2.5090330350997936e-05, "loss": 0.8048, "step": 289550 }, { "epoch": 49.82794218857536, "grad_norm": 10.018510818481445, "learning_rate": 2.5086028905712324e-05, "loss": 0.7179, "step": 289600 }, { "epoch": 49.836545079146596, "grad_norm": 21.81939125061035, "learning_rate": 2.5081727460426708e-05, "loss": 0.785, "step": 289650 }, { "epoch": 49.84514796971783, "grad_norm": 15.510241508483887, "learning_rate": 2.5077426015141086e-05, "loss": 0.7922, "step": 289700 }, { "epoch": 49.853750860289054, "grad_norm": 14.675703048706055, "learning_rate": 2.507312456985547e-05, "loss": 0.7361, "step": 289750 }, { "epoch": 49.86235375086029, "grad_norm": 9.062132835388184, "learning_rate": 2.506882312456986e-05, "loss": 0.8222, "step": 289800 }, { "epoch": 49.87095664143152, "grad_norm": 11.356934547424316, "learning_rate": 2.5064521679284243e-05, "loss": 0.8238, "step": 289850 }, { "epoch": 49.87955953200275, "grad_norm": 8.346104621887207, "learning_rate": 2.5060220233998628e-05, "loss": 0.7512, "step": 289900 }, { "epoch": 49.888162422573984, "grad_norm": 6.99303674697876, "learning_rate": 2.5055918788713005e-05, "loss": 0.7711, "step": 289950 }, { "epoch": 49.89676531314522, "grad_norm": 13.907303810119629, "learning_rate": 2.505161734342739e-05, "loss": 0.7905, "step": 290000 }, { "epoch": 49.90536820371645, "grad_norm": 11.212627410888672, "learning_rate": 2.5047315898141778e-05, "loss": 0.8002, "step": 290050 }, { "epoch": 49.91397109428768, "grad_norm": 8.052711486816406, "learning_rate": 2.5043014452856162e-05, "loss": 0.7318, "step": 290100 }, { "epoch": 49.922573984858914, "grad_norm": 14.100408554077148, "learning_rate": 2.5038713007570547e-05, "loss": 0.7933, "step": 290150 }, { "epoch": 49.93117687543015, "grad_norm": 16.473722457885742, "learning_rate": 2.5034411562284925e-05, "loss": 0.7827, "step": 290200 }, { "epoch": 49.93977976600138, "grad_norm": 12.910300254821777, "learning_rate": 2.5030110116999313e-05, "loss": 0.7755, "step": 290250 }, { "epoch": 49.94838265657261, "grad_norm": 8.22254467010498, "learning_rate": 2.5025808671713697e-05, "loss": 0.7444, "step": 290300 }, { "epoch": 49.95698554714384, "grad_norm": 9.928719520568848, "learning_rate": 2.5021507226428082e-05, "loss": 0.7895, "step": 290350 }, { "epoch": 49.96558843771507, "grad_norm": 13.921860694885254, "learning_rate": 2.5017205781142466e-05, "loss": 0.6775, "step": 290400 }, { "epoch": 49.9741913282863, "grad_norm": 12.827495574951172, "learning_rate": 2.501290433585685e-05, "loss": 0.8062, "step": 290450 }, { "epoch": 49.982794218857535, "grad_norm": 16.90026092529297, "learning_rate": 2.5008602890571232e-05, "loss": 0.8167, "step": 290500 }, { "epoch": 49.99139710942877, "grad_norm": 11.892571449279785, "learning_rate": 2.5004301445285617e-05, "loss": 0.7881, "step": 290550 }, { "epoch": 50.0, "grad_norm": 31.98631477355957, "learning_rate": 2.5e-05, "loss": 0.7742, "step": 290600 }, { "epoch": 50.0, "eval_accuracy": 0.5997246601273447, "eval_f1": 0.6006403938008589, "eval_f1_DuraRiadoRio_16x16": 0.5875299760191847, "eval_f1_Mole_16x16": 0.6000488639139995, "eval_f1_Quebrado_16x16": 0.7293139293139294, "eval_f1_RiadoRio_16x16": 0.4817429897115191, "eval_f1_RioFechado_16x16": 0.6045662100456621, "eval_loss": 1.5161103010177612, "eval_precision": 0.6067058448249324, "eval_precision_DuraRiadoRio_16x16": 0.5444444444444444, "eval_precision_Mole_16x16": 0.6864169927333706, "eval_precision_Quebrado_16x16": 0.6999201915403033, "eval_precision_RiadoRio_16x16": 0.47287128712871285, "eval_precision_RioFechado_16x16": 0.6298763082778306, "eval_recall": 0.6008914406296691, "eval_recall_DuraRiadoRio_16x16": 0.6380208333333334, "eval_recall_Mole_16x16": 0.5329861111111112, "eval_recall_Quebrado_16x16": 0.7612847222222222, "eval_recall_RiadoRio_16x16": 0.4909539473684211, "eval_recall_RioFechado_16x16": 0.5812115891132572, "eval_runtime": 46.5065, "eval_samples_per_second": 249.9, "eval_steps_per_second": 15.632, "step": 290600 }, { "epoch": 50.00860289057123, "grad_norm": 10.245528221130371, "learning_rate": 2.4995698554714386e-05, "loss": 0.7465, "step": 290650 }, { "epoch": 50.017205781142465, "grad_norm": 8.026317596435547, "learning_rate": 2.4991397109428767e-05, "loss": 0.7536, "step": 290700 }, { "epoch": 50.0258086717137, "grad_norm": 12.883636474609375, "learning_rate": 2.4987095664143155e-05, "loss": 0.775, "step": 290750 }, { "epoch": 50.03441156228493, "grad_norm": 18.11627197265625, "learning_rate": 2.498279421885754e-05, "loss": 0.778, "step": 290800 }, { "epoch": 50.04301445285616, "grad_norm": 12.078458786010742, "learning_rate": 2.497849277357192e-05, "loss": 0.7945, "step": 290850 }, { "epoch": 50.05161734342739, "grad_norm": 15.556851387023926, "learning_rate": 2.4974191328286305e-05, "loss": 0.7798, "step": 290900 }, { "epoch": 50.06022023399862, "grad_norm": 8.715314865112305, "learning_rate": 2.496988988300069e-05, "loss": 0.7788, "step": 290950 }, { "epoch": 50.06882312456985, "grad_norm": 9.057462692260742, "learning_rate": 2.4965588437715074e-05, "loss": 0.7889, "step": 291000 }, { "epoch": 50.077426015141086, "grad_norm": 16.906578063964844, "learning_rate": 2.496128699242946e-05, "loss": 0.7779, "step": 291050 }, { "epoch": 50.08602890571232, "grad_norm": 17.015283584594727, "learning_rate": 2.495698554714384e-05, "loss": 0.8048, "step": 291100 }, { "epoch": 50.09463179628355, "grad_norm": 17.025005340576172, "learning_rate": 2.4952684101858225e-05, "loss": 0.7221, "step": 291150 }, { "epoch": 50.10323468685478, "grad_norm": 13.619422912597656, "learning_rate": 2.4948382656572613e-05, "loss": 0.7649, "step": 291200 }, { "epoch": 50.111837577426016, "grad_norm": 13.978561401367188, "learning_rate": 2.4944081211286994e-05, "loss": 0.7933, "step": 291250 }, { "epoch": 50.12044046799725, "grad_norm": 12.805379867553711, "learning_rate": 2.4939779766001378e-05, "loss": 0.8139, "step": 291300 }, { "epoch": 50.12904335856848, "grad_norm": 8.511323928833008, "learning_rate": 2.493547832071576e-05, "loss": 0.7508, "step": 291350 }, { "epoch": 50.13764624913971, "grad_norm": 14.502440452575684, "learning_rate": 2.4931176875430144e-05, "loss": 0.75, "step": 291400 }, { "epoch": 50.146249139710946, "grad_norm": 16.19222640991211, "learning_rate": 2.4926875430144532e-05, "loss": 0.7411, "step": 291450 }, { "epoch": 50.15485203028217, "grad_norm": 14.662142753601074, "learning_rate": 2.4922573984858913e-05, "loss": 0.7319, "step": 291500 }, { "epoch": 50.163454920853404, "grad_norm": 8.865581512451172, "learning_rate": 2.4918272539573298e-05, "loss": 0.7771, "step": 291550 }, { "epoch": 50.17205781142464, "grad_norm": 17.84083366394043, "learning_rate": 2.4913971094287682e-05, "loss": 0.7218, "step": 291600 }, { "epoch": 50.18066070199587, "grad_norm": 11.267870903015137, "learning_rate": 2.4909669649002067e-05, "loss": 0.7889, "step": 291650 }, { "epoch": 50.1892635925671, "grad_norm": 10.110628128051758, "learning_rate": 2.490536820371645e-05, "loss": 0.8137, "step": 291700 }, { "epoch": 50.197866483138334, "grad_norm": 9.35427474975586, "learning_rate": 2.4901066758430833e-05, "loss": 0.7703, "step": 291750 }, { "epoch": 50.20646937370957, "grad_norm": 10.615189552307129, "learning_rate": 2.4896765313145217e-05, "loss": 0.7797, "step": 291800 }, { "epoch": 50.2150722642808, "grad_norm": 13.923439979553223, "learning_rate": 2.48924638678596e-05, "loss": 0.764, "step": 291850 }, { "epoch": 50.22367515485203, "grad_norm": 13.804241180419922, "learning_rate": 2.4888162422573986e-05, "loss": 0.7494, "step": 291900 }, { "epoch": 50.232278045423264, "grad_norm": 11.26194953918457, "learning_rate": 2.488386097728837e-05, "loss": 0.771, "step": 291950 }, { "epoch": 50.2408809359945, "grad_norm": 10.889945983886719, "learning_rate": 2.4879559532002755e-05, "loss": 0.7381, "step": 292000 }, { "epoch": 50.24948382656573, "grad_norm": 12.131166458129883, "learning_rate": 2.4875258086717137e-05, "loss": 0.7868, "step": 292050 }, { "epoch": 50.258086717136955, "grad_norm": 10.924623489379883, "learning_rate": 2.487095664143152e-05, "loss": 0.7219, "step": 292100 }, { "epoch": 50.26668960770819, "grad_norm": 17.35066795349121, "learning_rate": 2.4866655196145906e-05, "loss": 0.7948, "step": 292150 }, { "epoch": 50.27529249827942, "grad_norm": 16.07591438293457, "learning_rate": 2.486235375086029e-05, "loss": 0.7953, "step": 292200 }, { "epoch": 50.28389538885065, "grad_norm": 14.236689567565918, "learning_rate": 2.4858052305574675e-05, "loss": 0.7496, "step": 292250 }, { "epoch": 50.292498279421885, "grad_norm": 12.553936958312988, "learning_rate": 2.4853750860289056e-05, "loss": 0.7625, "step": 292300 }, { "epoch": 50.30110116999312, "grad_norm": 8.874856948852539, "learning_rate": 2.4849449415003444e-05, "loss": 0.7172, "step": 292350 }, { "epoch": 50.30970406056435, "grad_norm": 12.248892784118652, "learning_rate": 2.484514796971783e-05, "loss": 0.745, "step": 292400 }, { "epoch": 50.31830695113558, "grad_norm": 11.261933326721191, "learning_rate": 2.484084652443221e-05, "loss": 0.7868, "step": 292450 }, { "epoch": 50.326909841706815, "grad_norm": 15.841104507446289, "learning_rate": 2.4836545079146594e-05, "loss": 0.7598, "step": 292500 }, { "epoch": 50.33551273227805, "grad_norm": 17.86539649963379, "learning_rate": 2.483224363386098e-05, "loss": 0.7651, "step": 292550 }, { "epoch": 50.34411562284928, "grad_norm": 13.614448547363281, "learning_rate": 2.4827942188575363e-05, "loss": 0.7017, "step": 292600 }, { "epoch": 50.35271851342051, "grad_norm": 16.526470184326172, "learning_rate": 2.4823640743289748e-05, "loss": 0.743, "step": 292650 }, { "epoch": 50.36132140399174, "grad_norm": 13.104747772216797, "learning_rate": 2.481933929800413e-05, "loss": 0.7797, "step": 292700 }, { "epoch": 50.36992429456297, "grad_norm": 12.189225196838379, "learning_rate": 2.4815037852718514e-05, "loss": 0.7357, "step": 292750 }, { "epoch": 50.3785271851342, "grad_norm": 16.988327026367188, "learning_rate": 2.4810736407432898e-05, "loss": 0.7274, "step": 292800 }, { "epoch": 50.387130075705436, "grad_norm": 17.087923049926758, "learning_rate": 2.4806434962147283e-05, "loss": 0.7612, "step": 292850 }, { "epoch": 50.39573296627667, "grad_norm": 17.28749656677246, "learning_rate": 2.4802133516861667e-05, "loss": 0.754, "step": 292900 }, { "epoch": 50.4043358568479, "grad_norm": 9.720795631408691, "learning_rate": 2.4797832071576052e-05, "loss": 0.7548, "step": 292950 }, { "epoch": 50.41293874741913, "grad_norm": 11.390569686889648, "learning_rate": 2.4793530626290433e-05, "loss": 0.8037, "step": 293000 }, { "epoch": 50.421541637990366, "grad_norm": 7.9909796714782715, "learning_rate": 2.478922918100482e-05, "loss": 0.7897, "step": 293050 }, { "epoch": 50.4301445285616, "grad_norm": 11.264514923095703, "learning_rate": 2.4784927735719202e-05, "loss": 0.7727, "step": 293100 }, { "epoch": 50.43874741913283, "grad_norm": 10.609395980834961, "learning_rate": 2.4780626290433587e-05, "loss": 0.7884, "step": 293150 }, { "epoch": 50.44735030970406, "grad_norm": 14.256702423095703, "learning_rate": 2.477632484514797e-05, "loss": 0.7947, "step": 293200 }, { "epoch": 50.455953200275296, "grad_norm": 15.825361251831055, "learning_rate": 2.4772023399862352e-05, "loss": 0.8023, "step": 293250 }, { "epoch": 50.46455609084652, "grad_norm": 13.596331596374512, "learning_rate": 2.476772195457674e-05, "loss": 0.7562, "step": 293300 }, { "epoch": 50.473158981417754, "grad_norm": 14.30758285522461, "learning_rate": 2.4763420509291125e-05, "loss": 0.7427, "step": 293350 }, { "epoch": 50.48176187198899, "grad_norm": 13.24466609954834, "learning_rate": 2.4759119064005506e-05, "loss": 0.7747, "step": 293400 }, { "epoch": 50.49036476256022, "grad_norm": 23.05093765258789, "learning_rate": 2.475481761871989e-05, "loss": 0.8136, "step": 293450 }, { "epoch": 50.49896765313145, "grad_norm": 13.15686321258545, "learning_rate": 2.4750516173434275e-05, "loss": 0.78, "step": 293500 }, { "epoch": 50.507570543702684, "grad_norm": 14.95875072479248, "learning_rate": 2.474621472814866e-05, "loss": 0.7683, "step": 293550 }, { "epoch": 50.51617343427392, "grad_norm": 17.761898040771484, "learning_rate": 2.4741913282863044e-05, "loss": 0.7999, "step": 293600 }, { "epoch": 50.52477632484515, "grad_norm": 13.978631019592285, "learning_rate": 2.4737611837577426e-05, "loss": 0.76, "step": 293650 }, { "epoch": 50.53337921541638, "grad_norm": 14.894869804382324, "learning_rate": 2.473331039229181e-05, "loss": 0.783, "step": 293700 }, { "epoch": 50.541982105987614, "grad_norm": 19.8994197845459, "learning_rate": 2.4729008947006198e-05, "loss": 0.8251, "step": 293750 }, { "epoch": 50.55058499655885, "grad_norm": 12.393366813659668, "learning_rate": 2.472470750172058e-05, "loss": 0.7638, "step": 293800 }, { "epoch": 50.55918788713008, "grad_norm": 13.88259506225586, "learning_rate": 2.4720406056434964e-05, "loss": 0.7876, "step": 293850 }, { "epoch": 50.567790777701305, "grad_norm": 14.820663452148438, "learning_rate": 2.4716104611149345e-05, "loss": 0.8087, "step": 293900 }, { "epoch": 50.57639366827254, "grad_norm": 9.285524368286133, "learning_rate": 2.471180316586373e-05, "loss": 0.7989, "step": 293950 }, { "epoch": 50.58499655884377, "grad_norm": 13.644539833068848, "learning_rate": 2.4707501720578117e-05, "loss": 0.7905, "step": 294000 }, { "epoch": 50.593599449415, "grad_norm": 15.458157539367676, "learning_rate": 2.47032002752925e-05, "loss": 0.8182, "step": 294050 }, { "epoch": 50.602202339986235, "grad_norm": 20.065967559814453, "learning_rate": 2.4698898830006883e-05, "loss": 0.7992, "step": 294100 }, { "epoch": 50.61080523055747, "grad_norm": 7.557967185974121, "learning_rate": 2.4694597384721268e-05, "loss": 0.7915, "step": 294150 }, { "epoch": 50.6194081211287, "grad_norm": 15.041118621826172, "learning_rate": 2.4690295939435652e-05, "loss": 0.8276, "step": 294200 }, { "epoch": 50.62801101169993, "grad_norm": 16.35208511352539, "learning_rate": 2.4685994494150037e-05, "loss": 0.7566, "step": 294250 }, { "epoch": 50.636613902271165, "grad_norm": 11.658527374267578, "learning_rate": 2.4681693048864418e-05, "loss": 0.7952, "step": 294300 }, { "epoch": 50.6452167928424, "grad_norm": 14.771836280822754, "learning_rate": 2.4677391603578803e-05, "loss": 0.7657, "step": 294350 }, { "epoch": 50.65381968341363, "grad_norm": 8.932518005371094, "learning_rate": 2.4673090158293187e-05, "loss": 0.7927, "step": 294400 }, { "epoch": 50.662422573984855, "grad_norm": 12.715476989746094, "learning_rate": 2.466878871300757e-05, "loss": 0.8145, "step": 294450 }, { "epoch": 50.67102546455609, "grad_norm": 19.275836944580078, "learning_rate": 2.4664487267721956e-05, "loss": 0.7744, "step": 294500 }, { "epoch": 50.67962835512732, "grad_norm": 9.309185028076172, "learning_rate": 2.466018582243634e-05, "loss": 0.7506, "step": 294550 }, { "epoch": 50.68823124569855, "grad_norm": 10.064460754394531, "learning_rate": 2.4655884377150722e-05, "loss": 0.7829, "step": 294600 }, { "epoch": 50.696834136269786, "grad_norm": 17.47478485107422, "learning_rate": 2.4651582931865107e-05, "loss": 0.7722, "step": 294650 }, { "epoch": 50.70543702684102, "grad_norm": 19.582746505737305, "learning_rate": 2.4647281486579494e-05, "loss": 0.7692, "step": 294700 }, { "epoch": 50.71403991741225, "grad_norm": 15.520309448242188, "learning_rate": 2.4642980041293876e-05, "loss": 0.745, "step": 294750 }, { "epoch": 50.72264280798348, "grad_norm": 17.792957305908203, "learning_rate": 2.463867859600826e-05, "loss": 0.7995, "step": 294800 }, { "epoch": 50.731245698554716, "grad_norm": 14.506513595581055, "learning_rate": 2.463437715072264e-05, "loss": 0.8265, "step": 294850 }, { "epoch": 50.73984858912595, "grad_norm": 9.428252220153809, "learning_rate": 2.463007570543703e-05, "loss": 0.7741, "step": 294900 }, { "epoch": 50.74845147969718, "grad_norm": 14.552600860595703, "learning_rate": 2.4625774260151414e-05, "loss": 0.7901, "step": 294950 }, { "epoch": 50.75705437026841, "grad_norm": 21.335973739624023, "learning_rate": 2.4621472814865795e-05, "loss": 0.8229, "step": 295000 }, { "epoch": 50.76565726083964, "grad_norm": 17.02912139892578, "learning_rate": 2.461717136958018e-05, "loss": 0.8208, "step": 295050 }, { "epoch": 50.77426015141087, "grad_norm": 12.608941078186035, "learning_rate": 2.4612869924294564e-05, "loss": 0.7761, "step": 295100 }, { "epoch": 50.782863041982104, "grad_norm": 14.636415481567383, "learning_rate": 2.460856847900895e-05, "loss": 0.7725, "step": 295150 }, { "epoch": 50.791465932553336, "grad_norm": 18.20149803161621, "learning_rate": 2.4604267033723333e-05, "loss": 0.7969, "step": 295200 }, { "epoch": 50.80006882312457, "grad_norm": 11.4070463180542, "learning_rate": 2.4599965588437714e-05, "loss": 0.7824, "step": 295250 }, { "epoch": 50.8086717136958, "grad_norm": 11.977189064025879, "learning_rate": 2.45956641431521e-05, "loss": 0.7502, "step": 295300 }, { "epoch": 50.817274604267034, "grad_norm": 7.8033857345581055, "learning_rate": 2.4591362697866484e-05, "loss": 0.7438, "step": 295350 }, { "epoch": 50.82587749483827, "grad_norm": 15.350400924682617, "learning_rate": 2.4587061252580868e-05, "loss": 0.8013, "step": 295400 }, { "epoch": 50.8344803854095, "grad_norm": 11.280023574829102, "learning_rate": 2.4582759807295253e-05, "loss": 0.7494, "step": 295450 }, { "epoch": 50.84308327598073, "grad_norm": 11.201798439025879, "learning_rate": 2.4578458362009637e-05, "loss": 0.7394, "step": 295500 }, { "epoch": 50.851686166551964, "grad_norm": 16.925960540771484, "learning_rate": 2.457415691672402e-05, "loss": 0.751, "step": 295550 }, { "epoch": 50.8602890571232, "grad_norm": 16.65064811706543, "learning_rate": 2.4569855471438406e-05, "loss": 0.7544, "step": 295600 }, { "epoch": 50.86889194769442, "grad_norm": 9.630936622619629, "learning_rate": 2.4565554026152788e-05, "loss": 0.7942, "step": 295650 }, { "epoch": 50.877494838265655, "grad_norm": 19.668136596679688, "learning_rate": 2.4561252580867172e-05, "loss": 0.7708, "step": 295700 }, { "epoch": 50.88609772883689, "grad_norm": 13.852607727050781, "learning_rate": 2.4556951135581557e-05, "loss": 0.8093, "step": 295750 }, { "epoch": 50.89470061940812, "grad_norm": 13.884551048278809, "learning_rate": 2.455264969029594e-05, "loss": 0.8183, "step": 295800 }, { "epoch": 50.90330350997935, "grad_norm": 12.541149139404297, "learning_rate": 2.4548348245010326e-05, "loss": 0.7602, "step": 295850 }, { "epoch": 50.911906400550585, "grad_norm": 13.614590644836426, "learning_rate": 2.454404679972471e-05, "loss": 0.7789, "step": 295900 }, { "epoch": 50.92050929112182, "grad_norm": 12.431035041809082, "learning_rate": 2.453974535443909e-05, "loss": 0.8029, "step": 295950 }, { "epoch": 50.92911218169305, "grad_norm": 10.72928237915039, "learning_rate": 2.4535443909153476e-05, "loss": 0.745, "step": 296000 }, { "epoch": 50.93771507226428, "grad_norm": 15.868298530578613, "learning_rate": 2.453114246386786e-05, "loss": 0.7543, "step": 296050 }, { "epoch": 50.946317962835515, "grad_norm": 19.618528366088867, "learning_rate": 2.4526841018582245e-05, "loss": 0.7482, "step": 296100 }, { "epoch": 50.95492085340675, "grad_norm": 16.006664276123047, "learning_rate": 2.452253957329663e-05, "loss": 0.7791, "step": 296150 }, { "epoch": 50.96352374397798, "grad_norm": 13.710865020751953, "learning_rate": 2.451823812801101e-05, "loss": 0.7411, "step": 296200 }, { "epoch": 50.972126634549205, "grad_norm": 12.650625228881836, "learning_rate": 2.4513936682725396e-05, "loss": 0.7695, "step": 296250 }, { "epoch": 50.98072952512044, "grad_norm": 21.642879486083984, "learning_rate": 2.4509635237439783e-05, "loss": 0.7743, "step": 296300 }, { "epoch": 50.98933241569167, "grad_norm": 18.63003921508789, "learning_rate": 2.4505333792154165e-05, "loss": 0.7582, "step": 296350 }, { "epoch": 50.9979353062629, "grad_norm": 9.86292552947998, "learning_rate": 2.450103234686855e-05, "loss": 0.7774, "step": 296400 }, { "epoch": 51.0, "eval_accuracy": 0.4461366374118052, "eval_f1": 0.4202032766109102, "eval_f1_DuraRiadoRio_16x16": 0.24388514711095358, "eval_f1_Mole_16x16": 0.5361216730038023, "eval_f1_Quebrado_16x16": 0.6539765319426336, "eval_f1_RiadoRio_16x16": 0.21045576407506703, "eval_f1_RioFechado_16x16": 0.4565772669220945, "eval_loss": 4.290386199951172, "eval_precision": 0.5894151851203343, "eval_precision_DuraRiadoRio_16x16": 0.6653771760154739, "eval_precision_Mole_16x16": 0.592436974789916, "eval_precision_Quebrado_16x16": 0.8190725016329197, "eval_precision_RiadoRio_16x16": 0.5688405797101449, "eval_precision_RioFechado_16x16": 0.3013486934532172, "eval_recall": 0.45077740329569904, "eval_recall_DuraRiadoRio_16x16": 0.14930555555555555, "eval_recall_Mole_16x16": 0.4895833333333333, "eval_recall_Quebrado_16x16": 0.5442708333333334, "eval_recall_RiadoRio_16x16": 0.12911184210526316, "eval_recall_RioFechado_16x16": 0.9416154521510096, "eval_runtime": 45.9852, "eval_samples_per_second": 252.734, "eval_steps_per_second": 15.809, "step": 296412 }, { "epoch": 51.006538196834136, "grad_norm": 15.629107475280762, "learning_rate": 2.4496730901582934e-05, "loss": 0.7624, "step": 296450 }, { "epoch": 51.01514108740537, "grad_norm": 14.150317192077637, "learning_rate": 2.449242945629732e-05, "loss": 0.743, "step": 296500 }, { "epoch": 51.0237439779766, "grad_norm": 16.561594009399414, "learning_rate": 2.4488128011011703e-05, "loss": 0.8102, "step": 296550 }, { "epoch": 51.03234686854783, "grad_norm": 15.204049110412598, "learning_rate": 2.4483826565726084e-05, "loss": 0.7981, "step": 296600 }, { "epoch": 51.040949759119066, "grad_norm": 18.412694931030273, "learning_rate": 2.447952512044047e-05, "loss": 0.7579, "step": 296650 }, { "epoch": 51.0495526496903, "grad_norm": 12.526202201843262, "learning_rate": 2.4475223675154853e-05, "loss": 0.7697, "step": 296700 }, { "epoch": 51.05815554026153, "grad_norm": 18.28898811340332, "learning_rate": 2.4470922229869238e-05, "loss": 0.8095, "step": 296750 }, { "epoch": 51.06675843083276, "grad_norm": 11.0765962600708, "learning_rate": 2.4466620784583622e-05, "loss": 0.7409, "step": 296800 }, { "epoch": 51.07536132140399, "grad_norm": 14.948156356811523, "learning_rate": 2.4462319339298007e-05, "loss": 0.7622, "step": 296850 }, { "epoch": 51.08396421197522, "grad_norm": 11.619377136230469, "learning_rate": 2.4458017894012388e-05, "loss": 0.797, "step": 296900 }, { "epoch": 51.092567102546454, "grad_norm": 14.735066413879395, "learning_rate": 2.4453716448726773e-05, "loss": 0.7839, "step": 296950 }, { "epoch": 51.101169993117686, "grad_norm": 9.65258502960205, "learning_rate": 2.4449415003441157e-05, "loss": 0.7925, "step": 297000 }, { "epoch": 51.10977288368892, "grad_norm": 18.488733291625977, "learning_rate": 2.4445113558155542e-05, "loss": 0.809, "step": 297050 }, { "epoch": 51.11837577426015, "grad_norm": 9.406966209411621, "learning_rate": 2.4440812112869926e-05, "loss": 0.7404, "step": 297100 }, { "epoch": 51.126978664831384, "grad_norm": 15.730734825134277, "learning_rate": 2.4436510667584307e-05, "loss": 0.7623, "step": 297150 }, { "epoch": 51.13558155540262, "grad_norm": 10.564815521240234, "learning_rate": 2.4432209222298692e-05, "loss": 0.8038, "step": 297200 }, { "epoch": 51.14418444597385, "grad_norm": 6.825389385223389, "learning_rate": 2.442790777701308e-05, "loss": 0.749, "step": 297250 }, { "epoch": 51.15278733654508, "grad_norm": 19.525959014892578, "learning_rate": 2.442360633172746e-05, "loss": 0.8046, "step": 297300 }, { "epoch": 51.161390227116314, "grad_norm": 12.621420860290527, "learning_rate": 2.4419304886441846e-05, "loss": 0.7538, "step": 297350 }, { "epoch": 51.16999311768754, "grad_norm": 20.7467098236084, "learning_rate": 2.4415003441156227e-05, "loss": 0.7585, "step": 297400 }, { "epoch": 51.17859600825877, "grad_norm": 12.941986083984375, "learning_rate": 2.4410701995870615e-05, "loss": 0.8231, "step": 297450 }, { "epoch": 51.187198898830005, "grad_norm": 13.436692237854004, "learning_rate": 2.4406400550585e-05, "loss": 0.7365, "step": 297500 }, { "epoch": 51.19580178940124, "grad_norm": 22.198959350585938, "learning_rate": 2.440209910529938e-05, "loss": 0.7244, "step": 297550 }, { "epoch": 51.20440467997247, "grad_norm": 6.792238712310791, "learning_rate": 2.4397797660013765e-05, "loss": 0.7238, "step": 297600 }, { "epoch": 51.2130075705437, "grad_norm": 10.233999252319336, "learning_rate": 2.439349621472815e-05, "loss": 0.7488, "step": 297650 }, { "epoch": 51.221610461114935, "grad_norm": 10.4973726272583, "learning_rate": 2.4389194769442534e-05, "loss": 0.7332, "step": 297700 }, { "epoch": 51.23021335168617, "grad_norm": 12.00600528717041, "learning_rate": 2.438489332415692e-05, "loss": 0.7608, "step": 297750 }, { "epoch": 51.2388162422574, "grad_norm": 14.488578796386719, "learning_rate": 2.43805918788713e-05, "loss": 0.8064, "step": 297800 }, { "epoch": 51.24741913282863, "grad_norm": 11.280583381652832, "learning_rate": 2.4376290433585685e-05, "loss": 0.7734, "step": 297850 }, { "epoch": 51.256022023399865, "grad_norm": 15.2959566116333, "learning_rate": 2.437198898830007e-05, "loss": 0.7765, "step": 297900 }, { "epoch": 51.2646249139711, "grad_norm": 11.560054779052734, "learning_rate": 2.4367687543014454e-05, "loss": 0.7889, "step": 297950 }, { "epoch": 51.27322780454232, "grad_norm": 11.409329414367676, "learning_rate": 2.4363386097728838e-05, "loss": 0.7915, "step": 298000 }, { "epoch": 51.281830695113555, "grad_norm": 16.321842193603516, "learning_rate": 2.4359084652443223e-05, "loss": 0.7695, "step": 298050 }, { "epoch": 51.29043358568479, "grad_norm": 17.14271354675293, "learning_rate": 2.4354783207157604e-05, "loss": 0.7345, "step": 298100 }, { "epoch": 51.29903647625602, "grad_norm": 9.585871696472168, "learning_rate": 2.4350481761871992e-05, "loss": 0.7822, "step": 298150 }, { "epoch": 51.30763936682725, "grad_norm": 16.081771850585938, "learning_rate": 2.4346180316586376e-05, "loss": 0.7652, "step": 298200 }, { "epoch": 51.316242257398486, "grad_norm": 12.09781551361084, "learning_rate": 2.4341878871300758e-05, "loss": 0.7765, "step": 298250 }, { "epoch": 51.32484514796972, "grad_norm": 9.645194053649902, "learning_rate": 2.4337577426015142e-05, "loss": 0.7837, "step": 298300 }, { "epoch": 51.33344803854095, "grad_norm": 10.146745681762695, "learning_rate": 2.4333275980729527e-05, "loss": 0.7799, "step": 298350 }, { "epoch": 51.34205092911218, "grad_norm": 12.682882308959961, "learning_rate": 2.432897453544391e-05, "loss": 0.7513, "step": 298400 }, { "epoch": 51.350653819683416, "grad_norm": 13.290048599243164, "learning_rate": 2.4324673090158296e-05, "loss": 0.7601, "step": 298450 }, { "epoch": 51.35925671025465, "grad_norm": 14.093371391296387, "learning_rate": 2.4320371644872677e-05, "loss": 0.7862, "step": 298500 }, { "epoch": 51.36785960082588, "grad_norm": 15.221283912658691, "learning_rate": 2.431607019958706e-05, "loss": 0.7728, "step": 298550 }, { "epoch": 51.376462491397106, "grad_norm": 18.864654541015625, "learning_rate": 2.4311768754301446e-05, "loss": 0.7528, "step": 298600 }, { "epoch": 51.38506538196834, "grad_norm": 15.181793212890625, "learning_rate": 2.430746730901583e-05, "loss": 0.7759, "step": 298650 }, { "epoch": 51.39366827253957, "grad_norm": 13.194987297058105, "learning_rate": 2.4303165863730215e-05, "loss": 0.7691, "step": 298700 }, { "epoch": 51.402271163110804, "grad_norm": 10.618470191955566, "learning_rate": 2.4298864418444596e-05, "loss": 0.7321, "step": 298750 }, { "epoch": 51.410874053682036, "grad_norm": 14.04926586151123, "learning_rate": 2.429456297315898e-05, "loss": 0.7287, "step": 298800 }, { "epoch": 51.41947694425327, "grad_norm": 24.524917602539062, "learning_rate": 2.429026152787337e-05, "loss": 0.7576, "step": 298850 }, { "epoch": 51.4280798348245, "grad_norm": 12.837124824523926, "learning_rate": 2.428596008258775e-05, "loss": 0.8219, "step": 298900 }, { "epoch": 51.436682725395734, "grad_norm": 14.50173568725586, "learning_rate": 2.4281658637302135e-05, "loss": 0.7216, "step": 298950 }, { "epoch": 51.44528561596697, "grad_norm": 14.34744644165039, "learning_rate": 2.427735719201652e-05, "loss": 0.8293, "step": 299000 }, { "epoch": 51.4538885065382, "grad_norm": 12.109803199768066, "learning_rate": 2.4273055746730904e-05, "loss": 0.763, "step": 299050 }, { "epoch": 51.46249139710943, "grad_norm": 12.40375804901123, "learning_rate": 2.426875430144529e-05, "loss": 0.8551, "step": 299100 }, { "epoch": 51.471094287680664, "grad_norm": 8.563041687011719, "learning_rate": 2.426445285615967e-05, "loss": 0.759, "step": 299150 }, { "epoch": 51.47969717825189, "grad_norm": 14.555932998657227, "learning_rate": 2.4260151410874054e-05, "loss": 0.7574, "step": 299200 }, { "epoch": 51.48830006882312, "grad_norm": 15.35213565826416, "learning_rate": 2.425584996558844e-05, "loss": 0.7537, "step": 299250 }, { "epoch": 51.496902959394355, "grad_norm": 9.550514221191406, "learning_rate": 2.4251548520302823e-05, "loss": 0.7829, "step": 299300 }, { "epoch": 51.50550584996559, "grad_norm": 15.838205337524414, "learning_rate": 2.4247247075017208e-05, "loss": 0.7963, "step": 299350 }, { "epoch": 51.51410874053682, "grad_norm": 8.773026466369629, "learning_rate": 2.4242945629731592e-05, "loss": 0.7709, "step": 299400 }, { "epoch": 51.52271163110805, "grad_norm": 13.093844413757324, "learning_rate": 2.4238644184445974e-05, "loss": 0.7448, "step": 299450 }, { "epoch": 51.531314521679285, "grad_norm": 14.698826789855957, "learning_rate": 2.4234342739160358e-05, "loss": 0.7424, "step": 299500 }, { "epoch": 51.53991741225052, "grad_norm": 13.37062931060791, "learning_rate": 2.4230041293874743e-05, "loss": 0.7603, "step": 299550 }, { "epoch": 51.54852030282175, "grad_norm": 11.976445198059082, "learning_rate": 2.4225739848589127e-05, "loss": 0.7638, "step": 299600 }, { "epoch": 51.55712319339298, "grad_norm": 15.297161102294922, "learning_rate": 2.4221438403303512e-05, "loss": 0.7683, "step": 299650 }, { "epoch": 51.565726083964215, "grad_norm": 12.424463272094727, "learning_rate": 2.4217136958017893e-05, "loss": 0.7896, "step": 299700 }, { "epoch": 51.57432897453545, "grad_norm": 9.840136528015137, "learning_rate": 2.421283551273228e-05, "loss": 0.8082, "step": 299750 }, { "epoch": 51.58293186510667, "grad_norm": 13.171178817749023, "learning_rate": 2.4208534067446665e-05, "loss": 0.7833, "step": 299800 }, { "epoch": 51.591534755677905, "grad_norm": 10.760997772216797, "learning_rate": 2.4204232622161047e-05, "loss": 0.718, "step": 299850 }, { "epoch": 51.60013764624914, "grad_norm": 12.232828140258789, "learning_rate": 2.419993117687543e-05, "loss": 0.7569, "step": 299900 }, { "epoch": 51.60874053682037, "grad_norm": 14.65766429901123, "learning_rate": 2.4195629731589816e-05, "loss": 0.7606, "step": 299950 }, { "epoch": 51.6173434273916, "grad_norm": 15.880827903747559, "learning_rate": 2.41913282863042e-05, "loss": 0.7846, "step": 300000 }, { "epoch": 51.625946317962836, "grad_norm": 11.956894874572754, "learning_rate": 2.4187026841018585e-05, "loss": 0.7701, "step": 300050 }, { "epoch": 51.63454920853407, "grad_norm": 12.168726921081543, "learning_rate": 2.4182725395732966e-05, "loss": 0.7968, "step": 300100 }, { "epoch": 51.6431520991053, "grad_norm": 8.523087501525879, "learning_rate": 2.417842395044735e-05, "loss": 0.7778, "step": 300150 }, { "epoch": 51.65175498967653, "grad_norm": 9.223657608032227, "learning_rate": 2.4174122505161735e-05, "loss": 0.7487, "step": 300200 }, { "epoch": 51.660357880247766, "grad_norm": 17.185081481933594, "learning_rate": 2.416982105987612e-05, "loss": 0.7829, "step": 300250 }, { "epoch": 51.668960770819, "grad_norm": 14.04483699798584, "learning_rate": 2.4165519614590504e-05, "loss": 0.7913, "step": 300300 }, { "epoch": 51.677563661390224, "grad_norm": 15.633652687072754, "learning_rate": 2.416121816930489e-05, "loss": 0.7869, "step": 300350 }, { "epoch": 51.686166551961456, "grad_norm": 10.091538429260254, "learning_rate": 2.415691672401927e-05, "loss": 0.7734, "step": 300400 }, { "epoch": 51.69476944253269, "grad_norm": 14.107457160949707, "learning_rate": 2.4152615278733658e-05, "loss": 0.7843, "step": 300450 }, { "epoch": 51.70337233310392, "grad_norm": 8.989644050598145, "learning_rate": 2.414831383344804e-05, "loss": 0.7179, "step": 300500 }, { "epoch": 51.711975223675154, "grad_norm": 9.965363502502441, "learning_rate": 2.4144012388162424e-05, "loss": 0.8085, "step": 300550 }, { "epoch": 51.720578114246386, "grad_norm": 14.104440689086914, "learning_rate": 2.4139710942876808e-05, "loss": 0.76, "step": 300600 }, { "epoch": 51.72918100481762, "grad_norm": 13.695221900939941, "learning_rate": 2.413540949759119e-05, "loss": 0.7544, "step": 300650 }, { "epoch": 51.73778389538885, "grad_norm": 12.408438682556152, "learning_rate": 2.4131108052305577e-05, "loss": 0.8481, "step": 300700 }, { "epoch": 51.746386785960084, "grad_norm": 10.065352439880371, "learning_rate": 2.4126806607019962e-05, "loss": 0.7908, "step": 300750 }, { "epoch": 51.754989676531316, "grad_norm": 10.65684700012207, "learning_rate": 2.4122505161734343e-05, "loss": 0.7391, "step": 300800 }, { "epoch": 51.76359256710255, "grad_norm": 12.42164421081543, "learning_rate": 2.4118203716448728e-05, "loss": 0.7299, "step": 300850 }, { "epoch": 51.77219545767378, "grad_norm": 13.21921157836914, "learning_rate": 2.4113902271163112e-05, "loss": 0.7889, "step": 300900 }, { "epoch": 51.78079834824501, "grad_norm": 13.215250015258789, "learning_rate": 2.4109600825877497e-05, "loss": 0.7887, "step": 300950 }, { "epoch": 51.78940123881624, "grad_norm": 13.859885215759277, "learning_rate": 2.410529938059188e-05, "loss": 0.7837, "step": 301000 }, { "epoch": 51.79800412938747, "grad_norm": 10.141338348388672, "learning_rate": 2.4100997935306263e-05, "loss": 0.7769, "step": 301050 }, { "epoch": 51.806607019958705, "grad_norm": 12.97364330291748, "learning_rate": 2.4096696490020647e-05, "loss": 0.749, "step": 301100 }, { "epoch": 51.81520991052994, "grad_norm": 11.852919578552246, "learning_rate": 2.409239504473503e-05, "loss": 0.7902, "step": 301150 }, { "epoch": 51.82381280110117, "grad_norm": 16.208675384521484, "learning_rate": 2.4088093599449416e-05, "loss": 0.7598, "step": 301200 }, { "epoch": 51.8324156916724, "grad_norm": 18.065711975097656, "learning_rate": 2.40837921541638e-05, "loss": 0.7418, "step": 301250 }, { "epoch": 51.841018582243635, "grad_norm": 9.789807319641113, "learning_rate": 2.4079490708878182e-05, "loss": 0.7566, "step": 301300 }, { "epoch": 51.84962147281487, "grad_norm": 14.218421936035156, "learning_rate": 2.4075189263592566e-05, "loss": 0.793, "step": 301350 }, { "epoch": 51.8582243633861, "grad_norm": 14.160931587219238, "learning_rate": 2.4070887818306954e-05, "loss": 0.8028, "step": 301400 }, { "epoch": 51.86682725395733, "grad_norm": 6.82868766784668, "learning_rate": 2.4066586373021336e-05, "loss": 0.7682, "step": 301450 }, { "epoch": 51.875430144528565, "grad_norm": 8.292623519897461, "learning_rate": 2.406228492773572e-05, "loss": 0.7363, "step": 301500 }, { "epoch": 51.88403303509979, "grad_norm": 9.861109733581543, "learning_rate": 2.4057983482450105e-05, "loss": 0.7843, "step": 301550 }, { "epoch": 51.89263592567102, "grad_norm": 15.961164474487305, "learning_rate": 2.405368203716449e-05, "loss": 0.7877, "step": 301600 }, { "epoch": 51.901238816242255, "grad_norm": 15.68397045135498, "learning_rate": 2.4049380591878874e-05, "loss": 0.7807, "step": 301650 }, { "epoch": 51.90984170681349, "grad_norm": 17.308698654174805, "learning_rate": 2.4045079146593255e-05, "loss": 0.7651, "step": 301700 }, { "epoch": 51.91844459738472, "grad_norm": 18.87973403930664, "learning_rate": 2.404077770130764e-05, "loss": 0.7646, "step": 301750 }, { "epoch": 51.92704748795595, "grad_norm": 16.995756149291992, "learning_rate": 2.4036476256022024e-05, "loss": 0.7494, "step": 301800 }, { "epoch": 51.935650378527185, "grad_norm": 15.620753288269043, "learning_rate": 2.403217481073641e-05, "loss": 0.8084, "step": 301850 }, { "epoch": 51.94425326909842, "grad_norm": 17.963764190673828, "learning_rate": 2.4027873365450793e-05, "loss": 0.7628, "step": 301900 }, { "epoch": 51.95285615966965, "grad_norm": 13.632562637329102, "learning_rate": 2.4023571920165178e-05, "loss": 0.7309, "step": 301950 }, { "epoch": 51.96145905024088, "grad_norm": 15.50508975982666, "learning_rate": 2.401927047487956e-05, "loss": 0.7571, "step": 302000 }, { "epoch": 51.970061940812116, "grad_norm": 13.678874969482422, "learning_rate": 2.4014969029593944e-05, "loss": 0.753, "step": 302050 }, { "epoch": 51.97866483138335, "grad_norm": 16.507688522338867, "learning_rate": 2.401066758430833e-05, "loss": 0.7962, "step": 302100 }, { "epoch": 51.987267721954574, "grad_norm": 10.40715217590332, "learning_rate": 2.4006366139022713e-05, "loss": 0.7945, "step": 302150 }, { "epoch": 51.995870612525806, "grad_norm": 22.953458786010742, "learning_rate": 2.4002064693737097e-05, "loss": 0.7166, "step": 302200 }, { "epoch": 52.0, "eval_accuracy": 0.5367406642574428, "eval_f1": 0.512233587371524, "eval_f1_DuraRiadoRio_16x16": 0.4539911308203991, "eval_f1_Mole_16x16": 0.5708905030938929, "eval_f1_Quebrado_16x16": 0.7344021634151052, "eval_f1_RiadoRio_16x16": 0.3127053669222344, "eval_f1_RioFechado_16x16": 0.48917877260598874, "eval_loss": 2.6085867881774902, "eval_precision": 0.5849695684161238, "eval_precision_DuraRiadoRio_16x16": 0.6280674846625767, "eval_precision_Mole_16x16": 0.41364522417154, "eval_precision_Quebrado_16x16": 0.6616776888270101, "eval_precision_RiadoRio_16x16": 0.4680327868852459, "eval_precision_RioFechado_16x16": 0.7534246575342466, "eval_recall": 0.5397016946998753, "eval_recall_DuraRiadoRio_16x16": 0.35546875, "eval_recall_Mole_16x16": 0.9210069444444444, "eval_recall_Quebrado_16x16": 0.8250868055555556, "eval_recall_RiadoRio_16x16": 0.23478618421052633, "eval_recall_RioFechado_16x16": 0.36215978928884984, "eval_runtime": 46.1014, "eval_samples_per_second": 252.097, "eval_steps_per_second": 15.77, "step": 302224 }, { "epoch": 52.00447350309704, "grad_norm": 12.418295860290527, "learning_rate": 2.399776324845148e-05, "loss": 0.7785, "step": 302250 }, { "epoch": 52.01307639366827, "grad_norm": 9.948176383972168, "learning_rate": 2.3993461803165866e-05, "loss": 0.7319, "step": 302300 }, { "epoch": 52.021679284239504, "grad_norm": 17.838016510009766, "learning_rate": 2.398916035788025e-05, "loss": 0.7426, "step": 302350 }, { "epoch": 52.030282174810736, "grad_norm": 19.595731735229492, "learning_rate": 2.3984858912594632e-05, "loss": 0.783, "step": 302400 }, { "epoch": 52.03888506538197, "grad_norm": 7.078357219696045, "learning_rate": 2.3980557467309017e-05, "loss": 0.7254, "step": 302450 }, { "epoch": 52.0474879559532, "grad_norm": 14.584649085998535, "learning_rate": 2.39762560220234e-05, "loss": 0.7329, "step": 302500 }, { "epoch": 52.056090846524434, "grad_norm": 11.408482551574707, "learning_rate": 2.3971954576737786e-05, "loss": 0.7368, "step": 302550 }, { "epoch": 52.064693737095666, "grad_norm": 14.552284240722656, "learning_rate": 2.396765313145217e-05, "loss": 0.7528, "step": 302600 }, { "epoch": 52.0732966276669, "grad_norm": 10.403069496154785, "learning_rate": 2.396335168616655e-05, "loss": 0.7523, "step": 302650 }, { "epoch": 52.08189951823813, "grad_norm": 24.032329559326172, "learning_rate": 2.3959050240880936e-05, "loss": 0.7585, "step": 302700 }, { "epoch": 52.09050240880936, "grad_norm": 16.991575241088867, "learning_rate": 2.395474879559532e-05, "loss": 0.8074, "step": 302750 }, { "epoch": 52.09910529938059, "grad_norm": 11.863080978393555, "learning_rate": 2.3950447350309705e-05, "loss": 0.764, "step": 302800 }, { "epoch": 52.10770818995182, "grad_norm": 11.782971382141113, "learning_rate": 2.394614590502409e-05, "loss": 0.7898, "step": 302850 }, { "epoch": 52.116311080523054, "grad_norm": 13.652034759521484, "learning_rate": 2.3941844459738474e-05, "loss": 0.804, "step": 302900 }, { "epoch": 52.12491397109429, "grad_norm": 14.41680908203125, "learning_rate": 2.3937543014452855e-05, "loss": 0.7677, "step": 302950 }, { "epoch": 52.13351686166552, "grad_norm": 20.002540588378906, "learning_rate": 2.3933241569167243e-05, "loss": 0.746, "step": 303000 }, { "epoch": 52.14211975223675, "grad_norm": 11.224177360534668, "learning_rate": 2.3928940123881625e-05, "loss": 0.7838, "step": 303050 }, { "epoch": 52.150722642807985, "grad_norm": 18.71002960205078, "learning_rate": 2.392463867859601e-05, "loss": 0.8601, "step": 303100 }, { "epoch": 52.15932553337922, "grad_norm": 17.111162185668945, "learning_rate": 2.3920337233310394e-05, "loss": 0.7842, "step": 303150 }, { "epoch": 52.16792842395045, "grad_norm": 10.87602710723877, "learning_rate": 2.3916035788024775e-05, "loss": 0.7784, "step": 303200 }, { "epoch": 52.17653131452168, "grad_norm": 8.144567489624023, "learning_rate": 2.3911734342739163e-05, "loss": 0.8019, "step": 303250 }, { "epoch": 52.18513420509291, "grad_norm": 11.692886352539062, "learning_rate": 2.3907432897453547e-05, "loss": 0.7397, "step": 303300 }, { "epoch": 52.19373709566414, "grad_norm": 10.910934448242188, "learning_rate": 2.390313145216793e-05, "loss": 0.8116, "step": 303350 }, { "epoch": 52.20233998623537, "grad_norm": 14.33903980255127, "learning_rate": 2.3898830006882313e-05, "loss": 0.7285, "step": 303400 }, { "epoch": 52.210942876806605, "grad_norm": 12.175540924072266, "learning_rate": 2.3894528561596698e-05, "loss": 0.7167, "step": 303450 }, { "epoch": 52.21954576737784, "grad_norm": 8.735443115234375, "learning_rate": 2.3890227116311082e-05, "loss": 0.7737, "step": 303500 }, { "epoch": 52.22814865794907, "grad_norm": 14.354621887207031, "learning_rate": 2.3885925671025467e-05, "loss": 0.7388, "step": 303550 }, { "epoch": 52.2367515485203, "grad_norm": 21.36038589477539, "learning_rate": 2.3881624225739848e-05, "loss": 0.7655, "step": 303600 }, { "epoch": 52.245354439091535, "grad_norm": 15.365500450134277, "learning_rate": 2.3877322780454233e-05, "loss": 0.7755, "step": 303650 }, { "epoch": 52.25395732966277, "grad_norm": 17.10702896118164, "learning_rate": 2.387302133516862e-05, "loss": 0.7176, "step": 303700 }, { "epoch": 52.262560220234, "grad_norm": 12.157805442810059, "learning_rate": 2.3868719889883e-05, "loss": 0.7572, "step": 303750 }, { "epoch": 52.27116311080523, "grad_norm": 11.154520034790039, "learning_rate": 2.3864418444597386e-05, "loss": 0.7675, "step": 303800 }, { "epoch": 52.279766001376466, "grad_norm": 13.914887428283691, "learning_rate": 2.386011699931177e-05, "loss": 0.7272, "step": 303850 }, { "epoch": 52.28836889194769, "grad_norm": 11.86023998260498, "learning_rate": 2.3855815554026152e-05, "loss": 0.8182, "step": 303900 }, { "epoch": 52.29697178251892, "grad_norm": 11.25808334350586, "learning_rate": 2.385151410874054e-05, "loss": 0.7952, "step": 303950 }, { "epoch": 52.305574673090156, "grad_norm": 12.751522064208984, "learning_rate": 2.384721266345492e-05, "loss": 0.7859, "step": 304000 }, { "epoch": 52.31417756366139, "grad_norm": 9.549040794372559, "learning_rate": 2.3842911218169306e-05, "loss": 0.7788, "step": 304050 }, { "epoch": 52.32278045423262, "grad_norm": 12.214509963989258, "learning_rate": 2.383860977288369e-05, "loss": 0.7537, "step": 304100 }, { "epoch": 52.331383344803854, "grad_norm": 13.286605834960938, "learning_rate": 2.3834308327598075e-05, "loss": 0.799, "step": 304150 }, { "epoch": 52.339986235375086, "grad_norm": 13.713906288146973, "learning_rate": 2.383000688231246e-05, "loss": 0.7428, "step": 304200 }, { "epoch": 52.34858912594632, "grad_norm": 10.80893325805664, "learning_rate": 2.3825705437026844e-05, "loss": 0.7729, "step": 304250 }, { "epoch": 52.35719201651755, "grad_norm": 24.65451431274414, "learning_rate": 2.3821403991741225e-05, "loss": 0.7545, "step": 304300 }, { "epoch": 52.365794907088784, "grad_norm": 13.946114540100098, "learning_rate": 2.381710254645561e-05, "loss": 0.8145, "step": 304350 }, { "epoch": 52.374397797660016, "grad_norm": 9.967823028564453, "learning_rate": 2.3812801101169994e-05, "loss": 0.7971, "step": 304400 }, { "epoch": 52.38300068823125, "grad_norm": 13.65175724029541, "learning_rate": 2.380849965588438e-05, "loss": 0.7869, "step": 304450 }, { "epoch": 52.391603578802474, "grad_norm": 6.90951681137085, "learning_rate": 2.3804198210598763e-05, "loss": 0.8126, "step": 304500 }, { "epoch": 52.40020646937371, "grad_norm": 11.697327613830566, "learning_rate": 2.3799896765313144e-05, "loss": 0.7581, "step": 304550 }, { "epoch": 52.40880935994494, "grad_norm": 9.734108924865723, "learning_rate": 2.379559532002753e-05, "loss": 0.7578, "step": 304600 }, { "epoch": 52.41741225051617, "grad_norm": 13.872208595275879, "learning_rate": 2.3791293874741917e-05, "loss": 0.7233, "step": 304650 }, { "epoch": 52.426015141087404, "grad_norm": 16.15709686279297, "learning_rate": 2.3786992429456298e-05, "loss": 0.8354, "step": 304700 }, { "epoch": 52.43461803165864, "grad_norm": 13.18346118927002, "learning_rate": 2.3782690984170683e-05, "loss": 0.7214, "step": 304750 }, { "epoch": 52.44322092222987, "grad_norm": 14.848281860351562, "learning_rate": 2.3778389538885064e-05, "loss": 0.7481, "step": 304800 }, { "epoch": 52.4518238128011, "grad_norm": 11.482230186462402, "learning_rate": 2.3774088093599452e-05, "loss": 0.7848, "step": 304850 }, { "epoch": 52.460426703372335, "grad_norm": 10.858389854431152, "learning_rate": 2.3769786648313836e-05, "loss": 0.6946, "step": 304900 }, { "epoch": 52.46902959394357, "grad_norm": 15.385319709777832, "learning_rate": 2.3765485203028218e-05, "loss": 0.7818, "step": 304950 }, { "epoch": 52.4776324845148, "grad_norm": 12.792436599731445, "learning_rate": 2.3761183757742602e-05, "loss": 0.7721, "step": 305000 }, { "epoch": 52.48623537508603, "grad_norm": 14.159947395324707, "learning_rate": 2.3756882312456987e-05, "loss": 0.746, "step": 305050 }, { "epoch": 52.49483826565726, "grad_norm": 10.376861572265625, "learning_rate": 2.375258086717137e-05, "loss": 0.7435, "step": 305100 }, { "epoch": 52.50344115622849, "grad_norm": 18.617406845092773, "learning_rate": 2.3748279421885756e-05, "loss": 0.7393, "step": 305150 }, { "epoch": 52.51204404679972, "grad_norm": 11.21811580657959, "learning_rate": 2.3743977976600137e-05, "loss": 0.7565, "step": 305200 }, { "epoch": 52.520646937370955, "grad_norm": 22.973642349243164, "learning_rate": 2.373967653131452e-05, "loss": 0.7812, "step": 305250 }, { "epoch": 52.52924982794219, "grad_norm": 15.252547264099121, "learning_rate": 2.3735375086028906e-05, "loss": 0.8127, "step": 305300 }, { "epoch": 52.53785271851342, "grad_norm": 18.689233779907227, "learning_rate": 2.373107364074329e-05, "loss": 0.757, "step": 305350 }, { "epoch": 52.54645560908465, "grad_norm": 13.310492515563965, "learning_rate": 2.3726772195457675e-05, "loss": 0.755, "step": 305400 }, { "epoch": 52.555058499655885, "grad_norm": 12.322959899902344, "learning_rate": 2.372247075017206e-05, "loss": 0.776, "step": 305450 }, { "epoch": 52.56366139022712, "grad_norm": 12.155099868774414, "learning_rate": 2.371816930488644e-05, "loss": 0.7759, "step": 305500 }, { "epoch": 52.57226428079835, "grad_norm": 16.569896697998047, "learning_rate": 2.371386785960083e-05, "loss": 0.7464, "step": 305550 }, { "epoch": 52.58086717136958, "grad_norm": 12.533021926879883, "learning_rate": 2.3709566414315213e-05, "loss": 0.7681, "step": 305600 }, { "epoch": 52.589470061940816, "grad_norm": 7.816224098205566, "learning_rate": 2.3705264969029595e-05, "loss": 0.7686, "step": 305650 }, { "epoch": 52.59807295251204, "grad_norm": 11.866289138793945, "learning_rate": 2.370096352374398e-05, "loss": 0.7529, "step": 305700 }, { "epoch": 52.60667584308327, "grad_norm": 8.610618591308594, "learning_rate": 2.369666207845836e-05, "loss": 0.7402, "step": 305750 }, { "epoch": 52.615278733654506, "grad_norm": 11.697550773620605, "learning_rate": 2.369236063317275e-05, "loss": 0.7664, "step": 305800 }, { "epoch": 52.62388162422574, "grad_norm": 14.873139381408691, "learning_rate": 2.3688059187887133e-05, "loss": 0.7958, "step": 305850 }, { "epoch": 52.63248451479697, "grad_norm": 25.41938018798828, "learning_rate": 2.3683757742601514e-05, "loss": 0.7939, "step": 305900 }, { "epoch": 52.641087405368204, "grad_norm": 9.559340476989746, "learning_rate": 2.36794562973159e-05, "loss": 0.776, "step": 305950 }, { "epoch": 52.649690295939436, "grad_norm": 10.90863037109375, "learning_rate": 2.3675154852030283e-05, "loss": 0.7553, "step": 306000 }, { "epoch": 52.65829318651067, "grad_norm": 17.798717498779297, "learning_rate": 2.3670853406744668e-05, "loss": 0.7306, "step": 306050 }, { "epoch": 52.6668960770819, "grad_norm": 12.69980239868164, "learning_rate": 2.3666551961459052e-05, "loss": 0.8115, "step": 306100 }, { "epoch": 52.675498967653134, "grad_norm": 16.138500213623047, "learning_rate": 2.3662250516173433e-05, "loss": 0.8488, "step": 306150 }, { "epoch": 52.684101858224366, "grad_norm": 16.222562789916992, "learning_rate": 2.3657949070887818e-05, "loss": 0.7518, "step": 306200 }, { "epoch": 52.69270474879559, "grad_norm": 15.639348030090332, "learning_rate": 2.3653647625602206e-05, "loss": 0.7907, "step": 306250 }, { "epoch": 52.701307639366824, "grad_norm": 12.335906982421875, "learning_rate": 2.3649346180316587e-05, "loss": 0.7863, "step": 306300 }, { "epoch": 52.70991052993806, "grad_norm": 13.428690910339355, "learning_rate": 2.3645044735030972e-05, "loss": 0.7778, "step": 306350 }, { "epoch": 52.71851342050929, "grad_norm": 10.461819648742676, "learning_rate": 2.3640743289745356e-05, "loss": 0.7636, "step": 306400 }, { "epoch": 52.72711631108052, "grad_norm": 14.620655059814453, "learning_rate": 2.3636441844459737e-05, "loss": 0.7394, "step": 306450 }, { "epoch": 52.735719201651754, "grad_norm": 14.748247146606445, "learning_rate": 2.3632140399174125e-05, "loss": 0.7679, "step": 306500 }, { "epoch": 52.74432209222299, "grad_norm": 11.24390983581543, "learning_rate": 2.3627838953888507e-05, "loss": 0.7909, "step": 306550 }, { "epoch": 52.75292498279422, "grad_norm": 17.099130630493164, "learning_rate": 2.362353750860289e-05, "loss": 0.7854, "step": 306600 }, { "epoch": 52.76152787336545, "grad_norm": 12.475773811340332, "learning_rate": 2.3619236063317276e-05, "loss": 0.7567, "step": 306650 }, { "epoch": 52.770130763936685, "grad_norm": 19.405492782592773, "learning_rate": 2.361493461803166e-05, "loss": 0.8273, "step": 306700 }, { "epoch": 52.77873365450792, "grad_norm": 20.430673599243164, "learning_rate": 2.3610633172746045e-05, "loss": 0.7441, "step": 306750 }, { "epoch": 52.78733654507915, "grad_norm": 10.449464797973633, "learning_rate": 2.360633172746043e-05, "loss": 0.7348, "step": 306800 }, { "epoch": 52.795939435650375, "grad_norm": 11.662518501281738, "learning_rate": 2.360203028217481e-05, "loss": 0.7943, "step": 306850 }, { "epoch": 52.80454232622161, "grad_norm": 11.564478874206543, "learning_rate": 2.3597728836889195e-05, "loss": 0.7444, "step": 306900 }, { "epoch": 52.81314521679284, "grad_norm": 12.058732032775879, "learning_rate": 2.359342739160358e-05, "loss": 0.7872, "step": 306950 }, { "epoch": 52.82174810736407, "grad_norm": 11.910516738891602, "learning_rate": 2.3589125946317964e-05, "loss": 0.7834, "step": 307000 }, { "epoch": 52.830350997935305, "grad_norm": 14.726624488830566, "learning_rate": 2.358482450103235e-05, "loss": 0.8184, "step": 307050 }, { "epoch": 52.83895388850654, "grad_norm": 11.858973503112793, "learning_rate": 2.358052305574673e-05, "loss": 0.7733, "step": 307100 }, { "epoch": 52.84755677907777, "grad_norm": 10.9407320022583, "learning_rate": 2.3576221610461115e-05, "loss": 0.7682, "step": 307150 }, { "epoch": 52.856159669649, "grad_norm": 10.248597145080566, "learning_rate": 2.3571920165175502e-05, "loss": 0.7567, "step": 307200 }, { "epoch": 52.864762560220235, "grad_norm": 18.780954360961914, "learning_rate": 2.3567618719889884e-05, "loss": 0.7646, "step": 307250 }, { "epoch": 52.87336545079147, "grad_norm": 12.632094383239746, "learning_rate": 2.3563317274604268e-05, "loss": 0.7898, "step": 307300 }, { "epoch": 52.8819683413627, "grad_norm": 9.268349647521973, "learning_rate": 2.355901582931865e-05, "loss": 0.7666, "step": 307350 }, { "epoch": 52.89057123193393, "grad_norm": 12.314298629760742, "learning_rate": 2.3554714384033037e-05, "loss": 0.7594, "step": 307400 }, { "epoch": 52.89917412250516, "grad_norm": 11.156058311462402, "learning_rate": 2.3550412938747422e-05, "loss": 0.7485, "step": 307450 }, { "epoch": 52.90777701307639, "grad_norm": 18.99005889892578, "learning_rate": 2.3546111493461803e-05, "loss": 0.7845, "step": 307500 }, { "epoch": 52.91637990364762, "grad_norm": 15.489055633544922, "learning_rate": 2.3541810048176188e-05, "loss": 0.8179, "step": 307550 }, { "epoch": 52.924982794218856, "grad_norm": 15.53054428100586, "learning_rate": 2.3537508602890572e-05, "loss": 0.7875, "step": 307600 }, { "epoch": 52.93358568479009, "grad_norm": 12.224893569946289, "learning_rate": 2.3533207157604957e-05, "loss": 0.7234, "step": 307650 }, { "epoch": 52.94218857536132, "grad_norm": 18.173694610595703, "learning_rate": 2.352890571231934e-05, "loss": 0.798, "step": 307700 }, { "epoch": 52.950791465932554, "grad_norm": 10.915633201599121, "learning_rate": 2.3524604267033726e-05, "loss": 0.8092, "step": 307750 }, { "epoch": 52.959394356503786, "grad_norm": 15.89871883392334, "learning_rate": 2.3520302821748107e-05, "loss": 0.7481, "step": 307800 }, { "epoch": 52.96799724707502, "grad_norm": 14.76341724395752, "learning_rate": 2.351600137646249e-05, "loss": 0.7955, "step": 307850 }, { "epoch": 52.97660013764625, "grad_norm": 11.476591110229492, "learning_rate": 2.3511699931176876e-05, "loss": 0.7515, "step": 307900 }, { "epoch": 52.985203028217484, "grad_norm": 14.84432315826416, "learning_rate": 2.350739848589126e-05, "loss": 0.7578, "step": 307950 }, { "epoch": 52.993805918788716, "grad_norm": 17.80625343322754, "learning_rate": 2.3503097040605645e-05, "loss": 0.7684, "step": 308000 }, { "epoch": 53.0, "eval_accuracy": 0.5268456375838926, "eval_f1": 0.4902572186935166, "eval_f1_DuraRiadoRio_16x16": 0.2692837465564738, "eval_f1_Mole_16x16": 0.5977703061405061, "eval_f1_Quebrado_16x16": 0.6315164117035497, "eval_f1_RiadoRio_16x16": 0.37836397548627765, "eval_f1_RioFechado_16x16": 0.5743516535807757, "eval_loss": 2.5523035526275635, "eval_precision": 0.5601600683419872, "eval_precision_DuraRiadoRio_16x16": 0.6516666666666666, "eval_precision_Mole_16x16": 0.5046310128473259, "eval_precision_Quebrado_16x16": 0.4800180627681192, "eval_precision_RiadoRio_16x16": 0.5374716124148372, "eval_precision_RioFechado_16x16": 0.627012987012987, "eval_recall": 0.5294624738151349, "eval_recall_DuraRiadoRio_16x16": 0.1697048611111111, "eval_recall_Mole_16x16": 0.7330729166666666, "eval_recall_Quebrado_16x16": 0.9227430555555556, "eval_recall_RiadoRio_16x16": 0.29194078947368424, "eval_recall_RioFechado_16x16": 0.5298507462686567, "eval_runtime": 46.5131, "eval_samples_per_second": 249.865, "eval_steps_per_second": 15.63, "step": 308036 }, { "epoch": 53.00240880935994, "grad_norm": 15.939867973327637, "learning_rate": 2.3498795595320026e-05, "loss": 0.7566, "step": 308050 }, { "epoch": 53.011011699931174, "grad_norm": 16.98733139038086, "learning_rate": 2.3494494150034414e-05, "loss": 0.802, "step": 308100 }, { "epoch": 53.01961459050241, "grad_norm": 13.769038200378418, "learning_rate": 2.34901927047488e-05, "loss": 0.6718, "step": 308150 }, { "epoch": 53.02821748107364, "grad_norm": 9.89223575592041, "learning_rate": 2.348589125946318e-05, "loss": 0.6819, "step": 308200 }, { "epoch": 53.03682037164487, "grad_norm": 20.047216415405273, "learning_rate": 2.3481589814177565e-05, "loss": 0.741, "step": 308250 }, { "epoch": 53.045423262216104, "grad_norm": 15.652968406677246, "learning_rate": 2.347728836889195e-05, "loss": 0.8479, "step": 308300 }, { "epoch": 53.05402615278734, "grad_norm": 14.38909912109375, "learning_rate": 2.3472986923606334e-05, "loss": 0.7675, "step": 308350 }, { "epoch": 53.06262904335857, "grad_norm": 22.521135330200195, "learning_rate": 2.346868547832072e-05, "loss": 0.7308, "step": 308400 }, { "epoch": 53.0712319339298, "grad_norm": 15.249810218811035, "learning_rate": 2.34643840330351e-05, "loss": 0.7023, "step": 308450 }, { "epoch": 53.079834824501035, "grad_norm": 7.447800636291504, "learning_rate": 2.3460082587749484e-05, "loss": 0.7504, "step": 308500 }, { "epoch": 53.08843771507227, "grad_norm": 11.006595611572266, "learning_rate": 2.345578114246387e-05, "loss": 0.7224, "step": 308550 }, { "epoch": 53.0970406056435, "grad_norm": 15.788049697875977, "learning_rate": 2.3451479697178253e-05, "loss": 0.7727, "step": 308600 }, { "epoch": 53.105643496214725, "grad_norm": 13.265034675598145, "learning_rate": 2.3447178251892638e-05, "loss": 0.7331, "step": 308650 }, { "epoch": 53.11424638678596, "grad_norm": 12.940879821777344, "learning_rate": 2.344287680660702e-05, "loss": 0.7841, "step": 308700 }, { "epoch": 53.12284927735719, "grad_norm": 21.21038055419922, "learning_rate": 2.3438575361321404e-05, "loss": 0.7493, "step": 308750 }, { "epoch": 53.13145216792842, "grad_norm": 18.015817642211914, "learning_rate": 2.343427391603579e-05, "loss": 0.7051, "step": 308800 }, { "epoch": 53.140055058499655, "grad_norm": 11.523412704467773, "learning_rate": 2.3429972470750173e-05, "loss": 0.7333, "step": 308850 }, { "epoch": 53.14865794907089, "grad_norm": 10.60586929321289, "learning_rate": 2.3425671025464557e-05, "loss": 0.7482, "step": 308900 }, { "epoch": 53.15726083964212, "grad_norm": 15.611838340759277, "learning_rate": 2.3421369580178942e-05, "loss": 0.8288, "step": 308950 }, { "epoch": 53.16586373021335, "grad_norm": 13.66171932220459, "learning_rate": 2.3417068134893326e-05, "loss": 0.8183, "step": 309000 }, { "epoch": 53.174466620784585, "grad_norm": 8.10567855834961, "learning_rate": 2.341276668960771e-05, "loss": 0.7379, "step": 309050 }, { "epoch": 53.18306951135582, "grad_norm": 11.211870193481445, "learning_rate": 2.3408465244322092e-05, "loss": 0.784, "step": 309100 }, { "epoch": 53.19167240192705, "grad_norm": 17.854694366455078, "learning_rate": 2.3404163799036477e-05, "loss": 0.7626, "step": 309150 }, { "epoch": 53.20027529249828, "grad_norm": 13.009683609008789, "learning_rate": 2.339986235375086e-05, "loss": 0.7393, "step": 309200 }, { "epoch": 53.20887818306951, "grad_norm": 19.042388916015625, "learning_rate": 2.3395560908465246e-05, "loss": 0.714, "step": 309250 }, { "epoch": 53.21748107364074, "grad_norm": 12.386651992797852, "learning_rate": 2.339125946317963e-05, "loss": 0.873, "step": 309300 }, { "epoch": 53.22608396421197, "grad_norm": 10.329297065734863, "learning_rate": 2.3386958017894015e-05, "loss": 0.7223, "step": 309350 }, { "epoch": 53.234686854783206, "grad_norm": 14.563817977905273, "learning_rate": 2.3382656572608396e-05, "loss": 0.7305, "step": 309400 }, { "epoch": 53.24328974535444, "grad_norm": 9.451186180114746, "learning_rate": 2.337835512732278e-05, "loss": 0.7444, "step": 309450 }, { "epoch": 53.25189263592567, "grad_norm": 6.93917179107666, "learning_rate": 2.337405368203717e-05, "loss": 0.7142, "step": 309500 }, { "epoch": 53.2604955264969, "grad_norm": 13.02878189086914, "learning_rate": 2.336975223675155e-05, "loss": 0.7666, "step": 309550 }, { "epoch": 53.269098417068136, "grad_norm": 16.961994171142578, "learning_rate": 2.3365450791465934e-05, "loss": 0.72, "step": 309600 }, { "epoch": 53.27770130763937, "grad_norm": 16.81786346435547, "learning_rate": 2.3361149346180315e-05, "loss": 0.8191, "step": 309650 }, { "epoch": 53.2863041982106, "grad_norm": 16.277494430541992, "learning_rate": 2.33568479008947e-05, "loss": 0.6986, "step": 309700 }, { "epoch": 53.294907088781834, "grad_norm": 8.261411666870117, "learning_rate": 2.3352546455609088e-05, "loss": 0.7865, "step": 309750 }, { "epoch": 53.30350997935306, "grad_norm": 18.533918380737305, "learning_rate": 2.334824501032347e-05, "loss": 0.7681, "step": 309800 }, { "epoch": 53.31211286992429, "grad_norm": 20.951173782348633, "learning_rate": 2.3343943565037854e-05, "loss": 0.7504, "step": 309850 }, { "epoch": 53.320715760495524, "grad_norm": 14.577421188354492, "learning_rate": 2.3339642119752238e-05, "loss": 0.7629, "step": 309900 }, { "epoch": 53.32931865106676, "grad_norm": 16.776229858398438, "learning_rate": 2.3335340674466623e-05, "loss": 0.8129, "step": 309950 }, { "epoch": 53.33792154163799, "grad_norm": 11.101073265075684, "learning_rate": 2.3331039229181007e-05, "loss": 0.7451, "step": 310000 }, { "epoch": 53.34652443220922, "grad_norm": 12.12421703338623, "learning_rate": 2.332673778389539e-05, "loss": 0.7078, "step": 310050 }, { "epoch": 53.355127322780454, "grad_norm": 13.900710105895996, "learning_rate": 2.3322436338609773e-05, "loss": 0.7241, "step": 310100 }, { "epoch": 53.36373021335169, "grad_norm": 21.00037384033203, "learning_rate": 2.3318134893324158e-05, "loss": 0.7814, "step": 310150 }, { "epoch": 53.37233310392292, "grad_norm": 13.534565925598145, "learning_rate": 2.3313833448038542e-05, "loss": 0.7288, "step": 310200 }, { "epoch": 53.38093599449415, "grad_norm": 16.995569229125977, "learning_rate": 2.3309532002752927e-05, "loss": 0.7462, "step": 310250 }, { "epoch": 53.389538885065384, "grad_norm": 21.910186767578125, "learning_rate": 2.330523055746731e-05, "loss": 0.7976, "step": 310300 }, { "epoch": 53.39814177563662, "grad_norm": 16.5520076751709, "learning_rate": 2.3300929112181692e-05, "loss": 0.7602, "step": 310350 }, { "epoch": 53.40674466620784, "grad_norm": 14.575094223022461, "learning_rate": 2.3296627666896077e-05, "loss": 0.7188, "step": 310400 }, { "epoch": 53.415347556779075, "grad_norm": 17.160459518432617, "learning_rate": 2.329232622161046e-05, "loss": 0.7891, "step": 310450 }, { "epoch": 53.42395044735031, "grad_norm": 12.112200736999512, "learning_rate": 2.3288024776324846e-05, "loss": 0.7217, "step": 310500 }, { "epoch": 53.43255333792154, "grad_norm": 19.971006393432617, "learning_rate": 2.328372333103923e-05, "loss": 0.7586, "step": 310550 }, { "epoch": 53.44115622849277, "grad_norm": 14.09975814819336, "learning_rate": 2.3279421885753612e-05, "loss": 0.7632, "step": 310600 }, { "epoch": 53.449759119064005, "grad_norm": 11.637717247009277, "learning_rate": 2.3275120440468e-05, "loss": 0.7431, "step": 310650 }, { "epoch": 53.45836200963524, "grad_norm": 17.269590377807617, "learning_rate": 2.3270818995182384e-05, "loss": 0.8096, "step": 310700 }, { "epoch": 53.46696490020647, "grad_norm": 14.955405235290527, "learning_rate": 2.3266517549896766e-05, "loss": 0.7916, "step": 310750 }, { "epoch": 53.4755677907777, "grad_norm": 18.20618438720703, "learning_rate": 2.326221610461115e-05, "loss": 0.7364, "step": 310800 }, { "epoch": 53.484170681348935, "grad_norm": 12.734408378601074, "learning_rate": 2.3257914659325535e-05, "loss": 0.7576, "step": 310850 }, { "epoch": 53.49277357192017, "grad_norm": 11.366917610168457, "learning_rate": 2.325361321403992e-05, "loss": 0.7734, "step": 310900 }, { "epoch": 53.5013764624914, "grad_norm": 13.774531364440918, "learning_rate": 2.3249311768754304e-05, "loss": 0.7868, "step": 310950 }, { "epoch": 53.509979353062626, "grad_norm": 11.484539031982422, "learning_rate": 2.3245010323468685e-05, "loss": 0.7895, "step": 311000 }, { "epoch": 53.51858224363386, "grad_norm": 14.695635795593262, "learning_rate": 2.324070887818307e-05, "loss": 0.787, "step": 311050 }, { "epoch": 53.52718513420509, "grad_norm": 15.069669723510742, "learning_rate": 2.3236407432897454e-05, "loss": 0.8001, "step": 311100 }, { "epoch": 53.53578802477632, "grad_norm": 13.471025466918945, "learning_rate": 2.323210598761184e-05, "loss": 0.7326, "step": 311150 }, { "epoch": 53.544390915347556, "grad_norm": 13.472294807434082, "learning_rate": 2.3227804542326223e-05, "loss": 0.7809, "step": 311200 }, { "epoch": 53.55299380591879, "grad_norm": 10.326150894165039, "learning_rate": 2.3223503097040608e-05, "loss": 0.7481, "step": 311250 }, { "epoch": 53.56159669649002, "grad_norm": 13.76574993133545, "learning_rate": 2.321920165175499e-05, "loss": 0.7549, "step": 311300 }, { "epoch": 53.57019958706125, "grad_norm": 16.964181900024414, "learning_rate": 2.3214900206469377e-05, "loss": 0.7443, "step": 311350 }, { "epoch": 53.578802477632486, "grad_norm": 18.123342514038086, "learning_rate": 2.3210598761183758e-05, "loss": 0.7375, "step": 311400 }, { "epoch": 53.58740536820372, "grad_norm": 14.335343360900879, "learning_rate": 2.3206297315898143e-05, "loss": 0.748, "step": 311450 }, { "epoch": 53.59600825877495, "grad_norm": 10.643555641174316, "learning_rate": 2.3201995870612527e-05, "loss": 0.7891, "step": 311500 }, { "epoch": 53.604611149346184, "grad_norm": 16.265670776367188, "learning_rate": 2.3197694425326912e-05, "loss": 0.8045, "step": 311550 }, { "epoch": 53.61321403991741, "grad_norm": 16.215343475341797, "learning_rate": 2.3193392980041296e-05, "loss": 0.7168, "step": 311600 }, { "epoch": 53.62181693048864, "grad_norm": 7.132363319396973, "learning_rate": 2.318909153475568e-05, "loss": 0.7415, "step": 311650 }, { "epoch": 53.630419821059874, "grad_norm": 20.908540725708008, "learning_rate": 2.3184790089470062e-05, "loss": 0.7669, "step": 311700 }, { "epoch": 53.63902271163111, "grad_norm": 16.749780654907227, "learning_rate": 2.3180488644184447e-05, "loss": 0.7916, "step": 311750 }, { "epoch": 53.64762560220234, "grad_norm": 10.24003791809082, "learning_rate": 2.317618719889883e-05, "loss": 0.773, "step": 311800 }, { "epoch": 53.65622849277357, "grad_norm": 8.523088455200195, "learning_rate": 2.3171885753613216e-05, "loss": 0.738, "step": 311850 }, { "epoch": 53.664831383344804, "grad_norm": 12.432271957397461, "learning_rate": 2.31675843083276e-05, "loss": 0.7573, "step": 311900 }, { "epoch": 53.67343427391604, "grad_norm": 12.423921585083008, "learning_rate": 2.316328286304198e-05, "loss": 0.7593, "step": 311950 }, { "epoch": 53.68203716448727, "grad_norm": 18.417095184326172, "learning_rate": 2.3158981417756366e-05, "loss": 0.7798, "step": 312000 }, { "epoch": 53.6906400550585, "grad_norm": 16.51573371887207, "learning_rate": 2.3154679972470754e-05, "loss": 0.8168, "step": 312050 }, { "epoch": 53.699242945629734, "grad_norm": 14.112520217895508, "learning_rate": 2.3150378527185135e-05, "loss": 0.7477, "step": 312100 }, { "epoch": 53.70784583620097, "grad_norm": 10.46467399597168, "learning_rate": 2.314607708189952e-05, "loss": 0.7362, "step": 312150 }, { "epoch": 53.71644872677219, "grad_norm": 8.029268264770508, "learning_rate": 2.31417756366139e-05, "loss": 0.7637, "step": 312200 }, { "epoch": 53.725051617343425, "grad_norm": 16.693931579589844, "learning_rate": 2.313747419132829e-05, "loss": 0.7679, "step": 312250 }, { "epoch": 53.73365450791466, "grad_norm": 11.63377571105957, "learning_rate": 2.3133172746042673e-05, "loss": 0.7077, "step": 312300 }, { "epoch": 53.74225739848589, "grad_norm": 18.334218978881836, "learning_rate": 2.3128871300757055e-05, "loss": 0.786, "step": 312350 }, { "epoch": 53.75086028905712, "grad_norm": 9.556014060974121, "learning_rate": 2.312456985547144e-05, "loss": 0.7857, "step": 312400 }, { "epoch": 53.759463179628355, "grad_norm": 12.749064445495605, "learning_rate": 2.3120268410185824e-05, "loss": 0.775, "step": 312450 }, { "epoch": 53.76806607019959, "grad_norm": 17.12933349609375, "learning_rate": 2.3115966964900208e-05, "loss": 0.7592, "step": 312500 }, { "epoch": 53.77666896077082, "grad_norm": 15.919258117675781, "learning_rate": 2.3111665519614593e-05, "loss": 0.693, "step": 312550 }, { "epoch": 53.78527185134205, "grad_norm": 12.240635871887207, "learning_rate": 2.3107364074328974e-05, "loss": 0.7435, "step": 312600 }, { "epoch": 53.793874741913285, "grad_norm": 18.168916702270508, "learning_rate": 2.310306262904336e-05, "loss": 0.7945, "step": 312650 }, { "epoch": 53.80247763248452, "grad_norm": 11.093887329101562, "learning_rate": 2.3098761183757743e-05, "loss": 0.6898, "step": 312700 }, { "epoch": 53.81108052305575, "grad_norm": 17.296621322631836, "learning_rate": 2.3094459738472128e-05, "loss": 0.7444, "step": 312750 }, { "epoch": 53.819683413626976, "grad_norm": 21.823102951049805, "learning_rate": 2.3090158293186512e-05, "loss": 0.7405, "step": 312800 }, { "epoch": 53.82828630419821, "grad_norm": 17.483726501464844, "learning_rate": 2.3085856847900897e-05, "loss": 0.7871, "step": 312850 }, { "epoch": 53.83688919476944, "grad_norm": 16.36345863342285, "learning_rate": 2.3081555402615278e-05, "loss": 0.7558, "step": 312900 }, { "epoch": 53.84549208534067, "grad_norm": 12.0330810546875, "learning_rate": 2.3077253957329663e-05, "loss": 0.7822, "step": 312950 }, { "epoch": 53.854094975911906, "grad_norm": 13.593245506286621, "learning_rate": 2.307295251204405e-05, "loss": 0.7033, "step": 313000 }, { "epoch": 53.86269786648314, "grad_norm": 10.099747657775879, "learning_rate": 2.306865106675843e-05, "loss": 0.7476, "step": 313050 }, { "epoch": 53.87130075705437, "grad_norm": 10.539434432983398, "learning_rate": 2.3064349621472816e-05, "loss": 0.7371, "step": 313100 }, { "epoch": 53.8799036476256, "grad_norm": 10.553557395935059, "learning_rate": 2.3060048176187197e-05, "loss": 0.755, "step": 313150 }, { "epoch": 53.888506538196836, "grad_norm": 12.97243595123291, "learning_rate": 2.3055746730901585e-05, "loss": 0.7562, "step": 313200 }, { "epoch": 53.89710942876807, "grad_norm": 14.573647499084473, "learning_rate": 2.305144528561597e-05, "loss": 0.7567, "step": 313250 }, { "epoch": 53.9057123193393, "grad_norm": 8.006017684936523, "learning_rate": 2.304714384033035e-05, "loss": 0.7252, "step": 313300 }, { "epoch": 53.91431520991053, "grad_norm": 6.959365367889404, "learning_rate": 2.3042842395044736e-05, "loss": 0.7428, "step": 313350 }, { "epoch": 53.92291810048176, "grad_norm": 12.218621253967285, "learning_rate": 2.303854094975912e-05, "loss": 0.744, "step": 313400 }, { "epoch": 53.93152099105299, "grad_norm": 9.451750755310059, "learning_rate": 2.3034239504473505e-05, "loss": 0.7807, "step": 313450 }, { "epoch": 53.940123881624224, "grad_norm": 11.680028915405273, "learning_rate": 2.302993805918789e-05, "loss": 0.8109, "step": 313500 }, { "epoch": 53.94872677219546, "grad_norm": 11.191805839538574, "learning_rate": 2.302563661390227e-05, "loss": 0.7549, "step": 313550 }, { "epoch": 53.95732966276669, "grad_norm": 15.206026077270508, "learning_rate": 2.3021335168616655e-05, "loss": 0.7268, "step": 313600 }, { "epoch": 53.96593255333792, "grad_norm": 12.616798400878906, "learning_rate": 2.301703372333104e-05, "loss": 0.7472, "step": 313650 }, { "epoch": 53.974535443909154, "grad_norm": 13.436003684997559, "learning_rate": 2.3012732278045424e-05, "loss": 0.7816, "step": 313700 }, { "epoch": 53.98313833448039, "grad_norm": 16.410287857055664, "learning_rate": 2.300843083275981e-05, "loss": 0.7546, "step": 313750 }, { "epoch": 53.99174122505162, "grad_norm": 17.821826934814453, "learning_rate": 2.3004129387474193e-05, "loss": 0.7749, "step": 313800 }, { "epoch": 54.0, "eval_accuracy": 0.532266391326794, "eval_f1": 0.5273658160415824, "eval_f1_DuraRiadoRio_16x16": 0.49008642602948654, "eval_f1_Mole_16x16": 0.5428020919350399, "eval_f1_Quebrado_16x16": 0.7199063231850117, "eval_f1_RiadoRio_16x16": 0.35243475921442025, "eval_f1_RioFechado_16x16": 0.5315994798439532, "eval_loss": 2.167929172515869, "eval_precision": 0.6005039099230574, "eval_precision_DuraRiadoRio_16x16": 0.5914110429447853, "eval_precision_Mole_16x16": 0.7419112114371708, "eval_precision_Quebrado_16x16": 0.7817904374364191, "eval_precision_RiadoRio_16x16": 0.5097276264591439, "eval_precision_RioFechado_16x16": 0.37767923133776793, "eval_recall": 0.5360117666633293, "eval_recall_DuraRiadoRio_16x16": 0.4184027777777778, "eval_recall_Mole_16x16": 0.4279513888888889, "eval_recall_Quebrado_16x16": 0.6671006944444444, "eval_recall_RiadoRio_16x16": 0.26932565789473684, "eval_recall_RioFechado_16x16": 0.897278314310799, "eval_runtime": 46.7901, "eval_samples_per_second": 248.386, "eval_steps_per_second": 15.537, "step": 313848 }, { "epoch": 54.00034411562285, "grad_norm": 20.60027503967285, "learning_rate": 2.2999827942188574e-05, "loss": 0.785, "step": 313850 }, { "epoch": 54.008947006194084, "grad_norm": 19.911426544189453, "learning_rate": 2.2995526496902962e-05, "loss": 0.792, "step": 313900 }, { "epoch": 54.01754989676531, "grad_norm": 14.687210083007812, "learning_rate": 2.2991225051617344e-05, "loss": 0.7466, "step": 313950 }, { "epoch": 54.02615278733654, "grad_norm": 13.074063301086426, "learning_rate": 2.2986923606331728e-05, "loss": 0.7226, "step": 314000 }, { "epoch": 54.034755677907775, "grad_norm": 17.231351852416992, "learning_rate": 2.2982622161046113e-05, "loss": 0.7241, "step": 314050 }, { "epoch": 54.04335856847901, "grad_norm": 13.139083862304688, "learning_rate": 2.2978320715760497e-05, "loss": 0.7866, "step": 314100 }, { "epoch": 54.05196145905024, "grad_norm": 13.383219718933105, "learning_rate": 2.2974019270474882e-05, "loss": 0.7203, "step": 314150 }, { "epoch": 54.06056434962147, "grad_norm": 12.540266036987305, "learning_rate": 2.2969717825189266e-05, "loss": 0.7301, "step": 314200 }, { "epoch": 54.069167240192705, "grad_norm": 14.771297454833984, "learning_rate": 2.2965416379903648e-05, "loss": 0.7542, "step": 314250 }, { "epoch": 54.07777013076394, "grad_norm": 13.029250144958496, "learning_rate": 2.2961114934618032e-05, "loss": 0.802, "step": 314300 }, { "epoch": 54.08637302133517, "grad_norm": 6.42345666885376, "learning_rate": 2.2956813489332417e-05, "loss": 0.7753, "step": 314350 }, { "epoch": 54.0949759119064, "grad_norm": 15.06065559387207, "learning_rate": 2.29525120440468e-05, "loss": 0.8244, "step": 314400 }, { "epoch": 54.103578802477635, "grad_norm": 8.3702392578125, "learning_rate": 2.2948210598761186e-05, "loss": 0.7219, "step": 314450 }, { "epoch": 54.11218169304887, "grad_norm": 9.65748119354248, "learning_rate": 2.2943909153475567e-05, "loss": 0.7364, "step": 314500 }, { "epoch": 54.12078458362009, "grad_norm": 10.05184555053711, "learning_rate": 2.293960770818995e-05, "loss": 0.7831, "step": 314550 }, { "epoch": 54.129387474191326, "grad_norm": 23.851911544799805, "learning_rate": 2.293530626290434e-05, "loss": 0.8009, "step": 314600 }, { "epoch": 54.13799036476256, "grad_norm": 11.891393661499023, "learning_rate": 2.293100481761872e-05, "loss": 0.8051, "step": 314650 }, { "epoch": 54.14659325533379, "grad_norm": 10.751638412475586, "learning_rate": 2.2926703372333105e-05, "loss": 0.7216, "step": 314700 }, { "epoch": 54.15519614590502, "grad_norm": 13.775612831115723, "learning_rate": 2.2922401927047486e-05, "loss": 0.7254, "step": 314750 }, { "epoch": 54.163799036476256, "grad_norm": 17.036727905273438, "learning_rate": 2.2918100481761874e-05, "loss": 0.7961, "step": 314800 }, { "epoch": 54.17240192704749, "grad_norm": 10.225911140441895, "learning_rate": 2.291379903647626e-05, "loss": 0.7439, "step": 314850 }, { "epoch": 54.18100481761872, "grad_norm": 15.43553638458252, "learning_rate": 2.290949759119064e-05, "loss": 0.7299, "step": 314900 }, { "epoch": 54.18960770818995, "grad_norm": 10.22402572631836, "learning_rate": 2.2905196145905025e-05, "loss": 0.7356, "step": 314950 }, { "epoch": 54.198210598761186, "grad_norm": 10.183754920959473, "learning_rate": 2.290089470061941e-05, "loss": 0.7363, "step": 315000 }, { "epoch": 54.20681348933242, "grad_norm": 16.210981369018555, "learning_rate": 2.2896593255333794e-05, "loss": 0.7034, "step": 315050 }, { "epoch": 54.21541637990365, "grad_norm": 19.885318756103516, "learning_rate": 2.2892291810048178e-05, "loss": 0.7802, "step": 315100 }, { "epoch": 54.22401927047488, "grad_norm": 13.581716537475586, "learning_rate": 2.2887990364762563e-05, "loss": 0.7864, "step": 315150 }, { "epoch": 54.23262216104611, "grad_norm": 12.410229682922363, "learning_rate": 2.2883688919476944e-05, "loss": 0.8059, "step": 315200 }, { "epoch": 54.24122505161734, "grad_norm": 9.859700202941895, "learning_rate": 2.287938747419133e-05, "loss": 0.8074, "step": 315250 }, { "epoch": 54.249827942188574, "grad_norm": 9.096303939819336, "learning_rate": 2.2875086028905713e-05, "loss": 0.7861, "step": 315300 }, { "epoch": 54.25843083275981, "grad_norm": 15.314123153686523, "learning_rate": 2.2870784583620098e-05, "loss": 0.771, "step": 315350 }, { "epoch": 54.26703372333104, "grad_norm": 12.67319107055664, "learning_rate": 2.2866483138334482e-05, "loss": 0.7396, "step": 315400 }, { "epoch": 54.27563661390227, "grad_norm": 15.548863410949707, "learning_rate": 2.2862181693048863e-05, "loss": 0.7802, "step": 315450 }, { "epoch": 54.284239504473504, "grad_norm": 17.864404678344727, "learning_rate": 2.285788024776325e-05, "loss": 0.7293, "step": 315500 }, { "epoch": 54.29284239504474, "grad_norm": 13.010811805725098, "learning_rate": 2.2853578802477636e-05, "loss": 0.7894, "step": 315550 }, { "epoch": 54.30144528561597, "grad_norm": 11.726652145385742, "learning_rate": 2.2849277357192017e-05, "loss": 0.7613, "step": 315600 }, { "epoch": 54.3100481761872, "grad_norm": 16.479162216186523, "learning_rate": 2.28449759119064e-05, "loss": 0.7365, "step": 315650 }, { "epoch": 54.318651066758434, "grad_norm": 16.307998657226562, "learning_rate": 2.2840674466620783e-05, "loss": 0.7197, "step": 315700 }, { "epoch": 54.32725395732966, "grad_norm": 10.491943359375, "learning_rate": 2.283637302133517e-05, "loss": 0.7234, "step": 315750 }, { "epoch": 54.33585684790089, "grad_norm": 13.791342735290527, "learning_rate": 2.2832071576049555e-05, "loss": 0.7997, "step": 315800 }, { "epoch": 54.344459738472125, "grad_norm": 12.277112007141113, "learning_rate": 2.2827770130763937e-05, "loss": 0.746, "step": 315850 }, { "epoch": 54.35306262904336, "grad_norm": 14.81520938873291, "learning_rate": 2.282346868547832e-05, "loss": 0.8382, "step": 315900 }, { "epoch": 54.36166551961459, "grad_norm": 8.917716979980469, "learning_rate": 2.2819167240192706e-05, "loss": 0.7296, "step": 315950 }, { "epoch": 54.37026841018582, "grad_norm": 13.4008150100708, "learning_rate": 2.281486579490709e-05, "loss": 0.7664, "step": 316000 }, { "epoch": 54.378871300757055, "grad_norm": 15.780750274658203, "learning_rate": 2.2810564349621475e-05, "loss": 0.7426, "step": 316050 }, { "epoch": 54.38747419132829, "grad_norm": 18.082355499267578, "learning_rate": 2.2806262904335856e-05, "loss": 0.7099, "step": 316100 }, { "epoch": 54.39607708189952, "grad_norm": 8.046965599060059, "learning_rate": 2.280196145905024e-05, "loss": 0.732, "step": 316150 }, { "epoch": 54.40467997247075, "grad_norm": 23.645023345947266, "learning_rate": 2.279766001376463e-05, "loss": 0.763, "step": 316200 }, { "epoch": 54.413282863041985, "grad_norm": 15.912489891052246, "learning_rate": 2.279335856847901e-05, "loss": 0.7892, "step": 316250 }, { "epoch": 54.42188575361321, "grad_norm": 18.321531295776367, "learning_rate": 2.2789057123193394e-05, "loss": 0.7988, "step": 316300 }, { "epoch": 54.43048864418444, "grad_norm": 15.996809959411621, "learning_rate": 2.278475567790778e-05, "loss": 0.7762, "step": 316350 }, { "epoch": 54.439091534755676, "grad_norm": 11.118760108947754, "learning_rate": 2.278045423262216e-05, "loss": 0.7543, "step": 316400 }, { "epoch": 54.44769442532691, "grad_norm": 11.355432510375977, "learning_rate": 2.2776152787336548e-05, "loss": 0.746, "step": 316450 }, { "epoch": 54.45629731589814, "grad_norm": 11.355931282043457, "learning_rate": 2.277185134205093e-05, "loss": 0.7819, "step": 316500 }, { "epoch": 54.46490020646937, "grad_norm": 16.868637084960938, "learning_rate": 2.2767549896765314e-05, "loss": 0.7472, "step": 316550 }, { "epoch": 54.473503097040606, "grad_norm": 13.744749069213867, "learning_rate": 2.2763248451479698e-05, "loss": 0.7658, "step": 316600 }, { "epoch": 54.48210598761184, "grad_norm": 11.634510040283203, "learning_rate": 2.2758947006194083e-05, "loss": 0.7594, "step": 316650 }, { "epoch": 54.49070887818307, "grad_norm": 12.451416969299316, "learning_rate": 2.2754645560908467e-05, "loss": 0.7347, "step": 316700 }, { "epoch": 54.4993117687543, "grad_norm": 10.516653060913086, "learning_rate": 2.2750344115622852e-05, "loss": 0.7236, "step": 316750 }, { "epoch": 54.507914659325536, "grad_norm": 15.594688415527344, "learning_rate": 2.2746042670337233e-05, "loss": 0.7134, "step": 316800 }, { "epoch": 54.51651754989677, "grad_norm": 19.547199249267578, "learning_rate": 2.2741741225051618e-05, "loss": 0.7673, "step": 316850 }, { "epoch": 54.525120440467994, "grad_norm": 14.02823257446289, "learning_rate": 2.2737439779766002e-05, "loss": 0.7515, "step": 316900 }, { "epoch": 54.533723331039226, "grad_norm": 11.600542068481445, "learning_rate": 2.2733138334480387e-05, "loss": 0.8219, "step": 316950 }, { "epoch": 54.54232622161046, "grad_norm": 9.409997940063477, "learning_rate": 2.272883688919477e-05, "loss": 0.7804, "step": 317000 }, { "epoch": 54.55092911218169, "grad_norm": 13.749889373779297, "learning_rate": 2.2724535443909152e-05, "loss": 0.7399, "step": 317050 }, { "epoch": 54.559532002752924, "grad_norm": 13.241521835327148, "learning_rate": 2.2720233998623537e-05, "loss": 0.802, "step": 317100 }, { "epoch": 54.56813489332416, "grad_norm": 13.127296447753906, "learning_rate": 2.2715932553337925e-05, "loss": 0.7524, "step": 317150 }, { "epoch": 54.57673778389539, "grad_norm": 17.30290412902832, "learning_rate": 2.2711631108052306e-05, "loss": 0.7587, "step": 317200 }, { "epoch": 54.58534067446662, "grad_norm": 13.326567649841309, "learning_rate": 2.270732966276669e-05, "loss": 0.7025, "step": 317250 }, { "epoch": 54.593943565037854, "grad_norm": 12.457071304321289, "learning_rate": 2.2703028217481075e-05, "loss": 0.7598, "step": 317300 }, { "epoch": 54.60254645560909, "grad_norm": 17.582555770874023, "learning_rate": 2.269872677219546e-05, "loss": 0.754, "step": 317350 }, { "epoch": 54.61114934618032, "grad_norm": 7.032476902008057, "learning_rate": 2.2694425326909844e-05, "loss": 0.7736, "step": 317400 }, { "epoch": 54.61975223675155, "grad_norm": 11.224843978881836, "learning_rate": 2.2690123881624226e-05, "loss": 0.7428, "step": 317450 }, { "epoch": 54.62835512732278, "grad_norm": 6.600708961486816, "learning_rate": 2.268582243633861e-05, "loss": 0.7434, "step": 317500 }, { "epoch": 54.63695801789401, "grad_norm": 16.200786590576172, "learning_rate": 2.2681520991052995e-05, "loss": 0.7406, "step": 317550 }, { "epoch": 54.64556090846524, "grad_norm": 10.365516662597656, "learning_rate": 2.267721954576738e-05, "loss": 0.763, "step": 317600 }, { "epoch": 54.654163799036475, "grad_norm": 12.466631889343262, "learning_rate": 2.2672918100481764e-05, "loss": 0.7618, "step": 317650 }, { "epoch": 54.66276668960771, "grad_norm": 13.371328353881836, "learning_rate": 2.266861665519615e-05, "loss": 0.6982, "step": 317700 }, { "epoch": 54.67136958017894, "grad_norm": 18.748319625854492, "learning_rate": 2.266431520991053e-05, "loss": 0.7689, "step": 317750 }, { "epoch": 54.67997247075017, "grad_norm": 20.989295959472656, "learning_rate": 2.2660013764624914e-05, "loss": 0.7864, "step": 317800 }, { "epoch": 54.688575361321405, "grad_norm": 14.654935836791992, "learning_rate": 2.26557123193393e-05, "loss": 0.7929, "step": 317850 }, { "epoch": 54.69717825189264, "grad_norm": 11.793642044067383, "learning_rate": 2.2651410874053683e-05, "loss": 0.6876, "step": 317900 }, { "epoch": 54.70578114246387, "grad_norm": 18.881933212280273, "learning_rate": 2.2647109428768068e-05, "loss": 0.7741, "step": 317950 }, { "epoch": 54.7143840330351, "grad_norm": 12.841501235961914, "learning_rate": 2.264280798348245e-05, "loss": 0.7867, "step": 318000 }, { "epoch": 54.722986923606335, "grad_norm": 17.47322654724121, "learning_rate": 2.2638506538196837e-05, "loss": 0.7274, "step": 318050 }, { "epoch": 54.73158981417756, "grad_norm": 17.03144645690918, "learning_rate": 2.263420509291122e-05, "loss": 0.7505, "step": 318100 }, { "epoch": 54.74019270474879, "grad_norm": 11.593865394592285, "learning_rate": 2.2629903647625603e-05, "loss": 0.7842, "step": 318150 }, { "epoch": 54.748795595320026, "grad_norm": 11.721611022949219, "learning_rate": 2.2625602202339987e-05, "loss": 0.6981, "step": 318200 }, { "epoch": 54.75739848589126, "grad_norm": 16.159460067749023, "learning_rate": 2.262130075705437e-05, "loss": 0.7901, "step": 318250 }, { "epoch": 54.76600137646249, "grad_norm": 10.27210521697998, "learning_rate": 2.2616999311768756e-05, "loss": 0.7471, "step": 318300 }, { "epoch": 54.77460426703372, "grad_norm": 8.345108985900879, "learning_rate": 2.261269786648314e-05, "loss": 0.7875, "step": 318350 }, { "epoch": 54.783207157604956, "grad_norm": 15.654830932617188, "learning_rate": 2.2608396421197522e-05, "loss": 0.7786, "step": 318400 }, { "epoch": 54.79181004817619, "grad_norm": 13.196956634521484, "learning_rate": 2.2604094975911907e-05, "loss": 0.7322, "step": 318450 }, { "epoch": 54.80041293874742, "grad_norm": 9.540769577026367, "learning_rate": 2.259979353062629e-05, "loss": 0.7781, "step": 318500 }, { "epoch": 54.80901582931865, "grad_norm": 19.340389251708984, "learning_rate": 2.2595492085340676e-05, "loss": 0.7287, "step": 318550 }, { "epoch": 54.817618719889886, "grad_norm": 14.424331665039062, "learning_rate": 2.259119064005506e-05, "loss": 0.7309, "step": 318600 }, { "epoch": 54.82622161046112, "grad_norm": 14.98478889465332, "learning_rate": 2.2586889194769445e-05, "loss": 0.7171, "step": 318650 }, { "epoch": 54.834824501032344, "grad_norm": 19.40909194946289, "learning_rate": 2.2582587749483826e-05, "loss": 0.7976, "step": 318700 }, { "epoch": 54.843427391603576, "grad_norm": 25.45996856689453, "learning_rate": 2.2578286304198214e-05, "loss": 0.7924, "step": 318750 }, { "epoch": 54.85203028217481, "grad_norm": 23.98723030090332, "learning_rate": 2.2573984858912595e-05, "loss": 0.7314, "step": 318800 }, { "epoch": 54.86063317274604, "grad_norm": 13.686993598937988, "learning_rate": 2.256968341362698e-05, "loss": 0.7408, "step": 318850 }, { "epoch": 54.869236063317274, "grad_norm": 14.174334526062012, "learning_rate": 2.2565381968341364e-05, "loss": 0.7737, "step": 318900 }, { "epoch": 54.87783895388851, "grad_norm": 10.377445220947266, "learning_rate": 2.2561080523055745e-05, "loss": 0.7664, "step": 318950 }, { "epoch": 54.88644184445974, "grad_norm": 14.7946138381958, "learning_rate": 2.2556779077770133e-05, "loss": 0.7392, "step": 319000 }, { "epoch": 54.89504473503097, "grad_norm": 14.072505950927734, "learning_rate": 2.2552477632484518e-05, "loss": 0.7803, "step": 319050 }, { "epoch": 54.903647625602204, "grad_norm": 11.285295486450195, "learning_rate": 2.25481761871989e-05, "loss": 0.7655, "step": 319100 }, { "epoch": 54.91225051617344, "grad_norm": 13.106193542480469, "learning_rate": 2.2543874741913284e-05, "loss": 0.7855, "step": 319150 }, { "epoch": 54.92085340674467, "grad_norm": 9.3126802444458, "learning_rate": 2.2539573296627668e-05, "loss": 0.7823, "step": 319200 }, { "epoch": 54.929456297315895, "grad_norm": 9.85348892211914, "learning_rate": 2.2535271851342053e-05, "loss": 0.751, "step": 319250 }, { "epoch": 54.93805918788713, "grad_norm": 17.443708419799805, "learning_rate": 2.2530970406056437e-05, "loss": 0.7838, "step": 319300 }, { "epoch": 54.94666207845836, "grad_norm": 16.882450103759766, "learning_rate": 2.252666896077082e-05, "loss": 0.7615, "step": 319350 }, { "epoch": 54.95526496902959, "grad_norm": 11.538250923156738, "learning_rate": 2.2522367515485203e-05, "loss": 0.6701, "step": 319400 }, { "epoch": 54.963867859600825, "grad_norm": 13.759439468383789, "learning_rate": 2.251806607019959e-05, "loss": 0.8052, "step": 319450 }, { "epoch": 54.97247075017206, "grad_norm": 12.91269302368164, "learning_rate": 2.2513764624913972e-05, "loss": 0.7656, "step": 319500 }, { "epoch": 54.98107364074329, "grad_norm": 13.721251487731934, "learning_rate": 2.2509463179628357e-05, "loss": 0.7673, "step": 319550 }, { "epoch": 54.98967653131452, "grad_norm": 12.37458324432373, "learning_rate": 2.2505161734342738e-05, "loss": 0.7461, "step": 319600 }, { "epoch": 54.998279421885755, "grad_norm": 12.3622465133667, "learning_rate": 2.2500860289057122e-05, "loss": 0.7422, "step": 319650 }, { "epoch": 55.0, "eval_accuracy": 0.6202030631560833, "eval_f1": 0.6235375647136714, "eval_f1_DuraRiadoRio_16x16": 0.6101569713758079, "eval_f1_Mole_16x16": 0.585511982570806, "eval_f1_Quebrado_16x16": 0.7546210720887245, "eval_f1_RiadoRio_16x16": 0.568178176574267, "eval_f1_RioFechado_16x16": 0.5992196209587514, "eval_loss": 1.3643860816955566, "eval_precision": 0.6819486100547149, "eval_precision_DuraRiadoRio_16x16": 0.5310189649630344, "eval_precision_Mole_16x16": 0.7858187134502924, "eval_precision_Quebrado_16x16": 0.8068181818181818, "eval_precision_RiadoRio_16x16": 0.46547650301916516, "eval_precision_RioFechado_16x16": 0.8206106870229007, "eval_recall": 0.6186591792713421, "eval_recall_DuraRiadoRio_16x16": 0.7170138888888888, "eval_recall_Mole_16x16": 0.4665798611111111, "eval_recall_Quebrado_16x16": 0.7087673611111112, "eval_recall_RiadoRio_16x16": 0.7290296052631579, "eval_recall_RioFechado_16x16": 0.47190517998244075, "eval_runtime": 46.4161, "eval_samples_per_second": 250.387, "eval_steps_per_second": 15.663, "step": 319660 }, { "epoch": 55.00688231245699, "grad_norm": 13.523241996765137, "learning_rate": 2.249655884377151e-05, "loss": 0.7546, "step": 319700 }, { "epoch": 55.01548520302822, "grad_norm": 11.15059757232666, "learning_rate": 2.249225739848589e-05, "loss": 0.7452, "step": 319750 }, { "epoch": 55.02408809359945, "grad_norm": 14.503946304321289, "learning_rate": 2.2487955953200276e-05, "loss": 0.7309, "step": 319800 }, { "epoch": 55.03269098417068, "grad_norm": 21.391469955444336, "learning_rate": 2.248365450791466e-05, "loss": 0.745, "step": 319850 }, { "epoch": 55.04129387474191, "grad_norm": 16.005884170532227, "learning_rate": 2.2479353062629045e-05, "loss": 0.719, "step": 319900 }, { "epoch": 55.04989676531314, "grad_norm": 20.39803123474121, "learning_rate": 2.247505161734343e-05, "loss": 0.7671, "step": 319950 }, { "epoch": 55.058499655884376, "grad_norm": 8.459964752197266, "learning_rate": 2.247075017205781e-05, "loss": 0.7357, "step": 320000 }, { "epoch": 55.06710254645561, "grad_norm": 15.430909156799316, "learning_rate": 2.2466448726772196e-05, "loss": 0.7548, "step": 320050 }, { "epoch": 55.07570543702684, "grad_norm": 11.637102127075195, "learning_rate": 2.246214728148658e-05, "loss": 0.7549, "step": 320100 }, { "epoch": 55.08430832759807, "grad_norm": 12.32205867767334, "learning_rate": 2.2457845836200965e-05, "loss": 0.7631, "step": 320150 }, { "epoch": 55.092911218169306, "grad_norm": 14.69284725189209, "learning_rate": 2.245354439091535e-05, "loss": 0.7705, "step": 320200 }, { "epoch": 55.10151410874054, "grad_norm": 9.980154991149902, "learning_rate": 2.2449242945629734e-05, "loss": 0.7726, "step": 320250 }, { "epoch": 55.11011699931177, "grad_norm": 10.765692710876465, "learning_rate": 2.2444941500344115e-05, "loss": 0.763, "step": 320300 }, { "epoch": 55.118719889883, "grad_norm": 17.068485260009766, "learning_rate": 2.24406400550585e-05, "loss": 0.711, "step": 320350 }, { "epoch": 55.127322780454236, "grad_norm": 17.078754425048828, "learning_rate": 2.2436338609772884e-05, "loss": 0.7792, "step": 320400 }, { "epoch": 55.13592567102546, "grad_norm": 13.229893684387207, "learning_rate": 2.243203716448727e-05, "loss": 0.7882, "step": 320450 }, { "epoch": 55.144528561596694, "grad_norm": 7.750113487243652, "learning_rate": 2.2427735719201653e-05, "loss": 0.77, "step": 320500 }, { "epoch": 55.153131452167926, "grad_norm": 8.609209060668945, "learning_rate": 2.2423434273916034e-05, "loss": 0.7442, "step": 320550 }, { "epoch": 55.16173434273916, "grad_norm": 15.463835716247559, "learning_rate": 2.2419132828630422e-05, "loss": 0.7286, "step": 320600 }, { "epoch": 55.17033723331039, "grad_norm": 11.08152961730957, "learning_rate": 2.2414831383344807e-05, "loss": 0.6905, "step": 320650 }, { "epoch": 55.178940123881624, "grad_norm": 13.652841567993164, "learning_rate": 2.2410529938059188e-05, "loss": 0.7308, "step": 320700 }, { "epoch": 55.18754301445286, "grad_norm": 13.550931930541992, "learning_rate": 2.2406228492773573e-05, "loss": 0.7767, "step": 320750 }, { "epoch": 55.19614590502409, "grad_norm": 12.613743782043457, "learning_rate": 2.2401927047487957e-05, "loss": 0.7723, "step": 320800 }, { "epoch": 55.20474879559532, "grad_norm": 22.943002700805664, "learning_rate": 2.2397625602202342e-05, "loss": 0.7583, "step": 320850 }, { "epoch": 55.213351686166554, "grad_norm": 16.80467987060547, "learning_rate": 2.2393324156916726e-05, "loss": 0.7454, "step": 320900 }, { "epoch": 55.22195457673779, "grad_norm": 10.072977066040039, "learning_rate": 2.2389022711631107e-05, "loss": 0.7606, "step": 320950 }, { "epoch": 55.23055746730902, "grad_norm": 8.853642463684082, "learning_rate": 2.2384721266345492e-05, "loss": 0.755, "step": 321000 }, { "epoch": 55.239160357880245, "grad_norm": 10.460075378417969, "learning_rate": 2.2380419821059877e-05, "loss": 0.7735, "step": 321050 }, { "epoch": 55.24776324845148, "grad_norm": 15.83518123626709, "learning_rate": 2.237611837577426e-05, "loss": 0.8571, "step": 321100 }, { "epoch": 55.25636613902271, "grad_norm": 7.295064449310303, "learning_rate": 2.2371816930488646e-05, "loss": 0.7188, "step": 321150 }, { "epoch": 55.26496902959394, "grad_norm": 12.87247085571289, "learning_rate": 2.236751548520303e-05, "loss": 0.7595, "step": 321200 }, { "epoch": 55.273571920165175, "grad_norm": 13.698345184326172, "learning_rate": 2.236321403991741e-05, "loss": 0.7996, "step": 321250 }, { "epoch": 55.28217481073641, "grad_norm": 16.285179138183594, "learning_rate": 2.23589125946318e-05, "loss": 0.7553, "step": 321300 }, { "epoch": 55.29077770130764, "grad_norm": 9.48444938659668, "learning_rate": 2.235461114934618e-05, "loss": 0.752, "step": 321350 }, { "epoch": 55.29938059187887, "grad_norm": 19.327863693237305, "learning_rate": 2.2350309704060565e-05, "loss": 0.7376, "step": 321400 }, { "epoch": 55.307983482450105, "grad_norm": 17.894657135009766, "learning_rate": 2.234600825877495e-05, "loss": 0.8158, "step": 321450 }, { "epoch": 55.31658637302134, "grad_norm": 13.500619888305664, "learning_rate": 2.2341706813489334e-05, "loss": 0.7372, "step": 321500 }, { "epoch": 55.32518926359257, "grad_norm": 18.630586624145508, "learning_rate": 2.233740536820372e-05, "loss": 0.7604, "step": 321550 }, { "epoch": 55.3337921541638, "grad_norm": 16.884218215942383, "learning_rate": 2.2333103922918103e-05, "loss": 0.7507, "step": 321600 }, { "epoch": 55.34239504473503, "grad_norm": 16.24567985534668, "learning_rate": 2.2328802477632485e-05, "loss": 0.7467, "step": 321650 }, { "epoch": 55.35099793530626, "grad_norm": 8.027886390686035, "learning_rate": 2.232450103234687e-05, "loss": 0.7, "step": 321700 }, { "epoch": 55.35960082587749, "grad_norm": 10.061492919921875, "learning_rate": 2.2320199587061254e-05, "loss": 0.7684, "step": 321750 }, { "epoch": 55.368203716448726, "grad_norm": 7.468623161315918, "learning_rate": 2.2315898141775638e-05, "loss": 0.6985, "step": 321800 }, { "epoch": 55.37680660701996, "grad_norm": 15.173102378845215, "learning_rate": 2.2311596696490023e-05, "loss": 0.7613, "step": 321850 }, { "epoch": 55.38540949759119, "grad_norm": 12.972891807556152, "learning_rate": 2.2307295251204404e-05, "loss": 0.7741, "step": 321900 }, { "epoch": 55.39401238816242, "grad_norm": 20.147106170654297, "learning_rate": 2.230299380591879e-05, "loss": 0.7188, "step": 321950 }, { "epoch": 55.402615278733656, "grad_norm": 11.726993560791016, "learning_rate": 2.2298692360633176e-05, "loss": 0.7419, "step": 322000 }, { "epoch": 55.41121816930489, "grad_norm": 10.453272819519043, "learning_rate": 2.2294390915347558e-05, "loss": 0.7821, "step": 322050 }, { "epoch": 55.41982105987612, "grad_norm": 14.909049034118652, "learning_rate": 2.2290089470061942e-05, "loss": 0.7348, "step": 322100 }, { "epoch": 55.42842395044735, "grad_norm": 14.821828842163086, "learning_rate": 2.2285788024776323e-05, "loss": 0.7583, "step": 322150 }, { "epoch": 55.43702684101858, "grad_norm": 13.537711143493652, "learning_rate": 2.2281486579490708e-05, "loss": 0.7548, "step": 322200 }, { "epoch": 55.44562973158981, "grad_norm": 16.457300186157227, "learning_rate": 2.2277185134205096e-05, "loss": 0.702, "step": 322250 }, { "epoch": 55.454232622161044, "grad_norm": 9.768155097961426, "learning_rate": 2.2272883688919477e-05, "loss": 0.7577, "step": 322300 }, { "epoch": 55.462835512732276, "grad_norm": 11.358742713928223, "learning_rate": 2.226858224363386e-05, "loss": 0.7338, "step": 322350 }, { "epoch": 55.47143840330351, "grad_norm": 21.66604995727539, "learning_rate": 2.2264280798348246e-05, "loss": 0.7406, "step": 322400 }, { "epoch": 55.48004129387474, "grad_norm": 18.198413848876953, "learning_rate": 2.225997935306263e-05, "loss": 0.799, "step": 322450 }, { "epoch": 55.488644184445974, "grad_norm": 15.12946605682373, "learning_rate": 2.2255677907777015e-05, "loss": 0.7315, "step": 322500 }, { "epoch": 55.497247075017206, "grad_norm": 6.747973442077637, "learning_rate": 2.22513764624914e-05, "loss": 0.8601, "step": 322550 }, { "epoch": 55.50584996558844, "grad_norm": 13.658706665039062, "learning_rate": 2.224707501720578e-05, "loss": 0.7457, "step": 322600 }, { "epoch": 55.51445285615967, "grad_norm": 18.614992141723633, "learning_rate": 2.2242773571920166e-05, "loss": 0.8302, "step": 322650 }, { "epoch": 55.523055746730904, "grad_norm": 10.87449836730957, "learning_rate": 2.223847212663455e-05, "loss": 0.7109, "step": 322700 }, { "epoch": 55.53165863730214, "grad_norm": 15.628812789916992, "learning_rate": 2.2234170681348935e-05, "loss": 0.7569, "step": 322750 }, { "epoch": 55.54026152787336, "grad_norm": 24.156356811523438, "learning_rate": 2.222986923606332e-05, "loss": 0.7851, "step": 322800 }, { "epoch": 55.548864418444595, "grad_norm": 9.519055366516113, "learning_rate": 2.22255677907777e-05, "loss": 0.7698, "step": 322850 }, { "epoch": 55.55746730901583, "grad_norm": 16.987163543701172, "learning_rate": 2.2221266345492085e-05, "loss": 0.7513, "step": 322900 }, { "epoch": 55.56607019958706, "grad_norm": 13.585576057434082, "learning_rate": 2.2216964900206473e-05, "loss": 0.7608, "step": 322950 }, { "epoch": 55.57467309015829, "grad_norm": 17.974855422973633, "learning_rate": 2.2212663454920854e-05, "loss": 0.7593, "step": 323000 }, { "epoch": 55.583275980729525, "grad_norm": 17.008474349975586, "learning_rate": 2.220836200963524e-05, "loss": 0.7742, "step": 323050 }, { "epoch": 55.59187887130076, "grad_norm": 12.695636749267578, "learning_rate": 2.220406056434962e-05, "loss": 0.7674, "step": 323100 }, { "epoch": 55.60048176187199, "grad_norm": 10.207210540771484, "learning_rate": 2.2199759119064008e-05, "loss": 0.7401, "step": 323150 }, { "epoch": 55.60908465244322, "grad_norm": 22.54661750793457, "learning_rate": 2.2195457673778392e-05, "loss": 0.7369, "step": 323200 }, { "epoch": 55.617687543014455, "grad_norm": 12.222530364990234, "learning_rate": 2.2191156228492774e-05, "loss": 0.7896, "step": 323250 }, { "epoch": 55.62629043358569, "grad_norm": 8.611534118652344, "learning_rate": 2.2186854783207158e-05, "loss": 0.7567, "step": 323300 }, { "epoch": 55.63489332415692, "grad_norm": 11.919578552246094, "learning_rate": 2.2182553337921543e-05, "loss": 0.771, "step": 323350 }, { "epoch": 55.643496214728145, "grad_norm": 12.335355758666992, "learning_rate": 2.2178251892635927e-05, "loss": 0.7639, "step": 323400 }, { "epoch": 55.65209910529938, "grad_norm": 18.110483169555664, "learning_rate": 2.2173950447350312e-05, "loss": 0.7385, "step": 323450 }, { "epoch": 55.66070199587061, "grad_norm": 17.813350677490234, "learning_rate": 2.2169649002064693e-05, "loss": 0.7573, "step": 323500 }, { "epoch": 55.66930488644184, "grad_norm": 12.159749984741211, "learning_rate": 2.2165347556779078e-05, "loss": 0.8225, "step": 323550 }, { "epoch": 55.677907777013075, "grad_norm": 14.693196296691895, "learning_rate": 2.2161046111493462e-05, "loss": 0.7552, "step": 323600 }, { "epoch": 55.68651066758431, "grad_norm": 8.998162269592285, "learning_rate": 2.2156744666207847e-05, "loss": 0.7832, "step": 323650 }, { "epoch": 55.69511355815554, "grad_norm": 7.49350643157959, "learning_rate": 2.215244322092223e-05, "loss": 0.7847, "step": 323700 }, { "epoch": 55.70371644872677, "grad_norm": 12.422843933105469, "learning_rate": 2.2148141775636616e-05, "loss": 0.7195, "step": 323750 }, { "epoch": 55.712319339298006, "grad_norm": 9.472850799560547, "learning_rate": 2.2143840330350997e-05, "loss": 0.7773, "step": 323800 }, { "epoch": 55.72092222986924, "grad_norm": 15.577564239501953, "learning_rate": 2.2139538885065385e-05, "loss": 0.7122, "step": 323850 }, { "epoch": 55.72952512044047, "grad_norm": 10.055685997009277, "learning_rate": 2.2135237439779766e-05, "loss": 0.8241, "step": 323900 }, { "epoch": 55.7381280110117, "grad_norm": 12.520206451416016, "learning_rate": 2.213093599449415e-05, "loss": 0.7621, "step": 323950 }, { "epoch": 55.74673090158293, "grad_norm": 8.654446601867676, "learning_rate": 2.2126634549208535e-05, "loss": 0.7768, "step": 324000 }, { "epoch": 55.75533379215416, "grad_norm": 19.36569595336914, "learning_rate": 2.212233310392292e-05, "loss": 0.7145, "step": 324050 }, { "epoch": 55.763936682725394, "grad_norm": 16.597139358520508, "learning_rate": 2.2118031658637304e-05, "loss": 0.7799, "step": 324100 }, { "epoch": 55.772539573296626, "grad_norm": 16.98992347717285, "learning_rate": 2.211373021335169e-05, "loss": 0.6799, "step": 324150 }, { "epoch": 55.78114246386786, "grad_norm": 17.377517700195312, "learning_rate": 2.210942876806607e-05, "loss": 0.7751, "step": 324200 }, { "epoch": 55.78974535443909, "grad_norm": 10.644232749938965, "learning_rate": 2.2105127322780455e-05, "loss": 0.7273, "step": 324250 }, { "epoch": 55.798348245010324, "grad_norm": 10.078774452209473, "learning_rate": 2.210082587749484e-05, "loss": 0.7603, "step": 324300 }, { "epoch": 55.806951135581556, "grad_norm": 12.099966049194336, "learning_rate": 2.2096524432209224e-05, "loss": 0.7168, "step": 324350 }, { "epoch": 55.81555402615279, "grad_norm": 13.665355682373047, "learning_rate": 2.2092222986923608e-05, "loss": 0.7815, "step": 324400 }, { "epoch": 55.82415691672402, "grad_norm": 17.785158157348633, "learning_rate": 2.208792154163799e-05, "loss": 0.7768, "step": 324450 }, { "epoch": 55.832759807295254, "grad_norm": 11.525891304016113, "learning_rate": 2.2083620096352374e-05, "loss": 0.6996, "step": 324500 }, { "epoch": 55.84136269786649, "grad_norm": 27.302120208740234, "learning_rate": 2.2079318651066762e-05, "loss": 0.7644, "step": 324550 }, { "epoch": 55.84996558843771, "grad_norm": 11.211467742919922, "learning_rate": 2.2075017205781143e-05, "loss": 0.7712, "step": 324600 }, { "epoch": 55.858568479008945, "grad_norm": 18.97047996520996, "learning_rate": 2.2070715760495528e-05, "loss": 0.8226, "step": 324650 }, { "epoch": 55.86717136958018, "grad_norm": 14.22691822052002, "learning_rate": 2.2066414315209912e-05, "loss": 0.7667, "step": 324700 }, { "epoch": 55.87577426015141, "grad_norm": 6.830941200256348, "learning_rate": 2.2062112869924297e-05, "loss": 0.7598, "step": 324750 }, { "epoch": 55.88437715072264, "grad_norm": 12.325485229492188, "learning_rate": 2.205781142463868e-05, "loss": 0.762, "step": 324800 }, { "epoch": 55.892980041293875, "grad_norm": 15.047521591186523, "learning_rate": 2.2053509979353063e-05, "loss": 0.8143, "step": 324850 }, { "epoch": 55.90158293186511, "grad_norm": 17.90635108947754, "learning_rate": 2.2049208534067447e-05, "loss": 0.7026, "step": 324900 }, { "epoch": 55.91018582243634, "grad_norm": 20.9141788482666, "learning_rate": 2.204490708878183e-05, "loss": 0.7678, "step": 324950 }, { "epoch": 55.91878871300757, "grad_norm": 15.370464324951172, "learning_rate": 2.2040605643496216e-05, "loss": 0.7423, "step": 325000 }, { "epoch": 55.927391603578805, "grad_norm": 14.629839897155762, "learning_rate": 2.20363041982106e-05, "loss": 0.7781, "step": 325050 }, { "epoch": 55.93599449415004, "grad_norm": 15.168642044067383, "learning_rate": 2.2032002752924985e-05, "loss": 0.7316, "step": 325100 }, { "epoch": 55.94459738472126, "grad_norm": 13.788406372070312, "learning_rate": 2.2027701307639367e-05, "loss": 0.8257, "step": 325150 }, { "epoch": 55.953200275292495, "grad_norm": 10.996456146240234, "learning_rate": 2.202339986235375e-05, "loss": 0.7239, "step": 325200 }, { "epoch": 55.96180316586373, "grad_norm": 19.075136184692383, "learning_rate": 2.2019098417068136e-05, "loss": 0.7257, "step": 325250 }, { "epoch": 55.97040605643496, "grad_norm": 8.990406036376953, "learning_rate": 2.201479697178252e-05, "loss": 0.7372, "step": 325300 }, { "epoch": 55.97900894700619, "grad_norm": 14.393460273742676, "learning_rate": 2.2010495526496905e-05, "loss": 0.7626, "step": 325350 }, { "epoch": 55.987611837577425, "grad_norm": 15.73536205291748, "learning_rate": 2.2006194081211286e-05, "loss": 0.7782, "step": 325400 }, { "epoch": 55.99621472814866, "grad_norm": 10.898555755615234, "learning_rate": 2.200189263592567e-05, "loss": 0.7326, "step": 325450 }, { "epoch": 56.0, "eval_accuracy": 0.6007571846498021, "eval_f1": 0.5894877731982711, "eval_f1_DuraRiadoRio_16x16": 0.5170469798657719, "eval_f1_Mole_16x16": 0.48446455505279035, "eval_f1_Quebrado_16x16": 0.7815762538382804, "eval_f1_RiadoRio_16x16": 0.5387618516452872, "eval_f1_RioFechado_16x16": 0.6255892255892256, "eval_loss": 1.694771409034729, "eval_precision": 0.6421300306580301, "eval_precision_DuraRiadoRio_16x16": 0.6776917663617171, "eval_precision_Mole_16x16": 0.7942631058358062, "eval_precision_Quebrado_16x16": 0.7396358000774893, "eval_precision_RiadoRio_16x16": 0.4916864608076009, "eval_precision_RioFechado_16x16": 0.5073730202075368, "eval_recall": 0.6012971496041464, "eval_recall_DuraRiadoRio_16x16": 0.41796875, "eval_recall_Mole_16x16": 0.3485243055555556, "eval_recall_Quebrado_16x16": 0.8285590277777778, "eval_recall_RiadoRio_16x16": 0.5958059210526315, "eval_recall_RioFechado_16x16": 0.8156277436347673, "eval_runtime": 45.5019, "eval_samples_per_second": 255.418, "eval_steps_per_second": 15.977, "step": 325472 }, { "epoch": 56.00481761871989, "grad_norm": 11.944744110107422, "learning_rate": 2.199759119064006e-05, "loss": 0.7412, "step": 325500 }, { "epoch": 56.01342050929112, "grad_norm": 20.13261604309082, "learning_rate": 2.199328974535444e-05, "loss": 0.7463, "step": 325550 }, { "epoch": 56.022023399862356, "grad_norm": 12.014873504638672, "learning_rate": 2.1988988300068824e-05, "loss": 0.7001, "step": 325600 }, { "epoch": 56.03062629043359, "grad_norm": 14.081549644470215, "learning_rate": 2.1984686854783205e-05, "loss": 0.7462, "step": 325650 }, { "epoch": 56.03922918100482, "grad_norm": 8.078246116638184, "learning_rate": 2.1980385409497593e-05, "loss": 0.7902, "step": 325700 }, { "epoch": 56.047832071576046, "grad_norm": 11.752195358276367, "learning_rate": 2.1976083964211978e-05, "loss": 0.7774, "step": 325750 }, { "epoch": 56.05643496214728, "grad_norm": 10.894808769226074, "learning_rate": 2.197178251892636e-05, "loss": 0.713, "step": 325800 }, { "epoch": 56.06503785271851, "grad_norm": 13.446453094482422, "learning_rate": 2.1967481073640744e-05, "loss": 0.7194, "step": 325850 }, { "epoch": 56.073640743289744, "grad_norm": 9.742216110229492, "learning_rate": 2.1963179628355128e-05, "loss": 0.8361, "step": 325900 }, { "epoch": 56.082243633860976, "grad_norm": 17.523517608642578, "learning_rate": 2.1958878183069513e-05, "loss": 0.7821, "step": 325950 }, { "epoch": 56.09084652443221, "grad_norm": 19.786888122558594, "learning_rate": 2.1954576737783897e-05, "loss": 0.7735, "step": 326000 }, { "epoch": 56.09944941500344, "grad_norm": 18.911548614501953, "learning_rate": 2.1950275292498282e-05, "loss": 0.6649, "step": 326050 }, { "epoch": 56.108052305574674, "grad_norm": 14.901814460754395, "learning_rate": 2.1945973847212663e-05, "loss": 0.7986, "step": 326100 }, { "epoch": 56.116655196145906, "grad_norm": 14.783246040344238, "learning_rate": 2.1941672401927048e-05, "loss": 0.7467, "step": 326150 }, { "epoch": 56.12525808671714, "grad_norm": 13.413820266723633, "learning_rate": 2.1937370956641432e-05, "loss": 0.7365, "step": 326200 }, { "epoch": 56.13386097728837, "grad_norm": 19.359046936035156, "learning_rate": 2.1933069511355817e-05, "loss": 0.7049, "step": 326250 }, { "epoch": 56.142463867859604, "grad_norm": 11.9705810546875, "learning_rate": 2.19287680660702e-05, "loss": 0.7535, "step": 326300 }, { "epoch": 56.15106675843083, "grad_norm": 6.159602165222168, "learning_rate": 2.1924466620784582e-05, "loss": 0.7771, "step": 326350 }, { "epoch": 56.15966964900206, "grad_norm": 11.037514686584473, "learning_rate": 2.192016517549897e-05, "loss": 0.7283, "step": 326400 }, { "epoch": 56.168272539573294, "grad_norm": 8.029251098632812, "learning_rate": 2.1915863730213355e-05, "loss": 0.7483, "step": 326450 }, { "epoch": 56.17687543014453, "grad_norm": 20.76032066345215, "learning_rate": 2.1911562284927736e-05, "loss": 0.7011, "step": 326500 }, { "epoch": 56.18547832071576, "grad_norm": 16.388439178466797, "learning_rate": 2.190726083964212e-05, "loss": 0.7306, "step": 326550 }, { "epoch": 56.19408121128699, "grad_norm": 12.437400817871094, "learning_rate": 2.1902959394356505e-05, "loss": 0.7882, "step": 326600 }, { "epoch": 56.202684101858225, "grad_norm": 14.915793418884277, "learning_rate": 2.189865794907089e-05, "loss": 0.7105, "step": 326650 }, { "epoch": 56.21128699242946, "grad_norm": 23.553970336914062, "learning_rate": 2.1894356503785274e-05, "loss": 0.7849, "step": 326700 }, { "epoch": 56.21988988300069, "grad_norm": 12.169373512268066, "learning_rate": 2.1890055058499656e-05, "loss": 0.7788, "step": 326750 }, { "epoch": 56.22849277357192, "grad_norm": 13.590761184692383, "learning_rate": 2.188575361321404e-05, "loss": 0.7271, "step": 326800 }, { "epoch": 56.237095664143155, "grad_norm": 12.281301498413086, "learning_rate": 2.1881452167928425e-05, "loss": 0.6973, "step": 326850 }, { "epoch": 56.24569855471439, "grad_norm": 17.776824951171875, "learning_rate": 2.187715072264281e-05, "loss": 0.7503, "step": 326900 }, { "epoch": 56.25430144528561, "grad_norm": 18.423145294189453, "learning_rate": 2.1872849277357194e-05, "loss": 0.7368, "step": 326950 }, { "epoch": 56.262904335856845, "grad_norm": 19.319839477539062, "learning_rate": 2.1868547832071575e-05, "loss": 0.7361, "step": 327000 }, { "epoch": 56.27150722642808, "grad_norm": 11.279186248779297, "learning_rate": 2.186424638678596e-05, "loss": 0.8385, "step": 327050 }, { "epoch": 56.28011011699931, "grad_norm": 13.865184783935547, "learning_rate": 2.1859944941500347e-05, "loss": 0.7265, "step": 327100 }, { "epoch": 56.28871300757054, "grad_norm": 23.87509536743164, "learning_rate": 2.185564349621473e-05, "loss": 0.7178, "step": 327150 }, { "epoch": 56.297315898141775, "grad_norm": 22.678146362304688, "learning_rate": 2.1851342050929113e-05, "loss": 0.7073, "step": 327200 }, { "epoch": 56.30591878871301, "grad_norm": 17.328664779663086, "learning_rate": 2.1847040605643498e-05, "loss": 0.747, "step": 327250 }, { "epoch": 56.31452167928424, "grad_norm": 12.744461059570312, "learning_rate": 2.1842739160357882e-05, "loss": 0.8311, "step": 327300 }, { "epoch": 56.32312456985547, "grad_norm": 11.622862815856934, "learning_rate": 2.1838437715072267e-05, "loss": 0.7072, "step": 327350 }, { "epoch": 56.331727460426706, "grad_norm": 17.443115234375, "learning_rate": 2.1834136269786648e-05, "loss": 0.7447, "step": 327400 }, { "epoch": 56.34033035099794, "grad_norm": 12.876482963562012, "learning_rate": 2.1829834824501033e-05, "loss": 0.6922, "step": 327450 }, { "epoch": 56.34893324156917, "grad_norm": 11.688224792480469, "learning_rate": 2.1825533379215417e-05, "loss": 0.7685, "step": 327500 }, { "epoch": 56.357536132140396, "grad_norm": 14.653177261352539, "learning_rate": 2.1821231933929802e-05, "loss": 0.8082, "step": 327550 }, { "epoch": 56.36613902271163, "grad_norm": 11.765002250671387, "learning_rate": 2.1816930488644186e-05, "loss": 0.745, "step": 327600 }, { "epoch": 56.37474191328286, "grad_norm": 12.402071952819824, "learning_rate": 2.181262904335857e-05, "loss": 0.7335, "step": 327650 }, { "epoch": 56.383344803854094, "grad_norm": 11.960524559020996, "learning_rate": 2.1808327598072952e-05, "loss": 0.7247, "step": 327700 }, { "epoch": 56.391947694425326, "grad_norm": 15.123931884765625, "learning_rate": 2.1804026152787337e-05, "loss": 0.7002, "step": 327750 }, { "epoch": 56.40055058499656, "grad_norm": 14.214808464050293, "learning_rate": 2.179972470750172e-05, "loss": 0.7069, "step": 327800 }, { "epoch": 56.40915347556779, "grad_norm": 10.868378639221191, "learning_rate": 2.1795423262216106e-05, "loss": 0.7501, "step": 327850 }, { "epoch": 56.417756366139024, "grad_norm": 15.082490921020508, "learning_rate": 2.179112181693049e-05, "loss": 0.8096, "step": 327900 }, { "epoch": 56.426359256710256, "grad_norm": 14.297953605651855, "learning_rate": 2.178682037164487e-05, "loss": 0.7522, "step": 327950 }, { "epoch": 56.43496214728149, "grad_norm": 15.895206451416016, "learning_rate": 2.178251892635926e-05, "loss": 0.7292, "step": 328000 }, { "epoch": 56.44356503785272, "grad_norm": 19.46371078491211, "learning_rate": 2.1778217481073644e-05, "loss": 0.8563, "step": 328050 }, { "epoch": 56.452167928423954, "grad_norm": 15.03677749633789, "learning_rate": 2.1773916035788025e-05, "loss": 0.7746, "step": 328100 }, { "epoch": 56.46077081899518, "grad_norm": 11.820414543151855, "learning_rate": 2.176961459050241e-05, "loss": 0.7793, "step": 328150 }, { "epoch": 56.46937370956641, "grad_norm": 14.511885643005371, "learning_rate": 2.1765313145216794e-05, "loss": 0.7243, "step": 328200 }, { "epoch": 56.477976600137644, "grad_norm": 11.10456657409668, "learning_rate": 2.176101169993118e-05, "loss": 0.7456, "step": 328250 }, { "epoch": 56.48657949070888, "grad_norm": 9.736360549926758, "learning_rate": 2.1756710254645563e-05, "loss": 0.7305, "step": 328300 }, { "epoch": 56.49518238128011, "grad_norm": 11.46731948852539, "learning_rate": 2.1752408809359945e-05, "loss": 0.7898, "step": 328350 }, { "epoch": 56.50378527185134, "grad_norm": 14.585742950439453, "learning_rate": 2.174810736407433e-05, "loss": 0.7514, "step": 328400 }, { "epoch": 56.512388162422575, "grad_norm": 13.374059677124023, "learning_rate": 2.1743805918788714e-05, "loss": 0.746, "step": 328450 }, { "epoch": 56.52099105299381, "grad_norm": 9.492286682128906, "learning_rate": 2.1739504473503098e-05, "loss": 0.7949, "step": 328500 }, { "epoch": 56.52959394356504, "grad_norm": 10.383533477783203, "learning_rate": 2.1735203028217483e-05, "loss": 0.7584, "step": 328550 }, { "epoch": 56.53819683413627, "grad_norm": 18.120981216430664, "learning_rate": 2.1730901582931867e-05, "loss": 0.7512, "step": 328600 }, { "epoch": 56.546799724707505, "grad_norm": 19.555524826049805, "learning_rate": 2.172660013764625e-05, "loss": 0.6753, "step": 328650 }, { "epoch": 56.55540261527874, "grad_norm": 18.769405364990234, "learning_rate": 2.1722298692360636e-05, "loss": 0.7892, "step": 328700 }, { "epoch": 56.56400550584996, "grad_norm": 13.410124778747559, "learning_rate": 2.1717997247075018e-05, "loss": 0.74, "step": 328750 }, { "epoch": 56.572608396421195, "grad_norm": 17.020885467529297, "learning_rate": 2.1713695801789402e-05, "loss": 0.7203, "step": 328800 }, { "epoch": 56.58121128699243, "grad_norm": 16.805252075195312, "learning_rate": 2.1709394356503787e-05, "loss": 0.731, "step": 328850 }, { "epoch": 56.58981417756366, "grad_norm": 11.802996635437012, "learning_rate": 2.1705092911218168e-05, "loss": 0.73, "step": 328900 }, { "epoch": 56.59841706813489, "grad_norm": 8.064237594604492, "learning_rate": 2.1700791465932556e-05, "loss": 0.7331, "step": 328950 }, { "epoch": 56.607019958706125, "grad_norm": 13.761898040771484, "learning_rate": 2.169649002064694e-05, "loss": 0.7729, "step": 329000 }, { "epoch": 56.61562284927736, "grad_norm": 19.000442504882812, "learning_rate": 2.169218857536132e-05, "loss": 0.7608, "step": 329050 }, { "epoch": 56.62422573984859, "grad_norm": 12.299666404724121, "learning_rate": 2.1687887130075706e-05, "loss": 0.7117, "step": 329100 }, { "epoch": 56.63282863041982, "grad_norm": 12.398569107055664, "learning_rate": 2.168358568479009e-05, "loss": 0.833, "step": 329150 }, { "epoch": 56.641431520991055, "grad_norm": 16.028526306152344, "learning_rate": 2.1679284239504475e-05, "loss": 0.7678, "step": 329200 }, { "epoch": 56.65003441156229, "grad_norm": 16.986597061157227, "learning_rate": 2.167498279421886e-05, "loss": 0.7676, "step": 329250 }, { "epoch": 56.65863730213351, "grad_norm": 13.795456886291504, "learning_rate": 2.167068134893324e-05, "loss": 0.7497, "step": 329300 }, { "epoch": 56.667240192704746, "grad_norm": 18.949081420898438, "learning_rate": 2.1666379903647626e-05, "loss": 0.7808, "step": 329350 }, { "epoch": 56.67584308327598, "grad_norm": 12.829408645629883, "learning_rate": 2.166207845836201e-05, "loss": 0.7606, "step": 329400 }, { "epoch": 56.68444597384721, "grad_norm": 19.135141372680664, "learning_rate": 2.1657777013076395e-05, "loss": 0.7686, "step": 329450 }, { "epoch": 56.693048864418444, "grad_norm": 15.07206916809082, "learning_rate": 2.165347556779078e-05, "loss": 0.7219, "step": 329500 }, { "epoch": 56.701651754989676, "grad_norm": 14.249181747436523, "learning_rate": 2.164917412250516e-05, "loss": 0.6991, "step": 329550 }, { "epoch": 56.71025464556091, "grad_norm": 8.249271392822266, "learning_rate": 2.1644872677219545e-05, "loss": 0.792, "step": 329600 }, { "epoch": 56.71885753613214, "grad_norm": 14.414939880371094, "learning_rate": 2.1640571231933933e-05, "loss": 0.7369, "step": 329650 }, { "epoch": 56.727460426703374, "grad_norm": 13.494241714477539, "learning_rate": 2.1636269786648314e-05, "loss": 0.7897, "step": 329700 }, { "epoch": 56.736063317274606, "grad_norm": 11.761247634887695, "learning_rate": 2.16319683413627e-05, "loss": 0.7759, "step": 329750 }, { "epoch": 56.74466620784584, "grad_norm": 10.261711120605469, "learning_rate": 2.1627666896077083e-05, "loss": 0.7206, "step": 329800 }, { "epoch": 56.75326909841707, "grad_norm": 11.479698181152344, "learning_rate": 2.1623365450791468e-05, "loss": 0.7593, "step": 329850 }, { "epoch": 56.7618719889883, "grad_norm": 18.143861770629883, "learning_rate": 2.1619064005505852e-05, "loss": 0.7283, "step": 329900 }, { "epoch": 56.77047487955953, "grad_norm": 14.91513729095459, "learning_rate": 2.1614762560220237e-05, "loss": 0.7878, "step": 329950 }, { "epoch": 56.77907777013076, "grad_norm": 15.723901748657227, "learning_rate": 2.1610461114934618e-05, "loss": 0.7543, "step": 330000 }, { "epoch": 56.787680660701994, "grad_norm": 17.6783504486084, "learning_rate": 2.1606159669649003e-05, "loss": 0.7386, "step": 330050 }, { "epoch": 56.79628355127323, "grad_norm": 17.769184112548828, "learning_rate": 2.1601858224363387e-05, "loss": 0.7493, "step": 330100 }, { "epoch": 56.80488644184446, "grad_norm": 14.189520835876465, "learning_rate": 2.1597556779077772e-05, "loss": 0.6998, "step": 330150 }, { "epoch": 56.81348933241569, "grad_norm": 8.861952781677246, "learning_rate": 2.1593255333792156e-05, "loss": 0.7384, "step": 330200 }, { "epoch": 56.822092222986925, "grad_norm": 12.532812118530273, "learning_rate": 2.1588953888506537e-05, "loss": 0.7455, "step": 330250 }, { "epoch": 56.83069511355816, "grad_norm": 14.203064918518066, "learning_rate": 2.1584652443220922e-05, "loss": 0.7644, "step": 330300 }, { "epoch": 56.83929800412939, "grad_norm": 14.229214668273926, "learning_rate": 2.158035099793531e-05, "loss": 0.7583, "step": 330350 }, { "epoch": 56.84790089470062, "grad_norm": 21.619403839111328, "learning_rate": 2.157604955264969e-05, "loss": 0.7496, "step": 330400 }, { "epoch": 56.856503785271855, "grad_norm": 15.600422859191895, "learning_rate": 2.1571748107364076e-05, "loss": 0.7209, "step": 330450 }, { "epoch": 56.86510667584308, "grad_norm": 14.993415832519531, "learning_rate": 2.1567446662078457e-05, "loss": 0.7855, "step": 330500 }, { "epoch": 56.87370956641431, "grad_norm": 9.327057838439941, "learning_rate": 2.1563145216792845e-05, "loss": 0.7255, "step": 330550 }, { "epoch": 56.882312456985545, "grad_norm": 10.326887130737305, "learning_rate": 2.155884377150723e-05, "loss": 0.8028, "step": 330600 }, { "epoch": 56.89091534755678, "grad_norm": 22.57769012451172, "learning_rate": 2.155454232622161e-05, "loss": 0.7674, "step": 330650 }, { "epoch": 56.89951823812801, "grad_norm": 20.71199607849121, "learning_rate": 2.1550240880935995e-05, "loss": 0.7063, "step": 330700 }, { "epoch": 56.90812112869924, "grad_norm": 11.002126693725586, "learning_rate": 2.154593943565038e-05, "loss": 0.7665, "step": 330750 }, { "epoch": 56.916724019270475, "grad_norm": 16.41135025024414, "learning_rate": 2.1541637990364764e-05, "loss": 0.7672, "step": 330800 }, { "epoch": 56.92532690984171, "grad_norm": 17.381900787353516, "learning_rate": 2.153733654507915e-05, "loss": 0.7246, "step": 330850 }, { "epoch": 56.93392980041294, "grad_norm": 10.216157913208008, "learning_rate": 2.153303509979353e-05, "loss": 0.7545, "step": 330900 }, { "epoch": 56.94253269098417, "grad_norm": 9.839737892150879, "learning_rate": 2.1528733654507915e-05, "loss": 0.7648, "step": 330950 }, { "epoch": 56.951135581555405, "grad_norm": 16.061521530151367, "learning_rate": 2.15244322092223e-05, "loss": 0.7298, "step": 331000 }, { "epoch": 56.95973847212664, "grad_norm": 13.050433158874512, "learning_rate": 2.1520130763936684e-05, "loss": 0.7483, "step": 331050 }, { "epoch": 56.96834136269786, "grad_norm": 16.785375595092773, "learning_rate": 2.1515829318651068e-05, "loss": 0.7456, "step": 331100 }, { "epoch": 56.976944253269096, "grad_norm": 14.47414779663086, "learning_rate": 2.1511527873365453e-05, "loss": 0.7368, "step": 331150 }, { "epoch": 56.98554714384033, "grad_norm": 13.809205055236816, "learning_rate": 2.1507226428079834e-05, "loss": 0.7867, "step": 331200 }, { "epoch": 56.99415003441156, "grad_norm": 15.478337287902832, "learning_rate": 2.1502924982794222e-05, "loss": 0.7648, "step": 331250 }, { "epoch": 57.0, "eval_accuracy": 0.5460333849595594, "eval_f1": 0.5448631014849712, "eval_f1_DuraRiadoRio_16x16": 0.5162287480680062, "eval_f1_Mole_16x16": 0.4665482534043813, "eval_f1_Quebrado_16x16": 0.6987951807228916, "eval_f1_RiadoRio_16x16": 0.4633149760170545, "eval_f1_RioFechado_16x16": 0.5794283492125218, "eval_loss": 1.9123625755310059, "eval_precision": 0.5784982599544675, "eval_precision_DuraRiadoRio_16x16": 0.5253932584269663, "eval_precision_Mole_16x16": 0.7337057728119181, "eval_precision_Quebrado_16x16": 0.7054400707651481, "eval_precision_RiadoRio_16x16": 0.4078823897403816, "eval_precision_RioFechado_16x16": 0.5200698080279232, "eval_recall": 0.5463866811453568, "eval_recall_DuraRiadoRio_16x16": 0.5073784722222222, "eval_recall_Mole_16x16": 0.3420138888888889, "eval_recall_Quebrado_16x16": 0.6922743055555556, "eval_recall_RiadoRio_16x16": 0.5361842105263158, "eval_recall_RioFechado_16x16": 0.6540825285338016, "eval_runtime": 45.4149, "eval_samples_per_second": 255.907, "eval_steps_per_second": 16.008, "step": 331284 }, { "epoch": 57.002752924982794, "grad_norm": 13.309732437133789, "learning_rate": 2.1498623537508603e-05, "loss": 0.733, "step": 331300 }, { "epoch": 57.011355815554026, "grad_norm": 9.975556373596191, "learning_rate": 2.1494322092222988e-05, "loss": 0.7375, "step": 331350 }, { "epoch": 57.01995870612526, "grad_norm": 8.262561798095703, "learning_rate": 2.1490020646937372e-05, "loss": 0.7398, "step": 331400 }, { "epoch": 57.02856159669649, "grad_norm": 15.46056842803955, "learning_rate": 2.1485719201651753e-05, "loss": 0.7032, "step": 331450 }, { "epoch": 57.037164487267724, "grad_norm": 9.900655746459961, "learning_rate": 2.148141775636614e-05, "loss": 0.7349, "step": 331500 }, { "epoch": 57.045767377838956, "grad_norm": 24.27967071533203, "learning_rate": 2.1477116311080526e-05, "loss": 0.7283, "step": 331550 }, { "epoch": 57.05437026841019, "grad_norm": 10.664969444274902, "learning_rate": 2.1472814865794907e-05, "loss": 0.7139, "step": 331600 }, { "epoch": 57.06297315898142, "grad_norm": 8.885272026062012, "learning_rate": 2.146851342050929e-05, "loss": 0.7355, "step": 331650 }, { "epoch": 57.07157604955265, "grad_norm": 15.218502044677734, "learning_rate": 2.1464211975223676e-05, "loss": 0.7521, "step": 331700 }, { "epoch": 57.08017894012388, "grad_norm": 7.771664619445801, "learning_rate": 2.145991052993806e-05, "loss": 0.7649, "step": 331750 }, { "epoch": 57.08878183069511, "grad_norm": 13.366262435913086, "learning_rate": 2.1455609084652445e-05, "loss": 0.7491, "step": 331800 }, { "epoch": 57.097384721266344, "grad_norm": 17.28685760498047, "learning_rate": 2.1451307639366826e-05, "loss": 0.7722, "step": 331850 }, { "epoch": 57.10598761183758, "grad_norm": 9.133752822875977, "learning_rate": 2.144700619408121e-05, "loss": 0.7166, "step": 331900 }, { "epoch": 57.11459050240881, "grad_norm": 18.749540328979492, "learning_rate": 2.14427047487956e-05, "loss": 0.6995, "step": 331950 }, { "epoch": 57.12319339298004, "grad_norm": 18.548114776611328, "learning_rate": 2.143840330350998e-05, "loss": 0.7854, "step": 332000 }, { "epoch": 57.131796283551274, "grad_norm": 17.490999221801758, "learning_rate": 2.1434101858224365e-05, "loss": 0.7731, "step": 332050 }, { "epoch": 57.14039917412251, "grad_norm": 13.709226608276367, "learning_rate": 2.142980041293875e-05, "loss": 0.7118, "step": 332100 }, { "epoch": 57.14900206469374, "grad_norm": 20.964637756347656, "learning_rate": 2.142549896765313e-05, "loss": 0.7844, "step": 332150 }, { "epoch": 57.15760495526497, "grad_norm": 12.658946990966797, "learning_rate": 2.142119752236752e-05, "loss": 0.6736, "step": 332200 }, { "epoch": 57.1662078458362, "grad_norm": 8.345199584960938, "learning_rate": 2.14168960770819e-05, "loss": 0.7355, "step": 332250 }, { "epoch": 57.17481073640743, "grad_norm": 24.756311416625977, "learning_rate": 2.1412594631796284e-05, "loss": 0.7764, "step": 332300 }, { "epoch": 57.18341362697866, "grad_norm": 12.045780181884766, "learning_rate": 2.140829318651067e-05, "loss": 0.7037, "step": 332350 }, { "epoch": 57.192016517549895, "grad_norm": 12.782896995544434, "learning_rate": 2.1403991741225053e-05, "loss": 0.7136, "step": 332400 }, { "epoch": 57.20061940812113, "grad_norm": 12.124938011169434, "learning_rate": 2.1399690295939438e-05, "loss": 0.7349, "step": 332450 }, { "epoch": 57.20922229869236, "grad_norm": 18.18598747253418, "learning_rate": 2.1395388850653822e-05, "loss": 0.8002, "step": 332500 }, { "epoch": 57.21782518926359, "grad_norm": 13.965160369873047, "learning_rate": 2.1391087405368204e-05, "loss": 0.7347, "step": 332550 }, { "epoch": 57.226428079834825, "grad_norm": 13.036042213439941, "learning_rate": 2.1386785960082588e-05, "loss": 0.7116, "step": 332600 }, { "epoch": 57.23503097040606, "grad_norm": 18.160781860351562, "learning_rate": 2.1382484514796973e-05, "loss": 0.7618, "step": 332650 }, { "epoch": 57.24363386097729, "grad_norm": 18.55841827392578, "learning_rate": 2.1378183069511357e-05, "loss": 0.8093, "step": 332700 }, { "epoch": 57.25223675154852, "grad_norm": 12.8931884765625, "learning_rate": 2.1373881624225742e-05, "loss": 0.6811, "step": 332750 }, { "epoch": 57.260839642119755, "grad_norm": 13.004094123840332, "learning_rate": 2.1369580178940123e-05, "loss": 0.7326, "step": 332800 }, { "epoch": 57.26944253269098, "grad_norm": 9.586027145385742, "learning_rate": 2.1365278733654508e-05, "loss": 0.7036, "step": 332850 }, { "epoch": 57.27804542326221, "grad_norm": 25.376283645629883, "learning_rate": 2.1360977288368895e-05, "loss": 0.7496, "step": 332900 }, { "epoch": 57.286648313833446, "grad_norm": 12.252067565917969, "learning_rate": 2.1356675843083277e-05, "loss": 0.7399, "step": 332950 }, { "epoch": 57.29525120440468, "grad_norm": 14.43028736114502, "learning_rate": 2.135237439779766e-05, "loss": 0.7314, "step": 333000 }, { "epoch": 57.30385409497591, "grad_norm": 12.875117301940918, "learning_rate": 2.1348072952512042e-05, "loss": 0.7812, "step": 333050 }, { "epoch": 57.31245698554714, "grad_norm": 15.535438537597656, "learning_rate": 2.134377150722643e-05, "loss": 0.8025, "step": 333100 }, { "epoch": 57.321059876118376, "grad_norm": 11.02721881866455, "learning_rate": 2.1339470061940815e-05, "loss": 0.7542, "step": 333150 }, { "epoch": 57.32966276668961, "grad_norm": 15.171525001525879, "learning_rate": 2.1335168616655196e-05, "loss": 0.7606, "step": 333200 }, { "epoch": 57.33826565726084, "grad_norm": 6.441656112670898, "learning_rate": 2.133086717136958e-05, "loss": 0.751, "step": 333250 }, { "epoch": 57.346868547832074, "grad_norm": 14.328540802001953, "learning_rate": 2.1326565726083965e-05, "loss": 0.7758, "step": 333300 }, { "epoch": 57.355471438403306, "grad_norm": 13.85679817199707, "learning_rate": 2.132226428079835e-05, "loss": 0.7126, "step": 333350 }, { "epoch": 57.36407432897454, "grad_norm": 14.439485549926758, "learning_rate": 2.1317962835512734e-05, "loss": 0.7598, "step": 333400 }, { "epoch": 57.372677219545764, "grad_norm": 16.979724884033203, "learning_rate": 2.131366139022712e-05, "loss": 0.7836, "step": 333450 }, { "epoch": 57.381280110117, "grad_norm": 11.619569778442383, "learning_rate": 2.13093599449415e-05, "loss": 0.7524, "step": 333500 }, { "epoch": 57.38988300068823, "grad_norm": 13.816096305847168, "learning_rate": 2.1305058499655885e-05, "loss": 0.7386, "step": 333550 }, { "epoch": 57.39848589125946, "grad_norm": 22.751121520996094, "learning_rate": 2.130075705437027e-05, "loss": 0.6756, "step": 333600 }, { "epoch": 57.407088781830694, "grad_norm": 9.692771911621094, "learning_rate": 2.1296455609084654e-05, "loss": 0.7522, "step": 333650 }, { "epoch": 57.41569167240193, "grad_norm": 9.00556468963623, "learning_rate": 2.1292154163799038e-05, "loss": 0.8151, "step": 333700 }, { "epoch": 57.42429456297316, "grad_norm": 21.61408805847168, "learning_rate": 2.128785271851342e-05, "loss": 0.7653, "step": 333750 }, { "epoch": 57.43289745354439, "grad_norm": 17.256668090820312, "learning_rate": 2.1283551273227807e-05, "loss": 0.7625, "step": 333800 }, { "epoch": 57.441500344115624, "grad_norm": 13.026637077331543, "learning_rate": 2.1279249827942192e-05, "loss": 0.7092, "step": 333850 }, { "epoch": 57.45010323468686, "grad_norm": 15.274076461791992, "learning_rate": 2.1274948382656573e-05, "loss": 0.7233, "step": 333900 }, { "epoch": 57.45870612525809, "grad_norm": 16.90962028503418, "learning_rate": 2.1270646937370958e-05, "loss": 0.773, "step": 333950 }, { "epoch": 57.46730901582932, "grad_norm": 13.220916748046875, "learning_rate": 2.1266345492085342e-05, "loss": 0.7512, "step": 334000 }, { "epoch": 57.47591190640055, "grad_norm": 13.084602355957031, "learning_rate": 2.1262044046799727e-05, "loss": 0.7286, "step": 334050 }, { "epoch": 57.48451479697178, "grad_norm": 14.20023250579834, "learning_rate": 2.125774260151411e-05, "loss": 0.7397, "step": 334100 }, { "epoch": 57.49311768754301, "grad_norm": 7.8757853507995605, "learning_rate": 2.1253441156228493e-05, "loss": 0.7113, "step": 334150 }, { "epoch": 57.501720578114245, "grad_norm": 9.667074203491211, "learning_rate": 2.1249139710942877e-05, "loss": 0.7638, "step": 334200 }, { "epoch": 57.51032346868548, "grad_norm": 13.479400634765625, "learning_rate": 2.124483826565726e-05, "loss": 0.7999, "step": 334250 }, { "epoch": 57.51892635925671, "grad_norm": 19.40728759765625, "learning_rate": 2.1240536820371646e-05, "loss": 0.8049, "step": 334300 }, { "epoch": 57.52752924982794, "grad_norm": 8.030006408691406, "learning_rate": 2.123623537508603e-05, "loss": 0.7738, "step": 334350 }, { "epoch": 57.536132140399175, "grad_norm": 12.490761756896973, "learning_rate": 2.1231933929800412e-05, "loss": 0.767, "step": 334400 }, { "epoch": 57.54473503097041, "grad_norm": 10.903556823730469, "learning_rate": 2.1227632484514796e-05, "loss": 0.7383, "step": 334450 }, { "epoch": 57.55333792154164, "grad_norm": 9.886442184448242, "learning_rate": 2.1223331039229184e-05, "loss": 0.7547, "step": 334500 }, { "epoch": 57.56194081211287, "grad_norm": 17.79019546508789, "learning_rate": 2.1219029593943566e-05, "loss": 0.7563, "step": 334550 }, { "epoch": 57.570543702684105, "grad_norm": 17.273841857910156, "learning_rate": 2.121472814865795e-05, "loss": 0.7377, "step": 334600 }, { "epoch": 57.57914659325533, "grad_norm": 17.53286361694336, "learning_rate": 2.1210426703372335e-05, "loss": 0.7684, "step": 334650 }, { "epoch": 57.58774948382656, "grad_norm": 6.697093963623047, "learning_rate": 2.1206125258086716e-05, "loss": 0.7612, "step": 334700 }, { "epoch": 57.596352374397796, "grad_norm": 8.514906883239746, "learning_rate": 2.1201823812801104e-05, "loss": 0.771, "step": 334750 }, { "epoch": 57.60495526496903, "grad_norm": 18.22245216369629, "learning_rate": 2.1197522367515485e-05, "loss": 0.7624, "step": 334800 }, { "epoch": 57.61355815554026, "grad_norm": 15.768082618713379, "learning_rate": 2.119322092222987e-05, "loss": 0.7702, "step": 334850 }, { "epoch": 57.62216104611149, "grad_norm": 13.367298126220703, "learning_rate": 2.1188919476944254e-05, "loss": 0.811, "step": 334900 }, { "epoch": 57.630763936682726, "grad_norm": 6.811519622802734, "learning_rate": 2.118461803165864e-05, "loss": 0.7831, "step": 334950 }, { "epoch": 57.63936682725396, "grad_norm": 10.033526420593262, "learning_rate": 2.1180316586373023e-05, "loss": 0.7553, "step": 335000 }, { "epoch": 57.64796971782519, "grad_norm": 9.741471290588379, "learning_rate": 2.1176015141087408e-05, "loss": 0.7374, "step": 335050 }, { "epoch": 57.656572608396424, "grad_norm": 12.35171127319336, "learning_rate": 2.117171369580179e-05, "loss": 0.7361, "step": 335100 }, { "epoch": 57.665175498967656, "grad_norm": 13.112631797790527, "learning_rate": 2.1167412250516174e-05, "loss": 0.6817, "step": 335150 }, { "epoch": 57.67377838953888, "grad_norm": 14.646646499633789, "learning_rate": 2.1163110805230558e-05, "loss": 0.7503, "step": 335200 }, { "epoch": 57.682381280110114, "grad_norm": 14.633530616760254, "learning_rate": 2.1158809359944943e-05, "loss": 0.7474, "step": 335250 }, { "epoch": 57.69098417068135, "grad_norm": 10.736250877380371, "learning_rate": 2.1154507914659327e-05, "loss": 0.7422, "step": 335300 }, { "epoch": 57.69958706125258, "grad_norm": 14.726401329040527, "learning_rate": 2.115020646937371e-05, "loss": 0.7159, "step": 335350 }, { "epoch": 57.70818995182381, "grad_norm": 18.851627349853516, "learning_rate": 2.1145905024088093e-05, "loss": 0.7856, "step": 335400 }, { "epoch": 57.716792842395044, "grad_norm": 15.512848854064941, "learning_rate": 2.114160357880248e-05, "loss": 0.8306, "step": 335450 }, { "epoch": 57.72539573296628, "grad_norm": 15.200379371643066, "learning_rate": 2.1137302133516862e-05, "loss": 0.7069, "step": 335500 }, { "epoch": 57.73399862353751, "grad_norm": 7.012669086456299, "learning_rate": 2.1133000688231247e-05, "loss": 0.7516, "step": 335550 }, { "epoch": 57.74260151410874, "grad_norm": 9.086288452148438, "learning_rate": 2.112869924294563e-05, "loss": 0.7269, "step": 335600 }, { "epoch": 57.751204404679974, "grad_norm": 9.194598197937012, "learning_rate": 2.1124397797660016e-05, "loss": 0.7553, "step": 335650 }, { "epoch": 57.75980729525121, "grad_norm": 18.176658630371094, "learning_rate": 2.11200963523744e-05, "loss": 0.7159, "step": 335700 }, { "epoch": 57.76841018582244, "grad_norm": 16.899452209472656, "learning_rate": 2.111579490708878e-05, "loss": 0.7784, "step": 335750 }, { "epoch": 57.777013076393665, "grad_norm": 13.540183067321777, "learning_rate": 2.1111493461803166e-05, "loss": 0.7085, "step": 335800 }, { "epoch": 57.7856159669649, "grad_norm": 14.42415714263916, "learning_rate": 2.110719201651755e-05, "loss": 0.7342, "step": 335850 }, { "epoch": 57.79421885753613, "grad_norm": 12.514610290527344, "learning_rate": 2.1102890571231935e-05, "loss": 0.6734, "step": 335900 }, { "epoch": 57.80282174810736, "grad_norm": 11.953702926635742, "learning_rate": 2.109858912594632e-05, "loss": 0.7548, "step": 335950 }, { "epoch": 57.811424638678595, "grad_norm": 12.363713264465332, "learning_rate": 2.1094287680660704e-05, "loss": 0.7211, "step": 336000 }, { "epoch": 57.82002752924983, "grad_norm": 19.47869873046875, "learning_rate": 2.1089986235375085e-05, "loss": 0.722, "step": 336050 }, { "epoch": 57.82863041982106, "grad_norm": 12.189957618713379, "learning_rate": 2.108568479008947e-05, "loss": 0.7907, "step": 336100 }, { "epoch": 57.83723331039229, "grad_norm": 19.968345642089844, "learning_rate": 2.1081383344803855e-05, "loss": 0.7706, "step": 336150 }, { "epoch": 57.845836200963525, "grad_norm": 15.496025085449219, "learning_rate": 2.107708189951824e-05, "loss": 0.7701, "step": 336200 }, { "epoch": 57.85443909153476, "grad_norm": 12.686484336853027, "learning_rate": 2.1072780454232624e-05, "loss": 0.7338, "step": 336250 }, { "epoch": 57.86304198210599, "grad_norm": 20.40520668029785, "learning_rate": 2.1068479008947005e-05, "loss": 0.758, "step": 336300 }, { "epoch": 57.87164487267722, "grad_norm": 10.410775184631348, "learning_rate": 2.1064177563661393e-05, "loss": 0.7572, "step": 336350 }, { "epoch": 57.88024776324845, "grad_norm": 18.275524139404297, "learning_rate": 2.1059876118375777e-05, "loss": 0.7503, "step": 336400 }, { "epoch": 57.88885065381968, "grad_norm": 15.672919273376465, "learning_rate": 2.105557467309016e-05, "loss": 0.7859, "step": 336450 }, { "epoch": 57.89745354439091, "grad_norm": 14.010963439941406, "learning_rate": 2.1051273227804543e-05, "loss": 0.7553, "step": 336500 }, { "epoch": 57.906056434962146, "grad_norm": 14.867149353027344, "learning_rate": 2.1046971782518928e-05, "loss": 0.7034, "step": 336550 }, { "epoch": 57.91465932553338, "grad_norm": 11.974017143249512, "learning_rate": 2.1042670337233312e-05, "loss": 0.6924, "step": 336600 }, { "epoch": 57.92326221610461, "grad_norm": 25.65148162841797, "learning_rate": 2.1038368891947697e-05, "loss": 0.7397, "step": 336650 }, { "epoch": 57.93186510667584, "grad_norm": 19.52814292907715, "learning_rate": 2.1034067446662078e-05, "loss": 0.7779, "step": 336700 }, { "epoch": 57.940467997247076, "grad_norm": 10.216012954711914, "learning_rate": 2.1029766001376463e-05, "loss": 0.7495, "step": 336750 }, { "epoch": 57.94907088781831, "grad_norm": 8.479580879211426, "learning_rate": 2.1025464556090847e-05, "loss": 0.7689, "step": 336800 }, { "epoch": 57.95767377838954, "grad_norm": 16.040271759033203, "learning_rate": 2.102116311080523e-05, "loss": 0.7534, "step": 336850 }, { "epoch": 57.966276668960774, "grad_norm": 16.329565048217773, "learning_rate": 2.1016861665519616e-05, "loss": 0.7894, "step": 336900 }, { "epoch": 57.974879559532006, "grad_norm": 16.986818313598633, "learning_rate": 2.1012560220233997e-05, "loss": 0.7699, "step": 336950 }, { "epoch": 57.98348245010323, "grad_norm": 8.430131912231445, "learning_rate": 2.1008258774948382e-05, "loss": 0.7265, "step": 337000 }, { "epoch": 57.992085340674464, "grad_norm": 11.905317306518555, "learning_rate": 2.100395732966277e-05, "loss": 0.7768, "step": 337050 }, { "epoch": 58.0, "eval_accuracy": 0.5514541387024608, "eval_f1": 0.5313816344477386, "eval_f1_DuraRiadoRio_16x16": 0.55103668261563, "eval_f1_Mole_16x16": 0.3821138211382114, "eval_f1_Quebrado_16x16": 0.7416805324459235, "eval_f1_RiadoRio_16x16": 0.42045145499048137, "eval_f1_RioFechado_16x16": 0.5616256810484465, "eval_loss": 2.0000569820404053, "eval_precision": 0.6270917370842093, "eval_precision_DuraRiadoRio_16x16": 0.5095870206489675, "eval_precision_Mole_16x16": 0.8703703703703703, "eval_precision_Quebrado_16x16": 0.7120607028753994, "eval_precision_RiadoRio_16x16": 0.6208835341365462, "eval_precision_RioFechado_16x16": 0.4225570573897629, "eval_recall": 0.5546945636561773, "eval_recall_DuraRiadoRio_16x16": 0.5998263888888888, "eval_recall_Mole_16x16": 0.24479166666666666, "eval_recall_Quebrado_16x16": 0.7738715277777778, "eval_recall_RiadoRio_16x16": 0.3178453947368421, "eval_recall_RioFechado_16x16": 0.8371378402107111, "eval_runtime": 46.2219, "eval_samples_per_second": 251.439, "eval_steps_per_second": 15.728, "step": 337096 }, { "epoch": 58.0006882312457, "grad_norm": 18.45948600769043, "learning_rate": 2.099965588437715e-05, "loss": 0.7638, "step": 337100 }, { "epoch": 58.00929112181693, "grad_norm": 11.112471580505371, "learning_rate": 2.0995354439091536e-05, "loss": 0.7607, "step": 337150 }, { "epoch": 58.01789401238816, "grad_norm": 12.861695289611816, "learning_rate": 2.099105299380592e-05, "loss": 0.7505, "step": 337200 }, { "epoch": 58.026496902959394, "grad_norm": 17.13945960998535, "learning_rate": 2.0986751548520305e-05, "loss": 0.7858, "step": 337250 }, { "epoch": 58.03509979353063, "grad_norm": 13.964406967163086, "learning_rate": 2.098245010323469e-05, "loss": 0.7122, "step": 337300 }, { "epoch": 58.04370268410186, "grad_norm": 19.446746826171875, "learning_rate": 2.0978148657949074e-05, "loss": 0.8309, "step": 337350 }, { "epoch": 58.05230557467309, "grad_norm": 12.579668998718262, "learning_rate": 2.0973847212663455e-05, "loss": 0.7887, "step": 337400 }, { "epoch": 58.060908465244324, "grad_norm": 10.4732084274292, "learning_rate": 2.096954576737784e-05, "loss": 0.7036, "step": 337450 }, { "epoch": 58.06951135581556, "grad_norm": 11.111150741577148, "learning_rate": 2.0965244322092224e-05, "loss": 0.7569, "step": 337500 }, { "epoch": 58.07811424638679, "grad_norm": 20.16413116455078, "learning_rate": 2.096094287680661e-05, "loss": 0.7158, "step": 337550 }, { "epoch": 58.086717136958015, "grad_norm": 8.109871864318848, "learning_rate": 2.0956641431520993e-05, "loss": 0.6988, "step": 337600 }, { "epoch": 58.09532002752925, "grad_norm": 12.359088897705078, "learning_rate": 2.0952339986235374e-05, "loss": 0.6839, "step": 337650 }, { "epoch": 58.10392291810048, "grad_norm": 11.323436737060547, "learning_rate": 2.094803854094976e-05, "loss": 0.7163, "step": 337700 }, { "epoch": 58.11252580867171, "grad_norm": 10.090826034545898, "learning_rate": 2.0943737095664147e-05, "loss": 0.7669, "step": 337750 }, { "epoch": 58.121128699242945, "grad_norm": 12.3829345703125, "learning_rate": 2.0939435650378528e-05, "loss": 0.7141, "step": 337800 }, { "epoch": 58.12973158981418, "grad_norm": 15.287969589233398, "learning_rate": 2.0935134205092913e-05, "loss": 0.725, "step": 337850 }, { "epoch": 58.13833448038541, "grad_norm": 10.993062973022461, "learning_rate": 2.0930832759807294e-05, "loss": 0.6963, "step": 337900 }, { "epoch": 58.14693737095664, "grad_norm": 17.292314529418945, "learning_rate": 2.092653131452168e-05, "loss": 0.7417, "step": 337950 }, { "epoch": 58.155540261527875, "grad_norm": 9.948911666870117, "learning_rate": 2.0922229869236066e-05, "loss": 0.7354, "step": 338000 }, { "epoch": 58.16414315209911, "grad_norm": 14.593234062194824, "learning_rate": 2.0917928423950448e-05, "loss": 0.761, "step": 338050 }, { "epoch": 58.17274604267034, "grad_norm": 11.848487854003906, "learning_rate": 2.0913626978664832e-05, "loss": 0.753, "step": 338100 }, { "epoch": 58.181348933241566, "grad_norm": 13.397107124328613, "learning_rate": 2.0909325533379217e-05, "loss": 0.7649, "step": 338150 }, { "epoch": 58.1899518238128, "grad_norm": 13.911908149719238, "learning_rate": 2.09050240880936e-05, "loss": 0.7221, "step": 338200 }, { "epoch": 58.19855471438403, "grad_norm": 11.62131118774414, "learning_rate": 2.0900722642807986e-05, "loss": 0.6643, "step": 338250 }, { "epoch": 58.20715760495526, "grad_norm": 20.84575843811035, "learning_rate": 2.0896421197522367e-05, "loss": 0.739, "step": 338300 }, { "epoch": 58.215760495526496, "grad_norm": 14.327361106872559, "learning_rate": 2.089211975223675e-05, "loss": 0.7576, "step": 338350 }, { "epoch": 58.22436338609773, "grad_norm": 13.78609848022461, "learning_rate": 2.0887818306951136e-05, "loss": 0.6561, "step": 338400 }, { "epoch": 58.23296627666896, "grad_norm": 15.964612007141113, "learning_rate": 2.088351686166552e-05, "loss": 0.6801, "step": 338450 }, { "epoch": 58.24156916724019, "grad_norm": 12.462899208068848, "learning_rate": 2.0879215416379905e-05, "loss": 0.7388, "step": 338500 }, { "epoch": 58.250172057811426, "grad_norm": 7.77956485748291, "learning_rate": 2.087491397109429e-05, "loss": 0.6846, "step": 338550 }, { "epoch": 58.25877494838266, "grad_norm": 18.785776138305664, "learning_rate": 2.087061252580867e-05, "loss": 0.8018, "step": 338600 }, { "epoch": 58.26737783895389, "grad_norm": 16.278770446777344, "learning_rate": 2.0866311080523056e-05, "loss": 0.7094, "step": 338650 }, { "epoch": 58.27598072952512, "grad_norm": 18.490821838378906, "learning_rate": 2.086200963523744e-05, "loss": 0.7399, "step": 338700 }, { "epoch": 58.28458362009635, "grad_norm": 18.46456527709961, "learning_rate": 2.0857708189951825e-05, "loss": 0.7142, "step": 338750 }, { "epoch": 58.29318651066758, "grad_norm": 9.501691818237305, "learning_rate": 2.085340674466621e-05, "loss": 0.7366, "step": 338800 }, { "epoch": 58.301789401238814, "grad_norm": 12.474523544311523, "learning_rate": 2.084910529938059e-05, "loss": 0.7327, "step": 338850 }, { "epoch": 58.31039229181005, "grad_norm": 11.07871150970459, "learning_rate": 2.084480385409498e-05, "loss": 0.7485, "step": 338900 }, { "epoch": 58.31899518238128, "grad_norm": 15.166801452636719, "learning_rate": 2.0840502408809363e-05, "loss": 0.7631, "step": 338950 }, { "epoch": 58.32759807295251, "grad_norm": 14.496978759765625, "learning_rate": 2.0836200963523744e-05, "loss": 0.701, "step": 339000 }, { "epoch": 58.336200963523744, "grad_norm": 14.268905639648438, "learning_rate": 2.083189951823813e-05, "loss": 0.7565, "step": 339050 }, { "epoch": 58.34480385409498, "grad_norm": 18.27751350402832, "learning_rate": 2.0827598072952513e-05, "loss": 0.7914, "step": 339100 }, { "epoch": 58.35340674466621, "grad_norm": 9.100638389587402, "learning_rate": 2.0823296627666898e-05, "loss": 0.7746, "step": 339150 }, { "epoch": 58.36200963523744, "grad_norm": 8.620205879211426, "learning_rate": 2.0818995182381282e-05, "loss": 0.7368, "step": 339200 }, { "epoch": 58.370612525808674, "grad_norm": 8.666815757751465, "learning_rate": 2.0814693737095663e-05, "loss": 0.757, "step": 339250 }, { "epoch": 58.37921541637991, "grad_norm": 17.34161376953125, "learning_rate": 2.0810392291810048e-05, "loss": 0.7244, "step": 339300 }, { "epoch": 58.38781830695113, "grad_norm": 14.199873924255371, "learning_rate": 2.0806090846524433e-05, "loss": 0.7867, "step": 339350 }, { "epoch": 58.396421197522365, "grad_norm": 10.834476470947266, "learning_rate": 2.0801789401238817e-05, "loss": 0.7384, "step": 339400 }, { "epoch": 58.4050240880936, "grad_norm": 15.042659759521484, "learning_rate": 2.0797487955953202e-05, "loss": 0.7065, "step": 339450 }, { "epoch": 58.41362697866483, "grad_norm": 15.556252479553223, "learning_rate": 2.0793186510667586e-05, "loss": 0.7463, "step": 339500 }, { "epoch": 58.42222986923606, "grad_norm": 13.676641464233398, "learning_rate": 2.0788885065381967e-05, "loss": 0.7431, "step": 339550 }, { "epoch": 58.430832759807295, "grad_norm": 15.619707107543945, "learning_rate": 2.0784583620096355e-05, "loss": 0.7069, "step": 339600 }, { "epoch": 58.43943565037853, "grad_norm": 19.349645614624023, "learning_rate": 2.0780282174810737e-05, "loss": 0.7669, "step": 339650 }, { "epoch": 58.44803854094976, "grad_norm": 11.061793327331543, "learning_rate": 2.077598072952512e-05, "loss": 0.7469, "step": 339700 }, { "epoch": 58.45664143152099, "grad_norm": 11.054930686950684, "learning_rate": 2.0771679284239506e-05, "loss": 0.7333, "step": 339750 }, { "epoch": 58.465244322092225, "grad_norm": 9.77147388458252, "learning_rate": 2.076737783895389e-05, "loss": 0.7681, "step": 339800 }, { "epoch": 58.47384721266346, "grad_norm": 18.100255966186523, "learning_rate": 2.0763076393668275e-05, "loss": 0.7221, "step": 339850 }, { "epoch": 58.48245010323469, "grad_norm": 9.72760009765625, "learning_rate": 2.075877494838266e-05, "loss": 0.7247, "step": 339900 }, { "epoch": 58.491052993805916, "grad_norm": 12.544808387756348, "learning_rate": 2.075447350309704e-05, "loss": 0.7221, "step": 339950 }, { "epoch": 58.49965588437715, "grad_norm": 13.47618579864502, "learning_rate": 2.0750172057811425e-05, "loss": 0.729, "step": 340000 }, { "epoch": 58.50825877494838, "grad_norm": 13.98698902130127, "learning_rate": 2.074587061252581e-05, "loss": 0.7423, "step": 340050 }, { "epoch": 58.51686166551961, "grad_norm": 11.020320892333984, "learning_rate": 2.0741569167240194e-05, "loss": 0.767, "step": 340100 }, { "epoch": 58.525464556090846, "grad_norm": 11.076303482055664, "learning_rate": 2.073726772195458e-05, "loss": 0.7385, "step": 340150 }, { "epoch": 58.53406744666208, "grad_norm": 23.13214111328125, "learning_rate": 2.073296627666896e-05, "loss": 0.7635, "step": 340200 }, { "epoch": 58.54267033723331, "grad_norm": 10.413269996643066, "learning_rate": 2.0728664831383345e-05, "loss": 0.6656, "step": 340250 }, { "epoch": 58.55127322780454, "grad_norm": 16.825286865234375, "learning_rate": 2.0724363386097732e-05, "loss": 0.7436, "step": 340300 }, { "epoch": 58.559876118375776, "grad_norm": 17.658260345458984, "learning_rate": 2.0720061940812114e-05, "loss": 0.7228, "step": 340350 }, { "epoch": 58.56847900894701, "grad_norm": 16.93277359008789, "learning_rate": 2.0715760495526498e-05, "loss": 0.7485, "step": 340400 }, { "epoch": 58.57708189951824, "grad_norm": 12.80220890045166, "learning_rate": 2.071145905024088e-05, "loss": 0.7962, "step": 340450 }, { "epoch": 58.58568479008947, "grad_norm": 22.196002960205078, "learning_rate": 2.0707157604955267e-05, "loss": 0.7913, "step": 340500 }, { "epoch": 58.5942876806607, "grad_norm": 13.317195892333984, "learning_rate": 2.0702856159669652e-05, "loss": 0.7447, "step": 340550 }, { "epoch": 58.60289057123193, "grad_norm": 13.379812240600586, "learning_rate": 2.0698554714384033e-05, "loss": 0.7721, "step": 340600 }, { "epoch": 58.611493461803164, "grad_norm": 18.407155990600586, "learning_rate": 2.0694253269098418e-05, "loss": 0.7324, "step": 340650 }, { "epoch": 58.6200963523744, "grad_norm": 23.834829330444336, "learning_rate": 2.0689951823812802e-05, "loss": 0.7314, "step": 340700 }, { "epoch": 58.62869924294563, "grad_norm": 12.901362419128418, "learning_rate": 2.0685650378527187e-05, "loss": 0.7546, "step": 340750 }, { "epoch": 58.63730213351686, "grad_norm": 20.70806884765625, "learning_rate": 2.068134893324157e-05, "loss": 0.7557, "step": 340800 }, { "epoch": 58.645905024088094, "grad_norm": 13.947811126708984, "learning_rate": 2.0677047487955952e-05, "loss": 0.7682, "step": 340850 }, { "epoch": 58.65450791465933, "grad_norm": 22.483861923217773, "learning_rate": 2.0672746042670337e-05, "loss": 0.7384, "step": 340900 }, { "epoch": 58.66311080523056, "grad_norm": 21.080472946166992, "learning_rate": 2.066844459738472e-05, "loss": 0.7701, "step": 340950 }, { "epoch": 58.67171369580179, "grad_norm": 10.169876098632812, "learning_rate": 2.0664143152099106e-05, "loss": 0.7376, "step": 341000 }, { "epoch": 58.680316586373024, "grad_norm": 20.48004913330078, "learning_rate": 2.065984170681349e-05, "loss": 0.7244, "step": 341050 }, { "epoch": 58.68891947694425, "grad_norm": 15.4773530960083, "learning_rate": 2.0655540261527875e-05, "loss": 0.774, "step": 341100 }, { "epoch": 58.69752236751548, "grad_norm": 11.271245002746582, "learning_rate": 2.0651238816242256e-05, "loss": 0.7175, "step": 341150 }, { "epoch": 58.706125258086715, "grad_norm": 15.375006675720215, "learning_rate": 2.0646937370956644e-05, "loss": 0.7041, "step": 341200 }, { "epoch": 58.71472814865795, "grad_norm": 16.20039939880371, "learning_rate": 2.064263592567103e-05, "loss": 0.8221, "step": 341250 }, { "epoch": 58.72333103922918, "grad_norm": 15.984435081481934, "learning_rate": 2.063833448038541e-05, "loss": 0.7928, "step": 341300 }, { "epoch": 58.73193392980041, "grad_norm": 16.71038055419922, "learning_rate": 2.0634033035099795e-05, "loss": 0.717, "step": 341350 }, { "epoch": 58.740536820371645, "grad_norm": 12.002765655517578, "learning_rate": 2.0629731589814176e-05, "loss": 0.7324, "step": 341400 }, { "epoch": 58.74913971094288, "grad_norm": 16.05768394470215, "learning_rate": 2.0625430144528564e-05, "loss": 0.7092, "step": 341450 }, { "epoch": 58.75774260151411, "grad_norm": 20.251985549926758, "learning_rate": 2.062112869924295e-05, "loss": 0.7379, "step": 341500 }, { "epoch": 58.76634549208534, "grad_norm": 15.80126953125, "learning_rate": 2.061682725395733e-05, "loss": 0.7259, "step": 341550 }, { "epoch": 58.774948382656575, "grad_norm": 16.098609924316406, "learning_rate": 2.0612525808671714e-05, "loss": 0.7464, "step": 341600 }, { "epoch": 58.78355127322781, "grad_norm": 11.264914512634277, "learning_rate": 2.06082243633861e-05, "loss": 0.7262, "step": 341650 }, { "epoch": 58.79215416379903, "grad_norm": 18.674219131469727, "learning_rate": 2.0603922918100483e-05, "loss": 0.7644, "step": 341700 }, { "epoch": 58.800757054370266, "grad_norm": 10.141558647155762, "learning_rate": 2.0599621472814868e-05, "loss": 0.6539, "step": 341750 }, { "epoch": 58.8093599449415, "grad_norm": 18.81056785583496, "learning_rate": 2.059532002752925e-05, "loss": 0.7543, "step": 341800 }, { "epoch": 58.81796283551273, "grad_norm": 15.665589332580566, "learning_rate": 2.0591018582243634e-05, "loss": 0.7519, "step": 341850 }, { "epoch": 58.82656572608396, "grad_norm": 21.89398193359375, "learning_rate": 2.0586717136958018e-05, "loss": 0.767, "step": 341900 }, { "epoch": 58.835168616655196, "grad_norm": 11.811440467834473, "learning_rate": 2.0582415691672403e-05, "loss": 0.7976, "step": 341950 }, { "epoch": 58.84377150722643, "grad_norm": 15.315133094787598, "learning_rate": 2.0578114246386787e-05, "loss": 0.7115, "step": 342000 }, { "epoch": 58.85237439779766, "grad_norm": 12.596988677978516, "learning_rate": 2.0573812801101172e-05, "loss": 0.739, "step": 342050 }, { "epoch": 58.86097728836889, "grad_norm": 19.625179290771484, "learning_rate": 2.0569511355815553e-05, "loss": 0.7315, "step": 342100 }, { "epoch": 58.869580178940126, "grad_norm": 28.587936401367188, "learning_rate": 2.056520991052994e-05, "loss": 0.7432, "step": 342150 }, { "epoch": 58.87818306951136, "grad_norm": 14.382222175598145, "learning_rate": 2.0560908465244322e-05, "loss": 0.7701, "step": 342200 }, { "epoch": 58.88678596008259, "grad_norm": 12.582257270812988, "learning_rate": 2.0556607019958707e-05, "loss": 0.8189, "step": 342250 }, { "epoch": 58.895388850653816, "grad_norm": 15.900382041931152, "learning_rate": 2.055230557467309e-05, "loss": 0.7124, "step": 342300 }, { "epoch": 58.90399174122505, "grad_norm": 12.896328926086426, "learning_rate": 2.0548004129387476e-05, "loss": 0.7162, "step": 342350 }, { "epoch": 58.91259463179628, "grad_norm": 11.081790924072266, "learning_rate": 2.054370268410186e-05, "loss": 0.734, "step": 342400 }, { "epoch": 58.921197522367514, "grad_norm": 18.824600219726562, "learning_rate": 2.0539401238816245e-05, "loss": 0.7534, "step": 342450 }, { "epoch": 58.92980041293875, "grad_norm": 12.467455863952637, "learning_rate": 2.0535099793530626e-05, "loss": 0.7337, "step": 342500 }, { "epoch": 58.93840330350998, "grad_norm": 15.794851303100586, "learning_rate": 2.053079834824501e-05, "loss": 0.7465, "step": 342550 }, { "epoch": 58.94700619408121, "grad_norm": 14.597437858581543, "learning_rate": 2.0526496902959395e-05, "loss": 0.7752, "step": 342600 }, { "epoch": 58.955609084652444, "grad_norm": 10.337575912475586, "learning_rate": 2.052219545767378e-05, "loss": 0.7195, "step": 342650 }, { "epoch": 58.96421197522368, "grad_norm": 16.082294464111328, "learning_rate": 2.0517894012388164e-05, "loss": 0.815, "step": 342700 }, { "epoch": 58.97281486579491, "grad_norm": 17.734437942504883, "learning_rate": 2.0513592567102545e-05, "loss": 0.7839, "step": 342750 }, { "epoch": 58.98141775636614, "grad_norm": 15.511224746704102, "learning_rate": 2.050929112181693e-05, "loss": 0.8178, "step": 342800 }, { "epoch": 58.990020646937374, "grad_norm": 14.50933837890625, "learning_rate": 2.0504989676531318e-05, "loss": 0.7902, "step": 342850 }, { "epoch": 58.9986235375086, "grad_norm": 15.562018394470215, "learning_rate": 2.05006882312457e-05, "loss": 0.7521, "step": 342900 }, { "epoch": 59.0, "eval_accuracy": 0.5616072965066253, "eval_f1": 0.5561522068757871, "eval_f1_DuraRiadoRio_16x16": 0.5311013886246909, "eval_f1_Mole_16x16": 0.5987831066571224, "eval_f1_Quebrado_16x16": 0.741970252209528, "eval_f1_RiadoRio_16x16": 0.3914358855015127, "eval_f1_RioFechado_16x16": 0.5174704013860815, "eval_loss": 2.132805347442627, "eval_precision": 0.5852561487351997, "eval_precision_DuraRiadoRio_16x16": 0.4727395868608195, "eval_precision_Mole_16x16": 0.5094397076735688, "eval_precision_Quebrado_16x16": 0.7370449678800857, "eval_precision_RiadoRio_16x16": 0.4509383378016086, "eval_precision_RioFechado_16x16": 0.7561181434599156, "eval_recall": 0.5636252913708033, "eval_recall_DuraRiadoRio_16x16": 0.6059027777777778, "eval_recall_Mole_16x16": 0.7261284722222222, "eval_recall_Quebrado_16x16": 0.7469618055555556, "eval_recall_RiadoRio_16x16": 0.3458059210526316, "eval_recall_RioFechado_16x16": 0.3933274802458297, "eval_runtime": 46.3871, "eval_samples_per_second": 250.544, "eval_steps_per_second": 15.672, "step": 342908 }, { "epoch": 59.00722642807983, "grad_norm": 7.817824840545654, "learning_rate": 2.0496386785960084e-05, "loss": 0.6971, "step": 342950 }, { "epoch": 59.015829318651065, "grad_norm": 17.167625427246094, "learning_rate": 2.0492085340674468e-05, "loss": 0.7277, "step": 343000 }, { "epoch": 59.0244322092223, "grad_norm": 18.04184341430664, "learning_rate": 2.0487783895388853e-05, "loss": 0.7244, "step": 343050 }, { "epoch": 59.03303509979353, "grad_norm": 14.046208381652832, "learning_rate": 2.0483482450103237e-05, "loss": 0.7386, "step": 343100 }, { "epoch": 59.04163799036476, "grad_norm": 18.615379333496094, "learning_rate": 2.047918100481762e-05, "loss": 0.7264, "step": 343150 }, { "epoch": 59.050240880935995, "grad_norm": 13.879512786865234, "learning_rate": 2.0474879559532003e-05, "loss": 0.7794, "step": 343200 }, { "epoch": 59.05884377150723, "grad_norm": 15.2373685836792, "learning_rate": 2.0470578114246388e-05, "loss": 0.7142, "step": 343250 }, { "epoch": 59.06744666207846, "grad_norm": 15.43734359741211, "learning_rate": 2.0466276668960772e-05, "loss": 0.732, "step": 343300 }, { "epoch": 59.07604955264969, "grad_norm": 8.963787078857422, "learning_rate": 2.0461975223675157e-05, "loss": 0.7206, "step": 343350 }, { "epoch": 59.084652443220925, "grad_norm": 15.035676002502441, "learning_rate": 2.045767377838954e-05, "loss": 0.7527, "step": 343400 }, { "epoch": 59.09325533379216, "grad_norm": 10.732107162475586, "learning_rate": 2.0453372333103923e-05, "loss": 0.7398, "step": 343450 }, { "epoch": 59.10185822436338, "grad_norm": 15.693502426147461, "learning_rate": 2.0449070887818307e-05, "loss": 0.7892, "step": 343500 }, { "epoch": 59.110461114934616, "grad_norm": 10.905645370483398, "learning_rate": 2.044476944253269e-05, "loss": 0.7832, "step": 343550 }, { "epoch": 59.11906400550585, "grad_norm": 12.824884414672852, "learning_rate": 2.0440467997247076e-05, "loss": 0.7197, "step": 343600 }, { "epoch": 59.12766689607708, "grad_norm": 14.507590293884277, "learning_rate": 2.043616655196146e-05, "loss": 0.769, "step": 343650 }, { "epoch": 59.13626978664831, "grad_norm": 13.159537315368652, "learning_rate": 2.0431865106675842e-05, "loss": 0.7219, "step": 343700 }, { "epoch": 59.144872677219546, "grad_norm": 14.902888298034668, "learning_rate": 2.042756366139023e-05, "loss": 0.7898, "step": 343750 }, { "epoch": 59.15347556779078, "grad_norm": 15.260177612304688, "learning_rate": 2.0423262216104614e-05, "loss": 0.7325, "step": 343800 }, { "epoch": 59.16207845836201, "grad_norm": 15.72882080078125, "learning_rate": 2.0418960770818996e-05, "loss": 0.7208, "step": 343850 }, { "epoch": 59.17068134893324, "grad_norm": 10.265575408935547, "learning_rate": 2.041465932553338e-05, "loss": 0.7209, "step": 343900 }, { "epoch": 59.179284239504476, "grad_norm": 15.913309097290039, "learning_rate": 2.041035788024776e-05, "loss": 0.7491, "step": 343950 }, { "epoch": 59.18788713007571, "grad_norm": 16.047191619873047, "learning_rate": 2.040605643496215e-05, "loss": 0.7507, "step": 344000 }, { "epoch": 59.19649002064694, "grad_norm": 18.993873596191406, "learning_rate": 2.0401754989676534e-05, "loss": 0.7369, "step": 344050 }, { "epoch": 59.205092911218166, "grad_norm": 11.19983196258545, "learning_rate": 2.0397453544390915e-05, "loss": 0.7451, "step": 344100 }, { "epoch": 59.2136958017894, "grad_norm": 11.035270690917969, "learning_rate": 2.03931520991053e-05, "loss": 0.7786, "step": 344150 }, { "epoch": 59.22229869236063, "grad_norm": 18.55753517150879, "learning_rate": 2.0388850653819684e-05, "loss": 0.7771, "step": 344200 }, { "epoch": 59.230901582931864, "grad_norm": 20.140731811523438, "learning_rate": 2.038454920853407e-05, "loss": 0.7661, "step": 344250 }, { "epoch": 59.2395044735031, "grad_norm": 24.871421813964844, "learning_rate": 2.0380247763248453e-05, "loss": 0.7333, "step": 344300 }, { "epoch": 59.24810736407433, "grad_norm": 15.984467506408691, "learning_rate": 2.0375946317962834e-05, "loss": 0.7442, "step": 344350 }, { "epoch": 59.25671025464556, "grad_norm": 13.344985961914062, "learning_rate": 2.037164487267722e-05, "loss": 0.74, "step": 344400 }, { "epoch": 59.265313145216794, "grad_norm": 17.773405075073242, "learning_rate": 2.0367343427391607e-05, "loss": 0.7262, "step": 344450 }, { "epoch": 59.27391603578803, "grad_norm": 13.961681365966797, "learning_rate": 2.0363041982105988e-05, "loss": 0.7661, "step": 344500 }, { "epoch": 59.28251892635926, "grad_norm": 11.352237701416016, "learning_rate": 2.0358740536820373e-05, "loss": 0.7225, "step": 344550 }, { "epoch": 59.29112181693049, "grad_norm": 17.04424285888672, "learning_rate": 2.0354439091534757e-05, "loss": 0.7319, "step": 344600 }, { "epoch": 59.29972470750172, "grad_norm": 11.100432395935059, "learning_rate": 2.035013764624914e-05, "loss": 0.7587, "step": 344650 }, { "epoch": 59.30832759807295, "grad_norm": 10.430562973022461, "learning_rate": 2.0345836200963526e-05, "loss": 0.6851, "step": 344700 }, { "epoch": 59.31693048864418, "grad_norm": 11.648191452026367, "learning_rate": 2.034153475567791e-05, "loss": 0.7795, "step": 344750 }, { "epoch": 59.325533379215415, "grad_norm": 22.378097534179688, "learning_rate": 2.0337233310392292e-05, "loss": 0.779, "step": 344800 }, { "epoch": 59.33413626978665, "grad_norm": 16.13587760925293, "learning_rate": 2.0332931865106677e-05, "loss": 0.7595, "step": 344850 }, { "epoch": 59.34273916035788, "grad_norm": 21.702024459838867, "learning_rate": 2.032863041982106e-05, "loss": 0.7565, "step": 344900 }, { "epoch": 59.35134205092911, "grad_norm": 12.382328033447266, "learning_rate": 2.0324328974535446e-05, "loss": 0.6971, "step": 344950 }, { "epoch": 59.359944941500345, "grad_norm": 11.371623039245605, "learning_rate": 2.032002752924983e-05, "loss": 0.7627, "step": 345000 }, { "epoch": 59.36854783207158, "grad_norm": 9.550614356994629, "learning_rate": 2.031572608396421e-05, "loss": 0.6992, "step": 345050 }, { "epoch": 59.37715072264281, "grad_norm": 17.3050594329834, "learning_rate": 2.0311424638678596e-05, "loss": 0.7333, "step": 345100 }, { "epoch": 59.38575361321404, "grad_norm": 17.944347381591797, "learning_rate": 2.030712319339298e-05, "loss": 0.7554, "step": 345150 }, { "epoch": 59.394356503785275, "grad_norm": 9.864151954650879, "learning_rate": 2.0302821748107365e-05, "loss": 0.7382, "step": 345200 }, { "epoch": 59.4029593943565, "grad_norm": 8.915443420410156, "learning_rate": 2.029852030282175e-05, "loss": 0.7163, "step": 345250 }, { "epoch": 59.41156228492773, "grad_norm": 10.671365737915039, "learning_rate": 2.029421885753613e-05, "loss": 0.7115, "step": 345300 }, { "epoch": 59.420165175498965, "grad_norm": 9.496668815612793, "learning_rate": 2.0289917412250515e-05, "loss": 0.7445, "step": 345350 }, { "epoch": 59.4287680660702, "grad_norm": 13.826031684875488, "learning_rate": 2.0285615966964903e-05, "loss": 0.6839, "step": 345400 }, { "epoch": 59.43737095664143, "grad_norm": 15.458892822265625, "learning_rate": 2.0281314521679285e-05, "loss": 0.746, "step": 345450 }, { "epoch": 59.44597384721266, "grad_norm": 12.730711936950684, "learning_rate": 2.027701307639367e-05, "loss": 0.7313, "step": 345500 }, { "epoch": 59.454576737783896, "grad_norm": 13.290470123291016, "learning_rate": 2.0272711631108054e-05, "loss": 0.7295, "step": 345550 }, { "epoch": 59.46317962835513, "grad_norm": 14.591571807861328, "learning_rate": 2.0268410185822438e-05, "loss": 0.7788, "step": 345600 }, { "epoch": 59.47178251892636, "grad_norm": 17.645923614501953, "learning_rate": 2.0264108740536823e-05, "loss": 0.7246, "step": 345650 }, { "epoch": 59.48038540949759, "grad_norm": 11.566411018371582, "learning_rate": 2.0259807295251204e-05, "loss": 0.7198, "step": 345700 }, { "epoch": 59.488988300068826, "grad_norm": 11.482674598693848, "learning_rate": 2.025550584996559e-05, "loss": 0.7053, "step": 345750 }, { "epoch": 59.49759119064006, "grad_norm": 16.530990600585938, "learning_rate": 2.0251204404679973e-05, "loss": 0.7223, "step": 345800 }, { "epoch": 59.506194081211284, "grad_norm": 12.134199142456055, "learning_rate": 2.0246902959394358e-05, "loss": 0.7895, "step": 345850 }, { "epoch": 59.514796971782516, "grad_norm": 14.150046348571777, "learning_rate": 2.0242601514108742e-05, "loss": 0.7437, "step": 345900 }, { "epoch": 59.52339986235375, "grad_norm": 12.211923599243164, "learning_rate": 2.0238300068823127e-05, "loss": 0.7538, "step": 345950 }, { "epoch": 59.53200275292498, "grad_norm": 13.1089506149292, "learning_rate": 2.0233998623537508e-05, "loss": 0.7245, "step": 346000 }, { "epoch": 59.540605643496214, "grad_norm": 10.286578178405762, "learning_rate": 2.0229697178251893e-05, "loss": 0.7235, "step": 346050 }, { "epoch": 59.549208534067446, "grad_norm": 19.36136245727539, "learning_rate": 2.0225395732966277e-05, "loss": 0.7171, "step": 346100 }, { "epoch": 59.55781142463868, "grad_norm": 12.616678237915039, "learning_rate": 2.022109428768066e-05, "loss": 0.7191, "step": 346150 }, { "epoch": 59.56641431520991, "grad_norm": 10.595259666442871, "learning_rate": 2.0216792842395046e-05, "loss": 0.7063, "step": 346200 }, { "epoch": 59.575017205781144, "grad_norm": 13.942296981811523, "learning_rate": 2.0212491397109427e-05, "loss": 0.7297, "step": 346250 }, { "epoch": 59.58362009635238, "grad_norm": 15.760912895202637, "learning_rate": 2.0208189951823815e-05, "loss": 0.7055, "step": 346300 }, { "epoch": 59.59222298692361, "grad_norm": 12.313725471496582, "learning_rate": 2.02038885065382e-05, "loss": 0.7368, "step": 346350 }, { "epoch": 59.60082587749484, "grad_norm": 11.098084449768066, "learning_rate": 2.019958706125258e-05, "loss": 0.7181, "step": 346400 }, { "epoch": 59.60942876806607, "grad_norm": 27.808420181274414, "learning_rate": 2.0195285615966966e-05, "loss": 0.7408, "step": 346450 }, { "epoch": 59.6180316586373, "grad_norm": 21.19243049621582, "learning_rate": 2.019098417068135e-05, "loss": 0.7706, "step": 346500 }, { "epoch": 59.62663454920853, "grad_norm": 12.607558250427246, "learning_rate": 2.0186682725395735e-05, "loss": 0.7603, "step": 346550 }, { "epoch": 59.635237439779765, "grad_norm": 15.622058868408203, "learning_rate": 2.018238128011012e-05, "loss": 0.7939, "step": 346600 }, { "epoch": 59.643840330351, "grad_norm": 20.028059005737305, "learning_rate": 2.01780798348245e-05, "loss": 0.6659, "step": 346650 }, { "epoch": 59.65244322092223, "grad_norm": 17.769474029541016, "learning_rate": 2.0173778389538885e-05, "loss": 0.8108, "step": 346700 }, { "epoch": 59.66104611149346, "grad_norm": 17.860065460205078, "learning_rate": 2.016947694425327e-05, "loss": 0.7948, "step": 346750 }, { "epoch": 59.669649002064695, "grad_norm": 8.436922073364258, "learning_rate": 2.0165175498967654e-05, "loss": 0.7729, "step": 346800 }, { "epoch": 59.67825189263593, "grad_norm": 19.750091552734375, "learning_rate": 2.016087405368204e-05, "loss": 0.7358, "step": 346850 }, { "epoch": 59.68685478320716, "grad_norm": 11.387706756591797, "learning_rate": 2.0156572608396423e-05, "loss": 0.7254, "step": 346900 }, { "epoch": 59.69545767377839, "grad_norm": 15.618230819702148, "learning_rate": 2.0152271163110804e-05, "loss": 0.7105, "step": 346950 }, { "epoch": 59.704060564349625, "grad_norm": 21.005563735961914, "learning_rate": 2.0147969717825192e-05, "loss": 0.763, "step": 347000 }, { "epoch": 59.71266345492085, "grad_norm": 9.173162460327148, "learning_rate": 2.0143668272539574e-05, "loss": 0.753, "step": 347050 }, { "epoch": 59.72126634549208, "grad_norm": 7.5767741203308105, "learning_rate": 2.0139366827253958e-05, "loss": 0.733, "step": 347100 }, { "epoch": 59.729869236063315, "grad_norm": 8.505270004272461, "learning_rate": 2.0135065381968343e-05, "loss": 0.7372, "step": 347150 }, { "epoch": 59.73847212663455, "grad_norm": 12.530915260314941, "learning_rate": 2.0130763936682724e-05, "loss": 0.7009, "step": 347200 }, { "epoch": 59.74707501720578, "grad_norm": 12.343570709228516, "learning_rate": 2.0126462491397112e-05, "loss": 0.7375, "step": 347250 }, { "epoch": 59.75567790777701, "grad_norm": 10.735816955566406, "learning_rate": 2.0122161046111496e-05, "loss": 0.7306, "step": 347300 }, { "epoch": 59.764280798348246, "grad_norm": 12.804740905761719, "learning_rate": 2.0117859600825878e-05, "loss": 0.6898, "step": 347350 }, { "epoch": 59.77288368891948, "grad_norm": 7.618077754974365, "learning_rate": 2.0113558155540262e-05, "loss": 0.7534, "step": 347400 }, { "epoch": 59.78148657949071, "grad_norm": 18.36615562438965, "learning_rate": 2.0109256710254647e-05, "loss": 0.7721, "step": 347450 }, { "epoch": 59.79008947006194, "grad_norm": 13.329477310180664, "learning_rate": 2.010495526496903e-05, "loss": 0.7628, "step": 347500 }, { "epoch": 59.798692360633176, "grad_norm": 11.859374046325684, "learning_rate": 2.0100653819683416e-05, "loss": 0.736, "step": 347550 }, { "epoch": 59.80729525120441, "grad_norm": 11.57689094543457, "learning_rate": 2.0096352374397797e-05, "loss": 0.7249, "step": 347600 }, { "epoch": 59.815898141775634, "grad_norm": 10.823155403137207, "learning_rate": 2.009205092911218e-05, "loss": 0.7309, "step": 347650 }, { "epoch": 59.824501032346866, "grad_norm": 12.49885082244873, "learning_rate": 2.008774948382657e-05, "loss": 0.8131, "step": 347700 }, { "epoch": 59.8331039229181, "grad_norm": 12.708916664123535, "learning_rate": 2.008344803854095e-05, "loss": 0.7608, "step": 347750 }, { "epoch": 59.84170681348933, "grad_norm": 10.682587623596191, "learning_rate": 2.0079146593255335e-05, "loss": 0.7141, "step": 347800 }, { "epoch": 59.850309704060564, "grad_norm": 14.350449562072754, "learning_rate": 2.0074845147969716e-05, "loss": 0.7653, "step": 347850 }, { "epoch": 59.858912594631796, "grad_norm": 13.067445755004883, "learning_rate": 2.00705437026841e-05, "loss": 0.7128, "step": 347900 }, { "epoch": 59.86751548520303, "grad_norm": 11.784612655639648, "learning_rate": 2.006624225739849e-05, "loss": 0.7326, "step": 347950 }, { "epoch": 59.87611837577426, "grad_norm": 14.263835906982422, "learning_rate": 2.006194081211287e-05, "loss": 0.7586, "step": 348000 }, { "epoch": 59.884721266345494, "grad_norm": 16.333419799804688, "learning_rate": 2.0057639366827255e-05, "loss": 0.7576, "step": 348050 }, { "epoch": 59.89332415691673, "grad_norm": 12.713354110717773, "learning_rate": 2.005333792154164e-05, "loss": 0.7592, "step": 348100 }, { "epoch": 59.90192704748796, "grad_norm": 22.636009216308594, "learning_rate": 2.0049036476256024e-05, "loss": 0.7771, "step": 348150 }, { "epoch": 59.910529938059184, "grad_norm": 20.167409896850586, "learning_rate": 2.004473503097041e-05, "loss": 0.7804, "step": 348200 }, { "epoch": 59.91913282863042, "grad_norm": 14.882535934448242, "learning_rate": 2.004043358568479e-05, "loss": 0.7527, "step": 348250 }, { "epoch": 59.92773571920165, "grad_norm": 13.550522804260254, "learning_rate": 2.0036132140399174e-05, "loss": 0.6933, "step": 348300 }, { "epoch": 59.93633860977288, "grad_norm": 14.887367248535156, "learning_rate": 2.003183069511356e-05, "loss": 0.6962, "step": 348350 }, { "epoch": 59.944941500344115, "grad_norm": 10.990434646606445, "learning_rate": 2.0027529249827943e-05, "loss": 0.7434, "step": 348400 }, { "epoch": 59.95354439091535, "grad_norm": 13.933392524719238, "learning_rate": 2.0023227804542328e-05, "loss": 0.7547, "step": 348450 }, { "epoch": 59.96214728148658, "grad_norm": 9.250371932983398, "learning_rate": 2.0018926359256712e-05, "loss": 0.7452, "step": 348500 }, { "epoch": 59.97075017205781, "grad_norm": 14.893486976623535, "learning_rate": 2.0014624913971093e-05, "loss": 0.7597, "step": 348550 }, { "epoch": 59.979353062629045, "grad_norm": 10.039148330688477, "learning_rate": 2.0010323468685478e-05, "loss": 0.6894, "step": 348600 }, { "epoch": 59.98795595320028, "grad_norm": 14.594877243041992, "learning_rate": 2.0006022023399866e-05, "loss": 0.7425, "step": 348650 }, { "epoch": 59.99655884377151, "grad_norm": 11.654862403869629, "learning_rate": 2.0001720578114247e-05, "loss": 0.7693, "step": 348700 }, { "epoch": 60.0, "eval_accuracy": 0.5387196695921528, "eval_f1": 0.5341337819425759, "eval_f1_DuraRiadoRio_16x16": 0.4577259475218659, "eval_f1_Mole_16x16": 0.60431654676259, "eval_f1_Quebrado_16x16": 0.6145910095799558, "eval_f1_RiadoRio_16x16": 0.41019607843137257, "eval_f1_RioFechado_16x16": 0.5838393274170948, "eval_loss": 2.017812490463257, "eval_precision": 0.5613790815206289, "eval_precision_DuraRiadoRio_16x16": 0.6971580817051509, "eval_precision_Mole_16x16": 0.56, "eval_precision_Quebrado_16x16": 0.5339308578745199, "eval_precision_RiadoRio_16x16": 0.39205397301349326, "eval_precision_RioFechado_16x16": 0.6237524950099801, "eval_recall": 0.5399491553134739, "eval_recall_DuraRiadoRio_16x16": 0.3407118055555556, "eval_recall_Mole_16x16": 0.65625, "eval_recall_Quebrado_16x16": 0.7239583333333334, "eval_recall_RiadoRio_16x16": 0.4300986842105263, "eval_recall_RioFechado_16x16": 0.5487269534679543, "eval_runtime": 46.2117, "eval_samples_per_second": 251.495, "eval_steps_per_second": 15.732, "step": 348720 }, { "epoch": 60.00516173434274, "grad_norm": 11.105569839477539, "learning_rate": 1.9997419132828632e-05, "loss": 0.7141, "step": 348750 }, { "epoch": 60.01376462491397, "grad_norm": 11.985713958740234, "learning_rate": 1.9993117687543013e-05, "loss": 0.7213, "step": 348800 }, { "epoch": 60.0223675154852, "grad_norm": 18.581478118896484, "learning_rate": 1.99888162422574e-05, "loss": 0.7395, "step": 348850 }, { "epoch": 60.03097040605643, "grad_norm": 10.120575904846191, "learning_rate": 1.9984514796971785e-05, "loss": 0.713, "step": 348900 }, { "epoch": 60.039573296627665, "grad_norm": 15.394792556762695, "learning_rate": 1.9980213351686167e-05, "loss": 0.7158, "step": 348950 }, { "epoch": 60.0481761871989, "grad_norm": 17.999847412109375, "learning_rate": 1.997591190640055e-05, "loss": 0.6944, "step": 349000 }, { "epoch": 60.05677907777013, "grad_norm": 7.738016605377197, "learning_rate": 1.9971610461114936e-05, "loss": 0.7555, "step": 349050 }, { "epoch": 60.06538196834136, "grad_norm": 13.230429649353027, "learning_rate": 1.996730901582932e-05, "loss": 0.7477, "step": 349100 }, { "epoch": 60.073984858912596, "grad_norm": 16.228290557861328, "learning_rate": 1.9963007570543705e-05, "loss": 0.7645, "step": 349150 }, { "epoch": 60.08258774948383, "grad_norm": 17.662158966064453, "learning_rate": 1.9958706125258086e-05, "loss": 0.7466, "step": 349200 }, { "epoch": 60.09119064005506, "grad_norm": 19.67049789428711, "learning_rate": 1.995440467997247e-05, "loss": 0.7442, "step": 349250 }, { "epoch": 60.09979353062629, "grad_norm": 15.289934158325195, "learning_rate": 1.9950103234686855e-05, "loss": 0.6729, "step": 349300 }, { "epoch": 60.108396421197526, "grad_norm": 9.416385650634766, "learning_rate": 1.994580178940124e-05, "loss": 0.714, "step": 349350 }, { "epoch": 60.11699931176875, "grad_norm": 18.740419387817383, "learning_rate": 1.9941500344115624e-05, "loss": 0.7522, "step": 349400 }, { "epoch": 60.125602202339984, "grad_norm": 24.56776237487793, "learning_rate": 1.993719889883001e-05, "loss": 0.7298, "step": 349450 }, { "epoch": 60.134205092911216, "grad_norm": 18.508512496948242, "learning_rate": 1.993289745354439e-05, "loss": 0.7241, "step": 349500 }, { "epoch": 60.14280798348245, "grad_norm": 13.622574806213379, "learning_rate": 1.9928596008258778e-05, "loss": 0.7373, "step": 349550 }, { "epoch": 60.15141087405368, "grad_norm": 12.085604667663574, "learning_rate": 1.992429456297316e-05, "loss": 0.6403, "step": 349600 }, { "epoch": 60.160013764624914, "grad_norm": 14.684967041015625, "learning_rate": 1.9919993117687544e-05, "loss": 0.7055, "step": 349650 }, { "epoch": 60.168616655196146, "grad_norm": 16.021865844726562, "learning_rate": 1.9915691672401928e-05, "loss": 0.7423, "step": 349700 }, { "epoch": 60.17721954576738, "grad_norm": 14.82502269744873, "learning_rate": 1.9911390227116313e-05, "loss": 0.7151, "step": 349750 }, { "epoch": 60.18582243633861, "grad_norm": 19.942031860351562, "learning_rate": 1.9907088781830697e-05, "loss": 0.7189, "step": 349800 }, { "epoch": 60.194425326909844, "grad_norm": 18.73396873474121, "learning_rate": 1.9902787336545082e-05, "loss": 0.7317, "step": 349850 }, { "epoch": 60.20302821748108, "grad_norm": 13.509955406188965, "learning_rate": 1.9898485891259463e-05, "loss": 0.7361, "step": 349900 }, { "epoch": 60.21163110805231, "grad_norm": 17.29391098022461, "learning_rate": 1.9894184445973848e-05, "loss": 0.7005, "step": 349950 }, { "epoch": 60.220233998623534, "grad_norm": 12.239171981811523, "learning_rate": 1.9889883000688232e-05, "loss": 0.7616, "step": 350000 }, { "epoch": 60.22883688919477, "grad_norm": 14.943883895874023, "learning_rate": 1.9885581555402617e-05, "loss": 0.721, "step": 350050 }, { "epoch": 60.237439779766, "grad_norm": 9.476358413696289, "learning_rate": 1.9881280110117e-05, "loss": 0.7303, "step": 350100 }, { "epoch": 60.24604267033723, "grad_norm": 11.458776473999023, "learning_rate": 1.9876978664831382e-05, "loss": 0.7378, "step": 350150 }, { "epoch": 60.254645560908465, "grad_norm": 16.433095932006836, "learning_rate": 1.9872677219545767e-05, "loss": 0.7309, "step": 350200 }, { "epoch": 60.2632484514797, "grad_norm": 7.793898582458496, "learning_rate": 1.9868375774260155e-05, "loss": 0.7726, "step": 350250 }, { "epoch": 60.27185134205093, "grad_norm": 15.38895320892334, "learning_rate": 1.9864074328974536e-05, "loss": 0.6758, "step": 350300 }, { "epoch": 60.28045423262216, "grad_norm": 18.089033126831055, "learning_rate": 1.985977288368892e-05, "loss": 0.7095, "step": 350350 }, { "epoch": 60.289057123193395, "grad_norm": 23.600046157836914, "learning_rate": 1.9855471438403305e-05, "loss": 0.8077, "step": 350400 }, { "epoch": 60.29766001376463, "grad_norm": 16.68874740600586, "learning_rate": 1.9851169993117686e-05, "loss": 0.7829, "step": 350450 }, { "epoch": 60.30626290433586, "grad_norm": 9.603658676147461, "learning_rate": 1.9846868547832074e-05, "loss": 0.728, "step": 350500 }, { "epoch": 60.31486579490709, "grad_norm": 13.02255916595459, "learning_rate": 1.9842567102546456e-05, "loss": 0.7148, "step": 350550 }, { "epoch": 60.32346868547832, "grad_norm": 22.646526336669922, "learning_rate": 1.983826565726084e-05, "loss": 0.7392, "step": 350600 }, { "epoch": 60.33207157604955, "grad_norm": 17.41221046447754, "learning_rate": 1.9833964211975225e-05, "loss": 0.7212, "step": 350650 }, { "epoch": 60.34067446662078, "grad_norm": 18.527481079101562, "learning_rate": 1.982966276668961e-05, "loss": 0.7176, "step": 350700 }, { "epoch": 60.349277357192015, "grad_norm": 12.133559226989746, "learning_rate": 1.9825361321403994e-05, "loss": 0.7777, "step": 350750 }, { "epoch": 60.35788024776325, "grad_norm": 15.424663543701172, "learning_rate": 1.982105987611838e-05, "loss": 0.695, "step": 350800 }, { "epoch": 60.36648313833448, "grad_norm": 13.874470710754395, "learning_rate": 1.981675843083276e-05, "loss": 0.7243, "step": 350850 }, { "epoch": 60.37508602890571, "grad_norm": 13.951666831970215, "learning_rate": 1.9812456985547144e-05, "loss": 0.763, "step": 350900 }, { "epoch": 60.383688919476946, "grad_norm": 17.74287986755371, "learning_rate": 1.980815554026153e-05, "loss": 0.6744, "step": 350950 }, { "epoch": 60.39229181004818, "grad_norm": 17.836605072021484, "learning_rate": 1.9803854094975913e-05, "loss": 0.7458, "step": 351000 }, { "epoch": 60.40089470061941, "grad_norm": 14.945732116699219, "learning_rate": 1.9799552649690298e-05, "loss": 0.6894, "step": 351050 }, { "epoch": 60.40949759119064, "grad_norm": 19.959442138671875, "learning_rate": 1.979525120440468e-05, "loss": 0.7189, "step": 351100 }, { "epoch": 60.41810048176187, "grad_norm": 7.292649745941162, "learning_rate": 1.9790949759119063e-05, "loss": 0.7452, "step": 351150 }, { "epoch": 60.4267033723331, "grad_norm": 17.089263916015625, "learning_rate": 1.978664831383345e-05, "loss": 0.7507, "step": 351200 }, { "epoch": 60.435306262904334, "grad_norm": 13.819695472717285, "learning_rate": 1.9782346868547833e-05, "loss": 0.7053, "step": 351250 }, { "epoch": 60.443909153475566, "grad_norm": 12.785922050476074, "learning_rate": 1.9778045423262217e-05, "loss": 0.7439, "step": 351300 }, { "epoch": 60.4525120440468, "grad_norm": 16.639896392822266, "learning_rate": 1.97737439779766e-05, "loss": 0.7541, "step": 351350 }, { "epoch": 60.46111493461803, "grad_norm": 11.195837020874023, "learning_rate": 1.9769442532690986e-05, "loss": 0.7538, "step": 351400 }, { "epoch": 60.469717825189264, "grad_norm": 10.703227996826172, "learning_rate": 1.976514108740537e-05, "loss": 0.7154, "step": 351450 }, { "epoch": 60.478320715760496, "grad_norm": 10.138917922973633, "learning_rate": 1.9760839642119752e-05, "loss": 0.7121, "step": 351500 }, { "epoch": 60.48692360633173, "grad_norm": 14.955473899841309, "learning_rate": 1.9756538196834137e-05, "loss": 0.7898, "step": 351550 }, { "epoch": 60.49552649690296, "grad_norm": 15.496611595153809, "learning_rate": 1.975223675154852e-05, "loss": 0.7782, "step": 351600 }, { "epoch": 60.504129387474194, "grad_norm": 15.34811019897461, "learning_rate": 1.9747935306262906e-05, "loss": 0.7273, "step": 351650 }, { "epoch": 60.512732278045426, "grad_norm": 13.403105735778809, "learning_rate": 1.974363386097729e-05, "loss": 0.7777, "step": 351700 }, { "epoch": 60.52133516861665, "grad_norm": 12.960915565490723, "learning_rate": 1.973933241569167e-05, "loss": 0.7279, "step": 351750 }, { "epoch": 60.529938059187884, "grad_norm": 18.09662437438965, "learning_rate": 1.9735030970406056e-05, "loss": 0.7976, "step": 351800 }, { "epoch": 60.53854094975912, "grad_norm": 9.567028045654297, "learning_rate": 1.973072952512044e-05, "loss": 0.7103, "step": 351850 }, { "epoch": 60.54714384033035, "grad_norm": 13.616212844848633, "learning_rate": 1.9726428079834825e-05, "loss": 0.7082, "step": 351900 }, { "epoch": 60.55574673090158, "grad_norm": 15.946179389953613, "learning_rate": 1.972212663454921e-05, "loss": 0.75, "step": 351950 }, { "epoch": 60.564349621472815, "grad_norm": 11.34281063079834, "learning_rate": 1.9717825189263594e-05, "loss": 0.7527, "step": 352000 }, { "epoch": 60.57295251204405, "grad_norm": 10.903099060058594, "learning_rate": 1.9713523743977975e-05, "loss": 0.7074, "step": 352050 }, { "epoch": 60.58155540261528, "grad_norm": 16.55138397216797, "learning_rate": 1.9709222298692363e-05, "loss": 0.7419, "step": 352100 }, { "epoch": 60.59015829318651, "grad_norm": 43.85694122314453, "learning_rate": 1.9704920853406748e-05, "loss": 0.7631, "step": 352150 }, { "epoch": 60.598761183757745, "grad_norm": 12.541385650634766, "learning_rate": 1.970061940812113e-05, "loss": 0.6957, "step": 352200 }, { "epoch": 60.60736407432898, "grad_norm": 9.931825637817383, "learning_rate": 1.9696317962835514e-05, "loss": 0.7691, "step": 352250 }, { "epoch": 60.61596696490021, "grad_norm": 9.193642616271973, "learning_rate": 1.9692016517549898e-05, "loss": 0.691, "step": 352300 }, { "epoch": 60.624569855471435, "grad_norm": 7.657919883728027, "learning_rate": 1.9687715072264283e-05, "loss": 0.7799, "step": 352350 }, { "epoch": 60.63317274604267, "grad_norm": 14.330310821533203, "learning_rate": 1.9683413626978667e-05, "loss": 0.7477, "step": 352400 }, { "epoch": 60.6417756366139, "grad_norm": 11.463118553161621, "learning_rate": 1.967911218169305e-05, "loss": 0.6958, "step": 352450 }, { "epoch": 60.65037852718513, "grad_norm": 12.728002548217773, "learning_rate": 1.9674810736407433e-05, "loss": 0.8388, "step": 352500 }, { "epoch": 60.658981417756365, "grad_norm": 16.385576248168945, "learning_rate": 1.9670509291121818e-05, "loss": 0.745, "step": 352550 }, { "epoch": 60.6675843083276, "grad_norm": 15.163694381713867, "learning_rate": 1.9666207845836202e-05, "loss": 0.7283, "step": 352600 }, { "epoch": 60.67618719889883, "grad_norm": 18.830781936645508, "learning_rate": 1.9661906400550587e-05, "loss": 0.748, "step": 352650 }, { "epoch": 60.68479008947006, "grad_norm": 8.353418350219727, "learning_rate": 1.9657604955264968e-05, "loss": 0.7878, "step": 352700 }, { "epoch": 60.693392980041295, "grad_norm": 14.051593780517578, "learning_rate": 1.9653303509979352e-05, "loss": 0.7331, "step": 352750 }, { "epoch": 60.70199587061253, "grad_norm": 13.747638702392578, "learning_rate": 1.964900206469374e-05, "loss": 0.7773, "step": 352800 }, { "epoch": 60.71059876118376, "grad_norm": 9.26309871673584, "learning_rate": 1.964470061940812e-05, "loss": 0.7653, "step": 352850 }, { "epoch": 60.71920165175499, "grad_norm": 19.64033317565918, "learning_rate": 1.9640399174122506e-05, "loss": 0.7276, "step": 352900 }, { "epoch": 60.72780454232622, "grad_norm": 12.335857391357422, "learning_rate": 1.963609772883689e-05, "loss": 0.7333, "step": 352950 }, { "epoch": 60.73640743289745, "grad_norm": 14.616116523742676, "learning_rate": 1.9631796283551275e-05, "loss": 0.7822, "step": 353000 }, { "epoch": 60.745010323468684, "grad_norm": 12.533924102783203, "learning_rate": 1.962749483826566e-05, "loss": 0.7646, "step": 353050 }, { "epoch": 60.753613214039916, "grad_norm": 14.830239295959473, "learning_rate": 1.962319339298004e-05, "loss": 0.7123, "step": 353100 }, { "epoch": 60.76221610461115, "grad_norm": 15.385213851928711, "learning_rate": 1.9618891947694426e-05, "loss": 0.7625, "step": 353150 }, { "epoch": 60.77081899518238, "grad_norm": 10.745696067810059, "learning_rate": 1.961459050240881e-05, "loss": 0.7388, "step": 353200 }, { "epoch": 60.779421885753614, "grad_norm": 13.2278413772583, "learning_rate": 1.9610289057123195e-05, "loss": 0.7238, "step": 353250 }, { "epoch": 60.788024776324846, "grad_norm": 15.13698959350586, "learning_rate": 1.960598761183758e-05, "loss": 0.7213, "step": 353300 }, { "epoch": 60.79662766689608, "grad_norm": 11.936963081359863, "learning_rate": 1.9601686166551964e-05, "loss": 0.7418, "step": 353350 }, { "epoch": 60.80523055746731, "grad_norm": 14.389122009277344, "learning_rate": 1.9597384721266345e-05, "loss": 0.7522, "step": 353400 }, { "epoch": 60.813833448038544, "grad_norm": 11.801043510437012, "learning_rate": 1.959308327598073e-05, "loss": 0.7473, "step": 353450 }, { "epoch": 60.822436338609776, "grad_norm": 11.393340110778809, "learning_rate": 1.9588781830695114e-05, "loss": 0.6726, "step": 353500 }, { "epoch": 60.831039229181, "grad_norm": 21.923728942871094, "learning_rate": 1.95844803854095e-05, "loss": 0.6895, "step": 353550 }, { "epoch": 60.839642119752234, "grad_norm": 8.514999389648438, "learning_rate": 1.9580178940123883e-05, "loss": 0.7435, "step": 353600 }, { "epoch": 60.84824501032347, "grad_norm": 11.6276273727417, "learning_rate": 1.9575877494838264e-05, "loss": 0.7731, "step": 353650 }, { "epoch": 60.8568479008947, "grad_norm": 12.543994903564453, "learning_rate": 1.9571576049552652e-05, "loss": 0.6797, "step": 353700 }, { "epoch": 60.86545079146593, "grad_norm": 18.514551162719727, "learning_rate": 1.9567274604267037e-05, "loss": 0.6951, "step": 353750 }, { "epoch": 60.874053682037164, "grad_norm": 9.171262741088867, "learning_rate": 1.9562973158981418e-05, "loss": 0.7901, "step": 353800 }, { "epoch": 60.8826565726084, "grad_norm": 21.3415584564209, "learning_rate": 1.9558671713695803e-05, "loss": 0.7634, "step": 353850 }, { "epoch": 60.89125946317963, "grad_norm": 19.58678436279297, "learning_rate": 1.9554370268410184e-05, "loss": 0.7402, "step": 353900 }, { "epoch": 60.89986235375086, "grad_norm": 27.22899627685547, "learning_rate": 1.9550068823124572e-05, "loss": 0.7583, "step": 353950 }, { "epoch": 60.908465244322095, "grad_norm": 16.19724464416504, "learning_rate": 1.9545767377838956e-05, "loss": 0.7321, "step": 354000 }, { "epoch": 60.91706813489333, "grad_norm": 7.60676908493042, "learning_rate": 1.9541465932553337e-05, "loss": 0.7904, "step": 354050 }, { "epoch": 60.92567102546455, "grad_norm": 11.822858810424805, "learning_rate": 1.9537164487267722e-05, "loss": 0.6962, "step": 354100 }, { "epoch": 60.934273916035785, "grad_norm": 14.51525592803955, "learning_rate": 1.9532863041982107e-05, "loss": 0.7227, "step": 354150 }, { "epoch": 60.94287680660702, "grad_norm": 11.578539848327637, "learning_rate": 1.952856159669649e-05, "loss": 0.7149, "step": 354200 }, { "epoch": 60.95147969717825, "grad_norm": 27.155981063842773, "learning_rate": 1.9524260151410876e-05, "loss": 0.7632, "step": 354250 }, { "epoch": 60.96008258774948, "grad_norm": 13.233884811401367, "learning_rate": 1.951995870612526e-05, "loss": 0.7154, "step": 354300 }, { "epoch": 60.968685478320715, "grad_norm": 19.222328186035156, "learning_rate": 1.951565726083964e-05, "loss": 0.8059, "step": 354350 }, { "epoch": 60.97728836889195, "grad_norm": 17.42637825012207, "learning_rate": 1.9511355815554026e-05, "loss": 0.7119, "step": 354400 }, { "epoch": 60.98589125946318, "grad_norm": 11.431872367858887, "learning_rate": 1.950705437026841e-05, "loss": 0.7791, "step": 354450 }, { "epoch": 60.99449415003441, "grad_norm": 12.489340782165527, "learning_rate": 1.9502752924982795e-05, "loss": 0.7369, "step": 354500 }, { "epoch": 61.0, "eval_accuracy": 0.5869041473068318, "eval_f1": 0.5758639186428891, "eval_f1_DuraRiadoRio_16x16": 0.5106497708277163, "eval_f1_Mole_16x16": 0.6017830609212481, "eval_f1_Quebrado_16x16": 0.7348097671777399, "eval_f1_RiadoRio_16x16": 0.44850669412976313, "eval_f1_RioFechado_16x16": 0.5835703001579778, "eval_loss": 1.8309677839279175, "eval_precision": 0.6163921727156543, "eval_precision_DuraRiadoRio_16x16": 0.6740213523131673, "eval_precision_Mole_16x16": 0.7006920415224913, "eval_precision_Quebrado_16x16": 0.6515609264853978, "eval_precision_RiadoRio_16x16": 0.599862258953168, "eval_precision_RioFechado_16x16": 0.45582428430404737, "eval_recall": 0.5899512732069786, "eval_recall_DuraRiadoRio_16x16": 0.4110243055555556, "eval_recall_Mole_16x16": 0.52734375, "eval_recall_Quebrado_16x16": 0.8424479166666666, "eval_recall_RiadoRio_16x16": 0.3581414473684211, "eval_recall_RioFechado_16x16": 0.8107989464442493, "eval_runtime": 45.8223, "eval_samples_per_second": 253.632, "eval_steps_per_second": 15.866, "step": 354532 }, { "epoch": 61.003097040605645, "grad_norm": 9.207019805908203, "learning_rate": 1.949845147969718e-05, "loss": 0.7345, "step": 354550 }, { "epoch": 61.01169993117688, "grad_norm": 15.328265190124512, "learning_rate": 1.949415003441156e-05, "loss": 0.7255, "step": 354600 }, { "epoch": 61.02030282174811, "grad_norm": 9.568460464477539, "learning_rate": 1.948984858912595e-05, "loss": 0.7332, "step": 354650 }, { "epoch": 61.028905712319336, "grad_norm": 16.787593841552734, "learning_rate": 1.9485547143840333e-05, "loss": 0.7319, "step": 354700 }, { "epoch": 61.03750860289057, "grad_norm": 22.432830810546875, "learning_rate": 1.9481245698554715e-05, "loss": 0.7023, "step": 354750 }, { "epoch": 61.0461114934618, "grad_norm": 16.937620162963867, "learning_rate": 1.94769442532691e-05, "loss": 0.7883, "step": 354800 }, { "epoch": 61.05471438403303, "grad_norm": 15.13216781616211, "learning_rate": 1.9472642807983484e-05, "loss": 0.7011, "step": 354850 }, { "epoch": 61.063317274604266, "grad_norm": 7.815466403961182, "learning_rate": 1.9468341362697868e-05, "loss": 0.7228, "step": 354900 }, { "epoch": 61.0719201651755, "grad_norm": 14.582608222961426, "learning_rate": 1.9464039917412253e-05, "loss": 0.7277, "step": 354950 }, { "epoch": 61.08052305574673, "grad_norm": 14.656004905700684, "learning_rate": 1.9459738472126634e-05, "loss": 0.7305, "step": 355000 }, { "epoch": 61.089125946317964, "grad_norm": 18.971364974975586, "learning_rate": 1.945543702684102e-05, "loss": 0.703, "step": 355050 }, { "epoch": 61.097728836889196, "grad_norm": 10.934468269348145, "learning_rate": 1.9451135581555403e-05, "loss": 0.6895, "step": 355100 }, { "epoch": 61.10633172746043, "grad_norm": 19.855558395385742, "learning_rate": 1.9446834136269788e-05, "loss": 0.7121, "step": 355150 }, { "epoch": 61.11493461803166, "grad_norm": 15.321056365966797, "learning_rate": 1.9442532690984172e-05, "loss": 0.7683, "step": 355200 }, { "epoch": 61.123537508602894, "grad_norm": 13.261561393737793, "learning_rate": 1.9438231245698553e-05, "loss": 0.6923, "step": 355250 }, { "epoch": 61.13214039917412, "grad_norm": 16.417835235595703, "learning_rate": 1.9433929800412938e-05, "loss": 0.7178, "step": 355300 }, { "epoch": 61.14074328974535, "grad_norm": 14.837423324584961, "learning_rate": 1.9429628355127326e-05, "loss": 0.7334, "step": 355350 }, { "epoch": 61.149346180316584, "grad_norm": 25.346757888793945, "learning_rate": 1.9425326909841707e-05, "loss": 0.7656, "step": 355400 }, { "epoch": 61.15794907088782, "grad_norm": 14.546126365661621, "learning_rate": 1.942102546455609e-05, "loss": 0.7203, "step": 355450 }, { "epoch": 61.16655196145905, "grad_norm": 17.579383850097656, "learning_rate": 1.9416724019270476e-05, "loss": 0.722, "step": 355500 }, { "epoch": 61.17515485203028, "grad_norm": 16.704233169555664, "learning_rate": 1.941242257398486e-05, "loss": 0.7205, "step": 355550 }, { "epoch": 61.183757742601514, "grad_norm": 12.305168151855469, "learning_rate": 1.9408121128699245e-05, "loss": 0.7029, "step": 355600 }, { "epoch": 61.19236063317275, "grad_norm": 13.802740097045898, "learning_rate": 1.9403819683413626e-05, "loss": 0.6957, "step": 355650 }, { "epoch": 61.20096352374398, "grad_norm": 11.818889617919922, "learning_rate": 1.939951823812801e-05, "loss": 0.7457, "step": 355700 }, { "epoch": 61.20956641431521, "grad_norm": 13.121281623840332, "learning_rate": 1.9395216792842396e-05, "loss": 0.7772, "step": 355750 }, { "epoch": 61.218169304886445, "grad_norm": 14.625678062438965, "learning_rate": 1.939091534755678e-05, "loss": 0.7559, "step": 355800 }, { "epoch": 61.22677219545768, "grad_norm": 6.6475934982299805, "learning_rate": 1.9386613902271165e-05, "loss": 0.7322, "step": 355850 }, { "epoch": 61.2353750860289, "grad_norm": 12.515069961547852, "learning_rate": 1.938231245698555e-05, "loss": 0.6858, "step": 355900 }, { "epoch": 61.243977976600135, "grad_norm": 15.997177124023438, "learning_rate": 1.937801101169993e-05, "loss": 0.7384, "step": 355950 }, { "epoch": 61.25258086717137, "grad_norm": 11.435954093933105, "learning_rate": 1.9373709566414315e-05, "loss": 0.7479, "step": 356000 }, { "epoch": 61.2611837577426, "grad_norm": 10.33314037322998, "learning_rate": 1.9369408121128703e-05, "loss": 0.6643, "step": 356050 }, { "epoch": 61.26978664831383, "grad_norm": 18.27266502380371, "learning_rate": 1.9365106675843084e-05, "loss": 0.7219, "step": 356100 }, { "epoch": 61.278389538885065, "grad_norm": 22.689149856567383, "learning_rate": 1.936080523055747e-05, "loss": 0.7455, "step": 356150 }, { "epoch": 61.2869924294563, "grad_norm": 12.316512107849121, "learning_rate": 1.935650378527185e-05, "loss": 0.7662, "step": 356200 }, { "epoch": 61.29559532002753, "grad_norm": 16.220216751098633, "learning_rate": 1.9352202339986238e-05, "loss": 0.7307, "step": 356250 }, { "epoch": 61.30419821059876, "grad_norm": 13.26021671295166, "learning_rate": 1.9347900894700622e-05, "loss": 0.77, "step": 356300 }, { "epoch": 61.312801101169995, "grad_norm": 15.211923599243164, "learning_rate": 1.9343599449415004e-05, "loss": 0.7459, "step": 356350 }, { "epoch": 61.32140399174123, "grad_norm": 8.259087562561035, "learning_rate": 1.9339298004129388e-05, "loss": 0.693, "step": 356400 }, { "epoch": 61.33000688231246, "grad_norm": 18.350322723388672, "learning_rate": 1.9334996558843773e-05, "loss": 0.735, "step": 356450 }, { "epoch": 61.338609772883686, "grad_norm": 15.833524703979492, "learning_rate": 1.9330695113558157e-05, "loss": 0.7077, "step": 356500 }, { "epoch": 61.34721266345492, "grad_norm": 12.411090850830078, "learning_rate": 1.9326393668272542e-05, "loss": 0.7313, "step": 356550 }, { "epoch": 61.35581555402615, "grad_norm": 30.525617599487305, "learning_rate": 1.9322092222986923e-05, "loss": 0.7762, "step": 356600 }, { "epoch": 61.36441844459738, "grad_norm": 16.757883071899414, "learning_rate": 1.9317790777701308e-05, "loss": 0.7295, "step": 356650 }, { "epoch": 61.373021335168616, "grad_norm": 19.240142822265625, "learning_rate": 1.9313489332415692e-05, "loss": 0.6874, "step": 356700 }, { "epoch": 61.38162422573985, "grad_norm": 13.801507949829102, "learning_rate": 1.9309187887130077e-05, "loss": 0.7305, "step": 356750 }, { "epoch": 61.39022711631108, "grad_norm": 15.172185897827148, "learning_rate": 1.930488644184446e-05, "loss": 0.7833, "step": 356800 }, { "epoch": 61.398830006882314, "grad_norm": 17.08031463623047, "learning_rate": 1.9300584996558846e-05, "loss": 0.7354, "step": 356850 }, { "epoch": 61.407432897453546, "grad_norm": 13.681267738342285, "learning_rate": 1.9296283551273227e-05, "loss": 0.765, "step": 356900 }, { "epoch": 61.41603578802478, "grad_norm": 12.702320098876953, "learning_rate": 1.9291982105987615e-05, "loss": 0.7008, "step": 356950 }, { "epoch": 61.42463867859601, "grad_norm": 13.050666809082031, "learning_rate": 1.9287680660701996e-05, "loss": 0.7579, "step": 357000 }, { "epoch": 61.43324156916724, "grad_norm": 15.406116485595703, "learning_rate": 1.928337921541638e-05, "loss": 0.7522, "step": 357050 }, { "epoch": 61.44184445973847, "grad_norm": 16.363019943237305, "learning_rate": 1.9279077770130765e-05, "loss": 0.7851, "step": 357100 }, { "epoch": 61.4504473503097, "grad_norm": 11.503348350524902, "learning_rate": 1.9274776324845146e-05, "loss": 0.7389, "step": 357150 }, { "epoch": 61.459050240880934, "grad_norm": 15.76382827758789, "learning_rate": 1.9270474879559534e-05, "loss": 0.745, "step": 357200 }, { "epoch": 61.46765313145217, "grad_norm": 18.475332260131836, "learning_rate": 1.926617343427392e-05, "loss": 0.7488, "step": 357250 }, { "epoch": 61.4762560220234, "grad_norm": 10.522942543029785, "learning_rate": 1.92618719889883e-05, "loss": 0.6644, "step": 357300 }, { "epoch": 61.48485891259463, "grad_norm": 9.071102142333984, "learning_rate": 1.9257570543702685e-05, "loss": 0.799, "step": 357350 }, { "epoch": 61.493461803165864, "grad_norm": 12.426544189453125, "learning_rate": 1.925326909841707e-05, "loss": 0.7832, "step": 357400 }, { "epoch": 61.5020646937371, "grad_norm": 16.83087158203125, "learning_rate": 1.9248967653131454e-05, "loss": 0.7065, "step": 357450 }, { "epoch": 61.51066758430833, "grad_norm": 12.426697731018066, "learning_rate": 1.9244666207845838e-05, "loss": 0.6909, "step": 357500 }, { "epoch": 61.51927047487956, "grad_norm": 13.03799057006836, "learning_rate": 1.924036476256022e-05, "loss": 0.7163, "step": 357550 }, { "epoch": 61.527873365450795, "grad_norm": 9.338347434997559, "learning_rate": 1.9236063317274604e-05, "loss": 0.6843, "step": 357600 }, { "epoch": 61.53647625602202, "grad_norm": 15.63300895690918, "learning_rate": 1.923176187198899e-05, "loss": 0.7246, "step": 357650 }, { "epoch": 61.54507914659325, "grad_norm": 9.80479621887207, "learning_rate": 1.9227460426703373e-05, "loss": 0.7312, "step": 357700 }, { "epoch": 61.553682037164485, "grad_norm": 14.091988563537598, "learning_rate": 1.9223158981417758e-05, "loss": 0.7306, "step": 357750 }, { "epoch": 61.56228492773572, "grad_norm": 14.158916473388672, "learning_rate": 1.9218857536132142e-05, "loss": 0.7338, "step": 357800 }, { "epoch": 61.57088781830695, "grad_norm": 14.866846084594727, "learning_rate": 1.9214556090846523e-05, "loss": 0.6981, "step": 357850 }, { "epoch": 61.57949070887818, "grad_norm": 14.243191719055176, "learning_rate": 1.921025464556091e-05, "loss": 0.7953, "step": 357900 }, { "epoch": 61.588093599449415, "grad_norm": 9.427383422851562, "learning_rate": 1.9205953200275293e-05, "loss": 0.6904, "step": 357950 }, { "epoch": 61.59669649002065, "grad_norm": 15.301623344421387, "learning_rate": 1.9201651754989677e-05, "loss": 0.7178, "step": 358000 }, { "epoch": 61.60529938059188, "grad_norm": 13.076553344726562, "learning_rate": 1.919735030970406e-05, "loss": 0.728, "step": 358050 }, { "epoch": 61.61390227116311, "grad_norm": 14.462899208068848, "learning_rate": 1.9193048864418446e-05, "loss": 0.7324, "step": 358100 }, { "epoch": 61.622505161734345, "grad_norm": 14.756317138671875, "learning_rate": 1.918874741913283e-05, "loss": 0.7473, "step": 358150 }, { "epoch": 61.63110805230558, "grad_norm": 13.18458080291748, "learning_rate": 1.9184445973847215e-05, "loss": 0.7031, "step": 358200 }, { "epoch": 61.6397109428768, "grad_norm": 15.566884994506836, "learning_rate": 1.9180144528561597e-05, "loss": 0.7472, "step": 358250 }, { "epoch": 61.648313833448036, "grad_norm": 10.641741752624512, "learning_rate": 1.917584308327598e-05, "loss": 0.7601, "step": 358300 }, { "epoch": 61.65691672401927, "grad_norm": 25.533214569091797, "learning_rate": 1.9171541637990366e-05, "loss": 0.6866, "step": 358350 }, { "epoch": 61.6655196145905, "grad_norm": 14.290504455566406, "learning_rate": 1.916724019270475e-05, "loss": 0.6909, "step": 358400 }, { "epoch": 61.67412250516173, "grad_norm": 11.486348152160645, "learning_rate": 1.9162938747419135e-05, "loss": 0.7735, "step": 358450 }, { "epoch": 61.682725395732966, "grad_norm": 20.339536666870117, "learning_rate": 1.9158637302133516e-05, "loss": 0.713, "step": 358500 }, { "epoch": 61.6913282863042, "grad_norm": 12.489684104919434, "learning_rate": 1.91543358568479e-05, "loss": 0.7323, "step": 358550 }, { "epoch": 61.69993117687543, "grad_norm": 20.519161224365234, "learning_rate": 1.915003441156229e-05, "loss": 0.7527, "step": 358600 }, { "epoch": 61.708534067446664, "grad_norm": 18.22690200805664, "learning_rate": 1.914573296627667e-05, "loss": 0.7623, "step": 358650 }, { "epoch": 61.717136958017896, "grad_norm": 8.827290534973145, "learning_rate": 1.9141431520991054e-05, "loss": 0.7409, "step": 358700 }, { "epoch": 61.72573984858913, "grad_norm": 10.675494194030762, "learning_rate": 1.9137130075705435e-05, "loss": 0.7535, "step": 358750 }, { "epoch": 61.73434273916036, "grad_norm": 12.347105979919434, "learning_rate": 1.9132828630419823e-05, "loss": 0.707, "step": 358800 }, { "epoch": 61.74294562973159, "grad_norm": 15.543675422668457, "learning_rate": 1.9128527185134208e-05, "loss": 0.778, "step": 358850 }, { "epoch": 61.75154852030282, "grad_norm": 17.87061882019043, "learning_rate": 1.912422573984859e-05, "loss": 0.691, "step": 358900 }, { "epoch": 61.76015141087405, "grad_norm": 14.105626106262207, "learning_rate": 1.9119924294562974e-05, "loss": 0.6973, "step": 358950 }, { "epoch": 61.768754301445284, "grad_norm": 9.397385597229004, "learning_rate": 1.9115622849277358e-05, "loss": 0.712, "step": 359000 }, { "epoch": 61.77735719201652, "grad_norm": 12.812023162841797, "learning_rate": 1.9111321403991743e-05, "loss": 0.7944, "step": 359050 }, { "epoch": 61.78596008258775, "grad_norm": 10.492156982421875, "learning_rate": 1.9107019958706127e-05, "loss": 0.7193, "step": 359100 }, { "epoch": 61.79456297315898, "grad_norm": 17.055967330932617, "learning_rate": 1.910271851342051e-05, "loss": 0.673, "step": 359150 }, { "epoch": 61.803165863730214, "grad_norm": 11.883382797241211, "learning_rate": 1.9098417068134893e-05, "loss": 0.7346, "step": 359200 }, { "epoch": 61.81176875430145, "grad_norm": 10.989776611328125, "learning_rate": 1.9094115622849278e-05, "loss": 0.7104, "step": 359250 }, { "epoch": 61.82037164487268, "grad_norm": 14.157071113586426, "learning_rate": 1.9089814177563662e-05, "loss": 0.7374, "step": 359300 }, { "epoch": 61.82897453544391, "grad_norm": 12.062267303466797, "learning_rate": 1.9085512732278047e-05, "loss": 0.7262, "step": 359350 }, { "epoch": 61.837577426015145, "grad_norm": 13.763001441955566, "learning_rate": 1.908121128699243e-05, "loss": 0.771, "step": 359400 }, { "epoch": 61.84618031658637, "grad_norm": 16.05923843383789, "learning_rate": 1.9076909841706812e-05, "loss": 0.7383, "step": 359450 }, { "epoch": 61.8547832071576, "grad_norm": 18.08308982849121, "learning_rate": 1.90726083964212e-05, "loss": 0.7011, "step": 359500 }, { "epoch": 61.863386097728835, "grad_norm": 11.933406829833984, "learning_rate": 1.9068306951135585e-05, "loss": 0.719, "step": 359550 }, { "epoch": 61.87198898830007, "grad_norm": 11.327176094055176, "learning_rate": 1.9064005505849966e-05, "loss": 0.7629, "step": 359600 }, { "epoch": 61.8805918788713, "grad_norm": 20.631744384765625, "learning_rate": 1.905970406056435e-05, "loss": 0.7617, "step": 359650 }, { "epoch": 61.88919476944253, "grad_norm": 14.43688678741455, "learning_rate": 1.9055402615278732e-05, "loss": 0.7355, "step": 359700 }, { "epoch": 61.897797660013765, "grad_norm": 20.72100257873535, "learning_rate": 1.905110116999312e-05, "loss": 0.7244, "step": 359750 }, { "epoch": 61.906400550585, "grad_norm": 13.2079496383667, "learning_rate": 1.9046799724707504e-05, "loss": 0.7831, "step": 359800 }, { "epoch": 61.91500344115623, "grad_norm": 15.7333345413208, "learning_rate": 1.9042498279421886e-05, "loss": 0.7395, "step": 359850 }, { "epoch": 61.92360633172746, "grad_norm": 11.735152244567871, "learning_rate": 1.903819683413627e-05, "loss": 0.6981, "step": 359900 }, { "epoch": 61.932209222298695, "grad_norm": 17.377193450927734, "learning_rate": 1.9033895388850655e-05, "loss": 0.7571, "step": 359950 }, { "epoch": 61.94081211286992, "grad_norm": 23.62192726135254, "learning_rate": 1.902959394356504e-05, "loss": 0.7345, "step": 360000 }, { "epoch": 61.94941500344115, "grad_norm": 15.09108829498291, "learning_rate": 1.9025292498279424e-05, "loss": 0.7447, "step": 360050 }, { "epoch": 61.958017894012386, "grad_norm": 13.743229866027832, "learning_rate": 1.9020991052993805e-05, "loss": 0.7077, "step": 360100 }, { "epoch": 61.96662078458362, "grad_norm": 17.34562110900879, "learning_rate": 1.901668960770819e-05, "loss": 0.7599, "step": 360150 }, { "epoch": 61.97522367515485, "grad_norm": 20.640592575073242, "learning_rate": 1.9012388162422577e-05, "loss": 0.6944, "step": 360200 }, { "epoch": 61.98382656572608, "grad_norm": 7.847495079040527, "learning_rate": 1.900808671713696e-05, "loss": 0.7166, "step": 360250 }, { "epoch": 61.992429456297316, "grad_norm": 13.998196601867676, "learning_rate": 1.9003785271851343e-05, "loss": 0.7791, "step": 360300 }, { "epoch": 62.0, "eval_accuracy": 0.47960764068146616, "eval_f1": 0.4547314406124262, "eval_f1_DuraRiadoRio_16x16": 0.36040453570334047, "eval_f1_Mole_16x16": 0.40023859230539816, "eval_f1_Quebrado_16x16": 0.5862266857962697, "eval_f1_RiadoRio_16x16": 0.38937242327072835, "eval_f1_RioFechado_16x16": 0.5374149659863946, "eval_loss": 3.3121519088745117, "eval_precision": 0.5203891320687828, "eval_precision_DuraRiadoRio_16x16": 0.6131386861313869, "eval_precision_Mole_16x16": 0.6396568160152526, "eval_precision_Quebrado_16x16": 0.43784826403771965, "eval_precision_RiadoRio_16x16": 0.43950361944157185, "eval_precision_RioFechado_16x16": 0.47179827471798275, "eval_recall": 0.4813796166869471, "eval_recall_DuraRiadoRio_16x16": 0.2552083333333333, "eval_recall_Mole_16x16": 0.2912326388888889, "eval_recall_Quebrado_16x16": 0.88671875, "eval_recall_RiadoRio_16x16": 0.3495065789473684, "eval_recall_RioFechado_16x16": 0.6242317822651449, "eval_runtime": 46.7231, "eval_samples_per_second": 248.742, "eval_steps_per_second": 15.56, "step": 360344 }, { "epoch": 62.00103234686855, "grad_norm": 15.600974082946777, "learning_rate": 1.8999483826565728e-05, "loss": 0.6736, "step": 360350 }, { "epoch": 62.00963523743978, "grad_norm": 18.50875473022461, "learning_rate": 1.899518238128011e-05, "loss": 0.6974, "step": 360400 }, { "epoch": 62.01823812801101, "grad_norm": 10.830168724060059, "learning_rate": 1.8990880935994497e-05, "loss": 0.7214, "step": 360450 }, { "epoch": 62.026841018582246, "grad_norm": 20.397136688232422, "learning_rate": 1.8986579490708878e-05, "loss": 0.7109, "step": 360500 }, { "epoch": 62.03544390915348, "grad_norm": 10.34022331237793, "learning_rate": 1.8982278045423263e-05, "loss": 0.7693, "step": 360550 }, { "epoch": 62.044046799724704, "grad_norm": 9.848682403564453, "learning_rate": 1.8977976600137647e-05, "loss": 0.7083, "step": 360600 }, { "epoch": 62.05264969029594, "grad_norm": 9.21817684173584, "learning_rate": 1.8973675154852032e-05, "loss": 0.7512, "step": 360650 }, { "epoch": 62.06125258086717, "grad_norm": 10.423112869262695, "learning_rate": 1.8969373709566416e-05, "loss": 0.7043, "step": 360700 }, { "epoch": 62.0698554714384, "grad_norm": 16.38535499572754, "learning_rate": 1.89650722642808e-05, "loss": 0.7172, "step": 360750 }, { "epoch": 62.078458362009634, "grad_norm": 14.300573348999023, "learning_rate": 1.8960770818995182e-05, "loss": 0.7478, "step": 360800 }, { "epoch": 62.08706125258087, "grad_norm": 22.86810874938965, "learning_rate": 1.8956469373709567e-05, "loss": 0.7106, "step": 360850 }, { "epoch": 62.0956641431521, "grad_norm": 15.230693817138672, "learning_rate": 1.895216792842395e-05, "loss": 0.6917, "step": 360900 }, { "epoch": 62.10426703372333, "grad_norm": 14.648529052734375, "learning_rate": 1.8947866483138336e-05, "loss": 0.7564, "step": 360950 }, { "epoch": 62.112869924294564, "grad_norm": 13.377793312072754, "learning_rate": 1.894356503785272e-05, "loss": 0.6896, "step": 361000 }, { "epoch": 62.1214728148658, "grad_norm": 14.752655982971191, "learning_rate": 1.89392635925671e-05, "loss": 0.7103, "step": 361050 }, { "epoch": 62.13007570543703, "grad_norm": 12.133294105529785, "learning_rate": 1.8934962147281486e-05, "loss": 0.7822, "step": 361100 }, { "epoch": 62.13867859600826, "grad_norm": 18.181846618652344, "learning_rate": 1.8930660701995874e-05, "loss": 0.7102, "step": 361150 }, { "epoch": 62.14728148657949, "grad_norm": 17.212121963500977, "learning_rate": 1.8926359256710255e-05, "loss": 0.7008, "step": 361200 }, { "epoch": 62.15588437715072, "grad_norm": 14.678262710571289, "learning_rate": 1.892205781142464e-05, "loss": 0.7648, "step": 361250 }, { "epoch": 62.16448726772195, "grad_norm": 17.462095260620117, "learning_rate": 1.891775636613902e-05, "loss": 0.727, "step": 361300 }, { "epoch": 62.173090158293185, "grad_norm": 18.596202850341797, "learning_rate": 1.891345492085341e-05, "loss": 0.7706, "step": 361350 }, { "epoch": 62.18169304886442, "grad_norm": 12.999692916870117, "learning_rate": 1.8909153475567793e-05, "loss": 0.7129, "step": 361400 }, { "epoch": 62.19029593943565, "grad_norm": 18.617599487304688, "learning_rate": 1.8904852030282175e-05, "loss": 0.7281, "step": 361450 }, { "epoch": 62.19889883000688, "grad_norm": 9.731154441833496, "learning_rate": 1.890055058499656e-05, "loss": 0.7824, "step": 361500 }, { "epoch": 62.207501720578115, "grad_norm": 16.80521011352539, "learning_rate": 1.8896249139710944e-05, "loss": 0.6792, "step": 361550 }, { "epoch": 62.21610461114935, "grad_norm": 13.3337984085083, "learning_rate": 1.8891947694425328e-05, "loss": 0.7602, "step": 361600 }, { "epoch": 62.22470750172058, "grad_norm": 9.401312828063965, "learning_rate": 1.8887646249139713e-05, "loss": 0.7334, "step": 361650 }, { "epoch": 62.23331039229181, "grad_norm": 12.53073501586914, "learning_rate": 1.8883344803854097e-05, "loss": 0.7644, "step": 361700 }, { "epoch": 62.241913282863045, "grad_norm": 14.166015625, "learning_rate": 1.887904335856848e-05, "loss": 0.6728, "step": 361750 }, { "epoch": 62.25051617343427, "grad_norm": 9.555935859680176, "learning_rate": 1.8874741913282863e-05, "loss": 0.7302, "step": 361800 }, { "epoch": 62.2591190640055, "grad_norm": 11.082574844360352, "learning_rate": 1.8870440467997248e-05, "loss": 0.6967, "step": 361850 }, { "epoch": 62.267721954576736, "grad_norm": 9.433613777160645, "learning_rate": 1.8866139022711632e-05, "loss": 0.6936, "step": 361900 }, { "epoch": 62.27632484514797, "grad_norm": 25.124013900756836, "learning_rate": 1.8861837577426017e-05, "loss": 0.7675, "step": 361950 }, { "epoch": 62.2849277357192, "grad_norm": 16.30584716796875, "learning_rate": 1.8857536132140398e-05, "loss": 0.7282, "step": 362000 }, { "epoch": 62.29353062629043, "grad_norm": 17.91404914855957, "learning_rate": 1.8853234686854786e-05, "loss": 0.7072, "step": 362050 }, { "epoch": 62.302133516861666, "grad_norm": 16.413345336914062, "learning_rate": 1.884893324156917e-05, "loss": 0.6752, "step": 362100 }, { "epoch": 62.3107364074329, "grad_norm": 13.130006790161133, "learning_rate": 1.884463179628355e-05, "loss": 0.7332, "step": 362150 }, { "epoch": 62.31933929800413, "grad_norm": 12.094844818115234, "learning_rate": 1.8840330350997936e-05, "loss": 0.7124, "step": 362200 }, { "epoch": 62.32794218857536, "grad_norm": 15.984662055969238, "learning_rate": 1.883602890571232e-05, "loss": 0.7092, "step": 362250 }, { "epoch": 62.336545079146596, "grad_norm": 18.64299201965332, "learning_rate": 1.8831727460426705e-05, "loss": 0.7612, "step": 362300 }, { "epoch": 62.34514796971783, "grad_norm": 9.684000968933105, "learning_rate": 1.882742601514109e-05, "loss": 0.7283, "step": 362350 }, { "epoch": 62.353750860289054, "grad_norm": 5.833063125610352, "learning_rate": 1.882312456985547e-05, "loss": 0.7428, "step": 362400 }, { "epoch": 62.36235375086029, "grad_norm": 9.18907642364502, "learning_rate": 1.8818823124569856e-05, "loss": 0.7258, "step": 362450 }, { "epoch": 62.37095664143152, "grad_norm": 17.620006561279297, "learning_rate": 1.881452167928424e-05, "loss": 0.7111, "step": 362500 }, { "epoch": 62.37955953200275, "grad_norm": 12.53167724609375, "learning_rate": 1.8810220233998625e-05, "loss": 0.7241, "step": 362550 }, { "epoch": 62.388162422573984, "grad_norm": 14.494162559509277, "learning_rate": 1.880591878871301e-05, "loss": 0.6993, "step": 362600 }, { "epoch": 62.39676531314522, "grad_norm": 21.324260711669922, "learning_rate": 1.880161734342739e-05, "loss": 0.6716, "step": 362650 }, { "epoch": 62.40536820371645, "grad_norm": 10.554784774780273, "learning_rate": 1.8797315898141775e-05, "loss": 0.7196, "step": 362700 }, { "epoch": 62.41397109428768, "grad_norm": 7.809649467468262, "learning_rate": 1.8793014452856163e-05, "loss": 0.798, "step": 362750 }, { "epoch": 62.422573984858914, "grad_norm": 5.118525505065918, "learning_rate": 1.8788713007570544e-05, "loss": 0.6791, "step": 362800 }, { "epoch": 62.43117687543015, "grad_norm": 17.440319061279297, "learning_rate": 1.878441156228493e-05, "loss": 0.699, "step": 362850 }, { "epoch": 62.43977976600138, "grad_norm": 16.085311889648438, "learning_rate": 1.8780110116999313e-05, "loss": 0.7528, "step": 362900 }, { "epoch": 62.44838265657261, "grad_norm": 14.854063034057617, "learning_rate": 1.8775808671713694e-05, "loss": 0.6829, "step": 362950 }, { "epoch": 62.45698554714384, "grad_norm": 10.63318157196045, "learning_rate": 1.8771507226428082e-05, "loss": 0.7294, "step": 363000 }, { "epoch": 62.46558843771507, "grad_norm": 20.731542587280273, "learning_rate": 1.8767205781142464e-05, "loss": 0.7456, "step": 363050 }, { "epoch": 62.4741913282863, "grad_norm": 16.15457534790039, "learning_rate": 1.8762904335856848e-05, "loss": 0.7287, "step": 363100 }, { "epoch": 62.482794218857535, "grad_norm": 19.389617919921875, "learning_rate": 1.8758602890571233e-05, "loss": 0.6948, "step": 363150 }, { "epoch": 62.49139710942877, "grad_norm": 17.923789978027344, "learning_rate": 1.8754301445285617e-05, "loss": 0.7285, "step": 363200 }, { "epoch": 62.5, "grad_norm": 17.98126983642578, "learning_rate": 1.8750000000000002e-05, "loss": 0.7517, "step": 363250 }, { "epoch": 62.50860289057123, "grad_norm": 18.533376693725586, "learning_rate": 1.8745698554714386e-05, "loss": 0.7506, "step": 363300 }, { "epoch": 62.517205781142465, "grad_norm": 11.089354515075684, "learning_rate": 1.8741397109428767e-05, "loss": 0.7161, "step": 363350 }, { "epoch": 62.5258086717137, "grad_norm": 16.37259864807129, "learning_rate": 1.8737095664143152e-05, "loss": 0.7178, "step": 363400 }, { "epoch": 62.53441156228493, "grad_norm": 11.860180854797363, "learning_rate": 1.873279421885754e-05, "loss": 0.6952, "step": 363450 }, { "epoch": 62.54301445285616, "grad_norm": 20.178091049194336, "learning_rate": 1.872849277357192e-05, "loss": 0.7517, "step": 363500 }, { "epoch": 62.55161734342739, "grad_norm": 13.622063636779785, "learning_rate": 1.8724191328286306e-05, "loss": 0.7174, "step": 363550 }, { "epoch": 62.56022023399862, "grad_norm": 18.692840576171875, "learning_rate": 1.8719889883000687e-05, "loss": 0.759, "step": 363600 }, { "epoch": 62.56882312456985, "grad_norm": 11.423805236816406, "learning_rate": 1.871558843771507e-05, "loss": 0.7354, "step": 363650 }, { "epoch": 62.577426015141086, "grad_norm": 21.636646270751953, "learning_rate": 1.871128699242946e-05, "loss": 0.7297, "step": 363700 }, { "epoch": 62.58602890571232, "grad_norm": 17.259626388549805, "learning_rate": 1.870698554714384e-05, "loss": 0.6981, "step": 363750 }, { "epoch": 62.59463179628355, "grad_norm": 26.910123825073242, "learning_rate": 1.8702684101858225e-05, "loss": 0.7036, "step": 363800 }, { "epoch": 62.60323468685478, "grad_norm": 14.577966690063477, "learning_rate": 1.869838265657261e-05, "loss": 0.7117, "step": 363850 }, { "epoch": 62.611837577426016, "grad_norm": 15.041065216064453, "learning_rate": 1.8694081211286994e-05, "loss": 0.6976, "step": 363900 }, { "epoch": 62.62044046799725, "grad_norm": 15.330924987792969, "learning_rate": 1.868977976600138e-05, "loss": 0.7164, "step": 363950 }, { "epoch": 62.62904335856848, "grad_norm": 14.205082893371582, "learning_rate": 1.868547832071576e-05, "loss": 0.7487, "step": 364000 }, { "epoch": 62.63764624913971, "grad_norm": 9.728877067565918, "learning_rate": 1.8681176875430145e-05, "loss": 0.6978, "step": 364050 }, { "epoch": 62.646249139710946, "grad_norm": 14.648957252502441, "learning_rate": 1.867687543014453e-05, "loss": 0.7282, "step": 364100 }, { "epoch": 62.65485203028217, "grad_norm": 14.086647033691406, "learning_rate": 1.8672573984858914e-05, "loss": 0.7522, "step": 364150 }, { "epoch": 62.663454920853404, "grad_norm": 18.55470085144043, "learning_rate": 1.8668272539573298e-05, "loss": 0.7424, "step": 364200 }, { "epoch": 62.67205781142464, "grad_norm": 13.88919734954834, "learning_rate": 1.8663971094287683e-05, "loss": 0.7524, "step": 364250 }, { "epoch": 62.68066070199587, "grad_norm": 20.050857543945312, "learning_rate": 1.8659669649002064e-05, "loss": 0.8042, "step": 364300 }, { "epoch": 62.6892635925671, "grad_norm": 21.254667282104492, "learning_rate": 1.865536820371645e-05, "loss": 0.718, "step": 364350 }, { "epoch": 62.697866483138334, "grad_norm": 17.619436264038086, "learning_rate": 1.8651066758430833e-05, "loss": 0.7482, "step": 364400 }, { "epoch": 62.70646937370957, "grad_norm": 13.15510082244873, "learning_rate": 1.8646765313145218e-05, "loss": 0.7116, "step": 364450 }, { "epoch": 62.7150722642808, "grad_norm": 19.745328903198242, "learning_rate": 1.8642463867859602e-05, "loss": 0.7567, "step": 364500 }, { "epoch": 62.72367515485203, "grad_norm": 10.854877471923828, "learning_rate": 1.8638162422573983e-05, "loss": 0.7506, "step": 364550 }, { "epoch": 62.732278045423264, "grad_norm": 14.349150657653809, "learning_rate": 1.863386097728837e-05, "loss": 0.726, "step": 364600 }, { "epoch": 62.7408809359945, "grad_norm": 12.6108980178833, "learning_rate": 1.8629559532002756e-05, "loss": 0.7469, "step": 364650 }, { "epoch": 62.74948382656573, "grad_norm": 19.43683624267578, "learning_rate": 1.8625258086717137e-05, "loss": 0.7421, "step": 364700 }, { "epoch": 62.758086717136955, "grad_norm": 13.514820098876953, "learning_rate": 1.862095664143152e-05, "loss": 0.7789, "step": 364750 }, { "epoch": 62.76668960770819, "grad_norm": 24.299386978149414, "learning_rate": 1.8616655196145906e-05, "loss": 0.7253, "step": 364800 }, { "epoch": 62.77529249827942, "grad_norm": 7.286418914794922, "learning_rate": 1.861235375086029e-05, "loss": 0.6959, "step": 364850 }, { "epoch": 62.78389538885065, "grad_norm": 20.020713806152344, "learning_rate": 1.8608052305574675e-05, "loss": 0.7213, "step": 364900 }, { "epoch": 62.792498279421885, "grad_norm": 16.213245391845703, "learning_rate": 1.8603750860289056e-05, "loss": 0.7406, "step": 364950 }, { "epoch": 62.80110116999312, "grad_norm": 18.0719051361084, "learning_rate": 1.859944941500344e-05, "loss": 0.6962, "step": 365000 }, { "epoch": 62.80970406056435, "grad_norm": 12.155367851257324, "learning_rate": 1.8595147969717826e-05, "loss": 0.748, "step": 365050 }, { "epoch": 62.81830695113558, "grad_norm": 15.00395679473877, "learning_rate": 1.859084652443221e-05, "loss": 0.7291, "step": 365100 }, { "epoch": 62.826909841706815, "grad_norm": 13.949947357177734, "learning_rate": 1.8586545079146595e-05, "loss": 0.7598, "step": 365150 }, { "epoch": 62.83551273227805, "grad_norm": 19.817527770996094, "learning_rate": 1.858224363386098e-05, "loss": 0.6929, "step": 365200 }, { "epoch": 62.84411562284928, "grad_norm": 7.0252180099487305, "learning_rate": 1.857794218857536e-05, "loss": 0.7235, "step": 365250 }, { "epoch": 62.85271851342051, "grad_norm": 9.3187894821167, "learning_rate": 1.857364074328975e-05, "loss": 0.7747, "step": 365300 }, { "epoch": 62.86132140399174, "grad_norm": 19.74669075012207, "learning_rate": 1.856933929800413e-05, "loss": 0.7411, "step": 365350 }, { "epoch": 62.86992429456297, "grad_norm": 18.05797004699707, "learning_rate": 1.8565037852718514e-05, "loss": 0.7413, "step": 365400 }, { "epoch": 62.8785271851342, "grad_norm": 10.662177085876465, "learning_rate": 1.85607364074329e-05, "loss": 0.766, "step": 365450 }, { "epoch": 62.887130075705436, "grad_norm": 8.322937965393066, "learning_rate": 1.8556434962147283e-05, "loss": 0.734, "step": 365500 }, { "epoch": 62.89573296627667, "grad_norm": 9.973600387573242, "learning_rate": 1.8552133516861668e-05, "loss": 0.684, "step": 365550 }, { "epoch": 62.9043358568479, "grad_norm": 9.294499397277832, "learning_rate": 1.8547832071576052e-05, "loss": 0.725, "step": 365600 }, { "epoch": 62.91293874741913, "grad_norm": 18.71989631652832, "learning_rate": 1.8543530626290434e-05, "loss": 0.7622, "step": 365650 }, { "epoch": 62.921541637990366, "grad_norm": 22.609962463378906, "learning_rate": 1.8539229181004818e-05, "loss": 0.6864, "step": 365700 }, { "epoch": 62.9301445285616, "grad_norm": 21.35928726196289, "learning_rate": 1.8534927735719203e-05, "loss": 0.729, "step": 365750 }, { "epoch": 62.93874741913283, "grad_norm": 26.012483596801758, "learning_rate": 1.8530626290433587e-05, "loss": 0.6884, "step": 365800 }, { "epoch": 62.94735030970406, "grad_norm": 9.013676643371582, "learning_rate": 1.8526324845147972e-05, "loss": 0.7491, "step": 365850 }, { "epoch": 62.955953200275296, "grad_norm": 10.822497367858887, "learning_rate": 1.8522023399862353e-05, "loss": 0.6882, "step": 365900 }, { "epoch": 62.96455609084652, "grad_norm": 15.55201530456543, "learning_rate": 1.8517721954576738e-05, "loss": 0.7354, "step": 365950 }, { "epoch": 62.973158981417754, "grad_norm": 18.467464447021484, "learning_rate": 1.8513420509291125e-05, "loss": 0.7331, "step": 366000 }, { "epoch": 62.98176187198899, "grad_norm": 9.962837219238281, "learning_rate": 1.8509119064005507e-05, "loss": 0.7163, "step": 366050 }, { "epoch": 62.99036476256022, "grad_norm": 11.276124000549316, "learning_rate": 1.850481761871989e-05, "loss": 0.7227, "step": 366100 }, { "epoch": 62.99896765313145, "grad_norm": 20.547611236572266, "learning_rate": 1.8500516173434272e-05, "loss": 0.7385, "step": 366150 }, { "epoch": 63.0, "eval_accuracy": 0.5316640853553605, "eval_f1": 0.5203708619386024, "eval_f1_DuraRiadoRio_16x16": 0.5235438518187905, "eval_f1_Mole_16x16": 0.5810679611650486, "eval_f1_Quebrado_16x16": 0.6252637925836599, "eval_f1_RiadoRio_16x16": 0.389622641509434, "eval_f1_RioFechado_16x16": 0.48235606261607855, "eval_loss": 2.7403178215026855, "eval_precision": 0.5486912081026666, "eval_precision_DuraRiadoRio_16x16": 0.5388148828663298, "eval_precision_Mole_16x16": 0.6591409691629956, "eval_precision_Quebrado_16x16": 0.4789838337182448, "eval_precision_RiadoRio_16x16": 0.45685840707964603, "eval_precision_RioFechado_16x16": 0.6096579476861167, "eval_recall": 0.5334983685802156, "eval_recall_DuraRiadoRio_16x16": 0.5091145833333334, "eval_recall_Mole_16x16": 0.51953125, "eval_recall_Quebrado_16x16": 0.9001736111111112, "eval_recall_RiadoRio_16x16": 0.33963815789473684, "eval_recall_RioFechado_16x16": 0.3990342405618964, "eval_runtime": 46.545, "eval_samples_per_second": 249.694, "eval_steps_per_second": 15.619, "step": 366156 }, { "epoch": 63.007570543702684, "grad_norm": 17.26058578491211, "learning_rate": 1.849621472814866e-05, "loss": 0.6846, "step": 366200 }, { "epoch": 63.01617343427392, "grad_norm": 16.047653198242188, "learning_rate": 1.8491913282863045e-05, "loss": 0.7263, "step": 366250 }, { "epoch": 63.02477632484515, "grad_norm": 12.007616996765137, "learning_rate": 1.8487611837577426e-05, "loss": 0.7335, "step": 366300 }, { "epoch": 63.03337921541638, "grad_norm": 13.5363130569458, "learning_rate": 1.848331039229181e-05, "loss": 0.7377, "step": 366350 }, { "epoch": 63.041982105987614, "grad_norm": 23.76610565185547, "learning_rate": 1.8479008947006195e-05, "loss": 0.7173, "step": 366400 }, { "epoch": 63.05058499655885, "grad_norm": 9.59937572479248, "learning_rate": 1.847470750172058e-05, "loss": 0.7253, "step": 366450 }, { "epoch": 63.05918788713007, "grad_norm": 14.394841194152832, "learning_rate": 1.8470406056434964e-05, "loss": 0.7771, "step": 366500 }, { "epoch": 63.067790777701305, "grad_norm": 18.3297176361084, "learning_rate": 1.8466104611149345e-05, "loss": 0.7028, "step": 366550 }, { "epoch": 63.07639366827254, "grad_norm": 7.797086238861084, "learning_rate": 1.846180316586373e-05, "loss": 0.7113, "step": 366600 }, { "epoch": 63.08499655884377, "grad_norm": 17.71658706665039, "learning_rate": 1.8457501720578115e-05, "loss": 0.7449, "step": 366650 }, { "epoch": 63.093599449415, "grad_norm": 21.63690757751465, "learning_rate": 1.84532002752925e-05, "loss": 0.7306, "step": 366700 }, { "epoch": 63.102202339986235, "grad_norm": 14.844738006591797, "learning_rate": 1.8448898830006884e-05, "loss": 0.758, "step": 366750 }, { "epoch": 63.11080523055747, "grad_norm": 16.896865844726562, "learning_rate": 1.8444597384721268e-05, "loss": 0.7562, "step": 366800 }, { "epoch": 63.1194081211287, "grad_norm": 19.034496307373047, "learning_rate": 1.844029593943565e-05, "loss": 0.7129, "step": 366850 }, { "epoch": 63.12801101169993, "grad_norm": 15.663886070251465, "learning_rate": 1.8435994494150034e-05, "loss": 0.6792, "step": 366900 }, { "epoch": 63.136613902271165, "grad_norm": 9.229565620422363, "learning_rate": 1.843169304886442e-05, "loss": 0.7636, "step": 366950 }, { "epoch": 63.1452167928424, "grad_norm": 12.73244571685791, "learning_rate": 1.8427391603578803e-05, "loss": 0.6994, "step": 367000 }, { "epoch": 63.15381968341363, "grad_norm": 17.009422302246094, "learning_rate": 1.8423090158293188e-05, "loss": 0.72, "step": 367050 }, { "epoch": 63.162422573984855, "grad_norm": 15.964129447937012, "learning_rate": 1.841878871300757e-05, "loss": 0.7198, "step": 367100 }, { "epoch": 63.17102546455609, "grad_norm": 11.574772834777832, "learning_rate": 1.8414487267721957e-05, "loss": 0.6634, "step": 367150 }, { "epoch": 63.17962835512732, "grad_norm": 11.441336631774902, "learning_rate": 1.841018582243634e-05, "loss": 0.7165, "step": 367200 }, { "epoch": 63.18823124569855, "grad_norm": 13.751158714294434, "learning_rate": 1.8405884377150723e-05, "loss": 0.7586, "step": 367250 }, { "epoch": 63.196834136269786, "grad_norm": 20.13347625732422, "learning_rate": 1.8401582931865107e-05, "loss": 0.7119, "step": 367300 }, { "epoch": 63.20543702684102, "grad_norm": 8.140819549560547, "learning_rate": 1.839728148657949e-05, "loss": 0.7156, "step": 367350 }, { "epoch": 63.21403991741225, "grad_norm": 14.720640182495117, "learning_rate": 1.8392980041293876e-05, "loss": 0.7383, "step": 367400 }, { "epoch": 63.22264280798348, "grad_norm": 17.70050621032715, "learning_rate": 1.838867859600826e-05, "loss": 0.677, "step": 367450 }, { "epoch": 63.231245698554716, "grad_norm": 24.7393856048584, "learning_rate": 1.8384377150722642e-05, "loss": 0.7526, "step": 367500 }, { "epoch": 63.23984858912595, "grad_norm": 18.187828063964844, "learning_rate": 1.8380075705437027e-05, "loss": 0.767, "step": 367550 }, { "epoch": 63.24845147969718, "grad_norm": 11.557548522949219, "learning_rate": 1.837577426015141e-05, "loss": 0.7529, "step": 367600 }, { "epoch": 63.25705437026841, "grad_norm": 8.983370780944824, "learning_rate": 1.8371472814865796e-05, "loss": 0.6958, "step": 367650 }, { "epoch": 63.26565726083964, "grad_norm": 12.822332382202148, "learning_rate": 1.836717136958018e-05, "loss": 0.6483, "step": 367700 }, { "epoch": 63.27426015141087, "grad_norm": 15.450413703918457, "learning_rate": 1.8362869924294565e-05, "loss": 0.752, "step": 367750 }, { "epoch": 63.282863041982104, "grad_norm": 20.746814727783203, "learning_rate": 1.8358568479008946e-05, "loss": 0.7419, "step": 367800 }, { "epoch": 63.291465932553336, "grad_norm": 14.039488792419434, "learning_rate": 1.8354267033723334e-05, "loss": 0.7363, "step": 367850 }, { "epoch": 63.30006882312457, "grad_norm": 9.140575408935547, "learning_rate": 1.8349965588437715e-05, "loss": 0.7346, "step": 367900 }, { "epoch": 63.3086717136958, "grad_norm": 15.350196838378906, "learning_rate": 1.83456641431521e-05, "loss": 0.7513, "step": 367950 }, { "epoch": 63.317274604267034, "grad_norm": 6.707966327667236, "learning_rate": 1.8341362697866484e-05, "loss": 0.7117, "step": 368000 }, { "epoch": 63.32587749483827, "grad_norm": 11.084277153015137, "learning_rate": 1.833706125258087e-05, "loss": 0.7188, "step": 368050 }, { "epoch": 63.3344803854095, "grad_norm": 14.983917236328125, "learning_rate": 1.8332759807295253e-05, "loss": 0.7206, "step": 368100 }, { "epoch": 63.34308327598073, "grad_norm": 17.488611221313477, "learning_rate": 1.8328458362009638e-05, "loss": 0.6549, "step": 368150 }, { "epoch": 63.351686166551964, "grad_norm": 19.148914337158203, "learning_rate": 1.832415691672402e-05, "loss": 0.7537, "step": 368200 }, { "epoch": 63.3602890571232, "grad_norm": 16.58632469177246, "learning_rate": 1.8319855471438404e-05, "loss": 0.7448, "step": 368250 }, { "epoch": 63.36889194769442, "grad_norm": 17.693225860595703, "learning_rate": 1.8315554026152788e-05, "loss": 0.7178, "step": 368300 }, { "epoch": 63.377494838265655, "grad_norm": 15.079618453979492, "learning_rate": 1.8311252580867173e-05, "loss": 0.6771, "step": 368350 }, { "epoch": 63.38609772883689, "grad_norm": 16.61878204345703, "learning_rate": 1.8306951135581557e-05, "loss": 0.7329, "step": 368400 }, { "epoch": 63.39470061940812, "grad_norm": 15.845526695251465, "learning_rate": 1.830264969029594e-05, "loss": 0.7351, "step": 368450 }, { "epoch": 63.40330350997935, "grad_norm": 16.294246673583984, "learning_rate": 1.8298348245010323e-05, "loss": 0.6766, "step": 368500 }, { "epoch": 63.411906400550585, "grad_norm": 11.007339477539062, "learning_rate": 1.829404679972471e-05, "loss": 0.6872, "step": 368550 }, { "epoch": 63.42050929112182, "grad_norm": 19.09088706970215, "learning_rate": 1.8289745354439092e-05, "loss": 0.7402, "step": 368600 }, { "epoch": 63.42911218169305, "grad_norm": 22.66972541809082, "learning_rate": 1.8285443909153477e-05, "loss": 0.7727, "step": 368650 }, { "epoch": 63.43771507226428, "grad_norm": 10.54520034790039, "learning_rate": 1.8281142463867858e-05, "loss": 0.7268, "step": 368700 }, { "epoch": 63.446317962835515, "grad_norm": 7.8340067863464355, "learning_rate": 1.8276841018582246e-05, "loss": 0.7286, "step": 368750 }, { "epoch": 63.45492085340675, "grad_norm": 9.142826080322266, "learning_rate": 1.827253957329663e-05, "loss": 0.7226, "step": 368800 }, { "epoch": 63.46352374397798, "grad_norm": 6.9041547775268555, "learning_rate": 1.826823812801101e-05, "loss": 0.7508, "step": 368850 }, { "epoch": 63.472126634549205, "grad_norm": 9.040902137756348, "learning_rate": 1.8263936682725396e-05, "loss": 0.7659, "step": 368900 }, { "epoch": 63.48072952512044, "grad_norm": 20.574617385864258, "learning_rate": 1.825963523743978e-05, "loss": 0.77, "step": 368950 }, { "epoch": 63.48933241569167, "grad_norm": 14.477428436279297, "learning_rate": 1.8255333792154165e-05, "loss": 0.7354, "step": 369000 }, { "epoch": 63.4979353062629, "grad_norm": 14.624000549316406, "learning_rate": 1.825103234686855e-05, "loss": 0.7271, "step": 369050 }, { "epoch": 63.506538196834136, "grad_norm": 27.826488494873047, "learning_rate": 1.8246730901582934e-05, "loss": 0.7145, "step": 369100 }, { "epoch": 63.51514108740537, "grad_norm": 15.29755687713623, "learning_rate": 1.8242429456297315e-05, "loss": 0.7064, "step": 369150 }, { "epoch": 63.5237439779766, "grad_norm": 19.99287223815918, "learning_rate": 1.82381280110117e-05, "loss": 0.6831, "step": 369200 }, { "epoch": 63.53234686854783, "grad_norm": 9.293439865112305, "learning_rate": 1.8233826565726085e-05, "loss": 0.7178, "step": 369250 }, { "epoch": 63.540949759119066, "grad_norm": 16.279788970947266, "learning_rate": 1.822952512044047e-05, "loss": 0.6742, "step": 369300 }, { "epoch": 63.5495526496903, "grad_norm": 11.097611427307129, "learning_rate": 1.8225223675154854e-05, "loss": 0.7997, "step": 369350 }, { "epoch": 63.55815554026153, "grad_norm": 7.84420108795166, "learning_rate": 1.8220922229869235e-05, "loss": 0.7502, "step": 369400 }, { "epoch": 63.56675843083276, "grad_norm": 12.315799713134766, "learning_rate": 1.8216620784583623e-05, "loss": 0.7459, "step": 369450 }, { "epoch": 63.57536132140399, "grad_norm": 11.156281471252441, "learning_rate": 1.8212319339298007e-05, "loss": 0.7447, "step": 369500 }, { "epoch": 63.58396421197522, "grad_norm": 21.569856643676758, "learning_rate": 1.820801789401239e-05, "loss": 0.7338, "step": 369550 }, { "epoch": 63.592567102546454, "grad_norm": 16.956249237060547, "learning_rate": 1.8203716448726773e-05, "loss": 0.6812, "step": 369600 }, { "epoch": 63.601169993117686, "grad_norm": 23.583106994628906, "learning_rate": 1.8199415003441154e-05, "loss": 0.7173, "step": 369650 }, { "epoch": 63.60977288368892, "grad_norm": 24.42569351196289, "learning_rate": 1.8195113558155542e-05, "loss": 0.7001, "step": 369700 }, { "epoch": 63.61837577426015, "grad_norm": 15.415786743164062, "learning_rate": 1.8190812112869927e-05, "loss": 0.6805, "step": 369750 }, { "epoch": 63.626978664831384, "grad_norm": 17.971065521240234, "learning_rate": 1.8186510667584308e-05, "loss": 0.7249, "step": 369800 }, { "epoch": 63.63558155540262, "grad_norm": 13.886966705322266, "learning_rate": 1.8182209222298693e-05, "loss": 0.7055, "step": 369850 }, { "epoch": 63.64418444597385, "grad_norm": 13.623786926269531, "learning_rate": 1.8177907777013077e-05, "loss": 0.744, "step": 369900 }, { "epoch": 63.65278733654508, "grad_norm": 5.414880275726318, "learning_rate": 1.817360633172746e-05, "loss": 0.7445, "step": 369950 }, { "epoch": 63.661390227116314, "grad_norm": 10.857756614685059, "learning_rate": 1.8169304886441846e-05, "loss": 0.6752, "step": 370000 }, { "epoch": 63.66999311768754, "grad_norm": 10.778789520263672, "learning_rate": 1.8165003441156227e-05, "loss": 0.7332, "step": 370050 }, { "epoch": 63.67859600825877, "grad_norm": 9.413531303405762, "learning_rate": 1.8160701995870612e-05, "loss": 0.7419, "step": 370100 }, { "epoch": 63.687198898830005, "grad_norm": 10.162796020507812, "learning_rate": 1.8156400550584997e-05, "loss": 0.757, "step": 370150 }, { "epoch": 63.69580178940124, "grad_norm": 15.950993537902832, "learning_rate": 1.815209910529938e-05, "loss": 0.8082, "step": 370200 }, { "epoch": 63.70440467997247, "grad_norm": 12.064918518066406, "learning_rate": 1.8147797660013766e-05, "loss": 0.751, "step": 370250 }, { "epoch": 63.7130075705437, "grad_norm": 9.855006217956543, "learning_rate": 1.814349621472815e-05, "loss": 0.7537, "step": 370300 }, { "epoch": 63.721610461114935, "grad_norm": 13.42127513885498, "learning_rate": 1.813919476944253e-05, "loss": 0.697, "step": 370350 }, { "epoch": 63.73021335168617, "grad_norm": 11.19174861907959, "learning_rate": 1.813489332415692e-05, "loss": 0.7301, "step": 370400 }, { "epoch": 63.7388162422574, "grad_norm": 12.909928321838379, "learning_rate": 1.81305918788713e-05, "loss": 0.7202, "step": 370450 }, { "epoch": 63.74741913282863, "grad_norm": 15.423863410949707, "learning_rate": 1.8126290433585685e-05, "loss": 0.7393, "step": 370500 }, { "epoch": 63.756022023399865, "grad_norm": 9.753313064575195, "learning_rate": 1.812198898830007e-05, "loss": 0.7606, "step": 370550 }, { "epoch": 63.7646249139711, "grad_norm": 10.717524528503418, "learning_rate": 1.8117687543014454e-05, "loss": 0.6783, "step": 370600 }, { "epoch": 63.77322780454232, "grad_norm": 11.488759994506836, "learning_rate": 1.811338609772884e-05, "loss": 0.6854, "step": 370650 }, { "epoch": 63.781830695113555, "grad_norm": 12.856109619140625, "learning_rate": 1.8109084652443223e-05, "loss": 0.7048, "step": 370700 }, { "epoch": 63.79043358568479, "grad_norm": 9.581542015075684, "learning_rate": 1.8104783207157604e-05, "loss": 0.7414, "step": 370750 }, { "epoch": 63.79903647625602, "grad_norm": 12.045604705810547, "learning_rate": 1.810048176187199e-05, "loss": 0.7522, "step": 370800 }, { "epoch": 63.80763936682725, "grad_norm": 7.2233195304870605, "learning_rate": 1.8096180316586374e-05, "loss": 0.7567, "step": 370850 }, { "epoch": 63.816242257398486, "grad_norm": 17.206769943237305, "learning_rate": 1.8091878871300758e-05, "loss": 0.7061, "step": 370900 }, { "epoch": 63.82484514796972, "grad_norm": 16.689992904663086, "learning_rate": 1.8087577426015143e-05, "loss": 0.7392, "step": 370950 }, { "epoch": 63.83344803854095, "grad_norm": 11.93142318725586, "learning_rate": 1.8083275980729524e-05, "loss": 0.7654, "step": 371000 }, { "epoch": 63.84205092911218, "grad_norm": 27.979394912719727, "learning_rate": 1.807897453544391e-05, "loss": 0.7188, "step": 371050 }, { "epoch": 63.850653819683416, "grad_norm": 11.829843521118164, "learning_rate": 1.8074673090158296e-05, "loss": 0.7568, "step": 371100 }, { "epoch": 63.85925671025465, "grad_norm": 17.377460479736328, "learning_rate": 1.8070371644872678e-05, "loss": 0.7729, "step": 371150 }, { "epoch": 63.86785960082588, "grad_norm": 18.1741886138916, "learning_rate": 1.8066070199587062e-05, "loss": 0.7295, "step": 371200 }, { "epoch": 63.876462491397106, "grad_norm": 12.696059226989746, "learning_rate": 1.8061768754301447e-05, "loss": 0.7269, "step": 371250 }, { "epoch": 63.88506538196834, "grad_norm": 12.039657592773438, "learning_rate": 1.805746730901583e-05, "loss": 0.7931, "step": 371300 }, { "epoch": 63.89366827253957, "grad_norm": 18.159244537353516, "learning_rate": 1.8053165863730216e-05, "loss": 0.7334, "step": 371350 }, { "epoch": 63.902271163110804, "grad_norm": 19.1779842376709, "learning_rate": 1.8048864418444597e-05, "loss": 0.7149, "step": 371400 }, { "epoch": 63.910874053682036, "grad_norm": 15.18963623046875, "learning_rate": 1.804456297315898e-05, "loss": 0.7116, "step": 371450 }, { "epoch": 63.91947694425327, "grad_norm": 13.487147331237793, "learning_rate": 1.8040261527873366e-05, "loss": 0.6905, "step": 371500 }, { "epoch": 63.9280798348245, "grad_norm": 15.913947105407715, "learning_rate": 1.803596008258775e-05, "loss": 0.707, "step": 371550 }, { "epoch": 63.936682725395734, "grad_norm": 10.133556365966797, "learning_rate": 1.8031658637302135e-05, "loss": 0.7088, "step": 371600 }, { "epoch": 63.94528561596697, "grad_norm": 17.70921516418457, "learning_rate": 1.802735719201652e-05, "loss": 0.7312, "step": 371650 }, { "epoch": 63.9538885065382, "grad_norm": 12.146265029907227, "learning_rate": 1.80230557467309e-05, "loss": 0.6897, "step": 371700 }, { "epoch": 63.96249139710943, "grad_norm": 5.93372106552124, "learning_rate": 1.8018754301445286e-05, "loss": 0.725, "step": 371750 }, { "epoch": 63.971094287680664, "grad_norm": 12.946023941040039, "learning_rate": 1.801445285615967e-05, "loss": 0.703, "step": 371800 }, { "epoch": 63.97969717825189, "grad_norm": 16.922740936279297, "learning_rate": 1.8010151410874055e-05, "loss": 0.7444, "step": 371850 }, { "epoch": 63.98830006882312, "grad_norm": 13.572148323059082, "learning_rate": 1.800584996558844e-05, "loss": 0.7455, "step": 371900 }, { "epoch": 63.996902959394355, "grad_norm": 8.717278480529785, "learning_rate": 1.800154852030282e-05, "loss": 0.6977, "step": 371950 }, { "epoch": 64.0, "eval_accuracy": 0.45465496472207884, "eval_f1": 0.4240639661320576, "eval_f1_DuraRiadoRio_16x16": 0.28313856427378964, "eval_f1_Mole_16x16": 0.3788316946211683, "eval_f1_Quebrado_16x16": 0.662948503439301, "eval_f1_RiadoRio_16x16": 0.3161189358372457, "eval_f1_RioFechado_16x16": 0.4792821324887833, "eval_loss": 3.846553087234497, "eval_precision": 0.5071826053070783, "eval_precision_DuraRiadoRio_16x16": 0.613603473227207, "eval_precision_Mole_16x16": 0.5675909878682842, "eval_precision_Quebrado_16x16": 0.5798373983739837, "eval_precision_RiadoRio_16x16": 0.4322396576319544, "eval_precision_RioFechado_16x16": 0.34264150943396227, "eval_recall": 0.4577111299154383, "eval_recall_DuraRiadoRio_16x16": 0.1840277777777778, "eval_recall_Mole_16x16": 0.2842881944444444, "eval_recall_Quebrado_16x16": 0.7738715277777778, "eval_recall_RiadoRio_16x16": 0.24917763157894737, "eval_recall_RioFechado_16x16": 0.797190517998244, "eval_runtime": 46.1767, "eval_samples_per_second": 251.685, "eval_steps_per_second": 15.744, "step": 371968 }, { "epoch": 64.00550584996559, "grad_norm": 13.894721984863281, "learning_rate": 1.799724707501721e-05, "loss": 0.7866, "step": 372000 }, { "epoch": 64.01410874053683, "grad_norm": 11.306941032409668, "learning_rate": 1.7992945629731593e-05, "loss": 0.6893, "step": 372050 }, { "epoch": 64.02271163110805, "grad_norm": 11.017967224121094, "learning_rate": 1.7988644184445974e-05, "loss": 0.7034, "step": 372100 }, { "epoch": 64.03131452167928, "grad_norm": 15.462173461914062, "learning_rate": 1.798434273916036e-05, "loss": 0.6662, "step": 372150 }, { "epoch": 64.03991741225052, "grad_norm": 11.715109825134277, "learning_rate": 1.798004129387474e-05, "loss": 0.7392, "step": 372200 }, { "epoch": 64.04852030282174, "grad_norm": 14.740999221801758, "learning_rate": 1.7975739848589128e-05, "loss": 0.6572, "step": 372250 }, { "epoch": 64.05712319339298, "grad_norm": 14.527196884155273, "learning_rate": 1.7971438403303512e-05, "loss": 0.6699, "step": 372300 }, { "epoch": 64.06572608396421, "grad_norm": 12.955673217773438, "learning_rate": 1.7967136958017893e-05, "loss": 0.7581, "step": 372350 }, { "epoch": 64.07432897453545, "grad_norm": 15.469982147216797, "learning_rate": 1.7962835512732278e-05, "loss": 0.7256, "step": 372400 }, { "epoch": 64.08293186510667, "grad_norm": 20.90330696105957, "learning_rate": 1.7958534067446663e-05, "loss": 0.6911, "step": 372450 }, { "epoch": 64.09153475567791, "grad_norm": 9.26349925994873, "learning_rate": 1.7954232622161047e-05, "loss": 0.7586, "step": 372500 }, { "epoch": 64.10013764624914, "grad_norm": 15.428034782409668, "learning_rate": 1.7949931176875432e-05, "loss": 0.723, "step": 372550 }, { "epoch": 64.10874053682038, "grad_norm": 10.936187744140625, "learning_rate": 1.7945629731589816e-05, "loss": 0.7014, "step": 372600 }, { "epoch": 64.1173434273916, "grad_norm": 17.9011173248291, "learning_rate": 1.7941328286304197e-05, "loss": 0.7258, "step": 372650 }, { "epoch": 64.12594631796284, "grad_norm": 26.812969207763672, "learning_rate": 1.7937026841018585e-05, "loss": 0.7078, "step": 372700 }, { "epoch": 64.13454920853407, "grad_norm": 20.18167495727539, "learning_rate": 1.7932725395732967e-05, "loss": 0.6423, "step": 372750 }, { "epoch": 64.1431520991053, "grad_norm": 15.162300109863281, "learning_rate": 1.792842395044735e-05, "loss": 0.711, "step": 372800 }, { "epoch": 64.15175498967653, "grad_norm": 8.964308738708496, "learning_rate": 1.7924122505161736e-05, "loss": 0.7579, "step": 372850 }, { "epoch": 64.16035788024776, "grad_norm": 9.201009750366211, "learning_rate": 1.7919821059876117e-05, "loss": 0.7767, "step": 372900 }, { "epoch": 64.168960770819, "grad_norm": 15.991494178771973, "learning_rate": 1.7915519614590505e-05, "loss": 0.707, "step": 372950 }, { "epoch": 64.17756366139022, "grad_norm": 12.72122859954834, "learning_rate": 1.791121816930489e-05, "loss": 0.7801, "step": 373000 }, { "epoch": 64.18616655196146, "grad_norm": 13.08949089050293, "learning_rate": 1.790691672401927e-05, "loss": 0.6703, "step": 373050 }, { "epoch": 64.19476944253269, "grad_norm": 17.094200134277344, "learning_rate": 1.7902615278733655e-05, "loss": 0.7087, "step": 373100 }, { "epoch": 64.20337233310393, "grad_norm": 10.169160842895508, "learning_rate": 1.789831383344804e-05, "loss": 0.7239, "step": 373150 }, { "epoch": 64.21197522367515, "grad_norm": 19.599828720092773, "learning_rate": 1.7894012388162424e-05, "loss": 0.6908, "step": 373200 }, { "epoch": 64.2205781142464, "grad_norm": 12.76595401763916, "learning_rate": 1.788971094287681e-05, "loss": 0.7288, "step": 373250 }, { "epoch": 64.22918100481762, "grad_norm": 15.221271514892578, "learning_rate": 1.788540949759119e-05, "loss": 0.6742, "step": 373300 }, { "epoch": 64.23778389538884, "grad_norm": 25.862136840820312, "learning_rate": 1.7881108052305575e-05, "loss": 0.7972, "step": 373350 }, { "epoch": 64.24638678596008, "grad_norm": 12.387516975402832, "learning_rate": 1.7876806607019962e-05, "loss": 0.7076, "step": 373400 }, { "epoch": 64.25498967653131, "grad_norm": 8.496800422668457, "learning_rate": 1.7872505161734344e-05, "loss": 0.7097, "step": 373450 }, { "epoch": 64.26359256710255, "grad_norm": 16.24919891357422, "learning_rate": 1.7868203716448728e-05, "loss": 0.6879, "step": 373500 }, { "epoch": 64.27219545767377, "grad_norm": 9.512531280517578, "learning_rate": 1.786390227116311e-05, "loss": 0.7349, "step": 373550 }, { "epoch": 64.28079834824501, "grad_norm": 11.128376007080078, "learning_rate": 1.7859600825877494e-05, "loss": 0.6878, "step": 373600 }, { "epoch": 64.28940123881624, "grad_norm": 8.452259063720703, "learning_rate": 1.7855299380591882e-05, "loss": 0.7203, "step": 373650 }, { "epoch": 64.29800412938748, "grad_norm": 14.892341613769531, "learning_rate": 1.7850997935306263e-05, "loss": 0.7594, "step": 373700 }, { "epoch": 64.3066070199587, "grad_norm": 8.105586051940918, "learning_rate": 1.7846696490020648e-05, "loss": 0.7168, "step": 373750 }, { "epoch": 64.31520991052994, "grad_norm": 8.191200256347656, "learning_rate": 1.7842395044735032e-05, "loss": 0.71, "step": 373800 }, { "epoch": 64.32381280110117, "grad_norm": 12.815714836120605, "learning_rate": 1.7838093599449417e-05, "loss": 0.7024, "step": 373850 }, { "epoch": 64.3324156916724, "grad_norm": 20.31106185913086, "learning_rate": 1.78337921541638e-05, "loss": 0.7106, "step": 373900 }, { "epoch": 64.34101858224363, "grad_norm": 28.099369049072266, "learning_rate": 1.7829490708878182e-05, "loss": 0.7612, "step": 373950 }, { "epoch": 64.34962147281486, "grad_norm": 13.582399368286133, "learning_rate": 1.7825189263592567e-05, "loss": 0.7144, "step": 374000 }, { "epoch": 64.3582243633861, "grad_norm": 18.615665435791016, "learning_rate": 1.782088781830695e-05, "loss": 0.7527, "step": 374050 }, { "epoch": 64.36682725395733, "grad_norm": 17.70020294189453, "learning_rate": 1.7816586373021336e-05, "loss": 0.7252, "step": 374100 }, { "epoch": 64.37543014452856, "grad_norm": 24.354814529418945, "learning_rate": 1.781228492773572e-05, "loss": 0.7074, "step": 374150 }, { "epoch": 64.38403303509979, "grad_norm": 14.254433631896973, "learning_rate": 1.7807983482450105e-05, "loss": 0.7254, "step": 374200 }, { "epoch": 64.39263592567103, "grad_norm": 5.215162754058838, "learning_rate": 1.7803682037164486e-05, "loss": 0.7093, "step": 374250 }, { "epoch": 64.40123881624226, "grad_norm": 15.734113693237305, "learning_rate": 1.779938059187887e-05, "loss": 0.7269, "step": 374300 }, { "epoch": 64.4098417068135, "grad_norm": 12.412775039672852, "learning_rate": 1.7795079146593256e-05, "loss": 0.7199, "step": 374350 }, { "epoch": 64.41844459738472, "grad_norm": 20.200899124145508, "learning_rate": 1.779077770130764e-05, "loss": 0.7138, "step": 374400 }, { "epoch": 64.42704748795596, "grad_norm": 14.354203224182129, "learning_rate": 1.7786476256022025e-05, "loss": 0.716, "step": 374450 }, { "epoch": 64.43565037852719, "grad_norm": 16.2321834564209, "learning_rate": 1.7782174810736406e-05, "loss": 0.7188, "step": 374500 }, { "epoch": 64.44425326909841, "grad_norm": 19.462890625, "learning_rate": 1.7777873365450794e-05, "loss": 0.7343, "step": 374550 }, { "epoch": 64.45285615966965, "grad_norm": 16.397798538208008, "learning_rate": 1.777357192016518e-05, "loss": 0.738, "step": 374600 }, { "epoch": 64.46145905024088, "grad_norm": 18.21648597717285, "learning_rate": 1.776927047487956e-05, "loss": 0.6804, "step": 374650 }, { "epoch": 64.47006194081212, "grad_norm": 16.191455841064453, "learning_rate": 1.7764969029593944e-05, "loss": 0.7268, "step": 374700 }, { "epoch": 64.47866483138334, "grad_norm": 18.69513511657715, "learning_rate": 1.776066758430833e-05, "loss": 0.7443, "step": 374750 }, { "epoch": 64.48726772195458, "grad_norm": 10.469165802001953, "learning_rate": 1.7756366139022713e-05, "loss": 0.6695, "step": 374800 }, { "epoch": 64.4958706125258, "grad_norm": 13.721624374389648, "learning_rate": 1.7752064693737098e-05, "loss": 0.6903, "step": 374850 }, { "epoch": 64.50447350309705, "grad_norm": 14.738507270812988, "learning_rate": 1.774776324845148e-05, "loss": 0.749, "step": 374900 }, { "epoch": 64.51307639366827, "grad_norm": 12.977699279785156, "learning_rate": 1.7743461803165864e-05, "loss": 0.6864, "step": 374950 }, { "epoch": 64.52167928423951, "grad_norm": 10.011114120483398, "learning_rate": 1.7739160357880248e-05, "loss": 0.727, "step": 375000 }, { "epoch": 64.53028217481074, "grad_norm": 17.84636878967285, "learning_rate": 1.7734858912594633e-05, "loss": 0.7029, "step": 375050 }, { "epoch": 64.53888506538196, "grad_norm": 15.445840835571289, "learning_rate": 1.7730557467309017e-05, "loss": 0.7205, "step": 375100 }, { "epoch": 64.5474879559532, "grad_norm": 8.23295783996582, "learning_rate": 1.7726256022023402e-05, "loss": 0.6928, "step": 375150 }, { "epoch": 64.55609084652443, "grad_norm": 15.263738632202148, "learning_rate": 1.7721954576737783e-05, "loss": 0.7667, "step": 375200 }, { "epoch": 64.56469373709567, "grad_norm": 12.345664978027344, "learning_rate": 1.771765313145217e-05, "loss": 0.7441, "step": 375250 }, { "epoch": 64.57329662766689, "grad_norm": 13.891575813293457, "learning_rate": 1.7713351686166552e-05, "loss": 0.7163, "step": 375300 }, { "epoch": 64.58189951823813, "grad_norm": 15.758036613464355, "learning_rate": 1.7709050240880937e-05, "loss": 0.678, "step": 375350 }, { "epoch": 64.59050240880936, "grad_norm": 18.797380447387695, "learning_rate": 1.770474879559532e-05, "loss": 0.7183, "step": 375400 }, { "epoch": 64.5991052993806, "grad_norm": 11.665254592895508, "learning_rate": 1.7700447350309702e-05, "loss": 0.7121, "step": 375450 }, { "epoch": 64.60770818995182, "grad_norm": 14.193982124328613, "learning_rate": 1.769614590502409e-05, "loss": 0.7422, "step": 375500 }, { "epoch": 64.61631108052306, "grad_norm": 13.846742630004883, "learning_rate": 1.7691844459738475e-05, "loss": 0.7578, "step": 375550 }, { "epoch": 64.62491397109429, "grad_norm": 15.40621280670166, "learning_rate": 1.7687543014452856e-05, "loss": 0.7238, "step": 375600 }, { "epoch": 64.63351686166553, "grad_norm": 16.481714248657227, "learning_rate": 1.768324156916724e-05, "loss": 0.6664, "step": 375650 }, { "epoch": 64.64211975223675, "grad_norm": 10.014582633972168, "learning_rate": 1.7678940123881625e-05, "loss": 0.6943, "step": 375700 }, { "epoch": 64.65072264280798, "grad_norm": 24.062164306640625, "learning_rate": 1.767463867859601e-05, "loss": 0.6965, "step": 375750 }, { "epoch": 64.65932553337922, "grad_norm": 14.668082237243652, "learning_rate": 1.7670337233310394e-05, "loss": 0.7432, "step": 375800 }, { "epoch": 64.66792842395044, "grad_norm": 10.068825721740723, "learning_rate": 1.7666035788024775e-05, "loss": 0.6989, "step": 375850 }, { "epoch": 64.67653131452168, "grad_norm": 18.705364227294922, "learning_rate": 1.766173434273916e-05, "loss": 0.7055, "step": 375900 }, { "epoch": 64.68513420509291, "grad_norm": 17.218120574951172, "learning_rate": 1.7657432897453548e-05, "loss": 0.7298, "step": 375950 }, { "epoch": 64.69373709566415, "grad_norm": 15.407544136047363, "learning_rate": 1.765313145216793e-05, "loss": 0.7512, "step": 376000 }, { "epoch": 64.70233998623537, "grad_norm": 19.305234909057617, "learning_rate": 1.7648830006882314e-05, "loss": 0.7427, "step": 376050 }, { "epoch": 64.71094287680661, "grad_norm": 12.661364555358887, "learning_rate": 1.7644528561596695e-05, "loss": 0.7985, "step": 376100 }, { "epoch": 64.71954576737784, "grad_norm": 11.519777297973633, "learning_rate": 1.764022711631108e-05, "loss": 0.7387, "step": 376150 }, { "epoch": 64.72814865794908, "grad_norm": 16.5423583984375, "learning_rate": 1.7635925671025467e-05, "loss": 0.7337, "step": 376200 }, { "epoch": 64.7367515485203, "grad_norm": 16.853361129760742, "learning_rate": 1.763162422573985e-05, "loss": 0.7011, "step": 376250 }, { "epoch": 64.74535443909153, "grad_norm": 10.946318626403809, "learning_rate": 1.7627322780454233e-05, "loss": 0.7486, "step": 376300 }, { "epoch": 64.75395732966277, "grad_norm": 16.75688362121582, "learning_rate": 1.7623021335168618e-05, "loss": 0.7072, "step": 376350 }, { "epoch": 64.762560220234, "grad_norm": 12.994669914245605, "learning_rate": 1.7618719889883002e-05, "loss": 0.726, "step": 376400 }, { "epoch": 64.77116311080523, "grad_norm": 11.535650253295898, "learning_rate": 1.7614418444597387e-05, "loss": 0.685, "step": 376450 }, { "epoch": 64.77976600137646, "grad_norm": 16.528491973876953, "learning_rate": 1.761011699931177e-05, "loss": 0.6778, "step": 376500 }, { "epoch": 64.7883688919477, "grad_norm": 16.1497745513916, "learning_rate": 1.7605815554026153e-05, "loss": 0.7084, "step": 376550 }, { "epoch": 64.79697178251892, "grad_norm": 11.381715774536133, "learning_rate": 1.7601514108740537e-05, "loss": 0.7308, "step": 376600 }, { "epoch": 64.80557467309016, "grad_norm": 11.769692420959473, "learning_rate": 1.759721266345492e-05, "loss": 0.6815, "step": 376650 }, { "epoch": 64.81417756366139, "grad_norm": 16.91281509399414, "learning_rate": 1.7592911218169306e-05, "loss": 0.6792, "step": 376700 }, { "epoch": 64.82278045423263, "grad_norm": 11.230608940124512, "learning_rate": 1.758860977288369e-05, "loss": 0.7031, "step": 376750 }, { "epoch": 64.83138334480385, "grad_norm": 19.147253036499023, "learning_rate": 1.7584308327598072e-05, "loss": 0.7445, "step": 376800 }, { "epoch": 64.83998623537508, "grad_norm": 11.31320858001709, "learning_rate": 1.7580006882312456e-05, "loss": 0.6674, "step": 376850 }, { "epoch": 64.84858912594632, "grad_norm": 12.518843650817871, "learning_rate": 1.7575705437026844e-05, "loss": 0.7512, "step": 376900 }, { "epoch": 64.85719201651754, "grad_norm": 25.405017852783203, "learning_rate": 1.7571403991741226e-05, "loss": 0.738, "step": 376950 }, { "epoch": 64.86579490708878, "grad_norm": 23.876264572143555, "learning_rate": 1.756710254645561e-05, "loss": 0.7495, "step": 377000 }, { "epoch": 64.87439779766001, "grad_norm": 13.945097923278809, "learning_rate": 1.756280110116999e-05, "loss": 0.7259, "step": 377050 }, { "epoch": 64.88300068823125, "grad_norm": 12.334002494812012, "learning_rate": 1.755849965588438e-05, "loss": 0.6903, "step": 377100 }, { "epoch": 64.89160357880247, "grad_norm": 14.134892463684082, "learning_rate": 1.7554198210598764e-05, "loss": 0.7233, "step": 377150 }, { "epoch": 64.90020646937371, "grad_norm": 19.591800689697266, "learning_rate": 1.7549896765313145e-05, "loss": 0.7131, "step": 377200 }, { "epoch": 64.90880935994494, "grad_norm": 18.434724807739258, "learning_rate": 1.754559532002753e-05, "loss": 0.7784, "step": 377250 }, { "epoch": 64.91741225051618, "grad_norm": 10.719733238220215, "learning_rate": 1.7541293874741914e-05, "loss": 0.7383, "step": 377300 }, { "epoch": 64.9260151410874, "grad_norm": 12.785226821899414, "learning_rate": 1.75369924294563e-05, "loss": 0.7483, "step": 377350 }, { "epoch": 64.93461803165864, "grad_norm": 12.176015853881836, "learning_rate": 1.7532690984170683e-05, "loss": 0.7759, "step": 377400 }, { "epoch": 64.94322092222987, "grad_norm": 13.49488353729248, "learning_rate": 1.7528389538885064e-05, "loss": 0.7116, "step": 377450 }, { "epoch": 64.9518238128011, "grad_norm": 14.762614250183105, "learning_rate": 1.752408809359945e-05, "loss": 0.7316, "step": 377500 }, { "epoch": 64.96042670337233, "grad_norm": 14.540069580078125, "learning_rate": 1.7519786648313834e-05, "loss": 0.7185, "step": 377550 }, { "epoch": 64.96902959394356, "grad_norm": 21.41538429260254, "learning_rate": 1.7515485203028218e-05, "loss": 0.7329, "step": 377600 }, { "epoch": 64.9776324845148, "grad_norm": 11.093428611755371, "learning_rate": 1.7511183757742603e-05, "loss": 0.7807, "step": 377650 }, { "epoch": 64.98623537508603, "grad_norm": 23.243112564086914, "learning_rate": 1.7506882312456987e-05, "loss": 0.7826, "step": 377700 }, { "epoch": 64.99483826565726, "grad_norm": 10.848424911499023, "learning_rate": 1.750258086717137e-05, "loss": 0.7026, "step": 377750 }, { "epoch": 65.0, "eval_accuracy": 0.5336430906900705, "eval_f1": 0.5218697651928136, "eval_f1_DuraRiadoRio_16x16": 0.5461309523809523, "eval_f1_Mole_16x16": 0.5341200615700359, "eval_f1_Quebrado_16x16": 0.6, "eval_f1_RiadoRio_16x16": 0.4955823293172691, "eval_f1_RioFechado_16x16": 0.4335154826958106, "eval_loss": 2.244483232498169, "eval_precision": 0.5846460139423779, "eval_precision_DuraRiadoRio_16x16": 0.6371527777777778, "eval_precision_Mole_16x16": 0.6530740276035132, "eval_precision_Quebrado_16x16": 0.44594594594594594, "eval_precision_RiadoRio_16x16": 0.48430141287284145, "eval_precision_RioFechado_16x16": 0.702755905511811, "eval_recall": 0.5334376636554072, "eval_recall_DuraRiadoRio_16x16": 0.4778645833333333, "eval_recall_Mole_16x16": 0.4518229166666667, "eval_recall_Quebrado_16x16": 0.9166666666666666, "eval_recall_RiadoRio_16x16": 0.5074013157894737, "eval_recall_RioFechado_16x16": 0.31343283582089554, "eval_runtime": 45.773, "eval_samples_per_second": 253.905, "eval_steps_per_second": 15.883, "step": 377780 }, { "epoch": 65.00344115622849, "grad_norm": 9.08697509765625, "learning_rate": 1.7498279421885756e-05, "loss": 0.739, "step": 377800 }, { "epoch": 65.01204404679973, "grad_norm": 12.290207862854004, "learning_rate": 1.7493977976600138e-05, "loss": 0.6872, "step": 377850 }, { "epoch": 65.02064693737096, "grad_norm": 15.496720314025879, "learning_rate": 1.7489676531314522e-05, "loss": 0.7048, "step": 377900 }, { "epoch": 65.0292498279422, "grad_norm": 11.522099494934082, "learning_rate": 1.7485375086028907e-05, "loss": 0.7154, "step": 377950 }, { "epoch": 65.03785271851342, "grad_norm": 7.529290199279785, "learning_rate": 1.748107364074329e-05, "loss": 0.705, "step": 378000 }, { "epoch": 65.04645560908465, "grad_norm": 22.902284622192383, "learning_rate": 1.7476772195457676e-05, "loss": 0.7218, "step": 378050 }, { "epoch": 65.05505849965589, "grad_norm": 13.365254402160645, "learning_rate": 1.747247075017206e-05, "loss": 0.6979, "step": 378100 }, { "epoch": 65.06366139022711, "grad_norm": 14.218950271606445, "learning_rate": 1.746816930488644e-05, "loss": 0.7004, "step": 378150 }, { "epoch": 65.07226428079835, "grad_norm": 15.931007385253906, "learning_rate": 1.7463867859600826e-05, "loss": 0.7348, "step": 378200 }, { "epoch": 65.08086717136958, "grad_norm": 11.01070499420166, "learning_rate": 1.745956641431521e-05, "loss": 0.7036, "step": 378250 }, { "epoch": 65.08947006194082, "grad_norm": 8.773931503295898, "learning_rate": 1.7455264969029595e-05, "loss": 0.7138, "step": 378300 }, { "epoch": 65.09807295251204, "grad_norm": 16.323841094970703, "learning_rate": 1.745096352374398e-05, "loss": 0.7049, "step": 378350 }, { "epoch": 65.10667584308328, "grad_norm": 11.51109790802002, "learning_rate": 1.744666207845836e-05, "loss": 0.712, "step": 378400 }, { "epoch": 65.1152787336545, "grad_norm": 15.012592315673828, "learning_rate": 1.7442360633172745e-05, "loss": 0.681, "step": 378450 }, { "epoch": 65.12388162422575, "grad_norm": 12.214896202087402, "learning_rate": 1.7438059187887133e-05, "loss": 0.7406, "step": 378500 }, { "epoch": 65.13248451479697, "grad_norm": 14.943241119384766, "learning_rate": 1.7433757742601515e-05, "loss": 0.7137, "step": 378550 }, { "epoch": 65.14108740536821, "grad_norm": 21.78504753112793, "learning_rate": 1.74294562973159e-05, "loss": 0.7069, "step": 378600 }, { "epoch": 65.14969029593944, "grad_norm": 12.591924667358398, "learning_rate": 1.7425154852030284e-05, "loss": 0.7176, "step": 378650 }, { "epoch": 65.15829318651066, "grad_norm": 17.063364028930664, "learning_rate": 1.7420853406744668e-05, "loss": 0.6887, "step": 378700 }, { "epoch": 65.1668960770819, "grad_norm": 13.296332359313965, "learning_rate": 1.7416551961459053e-05, "loss": 0.6521, "step": 378750 }, { "epoch": 65.17549896765313, "grad_norm": 12.572980880737305, "learning_rate": 1.7412250516173434e-05, "loss": 0.7184, "step": 378800 }, { "epoch": 65.18410185822437, "grad_norm": 14.135180473327637, "learning_rate": 1.740794907088782e-05, "loss": 0.7311, "step": 378850 }, { "epoch": 65.19270474879559, "grad_norm": 11.338727951049805, "learning_rate": 1.7403647625602203e-05, "loss": 0.7182, "step": 378900 }, { "epoch": 65.20130763936683, "grad_norm": 9.128915786743164, "learning_rate": 1.7399346180316588e-05, "loss": 0.7114, "step": 378950 }, { "epoch": 65.20991052993806, "grad_norm": 19.07548713684082, "learning_rate": 1.7395044735030972e-05, "loss": 0.7088, "step": 379000 }, { "epoch": 65.2185134205093, "grad_norm": 15.92237377166748, "learning_rate": 1.7390743289745357e-05, "loss": 0.6794, "step": 379050 }, { "epoch": 65.22711631108052, "grad_norm": 11.433481216430664, "learning_rate": 1.7386441844459738e-05, "loss": 0.6987, "step": 379100 }, { "epoch": 65.23571920165176, "grad_norm": 12.770936012268066, "learning_rate": 1.7382140399174123e-05, "loss": 0.7087, "step": 379150 }, { "epoch": 65.24432209222299, "grad_norm": 20.596765518188477, "learning_rate": 1.7377838953888507e-05, "loss": 0.7076, "step": 379200 }, { "epoch": 65.25292498279421, "grad_norm": 17.922433853149414, "learning_rate": 1.737353750860289e-05, "loss": 0.6758, "step": 379250 }, { "epoch": 65.26152787336545, "grad_norm": 11.081007957458496, "learning_rate": 1.7369236063317276e-05, "loss": 0.7991, "step": 379300 }, { "epoch": 65.27013076393668, "grad_norm": 11.4227294921875, "learning_rate": 1.7364934618031657e-05, "loss": 0.6441, "step": 379350 }, { "epoch": 65.27873365450792, "grad_norm": 17.331851959228516, "learning_rate": 1.7360633172746042e-05, "loss": 0.74, "step": 379400 }, { "epoch": 65.28733654507914, "grad_norm": 10.926342010498047, "learning_rate": 1.735633172746043e-05, "loss": 0.6772, "step": 379450 }, { "epoch": 65.29593943565038, "grad_norm": 17.4522647857666, "learning_rate": 1.735203028217481e-05, "loss": 0.6997, "step": 379500 }, { "epoch": 65.30454232622161, "grad_norm": 18.727628707885742, "learning_rate": 1.7347728836889196e-05, "loss": 0.7388, "step": 379550 }, { "epoch": 65.31314521679285, "grad_norm": 12.487726211547852, "learning_rate": 1.7343427391603577e-05, "loss": 0.7551, "step": 379600 }, { "epoch": 65.32174810736407, "grad_norm": 17.715314865112305, "learning_rate": 1.7339125946317965e-05, "loss": 0.778, "step": 379650 }, { "epoch": 65.33035099793531, "grad_norm": 17.20648765563965, "learning_rate": 1.733482450103235e-05, "loss": 0.6993, "step": 379700 }, { "epoch": 65.33895388850654, "grad_norm": 15.066121101379395, "learning_rate": 1.733052305574673e-05, "loss": 0.6992, "step": 379750 }, { "epoch": 65.34755677907776, "grad_norm": 9.22723388671875, "learning_rate": 1.7326221610461115e-05, "loss": 0.6881, "step": 379800 }, { "epoch": 65.356159669649, "grad_norm": 13.915943145751953, "learning_rate": 1.73219201651755e-05, "loss": 0.6933, "step": 379850 }, { "epoch": 65.36476256022023, "grad_norm": 14.087850570678711, "learning_rate": 1.7317618719889884e-05, "loss": 0.7469, "step": 379900 }, { "epoch": 65.37336545079147, "grad_norm": 19.036897659301758, "learning_rate": 1.731331727460427e-05, "loss": 0.798, "step": 379950 }, { "epoch": 65.3819683413627, "grad_norm": 9.695699691772461, "learning_rate": 1.730901582931865e-05, "loss": 0.7264, "step": 380000 }, { "epoch": 65.39057123193393, "grad_norm": 11.615171432495117, "learning_rate": 1.7304714384033034e-05, "loss": 0.7159, "step": 380050 }, { "epoch": 65.39917412250516, "grad_norm": 20.803693771362305, "learning_rate": 1.730041293874742e-05, "loss": 0.6993, "step": 380100 }, { "epoch": 65.4077770130764, "grad_norm": 17.550935745239258, "learning_rate": 1.7296111493461804e-05, "loss": 0.7121, "step": 380150 }, { "epoch": 65.41637990364762, "grad_norm": 11.528557777404785, "learning_rate": 1.7291810048176188e-05, "loss": 0.7839, "step": 380200 }, { "epoch": 65.42498279421886, "grad_norm": 11.93471908569336, "learning_rate": 1.7287508602890573e-05, "loss": 0.7472, "step": 380250 }, { "epoch": 65.43358568479009, "grad_norm": 9.557151794433594, "learning_rate": 1.7283207157604954e-05, "loss": 0.7167, "step": 380300 }, { "epoch": 65.44218857536133, "grad_norm": 12.485918998718262, "learning_rate": 1.7278905712319342e-05, "loss": 0.7162, "step": 380350 }, { "epoch": 65.45079146593255, "grad_norm": 13.476380348205566, "learning_rate": 1.7274604267033726e-05, "loss": 0.692, "step": 380400 }, { "epoch": 65.45939435650378, "grad_norm": 8.850334167480469, "learning_rate": 1.7270302821748108e-05, "loss": 0.71, "step": 380450 }, { "epoch": 65.46799724707502, "grad_norm": 12.652323722839355, "learning_rate": 1.7266001376462492e-05, "loss": 0.6879, "step": 380500 }, { "epoch": 65.47660013764624, "grad_norm": 18.09718894958496, "learning_rate": 1.7261699931176877e-05, "loss": 0.7551, "step": 380550 }, { "epoch": 65.48520302821748, "grad_norm": 16.757736206054688, "learning_rate": 1.725739848589126e-05, "loss": 0.7303, "step": 380600 }, { "epoch": 65.49380591878871, "grad_norm": 15.625950813293457, "learning_rate": 1.7253097040605646e-05, "loss": 0.7431, "step": 380650 }, { "epoch": 65.50240880935995, "grad_norm": 17.460739135742188, "learning_rate": 1.7248795595320027e-05, "loss": 0.715, "step": 380700 }, { "epoch": 65.51101169993117, "grad_norm": 5.64289665222168, "learning_rate": 1.724449415003441e-05, "loss": 0.6902, "step": 380750 }, { "epoch": 65.51961459050241, "grad_norm": 23.719112396240234, "learning_rate": 1.7240192704748796e-05, "loss": 0.6883, "step": 380800 }, { "epoch": 65.52821748107364, "grad_norm": 10.712058067321777, "learning_rate": 1.723589125946318e-05, "loss": 0.7217, "step": 380850 }, { "epoch": 65.53682037164488, "grad_norm": 12.550654411315918, "learning_rate": 1.7231589814177565e-05, "loss": 0.714, "step": 380900 }, { "epoch": 65.5454232622161, "grad_norm": 11.17015552520752, "learning_rate": 1.7227288368891946e-05, "loss": 0.733, "step": 380950 }, { "epoch": 65.55402615278733, "grad_norm": 12.90907096862793, "learning_rate": 1.722298692360633e-05, "loss": 0.6841, "step": 381000 }, { "epoch": 65.56262904335857, "grad_norm": 14.972893714904785, "learning_rate": 1.721868547832072e-05, "loss": 0.749, "step": 381050 }, { "epoch": 65.5712319339298, "grad_norm": 12.882473945617676, "learning_rate": 1.72143840330351e-05, "loss": 0.7433, "step": 381100 }, { "epoch": 65.57983482450103, "grad_norm": 15.593786239624023, "learning_rate": 1.7210082587749485e-05, "loss": 0.7554, "step": 381150 }, { "epoch": 65.58843771507226, "grad_norm": 12.03397274017334, "learning_rate": 1.720578114246387e-05, "loss": 0.705, "step": 381200 }, { "epoch": 65.5970406056435, "grad_norm": 10.062023162841797, "learning_rate": 1.7201479697178254e-05, "loss": 0.6878, "step": 381250 }, { "epoch": 65.60564349621473, "grad_norm": 14.318254470825195, "learning_rate": 1.719717825189264e-05, "loss": 0.6733, "step": 381300 }, { "epoch": 65.61424638678596, "grad_norm": 8.614435195922852, "learning_rate": 1.719287680660702e-05, "loss": 0.736, "step": 381350 }, { "epoch": 65.62284927735719, "grad_norm": 12.901341438293457, "learning_rate": 1.7188575361321404e-05, "loss": 0.6939, "step": 381400 }, { "epoch": 65.63145216792843, "grad_norm": 13.936600685119629, "learning_rate": 1.718427391603579e-05, "loss": 0.6375, "step": 381450 }, { "epoch": 65.64005505849966, "grad_norm": 13.97054672241211, "learning_rate": 1.7179972470750173e-05, "loss": 0.7214, "step": 381500 }, { "epoch": 65.6486579490709, "grad_norm": 10.863178253173828, "learning_rate": 1.7175671025464558e-05, "loss": 0.7304, "step": 381550 }, { "epoch": 65.65726083964212, "grad_norm": 15.269760131835938, "learning_rate": 1.7171369580178942e-05, "loss": 0.7618, "step": 381600 }, { "epoch": 65.66586373021335, "grad_norm": 18.475177764892578, "learning_rate": 1.7167068134893323e-05, "loss": 0.7534, "step": 381650 }, { "epoch": 65.67446662078459, "grad_norm": 13.367215156555176, "learning_rate": 1.7162766689607708e-05, "loss": 0.6802, "step": 381700 }, { "epoch": 65.68306951135581, "grad_norm": 11.050200462341309, "learning_rate": 1.7158465244322093e-05, "loss": 0.744, "step": 381750 }, { "epoch": 65.69167240192705, "grad_norm": 9.101059913635254, "learning_rate": 1.7154163799036477e-05, "loss": 0.7314, "step": 381800 }, { "epoch": 65.70027529249828, "grad_norm": 15.809484481811523, "learning_rate": 1.7149862353750862e-05, "loss": 0.7591, "step": 381850 }, { "epoch": 65.70887818306952, "grad_norm": 23.5684814453125, "learning_rate": 1.7145560908465243e-05, "loss": 0.6793, "step": 381900 }, { "epoch": 65.71748107364074, "grad_norm": 12.324339866638184, "learning_rate": 1.714125946317963e-05, "loss": 0.6899, "step": 381950 }, { "epoch": 65.72608396421198, "grad_norm": 18.69995880126953, "learning_rate": 1.7136958017894015e-05, "loss": 0.6954, "step": 382000 }, { "epoch": 65.7346868547832, "grad_norm": 11.796587944030762, "learning_rate": 1.7132656572608397e-05, "loss": 0.729, "step": 382050 }, { "epoch": 65.74328974535445, "grad_norm": 20.77285385131836, "learning_rate": 1.712835512732278e-05, "loss": 0.7078, "step": 382100 }, { "epoch": 65.75189263592567, "grad_norm": 18.83816909790039, "learning_rate": 1.7124053682037166e-05, "loss": 0.7277, "step": 382150 }, { "epoch": 65.7604955264969, "grad_norm": 17.009065628051758, "learning_rate": 1.711975223675155e-05, "loss": 0.6762, "step": 382200 }, { "epoch": 65.76909841706814, "grad_norm": 8.801209449768066, "learning_rate": 1.7115450791465935e-05, "loss": 0.7475, "step": 382250 }, { "epoch": 65.77770130763936, "grad_norm": 13.687739372253418, "learning_rate": 1.7111149346180316e-05, "loss": 0.7453, "step": 382300 }, { "epoch": 65.7863041982106, "grad_norm": 13.384211540222168, "learning_rate": 1.71068479008947e-05, "loss": 0.7121, "step": 382350 }, { "epoch": 65.79490708878183, "grad_norm": 16.137117385864258, "learning_rate": 1.7102546455609085e-05, "loss": 0.7153, "step": 382400 }, { "epoch": 65.80350997935307, "grad_norm": 12.648897171020508, "learning_rate": 1.709824501032347e-05, "loss": 0.7151, "step": 382450 }, { "epoch": 65.81211286992429, "grad_norm": 19.673032760620117, "learning_rate": 1.7093943565037854e-05, "loss": 0.703, "step": 382500 }, { "epoch": 65.82071576049553, "grad_norm": 15.31014633178711, "learning_rate": 1.708964211975224e-05, "loss": 0.6565, "step": 382550 }, { "epoch": 65.82931865106676, "grad_norm": 17.542036056518555, "learning_rate": 1.708534067446662e-05, "loss": 0.7626, "step": 382600 }, { "epoch": 65.837921541638, "grad_norm": 16.348506927490234, "learning_rate": 1.7081039229181005e-05, "loss": 0.7047, "step": 382650 }, { "epoch": 65.84652443220922, "grad_norm": 8.281844139099121, "learning_rate": 1.707673778389539e-05, "loss": 0.6789, "step": 382700 }, { "epoch": 65.85512732278045, "grad_norm": 12.044588088989258, "learning_rate": 1.7072436338609774e-05, "loss": 0.7536, "step": 382750 }, { "epoch": 65.86373021335169, "grad_norm": 16.61647605895996, "learning_rate": 1.7068134893324158e-05, "loss": 0.7047, "step": 382800 }, { "epoch": 65.87233310392291, "grad_norm": 11.269796371459961, "learning_rate": 1.706383344803854e-05, "loss": 0.7158, "step": 382850 }, { "epoch": 65.88093599449415, "grad_norm": 17.707138061523438, "learning_rate": 1.7059532002752927e-05, "loss": 0.709, "step": 382900 }, { "epoch": 65.88953888506538, "grad_norm": 14.386630058288574, "learning_rate": 1.7055230557467312e-05, "loss": 0.7469, "step": 382950 }, { "epoch": 65.89814177563662, "grad_norm": 9.144037246704102, "learning_rate": 1.7050929112181693e-05, "loss": 0.6833, "step": 383000 }, { "epoch": 65.90674466620784, "grad_norm": 18.555728912353516, "learning_rate": 1.7046627666896078e-05, "loss": 0.7677, "step": 383050 }, { "epoch": 65.91534755677908, "grad_norm": 15.028829574584961, "learning_rate": 1.7042326221610462e-05, "loss": 0.7075, "step": 383100 }, { "epoch": 65.92395044735031, "grad_norm": 13.534529685974121, "learning_rate": 1.7038024776324847e-05, "loss": 0.7205, "step": 383150 }, { "epoch": 65.93255333792155, "grad_norm": 12.55201244354248, "learning_rate": 1.703372333103923e-05, "loss": 0.7723, "step": 383200 }, { "epoch": 65.94115622849277, "grad_norm": 11.054109573364258, "learning_rate": 1.7029421885753612e-05, "loss": 0.7075, "step": 383250 }, { "epoch": 65.94975911906401, "grad_norm": 13.230755805969238, "learning_rate": 1.7025120440467997e-05, "loss": 0.718, "step": 383300 }, { "epoch": 65.95836200963524, "grad_norm": 16.090534210205078, "learning_rate": 1.702081899518238e-05, "loss": 0.7074, "step": 383350 }, { "epoch": 65.96696490020646, "grad_norm": 16.793956756591797, "learning_rate": 1.7016517549896766e-05, "loss": 0.7443, "step": 383400 }, { "epoch": 65.9755677907777, "grad_norm": 13.312713623046875, "learning_rate": 1.701221610461115e-05, "loss": 0.7231, "step": 383450 }, { "epoch": 65.98417068134893, "grad_norm": 16.7473201751709, "learning_rate": 1.7007914659325532e-05, "loss": 0.7357, "step": 383500 }, { "epoch": 65.99277357192017, "grad_norm": 20.0018253326416, "learning_rate": 1.7003613214039916e-05, "loss": 0.7319, "step": 383550 }, { "epoch": 66.0, "eval_accuracy": 0.5920667699191189, "eval_f1": 0.5934103411749304, "eval_f1_DuraRiadoRio_16x16": 0.5554665598718462, "eval_f1_Mole_16x16": 0.6473988439306358, "eval_f1_Quebrado_16x16": 0.6744669786791472, "eval_f1_RiadoRio_16x16": 0.4594654296443561, "eval_f1_RioFechado_16x16": 0.6302538937486665, "eval_loss": 1.8517179489135742, "eval_precision": 0.6096773846769461, "eval_precision_DuraRiadoRio_16x16": 0.5156133828996282, "eval_precision_Mole_16x16": 0.5821205821205822, "eval_precision_Quebrado_16x16": 0.8411154345006485, "eval_precision_RiadoRio_16x16": 0.4964200477326969, "eval_precision_RioFechado_16x16": 0.6131174761311747, "eval_recall": 0.5940209138774651, "eval_recall_DuraRiadoRio_16x16": 0.6019965277777778, "eval_recall_Mole_16x16": 0.7291666666666666, "eval_recall_Quebrado_16x16": 0.5629340277777778, "eval_recall_RiadoRio_16x16": 0.4276315789473684, "eval_recall_RioFechado_16x16": 0.6483757682177349, "eval_runtime": 45.7735, "eval_samples_per_second": 253.902, "eval_steps_per_second": 15.883, "step": 383592 }, { "epoch": 66.0013764624914, "grad_norm": 11.202564239501953, "learning_rate": 1.6999311768754304e-05, "loss": 0.739, "step": 383600 }, { "epoch": 66.00997935306263, "grad_norm": 12.245973587036133, "learning_rate": 1.6995010323468686e-05, "loss": 0.6959, "step": 383650 }, { "epoch": 66.01858224363386, "grad_norm": 13.951194763183594, "learning_rate": 1.699070887818307e-05, "loss": 0.7457, "step": 383700 }, { "epoch": 66.0271851342051, "grad_norm": 13.93825626373291, "learning_rate": 1.6986407432897455e-05, "loss": 0.741, "step": 383750 }, { "epoch": 66.03578802477632, "grad_norm": 17.4287166595459, "learning_rate": 1.698210598761184e-05, "loss": 0.6717, "step": 383800 }, { "epoch": 66.04439091534756, "grad_norm": 17.965290069580078, "learning_rate": 1.6977804542326224e-05, "loss": 0.6986, "step": 383850 }, { "epoch": 66.05299380591879, "grad_norm": 15.351261138916016, "learning_rate": 1.697350309704061e-05, "loss": 0.7069, "step": 383900 }, { "epoch": 66.06159669649001, "grad_norm": 14.715812683105469, "learning_rate": 1.696920165175499e-05, "loss": 0.7395, "step": 383950 }, { "epoch": 66.07019958706125, "grad_norm": 16.069169998168945, "learning_rate": 1.6964900206469374e-05, "loss": 0.6997, "step": 384000 }, { "epoch": 66.07880247763248, "grad_norm": 12.45429515838623, "learning_rate": 1.696059876118376e-05, "loss": 0.6969, "step": 384050 }, { "epoch": 66.08740536820372, "grad_norm": 14.105769157409668, "learning_rate": 1.6956297315898143e-05, "loss": 0.7815, "step": 384100 }, { "epoch": 66.09600825877494, "grad_norm": 10.134699821472168, "learning_rate": 1.6951995870612528e-05, "loss": 0.724, "step": 384150 }, { "epoch": 66.10461114934618, "grad_norm": 10.085293769836426, "learning_rate": 1.694769442532691e-05, "loss": 0.6934, "step": 384200 }, { "epoch": 66.11321403991741, "grad_norm": 13.56592845916748, "learning_rate": 1.6943392980041293e-05, "loss": 0.7301, "step": 384250 }, { "epoch": 66.12181693048865, "grad_norm": 13.40754508972168, "learning_rate": 1.693909153475568e-05, "loss": 0.7452, "step": 384300 }, { "epoch": 66.13041982105987, "grad_norm": 13.296594619750977, "learning_rate": 1.6934790089470063e-05, "loss": 0.6846, "step": 384350 }, { "epoch": 66.13902271163111, "grad_norm": 6.354238033294678, "learning_rate": 1.6930488644184447e-05, "loss": 0.71, "step": 384400 }, { "epoch": 66.14762560220234, "grad_norm": 18.30087661743164, "learning_rate": 1.692618719889883e-05, "loss": 0.7089, "step": 384450 }, { "epoch": 66.15622849277358, "grad_norm": 9.589500427246094, "learning_rate": 1.6921885753613216e-05, "loss": 0.6767, "step": 384500 }, { "epoch": 66.1648313833448, "grad_norm": 15.705619812011719, "learning_rate": 1.69175843083276e-05, "loss": 0.6858, "step": 384550 }, { "epoch": 66.17343427391603, "grad_norm": 15.042218208312988, "learning_rate": 1.6913282863041982e-05, "loss": 0.7116, "step": 384600 }, { "epoch": 66.18203716448727, "grad_norm": 10.355679512023926, "learning_rate": 1.6908981417756367e-05, "loss": 0.7287, "step": 384650 }, { "epoch": 66.1906400550585, "grad_norm": 13.0565185546875, "learning_rate": 1.690467997247075e-05, "loss": 0.7131, "step": 384700 }, { "epoch": 66.19924294562973, "grad_norm": 11.109524726867676, "learning_rate": 1.6900378527185136e-05, "loss": 0.7459, "step": 384750 }, { "epoch": 66.20784583620096, "grad_norm": 10.150619506835938, "learning_rate": 1.689607708189952e-05, "loss": 0.6537, "step": 384800 }, { "epoch": 66.2164487267722, "grad_norm": 10.57193374633789, "learning_rate": 1.68917756366139e-05, "loss": 0.7485, "step": 384850 }, { "epoch": 66.22505161734342, "grad_norm": 18.461509704589844, "learning_rate": 1.6887474191328286e-05, "loss": 0.7148, "step": 384900 }, { "epoch": 66.23365450791466, "grad_norm": 7.636351108551025, "learning_rate": 1.688317274604267e-05, "loss": 0.6873, "step": 384950 }, { "epoch": 66.24225739848589, "grad_norm": 16.28044891357422, "learning_rate": 1.6878871300757055e-05, "loss": 0.7593, "step": 385000 }, { "epoch": 66.25086028905713, "grad_norm": 8.703317642211914, "learning_rate": 1.687456985547144e-05, "loss": 0.706, "step": 385050 }, { "epoch": 66.25946317962836, "grad_norm": 7.854639053344727, "learning_rate": 1.6870268410185824e-05, "loss": 0.7479, "step": 385100 }, { "epoch": 66.26806607019958, "grad_norm": 6.667532444000244, "learning_rate": 1.6865966964900205e-05, "loss": 0.7185, "step": 385150 }, { "epoch": 66.27666896077082, "grad_norm": 9.104005813598633, "learning_rate": 1.6861665519614593e-05, "loss": 0.7115, "step": 385200 }, { "epoch": 66.28527185134205, "grad_norm": 17.92078971862793, "learning_rate": 1.6857364074328975e-05, "loss": 0.6969, "step": 385250 }, { "epoch": 66.29387474191329, "grad_norm": 14.875105857849121, "learning_rate": 1.685306262904336e-05, "loss": 0.7073, "step": 385300 }, { "epoch": 66.30247763248451, "grad_norm": 26.516706466674805, "learning_rate": 1.6848761183757744e-05, "loss": 0.6747, "step": 385350 }, { "epoch": 66.31108052305575, "grad_norm": 12.272845268249512, "learning_rate": 1.6844459738472125e-05, "loss": 0.7749, "step": 385400 }, { "epoch": 66.31968341362698, "grad_norm": 23.08891487121582, "learning_rate": 1.6840158293186513e-05, "loss": 0.7164, "step": 385450 }, { "epoch": 66.32828630419822, "grad_norm": 14.964437484741211, "learning_rate": 1.6835856847900897e-05, "loss": 0.7032, "step": 385500 }, { "epoch": 66.33688919476944, "grad_norm": 14.399330139160156, "learning_rate": 1.683155540261528e-05, "loss": 0.7144, "step": 385550 }, { "epoch": 66.34549208534068, "grad_norm": 13.099337577819824, "learning_rate": 1.6827253957329663e-05, "loss": 0.6656, "step": 385600 }, { "epoch": 66.3540949759119, "grad_norm": 17.59469223022461, "learning_rate": 1.6822952512044048e-05, "loss": 0.7036, "step": 385650 }, { "epoch": 66.36269786648313, "grad_norm": 16.869102478027344, "learning_rate": 1.6818651066758432e-05, "loss": 0.7461, "step": 385700 }, { "epoch": 66.37130075705437, "grad_norm": 16.053892135620117, "learning_rate": 1.6814349621472817e-05, "loss": 0.6651, "step": 385750 }, { "epoch": 66.3799036476256, "grad_norm": 15.873509407043457, "learning_rate": 1.6810048176187198e-05, "loss": 0.7669, "step": 385800 }, { "epoch": 66.38850653819684, "grad_norm": 20.96152687072754, "learning_rate": 1.6805746730901582e-05, "loss": 0.7299, "step": 385850 }, { "epoch": 66.39710942876806, "grad_norm": 15.207547187805176, "learning_rate": 1.680144528561597e-05, "loss": 0.6506, "step": 385900 }, { "epoch": 66.4057123193393, "grad_norm": 13.80320930480957, "learning_rate": 1.679714384033035e-05, "loss": 0.6883, "step": 385950 }, { "epoch": 66.41431520991053, "grad_norm": 17.389463424682617, "learning_rate": 1.6792842395044736e-05, "loss": 0.7299, "step": 386000 }, { "epoch": 66.42291810048177, "grad_norm": 16.372695922851562, "learning_rate": 1.678854094975912e-05, "loss": 0.6978, "step": 386050 }, { "epoch": 66.43152099105299, "grad_norm": 19.157922744750977, "learning_rate": 1.6784239504473502e-05, "loss": 0.8045, "step": 386100 }, { "epoch": 66.44012388162423, "grad_norm": 19.0301456451416, "learning_rate": 1.677993805918789e-05, "loss": 0.7587, "step": 386150 }, { "epoch": 66.44872677219546, "grad_norm": 25.043336868286133, "learning_rate": 1.677563661390227e-05, "loss": 0.6969, "step": 386200 }, { "epoch": 66.4573296627667, "grad_norm": 18.61172103881836, "learning_rate": 1.6771335168616656e-05, "loss": 0.7336, "step": 386250 }, { "epoch": 66.46593255333792, "grad_norm": 22.096166610717773, "learning_rate": 1.676703372333104e-05, "loss": 0.7071, "step": 386300 }, { "epoch": 66.47453544390915, "grad_norm": 12.114983558654785, "learning_rate": 1.6762732278045425e-05, "loss": 0.7408, "step": 386350 }, { "epoch": 66.48313833448039, "grad_norm": 20.24579620361328, "learning_rate": 1.675843083275981e-05, "loss": 0.7192, "step": 386400 }, { "epoch": 66.49174122505161, "grad_norm": 11.043920516967773, "learning_rate": 1.6754129387474194e-05, "loss": 0.7349, "step": 386450 }, { "epoch": 66.50034411562285, "grad_norm": 10.344254493713379, "learning_rate": 1.6749827942188575e-05, "loss": 0.6565, "step": 386500 }, { "epoch": 66.50894700619408, "grad_norm": 12.707022666931152, "learning_rate": 1.674552649690296e-05, "loss": 0.6516, "step": 386550 }, { "epoch": 66.51754989676532, "grad_norm": 16.52663803100586, "learning_rate": 1.6741225051617344e-05, "loss": 0.7209, "step": 386600 }, { "epoch": 66.52615278733654, "grad_norm": 11.960389137268066, "learning_rate": 1.673692360633173e-05, "loss": 0.7075, "step": 386650 }, { "epoch": 66.53475567790778, "grad_norm": 13.033966064453125, "learning_rate": 1.6732622161046113e-05, "loss": 0.7513, "step": 386700 }, { "epoch": 66.54335856847901, "grad_norm": 19.91212272644043, "learning_rate": 1.6728320715760494e-05, "loss": 0.6892, "step": 386750 }, { "epoch": 66.55196145905025, "grad_norm": 12.82685661315918, "learning_rate": 1.672401927047488e-05, "loss": 0.6794, "step": 386800 }, { "epoch": 66.56056434962147, "grad_norm": 13.156272888183594, "learning_rate": 1.6719717825189267e-05, "loss": 0.6719, "step": 386850 }, { "epoch": 66.5691672401927, "grad_norm": 19.095218658447266, "learning_rate": 1.6715416379903648e-05, "loss": 0.7078, "step": 386900 }, { "epoch": 66.57777013076394, "grad_norm": 16.240203857421875, "learning_rate": 1.6711114934618033e-05, "loss": 0.6979, "step": 386950 }, { "epoch": 66.58637302133516, "grad_norm": 13.142480850219727, "learning_rate": 1.6706813489332414e-05, "loss": 0.6675, "step": 387000 }, { "epoch": 66.5949759119064, "grad_norm": 9.838066101074219, "learning_rate": 1.6702512044046802e-05, "loss": 0.6816, "step": 387050 }, { "epoch": 66.60357880247763, "grad_norm": 22.358184814453125, "learning_rate": 1.6698210598761186e-05, "loss": 0.6639, "step": 387100 }, { "epoch": 66.61218169304887, "grad_norm": 15.437363624572754, "learning_rate": 1.6693909153475568e-05, "loss": 0.7144, "step": 387150 }, { "epoch": 66.6207845836201, "grad_norm": 10.450594902038574, "learning_rate": 1.6689607708189952e-05, "loss": 0.7563, "step": 387200 }, { "epoch": 66.62938747419133, "grad_norm": 12.732502937316895, "learning_rate": 1.6685306262904337e-05, "loss": 0.7246, "step": 387250 }, { "epoch": 66.63799036476256, "grad_norm": 16.930889129638672, "learning_rate": 1.668100481761872e-05, "loss": 0.7127, "step": 387300 }, { "epoch": 66.6465932553338, "grad_norm": 25.97026824951172, "learning_rate": 1.6676703372333106e-05, "loss": 0.7266, "step": 387350 }, { "epoch": 66.65519614590502, "grad_norm": 8.57060432434082, "learning_rate": 1.6672401927047487e-05, "loss": 0.6588, "step": 387400 }, { "epoch": 66.66379903647626, "grad_norm": 23.423744201660156, "learning_rate": 1.666810048176187e-05, "loss": 0.7144, "step": 387450 }, { "epoch": 66.67240192704749, "grad_norm": 16.5234432220459, "learning_rate": 1.6663799036476256e-05, "loss": 0.7201, "step": 387500 }, { "epoch": 66.68100481761871, "grad_norm": 13.0784273147583, "learning_rate": 1.665949759119064e-05, "loss": 0.7526, "step": 387550 }, { "epoch": 66.68960770818995, "grad_norm": 11.475543975830078, "learning_rate": 1.6655196145905025e-05, "loss": 0.7635, "step": 387600 }, { "epoch": 66.69821059876118, "grad_norm": 17.636232376098633, "learning_rate": 1.665089470061941e-05, "loss": 0.732, "step": 387650 }, { "epoch": 66.70681348933242, "grad_norm": 11.802064895629883, "learning_rate": 1.664659325533379e-05, "loss": 0.7345, "step": 387700 }, { "epoch": 66.71541637990364, "grad_norm": 17.1914005279541, "learning_rate": 1.664229181004818e-05, "loss": 0.7726, "step": 387750 }, { "epoch": 66.72401927047488, "grad_norm": 19.1948184967041, "learning_rate": 1.6637990364762563e-05, "loss": 0.6908, "step": 387800 }, { "epoch": 66.73262216104611, "grad_norm": 12.628130912780762, "learning_rate": 1.6633688919476945e-05, "loss": 0.7324, "step": 387850 }, { "epoch": 66.74122505161735, "grad_norm": 11.484445571899414, "learning_rate": 1.662938747419133e-05, "loss": 0.6913, "step": 387900 }, { "epoch": 66.74982794218857, "grad_norm": 10.848382949829102, "learning_rate": 1.662508602890571e-05, "loss": 0.7467, "step": 387950 }, { "epoch": 66.75843083275981, "grad_norm": 9.188520431518555, "learning_rate": 1.6620784583620098e-05, "loss": 0.6736, "step": 388000 }, { "epoch": 66.76703372333104, "grad_norm": 18.287155151367188, "learning_rate": 1.6616483138334483e-05, "loss": 0.7292, "step": 388050 }, { "epoch": 66.77563661390226, "grad_norm": 18.9058895111084, "learning_rate": 1.6612181693048864e-05, "loss": 0.731, "step": 388100 }, { "epoch": 66.7842395044735, "grad_norm": 21.749242782592773, "learning_rate": 1.660788024776325e-05, "loss": 0.726, "step": 388150 }, { "epoch": 66.79284239504473, "grad_norm": 13.941513061523438, "learning_rate": 1.6603578802477633e-05, "loss": 0.7257, "step": 388200 }, { "epoch": 66.80144528561597, "grad_norm": 14.266083717346191, "learning_rate": 1.6599277357192018e-05, "loss": 0.6747, "step": 388250 }, { "epoch": 66.8100481761872, "grad_norm": 11.427578926086426, "learning_rate": 1.6594975911906402e-05, "loss": 0.738, "step": 388300 }, { "epoch": 66.81865106675843, "grad_norm": 7.67887544631958, "learning_rate": 1.6590674466620783e-05, "loss": 0.7096, "step": 388350 }, { "epoch": 66.82725395732966, "grad_norm": 18.791271209716797, "learning_rate": 1.6586373021335168e-05, "loss": 0.6912, "step": 388400 }, { "epoch": 66.8358568479009, "grad_norm": 21.359281539916992, "learning_rate": 1.6582071576049556e-05, "loss": 0.6693, "step": 388450 }, { "epoch": 66.84445973847212, "grad_norm": 13.768628120422363, "learning_rate": 1.6577770130763937e-05, "loss": 0.7008, "step": 388500 }, { "epoch": 66.85306262904336, "grad_norm": 14.311178207397461, "learning_rate": 1.657346868547832e-05, "loss": 0.7071, "step": 388550 }, { "epoch": 66.86166551961459, "grad_norm": 17.053115844726562, "learning_rate": 1.6569167240192706e-05, "loss": 0.7048, "step": 388600 }, { "epoch": 66.87026841018582, "grad_norm": 11.917977333068848, "learning_rate": 1.6564865794907087e-05, "loss": 0.6777, "step": 388650 }, { "epoch": 66.87887130075705, "grad_norm": 10.186111450195312, "learning_rate": 1.6560564349621475e-05, "loss": 0.7037, "step": 388700 }, { "epoch": 66.88747419132828, "grad_norm": 19.22406005859375, "learning_rate": 1.6556262904335857e-05, "loss": 0.6871, "step": 388750 }, { "epoch": 66.89607708189952, "grad_norm": 16.917943954467773, "learning_rate": 1.655196145905024e-05, "loss": 0.7189, "step": 388800 }, { "epoch": 66.90467997247075, "grad_norm": 17.893495559692383, "learning_rate": 1.6547660013764626e-05, "loss": 0.7129, "step": 388850 }, { "epoch": 66.91328286304199, "grad_norm": 14.901901245117188, "learning_rate": 1.654335856847901e-05, "loss": 0.6764, "step": 388900 }, { "epoch": 66.92188575361321, "grad_norm": 21.772838592529297, "learning_rate": 1.6539057123193395e-05, "loss": 0.7339, "step": 388950 }, { "epoch": 66.93048864418445, "grad_norm": 11.320022583007812, "learning_rate": 1.653475567790778e-05, "loss": 0.7753, "step": 389000 }, { "epoch": 66.93909153475568, "grad_norm": 15.849730491638184, "learning_rate": 1.653045423262216e-05, "loss": 0.7252, "step": 389050 }, { "epoch": 66.94769442532692, "grad_norm": 7.389357566833496, "learning_rate": 1.6526152787336545e-05, "loss": 0.7313, "step": 389100 }, { "epoch": 66.95629731589814, "grad_norm": 14.828728675842285, "learning_rate": 1.652185134205093e-05, "loss": 0.7153, "step": 389150 }, { "epoch": 66.96490020646938, "grad_norm": 13.471203804016113, "learning_rate": 1.6517549896765314e-05, "loss": 0.7021, "step": 389200 }, { "epoch": 66.9735030970406, "grad_norm": 23.307662963867188, "learning_rate": 1.65132484514797e-05, "loss": 0.7512, "step": 389250 }, { "epoch": 66.98210598761183, "grad_norm": 13.469773292541504, "learning_rate": 1.650894700619408e-05, "loss": 0.6866, "step": 389300 }, { "epoch": 66.99070887818307, "grad_norm": 14.912004470825195, "learning_rate": 1.6504645560908464e-05, "loss": 0.7352, "step": 389350 }, { "epoch": 66.9993117687543, "grad_norm": 11.49178695678711, "learning_rate": 1.6500344115622852e-05, "loss": 0.7037, "step": 389400 }, { "epoch": 67.0, "eval_accuracy": 0.6196868008948546, "eval_f1": 0.6210772893837172, "eval_f1_DuraRiadoRio_16x16": 0.5582898334576187, "eval_f1_Mole_16x16": 0.6166166166166166, "eval_f1_Quebrado_16x16": 0.7338551859099804, "eval_f1_RiadoRio_16x16": 0.5665421663062709, "eval_f1_RioFechado_16x16": 0.6300826446280992, "eval_loss": 1.603477954864502, "eval_precision": 0.6540556226736639, "eval_precision_DuraRiadoRio_16x16": 0.6532867946480512, "eval_precision_Mole_16x16": 0.7281323877068558, "eval_precision_Quebrado_16x16": 0.8408071748878924, "eval_precision_RiadoRio_16x16": 0.5427495291902071, "eval_precision_RioFechado_16x16": 0.5053022269353128, "eval_recall": 0.6204784778699383, "eval_recall_DuraRiadoRio_16x16": 0.4874131944444444, "eval_recall_Mole_16x16": 0.5347222222222222, "eval_recall_Quebrado_16x16": 0.6510416666666666, "eval_recall_RiadoRio_16x16": 0.592516447368421, "eval_recall_RioFechado_16x16": 0.8366988586479368, "eval_runtime": 45.9779, "eval_samples_per_second": 252.773, "eval_steps_per_second": 15.812, "step": 389404 }, { "epoch": 67.00791465932554, "grad_norm": 10.786299705505371, "learning_rate": 1.6496042670337234e-05, "loss": 0.7427, "step": 389450 }, { "epoch": 67.01651754989676, "grad_norm": 16.77695083618164, "learning_rate": 1.6491741225051618e-05, "loss": 0.755, "step": 389500 }, { "epoch": 67.025120440468, "grad_norm": 14.881721496582031, "learning_rate": 1.6487439779766003e-05, "loss": 0.6814, "step": 389550 }, { "epoch": 67.03372333103923, "grad_norm": 14.086237907409668, "learning_rate": 1.6483138334480387e-05, "loss": 0.6722, "step": 389600 }, { "epoch": 67.04232622161047, "grad_norm": 19.04656982421875, "learning_rate": 1.6478836889194772e-05, "loss": 0.7402, "step": 389650 }, { "epoch": 67.05092911218169, "grad_norm": 15.18547534942627, "learning_rate": 1.6474535443909153e-05, "loss": 0.6563, "step": 389700 }, { "epoch": 67.05953200275293, "grad_norm": 19.45865249633789, "learning_rate": 1.6470233998623538e-05, "loss": 0.7138, "step": 389750 }, { "epoch": 67.06813489332416, "grad_norm": 16.659381866455078, "learning_rate": 1.6465932553337922e-05, "loss": 0.7077, "step": 389800 }, { "epoch": 67.07673778389538, "grad_norm": 12.438114166259766, "learning_rate": 1.6461631108052307e-05, "loss": 0.7193, "step": 389850 }, { "epoch": 67.08534067446662, "grad_norm": 13.389045715332031, "learning_rate": 1.645732966276669e-05, "loss": 0.6503, "step": 389900 }, { "epoch": 67.09394356503785, "grad_norm": 18.566207885742188, "learning_rate": 1.6453028217481076e-05, "loss": 0.7049, "step": 389950 }, { "epoch": 67.10254645560909, "grad_norm": 20.643070220947266, "learning_rate": 1.6448726772195457e-05, "loss": 0.7369, "step": 390000 }, { "epoch": 67.11114934618031, "grad_norm": 9.214631080627441, "learning_rate": 1.644442532690984e-05, "loss": 0.6836, "step": 390050 }, { "epoch": 67.11975223675155, "grad_norm": 15.333890914916992, "learning_rate": 1.6440123881624226e-05, "loss": 0.7123, "step": 390100 }, { "epoch": 67.12835512732278, "grad_norm": 12.121803283691406, "learning_rate": 1.643582243633861e-05, "loss": 0.746, "step": 390150 }, { "epoch": 67.13695801789402, "grad_norm": 8.206504821777344, "learning_rate": 1.6431520991052995e-05, "loss": 0.7357, "step": 390200 }, { "epoch": 67.14556090846524, "grad_norm": 15.974823951721191, "learning_rate": 1.6427219545767376e-05, "loss": 0.6936, "step": 390250 }, { "epoch": 67.15416379903648, "grad_norm": 16.828886032104492, "learning_rate": 1.6422918100481764e-05, "loss": 0.6451, "step": 390300 }, { "epoch": 67.16276668960771, "grad_norm": 12.524911880493164, "learning_rate": 1.641861665519615e-05, "loss": 0.7463, "step": 390350 }, { "epoch": 67.17136958017895, "grad_norm": 16.300968170166016, "learning_rate": 1.641431520991053e-05, "loss": 0.6952, "step": 390400 }, { "epoch": 67.17997247075017, "grad_norm": 10.210455894470215, "learning_rate": 1.6410013764624915e-05, "loss": 0.7339, "step": 390450 }, { "epoch": 67.1885753613214, "grad_norm": 12.780023574829102, "learning_rate": 1.64057123193393e-05, "loss": 0.6917, "step": 390500 }, { "epoch": 67.19717825189264, "grad_norm": 14.754765510559082, "learning_rate": 1.6401410874053684e-05, "loss": 0.7098, "step": 390550 }, { "epoch": 67.20578114246386, "grad_norm": 11.58059024810791, "learning_rate": 1.6397109428768068e-05, "loss": 0.728, "step": 390600 }, { "epoch": 67.2143840330351, "grad_norm": 10.344597816467285, "learning_rate": 1.639280798348245e-05, "loss": 0.6865, "step": 390650 }, { "epoch": 67.22298692360633, "grad_norm": 17.79491424560547, "learning_rate": 1.6388506538196834e-05, "loss": 0.7317, "step": 390700 }, { "epoch": 67.23158981417757, "grad_norm": 18.065006256103516, "learning_rate": 1.638420509291122e-05, "loss": 0.6977, "step": 390750 }, { "epoch": 67.2401927047488, "grad_norm": 14.62398624420166, "learning_rate": 1.6379903647625603e-05, "loss": 0.6642, "step": 390800 }, { "epoch": 67.24879559532003, "grad_norm": 13.421514511108398, "learning_rate": 1.6375602202339988e-05, "loss": 0.7083, "step": 390850 }, { "epoch": 67.25739848589126, "grad_norm": 15.651066780090332, "learning_rate": 1.637130075705437e-05, "loss": 0.6635, "step": 390900 }, { "epoch": 67.2660013764625, "grad_norm": 10.272388458251953, "learning_rate": 1.6366999311768753e-05, "loss": 0.7279, "step": 390950 }, { "epoch": 67.27460426703372, "grad_norm": 13.625216484069824, "learning_rate": 1.636269786648314e-05, "loss": 0.7118, "step": 391000 }, { "epoch": 67.28320715760495, "grad_norm": 11.43275260925293, "learning_rate": 1.6358396421197523e-05, "loss": 0.7218, "step": 391050 }, { "epoch": 67.29181004817619, "grad_norm": 22.126195907592773, "learning_rate": 1.6354094975911907e-05, "loss": 0.7092, "step": 391100 }, { "epoch": 67.30041293874741, "grad_norm": 14.712627410888672, "learning_rate": 1.634979353062629e-05, "loss": 0.7208, "step": 391150 }, { "epoch": 67.30901582931865, "grad_norm": 23.434326171875, "learning_rate": 1.6345492085340676e-05, "loss": 0.7181, "step": 391200 }, { "epoch": 67.31761871988988, "grad_norm": 26.591157913208008, "learning_rate": 1.634119064005506e-05, "loss": 0.6821, "step": 391250 }, { "epoch": 67.32622161046112, "grad_norm": 10.159591674804688, "learning_rate": 1.6336889194769445e-05, "loss": 0.6935, "step": 391300 }, { "epoch": 67.33482450103234, "grad_norm": 9.744414329528809, "learning_rate": 1.6332587749483827e-05, "loss": 0.7424, "step": 391350 }, { "epoch": 67.34342739160358, "grad_norm": 11.58383846282959, "learning_rate": 1.632828630419821e-05, "loss": 0.6988, "step": 391400 }, { "epoch": 67.35203028217481, "grad_norm": 21.229053497314453, "learning_rate": 1.6323984858912596e-05, "loss": 0.7178, "step": 391450 }, { "epoch": 67.36063317274605, "grad_norm": 12.378365516662598, "learning_rate": 1.631968341362698e-05, "loss": 0.7198, "step": 391500 }, { "epoch": 67.36923606331727, "grad_norm": 8.20517635345459, "learning_rate": 1.6315381968341365e-05, "loss": 0.6932, "step": 391550 }, { "epoch": 67.3778389538885, "grad_norm": 20.47562599182129, "learning_rate": 1.6311080523055746e-05, "loss": 0.7298, "step": 391600 }, { "epoch": 67.38644184445974, "grad_norm": 13.907468795776367, "learning_rate": 1.630677907777013e-05, "loss": 0.6997, "step": 391650 }, { "epoch": 67.39504473503096, "grad_norm": 14.941664695739746, "learning_rate": 1.630247763248452e-05, "loss": 0.7153, "step": 391700 }, { "epoch": 67.4036476256022, "grad_norm": 15.635946273803711, "learning_rate": 1.62981761871989e-05, "loss": 0.6943, "step": 391750 }, { "epoch": 67.41225051617343, "grad_norm": 13.32741928100586, "learning_rate": 1.6293874741913284e-05, "loss": 0.6598, "step": 391800 }, { "epoch": 67.42085340674467, "grad_norm": 7.510952472686768, "learning_rate": 1.6289573296627665e-05, "loss": 0.7071, "step": 391850 }, { "epoch": 67.4294562973159, "grad_norm": 12.044915199279785, "learning_rate": 1.628527185134205e-05, "loss": 0.759, "step": 391900 }, { "epoch": 67.43805918788713, "grad_norm": 16.203025817871094, "learning_rate": 1.6280970406056438e-05, "loss": 0.714, "step": 391950 }, { "epoch": 67.44666207845836, "grad_norm": 18.7772274017334, "learning_rate": 1.627666896077082e-05, "loss": 0.7034, "step": 392000 }, { "epoch": 67.4552649690296, "grad_norm": 11.970983505249023, "learning_rate": 1.6272367515485204e-05, "loss": 0.7366, "step": 392050 }, { "epoch": 67.46386785960082, "grad_norm": 19.348360061645508, "learning_rate": 1.6268066070199588e-05, "loss": 0.6802, "step": 392100 }, { "epoch": 67.47247075017206, "grad_norm": 13.533859252929688, "learning_rate": 1.6263764624913973e-05, "loss": 0.7218, "step": 392150 }, { "epoch": 67.48107364074329, "grad_norm": 18.099620819091797, "learning_rate": 1.6259463179628357e-05, "loss": 0.7598, "step": 392200 }, { "epoch": 67.48967653131452, "grad_norm": 12.10698127746582, "learning_rate": 1.625516173434274e-05, "loss": 0.7144, "step": 392250 }, { "epoch": 67.49827942188575, "grad_norm": 16.98116683959961, "learning_rate": 1.6250860289057123e-05, "loss": 0.7375, "step": 392300 }, { "epoch": 67.50688231245698, "grad_norm": 9.14737606048584, "learning_rate": 1.6246558843771508e-05, "loss": 0.7111, "step": 392350 }, { "epoch": 67.51548520302822, "grad_norm": 14.104975700378418, "learning_rate": 1.6242257398485892e-05, "loss": 0.7112, "step": 392400 }, { "epoch": 67.52408809359945, "grad_norm": 9.884883880615234, "learning_rate": 1.6237955953200277e-05, "loss": 0.6535, "step": 392450 }, { "epoch": 67.53269098417069, "grad_norm": 15.766714096069336, "learning_rate": 1.623365450791466e-05, "loss": 0.6928, "step": 392500 }, { "epoch": 67.54129387474191, "grad_norm": 10.849631309509277, "learning_rate": 1.6229353062629042e-05, "loss": 0.7438, "step": 392550 }, { "epoch": 67.54989676531315, "grad_norm": 11.48650074005127, "learning_rate": 1.6225051617343427e-05, "loss": 0.7507, "step": 392600 }, { "epoch": 67.55849965588438, "grad_norm": 20.26283073425293, "learning_rate": 1.622075017205781e-05, "loss": 0.7373, "step": 392650 }, { "epoch": 67.56710254645562, "grad_norm": 11.230987548828125, "learning_rate": 1.6216448726772196e-05, "loss": 0.7461, "step": 392700 }, { "epoch": 67.57570543702684, "grad_norm": 14.98019790649414, "learning_rate": 1.621214728148658e-05, "loss": 0.6728, "step": 392750 }, { "epoch": 67.58430832759807, "grad_norm": 12.992609977722168, "learning_rate": 1.6207845836200962e-05, "loss": 0.7203, "step": 392800 }, { "epoch": 67.5929112181693, "grad_norm": 10.32664680480957, "learning_rate": 1.620354439091535e-05, "loss": 0.6685, "step": 392850 }, { "epoch": 67.60151410874053, "grad_norm": 11.946813583374023, "learning_rate": 1.6199242945629734e-05, "loss": 0.751, "step": 392900 }, { "epoch": 67.61011699931177, "grad_norm": 21.921297073364258, "learning_rate": 1.6194941500344116e-05, "loss": 0.7516, "step": 392950 }, { "epoch": 67.618719889883, "grad_norm": 10.91167163848877, "learning_rate": 1.61906400550585e-05, "loss": 0.7003, "step": 393000 }, { "epoch": 67.62732278045424, "grad_norm": 20.35734748840332, "learning_rate": 1.6186338609772885e-05, "loss": 0.7321, "step": 393050 }, { "epoch": 67.63592567102546, "grad_norm": 24.554492950439453, "learning_rate": 1.618203716448727e-05, "loss": 0.7008, "step": 393100 }, { "epoch": 67.6445285615967, "grad_norm": 9.262784957885742, "learning_rate": 1.6177735719201654e-05, "loss": 0.7245, "step": 393150 }, { "epoch": 67.65313145216793, "grad_norm": 22.106294631958008, "learning_rate": 1.6173434273916035e-05, "loss": 0.7394, "step": 393200 }, { "epoch": 67.66173434273917, "grad_norm": 13.955140113830566, "learning_rate": 1.616913282863042e-05, "loss": 0.7264, "step": 393250 }, { "epoch": 67.67033723331039, "grad_norm": 12.673361778259277, "learning_rate": 1.6164831383344804e-05, "loss": 0.7129, "step": 393300 }, { "epoch": 67.67894012388163, "grad_norm": 9.015937805175781, "learning_rate": 1.616052993805919e-05, "loss": 0.7007, "step": 393350 }, { "epoch": 67.68754301445286, "grad_norm": 13.586917877197266, "learning_rate": 1.6156228492773573e-05, "loss": 0.6738, "step": 393400 }, { "epoch": 67.69614590502408, "grad_norm": 13.107163429260254, "learning_rate": 1.6151927047487958e-05, "loss": 0.7329, "step": 393450 }, { "epoch": 67.70474879559532, "grad_norm": 10.300867080688477, "learning_rate": 1.614762560220234e-05, "loss": 0.687, "step": 393500 }, { "epoch": 67.71335168616655, "grad_norm": 17.5313720703125, "learning_rate": 1.6143324156916727e-05, "loss": 0.74, "step": 393550 }, { "epoch": 67.72195457673779, "grad_norm": 15.511218070983887, "learning_rate": 1.6139022711631108e-05, "loss": 0.7188, "step": 393600 }, { "epoch": 67.73055746730901, "grad_norm": 17.517425537109375, "learning_rate": 1.6134721266345493e-05, "loss": 0.6663, "step": 393650 }, { "epoch": 67.73916035788025, "grad_norm": 17.431442260742188, "learning_rate": 1.6130419821059877e-05, "loss": 0.7192, "step": 393700 }, { "epoch": 67.74776324845148, "grad_norm": 14.840222358703613, "learning_rate": 1.6126118375774262e-05, "loss": 0.7266, "step": 393750 }, { "epoch": 67.75636613902272, "grad_norm": 12.026097297668457, "learning_rate": 1.6121816930488646e-05, "loss": 0.7246, "step": 393800 }, { "epoch": 67.76496902959394, "grad_norm": 19.77341651916504, "learning_rate": 1.611751548520303e-05, "loss": 0.6757, "step": 393850 }, { "epoch": 67.77357192016518, "grad_norm": 19.977581024169922, "learning_rate": 1.6113214039917412e-05, "loss": 0.7117, "step": 393900 }, { "epoch": 67.78217481073641, "grad_norm": 9.561525344848633, "learning_rate": 1.6108912594631797e-05, "loss": 0.6661, "step": 393950 }, { "epoch": 67.79077770130763, "grad_norm": 19.727123260498047, "learning_rate": 1.610461114934618e-05, "loss": 0.7581, "step": 394000 }, { "epoch": 67.79938059187887, "grad_norm": 12.658164024353027, "learning_rate": 1.6100309704060566e-05, "loss": 0.7342, "step": 394050 }, { "epoch": 67.8079834824501, "grad_norm": 8.105978012084961, "learning_rate": 1.609600825877495e-05, "loss": 0.7412, "step": 394100 }, { "epoch": 67.81658637302134, "grad_norm": 15.640870094299316, "learning_rate": 1.609170681348933e-05, "loss": 0.7061, "step": 394150 }, { "epoch": 67.82518926359256, "grad_norm": 19.641014099121094, "learning_rate": 1.6087405368203716e-05, "loss": 0.7258, "step": 394200 }, { "epoch": 67.8337921541638, "grad_norm": 22.11599349975586, "learning_rate": 1.6083103922918104e-05, "loss": 0.6857, "step": 394250 }, { "epoch": 67.84239504473503, "grad_norm": 24.183547973632812, "learning_rate": 1.6078802477632485e-05, "loss": 0.7105, "step": 394300 }, { "epoch": 67.85099793530627, "grad_norm": 24.460163116455078, "learning_rate": 1.607450103234687e-05, "loss": 0.7553, "step": 394350 }, { "epoch": 67.8596008258775, "grad_norm": 18.452640533447266, "learning_rate": 1.607019958706125e-05, "loss": 0.6795, "step": 394400 }, { "epoch": 67.86820371644873, "grad_norm": 13.04937744140625, "learning_rate": 1.606589814177564e-05, "loss": 0.7154, "step": 394450 }, { "epoch": 67.87680660701996, "grad_norm": 14.165841102600098, "learning_rate": 1.6061596696490023e-05, "loss": 0.7011, "step": 394500 }, { "epoch": 67.88540949759118, "grad_norm": 17.68707847595215, "learning_rate": 1.6057295251204405e-05, "loss": 0.7331, "step": 394550 }, { "epoch": 67.89401238816242, "grad_norm": 22.831525802612305, "learning_rate": 1.605299380591879e-05, "loss": 0.7059, "step": 394600 }, { "epoch": 67.90261527873365, "grad_norm": 21.56911849975586, "learning_rate": 1.6048692360633174e-05, "loss": 0.7637, "step": 394650 }, { "epoch": 67.91121816930489, "grad_norm": 11.053203582763672, "learning_rate": 1.6044390915347558e-05, "loss": 0.7227, "step": 394700 }, { "epoch": 67.91982105987611, "grad_norm": 16.5186824798584, "learning_rate": 1.6040089470061943e-05, "loss": 0.698, "step": 394750 }, { "epoch": 67.92842395044735, "grad_norm": 25.213211059570312, "learning_rate": 1.6035788024776324e-05, "loss": 0.7905, "step": 394800 }, { "epoch": 67.93702684101858, "grad_norm": 12.87101936340332, "learning_rate": 1.603148657949071e-05, "loss": 0.7015, "step": 394850 }, { "epoch": 67.94562973158982, "grad_norm": 15.54085922241211, "learning_rate": 1.6027185134205093e-05, "loss": 0.6792, "step": 394900 }, { "epoch": 67.95423262216104, "grad_norm": 15.5801362991333, "learning_rate": 1.6022883688919478e-05, "loss": 0.6887, "step": 394950 }, { "epoch": 67.96283551273228, "grad_norm": 17.678050994873047, "learning_rate": 1.6018582243633862e-05, "loss": 0.6954, "step": 395000 }, { "epoch": 67.97143840330351, "grad_norm": 12.264019012451172, "learning_rate": 1.6014280798348247e-05, "loss": 0.7361, "step": 395050 }, { "epoch": 67.98004129387475, "grad_norm": 9.427718162536621, "learning_rate": 1.6009979353062628e-05, "loss": 0.7213, "step": 395100 }, { "epoch": 67.98864418444597, "grad_norm": 14.913168907165527, "learning_rate": 1.6005677907777012e-05, "loss": 0.6973, "step": 395150 }, { "epoch": 67.9972470750172, "grad_norm": 16.102584838867188, "learning_rate": 1.60013764624914e-05, "loss": 0.7307, "step": 395200 }, { "epoch": 68.0, "eval_accuracy": 0.5647048700739976, "eval_f1": 0.5428626843609914, "eval_f1_DuraRiadoRio_16x16": 0.547486033519553, "eval_f1_Mole_16x16": 0.506268656716418, "eval_f1_Quebrado_16x16": 0.7468253968253968, "eval_f1_RiadoRio_16x16": 0.33480176211453744, "eval_f1_RioFechado_16x16": 0.5789315726290516, "eval_loss": 2.4151573181152344, "eval_precision": 0.6316330273489332, "eval_precision_DuraRiadoRio_16x16": 0.5066469719350074, "eval_precision_Mole_16x16": 0.8107074569789675, "eval_precision_Quebrado_16x16": 0.6878654970760234, "eval_precision_RiadoRio_16x16": 0.7131367292225201, "eval_precision_RioFechado_16x16": 0.43980848153214774, "eval_recall": 0.5691854758072383, "eval_recall_DuraRiadoRio_16x16": 0.5954861111111112, "eval_recall_Mole_16x16": 0.3680555555555556, "eval_recall_Quebrado_16x16": 0.8168402777777778, "eval_recall_RiadoRio_16x16": 0.21875, "eval_recall_RioFechado_16x16": 0.8467954345917471, "eval_runtime": 45.6437, "eval_samples_per_second": 254.624, "eval_steps_per_second": 15.928, "step": 395216 }, { "epoch": 68.00584996558844, "grad_norm": 13.607502937316895, "learning_rate": 1.599707501720578e-05, "loss": 0.7052, "step": 395250 }, { "epoch": 68.01445285615966, "grad_norm": 15.946453094482422, "learning_rate": 1.5992773571920166e-05, "loss": 0.7356, "step": 395300 }, { "epoch": 68.0230557467309, "grad_norm": 20.23589515686035, "learning_rate": 1.5988472126634547e-05, "loss": 0.6989, "step": 395350 }, { "epoch": 68.03165863730213, "grad_norm": 24.38607406616211, "learning_rate": 1.5984170681348935e-05, "loss": 0.6344, "step": 395400 }, { "epoch": 68.04026152787337, "grad_norm": 14.063568115234375, "learning_rate": 1.597986923606332e-05, "loss": 0.6574, "step": 395450 }, { "epoch": 68.0488644184446, "grad_norm": 18.790515899658203, "learning_rate": 1.59755677907777e-05, "loss": 0.6882, "step": 395500 }, { "epoch": 68.05746730901583, "grad_norm": 14.866527557373047, "learning_rate": 1.5971266345492086e-05, "loss": 0.7094, "step": 395550 }, { "epoch": 68.06607019958706, "grad_norm": 17.488563537597656, "learning_rate": 1.596696490020647e-05, "loss": 0.7059, "step": 395600 }, { "epoch": 68.0746730901583, "grad_norm": 10.61436939239502, "learning_rate": 1.5962663454920855e-05, "loss": 0.6879, "step": 395650 }, { "epoch": 68.08327598072952, "grad_norm": 20.983606338500977, "learning_rate": 1.595836200963524e-05, "loss": 0.7365, "step": 395700 }, { "epoch": 68.09187887130075, "grad_norm": 13.771289825439453, "learning_rate": 1.595406056434962e-05, "loss": 0.7192, "step": 395750 }, { "epoch": 68.10048176187199, "grad_norm": 22.30669403076172, "learning_rate": 1.5949759119064005e-05, "loss": 0.7622, "step": 395800 }, { "epoch": 68.10908465244322, "grad_norm": 11.465353012084961, "learning_rate": 1.594545767377839e-05, "loss": 0.7666, "step": 395850 }, { "epoch": 68.11768754301445, "grad_norm": 16.2455997467041, "learning_rate": 1.5941156228492774e-05, "loss": 0.7003, "step": 395900 }, { "epoch": 68.12629043358568, "grad_norm": 10.304156303405762, "learning_rate": 1.593685478320716e-05, "loss": 0.6485, "step": 395950 }, { "epoch": 68.13489332415692, "grad_norm": 10.125899314880371, "learning_rate": 1.5932553337921543e-05, "loss": 0.6862, "step": 396000 }, { "epoch": 68.14349621472815, "grad_norm": 10.214363098144531, "learning_rate": 1.5928251892635924e-05, "loss": 0.6961, "step": 396050 }, { "epoch": 68.15209910529938, "grad_norm": 19.908191680908203, "learning_rate": 1.5923950447350312e-05, "loss": 0.7644, "step": 396100 }, { "epoch": 68.16070199587061, "grad_norm": 11.133197784423828, "learning_rate": 1.5919649002064694e-05, "loss": 0.7154, "step": 396150 }, { "epoch": 68.16930488644185, "grad_norm": 8.884573936462402, "learning_rate": 1.5915347556779078e-05, "loss": 0.6892, "step": 396200 }, { "epoch": 68.17790777701308, "grad_norm": 25.57292938232422, "learning_rate": 1.5911046111493463e-05, "loss": 0.7267, "step": 396250 }, { "epoch": 68.18651066758432, "grad_norm": 21.386770248413086, "learning_rate": 1.5906744666207847e-05, "loss": 0.6716, "step": 396300 }, { "epoch": 68.19511355815554, "grad_norm": 14.297759056091309, "learning_rate": 1.5902443220922232e-05, "loss": 0.6825, "step": 396350 }, { "epoch": 68.20371644872677, "grad_norm": 19.77266502380371, "learning_rate": 1.5898141775636616e-05, "loss": 0.7399, "step": 396400 }, { "epoch": 68.212319339298, "grad_norm": 10.0337495803833, "learning_rate": 1.5893840330350997e-05, "loss": 0.7, "step": 396450 }, { "epoch": 68.22092222986923, "grad_norm": 26.73702621459961, "learning_rate": 1.5889538885065382e-05, "loss": 0.6633, "step": 396500 }, { "epoch": 68.22952512044047, "grad_norm": 20.30940055847168, "learning_rate": 1.5885237439779767e-05, "loss": 0.7407, "step": 396550 }, { "epoch": 68.2381280110117, "grad_norm": 15.02544116973877, "learning_rate": 1.588093599449415e-05, "loss": 0.7067, "step": 396600 }, { "epoch": 68.24673090158294, "grad_norm": 13.479031562805176, "learning_rate": 1.5876634549208536e-05, "loss": 0.6459, "step": 396650 }, { "epoch": 68.25533379215416, "grad_norm": 15.200613975524902, "learning_rate": 1.5872333103922917e-05, "loss": 0.6272, "step": 396700 }, { "epoch": 68.2639366827254, "grad_norm": 9.059954643249512, "learning_rate": 1.58680316586373e-05, "loss": 0.7303, "step": 396750 }, { "epoch": 68.27253957329663, "grad_norm": 26.416194915771484, "learning_rate": 1.586373021335169e-05, "loss": 0.7089, "step": 396800 }, { "epoch": 68.28114246386787, "grad_norm": 19.31279182434082, "learning_rate": 1.585942876806607e-05, "loss": 0.7799, "step": 396850 }, { "epoch": 68.28974535443909, "grad_norm": 13.095500946044922, "learning_rate": 1.5855127322780455e-05, "loss": 0.6531, "step": 396900 }, { "epoch": 68.29834824501032, "grad_norm": 12.132549285888672, "learning_rate": 1.585082587749484e-05, "loss": 0.7229, "step": 396950 }, { "epoch": 68.30695113558156, "grad_norm": 11.541504859924316, "learning_rate": 1.5846524432209224e-05, "loss": 0.648, "step": 397000 }, { "epoch": 68.31555402615278, "grad_norm": 15.60581111907959, "learning_rate": 1.584222298692361e-05, "loss": 0.6958, "step": 397050 }, { "epoch": 68.32415691672402, "grad_norm": 11.67555046081543, "learning_rate": 1.583792154163799e-05, "loss": 0.7128, "step": 397100 }, { "epoch": 68.33275980729525, "grad_norm": 12.605448722839355, "learning_rate": 1.5833620096352375e-05, "loss": 0.6502, "step": 397150 }, { "epoch": 68.34136269786649, "grad_norm": 15.909997940063477, "learning_rate": 1.582931865106676e-05, "loss": 0.6975, "step": 397200 }, { "epoch": 68.34996558843771, "grad_norm": 16.544662475585938, "learning_rate": 1.5825017205781144e-05, "loss": 0.7121, "step": 397250 }, { "epoch": 68.35856847900895, "grad_norm": 16.413755416870117, "learning_rate": 1.5820715760495528e-05, "loss": 0.7327, "step": 397300 }, { "epoch": 68.36717136958018, "grad_norm": 17.169052124023438, "learning_rate": 1.5816414315209913e-05, "loss": 0.6638, "step": 397350 }, { "epoch": 68.37577426015142, "grad_norm": 11.2868070602417, "learning_rate": 1.5812112869924294e-05, "loss": 0.6713, "step": 397400 }, { "epoch": 68.38437715072264, "grad_norm": 13.506439208984375, "learning_rate": 1.580781142463868e-05, "loss": 0.7532, "step": 397450 }, { "epoch": 68.39298004129388, "grad_norm": 16.680580139160156, "learning_rate": 1.5803509979353063e-05, "loss": 0.6855, "step": 397500 }, { "epoch": 68.40158293186511, "grad_norm": 15.51053237915039, "learning_rate": 1.5799208534067448e-05, "loss": 0.7318, "step": 397550 }, { "epoch": 68.41018582243633, "grad_norm": 18.944774627685547, "learning_rate": 1.5794907088781832e-05, "loss": 0.7143, "step": 397600 }, { "epoch": 68.41878871300757, "grad_norm": 15.136820793151855, "learning_rate": 1.5790605643496213e-05, "loss": 0.6803, "step": 397650 }, { "epoch": 68.4273916035788, "grad_norm": 20.962448120117188, "learning_rate": 1.57863041982106e-05, "loss": 0.7127, "step": 397700 }, { "epoch": 68.43599449415004, "grad_norm": 13.209447860717773, "learning_rate": 1.5782002752924986e-05, "loss": 0.7409, "step": 397750 }, { "epoch": 68.44459738472126, "grad_norm": 14.116270065307617, "learning_rate": 1.5777701307639367e-05, "loss": 0.7238, "step": 397800 }, { "epoch": 68.4532002752925, "grad_norm": 16.414630889892578, "learning_rate": 1.577339986235375e-05, "loss": 0.743, "step": 397850 }, { "epoch": 68.46180316586373, "grad_norm": 15.686582565307617, "learning_rate": 1.5769098417068133e-05, "loss": 0.7126, "step": 397900 }, { "epoch": 68.47040605643497, "grad_norm": 11.369527816772461, "learning_rate": 1.576479697178252e-05, "loss": 0.7472, "step": 397950 }, { "epoch": 68.4790089470062, "grad_norm": 11.859050750732422, "learning_rate": 1.5760495526496905e-05, "loss": 0.7167, "step": 398000 }, { "epoch": 68.48761183757743, "grad_norm": 18.534137725830078, "learning_rate": 1.5756194081211286e-05, "loss": 0.6876, "step": 398050 }, { "epoch": 68.49621472814866, "grad_norm": 12.031258583068848, "learning_rate": 1.575189263592567e-05, "loss": 0.7149, "step": 398100 }, { "epoch": 68.50481761871988, "grad_norm": 10.60220718383789, "learning_rate": 1.5747591190640056e-05, "loss": 0.7188, "step": 398150 }, { "epoch": 68.51342050929112, "grad_norm": 10.076444625854492, "learning_rate": 1.574328974535444e-05, "loss": 0.7346, "step": 398200 }, { "epoch": 68.52202339986235, "grad_norm": 14.902754783630371, "learning_rate": 1.5738988300068825e-05, "loss": 0.6955, "step": 398250 }, { "epoch": 68.53062629043359, "grad_norm": 9.298827171325684, "learning_rate": 1.5734686854783206e-05, "loss": 0.6632, "step": 398300 }, { "epoch": 68.53922918100481, "grad_norm": 13.635764122009277, "learning_rate": 1.573038540949759e-05, "loss": 0.7148, "step": 398350 }, { "epoch": 68.54783207157605, "grad_norm": 14.144612312316895, "learning_rate": 1.572608396421198e-05, "loss": 0.6934, "step": 398400 }, { "epoch": 68.55643496214728, "grad_norm": 24.058639526367188, "learning_rate": 1.572178251892636e-05, "loss": 0.6496, "step": 398450 }, { "epoch": 68.56503785271852, "grad_norm": 12.433920860290527, "learning_rate": 1.5717481073640744e-05, "loss": 0.7733, "step": 398500 }, { "epoch": 68.57364074328974, "grad_norm": 9.8738431930542, "learning_rate": 1.571317962835513e-05, "loss": 0.6995, "step": 398550 }, { "epoch": 68.58224363386098, "grad_norm": 5.172061443328857, "learning_rate": 1.570887818306951e-05, "loss": 0.6599, "step": 398600 }, { "epoch": 68.59084652443221, "grad_norm": 13.09598159790039, "learning_rate": 1.5704576737783898e-05, "loss": 0.7218, "step": 398650 }, { "epoch": 68.59944941500343, "grad_norm": 16.3533992767334, "learning_rate": 1.5700275292498282e-05, "loss": 0.66, "step": 398700 }, { "epoch": 68.60805230557467, "grad_norm": 12.780060768127441, "learning_rate": 1.5695973847212664e-05, "loss": 0.7202, "step": 398750 }, { "epoch": 68.6166551961459, "grad_norm": 13.114072799682617, "learning_rate": 1.5691672401927048e-05, "loss": 0.7536, "step": 398800 }, { "epoch": 68.62525808671714, "grad_norm": 15.008917808532715, "learning_rate": 1.5687370956641433e-05, "loss": 0.6876, "step": 398850 }, { "epoch": 68.63386097728836, "grad_norm": 15.1415433883667, "learning_rate": 1.5683069511355817e-05, "loss": 0.7051, "step": 398900 }, { "epoch": 68.6424638678596, "grad_norm": 15.9806489944458, "learning_rate": 1.5678768066070202e-05, "loss": 0.7092, "step": 398950 }, { "epoch": 68.65106675843083, "grad_norm": 9.892770767211914, "learning_rate": 1.5674466620784583e-05, "loss": 0.7344, "step": 399000 }, { "epoch": 68.65966964900207, "grad_norm": 15.79366397857666, "learning_rate": 1.5670165175498968e-05, "loss": 0.7031, "step": 399050 }, { "epoch": 68.6682725395733, "grad_norm": 11.77059268951416, "learning_rate": 1.5665863730213352e-05, "loss": 0.7378, "step": 399100 }, { "epoch": 68.67687543014453, "grad_norm": 17.74004554748535, "learning_rate": 1.5661562284927737e-05, "loss": 0.7084, "step": 399150 }, { "epoch": 68.68547832071576, "grad_norm": 18.626379013061523, "learning_rate": 1.565726083964212e-05, "loss": 0.733, "step": 399200 }, { "epoch": 68.694081211287, "grad_norm": 27.38184356689453, "learning_rate": 1.5652959394356502e-05, "loss": 0.6331, "step": 399250 }, { "epoch": 68.70268410185822, "grad_norm": 20.069320678710938, "learning_rate": 1.5648657949070887e-05, "loss": 0.7092, "step": 399300 }, { "epoch": 68.71128699242945, "grad_norm": 18.656108856201172, "learning_rate": 1.5644356503785275e-05, "loss": 0.741, "step": 399350 }, { "epoch": 68.71988988300069, "grad_norm": 18.476844787597656, "learning_rate": 1.5640055058499656e-05, "loss": 0.7149, "step": 399400 }, { "epoch": 68.72849277357192, "grad_norm": 14.04045295715332, "learning_rate": 1.563575361321404e-05, "loss": 0.7222, "step": 399450 }, { "epoch": 68.73709566414315, "grad_norm": 13.083157539367676, "learning_rate": 1.5631452167928425e-05, "loss": 0.7108, "step": 399500 }, { "epoch": 68.74569855471438, "grad_norm": 18.088489532470703, "learning_rate": 1.562715072264281e-05, "loss": 0.7155, "step": 399550 }, { "epoch": 68.75430144528562, "grad_norm": 11.992610931396484, "learning_rate": 1.5622849277357194e-05, "loss": 0.701, "step": 399600 }, { "epoch": 68.76290433585685, "grad_norm": 12.990999221801758, "learning_rate": 1.5618547832071575e-05, "loss": 0.7319, "step": 399650 }, { "epoch": 68.77150722642808, "grad_norm": 13.44513988494873, "learning_rate": 1.561424638678596e-05, "loss": 0.7011, "step": 399700 }, { "epoch": 68.78011011699931, "grad_norm": 16.685714721679688, "learning_rate": 1.5609944941500345e-05, "loss": 0.7268, "step": 399750 }, { "epoch": 68.78871300757055, "grad_norm": 19.076242446899414, "learning_rate": 1.560564349621473e-05, "loss": 0.7213, "step": 399800 }, { "epoch": 68.79731589814178, "grad_norm": 11.473883628845215, "learning_rate": 1.5601342050929114e-05, "loss": 0.7838, "step": 399850 }, { "epoch": 68.805918788713, "grad_norm": 16.983659744262695, "learning_rate": 1.5597040605643498e-05, "loss": 0.6891, "step": 399900 }, { "epoch": 68.81452167928424, "grad_norm": 12.029990196228027, "learning_rate": 1.559273916035788e-05, "loss": 0.6966, "step": 399950 }, { "epoch": 68.82312456985547, "grad_norm": 15.072208404541016, "learning_rate": 1.5588437715072264e-05, "loss": 0.7244, "step": 400000 }, { "epoch": 68.8317274604267, "grad_norm": 17.6702823638916, "learning_rate": 1.558413626978665e-05, "loss": 0.645, "step": 400050 }, { "epoch": 68.84033035099793, "grad_norm": 8.728684425354004, "learning_rate": 1.5579834824501033e-05, "loss": 0.8011, "step": 400100 }, { "epoch": 68.84893324156917, "grad_norm": 22.000810623168945, "learning_rate": 1.5575533379215418e-05, "loss": 0.7726, "step": 400150 }, { "epoch": 68.8575361321404, "grad_norm": 16.26759147644043, "learning_rate": 1.55712319339298e-05, "loss": 0.7324, "step": 400200 }, { "epoch": 68.86613902271164, "grad_norm": 18.6248779296875, "learning_rate": 1.5566930488644187e-05, "loss": 0.7346, "step": 400250 }, { "epoch": 68.87474191328286, "grad_norm": 14.985101699829102, "learning_rate": 1.556262904335857e-05, "loss": 0.7585, "step": 400300 }, { "epoch": 68.8833448038541, "grad_norm": 17.184749603271484, "learning_rate": 1.5558327598072953e-05, "loss": 0.7165, "step": 400350 }, { "epoch": 68.89194769442533, "grad_norm": 20.241792678833008, "learning_rate": 1.5554026152787337e-05, "loss": 0.6777, "step": 400400 }, { "epoch": 68.90055058499657, "grad_norm": 14.412725448608398, "learning_rate": 1.5549724707501718e-05, "loss": 0.7181, "step": 400450 }, { "epoch": 68.90915347556779, "grad_norm": 10.295876502990723, "learning_rate": 1.5545423262216106e-05, "loss": 0.6841, "step": 400500 }, { "epoch": 68.91775636613902, "grad_norm": 12.575032234191895, "learning_rate": 1.554112181693049e-05, "loss": 0.6854, "step": 400550 }, { "epoch": 68.92635925671026, "grad_norm": 16.673805236816406, "learning_rate": 1.5536820371644872e-05, "loss": 0.6934, "step": 400600 }, { "epoch": 68.93496214728148, "grad_norm": 10.800286293029785, "learning_rate": 1.5532518926359257e-05, "loss": 0.6591, "step": 400650 }, { "epoch": 68.94356503785272, "grad_norm": 8.35517692565918, "learning_rate": 1.552821748107364e-05, "loss": 0.7331, "step": 400700 }, { "epoch": 68.95216792842395, "grad_norm": 21.72614288330078, "learning_rate": 1.5523916035788026e-05, "loss": 0.6444, "step": 400750 }, { "epoch": 68.96077081899519, "grad_norm": 19.749351501464844, "learning_rate": 1.551961459050241e-05, "loss": 0.7593, "step": 400800 }, { "epoch": 68.96937370956641, "grad_norm": 18.456010818481445, "learning_rate": 1.5515313145216795e-05, "loss": 0.6915, "step": 400850 }, { "epoch": 68.97797660013765, "grad_norm": 10.819467544555664, "learning_rate": 1.5511011699931176e-05, "loss": 0.7045, "step": 400900 }, { "epoch": 68.98657949070888, "grad_norm": 15.160398483276367, "learning_rate": 1.5506710254645564e-05, "loss": 0.6401, "step": 400950 }, { "epoch": 68.99518238128012, "grad_norm": 20.097808837890625, "learning_rate": 1.5502408809359945e-05, "loss": 0.731, "step": 401000 }, { "epoch": 69.0, "eval_accuracy": 0.4890724488039924, "eval_f1": 0.4684290778931331, "eval_f1_DuraRiadoRio_16x16": 0.4970414201183432, "eval_f1_Mole_16x16": 0.43648404575556893, "eval_f1_Quebrado_16x16": 0.5538423277791594, "eval_f1_RiadoRio_16x16": 0.32564925590895827, "eval_f1_RioFechado_16x16": 0.5291283399036356, "eval_loss": 3.0934336185455322, "eval_precision": 0.5598609707201899, "eval_precision_DuraRiadoRio_16x16": 0.6102337334175616, "eval_precision_Mole_16x16": 0.712180746561886, "eval_precision_Quebrado_16x16": 0.3881143255489718, "eval_precision_RiadoRio_16x16": 0.5608040201005026, "eval_precision_RioFechado_16x16": 0.527972027972028, "eval_recall": 0.49205027012768976, "eval_recall_DuraRiadoRio_16x16": 0.4192708333333333, "eval_recall_Mole_16x16": 0.3146701388888889, "eval_recall_Quebrado_16x16": 0.9665798611111112, "eval_recall_RiadoRio_16x16": 0.2294407894736842, "eval_recall_RioFechado_16x16": 0.5302897278314311, "eval_runtime": 46.3929, "eval_samples_per_second": 250.512, "eval_steps_per_second": 15.67, "step": 401028 }, { "epoch": 69.00378527185134, "grad_norm": 19.55902099609375, "learning_rate": 1.549810736407433e-05, "loss": 0.7061, "step": 401050 }, { "epoch": 69.01238816242257, "grad_norm": 11.418055534362793, "learning_rate": 1.5493805918788714e-05, "loss": 0.6955, "step": 401100 }, { "epoch": 69.0209910529938, "grad_norm": 13.689240455627441, "learning_rate": 1.5489504473503095e-05, "loss": 0.7212, "step": 401150 }, { "epoch": 69.02959394356503, "grad_norm": 27.58207130432129, "learning_rate": 1.5485203028217483e-05, "loss": 0.7061, "step": 401200 }, { "epoch": 69.03819683413627, "grad_norm": 9.036565780639648, "learning_rate": 1.5480901582931868e-05, "loss": 0.6904, "step": 401250 }, { "epoch": 69.0467997247075, "grad_norm": 20.547988891601562, "learning_rate": 1.547660013764625e-05, "loss": 0.6523, "step": 401300 }, { "epoch": 69.05540261527874, "grad_norm": 14.88006591796875, "learning_rate": 1.5472298692360634e-05, "loss": 0.6885, "step": 401350 }, { "epoch": 69.06400550584996, "grad_norm": 8.882986068725586, "learning_rate": 1.5467997247075018e-05, "loss": 0.6934, "step": 401400 }, { "epoch": 69.0726083964212, "grad_norm": 13.006462097167969, "learning_rate": 1.5463695801789403e-05, "loss": 0.6843, "step": 401450 }, { "epoch": 69.08121128699243, "grad_norm": 9.144857406616211, "learning_rate": 1.5459394356503787e-05, "loss": 0.6773, "step": 401500 }, { "epoch": 69.08981417756367, "grad_norm": 17.57782745361328, "learning_rate": 1.545509291121817e-05, "loss": 0.6988, "step": 401550 }, { "epoch": 69.09841706813489, "grad_norm": 18.834184646606445, "learning_rate": 1.5450791465932553e-05, "loss": 0.7065, "step": 401600 }, { "epoch": 69.10701995870612, "grad_norm": 13.306882858276367, "learning_rate": 1.544649002064694e-05, "loss": 0.719, "step": 401650 }, { "epoch": 69.11562284927736, "grad_norm": 11.412260055541992, "learning_rate": 1.5442188575361322e-05, "loss": 0.7174, "step": 401700 }, { "epoch": 69.12422573984858, "grad_norm": 25.453311920166016, "learning_rate": 1.5437887130075707e-05, "loss": 0.699, "step": 401750 }, { "epoch": 69.13282863041982, "grad_norm": 9.79060173034668, "learning_rate": 1.5433585684790088e-05, "loss": 0.6662, "step": 401800 }, { "epoch": 69.14143152099105, "grad_norm": 21.86773681640625, "learning_rate": 1.5429284239504472e-05, "loss": 0.7212, "step": 401850 }, { "epoch": 69.15003441156229, "grad_norm": 21.54037094116211, "learning_rate": 1.542498279421886e-05, "loss": 0.7456, "step": 401900 }, { "epoch": 69.15863730213351, "grad_norm": 18.340110778808594, "learning_rate": 1.542068134893324e-05, "loss": 0.7067, "step": 401950 }, { "epoch": 69.16724019270475, "grad_norm": 23.86601448059082, "learning_rate": 1.5416379903647626e-05, "loss": 0.7201, "step": 402000 }, { "epoch": 69.17584308327598, "grad_norm": 16.561025619506836, "learning_rate": 1.541207845836201e-05, "loss": 0.7132, "step": 402050 }, { "epoch": 69.18444597384722, "grad_norm": 12.918146133422852, "learning_rate": 1.5407777013076395e-05, "loss": 0.6284, "step": 402100 }, { "epoch": 69.19304886441844, "grad_norm": 15.437684059143066, "learning_rate": 1.540347556779078e-05, "loss": 0.6804, "step": 402150 }, { "epoch": 69.20165175498968, "grad_norm": 10.709959030151367, "learning_rate": 1.539917412250516e-05, "loss": 0.684, "step": 402200 }, { "epoch": 69.21025464556091, "grad_norm": 13.350627899169922, "learning_rate": 1.5394872677219546e-05, "loss": 0.655, "step": 402250 }, { "epoch": 69.21885753613213, "grad_norm": 7.697193145751953, "learning_rate": 1.539057123193393e-05, "loss": 0.731, "step": 402300 }, { "epoch": 69.22746042670337, "grad_norm": 24.701278686523438, "learning_rate": 1.5386269786648315e-05, "loss": 0.7447, "step": 402350 }, { "epoch": 69.2360633172746, "grad_norm": 14.020905494689941, "learning_rate": 1.53819683413627e-05, "loss": 0.6841, "step": 402400 }, { "epoch": 69.24466620784584, "grad_norm": 11.54292106628418, "learning_rate": 1.5377666896077084e-05, "loss": 0.7093, "step": 402450 }, { "epoch": 69.25326909841706, "grad_norm": 16.219327926635742, "learning_rate": 1.5373365450791465e-05, "loss": 0.707, "step": 402500 }, { "epoch": 69.2618719889883, "grad_norm": 11.108430862426758, "learning_rate": 1.536906400550585e-05, "loss": 0.7534, "step": 402550 }, { "epoch": 69.27047487955953, "grad_norm": 17.497713088989258, "learning_rate": 1.5364762560220237e-05, "loss": 0.7552, "step": 402600 }, { "epoch": 69.27907777013077, "grad_norm": 8.839812278747559, "learning_rate": 1.536046111493462e-05, "loss": 0.7075, "step": 402650 }, { "epoch": 69.287680660702, "grad_norm": 10.953859329223633, "learning_rate": 1.5356159669649003e-05, "loss": 0.6637, "step": 402700 }, { "epoch": 69.29628355127323, "grad_norm": 20.928667068481445, "learning_rate": 1.5351858224363384e-05, "loss": 0.7275, "step": 402750 }, { "epoch": 69.30488644184446, "grad_norm": 17.97658920288086, "learning_rate": 1.5347556779077772e-05, "loss": 0.6867, "step": 402800 }, { "epoch": 69.31348933241568, "grad_norm": 13.496051788330078, "learning_rate": 1.5343255333792157e-05, "loss": 0.7801, "step": 402850 }, { "epoch": 69.32209222298692, "grad_norm": 14.742976188659668, "learning_rate": 1.5338953888506538e-05, "loss": 0.6534, "step": 402900 }, { "epoch": 69.33069511355815, "grad_norm": 9.473394393920898, "learning_rate": 1.5334652443220923e-05, "loss": 0.7418, "step": 402950 }, { "epoch": 69.33929800412939, "grad_norm": 11.758743286132812, "learning_rate": 1.5330350997935307e-05, "loss": 0.7048, "step": 403000 }, { "epoch": 69.34790089470062, "grad_norm": 14.037487030029297, "learning_rate": 1.5326049552649692e-05, "loss": 0.6888, "step": 403050 }, { "epoch": 69.35650378527185, "grad_norm": 14.460260391235352, "learning_rate": 1.5321748107364076e-05, "loss": 0.7177, "step": 403100 }, { "epoch": 69.36510667584308, "grad_norm": 13.320086479187012, "learning_rate": 1.5317446662078457e-05, "loss": 0.6883, "step": 403150 }, { "epoch": 69.37370956641432, "grad_norm": 18.889484405517578, "learning_rate": 1.5313145216792842e-05, "loss": 0.6832, "step": 403200 }, { "epoch": 69.38231245698555, "grad_norm": 11.053770065307617, "learning_rate": 1.5308843771507227e-05, "loss": 0.6721, "step": 403250 }, { "epoch": 69.39091534755678, "grad_norm": 22.70392417907715, "learning_rate": 1.530454232622161e-05, "loss": 0.7149, "step": 403300 }, { "epoch": 69.39951823812801, "grad_norm": 19.64607048034668, "learning_rate": 1.5300240880935996e-05, "loss": 0.6876, "step": 403350 }, { "epoch": 69.40812112869925, "grad_norm": 16.741458892822266, "learning_rate": 1.529593943565038e-05, "loss": 0.6663, "step": 403400 }, { "epoch": 69.41672401927048, "grad_norm": 10.582462310791016, "learning_rate": 1.529163799036476e-05, "loss": 0.7114, "step": 403450 }, { "epoch": 69.4253269098417, "grad_norm": 14.949057579040527, "learning_rate": 1.528733654507915e-05, "loss": 0.7309, "step": 403500 }, { "epoch": 69.43392980041294, "grad_norm": 15.0554780960083, "learning_rate": 1.528303509979353e-05, "loss": 0.7425, "step": 403550 }, { "epoch": 69.44253269098417, "grad_norm": 11.823253631591797, "learning_rate": 1.5278733654507915e-05, "loss": 0.6826, "step": 403600 }, { "epoch": 69.4511355815554, "grad_norm": 12.073687553405762, "learning_rate": 1.52744322092223e-05, "loss": 0.7099, "step": 403650 }, { "epoch": 69.45973847212663, "grad_norm": 16.47577667236328, "learning_rate": 1.5270130763936684e-05, "loss": 0.6478, "step": 403700 }, { "epoch": 69.46834136269787, "grad_norm": 11.771547317504883, "learning_rate": 1.526582931865107e-05, "loss": 0.6507, "step": 403750 }, { "epoch": 69.4769442532691, "grad_norm": 13.83454418182373, "learning_rate": 1.5261527873365453e-05, "loss": 0.7432, "step": 403800 }, { "epoch": 69.48554714384034, "grad_norm": 24.336204528808594, "learning_rate": 1.5257226428079834e-05, "loss": 0.7003, "step": 403850 }, { "epoch": 69.49415003441156, "grad_norm": 16.49495506286621, "learning_rate": 1.525292498279422e-05, "loss": 0.7164, "step": 403900 }, { "epoch": 69.5027529249828, "grad_norm": 10.000326156616211, "learning_rate": 1.5248623537508602e-05, "loss": 0.6388, "step": 403950 }, { "epoch": 69.51135581555403, "grad_norm": 14.118398666381836, "learning_rate": 1.5244322092222988e-05, "loss": 0.7062, "step": 404000 }, { "epoch": 69.51995870612525, "grad_norm": 12.685416221618652, "learning_rate": 1.5240020646937373e-05, "loss": 0.65, "step": 404050 }, { "epoch": 69.52856159669649, "grad_norm": 6.642924785614014, "learning_rate": 1.5235719201651754e-05, "loss": 0.6858, "step": 404100 }, { "epoch": 69.53716448726772, "grad_norm": 21.029041290283203, "learning_rate": 1.523141775636614e-05, "loss": 0.6735, "step": 404150 }, { "epoch": 69.54576737783896, "grad_norm": 18.29841423034668, "learning_rate": 1.5227116311080525e-05, "loss": 0.6623, "step": 404200 }, { "epoch": 69.55437026841018, "grad_norm": 9.932893753051758, "learning_rate": 1.5222814865794908e-05, "loss": 0.7004, "step": 404250 }, { "epoch": 69.56297315898142, "grad_norm": 17.94831085205078, "learning_rate": 1.5218513420509292e-05, "loss": 0.7163, "step": 404300 }, { "epoch": 69.57157604955265, "grad_norm": 14.8526029586792, "learning_rate": 1.5214211975223677e-05, "loss": 0.6733, "step": 404350 }, { "epoch": 69.58017894012389, "grad_norm": 10.520980834960938, "learning_rate": 1.520991052993806e-05, "loss": 0.726, "step": 404400 }, { "epoch": 69.58878183069511, "grad_norm": 19.853660583496094, "learning_rate": 1.5205609084652444e-05, "loss": 0.6868, "step": 404450 }, { "epoch": 69.59738472126635, "grad_norm": 14.009587287902832, "learning_rate": 1.5201307639366827e-05, "loss": 0.6742, "step": 404500 }, { "epoch": 69.60598761183758, "grad_norm": 10.555907249450684, "learning_rate": 1.5197006194081212e-05, "loss": 0.7121, "step": 404550 }, { "epoch": 69.61459050240882, "grad_norm": 11.417182922363281, "learning_rate": 1.5192704748795598e-05, "loss": 0.7278, "step": 404600 }, { "epoch": 69.62319339298004, "grad_norm": 8.693785667419434, "learning_rate": 1.5188403303509979e-05, "loss": 0.6268, "step": 404650 }, { "epoch": 69.63179628355127, "grad_norm": 20.834081649780273, "learning_rate": 1.5184101858224364e-05, "loss": 0.7264, "step": 404700 }, { "epoch": 69.6403991741225, "grad_norm": 15.433826446533203, "learning_rate": 1.517980041293875e-05, "loss": 0.6936, "step": 404750 }, { "epoch": 69.64900206469373, "grad_norm": 12.440890312194824, "learning_rate": 1.5175498967653131e-05, "loss": 0.7526, "step": 404800 }, { "epoch": 69.65760495526497, "grad_norm": 13.552812576293945, "learning_rate": 1.5171197522367517e-05, "loss": 0.717, "step": 404850 }, { "epoch": 69.6662078458362, "grad_norm": 13.53485107421875, "learning_rate": 1.5166896077081898e-05, "loss": 0.6778, "step": 404900 }, { "epoch": 69.67481073640744, "grad_norm": 21.940935134887695, "learning_rate": 1.5162594631796285e-05, "loss": 0.7061, "step": 404950 }, { "epoch": 69.68341362697866, "grad_norm": 10.042893409729004, "learning_rate": 1.515829318651067e-05, "loss": 0.6875, "step": 405000 }, { "epoch": 69.6920165175499, "grad_norm": 19.037490844726562, "learning_rate": 1.5153991741225052e-05, "loss": 0.7285, "step": 405050 }, { "epoch": 69.70061940812113, "grad_norm": 16.156112670898438, "learning_rate": 1.5149690295939437e-05, "loss": 0.6768, "step": 405100 }, { "epoch": 69.70922229869237, "grad_norm": 21.618867874145508, "learning_rate": 1.5145388850653821e-05, "loss": 0.7334, "step": 405150 }, { "epoch": 69.71782518926359, "grad_norm": 14.62929916381836, "learning_rate": 1.5141087405368204e-05, "loss": 0.7371, "step": 405200 }, { "epoch": 69.72642807983482, "grad_norm": 8.216275215148926, "learning_rate": 1.5136785960082589e-05, "loss": 0.6797, "step": 405250 }, { "epoch": 69.73503097040606, "grad_norm": 17.934486389160156, "learning_rate": 1.5132484514796972e-05, "loss": 0.6628, "step": 405300 }, { "epoch": 69.74363386097728, "grad_norm": 11.468099594116211, "learning_rate": 1.5128183069511356e-05, "loss": 0.6746, "step": 405350 }, { "epoch": 69.75223675154852, "grad_norm": 11.444952964782715, "learning_rate": 1.512388162422574e-05, "loss": 0.7014, "step": 405400 }, { "epoch": 69.76083964211975, "grad_norm": 12.55447006225586, "learning_rate": 1.5119580178940123e-05, "loss": 0.6821, "step": 405450 }, { "epoch": 69.76944253269099, "grad_norm": 10.956100463867188, "learning_rate": 1.5115278733654508e-05, "loss": 0.6402, "step": 405500 }, { "epoch": 69.77804542326221, "grad_norm": 19.2869873046875, "learning_rate": 1.5110977288368894e-05, "loss": 0.7461, "step": 405550 }, { "epoch": 69.78664831383345, "grad_norm": 23.81093406677246, "learning_rate": 1.5106675843083275e-05, "loss": 0.7141, "step": 405600 }, { "epoch": 69.79525120440468, "grad_norm": 10.430275917053223, "learning_rate": 1.5102374397797662e-05, "loss": 0.7171, "step": 405650 }, { "epoch": 69.80385409497592, "grad_norm": 9.5851469039917, "learning_rate": 1.5098072952512043e-05, "loss": 0.7235, "step": 405700 }, { "epoch": 69.81245698554714, "grad_norm": 16.31774139404297, "learning_rate": 1.509377150722643e-05, "loss": 0.6918, "step": 405750 }, { "epoch": 69.82105987611837, "grad_norm": 16.449275970458984, "learning_rate": 1.5089470061940814e-05, "loss": 0.7099, "step": 405800 }, { "epoch": 69.82966276668961, "grad_norm": 13.971295356750488, "learning_rate": 1.5085168616655197e-05, "loss": 0.752, "step": 405850 }, { "epoch": 69.83826565726083, "grad_norm": 13.197493553161621, "learning_rate": 1.5080867171369581e-05, "loss": 0.71, "step": 405900 }, { "epoch": 69.84686854783207, "grad_norm": 22.58254051208496, "learning_rate": 1.5076565726083966e-05, "loss": 0.7207, "step": 405950 }, { "epoch": 69.8554714384033, "grad_norm": 17.483434677124023, "learning_rate": 1.5072264280798349e-05, "loss": 0.7266, "step": 406000 }, { "epoch": 69.86407432897454, "grad_norm": 12.024601936340332, "learning_rate": 1.5067962835512733e-05, "loss": 0.7062, "step": 406050 }, { "epoch": 69.87267721954576, "grad_norm": 17.33360481262207, "learning_rate": 1.5063661390227118e-05, "loss": 0.7047, "step": 406100 }, { "epoch": 69.881280110117, "grad_norm": 15.471051216125488, "learning_rate": 1.50593599449415e-05, "loss": 0.7092, "step": 406150 }, { "epoch": 69.88988300068823, "grad_norm": 10.13806438446045, "learning_rate": 1.5055058499655885e-05, "loss": 0.6665, "step": 406200 }, { "epoch": 69.89848589125947, "grad_norm": 18.81452751159668, "learning_rate": 1.5050757054370268e-05, "loss": 0.7033, "step": 406250 }, { "epoch": 69.9070887818307, "grad_norm": 16.276845932006836, "learning_rate": 1.5046455609084653e-05, "loss": 0.6449, "step": 406300 }, { "epoch": 69.91569167240193, "grad_norm": 13.956010818481445, "learning_rate": 1.5042154163799039e-05, "loss": 0.7595, "step": 406350 }, { "epoch": 69.92429456297316, "grad_norm": 14.04291820526123, "learning_rate": 1.503785271851342e-05, "loss": 0.7094, "step": 406400 }, { "epoch": 69.93289745354438, "grad_norm": 17.38360595703125, "learning_rate": 1.5033551273227806e-05, "loss": 0.6892, "step": 406450 }, { "epoch": 69.94150034411562, "grad_norm": 19.5562801361084, "learning_rate": 1.502924982794219e-05, "loss": 0.7072, "step": 406500 }, { "epoch": 69.95010323468685, "grad_norm": 27.905637741088867, "learning_rate": 1.5024948382656574e-05, "loss": 0.7055, "step": 406550 }, { "epoch": 69.95870612525809, "grad_norm": 27.799243927001953, "learning_rate": 1.5020646937370958e-05, "loss": 0.71, "step": 406600 }, { "epoch": 69.96730901582931, "grad_norm": 10.860299110412598, "learning_rate": 1.5016345492085341e-05, "loss": 0.7243, "step": 406650 }, { "epoch": 69.97591190640055, "grad_norm": 14.045604705810547, "learning_rate": 1.5012044046799726e-05, "loss": 0.7321, "step": 406700 }, { "epoch": 69.98451479697178, "grad_norm": 12.41484260559082, "learning_rate": 1.500774260151411e-05, "loss": 0.7005, "step": 406750 }, { "epoch": 69.99311768754302, "grad_norm": 16.62067222595215, "learning_rate": 1.5003441156228493e-05, "loss": 0.7365, "step": 406800 }, { "epoch": 70.0, "eval_accuracy": 0.5988642230252968, "eval_f1": 0.591317515572286, "eval_f1_DuraRiadoRio_16x16": 0.5500505561172901, "eval_f1_Mole_16x16": 0.637641877011689, "eval_f1_Quebrado_16x16": 0.7120947388856532, "eval_f1_RiadoRio_16x16": 0.4767415980413977, "eval_f1_RioFechado_16x16": 0.5800588078053996, "eval_loss": 1.770550012588501, "eval_precision": 0.6173965976935604, "eval_precision_DuraRiadoRio_16x16": 0.6585956416464891, "eval_precision_Mole_16x16": 0.52292303417616, "eval_precision_Quebrado_16x16": 0.6441868633649456, "eval_precision_RiadoRio_16x16": 0.519650655021834, "eval_precision_RioFechado_16x16": 0.7416267942583732, "eval_recall": 0.6003485459056626, "eval_recall_DuraRiadoRio_16x16": 0.4722222222222222, "eval_recall_Mole_16x16": 0.8168402777777778, "eval_recall_Quebrado_16x16": 0.7960069444444444, "eval_recall_RiadoRio_16x16": 0.44037828947368424, "eval_recall_RioFechado_16x16": 0.47629499561018435, "eval_runtime": 46.0497, "eval_samples_per_second": 252.38, "eval_steps_per_second": 15.787, "step": 406840 }, { "epoch": 70.00172057811425, "grad_norm": 14.450504302978516, "learning_rate": 1.4999139710942878e-05, "loss": 0.671, "step": 406850 }, { "epoch": 70.01032346868548, "grad_norm": 14.480411529541016, "learning_rate": 1.4994838265657262e-05, "loss": 0.6637, "step": 406900 }, { "epoch": 70.01892635925671, "grad_norm": 14.294744491577148, "learning_rate": 1.4990536820371645e-05, "loss": 0.7348, "step": 406950 }, { "epoch": 70.02752924982794, "grad_norm": 14.007848739624023, "learning_rate": 1.498623537508603e-05, "loss": 0.7367, "step": 407000 }, { "epoch": 70.03613214039918, "grad_norm": 24.7928524017334, "learning_rate": 1.4981933929800412e-05, "loss": 0.702, "step": 407050 }, { "epoch": 70.0447350309704, "grad_norm": 11.873159408569336, "learning_rate": 1.4977632484514797e-05, "loss": 0.7476, "step": 407100 }, { "epoch": 70.05333792154164, "grad_norm": 12.680473327636719, "learning_rate": 1.4973331039229183e-05, "loss": 0.6907, "step": 407150 }, { "epoch": 70.06194081211287, "grad_norm": 16.068143844604492, "learning_rate": 1.4969029593943564e-05, "loss": 0.6754, "step": 407200 }, { "epoch": 70.0705437026841, "grad_norm": 15.141552925109863, "learning_rate": 1.496472814865795e-05, "loss": 0.628, "step": 407250 }, { "epoch": 70.07914659325533, "grad_norm": 20.44547462463379, "learning_rate": 1.4960426703372335e-05, "loss": 0.7041, "step": 407300 }, { "epoch": 70.08774948382657, "grad_norm": 13.735649108886719, "learning_rate": 1.4956125258086716e-05, "loss": 0.6371, "step": 407350 }, { "epoch": 70.0963523743978, "grad_norm": 12.19234848022461, "learning_rate": 1.4951823812801103e-05, "loss": 0.6876, "step": 407400 }, { "epoch": 70.10495526496904, "grad_norm": 9.367653846740723, "learning_rate": 1.4947522367515484e-05, "loss": 0.73, "step": 407450 }, { "epoch": 70.11355815554026, "grad_norm": 21.360448837280273, "learning_rate": 1.494322092222987e-05, "loss": 0.6919, "step": 407500 }, { "epoch": 70.12216104611149, "grad_norm": 17.52753257751465, "learning_rate": 1.4938919476944255e-05, "loss": 0.6833, "step": 407550 }, { "epoch": 70.13076393668273, "grad_norm": 13.014265060424805, "learning_rate": 1.4934618031658638e-05, "loss": 0.7448, "step": 407600 }, { "epoch": 70.13936682725395, "grad_norm": 15.603608131408691, "learning_rate": 1.4930316586373022e-05, "loss": 0.6292, "step": 407650 }, { "epoch": 70.14796971782519, "grad_norm": 7.600451946258545, "learning_rate": 1.4926015141087407e-05, "loss": 0.7231, "step": 407700 }, { "epoch": 70.15657260839642, "grad_norm": 17.681188583374023, "learning_rate": 1.492171369580179e-05, "loss": 0.7307, "step": 407750 }, { "epoch": 70.16517549896766, "grad_norm": 23.822351455688477, "learning_rate": 1.4917412250516174e-05, "loss": 0.6784, "step": 407800 }, { "epoch": 70.17377838953888, "grad_norm": 12.404987335205078, "learning_rate": 1.4913110805230557e-05, "loss": 0.6846, "step": 407850 }, { "epoch": 70.18238128011012, "grad_norm": 12.614652633666992, "learning_rate": 1.4908809359944942e-05, "loss": 0.695, "step": 407900 }, { "epoch": 70.19098417068135, "grad_norm": 16.479976654052734, "learning_rate": 1.4904507914659326e-05, "loss": 0.7003, "step": 407950 }, { "epoch": 70.19958706125259, "grad_norm": 19.210834503173828, "learning_rate": 1.4900206469373709e-05, "loss": 0.714, "step": 408000 }, { "epoch": 70.20818995182381, "grad_norm": 13.987588882446289, "learning_rate": 1.4895905024088094e-05, "loss": 0.7241, "step": 408050 }, { "epoch": 70.21679284239505, "grad_norm": 18.657806396484375, "learning_rate": 1.489160357880248e-05, "loss": 0.676, "step": 408100 }, { "epoch": 70.22539573296628, "grad_norm": 16.878698348999023, "learning_rate": 1.4887302133516861e-05, "loss": 0.6992, "step": 408150 }, { "epoch": 70.2339986235375, "grad_norm": 14.032807350158691, "learning_rate": 1.4883000688231247e-05, "loss": 0.726, "step": 408200 }, { "epoch": 70.24260151410874, "grad_norm": 10.784789085388184, "learning_rate": 1.4878699242945632e-05, "loss": 0.6827, "step": 408250 }, { "epoch": 70.25120440467997, "grad_norm": 25.90644645690918, "learning_rate": 1.4874397797660015e-05, "loss": 0.6884, "step": 408300 }, { "epoch": 70.2598072952512, "grad_norm": 8.6668062210083, "learning_rate": 1.48700963523744e-05, "loss": 0.6672, "step": 408350 }, { "epoch": 70.26841018582243, "grad_norm": 20.95747184753418, "learning_rate": 1.4865794907088782e-05, "loss": 0.6984, "step": 408400 }, { "epoch": 70.27701307639367, "grad_norm": 19.095333099365234, "learning_rate": 1.4861493461803167e-05, "loss": 0.7086, "step": 408450 }, { "epoch": 70.2856159669649, "grad_norm": 10.110631942749023, "learning_rate": 1.4857192016517551e-05, "loss": 0.6772, "step": 408500 }, { "epoch": 70.29421885753614, "grad_norm": 14.081985473632812, "learning_rate": 1.4852890571231934e-05, "loss": 0.7139, "step": 408550 }, { "epoch": 70.30282174810736, "grad_norm": 22.059200286865234, "learning_rate": 1.4848589125946319e-05, "loss": 0.6896, "step": 408600 }, { "epoch": 70.3114246386786, "grad_norm": 20.96949577331543, "learning_rate": 1.4844287680660703e-05, "loss": 0.6964, "step": 408650 }, { "epoch": 70.32002752924983, "grad_norm": 6.166816711425781, "learning_rate": 1.4839986235375086e-05, "loss": 0.7005, "step": 408700 }, { "epoch": 70.32863041982105, "grad_norm": 13.835268020629883, "learning_rate": 1.483568479008947e-05, "loss": 0.674, "step": 408750 }, { "epoch": 70.33723331039229, "grad_norm": 10.887564659118652, "learning_rate": 1.4831383344803853e-05, "loss": 0.7243, "step": 408800 }, { "epoch": 70.34583620096352, "grad_norm": 10.922341346740723, "learning_rate": 1.4827081899518238e-05, "loss": 0.6674, "step": 408850 }, { "epoch": 70.35443909153476, "grad_norm": 9.863804817199707, "learning_rate": 1.4822780454232624e-05, "loss": 0.6758, "step": 408900 }, { "epoch": 70.36304198210598, "grad_norm": 9.097707748413086, "learning_rate": 1.4818479008947005e-05, "loss": 0.7551, "step": 408950 }, { "epoch": 70.37164487267722, "grad_norm": 13.091739654541016, "learning_rate": 1.4814177563661392e-05, "loss": 0.696, "step": 409000 }, { "epoch": 70.38024776324845, "grad_norm": 28.149900436401367, "learning_rate": 1.4809876118375776e-05, "loss": 0.7232, "step": 409050 }, { "epoch": 70.38885065381969, "grad_norm": 23.502662658691406, "learning_rate": 1.4805574673090159e-05, "loss": 0.727, "step": 409100 }, { "epoch": 70.39745354439091, "grad_norm": 16.59731674194336, "learning_rate": 1.4801273227804544e-05, "loss": 0.7207, "step": 409150 }, { "epoch": 70.40605643496215, "grad_norm": 17.839643478393555, "learning_rate": 1.4796971782518927e-05, "loss": 0.7411, "step": 409200 }, { "epoch": 70.41465932553338, "grad_norm": 21.736839294433594, "learning_rate": 1.4792670337233311e-05, "loss": 0.7638, "step": 409250 }, { "epoch": 70.42326221610462, "grad_norm": 15.914794921875, "learning_rate": 1.4788368891947696e-05, "loss": 0.6658, "step": 409300 }, { "epoch": 70.43186510667584, "grad_norm": 12.898957252502441, "learning_rate": 1.4784067446662079e-05, "loss": 0.6676, "step": 409350 }, { "epoch": 70.44046799724707, "grad_norm": 19.62517547607422, "learning_rate": 1.4779766001376463e-05, "loss": 0.7049, "step": 409400 }, { "epoch": 70.44907088781831, "grad_norm": 18.827041625976562, "learning_rate": 1.4775464556090848e-05, "loss": 0.7375, "step": 409450 }, { "epoch": 70.45767377838953, "grad_norm": 20.25049591064453, "learning_rate": 1.477116311080523e-05, "loss": 0.7098, "step": 409500 }, { "epoch": 70.46627666896077, "grad_norm": 15.387873649597168, "learning_rate": 1.4766861665519615e-05, "loss": 0.7495, "step": 409550 }, { "epoch": 70.474879559532, "grad_norm": 18.810420989990234, "learning_rate": 1.4762560220233998e-05, "loss": 0.7314, "step": 409600 }, { "epoch": 70.48348245010324, "grad_norm": 19.232563018798828, "learning_rate": 1.4758258774948383e-05, "loss": 0.6801, "step": 409650 }, { "epoch": 70.49208534067446, "grad_norm": 20.84612464904785, "learning_rate": 1.4753957329662769e-05, "loss": 0.7243, "step": 409700 }, { "epoch": 70.5006882312457, "grad_norm": 12.901121139526367, "learning_rate": 1.474965588437715e-05, "loss": 0.6656, "step": 409750 }, { "epoch": 70.50929112181693, "grad_norm": 12.328622817993164, "learning_rate": 1.4745354439091536e-05, "loss": 0.6521, "step": 409800 }, { "epoch": 70.51789401238817, "grad_norm": 11.620620727539062, "learning_rate": 1.474105299380592e-05, "loss": 0.7104, "step": 409850 }, { "epoch": 70.5264969029594, "grad_norm": 18.109643936157227, "learning_rate": 1.4736751548520304e-05, "loss": 0.6965, "step": 409900 }, { "epoch": 70.53509979353062, "grad_norm": 7.965305805206299, "learning_rate": 1.4732450103234688e-05, "loss": 0.6931, "step": 409950 }, { "epoch": 70.54370268410186, "grad_norm": 10.989516258239746, "learning_rate": 1.4728148657949073e-05, "loss": 0.6563, "step": 410000 }, { "epoch": 70.55230557467308, "grad_norm": 9.138676643371582, "learning_rate": 1.4723847212663456e-05, "loss": 0.6971, "step": 410050 }, { "epoch": 70.56090846524432, "grad_norm": 13.176525115966797, "learning_rate": 1.471954576737784e-05, "loss": 0.7185, "step": 410100 }, { "epoch": 70.56951135581555, "grad_norm": 12.758399963378906, "learning_rate": 1.4715244322092223e-05, "loss": 0.6907, "step": 410150 }, { "epoch": 70.57811424638679, "grad_norm": 18.479660034179688, "learning_rate": 1.4710942876806608e-05, "loss": 0.7614, "step": 410200 }, { "epoch": 70.58671713695801, "grad_norm": 11.317488670349121, "learning_rate": 1.4706641431520992e-05, "loss": 0.6847, "step": 410250 }, { "epoch": 70.59532002752925, "grad_norm": 12.6489896774292, "learning_rate": 1.4702339986235375e-05, "loss": 0.6846, "step": 410300 }, { "epoch": 70.60392291810048, "grad_norm": 12.974639892578125, "learning_rate": 1.469803854094976e-05, "loss": 0.7384, "step": 410350 }, { "epoch": 70.61252580867172, "grad_norm": 13.702531814575195, "learning_rate": 1.4693737095664146e-05, "loss": 0.7293, "step": 410400 }, { "epoch": 70.62112869924295, "grad_norm": 16.127674102783203, "learning_rate": 1.4689435650378527e-05, "loss": 0.746, "step": 410450 }, { "epoch": 70.62973158981418, "grad_norm": 12.456644058227539, "learning_rate": 1.4685134205092913e-05, "loss": 0.6441, "step": 410500 }, { "epoch": 70.63833448038541, "grad_norm": 24.302644729614258, "learning_rate": 1.4680832759807294e-05, "loss": 0.725, "step": 410550 }, { "epoch": 70.64693737095664, "grad_norm": 22.110153198242188, "learning_rate": 1.4676531314521679e-05, "loss": 0.6949, "step": 410600 }, { "epoch": 70.65554026152788, "grad_norm": 9.752877235412598, "learning_rate": 1.4672229869236065e-05, "loss": 0.7022, "step": 410650 }, { "epoch": 70.6641431520991, "grad_norm": 14.74313735961914, "learning_rate": 1.4667928423950446e-05, "loss": 0.6805, "step": 410700 }, { "epoch": 70.67274604267034, "grad_norm": 18.546077728271484, "learning_rate": 1.4663626978664833e-05, "loss": 0.7238, "step": 410750 }, { "epoch": 70.68134893324157, "grad_norm": 16.97316551208496, "learning_rate": 1.4659325533379217e-05, "loss": 0.7073, "step": 410800 }, { "epoch": 70.6899518238128, "grad_norm": 9.886799812316895, "learning_rate": 1.46550240880936e-05, "loss": 0.7099, "step": 410850 }, { "epoch": 70.69855471438403, "grad_norm": 12.063591957092285, "learning_rate": 1.4650722642807985e-05, "loss": 0.7226, "step": 410900 }, { "epoch": 70.70715760495527, "grad_norm": 10.991911888122559, "learning_rate": 1.4646421197522368e-05, "loss": 0.7082, "step": 410950 }, { "epoch": 70.7157604955265, "grad_norm": 10.584822654724121, "learning_rate": 1.4642119752236752e-05, "loss": 0.6834, "step": 411000 }, { "epoch": 70.72436338609774, "grad_norm": 21.011348724365234, "learning_rate": 1.4637818306951137e-05, "loss": 0.6526, "step": 411050 }, { "epoch": 70.73296627666896, "grad_norm": 18.27987289428711, "learning_rate": 1.463351686166552e-05, "loss": 0.6659, "step": 411100 }, { "epoch": 70.74156916724019, "grad_norm": 11.394633293151855, "learning_rate": 1.4629215416379904e-05, "loss": 0.6787, "step": 411150 }, { "epoch": 70.75017205781143, "grad_norm": 11.00064754486084, "learning_rate": 1.462491397109429e-05, "loss": 0.7231, "step": 411200 }, { "epoch": 70.75877494838265, "grad_norm": 18.674278259277344, "learning_rate": 1.4620612525808672e-05, "loss": 0.7297, "step": 411250 }, { "epoch": 70.76737783895389, "grad_norm": 22.97646713256836, "learning_rate": 1.4616311080523056e-05, "loss": 0.7237, "step": 411300 }, { "epoch": 70.77598072952512, "grad_norm": 11.046492576599121, "learning_rate": 1.4612009635237439e-05, "loss": 0.6904, "step": 411350 }, { "epoch": 70.78458362009636, "grad_norm": 11.21664047241211, "learning_rate": 1.4607708189951823e-05, "loss": 0.7217, "step": 411400 }, { "epoch": 70.79318651066758, "grad_norm": 12.138934135437012, "learning_rate": 1.460340674466621e-05, "loss": 0.6718, "step": 411450 }, { "epoch": 70.80178940123882, "grad_norm": 19.82110023498535, "learning_rate": 1.4599105299380591e-05, "loss": 0.6751, "step": 411500 }, { "epoch": 70.81039229181005, "grad_norm": 20.308181762695312, "learning_rate": 1.4594803854094977e-05, "loss": 0.6816, "step": 411550 }, { "epoch": 70.81899518238129, "grad_norm": 12.765812873840332, "learning_rate": 1.4590502408809362e-05, "loss": 0.6591, "step": 411600 }, { "epoch": 70.82759807295251, "grad_norm": 21.72567367553711, "learning_rate": 1.4586200963523745e-05, "loss": 0.7256, "step": 411650 }, { "epoch": 70.83620096352374, "grad_norm": 15.796318054199219, "learning_rate": 1.458189951823813e-05, "loss": 0.6593, "step": 411700 }, { "epoch": 70.84480385409498, "grad_norm": 15.004311561584473, "learning_rate": 1.4577598072952514e-05, "loss": 0.673, "step": 411750 }, { "epoch": 70.8534067446662, "grad_norm": 13.162912368774414, "learning_rate": 1.4573296627666897e-05, "loss": 0.7024, "step": 411800 }, { "epoch": 70.86200963523744, "grad_norm": 17.75739860534668, "learning_rate": 1.4568995182381281e-05, "loss": 0.6982, "step": 411850 }, { "epoch": 70.87061252580867, "grad_norm": 14.090348243713379, "learning_rate": 1.4564693737095664e-05, "loss": 0.6721, "step": 411900 }, { "epoch": 70.8792154163799, "grad_norm": 13.596837997436523, "learning_rate": 1.4560392291810049e-05, "loss": 0.7085, "step": 411950 }, { "epoch": 70.88781830695113, "grad_norm": 13.635163307189941, "learning_rate": 1.4556090846524433e-05, "loss": 0.7195, "step": 412000 }, { "epoch": 70.89642119752237, "grad_norm": 18.220550537109375, "learning_rate": 1.4551789401238816e-05, "loss": 0.6692, "step": 412050 }, { "epoch": 70.9050240880936, "grad_norm": 13.517500877380371, "learning_rate": 1.45474879559532e-05, "loss": 0.7677, "step": 412100 }, { "epoch": 70.91362697866484, "grad_norm": 17.76859474182129, "learning_rate": 1.4543186510667587e-05, "loss": 0.719, "step": 412150 }, { "epoch": 70.92222986923606, "grad_norm": 16.269344329833984, "learning_rate": 1.4538885065381968e-05, "loss": 0.7623, "step": 412200 }, { "epoch": 70.9308327598073, "grad_norm": 20.984941482543945, "learning_rate": 1.4534583620096354e-05, "loss": 0.6817, "step": 412250 }, { "epoch": 70.93943565037853, "grad_norm": 17.76770782470703, "learning_rate": 1.4530282174810735e-05, "loss": 0.7267, "step": 412300 }, { "epoch": 70.94803854094975, "grad_norm": 13.1541166305542, "learning_rate": 1.4525980729525122e-05, "loss": 0.688, "step": 412350 }, { "epoch": 70.95664143152099, "grad_norm": 17.65077781677246, "learning_rate": 1.4521679284239506e-05, "loss": 0.6811, "step": 412400 }, { "epoch": 70.96524432209222, "grad_norm": 8.67151927947998, "learning_rate": 1.4517377838953889e-05, "loss": 0.7011, "step": 412450 }, { "epoch": 70.97384721266346, "grad_norm": 9.628686904907227, "learning_rate": 1.4513076393668274e-05, "loss": 0.6504, "step": 412500 }, { "epoch": 70.98245010323468, "grad_norm": 13.15707015991211, "learning_rate": 1.4508774948382658e-05, "loss": 0.7056, "step": 412550 }, { "epoch": 70.99105299380592, "grad_norm": 22.86554527282715, "learning_rate": 1.4504473503097041e-05, "loss": 0.7019, "step": 412600 }, { "epoch": 70.99965588437715, "grad_norm": 11.42168140411377, "learning_rate": 1.4500172057811426e-05, "loss": 0.7129, "step": 412650 }, { "epoch": 71.0, "eval_accuracy": 0.551368094992256, "eval_f1": 0.5585947783356111, "eval_f1_DuraRiadoRio_16x16": 0.4731369150779896, "eval_f1_Mole_16x16": 0.5320796460176991, "eval_f1_Quebrado_16x16": 0.7317757009345794, "eval_f1_RiadoRio_16x16": 0.4929317762753534, "eval_f1_RioFechado_16x16": 0.5630498533724341, "eval_loss": 1.994749665260315, "eval_precision": 0.6602938682471725, "eval_precision_DuraRiadoRio_16x16": 0.7072538860103627, "eval_precision_Mole_16x16": 0.7332317073170732, "eval_precision_Quebrado_16x16": 0.7925101214574899, "eval_precision_RiadoRio_16x16": 0.35156934946519375, "eval_precision_RioFechado_16x16": 0.7169042769857433, "eval_recall": 0.5481359689234427, "eval_recall_DuraRiadoRio_16x16": 0.35546875, "eval_recall_Mole_16x16": 0.4175347222222222, "eval_recall_Quebrado_16x16": 0.6796875, "eval_recall_RiadoRio_16x16": 0.8244243421052632, "eval_recall_RioFechado_16x16": 0.4635645302897278, "eval_runtime": 46.4723, "eval_samples_per_second": 250.085, "eval_steps_per_second": 15.644, "step": 412652 }, { "epoch": 71.00825877494839, "grad_norm": 11.763453483581543, "learning_rate": 1.4495870612525809e-05, "loss": 0.6727, "step": 412700 }, { "epoch": 71.01686166551961, "grad_norm": 11.366218566894531, "learning_rate": 1.4491569167240193e-05, "loss": 0.701, "step": 412750 }, { "epoch": 71.02546455609085, "grad_norm": 16.414575576782227, "learning_rate": 1.4487267721954578e-05, "loss": 0.7283, "step": 412800 }, { "epoch": 71.03406744666208, "grad_norm": 17.556232452392578, "learning_rate": 1.448296627666896e-05, "loss": 0.6648, "step": 412850 }, { "epoch": 71.0426703372333, "grad_norm": 7.221403121948242, "learning_rate": 1.4478664831383345e-05, "loss": 0.6666, "step": 412900 }, { "epoch": 71.05127322780454, "grad_norm": 19.57260513305664, "learning_rate": 1.4474363386097731e-05, "loss": 0.7035, "step": 412950 }, { "epoch": 71.05987611837577, "grad_norm": 21.036706924438477, "learning_rate": 1.4470061940812112e-05, "loss": 0.6949, "step": 413000 }, { "epoch": 71.06847900894701, "grad_norm": 14.997758865356445, "learning_rate": 1.4465760495526499e-05, "loss": 0.6667, "step": 413050 }, { "epoch": 71.07708189951823, "grad_norm": 23.850046157836914, "learning_rate": 1.446145905024088e-05, "loss": 0.7089, "step": 413100 }, { "epoch": 71.08568479008947, "grad_norm": 12.861620903015137, "learning_rate": 1.4457157604955266e-05, "loss": 0.6776, "step": 413150 }, { "epoch": 71.0942876806607, "grad_norm": 12.545320510864258, "learning_rate": 1.445285615966965e-05, "loss": 0.7207, "step": 413200 }, { "epoch": 71.10289057123194, "grad_norm": 13.105843544006348, "learning_rate": 1.4448554714384032e-05, "loss": 0.7266, "step": 413250 }, { "epoch": 71.11149346180316, "grad_norm": 18.67549705505371, "learning_rate": 1.4444253269098418e-05, "loss": 0.751, "step": 413300 }, { "epoch": 71.1200963523744, "grad_norm": 10.95986557006836, "learning_rate": 1.4439951823812803e-05, "loss": 0.6893, "step": 413350 }, { "epoch": 71.12869924294563, "grad_norm": 14.915287971496582, "learning_rate": 1.4435650378527186e-05, "loss": 0.6729, "step": 413400 }, { "epoch": 71.13730213351687, "grad_norm": 15.501258850097656, "learning_rate": 1.443134893324157e-05, "loss": 0.7109, "step": 413450 }, { "epoch": 71.1459050240881, "grad_norm": 26.958023071289062, "learning_rate": 1.4427047487955953e-05, "loss": 0.655, "step": 413500 }, { "epoch": 71.15450791465932, "grad_norm": 10.58497142791748, "learning_rate": 1.4422746042670338e-05, "loss": 0.7034, "step": 413550 }, { "epoch": 71.16311080523056, "grad_norm": 11.290982246398926, "learning_rate": 1.4418444597384722e-05, "loss": 0.654, "step": 413600 }, { "epoch": 71.17171369580178, "grad_norm": 18.901939392089844, "learning_rate": 1.4414143152099105e-05, "loss": 0.6799, "step": 413650 }, { "epoch": 71.18031658637302, "grad_norm": 16.825471878051758, "learning_rate": 1.440984170681349e-05, "loss": 0.7165, "step": 413700 }, { "epoch": 71.18891947694425, "grad_norm": 15.864004135131836, "learning_rate": 1.4405540261527876e-05, "loss": 0.6726, "step": 413750 }, { "epoch": 71.19752236751549, "grad_norm": 11.8984956741333, "learning_rate": 1.4401238816242257e-05, "loss": 0.6619, "step": 413800 }, { "epoch": 71.20612525808671, "grad_norm": 23.69864273071289, "learning_rate": 1.4396937370956643e-05, "loss": 0.722, "step": 413850 }, { "epoch": 71.21472814865795, "grad_norm": 18.31846046447754, "learning_rate": 1.4392635925671028e-05, "loss": 0.6436, "step": 413900 }, { "epoch": 71.22333103922918, "grad_norm": 7.34660005569458, "learning_rate": 1.4388334480385409e-05, "loss": 0.6958, "step": 413950 }, { "epoch": 71.23193392980042, "grad_norm": 13.789754867553711, "learning_rate": 1.4384033035099795e-05, "loss": 0.7582, "step": 414000 }, { "epoch": 71.24053682037164, "grad_norm": 12.177610397338867, "learning_rate": 1.4379731589814176e-05, "loss": 0.6986, "step": 414050 }, { "epoch": 71.24913971094287, "grad_norm": 13.610553741455078, "learning_rate": 1.4375430144528563e-05, "loss": 0.6821, "step": 414100 }, { "epoch": 71.25774260151411, "grad_norm": 18.34286880493164, "learning_rate": 1.4371128699242947e-05, "loss": 0.7058, "step": 414150 }, { "epoch": 71.26634549208534, "grad_norm": 10.868509292602539, "learning_rate": 1.436682725395733e-05, "loss": 0.7296, "step": 414200 }, { "epoch": 71.27494838265658, "grad_norm": 11.555421829223633, "learning_rate": 1.4362525808671715e-05, "loss": 0.7154, "step": 414250 }, { "epoch": 71.2835512732278, "grad_norm": 19.382230758666992, "learning_rate": 1.43582243633861e-05, "loss": 0.6772, "step": 414300 }, { "epoch": 71.29215416379904, "grad_norm": 11.946433067321777, "learning_rate": 1.4353922918100482e-05, "loss": 0.655, "step": 414350 }, { "epoch": 71.30075705437027, "grad_norm": 32.020877838134766, "learning_rate": 1.4349621472814867e-05, "loss": 0.6523, "step": 414400 }, { "epoch": 71.3093599449415, "grad_norm": 25.97490119934082, "learning_rate": 1.434532002752925e-05, "loss": 0.708, "step": 414450 }, { "epoch": 71.31796283551273, "grad_norm": 13.421930313110352, "learning_rate": 1.4341018582243634e-05, "loss": 0.7262, "step": 414500 }, { "epoch": 71.32656572608397, "grad_norm": 11.63247013092041, "learning_rate": 1.4336717136958019e-05, "loss": 0.7354, "step": 414550 }, { "epoch": 71.3351686166552, "grad_norm": 11.650546073913574, "learning_rate": 1.4332415691672401e-05, "loss": 0.7585, "step": 414600 }, { "epoch": 71.34377150722642, "grad_norm": 9.966898918151855, "learning_rate": 1.4328114246386786e-05, "loss": 0.6833, "step": 414650 }, { "epoch": 71.35237439779766, "grad_norm": 15.202969551086426, "learning_rate": 1.4323812801101172e-05, "loss": 0.6929, "step": 414700 }, { "epoch": 71.36097728836889, "grad_norm": 15.184891700744629, "learning_rate": 1.4319511355815553e-05, "loss": 0.6837, "step": 414750 }, { "epoch": 71.36958017894013, "grad_norm": 8.863274574279785, "learning_rate": 1.431520991052994e-05, "loss": 0.6584, "step": 414800 }, { "epoch": 71.37818306951135, "grad_norm": 11.947077751159668, "learning_rate": 1.4310908465244321e-05, "loss": 0.73, "step": 414850 }, { "epoch": 71.38678596008259, "grad_norm": 11.637049674987793, "learning_rate": 1.4306607019958707e-05, "loss": 0.6773, "step": 414900 }, { "epoch": 71.39538885065382, "grad_norm": 35.493507385253906, "learning_rate": 1.4302305574673092e-05, "loss": 0.6991, "step": 414950 }, { "epoch": 71.40399174122506, "grad_norm": 15.726381301879883, "learning_rate": 1.4298004129387475e-05, "loss": 0.6968, "step": 415000 }, { "epoch": 71.41259463179628, "grad_norm": 13.467516899108887, "learning_rate": 1.4293702684101859e-05, "loss": 0.6645, "step": 415050 }, { "epoch": 71.42119752236752, "grad_norm": 12.259528160095215, "learning_rate": 1.4289401238816244e-05, "loss": 0.6877, "step": 415100 }, { "epoch": 71.42980041293875, "grad_norm": 11.67399787902832, "learning_rate": 1.4285099793530627e-05, "loss": 0.6777, "step": 415150 }, { "epoch": 71.43840330350999, "grad_norm": 11.540964126586914, "learning_rate": 1.4280798348245011e-05, "loss": 0.7399, "step": 415200 }, { "epoch": 71.44700619408121, "grad_norm": 20.970924377441406, "learning_rate": 1.4276496902959394e-05, "loss": 0.6935, "step": 415250 }, { "epoch": 71.45560908465244, "grad_norm": 15.057682991027832, "learning_rate": 1.4272195457673779e-05, "loss": 0.7323, "step": 415300 }, { "epoch": 71.46421197522368, "grad_norm": 16.256240844726562, "learning_rate": 1.4267894012388163e-05, "loss": 0.7052, "step": 415350 }, { "epoch": 71.4728148657949, "grad_norm": 17.0246524810791, "learning_rate": 1.4263592567102546e-05, "loss": 0.6744, "step": 415400 }, { "epoch": 71.48141775636614, "grad_norm": 16.10069465637207, "learning_rate": 1.425929112181693e-05, "loss": 0.6983, "step": 415450 }, { "epoch": 71.49002064693737, "grad_norm": 14.165087699890137, "learning_rate": 1.4254989676531317e-05, "loss": 0.6876, "step": 415500 }, { "epoch": 71.4986235375086, "grad_norm": 21.89678955078125, "learning_rate": 1.4250688231245698e-05, "loss": 0.7851, "step": 415550 }, { "epoch": 71.50722642807983, "grad_norm": 15.364922523498535, "learning_rate": 1.4246386785960084e-05, "loss": 0.7127, "step": 415600 }, { "epoch": 71.51582931865107, "grad_norm": 10.8438138961792, "learning_rate": 1.4242085340674469e-05, "loss": 0.6729, "step": 415650 }, { "epoch": 71.5244322092223, "grad_norm": 19.35379981994629, "learning_rate": 1.4237783895388852e-05, "loss": 0.7183, "step": 415700 }, { "epoch": 71.53303509979354, "grad_norm": 11.953607559204102, "learning_rate": 1.4233482450103236e-05, "loss": 0.7419, "step": 415750 }, { "epoch": 71.54163799036476, "grad_norm": 20.335546493530273, "learning_rate": 1.4229181004817619e-05, "loss": 0.7031, "step": 415800 }, { "epoch": 71.55024088093599, "grad_norm": 13.985740661621094, "learning_rate": 1.4224879559532004e-05, "loss": 0.6936, "step": 415850 }, { "epoch": 71.55884377150723, "grad_norm": 6.866407871246338, "learning_rate": 1.4220578114246388e-05, "loss": 0.638, "step": 415900 }, { "epoch": 71.56744666207845, "grad_norm": 12.648357391357422, "learning_rate": 1.4216276668960771e-05, "loss": 0.7124, "step": 415950 }, { "epoch": 71.57604955264969, "grad_norm": 17.8150634765625, "learning_rate": 1.4211975223675156e-05, "loss": 0.6927, "step": 416000 }, { "epoch": 71.58465244322092, "grad_norm": 19.21161460876465, "learning_rate": 1.420767377838954e-05, "loss": 0.6878, "step": 416050 }, { "epoch": 71.59325533379216, "grad_norm": 19.50210189819336, "learning_rate": 1.4203372333103923e-05, "loss": 0.733, "step": 416100 }, { "epoch": 71.60185822436338, "grad_norm": 21.02195167541504, "learning_rate": 1.4199070887818308e-05, "loss": 0.6614, "step": 416150 }, { "epoch": 71.61046111493462, "grad_norm": 18.60504722595215, "learning_rate": 1.419476944253269e-05, "loss": 0.7169, "step": 416200 }, { "epoch": 71.61906400550585, "grad_norm": 7.8173508644104, "learning_rate": 1.4190467997247075e-05, "loss": 0.67, "step": 416250 }, { "epoch": 71.62766689607709, "grad_norm": 14.932493209838867, "learning_rate": 1.4186166551961461e-05, "loss": 0.6845, "step": 416300 }, { "epoch": 71.63626978664831, "grad_norm": 17.449703216552734, "learning_rate": 1.4181865106675842e-05, "loss": 0.7297, "step": 416350 }, { "epoch": 71.64487267721955, "grad_norm": 19.928327560424805, "learning_rate": 1.4177563661390229e-05, "loss": 0.7086, "step": 416400 }, { "epoch": 71.65347556779078, "grad_norm": 13.672469139099121, "learning_rate": 1.4173262216104613e-05, "loss": 0.6995, "step": 416450 }, { "epoch": 71.662078458362, "grad_norm": 9.218276977539062, "learning_rate": 1.4168960770818996e-05, "loss": 0.7022, "step": 416500 }, { "epoch": 71.67068134893324, "grad_norm": 21.36471939086914, "learning_rate": 1.416465932553338e-05, "loss": 0.7152, "step": 416550 }, { "epoch": 71.67928423950447, "grad_norm": 10.040517807006836, "learning_rate": 1.4160357880247762e-05, "loss": 0.7048, "step": 416600 }, { "epoch": 71.68788713007571, "grad_norm": 15.495648384094238, "learning_rate": 1.4156056434962148e-05, "loss": 0.6681, "step": 416650 }, { "epoch": 71.69649002064693, "grad_norm": 17.024620056152344, "learning_rate": 1.4151754989676533e-05, "loss": 0.7719, "step": 416700 }, { "epoch": 71.70509291121817, "grad_norm": 14.272407531738281, "learning_rate": 1.4147453544390916e-05, "loss": 0.6993, "step": 416750 }, { "epoch": 71.7136958017894, "grad_norm": 18.09189796447754, "learning_rate": 1.41431520991053e-05, "loss": 0.7767, "step": 416800 }, { "epoch": 71.72229869236064, "grad_norm": 12.755119323730469, "learning_rate": 1.4138850653819685e-05, "loss": 0.7011, "step": 416850 }, { "epoch": 71.73090158293186, "grad_norm": 12.169929504394531, "learning_rate": 1.4134549208534068e-05, "loss": 0.7188, "step": 416900 }, { "epoch": 71.7395044735031, "grad_norm": 12.839855194091797, "learning_rate": 1.4130247763248452e-05, "loss": 0.6624, "step": 416950 }, { "epoch": 71.74810736407433, "grad_norm": 19.434738159179688, "learning_rate": 1.4125946317962835e-05, "loss": 0.7174, "step": 417000 }, { "epoch": 71.75671025464555, "grad_norm": 12.2906494140625, "learning_rate": 1.412164487267722e-05, "loss": 0.7269, "step": 417050 }, { "epoch": 71.7653131452168, "grad_norm": 10.8482666015625, "learning_rate": 1.4117343427391606e-05, "loss": 0.6344, "step": 417100 }, { "epoch": 71.77391603578802, "grad_norm": 15.507794380187988, "learning_rate": 1.4113041982105987e-05, "loss": 0.6657, "step": 417150 }, { "epoch": 71.78251892635926, "grad_norm": 10.939839363098145, "learning_rate": 1.4108740536820372e-05, "loss": 0.651, "step": 417200 }, { "epoch": 71.79112181693048, "grad_norm": 22.088056564331055, "learning_rate": 1.4104439091534758e-05, "loss": 0.6749, "step": 417250 }, { "epoch": 71.79972470750172, "grad_norm": 20.912052154541016, "learning_rate": 1.4100137646249139e-05, "loss": 0.6858, "step": 417300 }, { "epoch": 71.80832759807295, "grad_norm": 12.173185348510742, "learning_rate": 1.4095836200963525e-05, "loss": 0.6738, "step": 417350 }, { "epoch": 71.81693048864419, "grad_norm": 15.551214218139648, "learning_rate": 1.409153475567791e-05, "loss": 0.6154, "step": 417400 }, { "epoch": 71.82553337921541, "grad_norm": 14.465970993041992, "learning_rate": 1.4087233310392293e-05, "loss": 0.6954, "step": 417450 }, { "epoch": 71.83413626978665, "grad_norm": 12.21230411529541, "learning_rate": 1.4082931865106677e-05, "loss": 0.6618, "step": 417500 }, { "epoch": 71.84273916035788, "grad_norm": 15.096423149108887, "learning_rate": 1.407863041982106e-05, "loss": 0.7181, "step": 417550 }, { "epoch": 71.8513420509291, "grad_norm": 13.04273509979248, "learning_rate": 1.4074328974535445e-05, "loss": 0.7217, "step": 417600 }, { "epoch": 71.85994494150034, "grad_norm": 18.636474609375, "learning_rate": 1.407002752924983e-05, "loss": 0.7027, "step": 417650 }, { "epoch": 71.86854783207157, "grad_norm": 10.981657028198242, "learning_rate": 1.4065726083964212e-05, "loss": 0.6141, "step": 417700 }, { "epoch": 71.87715072264281, "grad_norm": 19.069616317749023, "learning_rate": 1.4061424638678597e-05, "loss": 0.7053, "step": 417750 }, { "epoch": 71.88575361321404, "grad_norm": 9.89016056060791, "learning_rate": 1.4057123193392981e-05, "loss": 0.6958, "step": 417800 }, { "epoch": 71.89435650378527, "grad_norm": 12.28982925415039, "learning_rate": 1.4052821748107364e-05, "loss": 0.738, "step": 417850 }, { "epoch": 71.9029593943565, "grad_norm": 10.089877128601074, "learning_rate": 1.4048520302821749e-05, "loss": 0.6515, "step": 417900 }, { "epoch": 71.91156228492774, "grad_norm": 10.354772567749023, "learning_rate": 1.4044218857536131e-05, "loss": 0.6857, "step": 417950 }, { "epoch": 71.92016517549897, "grad_norm": 16.14156150817871, "learning_rate": 1.4039917412250516e-05, "loss": 0.6876, "step": 418000 }, { "epoch": 71.9287680660702, "grad_norm": 13.66077995300293, "learning_rate": 1.4035615966964902e-05, "loss": 0.7096, "step": 418050 }, { "epoch": 71.93737095664143, "grad_norm": 15.987457275390625, "learning_rate": 1.4031314521679283e-05, "loss": 0.6902, "step": 418100 }, { "epoch": 71.94597384721267, "grad_norm": 14.29362964630127, "learning_rate": 1.402701307639367e-05, "loss": 0.6955, "step": 418150 }, { "epoch": 71.9545767377839, "grad_norm": 16.81549835205078, "learning_rate": 1.4022711631108054e-05, "loss": 0.6916, "step": 418200 }, { "epoch": 71.96317962835512, "grad_norm": 13.142071723937988, "learning_rate": 1.4018410185822437e-05, "loss": 0.6957, "step": 418250 }, { "epoch": 71.97178251892636, "grad_norm": 9.136984825134277, "learning_rate": 1.4014108740536822e-05, "loss": 0.6953, "step": 418300 }, { "epoch": 71.98038540949759, "grad_norm": 17.49303436279297, "learning_rate": 1.4009807295251205e-05, "loss": 0.7113, "step": 418350 }, { "epoch": 71.98898830006883, "grad_norm": 19.198759078979492, "learning_rate": 1.4005505849965589e-05, "loss": 0.6967, "step": 418400 }, { "epoch": 71.99759119064005, "grad_norm": 16.985950469970703, "learning_rate": 1.4001204404679974e-05, "loss": 0.6914, "step": 418450 }, { "epoch": 72.0, "eval_accuracy": 0.5166924797797281, "eval_f1": 0.5044289536538864, "eval_f1_DuraRiadoRio_16x16": 0.5638923543040456, "eval_f1_Mole_16x16": 0.456223942823109, "eval_f1_Quebrado_16x16": 0.6247640619101548, "eval_f1_RiadoRio_16x16": 0.46325167037861914, "eval_f1_RioFechado_16x16": 0.4140127388535032, "eval_loss": 2.576340913772583, "eval_precision": 0.5818060184428682, "eval_precision_DuraRiadoRio_16x16": 0.4783791956456002, "eval_precision_Mole_16x16": 0.7267552182163188, "eval_precision_Quebrado_16x16": 0.5527722110888443, "eval_precision_RiadoRio_16x16": 0.3970631424375918, "eval_precision_RioFechado_16x16": 0.7540603248259861, "eval_recall": 0.5157344525758719, "eval_recall_DuraRiadoRio_16x16": 0.6866319444444444, "eval_recall_Mole_16x16": 0.3324652777777778, "eval_recall_Quebrado_16x16": 0.7183159722222222, "eval_recall_RiadoRio_16x16": 0.555921052631579, "eval_recall_RioFechado_16x16": 0.2853380158033363, "eval_runtime": 46.514, "eval_samples_per_second": 249.86, "eval_steps_per_second": 15.63, "step": 418464 }, { "epoch": 72.00619408121129, "grad_norm": 25.57817840576172, "learning_rate": 1.3996902959394357e-05, "loss": 0.7116, "step": 418500 }, { "epoch": 72.01479697178252, "grad_norm": 12.402817726135254, "learning_rate": 1.3992601514108741e-05, "loss": 0.7068, "step": 418550 }, { "epoch": 72.02339986235376, "grad_norm": 24.084428787231445, "learning_rate": 1.3988300068823126e-05, "loss": 0.6832, "step": 418600 }, { "epoch": 72.03200275292498, "grad_norm": 22.780473709106445, "learning_rate": 1.3983998623537509e-05, "loss": 0.6806, "step": 418650 }, { "epoch": 72.04060564349622, "grad_norm": 14.515318870544434, "learning_rate": 1.3979697178251893e-05, "loss": 0.6695, "step": 418700 }, { "epoch": 72.04920853406745, "grad_norm": 10.353822708129883, "learning_rate": 1.3975395732966276e-05, "loss": 0.7217, "step": 418750 }, { "epoch": 72.05781142463867, "grad_norm": 19.341432571411133, "learning_rate": 1.397109428768066e-05, "loss": 0.7415, "step": 418800 }, { "epoch": 72.06641431520991, "grad_norm": 10.778374671936035, "learning_rate": 1.3966792842395047e-05, "loss": 0.7173, "step": 418850 }, { "epoch": 72.07501720578114, "grad_norm": 12.064221382141113, "learning_rate": 1.3962491397109428e-05, "loss": 0.6641, "step": 418900 }, { "epoch": 72.08362009635238, "grad_norm": 9.716400146484375, "learning_rate": 1.3958189951823814e-05, "loss": 0.7117, "step": 418950 }, { "epoch": 72.0922229869236, "grad_norm": 21.788373947143555, "learning_rate": 1.3953888506538199e-05, "loss": 0.6757, "step": 419000 }, { "epoch": 72.10082587749484, "grad_norm": 11.21207332611084, "learning_rate": 1.3949587061252582e-05, "loss": 0.7288, "step": 419050 }, { "epoch": 72.10942876806607, "grad_norm": 14.190617561340332, "learning_rate": 1.3945285615966966e-05, "loss": 0.6481, "step": 419100 }, { "epoch": 72.1180316586373, "grad_norm": 11.7478666305542, "learning_rate": 1.394098417068135e-05, "loss": 0.6804, "step": 419150 }, { "epoch": 72.12663454920853, "grad_norm": 9.265314102172852, "learning_rate": 1.3936682725395734e-05, "loss": 0.6764, "step": 419200 }, { "epoch": 72.13523743977977, "grad_norm": 12.028761863708496, "learning_rate": 1.3932381280110118e-05, "loss": 0.6524, "step": 419250 }, { "epoch": 72.143840330351, "grad_norm": 18.74740219116211, "learning_rate": 1.3928079834824501e-05, "loss": 0.6881, "step": 419300 }, { "epoch": 72.15244322092224, "grad_norm": 13.519957542419434, "learning_rate": 1.3923778389538886e-05, "loss": 0.6658, "step": 419350 }, { "epoch": 72.16104611149346, "grad_norm": 8.821571350097656, "learning_rate": 1.391947694425327e-05, "loss": 0.6869, "step": 419400 }, { "epoch": 72.16964900206469, "grad_norm": 18.25507926940918, "learning_rate": 1.3915175498967653e-05, "loss": 0.7135, "step": 419450 }, { "epoch": 72.17825189263593, "grad_norm": 12.99079704284668, "learning_rate": 1.3910874053682038e-05, "loss": 0.6759, "step": 419500 }, { "epoch": 72.18685478320715, "grad_norm": 15.362444877624512, "learning_rate": 1.3906572608396424e-05, "loss": 0.7048, "step": 419550 }, { "epoch": 72.19545767377839, "grad_norm": 15.47815990447998, "learning_rate": 1.3902271163110805e-05, "loss": 0.676, "step": 419600 }, { "epoch": 72.20406056434962, "grad_norm": 13.734200477600098, "learning_rate": 1.3897969717825191e-05, "loss": 0.6427, "step": 419650 }, { "epoch": 72.21266345492086, "grad_norm": 14.373022079467773, "learning_rate": 1.3893668272539572e-05, "loss": 0.6899, "step": 419700 }, { "epoch": 72.22126634549208, "grad_norm": 9.685423851013184, "learning_rate": 1.3889366827253959e-05, "loss": 0.6903, "step": 419750 }, { "epoch": 72.22986923606332, "grad_norm": 13.61600399017334, "learning_rate": 1.3885065381968343e-05, "loss": 0.7509, "step": 419800 }, { "epoch": 72.23847212663455, "grad_norm": 11.190386772155762, "learning_rate": 1.3880763936682724e-05, "loss": 0.7181, "step": 419850 }, { "epoch": 72.24707501720579, "grad_norm": 17.767101287841797, "learning_rate": 1.387646249139711e-05, "loss": 0.7401, "step": 419900 }, { "epoch": 72.25567790777701, "grad_norm": 11.241602897644043, "learning_rate": 1.3872161046111495e-05, "loss": 0.6703, "step": 419950 }, { "epoch": 72.26428079834824, "grad_norm": 19.06078338623047, "learning_rate": 1.3867859600825878e-05, "loss": 0.7681, "step": 420000 }, { "epoch": 72.27288368891948, "grad_norm": 17.89800262451172, "learning_rate": 1.3863558155540263e-05, "loss": 0.7272, "step": 420050 }, { "epoch": 72.2814865794907, "grad_norm": 8.664689064025879, "learning_rate": 1.3859256710254646e-05, "loss": 0.6956, "step": 420100 }, { "epoch": 72.29008947006194, "grad_norm": 15.252579689025879, "learning_rate": 1.385495526496903e-05, "loss": 0.6652, "step": 420150 }, { "epoch": 72.29869236063317, "grad_norm": 17.630704879760742, "learning_rate": 1.3850653819683415e-05, "loss": 0.6877, "step": 420200 }, { "epoch": 72.30729525120441, "grad_norm": 21.65201759338379, "learning_rate": 1.3846352374397798e-05, "loss": 0.64, "step": 420250 }, { "epoch": 72.31589814177563, "grad_norm": 21.158403396606445, "learning_rate": 1.3842050929112182e-05, "loss": 0.7068, "step": 420300 }, { "epoch": 72.32450103234687, "grad_norm": 15.770915985107422, "learning_rate": 1.3837749483826568e-05, "loss": 0.7162, "step": 420350 }, { "epoch": 72.3331039229181, "grad_norm": 12.654997825622559, "learning_rate": 1.383344803854095e-05, "loss": 0.6882, "step": 420400 }, { "epoch": 72.34170681348934, "grad_norm": 17.419214248657227, "learning_rate": 1.3829146593255334e-05, "loss": 0.6826, "step": 420450 }, { "epoch": 72.35030970406056, "grad_norm": 10.350272178649902, "learning_rate": 1.3824845147969717e-05, "loss": 0.6351, "step": 420500 }, { "epoch": 72.35891259463179, "grad_norm": 20.08740234375, "learning_rate": 1.3820543702684101e-05, "loss": 0.7076, "step": 420550 }, { "epoch": 72.36751548520303, "grad_norm": 17.64190673828125, "learning_rate": 1.3816242257398488e-05, "loss": 0.6981, "step": 420600 }, { "epoch": 72.37611837577425, "grad_norm": 8.952112197875977, "learning_rate": 1.3811940812112869e-05, "loss": 0.6634, "step": 420650 }, { "epoch": 72.3847212663455, "grad_norm": 17.99379539489746, "learning_rate": 1.3807639366827255e-05, "loss": 0.6714, "step": 420700 }, { "epoch": 72.39332415691672, "grad_norm": 20.750526428222656, "learning_rate": 1.380333792154164e-05, "loss": 0.6962, "step": 420750 }, { "epoch": 72.40192704748796, "grad_norm": 21.377540588378906, "learning_rate": 1.3799036476256023e-05, "loss": 0.7028, "step": 420800 }, { "epoch": 72.41052993805918, "grad_norm": 23.490633010864258, "learning_rate": 1.3794735030970407e-05, "loss": 0.656, "step": 420850 }, { "epoch": 72.41913282863042, "grad_norm": 15.001639366149902, "learning_rate": 1.379043358568479e-05, "loss": 0.6849, "step": 420900 }, { "epoch": 72.42773571920165, "grad_norm": 20.747913360595703, "learning_rate": 1.3786132140399175e-05, "loss": 0.6882, "step": 420950 }, { "epoch": 72.43633860977289, "grad_norm": 12.120789527893066, "learning_rate": 1.3781830695113559e-05, "loss": 0.727, "step": 421000 }, { "epoch": 72.44494150034411, "grad_norm": 18.199867248535156, "learning_rate": 1.3777529249827942e-05, "loss": 0.7088, "step": 421050 }, { "epoch": 72.45354439091535, "grad_norm": 9.847101211547852, "learning_rate": 1.3773227804542327e-05, "loss": 0.696, "step": 421100 }, { "epoch": 72.46214728148658, "grad_norm": 19.629308700561523, "learning_rate": 1.3768926359256711e-05, "loss": 0.6777, "step": 421150 }, { "epoch": 72.4707501720578, "grad_norm": 20.392179489135742, "learning_rate": 1.3764624913971094e-05, "loss": 0.6855, "step": 421200 }, { "epoch": 72.47935306262904, "grad_norm": 11.36807632446289, "learning_rate": 1.3760323468685479e-05, "loss": 0.6816, "step": 421250 }, { "epoch": 72.48795595320027, "grad_norm": 8.795166015625, "learning_rate": 1.3756022023399865e-05, "loss": 0.6757, "step": 421300 }, { "epoch": 72.49655884377151, "grad_norm": 8.885231018066406, "learning_rate": 1.3751720578114246e-05, "loss": 0.7026, "step": 421350 }, { "epoch": 72.50516173434274, "grad_norm": 11.588666915893555, "learning_rate": 1.3747419132828632e-05, "loss": 0.6996, "step": 421400 }, { "epoch": 72.51376462491397, "grad_norm": 19.339603424072266, "learning_rate": 1.3743117687543013e-05, "loss": 0.7471, "step": 421450 }, { "epoch": 72.5223675154852, "grad_norm": 17.046321868896484, "learning_rate": 1.37388162422574e-05, "loss": 0.7426, "step": 421500 }, { "epoch": 72.53097040605644, "grad_norm": 18.55270004272461, "learning_rate": 1.3734514796971784e-05, "loss": 0.6583, "step": 421550 }, { "epoch": 72.53957329662767, "grad_norm": 18.394210815429688, "learning_rate": 1.3730213351686167e-05, "loss": 0.7103, "step": 421600 }, { "epoch": 72.5481761871989, "grad_norm": 14.79342269897461, "learning_rate": 1.3725911906400552e-05, "loss": 0.6583, "step": 421650 }, { "epoch": 72.55677907777013, "grad_norm": 12.051802635192871, "learning_rate": 1.3721610461114936e-05, "loss": 0.6861, "step": 421700 }, { "epoch": 72.56538196834136, "grad_norm": 18.061016082763672, "learning_rate": 1.3717309015829319e-05, "loss": 0.7077, "step": 421750 }, { "epoch": 72.5739848589126, "grad_norm": 6.97931432723999, "learning_rate": 1.3713007570543704e-05, "loss": 0.6695, "step": 421800 }, { "epoch": 72.58258774948382, "grad_norm": 12.58266830444336, "learning_rate": 1.3708706125258087e-05, "loss": 0.6785, "step": 421850 }, { "epoch": 72.59119064005506, "grad_norm": 13.642914772033691, "learning_rate": 1.3704404679972471e-05, "loss": 0.7246, "step": 421900 }, { "epoch": 72.59979353062629, "grad_norm": 16.857585906982422, "learning_rate": 1.3700103234686856e-05, "loss": 0.6906, "step": 421950 }, { "epoch": 72.60839642119753, "grad_norm": 11.949477195739746, "learning_rate": 1.3695801789401238e-05, "loss": 0.6947, "step": 422000 }, { "epoch": 72.61699931176875, "grad_norm": 27.290664672851562, "learning_rate": 1.3691500344115623e-05, "loss": 0.7083, "step": 422050 }, { "epoch": 72.62560220233999, "grad_norm": 14.97168254852295, "learning_rate": 1.368719889883001e-05, "loss": 0.667, "step": 422100 }, { "epoch": 72.63420509291122, "grad_norm": 7.103188514709473, "learning_rate": 1.368289745354439e-05, "loss": 0.6604, "step": 422150 }, { "epoch": 72.64280798348246, "grad_norm": 14.470048904418945, "learning_rate": 1.3678596008258777e-05, "loss": 0.7106, "step": 422200 }, { "epoch": 72.65141087405368, "grad_norm": 22.89559555053711, "learning_rate": 1.3674294562973158e-05, "loss": 0.6812, "step": 422250 }, { "epoch": 72.66001376462492, "grad_norm": 12.686165809631348, "learning_rate": 1.3669993117687544e-05, "loss": 0.6722, "step": 422300 }, { "epoch": 72.66861665519615, "grad_norm": 19.596267700195312, "learning_rate": 1.3665691672401929e-05, "loss": 0.6977, "step": 422350 }, { "epoch": 72.67721954576737, "grad_norm": 12.496530532836914, "learning_rate": 1.3661390227116312e-05, "loss": 0.7138, "step": 422400 }, { "epoch": 72.68582243633861, "grad_norm": 12.853073120117188, "learning_rate": 1.3657088781830696e-05, "loss": 0.6835, "step": 422450 }, { "epoch": 72.69442532690984, "grad_norm": 17.717634201049805, "learning_rate": 1.365278733654508e-05, "loss": 0.6903, "step": 422500 }, { "epoch": 72.70302821748108, "grad_norm": 18.11112403869629, "learning_rate": 1.3648485891259464e-05, "loss": 0.6915, "step": 422550 }, { "epoch": 72.7116311080523, "grad_norm": 12.86721134185791, "learning_rate": 1.3644184445973848e-05, "loss": 0.7485, "step": 422600 }, { "epoch": 72.72023399862354, "grad_norm": 17.377016067504883, "learning_rate": 1.3639883000688231e-05, "loss": 0.6807, "step": 422650 }, { "epoch": 72.72883688919477, "grad_norm": 13.178459167480469, "learning_rate": 1.3635581555402616e-05, "loss": 0.6569, "step": 422700 }, { "epoch": 72.737439779766, "grad_norm": 18.19026756286621, "learning_rate": 1.3631280110117e-05, "loss": 0.6908, "step": 422750 }, { "epoch": 72.74604267033723, "grad_norm": 12.105598449707031, "learning_rate": 1.3626978664831383e-05, "loss": 0.686, "step": 422800 }, { "epoch": 72.75464556090847, "grad_norm": 13.906272888183594, "learning_rate": 1.3622677219545768e-05, "loss": 0.6746, "step": 422850 }, { "epoch": 72.7632484514797, "grad_norm": 20.637182235717773, "learning_rate": 1.3618375774260154e-05, "loss": 0.6608, "step": 422900 }, { "epoch": 72.77185134205092, "grad_norm": 13.446335792541504, "learning_rate": 1.3614074328974535e-05, "loss": 0.7092, "step": 422950 }, { "epoch": 72.78045423262216, "grad_norm": 18.341609954833984, "learning_rate": 1.3609772883688921e-05, "loss": 0.6654, "step": 423000 }, { "epoch": 72.78905712319339, "grad_norm": 11.649510383605957, "learning_rate": 1.3605471438403306e-05, "loss": 0.6889, "step": 423050 }, { "epoch": 72.79766001376463, "grad_norm": 25.396642684936523, "learning_rate": 1.3601169993117687e-05, "loss": 0.678, "step": 423100 }, { "epoch": 72.80626290433585, "grad_norm": 17.346092224121094, "learning_rate": 1.3596868547832073e-05, "loss": 0.6384, "step": 423150 }, { "epoch": 72.81486579490709, "grad_norm": 21.010953903198242, "learning_rate": 1.3592567102546454e-05, "loss": 0.6938, "step": 423200 }, { "epoch": 72.82346868547832, "grad_norm": 11.542535781860352, "learning_rate": 1.358826565726084e-05, "loss": 0.7233, "step": 423250 }, { "epoch": 72.83207157604956, "grad_norm": 15.120718955993652, "learning_rate": 1.3583964211975225e-05, "loss": 0.7527, "step": 423300 }, { "epoch": 72.84067446662078, "grad_norm": 10.543143272399902, "learning_rate": 1.3579662766689608e-05, "loss": 0.6993, "step": 423350 }, { "epoch": 72.84927735719202, "grad_norm": 18.829946517944336, "learning_rate": 1.3575361321403993e-05, "loss": 0.7026, "step": 423400 }, { "epoch": 72.85788024776325, "grad_norm": 10.119478225708008, "learning_rate": 1.3571059876118377e-05, "loss": 0.7369, "step": 423450 }, { "epoch": 72.86648313833447, "grad_norm": 19.775774002075195, "learning_rate": 1.356675843083276e-05, "loss": 0.7018, "step": 423500 }, { "epoch": 72.87508602890571, "grad_norm": 19.641450881958008, "learning_rate": 1.3562456985547145e-05, "loss": 0.7104, "step": 423550 }, { "epoch": 72.88368891947694, "grad_norm": 8.3142728805542, "learning_rate": 1.3558155540261527e-05, "loss": 0.7143, "step": 423600 }, { "epoch": 72.89229181004818, "grad_norm": 13.9188871383667, "learning_rate": 1.3553854094975912e-05, "loss": 0.7474, "step": 423650 }, { "epoch": 72.9008947006194, "grad_norm": 12.070127487182617, "learning_rate": 1.3549552649690298e-05, "loss": 0.7163, "step": 423700 }, { "epoch": 72.90949759119064, "grad_norm": 6.46368408203125, "learning_rate": 1.354525120440468e-05, "loss": 0.6723, "step": 423750 }, { "epoch": 72.91810048176187, "grad_norm": 14.607870101928711, "learning_rate": 1.3540949759119064e-05, "loss": 0.6999, "step": 423800 }, { "epoch": 72.92670337233311, "grad_norm": 19.24581527709961, "learning_rate": 1.353664831383345e-05, "loss": 0.747, "step": 423850 }, { "epoch": 72.93530626290433, "grad_norm": 17.3234920501709, "learning_rate": 1.3532346868547831e-05, "loss": 0.6945, "step": 423900 }, { "epoch": 72.94390915347557, "grad_norm": 12.813576698303223, "learning_rate": 1.3528045423262218e-05, "loss": 0.6975, "step": 423950 }, { "epoch": 72.9525120440468, "grad_norm": 17.197189331054688, "learning_rate": 1.3523743977976599e-05, "loss": 0.7054, "step": 424000 }, { "epoch": 72.96111493461804, "grad_norm": 12.779213905334473, "learning_rate": 1.3519442532690985e-05, "loss": 0.7322, "step": 424050 }, { "epoch": 72.96971782518926, "grad_norm": 12.914210319519043, "learning_rate": 1.351514108740537e-05, "loss": 0.7047, "step": 424100 }, { "epoch": 72.97832071576049, "grad_norm": 12.956335067749023, "learning_rate": 1.3510839642119753e-05, "loss": 0.6733, "step": 424150 }, { "epoch": 72.98692360633173, "grad_norm": 21.765235900878906, "learning_rate": 1.3506538196834137e-05, "loss": 0.7008, "step": 424200 }, { "epoch": 72.99552649690295, "grad_norm": 14.405112266540527, "learning_rate": 1.3502236751548522e-05, "loss": 0.6815, "step": 424250 }, { "epoch": 73.0, "eval_accuracy": 0.5550679745310618, "eval_f1": 0.5521078199754916, "eval_f1_DuraRiadoRio_16x16": 0.4205158968206359, "eval_f1_Mole_16x16": 0.6171821305841925, "eval_f1_Quebrado_16x16": 0.6999531176746366, "eval_f1_RiadoRio_16x16": 0.45333846449874926, "eval_f1_RioFechado_16x16": 0.5695494902992436, "eval_loss": 2.349670171737671, "eval_precision": 0.5952913885922252, "eval_precision_DuraRiadoRio_16x16": 0.6805825242718446, "eval_precision_Mole_16x16": 0.653566229985444, "eval_precision_Quebrado_16x16": 0.7609582059123343, "eval_precision_RiadoRio_16x16": 0.42603978300180834, "eval_precision_RioFechado_16x16": 0.45531019978969506, "eval_recall": 0.5563166855672618, "eval_recall_DuraRiadoRio_16x16": 0.3042534722222222, "eval_recall_Mole_16x16": 0.5846354166666666, "eval_recall_Quebrado_16x16": 0.6480034722222222, "eval_recall_RiadoRio_16x16": 0.484375, "eval_recall_RioFechado_16x16": 0.7603160667251976, "eval_runtime": 46.1703, "eval_samples_per_second": 251.72, "eval_steps_per_second": 15.746, "step": 424276 }, { "epoch": 73.0041293874742, "grad_norm": 14.086202621459961, "learning_rate": 1.3497935306262905e-05, "loss": 0.6791, "step": 424300 }, { "epoch": 73.01273227804542, "grad_norm": 12.276403427124023, "learning_rate": 1.3493633860977289e-05, "loss": 0.732, "step": 424350 }, { "epoch": 73.02133516861666, "grad_norm": 17.875524520874023, "learning_rate": 1.3489332415691672e-05, "loss": 0.7215, "step": 424400 }, { "epoch": 73.02993805918788, "grad_norm": 7.722130298614502, "learning_rate": 1.3485030970406057e-05, "loss": 0.69, "step": 424450 }, { "epoch": 73.03854094975912, "grad_norm": 16.74342155456543, "learning_rate": 1.3480729525120441e-05, "loss": 0.7305, "step": 424500 }, { "epoch": 73.04714384033035, "grad_norm": 13.390728950500488, "learning_rate": 1.3476428079834824e-05, "loss": 0.6432, "step": 424550 }, { "epoch": 73.05574673090159, "grad_norm": 18.59706687927246, "learning_rate": 1.3472126634549209e-05, "loss": 0.6414, "step": 424600 }, { "epoch": 73.06434962147281, "grad_norm": 10.943801879882812, "learning_rate": 1.3467825189263595e-05, "loss": 0.7134, "step": 424650 }, { "epoch": 73.07295251204404, "grad_norm": 14.09909439086914, "learning_rate": 1.3463523743977976e-05, "loss": 0.6637, "step": 424700 }, { "epoch": 73.08155540261528, "grad_norm": 8.60055160522461, "learning_rate": 1.3459222298692362e-05, "loss": 0.6699, "step": 424750 }, { "epoch": 73.0901582931865, "grad_norm": 10.180364608764648, "learning_rate": 1.3454920853406747e-05, "loss": 0.7259, "step": 424800 }, { "epoch": 73.09876118375774, "grad_norm": 6.607761859893799, "learning_rate": 1.345061940812113e-05, "loss": 0.7041, "step": 424850 }, { "epoch": 73.10736407432897, "grad_norm": 13.197408676147461, "learning_rate": 1.3446317962835514e-05, "loss": 0.6953, "step": 424900 }, { "epoch": 73.11596696490021, "grad_norm": 17.97299575805664, "learning_rate": 1.3442016517549897e-05, "loss": 0.6756, "step": 424950 }, { "epoch": 73.12456985547144, "grad_norm": 8.423301696777344, "learning_rate": 1.3437715072264282e-05, "loss": 0.6647, "step": 425000 }, { "epoch": 73.13317274604267, "grad_norm": 17.52586555480957, "learning_rate": 1.3433413626978666e-05, "loss": 0.6618, "step": 425050 }, { "epoch": 73.1417756366139, "grad_norm": 13.929011344909668, "learning_rate": 1.3429112181693049e-05, "loss": 0.6927, "step": 425100 }, { "epoch": 73.15037852718514, "grad_norm": 14.306713104248047, "learning_rate": 1.3424810736407434e-05, "loss": 0.6987, "step": 425150 }, { "epoch": 73.15898141775637, "grad_norm": 20.476289749145508, "learning_rate": 1.3420509291121818e-05, "loss": 0.681, "step": 425200 }, { "epoch": 73.1675843083276, "grad_norm": 10.79399299621582, "learning_rate": 1.3416207845836201e-05, "loss": 0.7672, "step": 425250 }, { "epoch": 73.17618719889883, "grad_norm": 17.868106842041016, "learning_rate": 1.3411906400550586e-05, "loss": 0.6431, "step": 425300 }, { "epoch": 73.18479008947006, "grad_norm": 16.647289276123047, "learning_rate": 1.3407604955264968e-05, "loss": 0.7345, "step": 425350 }, { "epoch": 73.1933929800413, "grad_norm": 19.671100616455078, "learning_rate": 1.3403303509979353e-05, "loss": 0.6794, "step": 425400 }, { "epoch": 73.20199587061252, "grad_norm": 8.039203643798828, "learning_rate": 1.339900206469374e-05, "loss": 0.653, "step": 425450 }, { "epoch": 73.21059876118376, "grad_norm": 5.613874912261963, "learning_rate": 1.339470061940812e-05, "loss": 0.6613, "step": 425500 }, { "epoch": 73.21920165175499, "grad_norm": 15.056013107299805, "learning_rate": 1.3390399174122507e-05, "loss": 0.6651, "step": 425550 }, { "epoch": 73.22780454232623, "grad_norm": 17.203004837036133, "learning_rate": 1.3386097728836891e-05, "loss": 0.7095, "step": 425600 }, { "epoch": 73.23640743289745, "grad_norm": 16.620527267456055, "learning_rate": 1.3381796283551274e-05, "loss": 0.6621, "step": 425650 }, { "epoch": 73.24501032346869, "grad_norm": 17.801219940185547, "learning_rate": 1.3377494838265659e-05, "loss": 0.6732, "step": 425700 }, { "epoch": 73.25361321403992, "grad_norm": 12.080306053161621, "learning_rate": 1.337319339298004e-05, "loss": 0.6918, "step": 425750 }, { "epoch": 73.26221610461116, "grad_norm": 22.649335861206055, "learning_rate": 1.3368891947694426e-05, "loss": 0.6796, "step": 425800 }, { "epoch": 73.27081899518238, "grad_norm": 20.325275421142578, "learning_rate": 1.336459050240881e-05, "loss": 0.721, "step": 425850 }, { "epoch": 73.2794218857536, "grad_norm": 14.666114807128906, "learning_rate": 1.3360289057123194e-05, "loss": 0.6909, "step": 425900 }, { "epoch": 73.28802477632485, "grad_norm": 16.218120574951172, "learning_rate": 1.3355987611837578e-05, "loss": 0.7096, "step": 425950 }, { "epoch": 73.29662766689607, "grad_norm": 11.663880348205566, "learning_rate": 1.3351686166551963e-05, "loss": 0.6597, "step": 426000 }, { "epoch": 73.30523055746731, "grad_norm": 16.061477661132812, "learning_rate": 1.3347384721266346e-05, "loss": 0.6853, "step": 426050 }, { "epoch": 73.31383344803854, "grad_norm": 14.460805892944336, "learning_rate": 1.334308327598073e-05, "loss": 0.6757, "step": 426100 }, { "epoch": 73.32243633860978, "grad_norm": 12.316174507141113, "learning_rate": 1.3338781830695113e-05, "loss": 0.6891, "step": 426150 }, { "epoch": 73.331039229181, "grad_norm": 16.780067443847656, "learning_rate": 1.3334480385409498e-05, "loss": 0.6915, "step": 426200 }, { "epoch": 73.33964211975224, "grad_norm": 13.846477508544922, "learning_rate": 1.3330178940123884e-05, "loss": 0.6714, "step": 426250 }, { "epoch": 73.34824501032347, "grad_norm": 11.74173355102539, "learning_rate": 1.3325877494838265e-05, "loss": 0.642, "step": 426300 }, { "epoch": 73.3568479008947, "grad_norm": 13.29859733581543, "learning_rate": 1.3321576049552651e-05, "loss": 0.6776, "step": 426350 }, { "epoch": 73.36545079146593, "grad_norm": 17.05843162536621, "learning_rate": 1.3317274604267036e-05, "loss": 0.7182, "step": 426400 }, { "epoch": 73.37405368203716, "grad_norm": 16.23416519165039, "learning_rate": 1.3312973158981417e-05, "loss": 0.7, "step": 426450 }, { "epoch": 73.3826565726084, "grad_norm": 17.76517105102539, "learning_rate": 1.3308671713695803e-05, "loss": 0.7222, "step": 426500 }, { "epoch": 73.39125946317962, "grad_norm": 20.9952392578125, "learning_rate": 1.3304370268410184e-05, "loss": 0.6752, "step": 426550 }, { "epoch": 73.39986235375086, "grad_norm": 18.445205688476562, "learning_rate": 1.330006882312457e-05, "loss": 0.7212, "step": 426600 }, { "epoch": 73.40846524432209, "grad_norm": 17.3917179107666, "learning_rate": 1.3295767377838955e-05, "loss": 0.7046, "step": 426650 }, { "epoch": 73.41706813489333, "grad_norm": 23.404672622680664, "learning_rate": 1.3291465932553338e-05, "loss": 0.7361, "step": 426700 }, { "epoch": 73.42567102546455, "grad_norm": 19.276657104492188, "learning_rate": 1.3287164487267723e-05, "loss": 0.6774, "step": 426750 }, { "epoch": 73.43427391603579, "grad_norm": 11.787214279174805, "learning_rate": 1.3282863041982107e-05, "loss": 0.6848, "step": 426800 }, { "epoch": 73.44287680660702, "grad_norm": 11.043753623962402, "learning_rate": 1.327856159669649e-05, "loss": 0.6518, "step": 426850 }, { "epoch": 73.45147969717826, "grad_norm": 13.881726264953613, "learning_rate": 1.3274260151410875e-05, "loss": 0.7338, "step": 426900 }, { "epoch": 73.46008258774948, "grad_norm": 9.997941017150879, "learning_rate": 1.3269958706125261e-05, "loss": 0.6663, "step": 426950 }, { "epoch": 73.46868547832072, "grad_norm": 13.748666763305664, "learning_rate": 1.3265657260839642e-05, "loss": 0.7244, "step": 427000 }, { "epoch": 73.47728836889195, "grad_norm": 17.176233291625977, "learning_rate": 1.3261355815554027e-05, "loss": 0.6851, "step": 427050 }, { "epoch": 73.48589125946317, "grad_norm": 7.078469276428223, "learning_rate": 1.325705437026841e-05, "loss": 0.6697, "step": 427100 }, { "epoch": 73.49449415003441, "grad_norm": 14.91309642791748, "learning_rate": 1.3252752924982794e-05, "loss": 0.6568, "step": 427150 }, { "epoch": 73.50309704060564, "grad_norm": 21.572649002075195, "learning_rate": 1.324845147969718e-05, "loss": 0.6922, "step": 427200 }, { "epoch": 73.51169993117688, "grad_norm": 9.211779594421387, "learning_rate": 1.3244150034411561e-05, "loss": 0.7246, "step": 427250 }, { "epoch": 73.5203028217481, "grad_norm": 26.22748565673828, "learning_rate": 1.3239848589125948e-05, "loss": 0.7409, "step": 427300 }, { "epoch": 73.52890571231934, "grad_norm": 10.961731910705566, "learning_rate": 1.3235547143840332e-05, "loss": 0.6672, "step": 427350 }, { "epoch": 73.53750860289057, "grad_norm": 9.515398979187012, "learning_rate": 1.3231245698554715e-05, "loss": 0.662, "step": 427400 }, { "epoch": 73.54611149346181, "grad_norm": 12.412062644958496, "learning_rate": 1.32269442532691e-05, "loss": 0.6718, "step": 427450 }, { "epoch": 73.55471438403303, "grad_norm": 13.575095176696777, "learning_rate": 1.3222642807983483e-05, "loss": 0.711, "step": 427500 }, { "epoch": 73.56331727460427, "grad_norm": 13.716934204101562, "learning_rate": 1.3218341362697867e-05, "loss": 0.6476, "step": 427550 }, { "epoch": 73.5719201651755, "grad_norm": 12.796754837036133, "learning_rate": 1.3214039917412252e-05, "loss": 0.6877, "step": 427600 }, { "epoch": 73.58052305574672, "grad_norm": 19.106714248657227, "learning_rate": 1.3209738472126635e-05, "loss": 0.7167, "step": 427650 }, { "epoch": 73.58912594631796, "grad_norm": 14.059812545776367, "learning_rate": 1.3205437026841019e-05, "loss": 0.7128, "step": 427700 }, { "epoch": 73.59772883688919, "grad_norm": 7.552093982696533, "learning_rate": 1.3201135581555404e-05, "loss": 0.6992, "step": 427750 }, { "epoch": 73.60633172746043, "grad_norm": 17.270702362060547, "learning_rate": 1.3196834136269787e-05, "loss": 0.7068, "step": 427800 }, { "epoch": 73.61493461803165, "grad_norm": 16.179054260253906, "learning_rate": 1.3192532690984171e-05, "loss": 0.6651, "step": 427850 }, { "epoch": 73.6235375086029, "grad_norm": 13.471710205078125, "learning_rate": 1.3188231245698554e-05, "loss": 0.68, "step": 427900 }, { "epoch": 73.63214039917412, "grad_norm": 8.907963752746582, "learning_rate": 1.3183929800412939e-05, "loss": 0.6954, "step": 427950 }, { "epoch": 73.64074328974536, "grad_norm": 12.619329452514648, "learning_rate": 1.3179628355127325e-05, "loss": 0.728, "step": 428000 }, { "epoch": 73.64934618031658, "grad_norm": 15.796602249145508, "learning_rate": 1.3175326909841706e-05, "loss": 0.7493, "step": 428050 }, { "epoch": 73.65794907088782, "grad_norm": 12.562566757202148, "learning_rate": 1.3171025464556092e-05, "loss": 0.7234, "step": 428100 }, { "epoch": 73.66655196145905, "grad_norm": 15.21731185913086, "learning_rate": 1.3166724019270477e-05, "loss": 0.6533, "step": 428150 }, { "epoch": 73.67515485203029, "grad_norm": 20.9017391204834, "learning_rate": 1.316242257398486e-05, "loss": 0.7093, "step": 428200 }, { "epoch": 73.68375774260151, "grad_norm": 17.772184371948242, "learning_rate": 1.3158121128699244e-05, "loss": 0.6717, "step": 428250 }, { "epoch": 73.69236063317274, "grad_norm": 15.377982139587402, "learning_rate": 1.3153819683413627e-05, "loss": 0.7026, "step": 428300 }, { "epoch": 73.70096352374398, "grad_norm": 15.0363187789917, "learning_rate": 1.3149518238128012e-05, "loss": 0.7007, "step": 428350 }, { "epoch": 73.7095664143152, "grad_norm": 16.649885177612305, "learning_rate": 1.3145216792842396e-05, "loss": 0.7073, "step": 428400 }, { "epoch": 73.71816930488644, "grad_norm": 20.639474868774414, "learning_rate": 1.3140915347556779e-05, "loss": 0.7067, "step": 428450 }, { "epoch": 73.72677219545767, "grad_norm": 13.364847183227539, "learning_rate": 1.3136613902271164e-05, "loss": 0.6928, "step": 428500 }, { "epoch": 73.73537508602891, "grad_norm": 17.799522399902344, "learning_rate": 1.3132312456985548e-05, "loss": 0.6599, "step": 428550 }, { "epoch": 73.74397797660014, "grad_norm": 20.423534393310547, "learning_rate": 1.3128011011699931e-05, "loss": 0.7353, "step": 428600 }, { "epoch": 73.75258086717137, "grad_norm": 18.310815811157227, "learning_rate": 1.3123709566414316e-05, "loss": 0.687, "step": 428650 }, { "epoch": 73.7611837577426, "grad_norm": 12.312110900878906, "learning_rate": 1.3119408121128702e-05, "loss": 0.7317, "step": 428700 }, { "epoch": 73.76978664831384, "grad_norm": 11.619399070739746, "learning_rate": 1.3115106675843083e-05, "loss": 0.6653, "step": 428750 }, { "epoch": 73.77838953888507, "grad_norm": 17.577159881591797, "learning_rate": 1.311080523055747e-05, "loss": 0.7504, "step": 428800 }, { "epoch": 73.78699242945629, "grad_norm": 12.001100540161133, "learning_rate": 1.310650378527185e-05, "loss": 0.7211, "step": 428850 }, { "epoch": 73.79559532002753, "grad_norm": 21.76767349243164, "learning_rate": 1.3102202339986237e-05, "loss": 0.715, "step": 428900 }, { "epoch": 73.80419821059876, "grad_norm": 18.25933265686035, "learning_rate": 1.3097900894700621e-05, "loss": 0.6764, "step": 428950 }, { "epoch": 73.81280110117, "grad_norm": 10.621068954467773, "learning_rate": 1.3093599449415004e-05, "loss": 0.6927, "step": 429000 }, { "epoch": 73.82140399174122, "grad_norm": 24.894969940185547, "learning_rate": 1.3089298004129389e-05, "loss": 0.671, "step": 429050 }, { "epoch": 73.83000688231246, "grad_norm": 7.154824256896973, "learning_rate": 1.3084996558843773e-05, "loss": 0.6792, "step": 429100 }, { "epoch": 73.83860977288369, "grad_norm": 15.388593673706055, "learning_rate": 1.3080695113558156e-05, "loss": 0.7243, "step": 429150 }, { "epoch": 73.84721266345493, "grad_norm": 23.096527099609375, "learning_rate": 1.307639366827254e-05, "loss": 0.674, "step": 429200 }, { "epoch": 73.85581555402615, "grad_norm": 9.14585018157959, "learning_rate": 1.3072092222986924e-05, "loss": 0.7151, "step": 429250 }, { "epoch": 73.86441844459739, "grad_norm": 14.579736709594727, "learning_rate": 1.3067790777701308e-05, "loss": 0.7052, "step": 429300 }, { "epoch": 73.87302133516862, "grad_norm": 19.835325241088867, "learning_rate": 1.3063489332415693e-05, "loss": 0.6961, "step": 429350 }, { "epoch": 73.88162422573984, "grad_norm": 14.395885467529297, "learning_rate": 1.3059187887130076e-05, "loss": 0.7353, "step": 429400 }, { "epoch": 73.89022711631108, "grad_norm": 15.608307838439941, "learning_rate": 1.305488644184446e-05, "loss": 0.641, "step": 429450 }, { "epoch": 73.8988300068823, "grad_norm": 19.222484588623047, "learning_rate": 1.3050584996558846e-05, "loss": 0.639, "step": 429500 }, { "epoch": 73.90743289745355, "grad_norm": 15.057713508605957, "learning_rate": 1.3046283551273227e-05, "loss": 0.6424, "step": 429550 }, { "epoch": 73.91603578802477, "grad_norm": 19.220691680908203, "learning_rate": 1.3041982105987614e-05, "loss": 0.6753, "step": 429600 }, { "epoch": 73.92463867859601, "grad_norm": 15.151973724365234, "learning_rate": 1.3037680660701995e-05, "loss": 0.6884, "step": 429650 }, { "epoch": 73.93324156916724, "grad_norm": 13.752676010131836, "learning_rate": 1.303337921541638e-05, "loss": 0.6825, "step": 429700 }, { "epoch": 73.94184445973848, "grad_norm": 10.699544906616211, "learning_rate": 1.3029077770130766e-05, "loss": 0.7067, "step": 429750 }, { "epoch": 73.9504473503097, "grad_norm": 9.217819213867188, "learning_rate": 1.3024776324845147e-05, "loss": 0.7261, "step": 429800 }, { "epoch": 73.95905024088094, "grad_norm": 14.315754890441895, "learning_rate": 1.3020474879559533e-05, "loss": 0.6767, "step": 429850 }, { "epoch": 73.96765313145217, "grad_norm": 17.258907318115234, "learning_rate": 1.3016173434273918e-05, "loss": 0.6533, "step": 429900 }, { "epoch": 73.9762560220234, "grad_norm": 15.366293907165527, "learning_rate": 1.30118719889883e-05, "loss": 0.7012, "step": 429950 }, { "epoch": 73.98485891259463, "grad_norm": 14.120633125305176, "learning_rate": 1.3007570543702685e-05, "loss": 0.648, "step": 430000 }, { "epoch": 73.99346180316586, "grad_norm": 11.717020988464355, "learning_rate": 1.3003269098417068e-05, "loss": 0.7438, "step": 430050 }, { "epoch": 74.0, "eval_accuracy": 0.48890036138358284, "eval_f1": 0.46104651455137535, "eval_f1_DuraRiadoRio_16x16": 0.4236170822585268, "eval_f1_Mole_16x16": 0.1806201550387597, "eval_f1_Quebrado_16x16": 0.7084139264990329, "eval_f1_RiadoRio_16x16": 0.4472893946578157, "eval_f1_RioFechado_16x16": 0.5452920143027413, "eval_loss": 3.655379056930542, "eval_precision": 0.6087020471608021, "eval_precision_DuraRiadoRio_16x16": 0.6236286919831223, "eval_precision_Mole_16x16": 0.8442028985507246, "eval_precision_Quebrado_16x16": 0.7996724890829694, "eval_precision_RiadoRio_16x16": 0.3632862644415918, "eval_precision_RioFechado_16x16": 0.41271989174560214, "eval_recall": 0.48857752244325325, "eval_recall_DuraRiadoRio_16x16": 0.3207465277777778, "eval_recall_Mole_16x16": 0.10112847222222222, "eval_recall_Quebrado_16x16": 0.6358506944444444, "eval_recall_RiadoRio_16x16": 0.5818256578947368, "eval_recall_RioFechado_16x16": 0.8033362598770851, "eval_runtime": 46.2439, "eval_samples_per_second": 251.32, "eval_steps_per_second": 15.721, "step": 430088 }, { "epoch": 74.0020646937371, "grad_norm": 19.219301223754883, "learning_rate": 1.2998967653131453e-05, "loss": 0.649, "step": 430100 }, { "epoch": 74.01066758430832, "grad_norm": 22.2974853515625, "learning_rate": 1.2994666207845837e-05, "loss": 0.6848, "step": 430150 }, { "epoch": 74.01927047487956, "grad_norm": 14.064760208129883, "learning_rate": 1.299036476256022e-05, "loss": 0.6853, "step": 430200 }, { "epoch": 74.02787336545079, "grad_norm": 16.98692512512207, "learning_rate": 1.2986063317274605e-05, "loss": 0.6721, "step": 430250 }, { "epoch": 74.03647625602203, "grad_norm": 13.814299583435059, "learning_rate": 1.2981761871988989e-05, "loss": 0.6668, "step": 430300 }, { "epoch": 74.04507914659325, "grad_norm": 17.961750030517578, "learning_rate": 1.2977460426703372e-05, "loss": 0.6426, "step": 430350 }, { "epoch": 74.05368203716449, "grad_norm": 20.1619873046875, "learning_rate": 1.2973158981417757e-05, "loss": 0.6569, "step": 430400 }, { "epoch": 74.06228492773572, "grad_norm": 13.231599807739258, "learning_rate": 1.2968857536132143e-05, "loss": 0.717, "step": 430450 }, { "epoch": 74.07088781830696, "grad_norm": 24.480981826782227, "learning_rate": 1.2964556090846524e-05, "loss": 0.67, "step": 430500 }, { "epoch": 74.07949070887818, "grad_norm": 12.606804847717285, "learning_rate": 1.296025464556091e-05, "loss": 0.6076, "step": 430550 }, { "epoch": 74.08809359944941, "grad_norm": 15.036300659179688, "learning_rate": 1.2955953200275291e-05, "loss": 0.7076, "step": 430600 }, { "epoch": 74.09669649002065, "grad_norm": 8.848679542541504, "learning_rate": 1.2951651754989678e-05, "loss": 0.6877, "step": 430650 }, { "epoch": 74.10529938059187, "grad_norm": 13.001049995422363, "learning_rate": 1.2947350309704062e-05, "loss": 0.6903, "step": 430700 }, { "epoch": 74.11390227116311, "grad_norm": 13.335180282592773, "learning_rate": 1.2943048864418445e-05, "loss": 0.7079, "step": 430750 }, { "epoch": 74.12250516173434, "grad_norm": 5.569223403930664, "learning_rate": 1.293874741913283e-05, "loss": 0.6055, "step": 430800 }, { "epoch": 74.13110805230558, "grad_norm": 18.313993453979492, "learning_rate": 1.2934445973847214e-05, "loss": 0.7112, "step": 430850 }, { "epoch": 74.1397109428768, "grad_norm": 19.85712242126465, "learning_rate": 1.2930144528561597e-05, "loss": 0.7209, "step": 430900 }, { "epoch": 74.14831383344804, "grad_norm": 13.76709270477295, "learning_rate": 1.2925843083275982e-05, "loss": 0.6479, "step": 430950 }, { "epoch": 74.15691672401927, "grad_norm": 14.36298656463623, "learning_rate": 1.2921541637990365e-05, "loss": 0.6883, "step": 431000 }, { "epoch": 74.16551961459051, "grad_norm": 16.48836898803711, "learning_rate": 1.2917240192704749e-05, "loss": 0.6645, "step": 431050 }, { "epoch": 74.17412250516173, "grad_norm": 17.394702911376953, "learning_rate": 1.2912938747419134e-05, "loss": 0.6729, "step": 431100 }, { "epoch": 74.18272539573297, "grad_norm": 11.608189582824707, "learning_rate": 1.2908637302133516e-05, "loss": 0.6619, "step": 431150 }, { "epoch": 74.1913282863042, "grad_norm": 15.361416816711426, "learning_rate": 1.2904335856847901e-05, "loss": 0.6383, "step": 431200 }, { "epoch": 74.19993117687542, "grad_norm": 12.741788864135742, "learning_rate": 1.2900034411562287e-05, "loss": 0.6909, "step": 431250 }, { "epoch": 74.20853406744666, "grad_norm": 14.580090522766113, "learning_rate": 1.2895732966276668e-05, "loss": 0.6482, "step": 431300 }, { "epoch": 74.21713695801789, "grad_norm": 21.993215560913086, "learning_rate": 1.2891431520991055e-05, "loss": 0.6815, "step": 431350 }, { "epoch": 74.22573984858913, "grad_norm": 12.041308403015137, "learning_rate": 1.2887130075705436e-05, "loss": 0.7265, "step": 431400 }, { "epoch": 74.23434273916035, "grad_norm": 14.7505521774292, "learning_rate": 1.2882828630419822e-05, "loss": 0.7269, "step": 431450 }, { "epoch": 74.2429456297316, "grad_norm": 13.200215339660645, "learning_rate": 1.2878527185134207e-05, "loss": 0.6467, "step": 431500 }, { "epoch": 74.25154852030282, "grad_norm": 16.039451599121094, "learning_rate": 1.287422573984859e-05, "loss": 0.6834, "step": 431550 }, { "epoch": 74.26015141087406, "grad_norm": 8.605850219726562, "learning_rate": 1.2869924294562974e-05, "loss": 0.6708, "step": 431600 }, { "epoch": 74.26875430144528, "grad_norm": 16.480775833129883, "learning_rate": 1.2865622849277359e-05, "loss": 0.674, "step": 431650 }, { "epoch": 74.27735719201652, "grad_norm": 7.347332000732422, "learning_rate": 1.2861321403991742e-05, "loss": 0.6974, "step": 431700 }, { "epoch": 74.28596008258775, "grad_norm": 22.867115020751953, "learning_rate": 1.2857019958706126e-05, "loss": 0.6864, "step": 431750 }, { "epoch": 74.29456297315897, "grad_norm": 10.993221282958984, "learning_rate": 1.2852718513420509e-05, "loss": 0.6868, "step": 431800 }, { "epoch": 74.30316586373021, "grad_norm": 9.515134811401367, "learning_rate": 1.2848417068134894e-05, "loss": 0.7251, "step": 431850 }, { "epoch": 74.31176875430144, "grad_norm": 15.280254364013672, "learning_rate": 1.2844115622849278e-05, "loss": 0.6339, "step": 431900 }, { "epoch": 74.32037164487268, "grad_norm": 12.186813354492188, "learning_rate": 1.2839814177563661e-05, "loss": 0.7353, "step": 431950 }, { "epoch": 74.3289745354439, "grad_norm": 13.715333938598633, "learning_rate": 1.2835512732278046e-05, "loss": 0.6995, "step": 432000 }, { "epoch": 74.33757742601514, "grad_norm": 8.151266098022461, "learning_rate": 1.2831211286992432e-05, "loss": 0.6447, "step": 432050 }, { "epoch": 74.34618031658637, "grad_norm": 13.398210525512695, "learning_rate": 1.2826909841706813e-05, "loss": 0.6903, "step": 432100 }, { "epoch": 74.35478320715761, "grad_norm": 14.52039623260498, "learning_rate": 1.28226083964212e-05, "loss": 0.6973, "step": 432150 }, { "epoch": 74.36338609772884, "grad_norm": 12.739805221557617, "learning_rate": 1.2818306951135584e-05, "loss": 0.6812, "step": 432200 }, { "epoch": 74.37198898830007, "grad_norm": 11.485918998718262, "learning_rate": 1.2814005505849967e-05, "loss": 0.7018, "step": 432250 }, { "epoch": 74.3805918788713, "grad_norm": 15.950337409973145, "learning_rate": 1.2809704060564351e-05, "loss": 0.7055, "step": 432300 }, { "epoch": 74.38919476944254, "grad_norm": 18.03301429748535, "learning_rate": 1.2805402615278732e-05, "loss": 0.6933, "step": 432350 }, { "epoch": 74.39779766001377, "grad_norm": 16.409046173095703, "learning_rate": 1.2801101169993119e-05, "loss": 0.7201, "step": 432400 }, { "epoch": 74.40640055058499, "grad_norm": 12.658576965332031, "learning_rate": 1.2796799724707503e-05, "loss": 0.7155, "step": 432450 }, { "epoch": 74.41500344115623, "grad_norm": 9.355297088623047, "learning_rate": 1.2792498279421886e-05, "loss": 0.701, "step": 432500 }, { "epoch": 74.42360633172746, "grad_norm": 18.164020538330078, "learning_rate": 1.278819683413627e-05, "loss": 0.7235, "step": 432550 }, { "epoch": 74.4322092222987, "grad_norm": 18.28485679626465, "learning_rate": 1.2783895388850655e-05, "loss": 0.6779, "step": 432600 }, { "epoch": 74.44081211286992, "grad_norm": 19.04896354675293, "learning_rate": 1.2779593943565038e-05, "loss": 0.7465, "step": 432650 }, { "epoch": 74.44941500344116, "grad_norm": 14.419571876525879, "learning_rate": 1.2775292498279423e-05, "loss": 0.671, "step": 432700 }, { "epoch": 74.45801789401239, "grad_norm": 13.33638858795166, "learning_rate": 1.2770991052993805e-05, "loss": 0.661, "step": 432750 }, { "epoch": 74.46662078458363, "grad_norm": 17.325010299682617, "learning_rate": 1.276668960770819e-05, "loss": 0.7047, "step": 432800 }, { "epoch": 74.47522367515485, "grad_norm": 10.39815616607666, "learning_rate": 1.2762388162422576e-05, "loss": 0.6624, "step": 432850 }, { "epoch": 74.48382656572609, "grad_norm": 12.640371322631836, "learning_rate": 1.2758086717136957e-05, "loss": 0.699, "step": 432900 }, { "epoch": 74.49242945629732, "grad_norm": 17.593507766723633, "learning_rate": 1.2753785271851342e-05, "loss": 0.7042, "step": 432950 }, { "epoch": 74.50103234686854, "grad_norm": 17.825105667114258, "learning_rate": 1.2749483826565728e-05, "loss": 0.7007, "step": 433000 }, { "epoch": 74.50963523743978, "grad_norm": 15.648697853088379, "learning_rate": 1.274518238128011e-05, "loss": 0.7163, "step": 433050 }, { "epoch": 74.518238128011, "grad_norm": 18.192089080810547, "learning_rate": 1.2740880935994496e-05, "loss": 0.6878, "step": 433100 }, { "epoch": 74.52684101858225, "grad_norm": 15.936273574829102, "learning_rate": 1.2736579490708877e-05, "loss": 0.6501, "step": 433150 }, { "epoch": 74.53544390915347, "grad_norm": 14.201883316040039, "learning_rate": 1.2732278045423263e-05, "loss": 0.6972, "step": 433200 }, { "epoch": 74.54404679972471, "grad_norm": 19.98264503479004, "learning_rate": 1.2727976600137648e-05, "loss": 0.6796, "step": 433250 }, { "epoch": 74.55264969029594, "grad_norm": 20.882049560546875, "learning_rate": 1.272367515485203e-05, "loss": 0.7113, "step": 433300 }, { "epoch": 74.56125258086718, "grad_norm": 22.260929107666016, "learning_rate": 1.2719373709566415e-05, "loss": 0.7243, "step": 433350 }, { "epoch": 74.5698554714384, "grad_norm": 11.835634231567383, "learning_rate": 1.27150722642808e-05, "loss": 0.6197, "step": 433400 }, { "epoch": 74.57845836200964, "grad_norm": 12.665645599365234, "learning_rate": 1.2710770818995183e-05, "loss": 0.7102, "step": 433450 }, { "epoch": 74.58706125258087, "grad_norm": 10.055181503295898, "learning_rate": 1.2706469373709567e-05, "loss": 0.6488, "step": 433500 }, { "epoch": 74.59566414315209, "grad_norm": 14.829768180847168, "learning_rate": 1.270216792842395e-05, "loss": 0.6562, "step": 433550 }, { "epoch": 74.60426703372333, "grad_norm": 21.095136642456055, "learning_rate": 1.2697866483138335e-05, "loss": 0.6461, "step": 433600 }, { "epoch": 74.61286992429456, "grad_norm": 14.786128997802734, "learning_rate": 1.2693565037852719e-05, "loss": 0.696, "step": 433650 }, { "epoch": 74.6214728148658, "grad_norm": 19.211124420166016, "learning_rate": 1.2689263592567102e-05, "loss": 0.5911, "step": 433700 }, { "epoch": 74.63007570543702, "grad_norm": 16.84640121459961, "learning_rate": 1.2684962147281487e-05, "loss": 0.6737, "step": 433750 }, { "epoch": 74.63867859600826, "grad_norm": 22.157419204711914, "learning_rate": 1.2680660701995873e-05, "loss": 0.713, "step": 433800 }, { "epoch": 74.64728148657949, "grad_norm": 17.937049865722656, "learning_rate": 1.2676359256710254e-05, "loss": 0.6794, "step": 433850 }, { "epoch": 74.65588437715073, "grad_norm": 10.787419319152832, "learning_rate": 1.267205781142464e-05, "loss": 0.766, "step": 433900 }, { "epoch": 74.66448726772195, "grad_norm": 14.014087677001953, "learning_rate": 1.2667756366139021e-05, "loss": 0.6898, "step": 433950 }, { "epoch": 74.67309015829319, "grad_norm": 19.856231689453125, "learning_rate": 1.2663454920853408e-05, "loss": 0.7172, "step": 434000 }, { "epoch": 74.68169304886442, "grad_norm": 14.416655540466309, "learning_rate": 1.2659153475567792e-05, "loss": 0.6303, "step": 434050 }, { "epoch": 74.69029593943566, "grad_norm": 21.682645797729492, "learning_rate": 1.2654852030282175e-05, "loss": 0.6832, "step": 434100 }, { "epoch": 74.69889883000688, "grad_norm": 8.790058135986328, "learning_rate": 1.265055058499656e-05, "loss": 0.6419, "step": 434150 }, { "epoch": 74.70750172057811, "grad_norm": 22.631370544433594, "learning_rate": 1.2646249139710944e-05, "loss": 0.6559, "step": 434200 }, { "epoch": 74.71610461114935, "grad_norm": 33.697139739990234, "learning_rate": 1.2641947694425327e-05, "loss": 0.7303, "step": 434250 }, { "epoch": 74.72470750172057, "grad_norm": 21.544597625732422, "learning_rate": 1.2637646249139712e-05, "loss": 0.7044, "step": 434300 }, { "epoch": 74.73331039229181, "grad_norm": 10.903624534606934, "learning_rate": 1.2633344803854096e-05, "loss": 0.7265, "step": 434350 }, { "epoch": 74.74191328286304, "grad_norm": 14.692438125610352, "learning_rate": 1.2629043358568479e-05, "loss": 0.6873, "step": 434400 }, { "epoch": 74.75051617343428, "grad_norm": 12.693252563476562, "learning_rate": 1.2624741913282864e-05, "loss": 0.7664, "step": 434450 }, { "epoch": 74.7591190640055, "grad_norm": 16.329896926879883, "learning_rate": 1.2620440467997246e-05, "loss": 0.7234, "step": 434500 }, { "epoch": 74.76772195457674, "grad_norm": 32.040401458740234, "learning_rate": 1.2616139022711631e-05, "loss": 0.6675, "step": 434550 }, { "epoch": 74.77632484514797, "grad_norm": 14.727715492248535, "learning_rate": 1.2611837577426017e-05, "loss": 0.6567, "step": 434600 }, { "epoch": 74.78492773571921, "grad_norm": 14.893020629882812, "learning_rate": 1.2607536132140398e-05, "loss": 0.6557, "step": 434650 }, { "epoch": 74.79353062629043, "grad_norm": 16.727840423583984, "learning_rate": 1.2603234686854785e-05, "loss": 0.7185, "step": 434700 }, { "epoch": 74.80213351686166, "grad_norm": 15.82629108428955, "learning_rate": 1.259893324156917e-05, "loss": 0.647, "step": 434750 }, { "epoch": 74.8107364074329, "grad_norm": 8.704681396484375, "learning_rate": 1.2594631796283552e-05, "loss": 0.6813, "step": 434800 }, { "epoch": 74.81933929800412, "grad_norm": 14.299966812133789, "learning_rate": 1.2590330350997937e-05, "loss": 0.7118, "step": 434850 }, { "epoch": 74.82794218857536, "grad_norm": 9.31298828125, "learning_rate": 1.258602890571232e-05, "loss": 0.6753, "step": 434900 }, { "epoch": 74.83654507914659, "grad_norm": 18.60012435913086, "learning_rate": 1.2581727460426704e-05, "loss": 0.7187, "step": 434950 }, { "epoch": 74.84514796971783, "grad_norm": 21.769527435302734, "learning_rate": 1.2577426015141089e-05, "loss": 0.6763, "step": 435000 }, { "epoch": 74.85375086028905, "grad_norm": 17.855819702148438, "learning_rate": 1.2573124569855472e-05, "loss": 0.688, "step": 435050 }, { "epoch": 74.8623537508603, "grad_norm": 19.108558654785156, "learning_rate": 1.2568823124569856e-05, "loss": 0.6971, "step": 435100 }, { "epoch": 74.87095664143152, "grad_norm": 18.40499496459961, "learning_rate": 1.256452167928424e-05, "loss": 0.7346, "step": 435150 }, { "epoch": 74.87955953200276, "grad_norm": 14.49284839630127, "learning_rate": 1.2560220233998624e-05, "loss": 0.6822, "step": 435200 }, { "epoch": 74.88816242257398, "grad_norm": 15.994492530822754, "learning_rate": 1.2555918788713008e-05, "loss": 0.6442, "step": 435250 }, { "epoch": 74.89676531314522, "grad_norm": 20.678266525268555, "learning_rate": 1.2551617343427391e-05, "loss": 0.7088, "step": 435300 }, { "epoch": 74.90536820371645, "grad_norm": 20.978755950927734, "learning_rate": 1.2547315898141776e-05, "loss": 0.717, "step": 435350 }, { "epoch": 74.91397109428767, "grad_norm": 13.39617919921875, "learning_rate": 1.2543014452856162e-05, "loss": 0.6862, "step": 435400 }, { "epoch": 74.92257398485891, "grad_norm": 18.85380744934082, "learning_rate": 1.2538713007570543e-05, "loss": 0.6791, "step": 435450 }, { "epoch": 74.93117687543014, "grad_norm": 19.809738159179688, "learning_rate": 1.253441156228493e-05, "loss": 0.6924, "step": 435500 }, { "epoch": 74.93977976600138, "grad_norm": 12.250764846801758, "learning_rate": 1.2530110116999314e-05, "loss": 0.7282, "step": 435550 }, { "epoch": 74.9483826565726, "grad_norm": 26.320669174194336, "learning_rate": 1.2525808671713695e-05, "loss": 0.7075, "step": 435600 }, { "epoch": 74.95698554714384, "grad_norm": 17.782947540283203, "learning_rate": 1.2521507226428081e-05, "loss": 0.713, "step": 435650 }, { "epoch": 74.96558843771507, "grad_norm": 9.811363220214844, "learning_rate": 1.2517205781142462e-05, "loss": 0.6483, "step": 435700 }, { "epoch": 74.97419132828631, "grad_norm": 8.5053071975708, "learning_rate": 1.2512904335856849e-05, "loss": 0.7117, "step": 435750 }, { "epoch": 74.98279421885753, "grad_norm": 17.142580032348633, "learning_rate": 1.2508602890571233e-05, "loss": 0.681, "step": 435800 }, { "epoch": 74.99139710942877, "grad_norm": 21.871728897094727, "learning_rate": 1.2504301445285616e-05, "loss": 0.6369, "step": 435850 }, { "epoch": 75.0, "grad_norm": 35.06055450439453, "learning_rate": 1.25e-05, "loss": 0.6709, "step": 435900 }, { "epoch": 75.0, "eval_accuracy": 0.6007571846498021, "eval_f1": 0.5883582334486256, "eval_f1_DuraRiadoRio_16x16": 0.49104436483879854, "eval_f1_Mole_16x16": 0.6419348227095716, "eval_f1_Quebrado_16x16": 0.714047751249306, "eval_f1_RiadoRio_16x16": 0.4616120545324085, "eval_f1_RioFechado_16x16": 0.6331521739130435, "eval_loss": 2.006608486175537, "eval_precision": 0.6044625242113483, "eval_precision_DuraRiadoRio_16x16": 0.6724528301886793, "eval_precision_Mole_16x16": 0.6081553398058253, "eval_precision_Quebrado_16x16": 0.6224588576960309, "eval_precision_RiadoRio_16x16": 0.551743853630646, "eval_precision_RioFechado_16x16": 0.5675017397355602, "eval_recall": 0.603283505075243, "eval_recall_DuraRiadoRio_16x16": 0.38671875, "eval_recall_Mole_16x16": 0.6796875, "eval_recall_Quebrado_16x16": 0.8372395833333334, "eval_recall_RiadoRio_16x16": 0.39679276315789475, "eval_recall_RioFechado_16x16": 0.7159789288849868, "eval_runtime": 45.8636, "eval_samples_per_second": 253.404, "eval_steps_per_second": 15.851, "step": 435900 }, { "epoch": 75.00860289057123, "grad_norm": 12.711277961730957, "learning_rate": 1.2495698554714383e-05, "loss": 0.625, "step": 435950 }, { "epoch": 75.01720578114247, "grad_norm": 14.381049156188965, "learning_rate": 1.249139710942877e-05, "loss": 0.7011, "step": 436000 }, { "epoch": 75.02580867171369, "grad_norm": 19.515031814575195, "learning_rate": 1.2487095664143153e-05, "loss": 0.6359, "step": 436050 }, { "epoch": 75.03441156228493, "grad_norm": 15.212631225585938, "learning_rate": 1.2482794218857537e-05, "loss": 0.639, "step": 436100 }, { "epoch": 75.04301445285616, "grad_norm": 16.238388061523438, "learning_rate": 1.247849277357192e-05, "loss": 0.7392, "step": 436150 }, { "epoch": 75.0516173434274, "grad_norm": 19.08497428894043, "learning_rate": 1.2474191328286306e-05, "loss": 0.702, "step": 436200 }, { "epoch": 75.06022023399862, "grad_norm": 9.622662544250488, "learning_rate": 1.2469889883000689e-05, "loss": 0.6297, "step": 436250 }, { "epoch": 75.06882312456986, "grad_norm": 18.549575805664062, "learning_rate": 1.2465588437715072e-05, "loss": 0.6614, "step": 436300 }, { "epoch": 75.07742601514109, "grad_norm": 17.58775520324707, "learning_rate": 1.2461286992429457e-05, "loss": 0.6715, "step": 436350 }, { "epoch": 75.08602890571233, "grad_norm": 19.393638610839844, "learning_rate": 1.2456985547143841e-05, "loss": 0.667, "step": 436400 }, { "epoch": 75.09463179628355, "grad_norm": 12.97643756866455, "learning_rate": 1.2452684101858226e-05, "loss": 0.6772, "step": 436450 }, { "epoch": 75.10323468685478, "grad_norm": 23.3187255859375, "learning_rate": 1.2448382656572609e-05, "loss": 0.6566, "step": 436500 }, { "epoch": 75.11183757742602, "grad_norm": 19.764150619506836, "learning_rate": 1.2444081211286993e-05, "loss": 0.6969, "step": 436550 }, { "epoch": 75.12044046799724, "grad_norm": 11.19465160369873, "learning_rate": 1.2439779766001378e-05, "loss": 0.6873, "step": 436600 }, { "epoch": 75.12904335856848, "grad_norm": 18.95055389404297, "learning_rate": 1.243547832071576e-05, "loss": 0.663, "step": 436650 }, { "epoch": 75.1376462491397, "grad_norm": 9.24363899230957, "learning_rate": 1.2431176875430145e-05, "loss": 0.6619, "step": 436700 }, { "epoch": 75.14624913971095, "grad_norm": 6.982728958129883, "learning_rate": 1.2426875430144528e-05, "loss": 0.6601, "step": 436750 }, { "epoch": 75.15485203028217, "grad_norm": 11.508416175842285, "learning_rate": 1.2422573984858914e-05, "loss": 0.7051, "step": 436800 }, { "epoch": 75.16345492085341, "grad_norm": 41.13494873046875, "learning_rate": 1.2418272539573297e-05, "loss": 0.7042, "step": 436850 }, { "epoch": 75.17205781142464, "grad_norm": 15.385159492492676, "learning_rate": 1.2413971094287682e-05, "loss": 0.6679, "step": 436900 }, { "epoch": 75.18066070199588, "grad_norm": 17.715042114257812, "learning_rate": 1.2409669649002065e-05, "loss": 0.6809, "step": 436950 }, { "epoch": 75.1892635925671, "grad_norm": 8.281517028808594, "learning_rate": 1.2405368203716449e-05, "loss": 0.6853, "step": 437000 }, { "epoch": 75.19786648313834, "grad_norm": 16.499698638916016, "learning_rate": 1.2401066758430834e-05, "loss": 0.691, "step": 437050 }, { "epoch": 75.20646937370957, "grad_norm": 13.181706428527832, "learning_rate": 1.2396765313145216e-05, "loss": 0.6795, "step": 437100 }, { "epoch": 75.21507226428079, "grad_norm": 15.745058059692383, "learning_rate": 1.2392463867859601e-05, "loss": 0.6465, "step": 437150 }, { "epoch": 75.22367515485203, "grad_norm": 18.942508697509766, "learning_rate": 1.2388162422573986e-05, "loss": 0.6775, "step": 437200 }, { "epoch": 75.23227804542326, "grad_norm": 20.9103946685791, "learning_rate": 1.238386097728837e-05, "loss": 0.6585, "step": 437250 }, { "epoch": 75.2408809359945, "grad_norm": 12.784262657165527, "learning_rate": 1.2379559532002753e-05, "loss": 0.6511, "step": 437300 }, { "epoch": 75.24948382656572, "grad_norm": 10.091845512390137, "learning_rate": 1.2375258086717138e-05, "loss": 0.663, "step": 437350 }, { "epoch": 75.25808671713696, "grad_norm": 11.17796516418457, "learning_rate": 1.2370956641431522e-05, "loss": 0.6551, "step": 437400 }, { "epoch": 75.26668960770819, "grad_norm": 28.255754470825195, "learning_rate": 1.2366655196145905e-05, "loss": 0.73, "step": 437450 }, { "epoch": 75.27529249827943, "grad_norm": 14.028192520141602, "learning_rate": 1.236235375086029e-05, "loss": 0.6647, "step": 437500 }, { "epoch": 75.28389538885065, "grad_norm": 21.49006462097168, "learning_rate": 1.2358052305574672e-05, "loss": 0.7384, "step": 437550 }, { "epoch": 75.29249827942189, "grad_norm": 8.776004791259766, "learning_rate": 1.2353750860289059e-05, "loss": 0.6619, "step": 437600 }, { "epoch": 75.30110116999312, "grad_norm": 9.002683639526367, "learning_rate": 1.2349449415003442e-05, "loss": 0.7124, "step": 437650 }, { "epoch": 75.30970406056434, "grad_norm": 16.86612319946289, "learning_rate": 1.2345147969717826e-05, "loss": 0.6863, "step": 437700 }, { "epoch": 75.31830695113558, "grad_norm": 17.302562713623047, "learning_rate": 1.2340846524432209e-05, "loss": 0.6639, "step": 437750 }, { "epoch": 75.32690984170681, "grad_norm": 6.432380676269531, "learning_rate": 1.2336545079146594e-05, "loss": 0.6498, "step": 437800 }, { "epoch": 75.33551273227805, "grad_norm": 12.33041000366211, "learning_rate": 1.2332243633860978e-05, "loss": 0.7328, "step": 437850 }, { "epoch": 75.34411562284927, "grad_norm": 16.408557891845703, "learning_rate": 1.2327942188575361e-05, "loss": 0.6741, "step": 437900 }, { "epoch": 75.35271851342051, "grad_norm": 13.85654067993164, "learning_rate": 1.2323640743289747e-05, "loss": 0.7141, "step": 437950 }, { "epoch": 75.36132140399174, "grad_norm": 17.906801223754883, "learning_rate": 1.231933929800413e-05, "loss": 0.7476, "step": 438000 }, { "epoch": 75.36992429456298, "grad_norm": 12.68820571899414, "learning_rate": 1.2315037852718515e-05, "loss": 0.6203, "step": 438050 }, { "epoch": 75.3785271851342, "grad_norm": 12.96099853515625, "learning_rate": 1.2310736407432898e-05, "loss": 0.6345, "step": 438100 }, { "epoch": 75.38713007570544, "grad_norm": 10.723170280456543, "learning_rate": 1.2306434962147282e-05, "loss": 0.6425, "step": 438150 }, { "epoch": 75.39573296627667, "grad_norm": 10.161581039428711, "learning_rate": 1.2302133516861667e-05, "loss": 0.6899, "step": 438200 }, { "epoch": 75.40433585684791, "grad_norm": 22.426467895507812, "learning_rate": 1.229783207157605e-05, "loss": 0.7479, "step": 438250 }, { "epoch": 75.41293874741913, "grad_norm": 13.067246437072754, "learning_rate": 1.2293530626290434e-05, "loss": 0.6695, "step": 438300 }, { "epoch": 75.42154163799036, "grad_norm": 17.21710205078125, "learning_rate": 1.2289229181004819e-05, "loss": 0.7182, "step": 438350 }, { "epoch": 75.4301445285616, "grad_norm": 11.734602928161621, "learning_rate": 1.2284927735719203e-05, "loss": 0.7004, "step": 438400 }, { "epoch": 75.43874741913282, "grad_norm": 15.584959030151367, "learning_rate": 1.2280626290433586e-05, "loss": 0.6575, "step": 438450 }, { "epoch": 75.44735030970406, "grad_norm": 10.889196395874023, "learning_rate": 1.227632484514797e-05, "loss": 0.6883, "step": 438500 }, { "epoch": 75.45595320027529, "grad_norm": 18.655715942382812, "learning_rate": 1.2272023399862355e-05, "loss": 0.6491, "step": 438550 }, { "epoch": 75.46455609084653, "grad_norm": 27.752037048339844, "learning_rate": 1.2267721954576738e-05, "loss": 0.7676, "step": 438600 }, { "epoch": 75.47315898141775, "grad_norm": 16.669300079345703, "learning_rate": 1.2263420509291123e-05, "loss": 0.7248, "step": 438650 }, { "epoch": 75.481761871989, "grad_norm": 21.530513763427734, "learning_rate": 1.2259119064005505e-05, "loss": 0.737, "step": 438700 }, { "epoch": 75.49036476256022, "grad_norm": 16.505613327026367, "learning_rate": 1.2254817618719892e-05, "loss": 0.6532, "step": 438750 }, { "epoch": 75.49896765313146, "grad_norm": 12.13188648223877, "learning_rate": 1.2250516173434275e-05, "loss": 0.7225, "step": 438800 }, { "epoch": 75.50757054370268, "grad_norm": 14.47838020324707, "learning_rate": 1.224621472814866e-05, "loss": 0.6543, "step": 438850 }, { "epoch": 75.51617343427391, "grad_norm": 15.86959171295166, "learning_rate": 1.2241913282863042e-05, "loss": 0.6776, "step": 438900 }, { "epoch": 75.52477632484515, "grad_norm": 12.748726844787598, "learning_rate": 1.2237611837577427e-05, "loss": 0.6501, "step": 438950 }, { "epoch": 75.53337921541637, "grad_norm": 14.773652076721191, "learning_rate": 1.2233310392291811e-05, "loss": 0.7116, "step": 439000 }, { "epoch": 75.54198210598761, "grad_norm": 17.840593338012695, "learning_rate": 1.2229008947006194e-05, "loss": 0.7209, "step": 439050 }, { "epoch": 75.55058499655884, "grad_norm": 7.495781898498535, "learning_rate": 1.2224707501720579e-05, "loss": 0.6761, "step": 439100 }, { "epoch": 75.55918788713008, "grad_norm": 5.960693836212158, "learning_rate": 1.2220406056434963e-05, "loss": 0.7009, "step": 439150 }, { "epoch": 75.5677907777013, "grad_norm": 9.127427101135254, "learning_rate": 1.2216104611149346e-05, "loss": 0.6692, "step": 439200 }, { "epoch": 75.57639366827254, "grad_norm": 13.617364883422852, "learning_rate": 1.221180316586373e-05, "loss": 0.7354, "step": 439250 }, { "epoch": 75.58499655884377, "grad_norm": 13.44190502166748, "learning_rate": 1.2207501720578113e-05, "loss": 0.6859, "step": 439300 }, { "epoch": 75.59359944941501, "grad_norm": 17.485925674438477, "learning_rate": 1.22032002752925e-05, "loss": 0.7134, "step": 439350 }, { "epoch": 75.60220233998623, "grad_norm": 13.262593269348145, "learning_rate": 1.2198898830006883e-05, "loss": 0.6872, "step": 439400 }, { "epoch": 75.61080523055746, "grad_norm": 14.954830169677734, "learning_rate": 1.2194597384721267e-05, "loss": 0.693, "step": 439450 }, { "epoch": 75.6194081211287, "grad_norm": 10.574771881103516, "learning_rate": 1.219029593943565e-05, "loss": 0.7184, "step": 439500 }, { "epoch": 75.62801101169993, "grad_norm": 23.856548309326172, "learning_rate": 1.2185994494150035e-05, "loss": 0.692, "step": 439550 }, { "epoch": 75.63661390227116, "grad_norm": 15.153326988220215, "learning_rate": 1.2181693048864419e-05, "loss": 0.6546, "step": 439600 }, { "epoch": 75.64521679284239, "grad_norm": 9.581445693969727, "learning_rate": 1.2177391603578802e-05, "loss": 0.6869, "step": 439650 }, { "epoch": 75.65381968341363, "grad_norm": 13.227323532104492, "learning_rate": 1.2173090158293188e-05, "loss": 0.6837, "step": 439700 }, { "epoch": 75.66242257398486, "grad_norm": 11.059216499328613, "learning_rate": 1.2168788713007571e-05, "loss": 0.6943, "step": 439750 }, { "epoch": 75.6710254645561, "grad_norm": 18.35938835144043, "learning_rate": 1.2164487267721956e-05, "loss": 0.7324, "step": 439800 }, { "epoch": 75.67962835512732, "grad_norm": 9.472108840942383, "learning_rate": 1.2160185822436339e-05, "loss": 0.6912, "step": 439850 }, { "epoch": 75.68823124569856, "grad_norm": 20.934112548828125, "learning_rate": 1.2155884377150723e-05, "loss": 0.6621, "step": 439900 }, { "epoch": 75.69683413626979, "grad_norm": 16.52595329284668, "learning_rate": 1.2151582931865108e-05, "loss": 0.6872, "step": 439950 }, { "epoch": 75.70543702684103, "grad_norm": 14.424139976501465, "learning_rate": 1.214728148657949e-05, "loss": 0.716, "step": 440000 }, { "epoch": 75.71403991741225, "grad_norm": 18.58774185180664, "learning_rate": 1.2142980041293875e-05, "loss": 0.7039, "step": 440050 }, { "epoch": 75.72264280798348, "grad_norm": 14.925125122070312, "learning_rate": 1.213867859600826e-05, "loss": 0.6975, "step": 440100 }, { "epoch": 75.73124569855472, "grad_norm": 27.44615364074707, "learning_rate": 1.2134377150722644e-05, "loss": 0.7102, "step": 440150 }, { "epoch": 75.73984858912594, "grad_norm": 22.837121963500977, "learning_rate": 1.2130075705437027e-05, "loss": 0.725, "step": 440200 }, { "epoch": 75.74845147969718, "grad_norm": 16.258459091186523, "learning_rate": 1.2125774260151412e-05, "loss": 0.6617, "step": 440250 }, { "epoch": 75.7570543702684, "grad_norm": 20.08734703063965, "learning_rate": 1.2121472814865796e-05, "loss": 0.6781, "step": 440300 }, { "epoch": 75.76565726083965, "grad_norm": 11.697669982910156, "learning_rate": 1.2117171369580179e-05, "loss": 0.6829, "step": 440350 }, { "epoch": 75.77426015141087, "grad_norm": 15.288185119628906, "learning_rate": 1.2112869924294564e-05, "loss": 0.7008, "step": 440400 }, { "epoch": 75.78286304198211, "grad_norm": 8.167585372924805, "learning_rate": 1.2108568479008946e-05, "loss": 0.667, "step": 440450 }, { "epoch": 75.79146593255334, "grad_norm": 17.80044174194336, "learning_rate": 1.2104267033723333e-05, "loss": 0.6746, "step": 440500 }, { "epoch": 75.80006882312458, "grad_norm": 19.964380264282227, "learning_rate": 1.2099965588437716e-05, "loss": 0.7501, "step": 440550 }, { "epoch": 75.8086717136958, "grad_norm": 12.36269474029541, "learning_rate": 1.20956641431521e-05, "loss": 0.7152, "step": 440600 }, { "epoch": 75.81727460426703, "grad_norm": 18.412967681884766, "learning_rate": 1.2091362697866483e-05, "loss": 0.6918, "step": 440650 }, { "epoch": 75.82587749483827, "grad_norm": 16.5703182220459, "learning_rate": 1.2087061252580868e-05, "loss": 0.6569, "step": 440700 }, { "epoch": 75.83448038540949, "grad_norm": 12.972833633422852, "learning_rate": 1.2082759807295252e-05, "loss": 0.6437, "step": 440750 }, { "epoch": 75.84308327598073, "grad_norm": 14.72881031036377, "learning_rate": 1.2078458362009635e-05, "loss": 0.6637, "step": 440800 }, { "epoch": 75.85168616655196, "grad_norm": 13.5518798828125, "learning_rate": 1.207415691672402e-05, "loss": 0.6595, "step": 440850 }, { "epoch": 75.8602890571232, "grad_norm": 12.977130889892578, "learning_rate": 1.2069855471438404e-05, "loss": 0.6833, "step": 440900 }, { "epoch": 75.86889194769442, "grad_norm": 20.958698272705078, "learning_rate": 1.2065554026152789e-05, "loss": 0.6877, "step": 440950 }, { "epoch": 75.87749483826566, "grad_norm": 13.850357055664062, "learning_rate": 1.2061252580867172e-05, "loss": 0.7089, "step": 441000 }, { "epoch": 75.88609772883689, "grad_norm": 11.858187675476074, "learning_rate": 1.2056951135581556e-05, "loss": 0.6844, "step": 441050 }, { "epoch": 75.89470061940813, "grad_norm": 22.10969352722168, "learning_rate": 1.205264969029594e-05, "loss": 0.6598, "step": 441100 }, { "epoch": 75.90330350997935, "grad_norm": 15.952227592468262, "learning_rate": 1.2048348245010324e-05, "loss": 0.695, "step": 441150 }, { "epoch": 75.91190640055059, "grad_norm": 10.439767837524414, "learning_rate": 1.2044046799724708e-05, "loss": 0.6989, "step": 441200 }, { "epoch": 75.92050929112182, "grad_norm": 18.379091262817383, "learning_rate": 1.2039745354439091e-05, "loss": 0.6659, "step": 441250 }, { "epoch": 75.92911218169304, "grad_norm": 12.877928733825684, "learning_rate": 1.2035443909153477e-05, "loss": 0.6732, "step": 441300 }, { "epoch": 75.93771507226428, "grad_norm": 20.11397933959961, "learning_rate": 1.203114246386786e-05, "loss": 0.7037, "step": 441350 }, { "epoch": 75.94631796283551, "grad_norm": 14.19643783569336, "learning_rate": 1.2026841018582245e-05, "loss": 0.6834, "step": 441400 }, { "epoch": 75.95492085340675, "grad_norm": 16.35695457458496, "learning_rate": 1.2022539573296628e-05, "loss": 0.6926, "step": 441450 }, { "epoch": 75.96352374397797, "grad_norm": 15.2689208984375, "learning_rate": 1.2018238128011012e-05, "loss": 0.7119, "step": 441500 }, { "epoch": 75.97212663454921, "grad_norm": 13.375205039978027, "learning_rate": 1.2013936682725397e-05, "loss": 0.7418, "step": 441550 }, { "epoch": 75.98072952512044, "grad_norm": 11.819114685058594, "learning_rate": 1.200963523743978e-05, "loss": 0.7039, "step": 441600 }, { "epoch": 75.98933241569168, "grad_norm": 18.493637084960938, "learning_rate": 1.2005333792154166e-05, "loss": 0.7109, "step": 441650 }, { "epoch": 75.9979353062629, "grad_norm": 14.59454345703125, "learning_rate": 1.2001032346868549e-05, "loss": 0.76, "step": 441700 }, { "epoch": 76.0, "eval_accuracy": 0.599208397866116, "eval_f1": 0.5949317604026003, "eval_f1_DuraRiadoRio_16x16": 0.5415701415701416, "eval_f1_Mole_16x16": 0.5299884659746251, "eval_f1_Quebrado_16x16": 0.7495437031028189, "eval_f1_RiadoRio_16x16": 0.5441127138544113, "eval_f1_RioFechado_16x16": 0.6094437775110044, "eval_loss": 1.766323447227478, "eval_precision": 0.6355602255708365, "eval_precision_DuraRiadoRio_16x16": 0.6654016445287793, "eval_precision_Mole_16x16": 0.7895189003436426, "eval_precision_Quebrado_16x16": 0.7034640274076894, "eval_precision_RiadoRio_16x16": 0.4594900849858357, "eval_precision_RioFechado_16x16": 0.5599264705882353, "eval_recall": 0.5986123585824746, "eval_recall_DuraRiadoRio_16x16": 0.4565972222222222, "eval_recall_Mole_16x16": 0.3988715277777778, "eval_recall_Quebrado_16x16": 0.8020833333333334, "eval_recall_RiadoRio_16x16": 0.6669407894736842, "eval_recall_RioFechado_16x16": 0.6685689201053556, "eval_runtime": 46.2785, "eval_samples_per_second": 251.132, "eval_steps_per_second": 15.709, "step": 441712 }, { "epoch": 76.00653819683414, "grad_norm": 10.847066879272461, "learning_rate": 1.1996730901582933e-05, "loss": 0.6884, "step": 441750 }, { "epoch": 76.01514108740537, "grad_norm": 23.074356079101562, "learning_rate": 1.1992429456297316e-05, "loss": 0.6855, "step": 441800 }, { "epoch": 76.0237439779766, "grad_norm": 20.213623046875, "learning_rate": 1.19881280110117e-05, "loss": 0.6533, "step": 441850 }, { "epoch": 76.03234686854783, "grad_norm": 19.081737518310547, "learning_rate": 1.1983826565726085e-05, "loss": 0.681, "step": 441900 }, { "epoch": 76.04094975911906, "grad_norm": 19.86202621459961, "learning_rate": 1.1979525120440468e-05, "loss": 0.6279, "step": 441950 }, { "epoch": 76.0495526496903, "grad_norm": 28.924358367919922, "learning_rate": 1.1975223675154853e-05, "loss": 0.6703, "step": 442000 }, { "epoch": 76.05815554026152, "grad_norm": 22.642297744750977, "learning_rate": 1.1970922229869237e-05, "loss": 0.663, "step": 442050 }, { "epoch": 76.06675843083276, "grad_norm": 23.66024398803711, "learning_rate": 1.1966620784583622e-05, "loss": 0.6963, "step": 442100 }, { "epoch": 76.07536132140399, "grad_norm": 14.237215042114258, "learning_rate": 1.1962319339298005e-05, "loss": 0.6696, "step": 442150 }, { "epoch": 76.08396421197523, "grad_norm": 18.26310157775879, "learning_rate": 1.1958017894012387e-05, "loss": 0.6497, "step": 442200 }, { "epoch": 76.09256710254645, "grad_norm": 18.04029655456543, "learning_rate": 1.1953716448726774e-05, "loss": 0.677, "step": 442250 }, { "epoch": 76.1011699931177, "grad_norm": 12.44274616241455, "learning_rate": 1.1949415003441157e-05, "loss": 0.7177, "step": 442300 }, { "epoch": 76.10977288368892, "grad_norm": 7.523771286010742, "learning_rate": 1.1945113558155541e-05, "loss": 0.6732, "step": 442350 }, { "epoch": 76.11837577426014, "grad_norm": 15.441924095153809, "learning_rate": 1.1940812112869924e-05, "loss": 0.7289, "step": 442400 }, { "epoch": 76.12697866483138, "grad_norm": 13.055009841918945, "learning_rate": 1.193651066758431e-05, "loss": 0.6402, "step": 442450 }, { "epoch": 76.13558155540261, "grad_norm": 11.975520133972168, "learning_rate": 1.1932209222298693e-05, "loss": 0.638, "step": 442500 }, { "epoch": 76.14418444597385, "grad_norm": 17.040115356445312, "learning_rate": 1.1927907777013076e-05, "loss": 0.6976, "step": 442550 }, { "epoch": 76.15278733654507, "grad_norm": 10.298079490661621, "learning_rate": 1.192360633172746e-05, "loss": 0.6939, "step": 442600 }, { "epoch": 76.16139022711631, "grad_norm": 9.925219535827637, "learning_rate": 1.1919304886441845e-05, "loss": 0.6485, "step": 442650 }, { "epoch": 76.16999311768754, "grad_norm": 12.295477867126465, "learning_rate": 1.191500344115623e-05, "loss": 0.6709, "step": 442700 }, { "epoch": 76.17859600825878, "grad_norm": 9.759729385375977, "learning_rate": 1.1910701995870613e-05, "loss": 0.6992, "step": 442750 }, { "epoch": 76.18719889883, "grad_norm": 12.753090858459473, "learning_rate": 1.1906400550584997e-05, "loss": 0.6557, "step": 442800 }, { "epoch": 76.19580178940124, "grad_norm": 13.654587745666504, "learning_rate": 1.1902099105299382e-05, "loss": 0.6722, "step": 442850 }, { "epoch": 76.20440467997247, "grad_norm": 15.990962982177734, "learning_rate": 1.1897797660013765e-05, "loss": 0.7083, "step": 442900 }, { "epoch": 76.21300757054371, "grad_norm": 18.432601928710938, "learning_rate": 1.1893496214728149e-05, "loss": 0.6904, "step": 442950 }, { "epoch": 76.22161046111493, "grad_norm": 25.38471031188965, "learning_rate": 1.1889194769442532e-05, "loss": 0.6658, "step": 443000 }, { "epoch": 76.23021335168616, "grad_norm": 19.529640197753906, "learning_rate": 1.1884893324156918e-05, "loss": 0.6474, "step": 443050 }, { "epoch": 76.2388162422574, "grad_norm": 22.967323303222656, "learning_rate": 1.1880591878871301e-05, "loss": 0.6481, "step": 443100 }, { "epoch": 76.24741913282863, "grad_norm": 15.3319730758667, "learning_rate": 1.1876290433585686e-05, "loss": 0.6705, "step": 443150 }, { "epoch": 76.25602202339986, "grad_norm": 16.04021453857422, "learning_rate": 1.1871988988300068e-05, "loss": 0.7278, "step": 443200 }, { "epoch": 76.26462491397109, "grad_norm": 15.887761116027832, "learning_rate": 1.1867687543014453e-05, "loss": 0.6842, "step": 443250 }, { "epoch": 76.27322780454233, "grad_norm": 7.751402854919434, "learning_rate": 1.1863386097728838e-05, "loss": 0.6816, "step": 443300 }, { "epoch": 76.28183069511356, "grad_norm": 15.69094467163086, "learning_rate": 1.185908465244322e-05, "loss": 0.6857, "step": 443350 }, { "epoch": 76.2904335856848, "grad_norm": 20.151844024658203, "learning_rate": 1.1854783207157607e-05, "loss": 0.6592, "step": 443400 }, { "epoch": 76.29903647625602, "grad_norm": 9.942814826965332, "learning_rate": 1.185048176187199e-05, "loss": 0.6469, "step": 443450 }, { "epoch": 76.30763936682726, "grad_norm": 23.15900230407715, "learning_rate": 1.1846180316586374e-05, "loss": 0.6914, "step": 443500 }, { "epoch": 76.31624225739849, "grad_norm": 17.42514991760254, "learning_rate": 1.1841878871300757e-05, "loss": 0.7272, "step": 443550 }, { "epoch": 76.32484514796971, "grad_norm": 14.397653579711914, "learning_rate": 1.1837577426015142e-05, "loss": 0.6528, "step": 443600 }, { "epoch": 76.33344803854095, "grad_norm": 20.94039535522461, "learning_rate": 1.1833275980729526e-05, "loss": 0.6864, "step": 443650 }, { "epoch": 76.34205092911218, "grad_norm": 15.656232833862305, "learning_rate": 1.1828974535443909e-05, "loss": 0.6389, "step": 443700 }, { "epoch": 76.35065381968342, "grad_norm": 21.23846435546875, "learning_rate": 1.1824673090158294e-05, "loss": 0.6741, "step": 443750 }, { "epoch": 76.35925671025464, "grad_norm": 10.625885963439941, "learning_rate": 1.1820371644872678e-05, "loss": 0.6617, "step": 443800 }, { "epoch": 76.36785960082588, "grad_norm": 12.742866516113281, "learning_rate": 1.1816070199587063e-05, "loss": 0.7426, "step": 443850 }, { "epoch": 76.3764624913971, "grad_norm": 21.410152435302734, "learning_rate": 1.1811768754301446e-05, "loss": 0.641, "step": 443900 }, { "epoch": 76.38506538196835, "grad_norm": 17.74333381652832, "learning_rate": 1.180746730901583e-05, "loss": 0.6169, "step": 443950 }, { "epoch": 76.39366827253957, "grad_norm": 13.082315444946289, "learning_rate": 1.1803165863730215e-05, "loss": 0.7277, "step": 444000 }, { "epoch": 76.40227116311081, "grad_norm": 16.99566078186035, "learning_rate": 1.1798864418444598e-05, "loss": 0.6932, "step": 444050 }, { "epoch": 76.41087405368204, "grad_norm": 24.87788200378418, "learning_rate": 1.1794562973158982e-05, "loss": 0.6584, "step": 444100 }, { "epoch": 76.41947694425328, "grad_norm": 13.313573837280273, "learning_rate": 1.1790261527873365e-05, "loss": 0.6534, "step": 444150 }, { "epoch": 76.4280798348245, "grad_norm": 18.421295166015625, "learning_rate": 1.1785960082587751e-05, "loss": 0.7695, "step": 444200 }, { "epoch": 76.43668272539573, "grad_norm": 10.888594627380371, "learning_rate": 1.1781658637302134e-05, "loss": 0.6927, "step": 444250 }, { "epoch": 76.44528561596697, "grad_norm": 17.045486450195312, "learning_rate": 1.1777357192016519e-05, "loss": 0.6875, "step": 444300 }, { "epoch": 76.45388850653819, "grad_norm": 14.783318519592285, "learning_rate": 1.1773055746730902e-05, "loss": 0.6732, "step": 444350 }, { "epoch": 76.46249139710943, "grad_norm": 9.589494705200195, "learning_rate": 1.1768754301445286e-05, "loss": 0.697, "step": 444400 }, { "epoch": 76.47109428768066, "grad_norm": 19.040151596069336, "learning_rate": 1.176445285615967e-05, "loss": 0.6702, "step": 444450 }, { "epoch": 76.4796971782519, "grad_norm": 18.62373924255371, "learning_rate": 1.1760151410874054e-05, "loss": 0.6665, "step": 444500 }, { "epoch": 76.48830006882312, "grad_norm": 18.642019271850586, "learning_rate": 1.1755849965588438e-05, "loss": 0.6941, "step": 444550 }, { "epoch": 76.49690295939436, "grad_norm": 14.707046508789062, "learning_rate": 1.1751548520302823e-05, "loss": 0.6424, "step": 444600 }, { "epoch": 76.50550584996559, "grad_norm": 16.577592849731445, "learning_rate": 1.1747247075017207e-05, "loss": 0.6945, "step": 444650 }, { "epoch": 76.51410874053683, "grad_norm": 16.26395034790039, "learning_rate": 1.174294562973159e-05, "loss": 0.7037, "step": 444700 }, { "epoch": 76.52271163110805, "grad_norm": 14.88575553894043, "learning_rate": 1.1738644184445975e-05, "loss": 0.6876, "step": 444750 }, { "epoch": 76.53131452167928, "grad_norm": 13.130255699157715, "learning_rate": 1.173434273916036e-05, "loss": 0.6963, "step": 444800 }, { "epoch": 76.53991741225052, "grad_norm": 18.232257843017578, "learning_rate": 1.1730041293874742e-05, "loss": 0.6351, "step": 444850 }, { "epoch": 76.54852030282174, "grad_norm": 15.19131088256836, "learning_rate": 1.1725739848589127e-05, "loss": 0.6778, "step": 444900 }, { "epoch": 76.55712319339298, "grad_norm": 17.366304397583008, "learning_rate": 1.172143840330351e-05, "loss": 0.7326, "step": 444950 }, { "epoch": 76.56572608396421, "grad_norm": 10.5128812789917, "learning_rate": 1.1717136958017896e-05, "loss": 0.7366, "step": 445000 }, { "epoch": 76.57432897453545, "grad_norm": 16.286094665527344, "learning_rate": 1.1712835512732279e-05, "loss": 0.6504, "step": 445050 }, { "epoch": 76.58293186510667, "grad_norm": 9.845297813415527, "learning_rate": 1.1708534067446663e-05, "loss": 0.6653, "step": 445100 }, { "epoch": 76.59153475567791, "grad_norm": 12.674320220947266, "learning_rate": 1.1704232622161046e-05, "loss": 0.6532, "step": 445150 }, { "epoch": 76.60013764624914, "grad_norm": 24.589031219482422, "learning_rate": 1.169993117687543e-05, "loss": 0.6588, "step": 445200 }, { "epoch": 76.60874053682038, "grad_norm": 11.931520462036133, "learning_rate": 1.1695629731589815e-05, "loss": 0.6615, "step": 445250 }, { "epoch": 76.6173434273916, "grad_norm": 14.024046897888184, "learning_rate": 1.1691328286304198e-05, "loss": 0.7083, "step": 445300 }, { "epoch": 76.62594631796284, "grad_norm": 13.101245880126953, "learning_rate": 1.1687026841018584e-05, "loss": 0.6573, "step": 445350 }, { "epoch": 76.63454920853407, "grad_norm": 18.827104568481445, "learning_rate": 1.1682725395732967e-05, "loss": 0.6902, "step": 445400 }, { "epoch": 76.6431520991053, "grad_norm": 16.756555557250977, "learning_rate": 1.167842395044735e-05, "loss": 0.6888, "step": 445450 }, { "epoch": 76.65175498967653, "grad_norm": 16.264232635498047, "learning_rate": 1.1674122505161735e-05, "loss": 0.6746, "step": 445500 }, { "epoch": 76.66035788024776, "grad_norm": 20.453649520874023, "learning_rate": 1.1669821059876119e-05, "loss": 0.6517, "step": 445550 }, { "epoch": 76.668960770819, "grad_norm": 12.098198890686035, "learning_rate": 1.1665519614590504e-05, "loss": 0.7709, "step": 445600 }, { "epoch": 76.67756366139022, "grad_norm": 12.386787414550781, "learning_rate": 1.1661218169304887e-05, "loss": 0.675, "step": 445650 }, { "epoch": 76.68616655196146, "grad_norm": 18.012588500976562, "learning_rate": 1.1656916724019271e-05, "loss": 0.6625, "step": 445700 }, { "epoch": 76.69476944253269, "grad_norm": 13.681233406066895, "learning_rate": 1.1652615278733656e-05, "loss": 0.67, "step": 445750 }, { "epoch": 76.70337233310393, "grad_norm": 12.92270278930664, "learning_rate": 1.1648313833448039e-05, "loss": 0.6585, "step": 445800 }, { "epoch": 76.71197522367515, "grad_norm": 26.964401245117188, "learning_rate": 1.1644012388162423e-05, "loss": 0.7326, "step": 445850 }, { "epoch": 76.7205781142464, "grad_norm": 21.043243408203125, "learning_rate": 1.1639710942876806e-05, "loss": 0.6765, "step": 445900 }, { "epoch": 76.72918100481762, "grad_norm": 25.092830657958984, "learning_rate": 1.1635409497591192e-05, "loss": 0.6567, "step": 445950 }, { "epoch": 76.73778389538884, "grad_norm": 11.81569766998291, "learning_rate": 1.1631108052305575e-05, "loss": 0.6789, "step": 446000 }, { "epoch": 76.74638678596008, "grad_norm": 19.848787307739258, "learning_rate": 1.162680660701996e-05, "loss": 0.6976, "step": 446050 }, { "epoch": 76.75498967653131, "grad_norm": 15.281746864318848, "learning_rate": 1.1622505161734343e-05, "loss": 0.6387, "step": 446100 }, { "epoch": 76.76359256710255, "grad_norm": 23.290328979492188, "learning_rate": 1.1618203716448727e-05, "loss": 0.7062, "step": 446150 }, { "epoch": 76.77219545767377, "grad_norm": 7.730297565460205, "learning_rate": 1.1613902271163112e-05, "loss": 0.6839, "step": 446200 }, { "epoch": 76.78079834824501, "grad_norm": 14.185111999511719, "learning_rate": 1.1609600825877494e-05, "loss": 0.6643, "step": 446250 }, { "epoch": 76.78940123881624, "grad_norm": 12.170747756958008, "learning_rate": 1.1605299380591879e-05, "loss": 0.7226, "step": 446300 }, { "epoch": 76.79800412938748, "grad_norm": 14.06230354309082, "learning_rate": 1.1600997935306264e-05, "loss": 0.7116, "step": 446350 }, { "epoch": 76.8066070199587, "grad_norm": 11.018351554870605, "learning_rate": 1.1596696490020648e-05, "loss": 0.704, "step": 446400 }, { "epoch": 76.81520991052994, "grad_norm": 15.967963218688965, "learning_rate": 1.1592395044735031e-05, "loss": 0.6956, "step": 446450 }, { "epoch": 76.82381280110117, "grad_norm": 18.01267433166504, "learning_rate": 1.1588093599449416e-05, "loss": 0.7134, "step": 446500 }, { "epoch": 76.8324156916724, "grad_norm": 15.873642921447754, "learning_rate": 1.15837921541638e-05, "loss": 0.6607, "step": 446550 }, { "epoch": 76.84101858224363, "grad_norm": 21.713748931884766, "learning_rate": 1.1579490708878183e-05, "loss": 0.7598, "step": 446600 }, { "epoch": 76.84962147281486, "grad_norm": 15.821956634521484, "learning_rate": 1.1575189263592568e-05, "loss": 0.6772, "step": 446650 }, { "epoch": 76.8582243633861, "grad_norm": 12.019572257995605, "learning_rate": 1.157088781830695e-05, "loss": 0.6821, "step": 446700 }, { "epoch": 76.86682725395733, "grad_norm": 15.759546279907227, "learning_rate": 1.1566586373021337e-05, "loss": 0.6628, "step": 446750 }, { "epoch": 76.87543014452856, "grad_norm": 24.959421157836914, "learning_rate": 1.156228492773572e-05, "loss": 0.7201, "step": 446800 }, { "epoch": 76.88403303509979, "grad_norm": 16.50052833557129, "learning_rate": 1.1557983482450104e-05, "loss": 0.6947, "step": 446850 }, { "epoch": 76.89263592567103, "grad_norm": 12.603623390197754, "learning_rate": 1.1553682037164487e-05, "loss": 0.6423, "step": 446900 }, { "epoch": 76.90123881624226, "grad_norm": 13.83243179321289, "learning_rate": 1.1549380591878872e-05, "loss": 0.651, "step": 446950 }, { "epoch": 76.9098417068135, "grad_norm": 14.609078407287598, "learning_rate": 1.1545079146593256e-05, "loss": 0.7025, "step": 447000 }, { "epoch": 76.91844459738472, "grad_norm": 21.857925415039062, "learning_rate": 1.1540777701307639e-05, "loss": 0.7536, "step": 447050 }, { "epoch": 76.92704748795596, "grad_norm": 11.631343841552734, "learning_rate": 1.1536476256022025e-05, "loss": 0.7035, "step": 447100 }, { "epoch": 76.93565037852719, "grad_norm": 12.328271865844727, "learning_rate": 1.1532174810736408e-05, "loss": 0.6874, "step": 447150 }, { "epoch": 76.94425326909841, "grad_norm": 11.836929321289062, "learning_rate": 1.1527873365450793e-05, "loss": 0.7015, "step": 447200 }, { "epoch": 76.95285615966965, "grad_norm": 13.737574577331543, "learning_rate": 1.1523571920165176e-05, "loss": 0.688, "step": 447250 }, { "epoch": 76.96145905024088, "grad_norm": 14.74349308013916, "learning_rate": 1.151927047487956e-05, "loss": 0.6422, "step": 447300 }, { "epoch": 76.97006194081212, "grad_norm": 17.354618072509766, "learning_rate": 1.1514969029593945e-05, "loss": 0.6466, "step": 447350 }, { "epoch": 76.97866483138334, "grad_norm": 18.48841094970703, "learning_rate": 1.1510667584308328e-05, "loss": 0.6736, "step": 447400 }, { "epoch": 76.98726772195458, "grad_norm": 17.104440689086914, "learning_rate": 1.1506366139022712e-05, "loss": 0.6821, "step": 447450 }, { "epoch": 76.9958706125258, "grad_norm": 13.92197036743164, "learning_rate": 1.1502064693737097e-05, "loss": 0.7148, "step": 447500 }, { "epoch": 77.0, "eval_accuracy": 0.6083290311478231, "eval_f1": 0.6120487677029489, "eval_f1_DuraRiadoRio_16x16": 0.5623944001930968, "eval_f1_Mole_16x16": 0.6448736998514116, "eval_f1_Quebrado_16x16": 0.7497296128055375, "eval_f1_RiadoRio_16x16": 0.5110356536502547, "eval_f1_RioFechado_16x16": 0.5922104720144442, "eval_loss": 2.068657159805298, "eval_precision": 0.633009751304958, "eval_precision_DuraRiadoRio_16x16": 0.6334964654703643, "eval_precision_Mole_16x16": 0.6310760282509348, "eval_precision_Quebrado_16x16": 0.7473048727899957, "eval_precision_RiadoRio_16x16": 0.4352226720647773, "eval_precision_RioFechado_16x16": 0.717948717948718, "eval_recall": 0.6079767530703039, "eval_recall_DuraRiadoRio_16x16": 0.5056423611111112, "eval_recall_Mole_16x16": 0.6592881944444444, "eval_recall_Quebrado_16x16": 0.7521701388888888, "eval_recall_RiadoRio_16x16": 0.6188322368421053, "eval_recall_RioFechado_16x16": 0.5039508340649693, "eval_runtime": 46.2658, "eval_samples_per_second": 251.201, "eval_steps_per_second": 15.714, "step": 447524 }, { "epoch": 77.00447350309705, "grad_norm": 9.344182968139648, "learning_rate": 1.1497763248451481e-05, "loss": 0.6424, "step": 447550 }, { "epoch": 77.01307639366827, "grad_norm": 22.419443130493164, "learning_rate": 1.1493461803165864e-05, "loss": 0.6464, "step": 447600 }, { "epoch": 77.02167928423951, "grad_norm": 29.017932891845703, "learning_rate": 1.1489160357880249e-05, "loss": 0.6529, "step": 447650 }, { "epoch": 77.03028217481074, "grad_norm": 30.08171272277832, "learning_rate": 1.1484858912594633e-05, "loss": 0.694, "step": 447700 }, { "epoch": 77.03888506538196, "grad_norm": 20.058317184448242, "learning_rate": 1.1480557467309016e-05, "loss": 0.6577, "step": 447750 }, { "epoch": 77.0474879559532, "grad_norm": 16.841642379760742, "learning_rate": 1.14762560220234e-05, "loss": 0.7028, "step": 447800 }, { "epoch": 77.05609084652443, "grad_norm": 13.031746864318848, "learning_rate": 1.1471954576737783e-05, "loss": 0.6859, "step": 447850 }, { "epoch": 77.06469373709567, "grad_norm": 8.416460037231445, "learning_rate": 1.146765313145217e-05, "loss": 0.6631, "step": 447900 }, { "epoch": 77.07329662766689, "grad_norm": 19.181413650512695, "learning_rate": 1.1463351686166553e-05, "loss": 0.7062, "step": 447950 }, { "epoch": 77.08189951823813, "grad_norm": 21.933191299438477, "learning_rate": 1.1459050240880937e-05, "loss": 0.6753, "step": 448000 }, { "epoch": 77.09050240880936, "grad_norm": 11.632862091064453, "learning_rate": 1.145474879559532e-05, "loss": 0.6872, "step": 448050 }, { "epoch": 77.0991052993806, "grad_norm": 27.86802101135254, "learning_rate": 1.1450447350309705e-05, "loss": 0.6744, "step": 448100 }, { "epoch": 77.10770818995182, "grad_norm": 16.400781631469727, "learning_rate": 1.1446145905024089e-05, "loss": 0.6246, "step": 448150 }, { "epoch": 77.11631108052306, "grad_norm": 12.261719703674316, "learning_rate": 1.1441844459738472e-05, "loss": 0.6723, "step": 448200 }, { "epoch": 77.12491397109429, "grad_norm": 19.682361602783203, "learning_rate": 1.1437543014452857e-05, "loss": 0.7029, "step": 448250 }, { "epoch": 77.13351686166553, "grad_norm": 18.433855056762695, "learning_rate": 1.1433241569167241e-05, "loss": 0.7072, "step": 448300 }, { "epoch": 77.14211975223675, "grad_norm": 9.700084686279297, "learning_rate": 1.1428940123881626e-05, "loss": 0.6089, "step": 448350 }, { "epoch": 77.15072264280798, "grad_norm": 15.849136352539062, "learning_rate": 1.1424638678596009e-05, "loss": 0.705, "step": 448400 }, { "epoch": 77.15932553337922, "grad_norm": 14.218981742858887, "learning_rate": 1.1420337233310391e-05, "loss": 0.6744, "step": 448450 }, { "epoch": 77.16792842395044, "grad_norm": 10.566424369812012, "learning_rate": 1.1416035788024778e-05, "loss": 0.7026, "step": 448500 }, { "epoch": 77.17653131452168, "grad_norm": 13.568634986877441, "learning_rate": 1.141173434273916e-05, "loss": 0.6631, "step": 448550 }, { "epoch": 77.18513420509291, "grad_norm": 10.748849868774414, "learning_rate": 1.1407432897453545e-05, "loss": 0.7067, "step": 448600 }, { "epoch": 77.19373709566415, "grad_norm": 9.743896484375, "learning_rate": 1.1403131452167928e-05, "loss": 0.684, "step": 448650 }, { "epoch": 77.20233998623537, "grad_norm": 11.820853233337402, "learning_rate": 1.1398830006882314e-05, "loss": 0.6727, "step": 448700 }, { "epoch": 77.21094287680661, "grad_norm": 6.30625057220459, "learning_rate": 1.1394528561596697e-05, "loss": 0.6855, "step": 448750 }, { "epoch": 77.21954576737784, "grad_norm": 12.429070472717285, "learning_rate": 1.139022711631108e-05, "loss": 0.6926, "step": 448800 }, { "epoch": 77.22814865794908, "grad_norm": 6.51588249206543, "learning_rate": 1.1385925671025465e-05, "loss": 0.6353, "step": 448850 }, { "epoch": 77.2367515485203, "grad_norm": 21.5793399810791, "learning_rate": 1.1381624225739849e-05, "loss": 0.6548, "step": 448900 }, { "epoch": 77.24535443909153, "grad_norm": 21.484071731567383, "learning_rate": 1.1377322780454234e-05, "loss": 0.6953, "step": 448950 }, { "epoch": 77.25395732966277, "grad_norm": 10.104816436767578, "learning_rate": 1.1373021335168617e-05, "loss": 0.6553, "step": 449000 }, { "epoch": 77.262560220234, "grad_norm": 9.82559871673584, "learning_rate": 1.1368719889883001e-05, "loss": 0.6522, "step": 449050 }, { "epoch": 77.27116311080523, "grad_norm": 16.721927642822266, "learning_rate": 1.1364418444597386e-05, "loss": 0.7045, "step": 449100 }, { "epoch": 77.27976600137646, "grad_norm": 24.97079086303711, "learning_rate": 1.1360116999311768e-05, "loss": 0.6708, "step": 449150 }, { "epoch": 77.2883688919477, "grad_norm": 15.615428924560547, "learning_rate": 1.1355815554026153e-05, "loss": 0.6843, "step": 449200 }, { "epoch": 77.29697178251892, "grad_norm": 15.174970626831055, "learning_rate": 1.1351514108740538e-05, "loss": 0.7585, "step": 449250 }, { "epoch": 77.30557467309016, "grad_norm": 10.207027435302734, "learning_rate": 1.1347212663454922e-05, "loss": 0.6318, "step": 449300 }, { "epoch": 77.31417756366139, "grad_norm": 8.229586601257324, "learning_rate": 1.1342911218169305e-05, "loss": 0.675, "step": 449350 }, { "epoch": 77.32278045423263, "grad_norm": 12.711627006530762, "learning_rate": 1.133860977288369e-05, "loss": 0.6091, "step": 449400 }, { "epoch": 77.33138334480385, "grad_norm": 8.300867080688477, "learning_rate": 1.1334308327598074e-05, "loss": 0.657, "step": 449450 }, { "epoch": 77.33998623537508, "grad_norm": 14.13022518157959, "learning_rate": 1.1330006882312457e-05, "loss": 0.6716, "step": 449500 }, { "epoch": 77.34858912594632, "grad_norm": 22.41037940979004, "learning_rate": 1.1325705437026842e-05, "loss": 0.7348, "step": 449550 }, { "epoch": 77.35719201651754, "grad_norm": 23.480186462402344, "learning_rate": 1.1321403991741224e-05, "loss": 0.7209, "step": 449600 }, { "epoch": 77.36579490708878, "grad_norm": 15.484097480773926, "learning_rate": 1.131710254645561e-05, "loss": 0.6573, "step": 449650 }, { "epoch": 77.37439779766001, "grad_norm": 12.705770492553711, "learning_rate": 1.1312801101169994e-05, "loss": 0.6476, "step": 449700 }, { "epoch": 77.38300068823125, "grad_norm": 20.27729606628418, "learning_rate": 1.1308499655884378e-05, "loss": 0.742, "step": 449750 }, { "epoch": 77.39160357880247, "grad_norm": 15.323966026306152, "learning_rate": 1.1304198210598761e-05, "loss": 0.6785, "step": 449800 }, { "epoch": 77.40020646937371, "grad_norm": 11.737505912780762, "learning_rate": 1.1299896765313146e-05, "loss": 0.6465, "step": 449850 }, { "epoch": 77.40880935994494, "grad_norm": 12.81900691986084, "learning_rate": 1.129559532002753e-05, "loss": 0.6549, "step": 449900 }, { "epoch": 77.41741225051618, "grad_norm": 9.391016960144043, "learning_rate": 1.1291293874741913e-05, "loss": 0.6676, "step": 449950 }, { "epoch": 77.4260151410874, "grad_norm": 23.442855834960938, "learning_rate": 1.1286992429456298e-05, "loss": 0.7063, "step": 450000 }, { "epoch": 77.43461803165864, "grad_norm": 19.580699920654297, "learning_rate": 1.1282690984170682e-05, "loss": 0.6561, "step": 450050 }, { "epoch": 77.44322092222987, "grad_norm": 14.327128410339355, "learning_rate": 1.1278389538885067e-05, "loss": 0.6633, "step": 450100 }, { "epoch": 77.4518238128011, "grad_norm": 14.899564743041992, "learning_rate": 1.127408809359945e-05, "loss": 0.6551, "step": 450150 }, { "epoch": 77.46042670337233, "grad_norm": 15.231908798217773, "learning_rate": 1.1269786648313834e-05, "loss": 0.6812, "step": 450200 }, { "epoch": 77.46902959394356, "grad_norm": 16.187402725219727, "learning_rate": 1.1265485203028219e-05, "loss": 0.65, "step": 450250 }, { "epoch": 77.4776324845148, "grad_norm": 23.814224243164062, "learning_rate": 1.1261183757742602e-05, "loss": 0.7171, "step": 450300 }, { "epoch": 77.48623537508603, "grad_norm": 11.657031059265137, "learning_rate": 1.1256882312456986e-05, "loss": 0.7428, "step": 450350 }, { "epoch": 77.49483826565726, "grad_norm": 17.00108528137207, "learning_rate": 1.1252580867171369e-05, "loss": 0.7129, "step": 450400 }, { "epoch": 77.50344115622849, "grad_norm": 15.491719245910645, "learning_rate": 1.1248279421885755e-05, "loss": 0.6865, "step": 450450 }, { "epoch": 77.51204404679973, "grad_norm": 15.634664535522461, "learning_rate": 1.1243977976600138e-05, "loss": 0.6947, "step": 450500 }, { "epoch": 77.52064693737096, "grad_norm": 17.082468032836914, "learning_rate": 1.1239676531314523e-05, "loss": 0.6753, "step": 450550 }, { "epoch": 77.5292498279422, "grad_norm": 26.094398498535156, "learning_rate": 1.1235375086028906e-05, "loss": 0.6761, "step": 450600 }, { "epoch": 77.53785271851342, "grad_norm": 11.704668045043945, "learning_rate": 1.123107364074329e-05, "loss": 0.6582, "step": 450650 }, { "epoch": 77.54645560908465, "grad_norm": 14.531861305236816, "learning_rate": 1.1226772195457675e-05, "loss": 0.6538, "step": 450700 }, { "epoch": 77.55505849965589, "grad_norm": 9.282621383666992, "learning_rate": 1.1222470750172057e-05, "loss": 0.694, "step": 450750 }, { "epoch": 77.56366139022711, "grad_norm": 11.3828763961792, "learning_rate": 1.1218169304886442e-05, "loss": 0.6775, "step": 450800 }, { "epoch": 77.57226428079835, "grad_norm": 13.719101905822754, "learning_rate": 1.1213867859600827e-05, "loss": 0.7022, "step": 450850 }, { "epoch": 77.58086717136958, "grad_norm": 12.641091346740723, "learning_rate": 1.1209566414315211e-05, "loss": 0.6702, "step": 450900 }, { "epoch": 77.58947006194082, "grad_norm": 18.321775436401367, "learning_rate": 1.1205264969029594e-05, "loss": 0.6575, "step": 450950 }, { "epoch": 77.59807295251204, "grad_norm": 18.17257308959961, "learning_rate": 1.1200963523743979e-05, "loss": 0.7115, "step": 451000 }, { "epoch": 77.60667584308328, "grad_norm": 12.746684074401855, "learning_rate": 1.1196662078458363e-05, "loss": 0.6238, "step": 451050 }, { "epoch": 77.6152787336545, "grad_norm": 17.206432342529297, "learning_rate": 1.1192360633172746e-05, "loss": 0.6654, "step": 451100 }, { "epoch": 77.62388162422575, "grad_norm": 11.80428695678711, "learning_rate": 1.118805918788713e-05, "loss": 0.665, "step": 451150 }, { "epoch": 77.63248451479697, "grad_norm": 17.73503875732422, "learning_rate": 1.1183757742601515e-05, "loss": 0.7266, "step": 451200 }, { "epoch": 77.64108740536821, "grad_norm": 14.572518348693848, "learning_rate": 1.11794562973159e-05, "loss": 0.665, "step": 451250 }, { "epoch": 77.64969029593944, "grad_norm": 15.935656547546387, "learning_rate": 1.1175154852030283e-05, "loss": 0.7227, "step": 451300 }, { "epoch": 77.65829318651066, "grad_norm": 13.100973129272461, "learning_rate": 1.1170853406744667e-05, "loss": 0.6568, "step": 451350 }, { "epoch": 77.6668960770819, "grad_norm": 12.081351280212402, "learning_rate": 1.1166551961459052e-05, "loss": 0.6849, "step": 451400 }, { "epoch": 77.67549896765313, "grad_norm": 18.54996109008789, "learning_rate": 1.1162250516173435e-05, "loss": 0.6013, "step": 451450 }, { "epoch": 77.68410185822437, "grad_norm": 16.119970321655273, "learning_rate": 1.1157949070887819e-05, "loss": 0.6539, "step": 451500 }, { "epoch": 77.69270474879559, "grad_norm": 22.529691696166992, "learning_rate": 1.1153647625602202e-05, "loss": 0.6764, "step": 451550 }, { "epoch": 77.70130763936683, "grad_norm": 10.869274139404297, "learning_rate": 1.1149346180316588e-05, "loss": 0.6339, "step": 451600 }, { "epoch": 77.70991052993806, "grad_norm": 11.844639778137207, "learning_rate": 1.1145044735030971e-05, "loss": 0.7193, "step": 451650 }, { "epoch": 77.7185134205093, "grad_norm": 14.163427352905273, "learning_rate": 1.1140743289745354e-05, "loss": 0.6698, "step": 451700 }, { "epoch": 77.72711631108052, "grad_norm": 14.917000770568848, "learning_rate": 1.1136441844459739e-05, "loss": 0.6914, "step": 451750 }, { "epoch": 77.73571920165176, "grad_norm": 24.166366577148438, "learning_rate": 1.1132140399174123e-05, "loss": 0.6432, "step": 451800 }, { "epoch": 77.74432209222299, "grad_norm": 15.046591758728027, "learning_rate": 1.1127838953888508e-05, "loss": 0.7195, "step": 451850 }, { "epoch": 77.75292498279421, "grad_norm": 15.217469215393066, "learning_rate": 1.112353750860289e-05, "loss": 0.6829, "step": 451900 }, { "epoch": 77.76152787336545, "grad_norm": 14.017075538635254, "learning_rate": 1.1119236063317275e-05, "loss": 0.6796, "step": 451950 }, { "epoch": 77.77013076393668, "grad_norm": 14.89627742767334, "learning_rate": 1.111493461803166e-05, "loss": 0.6662, "step": 452000 }, { "epoch": 77.77873365450792, "grad_norm": 15.265837669372559, "learning_rate": 1.1110633172746043e-05, "loss": 0.671, "step": 452050 }, { "epoch": 77.78733654507914, "grad_norm": 8.840740203857422, "learning_rate": 1.1106331727460427e-05, "loss": 0.7085, "step": 452100 }, { "epoch": 77.79593943565038, "grad_norm": 11.659348487854004, "learning_rate": 1.110203028217481e-05, "loss": 0.7213, "step": 452150 }, { "epoch": 77.80454232622161, "grad_norm": 11.570602416992188, "learning_rate": 1.1097728836889196e-05, "loss": 0.6572, "step": 452200 }, { "epoch": 77.81314521679285, "grad_norm": 22.739301681518555, "learning_rate": 1.1093427391603579e-05, "loss": 0.7113, "step": 452250 }, { "epoch": 77.82174810736407, "grad_norm": 15.895322799682617, "learning_rate": 1.1089125946317964e-05, "loss": 0.6779, "step": 452300 }, { "epoch": 77.83035099793531, "grad_norm": 15.208416938781738, "learning_rate": 1.1084824501032346e-05, "loss": 0.6486, "step": 452350 }, { "epoch": 77.83895388850654, "grad_norm": 8.759734153747559, "learning_rate": 1.1080523055746731e-05, "loss": 0.6135, "step": 452400 }, { "epoch": 77.84755677907776, "grad_norm": 17.729461669921875, "learning_rate": 1.1076221610461116e-05, "loss": 0.7021, "step": 452450 }, { "epoch": 77.856159669649, "grad_norm": 10.67471694946289, "learning_rate": 1.1071920165175498e-05, "loss": 0.641, "step": 452500 }, { "epoch": 77.86476256022023, "grad_norm": 20.093446731567383, "learning_rate": 1.1067618719889883e-05, "loss": 0.6754, "step": 452550 }, { "epoch": 77.87336545079147, "grad_norm": 17.36532974243164, "learning_rate": 1.1063317274604268e-05, "loss": 0.628, "step": 452600 }, { "epoch": 77.8819683413627, "grad_norm": 15.267143249511719, "learning_rate": 1.1059015829318652e-05, "loss": 0.6916, "step": 452650 }, { "epoch": 77.89057123193393, "grad_norm": 17.205883026123047, "learning_rate": 1.1054714384033035e-05, "loss": 0.6873, "step": 452700 }, { "epoch": 77.89917412250516, "grad_norm": 14.614529609680176, "learning_rate": 1.105041293874742e-05, "loss": 0.695, "step": 452750 }, { "epoch": 77.9077770130764, "grad_norm": 19.895401000976562, "learning_rate": 1.1046111493461804e-05, "loss": 0.6643, "step": 452800 }, { "epoch": 77.91637990364762, "grad_norm": 21.529067993164062, "learning_rate": 1.1041810048176187e-05, "loss": 0.7574, "step": 452850 }, { "epoch": 77.92498279421886, "grad_norm": 24.902271270751953, "learning_rate": 1.1037508602890572e-05, "loss": 0.6598, "step": 452900 }, { "epoch": 77.93358568479009, "grad_norm": 16.952251434326172, "learning_rate": 1.1033207157604956e-05, "loss": 0.6894, "step": 452950 }, { "epoch": 77.94218857536133, "grad_norm": 8.662456512451172, "learning_rate": 1.102890571231934e-05, "loss": 0.6785, "step": 453000 }, { "epoch": 77.95079146593255, "grad_norm": 19.320804595947266, "learning_rate": 1.1024604267033724e-05, "loss": 0.6887, "step": 453050 }, { "epoch": 77.95939435650378, "grad_norm": 27.81370735168457, "learning_rate": 1.1020302821748108e-05, "loss": 0.6718, "step": 453100 }, { "epoch": 77.96799724707502, "grad_norm": 8.388933181762695, "learning_rate": 1.1016001376462493e-05, "loss": 0.658, "step": 453150 }, { "epoch": 77.97660013764624, "grad_norm": 15.26375675201416, "learning_rate": 1.1011699931176876e-05, "loss": 0.7276, "step": 453200 }, { "epoch": 77.98520302821748, "grad_norm": 8.956306457519531, "learning_rate": 1.100739848589126e-05, "loss": 0.6697, "step": 453250 }, { "epoch": 77.99380591878871, "grad_norm": 19.06329345703125, "learning_rate": 1.1003097040605643e-05, "loss": 0.6471, "step": 453300 }, { "epoch": 78.0, "eval_accuracy": 0.48425400103252453, "eval_f1": 0.45553592794847103, "eval_f1_DuraRiadoRio_16x16": 0.2675250357653791, "eval_f1_Mole_16x16": 0.5232584524619923, "eval_f1_Quebrado_16x16": 0.5872093023255814, "eval_f1_RiadoRio_16x16": 0.40660566837759426, "eval_f1_RioFechado_16x16": 0.49308118081180813, "eval_loss": 3.5390477180480957, "eval_precision": 0.5407136036417303, "eval_precision_DuraRiadoRio_16x16": 0.7601626016260162, "eval_precision_Mole_16x16": 0.548264384213029, "eval_precision_Quebrado_16x16": 0.43109756097560975, "eval_precision_RiadoRio_16x16": 0.44460712542703756, "eval_precision_RioFechado_16x16": 0.5194363459669582, "eval_recall": 0.48543868794572037, "eval_recall_DuraRiadoRio_16x16": 0.1623263888888889, "eval_recall_Mole_16x16": 0.5004340277777778, "eval_recall_Quebrado_16x16": 0.9205729166666666, "eval_recall_RiadoRio_16x16": 0.3745888157894737, "eval_recall_RioFechado_16x16": 0.46927129060579453, "eval_runtime": 46.4177, "eval_samples_per_second": 250.379, "eval_steps_per_second": 15.662, "step": 453336 }, { "epoch": 78.00240880935995, "grad_norm": 13.337141990661621, "learning_rate": 1.099879559532003e-05, "loss": 0.7131, "step": 453350 }, { "epoch": 78.01101169993117, "grad_norm": 19.929515838623047, "learning_rate": 1.0994494150034412e-05, "loss": 0.6956, "step": 453400 }, { "epoch": 78.01961459050241, "grad_norm": 7.258577346801758, "learning_rate": 1.0990192704748797e-05, "loss": 0.6593, "step": 453450 }, { "epoch": 78.02821748107364, "grad_norm": 11.825085639953613, "learning_rate": 1.098589125946318e-05, "loss": 0.6078, "step": 453500 }, { "epoch": 78.03682037164488, "grad_norm": 21.498693466186523, "learning_rate": 1.0981589814177564e-05, "loss": 0.667, "step": 453550 }, { "epoch": 78.0454232622161, "grad_norm": 28.882461547851562, "learning_rate": 1.0977288368891949e-05, "loss": 0.6965, "step": 453600 }, { "epoch": 78.05402615278733, "grad_norm": 11.363053321838379, "learning_rate": 1.0972986923606332e-05, "loss": 0.6637, "step": 453650 }, { "epoch": 78.06262904335857, "grad_norm": 13.435772895812988, "learning_rate": 1.0968685478320716e-05, "loss": 0.6881, "step": 453700 }, { "epoch": 78.0712319339298, "grad_norm": 24.517454147338867, "learning_rate": 1.09643840330351e-05, "loss": 0.6722, "step": 453750 }, { "epoch": 78.07983482450103, "grad_norm": 14.117125511169434, "learning_rate": 1.0960082587749485e-05, "loss": 0.7032, "step": 453800 }, { "epoch": 78.08843771507226, "grad_norm": 12.669087409973145, "learning_rate": 1.0955781142463868e-05, "loss": 0.6997, "step": 453850 }, { "epoch": 78.0970406056435, "grad_norm": 9.175246238708496, "learning_rate": 1.0951479697178253e-05, "loss": 0.6464, "step": 453900 }, { "epoch": 78.10564349621473, "grad_norm": 24.786296844482422, "learning_rate": 1.0947178251892637e-05, "loss": 0.6934, "step": 453950 }, { "epoch": 78.11424638678596, "grad_norm": 19.66636085510254, "learning_rate": 1.094287680660702e-05, "loss": 0.6983, "step": 454000 }, { "epoch": 78.12284927735719, "grad_norm": 19.414682388305664, "learning_rate": 1.0938575361321405e-05, "loss": 0.6986, "step": 454050 }, { "epoch": 78.13145216792843, "grad_norm": 17.295639038085938, "learning_rate": 1.0934273916035787e-05, "loss": 0.674, "step": 454100 }, { "epoch": 78.14005505849966, "grad_norm": 20.30533790588379, "learning_rate": 1.0929972470750174e-05, "loss": 0.7034, "step": 454150 }, { "epoch": 78.1486579490709, "grad_norm": 10.435158729553223, "learning_rate": 1.0925671025464557e-05, "loss": 0.6662, "step": 454200 }, { "epoch": 78.15726083964212, "grad_norm": 20.157419204711914, "learning_rate": 1.0921369580178941e-05, "loss": 0.6864, "step": 454250 }, { "epoch": 78.16586373021335, "grad_norm": 20.292720794677734, "learning_rate": 1.0917068134893324e-05, "loss": 0.6414, "step": 454300 }, { "epoch": 78.17446662078459, "grad_norm": 14.959052085876465, "learning_rate": 1.0912766689607709e-05, "loss": 0.6151, "step": 454350 }, { "epoch": 78.18306951135581, "grad_norm": 17.281024932861328, "learning_rate": 1.0908465244322093e-05, "loss": 0.6966, "step": 454400 }, { "epoch": 78.19167240192705, "grad_norm": 9.782291412353516, "learning_rate": 1.0904163799036476e-05, "loss": 0.6813, "step": 454450 }, { "epoch": 78.20027529249828, "grad_norm": 9.763715744018555, "learning_rate": 1.089986235375086e-05, "loss": 0.6457, "step": 454500 }, { "epoch": 78.20887818306952, "grad_norm": 20.439498901367188, "learning_rate": 1.0895560908465245e-05, "loss": 0.656, "step": 454550 }, { "epoch": 78.21748107364074, "grad_norm": 12.405517578125, "learning_rate": 1.089125946317963e-05, "loss": 0.642, "step": 454600 }, { "epoch": 78.22608396421198, "grad_norm": 14.758041381835938, "learning_rate": 1.0886958017894013e-05, "loss": 0.6588, "step": 454650 }, { "epoch": 78.2346868547832, "grad_norm": 12.660143852233887, "learning_rate": 1.0882656572608397e-05, "loss": 0.6382, "step": 454700 }, { "epoch": 78.24328974535445, "grad_norm": 22.21129608154297, "learning_rate": 1.0878355127322782e-05, "loss": 0.721, "step": 454750 }, { "epoch": 78.25189263592567, "grad_norm": 12.839993476867676, "learning_rate": 1.0874053682037165e-05, "loss": 0.7523, "step": 454800 }, { "epoch": 78.2604955264969, "grad_norm": 19.976858139038086, "learning_rate": 1.0869752236751549e-05, "loss": 0.6469, "step": 454850 }, { "epoch": 78.26909841706814, "grad_norm": 18.156587600708008, "learning_rate": 1.0865450791465934e-05, "loss": 0.6912, "step": 454900 }, { "epoch": 78.27770130763936, "grad_norm": 13.685914039611816, "learning_rate": 1.0861149346180318e-05, "loss": 0.7013, "step": 454950 }, { "epoch": 78.2863041982106, "grad_norm": 7.1744160652160645, "learning_rate": 1.0856847900894701e-05, "loss": 0.7115, "step": 455000 }, { "epoch": 78.29490708878183, "grad_norm": 11.714698791503906, "learning_rate": 1.0852546455609084e-05, "loss": 0.6653, "step": 455050 }, { "epoch": 78.30350997935307, "grad_norm": 14.264583587646484, "learning_rate": 1.084824501032347e-05, "loss": 0.68, "step": 455100 }, { "epoch": 78.31211286992429, "grad_norm": 13.571578025817871, "learning_rate": 1.0843943565037853e-05, "loss": 0.7028, "step": 455150 }, { "epoch": 78.32071576049553, "grad_norm": 14.704729080200195, "learning_rate": 1.0839642119752238e-05, "loss": 0.7225, "step": 455200 }, { "epoch": 78.32931865106676, "grad_norm": 13.385514259338379, "learning_rate": 1.083534067446662e-05, "loss": 0.6539, "step": 455250 }, { "epoch": 78.337921541638, "grad_norm": 18.634302139282227, "learning_rate": 1.0831039229181005e-05, "loss": 0.6472, "step": 455300 }, { "epoch": 78.34652443220922, "grad_norm": 25.639883041381836, "learning_rate": 1.082673778389539e-05, "loss": 0.697, "step": 455350 }, { "epoch": 78.35512732278045, "grad_norm": 15.477303504943848, "learning_rate": 1.0822436338609772e-05, "loss": 0.7472, "step": 455400 }, { "epoch": 78.36373021335169, "grad_norm": 12.368717193603516, "learning_rate": 1.0818134893324157e-05, "loss": 0.6915, "step": 455450 }, { "epoch": 78.37233310392291, "grad_norm": 19.02264976501465, "learning_rate": 1.0813833448038542e-05, "loss": 0.698, "step": 455500 }, { "epoch": 78.38093599449415, "grad_norm": 35.94858169555664, "learning_rate": 1.0809532002752926e-05, "loss": 0.6582, "step": 455550 }, { "epoch": 78.38953888506538, "grad_norm": 12.082388877868652, "learning_rate": 1.0805230557467309e-05, "loss": 0.7058, "step": 455600 }, { "epoch": 78.39814177563662, "grad_norm": 12.113425254821777, "learning_rate": 1.0800929112181694e-05, "loss": 0.6338, "step": 455650 }, { "epoch": 78.40674466620784, "grad_norm": 21.323768615722656, "learning_rate": 1.0796627666896078e-05, "loss": 0.636, "step": 455700 }, { "epoch": 78.41534755677908, "grad_norm": 15.585256576538086, "learning_rate": 1.0792326221610461e-05, "loss": 0.6953, "step": 455750 }, { "epoch": 78.42395044735031, "grad_norm": 13.725298881530762, "learning_rate": 1.0788024776324846e-05, "loss": 0.6427, "step": 455800 }, { "epoch": 78.43255333792155, "grad_norm": 18.902992248535156, "learning_rate": 1.0783723331039228e-05, "loss": 0.6885, "step": 455850 }, { "epoch": 78.44115622849277, "grad_norm": 19.891908645629883, "learning_rate": 1.0779421885753615e-05, "loss": 0.6997, "step": 455900 }, { "epoch": 78.44975911906401, "grad_norm": 9.501873016357422, "learning_rate": 1.0775120440467998e-05, "loss": 0.6808, "step": 455950 }, { "epoch": 78.45836200963524, "grad_norm": 12.664923667907715, "learning_rate": 1.0770818995182382e-05, "loss": 0.6391, "step": 456000 }, { "epoch": 78.46696490020646, "grad_norm": 13.970789909362793, "learning_rate": 1.0766517549896765e-05, "loss": 0.7163, "step": 456050 }, { "epoch": 78.4755677907777, "grad_norm": 16.177776336669922, "learning_rate": 1.076221610461115e-05, "loss": 0.6307, "step": 456100 }, { "epoch": 78.48417068134893, "grad_norm": 12.426072120666504, "learning_rate": 1.0757914659325534e-05, "loss": 0.6781, "step": 456150 }, { "epoch": 78.49277357192017, "grad_norm": 19.231990814208984, "learning_rate": 1.0753613214039917e-05, "loss": 0.6527, "step": 456200 }, { "epoch": 78.5013764624914, "grad_norm": 17.974592208862305, "learning_rate": 1.0749311768754302e-05, "loss": 0.6867, "step": 456250 }, { "epoch": 78.50997935306263, "grad_norm": 6.559001922607422, "learning_rate": 1.0745010323468686e-05, "loss": 0.6189, "step": 456300 }, { "epoch": 78.51858224363386, "grad_norm": 13.519575119018555, "learning_rate": 1.074070887818307e-05, "loss": 0.657, "step": 456350 }, { "epoch": 78.5271851342051, "grad_norm": 11.935171127319336, "learning_rate": 1.0736407432897454e-05, "loss": 0.6674, "step": 456400 }, { "epoch": 78.53578802477632, "grad_norm": 21.41615867614746, "learning_rate": 1.0732105987611838e-05, "loss": 0.6796, "step": 456450 }, { "epoch": 78.54439091534756, "grad_norm": 19.141164779663086, "learning_rate": 1.0727804542326223e-05, "loss": 0.6444, "step": 456500 }, { "epoch": 78.55299380591879, "grad_norm": 23.963773727416992, "learning_rate": 1.0723503097040606e-05, "loss": 0.717, "step": 456550 }, { "epoch": 78.56159669649001, "grad_norm": 17.305068969726562, "learning_rate": 1.071920165175499e-05, "loss": 0.7099, "step": 456600 }, { "epoch": 78.57019958706125, "grad_norm": 14.670920372009277, "learning_rate": 1.0714900206469375e-05, "loss": 0.6465, "step": 456650 }, { "epoch": 78.57880247763248, "grad_norm": 14.774508476257324, "learning_rate": 1.071059876118376e-05, "loss": 0.6884, "step": 456700 }, { "epoch": 78.58740536820372, "grad_norm": 13.2818603515625, "learning_rate": 1.0706297315898142e-05, "loss": 0.649, "step": 456750 }, { "epoch": 78.59600825877494, "grad_norm": 18.911792755126953, "learning_rate": 1.0701995870612527e-05, "loss": 0.6802, "step": 456800 }, { "epoch": 78.60461114934618, "grad_norm": 14.223786354064941, "learning_rate": 1.0697694425326911e-05, "loss": 0.6431, "step": 456850 }, { "epoch": 78.61321403991741, "grad_norm": 20.174108505249023, "learning_rate": 1.0693392980041294e-05, "loss": 0.6858, "step": 456900 }, { "epoch": 78.62181693048865, "grad_norm": 8.511014938354492, "learning_rate": 1.0689091534755679e-05, "loss": 0.6293, "step": 456950 }, { "epoch": 78.63041982105987, "grad_norm": 24.123796463012695, "learning_rate": 1.0684790089470061e-05, "loss": 0.673, "step": 457000 }, { "epoch": 78.63902271163111, "grad_norm": 20.553102493286133, "learning_rate": 1.0680488644184448e-05, "loss": 0.6977, "step": 457050 }, { "epoch": 78.64762560220234, "grad_norm": 27.93335723876953, "learning_rate": 1.067618719889883e-05, "loss": 0.703, "step": 457100 }, { "epoch": 78.65622849277358, "grad_norm": 14.850741386413574, "learning_rate": 1.0671885753613215e-05, "loss": 0.658, "step": 457150 }, { "epoch": 78.6648313833448, "grad_norm": 11.030961990356445, "learning_rate": 1.0667584308327598e-05, "loss": 0.6421, "step": 457200 }, { "epoch": 78.67343427391603, "grad_norm": 15.681761741638184, "learning_rate": 1.0663282863041983e-05, "loss": 0.6677, "step": 457250 }, { "epoch": 78.68203716448727, "grad_norm": 12.360162734985352, "learning_rate": 1.0658981417756367e-05, "loss": 0.6654, "step": 457300 }, { "epoch": 78.6906400550585, "grad_norm": 13.610860824584961, "learning_rate": 1.065467997247075e-05, "loss": 0.7077, "step": 457350 }, { "epoch": 78.69924294562973, "grad_norm": 10.287308692932129, "learning_rate": 1.0650378527185135e-05, "loss": 0.6898, "step": 457400 }, { "epoch": 78.70784583620096, "grad_norm": 9.529984474182129, "learning_rate": 1.0646077081899519e-05, "loss": 0.7212, "step": 457450 }, { "epoch": 78.7164487267722, "grad_norm": 20.59107208251953, "learning_rate": 1.0641775636613904e-05, "loss": 0.6356, "step": 457500 }, { "epoch": 78.72505161734342, "grad_norm": 28.502517700195312, "learning_rate": 1.0637474191328287e-05, "loss": 0.699, "step": 457550 }, { "epoch": 78.73365450791466, "grad_norm": 19.058795928955078, "learning_rate": 1.0633172746042671e-05, "loss": 0.6836, "step": 457600 }, { "epoch": 78.74225739848589, "grad_norm": 8.843798637390137, "learning_rate": 1.0628871300757056e-05, "loss": 0.7127, "step": 457650 }, { "epoch": 78.75086028905713, "grad_norm": 14.599047660827637, "learning_rate": 1.0624569855471439e-05, "loss": 0.7042, "step": 457700 }, { "epoch": 78.75946317962836, "grad_norm": 19.053516387939453, "learning_rate": 1.0620268410185823e-05, "loss": 0.7061, "step": 457750 }, { "epoch": 78.76806607019958, "grad_norm": 12.262799263000488, "learning_rate": 1.0615966964900206e-05, "loss": 0.6922, "step": 457800 }, { "epoch": 78.77666896077082, "grad_norm": 21.28731346130371, "learning_rate": 1.0611665519614592e-05, "loss": 0.6846, "step": 457850 }, { "epoch": 78.78527185134205, "grad_norm": 13.62216854095459, "learning_rate": 1.0607364074328975e-05, "loss": 0.665, "step": 457900 }, { "epoch": 78.79387474191329, "grad_norm": 15.176942825317383, "learning_rate": 1.0603062629043358e-05, "loss": 0.6946, "step": 457950 }, { "epoch": 78.80247763248451, "grad_norm": 17.805438995361328, "learning_rate": 1.0598761183757743e-05, "loss": 0.6716, "step": 458000 }, { "epoch": 78.81108052305575, "grad_norm": 19.215530395507812, "learning_rate": 1.0594459738472127e-05, "loss": 0.6948, "step": 458050 }, { "epoch": 78.81968341362698, "grad_norm": 16.721120834350586, "learning_rate": 1.0590158293186512e-05, "loss": 0.7189, "step": 458100 }, { "epoch": 78.82828630419822, "grad_norm": 12.74980640411377, "learning_rate": 1.0585856847900895e-05, "loss": 0.6768, "step": 458150 }, { "epoch": 78.83688919476944, "grad_norm": 21.55829620361328, "learning_rate": 1.0581555402615279e-05, "loss": 0.6726, "step": 458200 }, { "epoch": 78.84549208534068, "grad_norm": 21.36546516418457, "learning_rate": 1.0577253957329664e-05, "loss": 0.7154, "step": 458250 }, { "epoch": 78.8540949759119, "grad_norm": 6.826078414916992, "learning_rate": 1.0572952512044046e-05, "loss": 0.6862, "step": 458300 }, { "epoch": 78.86269786648313, "grad_norm": 22.173219680786133, "learning_rate": 1.0568651066758431e-05, "loss": 0.7403, "step": 458350 }, { "epoch": 78.87130075705437, "grad_norm": 11.354109764099121, "learning_rate": 1.0564349621472816e-05, "loss": 0.6377, "step": 458400 }, { "epoch": 78.8799036476256, "grad_norm": 10.282560348510742, "learning_rate": 1.05600481761872e-05, "loss": 0.6354, "step": 458450 }, { "epoch": 78.88850653819684, "grad_norm": 24.26494789123535, "learning_rate": 1.0555746730901583e-05, "loss": 0.6636, "step": 458500 }, { "epoch": 78.89710942876806, "grad_norm": 22.955238342285156, "learning_rate": 1.0551445285615968e-05, "loss": 0.6991, "step": 458550 }, { "epoch": 78.9057123193393, "grad_norm": 9.435158729553223, "learning_rate": 1.0547143840330352e-05, "loss": 0.709, "step": 458600 }, { "epoch": 78.91431520991053, "grad_norm": 17.33127212524414, "learning_rate": 1.0542842395044735e-05, "loss": 0.7281, "step": 458650 }, { "epoch": 78.92291810048177, "grad_norm": 14.008867263793945, "learning_rate": 1.053854094975912e-05, "loss": 0.6635, "step": 458700 }, { "epoch": 78.93152099105299, "grad_norm": 16.738407135009766, "learning_rate": 1.0534239504473502e-05, "loss": 0.6565, "step": 458750 }, { "epoch": 78.94012388162423, "grad_norm": 12.946025848388672, "learning_rate": 1.0529938059187889e-05, "loss": 0.6495, "step": 458800 }, { "epoch": 78.94872677219546, "grad_norm": 13.645745277404785, "learning_rate": 1.0525636613902272e-05, "loss": 0.6743, "step": 458850 }, { "epoch": 78.9573296627667, "grad_norm": 13.916584968566895, "learning_rate": 1.0521335168616656e-05, "loss": 0.6832, "step": 458900 }, { "epoch": 78.96593255333792, "grad_norm": 22.89187240600586, "learning_rate": 1.0517033723331039e-05, "loss": 0.6878, "step": 458950 }, { "epoch": 78.97453544390915, "grad_norm": 18.768287658691406, "learning_rate": 1.0512732278045424e-05, "loss": 0.6737, "step": 459000 }, { "epoch": 78.98313833448039, "grad_norm": 18.082128524780273, "learning_rate": 1.0508430832759808e-05, "loss": 0.6882, "step": 459050 }, { "epoch": 78.99174122505161, "grad_norm": 10.910798072814941, "learning_rate": 1.0504129387474191e-05, "loss": 0.7424, "step": 459100 }, { "epoch": 79.0, "eval_accuracy": 0.5605747719841679, "eval_f1": 0.5556115844307431, "eval_f1_DuraRiadoRio_16x16": 0.5708282262300823, "eval_f1_Mole_16x16": 0.538082191780822, "eval_f1_Quebrado_16x16": 0.6632574789056507, "eval_f1_RiadoRio_16x16": 0.4144516129032258, "eval_f1_RioFechado_16x16": 0.5914384123339347, "eval_loss": 2.3235280513763428, "eval_precision": 0.608736500206712, "eval_precision_DuraRiadoRio_16x16": 0.478426768417963, "eval_precision_Mole_16x16": 0.7295690936106983, "eval_precision_Quebrado_16x16": 0.8070939639079029, "eval_precision_RiadoRio_16x16": 0.5564795564795565, "eval_precision_RioFechado_16x16": 0.4721131186174391, "eval_recall": 0.5636558524136285, "eval_recall_DuraRiadoRio_16x16": 0.7074652777777778, "eval_recall_Mole_16x16": 0.4262152777777778, "eval_recall_Quebrado_16x16": 0.5629340277777778, "eval_recall_RiadoRio_16x16": 0.3301809210526316, "eval_recall_RioFechado_16x16": 0.7914837576821774, "eval_runtime": 46.3471, "eval_samples_per_second": 250.76, "eval_steps_per_second": 15.686, "step": 459148 }, { "epoch": 79.00034411562285, "grad_norm": 15.579867362976074, "learning_rate": 1.0499827942188576e-05, "loss": 0.6458, "step": 459150 }, { "epoch": 79.00894700619408, "grad_norm": 11.436902046203613, "learning_rate": 1.049552649690296e-05, "loss": 0.6841, "step": 459200 }, { "epoch": 79.01754989676532, "grad_norm": 11.194107055664062, "learning_rate": 1.0491225051617345e-05, "loss": 0.6355, "step": 459250 }, { "epoch": 79.02615278733654, "grad_norm": 14.603218078613281, "learning_rate": 1.0486923606331728e-05, "loss": 0.6677, "step": 459300 }, { "epoch": 79.03475567790778, "grad_norm": 12.241548538208008, "learning_rate": 1.0482622161046112e-05, "loss": 0.6102, "step": 459350 }, { "epoch": 79.04335856847901, "grad_norm": 13.32249641418457, "learning_rate": 1.0478320715760497e-05, "loss": 0.6643, "step": 459400 }, { "epoch": 79.05196145905025, "grad_norm": 9.770744323730469, "learning_rate": 1.047401927047488e-05, "loss": 0.6268, "step": 459450 }, { "epoch": 79.06056434962147, "grad_norm": 15.06916618347168, "learning_rate": 1.0469717825189264e-05, "loss": 0.7242, "step": 459500 }, { "epoch": 79.0691672401927, "grad_norm": 8.862161636352539, "learning_rate": 1.0465416379903647e-05, "loss": 0.662, "step": 459550 }, { "epoch": 79.07777013076394, "grad_norm": 14.031639099121094, "learning_rate": 1.0461114934618033e-05, "loss": 0.6631, "step": 459600 }, { "epoch": 79.08637302133516, "grad_norm": 22.400699615478516, "learning_rate": 1.0456813489332416e-05, "loss": 0.6361, "step": 459650 }, { "epoch": 79.0949759119064, "grad_norm": 8.69157600402832, "learning_rate": 1.04525120440468e-05, "loss": 0.6564, "step": 459700 }, { "epoch": 79.10357880247763, "grad_norm": 15.76465892791748, "learning_rate": 1.0448210598761183e-05, "loss": 0.6538, "step": 459750 }, { "epoch": 79.11218169304887, "grad_norm": 15.062689781188965, "learning_rate": 1.0443909153475568e-05, "loss": 0.6989, "step": 459800 }, { "epoch": 79.1207845836201, "grad_norm": 12.076034545898438, "learning_rate": 1.0439607708189953e-05, "loss": 0.704, "step": 459850 }, { "epoch": 79.12938747419133, "grad_norm": 15.136847496032715, "learning_rate": 1.0435306262904335e-05, "loss": 0.6716, "step": 459900 }, { "epoch": 79.13799036476256, "grad_norm": 18.334136962890625, "learning_rate": 1.043100481761872e-05, "loss": 0.662, "step": 459950 }, { "epoch": 79.1465932553338, "grad_norm": 10.879618644714355, "learning_rate": 1.0426703372333105e-05, "loss": 0.6375, "step": 460000 }, { "epoch": 79.15519614590502, "grad_norm": 15.534621238708496, "learning_rate": 1.042240192704749e-05, "loss": 0.6906, "step": 460050 }, { "epoch": 79.16379903647626, "grad_norm": 14.317586898803711, "learning_rate": 1.0418100481761872e-05, "loss": 0.6541, "step": 460100 }, { "epoch": 79.17240192704749, "grad_norm": 6.649514198303223, "learning_rate": 1.0413799036476257e-05, "loss": 0.6684, "step": 460150 }, { "epoch": 79.18100481761871, "grad_norm": 23.28426742553711, "learning_rate": 1.0409497591190641e-05, "loss": 0.6642, "step": 460200 }, { "epoch": 79.18960770818995, "grad_norm": 18.038572311401367, "learning_rate": 1.0405196145905024e-05, "loss": 0.6833, "step": 460250 }, { "epoch": 79.19821059876118, "grad_norm": 8.0079984664917, "learning_rate": 1.0400894700619409e-05, "loss": 0.7079, "step": 460300 }, { "epoch": 79.20681348933242, "grad_norm": 26.926868438720703, "learning_rate": 1.0396593255333793e-05, "loss": 0.6917, "step": 460350 }, { "epoch": 79.21541637990364, "grad_norm": 13.433387756347656, "learning_rate": 1.0392291810048178e-05, "loss": 0.7249, "step": 460400 }, { "epoch": 79.22401927047488, "grad_norm": 17.45774269104004, "learning_rate": 1.038799036476256e-05, "loss": 0.6352, "step": 460450 }, { "epoch": 79.23262216104611, "grad_norm": 10.009540557861328, "learning_rate": 1.0383688919476945e-05, "loss": 0.6711, "step": 460500 }, { "epoch": 79.24122505161735, "grad_norm": 14.33299732208252, "learning_rate": 1.037938747419133e-05, "loss": 0.6901, "step": 460550 }, { "epoch": 79.24982794218857, "grad_norm": 20.883638381958008, "learning_rate": 1.0375086028905713e-05, "loss": 0.6814, "step": 460600 }, { "epoch": 79.25843083275981, "grad_norm": 25.19229507446289, "learning_rate": 1.0370784583620097e-05, "loss": 0.6436, "step": 460650 }, { "epoch": 79.26703372333104, "grad_norm": 16.331233978271484, "learning_rate": 1.036648313833448e-05, "loss": 0.6372, "step": 460700 }, { "epoch": 79.27563661390226, "grad_norm": 11.839695930480957, "learning_rate": 1.0362181693048866e-05, "loss": 0.7226, "step": 460750 }, { "epoch": 79.2842395044735, "grad_norm": 15.178011894226074, "learning_rate": 1.0357880247763249e-05, "loss": 0.7206, "step": 460800 }, { "epoch": 79.29284239504473, "grad_norm": 16.227991104125977, "learning_rate": 1.0353578802477634e-05, "loss": 0.7112, "step": 460850 }, { "epoch": 79.30144528561597, "grad_norm": 18.870330810546875, "learning_rate": 1.0349277357192017e-05, "loss": 0.7379, "step": 460900 }, { "epoch": 79.3100481761872, "grad_norm": 7.16285514831543, "learning_rate": 1.0344975911906401e-05, "loss": 0.6426, "step": 460950 }, { "epoch": 79.31865106675843, "grad_norm": 20.05997085571289, "learning_rate": 1.0340674466620786e-05, "loss": 0.7076, "step": 461000 }, { "epoch": 79.32725395732966, "grad_norm": 19.828903198242188, "learning_rate": 1.0336373021335169e-05, "loss": 0.6603, "step": 461050 }, { "epoch": 79.3358568479009, "grad_norm": 23.17127227783203, "learning_rate": 1.0332071576049553e-05, "loss": 0.6145, "step": 461100 }, { "epoch": 79.34445973847212, "grad_norm": 11.802285194396973, "learning_rate": 1.0327770130763938e-05, "loss": 0.6868, "step": 461150 }, { "epoch": 79.35306262904336, "grad_norm": 15.217665672302246, "learning_rate": 1.0323468685478322e-05, "loss": 0.686, "step": 461200 }, { "epoch": 79.36166551961459, "grad_norm": 11.712254524230957, "learning_rate": 1.0319167240192705e-05, "loss": 0.6926, "step": 461250 }, { "epoch": 79.37026841018582, "grad_norm": 18.920778274536133, "learning_rate": 1.0314865794907088e-05, "loss": 0.6564, "step": 461300 }, { "epoch": 79.37887130075705, "grad_norm": 13.987428665161133, "learning_rate": 1.0310564349621474e-05, "loss": 0.6823, "step": 461350 }, { "epoch": 79.38747419132828, "grad_norm": 7.810690879821777, "learning_rate": 1.0306262904335857e-05, "loss": 0.5973, "step": 461400 }, { "epoch": 79.39607708189952, "grad_norm": 19.325300216674805, "learning_rate": 1.0301961459050242e-05, "loss": 0.6627, "step": 461450 }, { "epoch": 79.40467997247075, "grad_norm": 24.669105529785156, "learning_rate": 1.0297660013764624e-05, "loss": 0.6461, "step": 461500 }, { "epoch": 79.41328286304199, "grad_norm": 8.20857048034668, "learning_rate": 1.0293358568479009e-05, "loss": 0.655, "step": 461550 }, { "epoch": 79.42188575361321, "grad_norm": 16.114776611328125, "learning_rate": 1.0289057123193394e-05, "loss": 0.64, "step": 461600 }, { "epoch": 79.43048864418445, "grad_norm": 21.23495864868164, "learning_rate": 1.0284755677907776e-05, "loss": 0.7276, "step": 461650 }, { "epoch": 79.43909153475568, "grad_norm": 14.39787483215332, "learning_rate": 1.0280454232622161e-05, "loss": 0.6277, "step": 461700 }, { "epoch": 79.44769442532692, "grad_norm": 13.23657512664795, "learning_rate": 1.0276152787336546e-05, "loss": 0.6579, "step": 461750 }, { "epoch": 79.45629731589814, "grad_norm": 21.123945236206055, "learning_rate": 1.027185134205093e-05, "loss": 0.6929, "step": 461800 }, { "epoch": 79.46490020646938, "grad_norm": 10.340636253356934, "learning_rate": 1.0267549896765313e-05, "loss": 0.6443, "step": 461850 }, { "epoch": 79.4735030970406, "grad_norm": 12.483193397521973, "learning_rate": 1.0263248451479698e-05, "loss": 0.657, "step": 461900 }, { "epoch": 79.48210598761183, "grad_norm": 20.430479049682617, "learning_rate": 1.0258947006194082e-05, "loss": 0.6533, "step": 461950 }, { "epoch": 79.49070887818307, "grad_norm": 28.732036590576172, "learning_rate": 1.0254645560908465e-05, "loss": 0.6667, "step": 462000 }, { "epoch": 79.4993117687543, "grad_norm": 12.757789611816406, "learning_rate": 1.025034411562285e-05, "loss": 0.677, "step": 462050 }, { "epoch": 79.50791465932554, "grad_norm": 13.784245491027832, "learning_rate": 1.0246042670337234e-05, "loss": 0.6377, "step": 462100 }, { "epoch": 79.51651754989676, "grad_norm": 17.633037567138672, "learning_rate": 1.0241741225051619e-05, "loss": 0.6631, "step": 462150 }, { "epoch": 79.525120440468, "grad_norm": 17.838651657104492, "learning_rate": 1.0237439779766002e-05, "loss": 0.6397, "step": 462200 }, { "epoch": 79.53372333103923, "grad_norm": 11.108759880065918, "learning_rate": 1.0233138334480386e-05, "loss": 0.6358, "step": 462250 }, { "epoch": 79.54232622161047, "grad_norm": 23.122940063476562, "learning_rate": 1.022883688919477e-05, "loss": 0.6812, "step": 462300 }, { "epoch": 79.55092911218169, "grad_norm": 15.093680381774902, "learning_rate": 1.0224535443909154e-05, "loss": 0.6267, "step": 462350 }, { "epoch": 79.55953200275293, "grad_norm": 16.33039093017578, "learning_rate": 1.0220233998623538e-05, "loss": 0.6934, "step": 462400 }, { "epoch": 79.56813489332416, "grad_norm": 20.45105743408203, "learning_rate": 1.0215932553337921e-05, "loss": 0.6397, "step": 462450 }, { "epoch": 79.57673778389538, "grad_norm": 17.214752197265625, "learning_rate": 1.0211631108052307e-05, "loss": 0.6413, "step": 462500 }, { "epoch": 79.58534067446662, "grad_norm": 7.75523042678833, "learning_rate": 1.020732966276669e-05, "loss": 0.726, "step": 462550 }, { "epoch": 79.59394356503785, "grad_norm": 8.616045951843262, "learning_rate": 1.0203028217481075e-05, "loss": 0.6611, "step": 462600 }, { "epoch": 79.60254645560909, "grad_norm": 11.766274452209473, "learning_rate": 1.0198726772195458e-05, "loss": 0.6394, "step": 462650 }, { "epoch": 79.61114934618031, "grad_norm": 14.629734992980957, "learning_rate": 1.0194425326909842e-05, "loss": 0.6786, "step": 462700 }, { "epoch": 79.61975223675155, "grad_norm": 8.431346893310547, "learning_rate": 1.0190123881624227e-05, "loss": 0.6829, "step": 462750 }, { "epoch": 79.62835512732278, "grad_norm": 16.89169692993164, "learning_rate": 1.018582243633861e-05, "loss": 0.6697, "step": 462800 }, { "epoch": 79.63695801789402, "grad_norm": 14.412796020507812, "learning_rate": 1.0181520991052994e-05, "loss": 0.6827, "step": 462850 }, { "epoch": 79.64556090846524, "grad_norm": 20.464651107788086, "learning_rate": 1.0177219545767379e-05, "loss": 0.6933, "step": 462900 }, { "epoch": 79.65416379903648, "grad_norm": 18.544633865356445, "learning_rate": 1.0172918100481763e-05, "loss": 0.6545, "step": 462950 }, { "epoch": 79.66276668960771, "grad_norm": 13.021044731140137, "learning_rate": 1.0168616655196146e-05, "loss": 0.632, "step": 463000 }, { "epoch": 79.67136958017895, "grad_norm": 17.90346336364746, "learning_rate": 1.016431520991053e-05, "loss": 0.6797, "step": 463050 }, { "epoch": 79.67997247075017, "grad_norm": 23.785327911376953, "learning_rate": 1.0160013764624915e-05, "loss": 0.6681, "step": 463100 }, { "epoch": 79.6885753613214, "grad_norm": 13.805627822875977, "learning_rate": 1.0155712319339298e-05, "loss": 0.6476, "step": 463150 }, { "epoch": 79.69717825189264, "grad_norm": 13.63866138458252, "learning_rate": 1.0151410874053683e-05, "loss": 0.6728, "step": 463200 }, { "epoch": 79.70578114246386, "grad_norm": 14.56448745727539, "learning_rate": 1.0147109428768065e-05, "loss": 0.6942, "step": 463250 }, { "epoch": 79.7143840330351, "grad_norm": 12.60762882232666, "learning_rate": 1.0142807983482452e-05, "loss": 0.6894, "step": 463300 }, { "epoch": 79.72298692360633, "grad_norm": 18.46689796447754, "learning_rate": 1.0138506538196835e-05, "loss": 0.6981, "step": 463350 }, { "epoch": 79.73158981417757, "grad_norm": 11.919595718383789, "learning_rate": 1.0134205092911219e-05, "loss": 0.6854, "step": 463400 }, { "epoch": 79.7401927047488, "grad_norm": 16.25169563293457, "learning_rate": 1.0129903647625602e-05, "loss": 0.6245, "step": 463450 }, { "epoch": 79.74879559532003, "grad_norm": 12.97283935546875, "learning_rate": 1.0125602202339987e-05, "loss": 0.7281, "step": 463500 }, { "epoch": 79.75739848589126, "grad_norm": 12.021403312683105, "learning_rate": 1.0121300757054371e-05, "loss": 0.6447, "step": 463550 }, { "epoch": 79.7660013764625, "grad_norm": 22.456907272338867, "learning_rate": 1.0116999311768754e-05, "loss": 0.6993, "step": 463600 }, { "epoch": 79.77460426703372, "grad_norm": 9.00274658203125, "learning_rate": 1.0112697866483139e-05, "loss": 0.6895, "step": 463650 }, { "epoch": 79.78320715760495, "grad_norm": 14.631280899047852, "learning_rate": 1.0108396421197523e-05, "loss": 0.6766, "step": 463700 }, { "epoch": 79.79181004817619, "grad_norm": 13.994339942932129, "learning_rate": 1.0104094975911908e-05, "loss": 0.6443, "step": 463750 }, { "epoch": 79.80041293874741, "grad_norm": 17.614017486572266, "learning_rate": 1.009979353062629e-05, "loss": 0.7067, "step": 463800 }, { "epoch": 79.80901582931865, "grad_norm": 19.808931350708008, "learning_rate": 1.0095492085340675e-05, "loss": 0.6655, "step": 463850 }, { "epoch": 79.81761871988988, "grad_norm": 12.556854248046875, "learning_rate": 1.009119064005506e-05, "loss": 0.6507, "step": 463900 }, { "epoch": 79.82622161046112, "grad_norm": 16.174184799194336, "learning_rate": 1.0086889194769443e-05, "loss": 0.665, "step": 463950 }, { "epoch": 79.83482450103234, "grad_norm": 16.269269943237305, "learning_rate": 1.0082587749483827e-05, "loss": 0.7096, "step": 464000 }, { "epoch": 79.84342739160358, "grad_norm": 22.53375244140625, "learning_rate": 1.0078286304198212e-05, "loss": 0.7145, "step": 464050 }, { "epoch": 79.85203028217481, "grad_norm": 13.137804985046387, "learning_rate": 1.0073984858912596e-05, "loss": 0.6486, "step": 464100 }, { "epoch": 79.86063317274605, "grad_norm": 16.03273582458496, "learning_rate": 1.0069683413626979e-05, "loss": 0.6053, "step": 464150 }, { "epoch": 79.86923606331727, "grad_norm": 11.090837478637695, "learning_rate": 1.0065381968341362e-05, "loss": 0.7011, "step": 464200 }, { "epoch": 79.8778389538885, "grad_norm": 18.3441162109375, "learning_rate": 1.0061080523055748e-05, "loss": 0.6834, "step": 464250 }, { "epoch": 79.88644184445974, "grad_norm": 14.693622589111328, "learning_rate": 1.0056779077770131e-05, "loss": 0.7174, "step": 464300 }, { "epoch": 79.89504473503096, "grad_norm": 8.204846382141113, "learning_rate": 1.0052477632484516e-05, "loss": 0.6716, "step": 464350 }, { "epoch": 79.9036476256022, "grad_norm": 7.501256465911865, "learning_rate": 1.0048176187198898e-05, "loss": 0.7385, "step": 464400 }, { "epoch": 79.91225051617343, "grad_norm": 23.971593856811523, "learning_rate": 1.0043874741913285e-05, "loss": 0.7161, "step": 464450 }, { "epoch": 79.92085340674467, "grad_norm": 13.97546100616455, "learning_rate": 1.0039573296627668e-05, "loss": 0.7168, "step": 464500 }, { "epoch": 79.9294562973159, "grad_norm": 14.552392959594727, "learning_rate": 1.003527185134205e-05, "loss": 0.6503, "step": 464550 }, { "epoch": 79.93805918788713, "grad_norm": 17.85015106201172, "learning_rate": 1.0030970406056435e-05, "loss": 0.6667, "step": 464600 }, { "epoch": 79.94666207845836, "grad_norm": 27.870220184326172, "learning_rate": 1.002666896077082e-05, "loss": 0.631, "step": 464650 }, { "epoch": 79.9552649690296, "grad_norm": 17.668212890625, "learning_rate": 1.0022367515485204e-05, "loss": 0.6663, "step": 464700 }, { "epoch": 79.96386785960082, "grad_norm": 11.499092102050781, "learning_rate": 1.0018066070199587e-05, "loss": 0.6642, "step": 464750 }, { "epoch": 79.97247075017206, "grad_norm": 11.870362281799316, "learning_rate": 1.0013764624913972e-05, "loss": 0.6877, "step": 464800 }, { "epoch": 79.98107364074329, "grad_norm": 14.641399383544922, "learning_rate": 1.0009463179628356e-05, "loss": 0.7161, "step": 464850 }, { "epoch": 79.98967653131452, "grad_norm": 7.396857738494873, "learning_rate": 1.0005161734342739e-05, "loss": 0.6776, "step": 464900 }, { "epoch": 79.99827942188575, "grad_norm": 17.24068832397461, "learning_rate": 1.0000860289057124e-05, "loss": 0.7054, "step": 464950 }, { "epoch": 80.0, "eval_accuracy": 0.4989674754775426, "eval_f1": 0.4717763372194046, "eval_f1_DuraRiadoRio_16x16": 0.5129181397878706, "eval_f1_Mole_16x16": 0.1588893173929811, "eval_f1_Quebrado_16x16": 0.7147979918718623, "eval_f1_RiadoRio_16x16": 0.4181502388189318, "eval_f1_RioFechado_16x16": 0.5541259982253771, "eval_loss": 3.6876909732818604, "eval_precision": 0.5769912117782854, "eval_precision_DuraRiadoRio_16x16": 0.37346534653465346, "eval_precision_Mole_16x16": 0.71280276816609, "eval_precision_Quebrado_16x16": 0.795635976583289, "eval_precision_RiadoRio_16x16": 0.4429622815087397, "eval_precision_RioFechado_16x16": 0.5600896860986547, "eval_recall": 0.5002232011061822, "eval_recall_DuraRiadoRio_16x16": 0.8185763888888888, "eval_recall_Mole_16x16": 0.08940972222222222, "eval_recall_Quebrado_16x16": 0.6488715277777778, "eval_recall_RiadoRio_16x16": 0.3959703947368421, "eval_recall_RioFechado_16x16": 0.54828797190518, "eval_runtime": 46.4064, "eval_samples_per_second": 250.439, "eval_steps_per_second": 15.666, "step": 464960 }, { "epoch": 80.00688231245698, "grad_norm": 10.1148042678833, "learning_rate": 9.996558843771506e-06, "loss": 0.6906, "step": 465000 }, { "epoch": 80.01548520302822, "grad_norm": 10.757468223571777, "learning_rate": 9.992257398485893e-06, "loss": 0.6686, "step": 465050 }, { "epoch": 80.02408809359945, "grad_norm": 14.145272254943848, "learning_rate": 9.987955953200276e-06, "loss": 0.6528, "step": 465100 }, { "epoch": 80.03269098417069, "grad_norm": 13.3259859085083, "learning_rate": 9.98365450791466e-06, "loss": 0.7152, "step": 465150 }, { "epoch": 80.04129387474191, "grad_norm": 10.45158576965332, "learning_rate": 9.979353062629043e-06, "loss": 0.6556, "step": 465200 }, { "epoch": 80.04989676531315, "grad_norm": 15.823253631591797, "learning_rate": 9.975051617343428e-06, "loss": 0.6784, "step": 465250 }, { "epoch": 80.05849965588438, "grad_norm": 17.47986602783203, "learning_rate": 9.970750172057812e-06, "loss": 0.6364, "step": 465300 }, { "epoch": 80.06710254645562, "grad_norm": 21.824377059936523, "learning_rate": 9.966448726772195e-06, "loss": 0.6667, "step": 465350 }, { "epoch": 80.07570543702684, "grad_norm": 16.63699722290039, "learning_rate": 9.96214728148658e-06, "loss": 0.6545, "step": 465400 }, { "epoch": 80.08430832759807, "grad_norm": 14.927364349365234, "learning_rate": 9.957845836200964e-06, "loss": 0.6702, "step": 465450 }, { "epoch": 80.0929112181693, "grad_norm": 33.247398376464844, "learning_rate": 9.953544390915349e-06, "loss": 0.7139, "step": 465500 }, { "epoch": 80.10151410874053, "grad_norm": 10.40831184387207, "learning_rate": 9.949242945629732e-06, "loss": 0.68, "step": 465550 }, { "epoch": 80.11011699931177, "grad_norm": 17.975440979003906, "learning_rate": 9.944941500344116e-06, "loss": 0.6715, "step": 465600 }, { "epoch": 80.118719889883, "grad_norm": 12.061767578125, "learning_rate": 9.9406400550585e-06, "loss": 0.6513, "step": 465650 }, { "epoch": 80.12732278045424, "grad_norm": 13.576888084411621, "learning_rate": 9.936338609772884e-06, "loss": 0.687, "step": 465700 }, { "epoch": 80.13592567102546, "grad_norm": 23.384347915649414, "learning_rate": 9.932037164487268e-06, "loss": 0.6649, "step": 465750 }, { "epoch": 80.1445285615967, "grad_norm": 8.918794631958008, "learning_rate": 9.927735719201653e-06, "loss": 0.6661, "step": 465800 }, { "epoch": 80.15313145216793, "grad_norm": 14.50122356414795, "learning_rate": 9.923434273916037e-06, "loss": 0.6554, "step": 465850 }, { "epoch": 80.16173434273917, "grad_norm": 5.794430255889893, "learning_rate": 9.91913282863042e-06, "loss": 0.6626, "step": 465900 }, { "epoch": 80.17033723331039, "grad_norm": 11.927450180053711, "learning_rate": 9.914831383344805e-06, "loss": 0.6892, "step": 465950 }, { "epoch": 80.17894012388163, "grad_norm": 11.304521560668945, "learning_rate": 9.91052993805919e-06, "loss": 0.6654, "step": 466000 }, { "epoch": 80.18754301445286, "grad_norm": 28.416166305541992, "learning_rate": 9.906228492773572e-06, "loss": 0.6941, "step": 466050 }, { "epoch": 80.19614590502408, "grad_norm": 12.81706428527832, "learning_rate": 9.901927047487957e-06, "loss": 0.6655, "step": 466100 }, { "epoch": 80.20474879559532, "grad_norm": 29.267864227294922, "learning_rate": 9.89762560220234e-06, "loss": 0.7269, "step": 466150 }, { "epoch": 80.21335168616655, "grad_norm": 15.890045166015625, "learning_rate": 9.893324156916726e-06, "loss": 0.6696, "step": 466200 }, { "epoch": 80.22195457673779, "grad_norm": 16.999755859375, "learning_rate": 9.889022711631109e-06, "loss": 0.6515, "step": 466250 }, { "epoch": 80.23055746730901, "grad_norm": 12.041934967041016, "learning_rate": 9.884721266345493e-06, "loss": 0.6813, "step": 466300 }, { "epoch": 80.23916035788025, "grad_norm": 11.71735668182373, "learning_rate": 9.880419821059876e-06, "loss": 0.703, "step": 466350 }, { "epoch": 80.24776324845148, "grad_norm": 21.877784729003906, "learning_rate": 9.87611837577426e-06, "loss": 0.6788, "step": 466400 }, { "epoch": 80.25636613902272, "grad_norm": 15.79531478881836, "learning_rate": 9.871816930488645e-06, "loss": 0.6408, "step": 466450 }, { "epoch": 80.26496902959394, "grad_norm": 20.070844650268555, "learning_rate": 9.867515485203028e-06, "loss": 0.6987, "step": 466500 }, { "epoch": 80.27357192016518, "grad_norm": 17.427797317504883, "learning_rate": 9.863214039917413e-06, "loss": 0.6591, "step": 466550 }, { "epoch": 80.28217481073641, "grad_norm": 12.98167610168457, "learning_rate": 9.858912594631797e-06, "loss": 0.6519, "step": 466600 }, { "epoch": 80.29077770130763, "grad_norm": 18.200519561767578, "learning_rate": 9.854611149346182e-06, "loss": 0.6829, "step": 466650 }, { "epoch": 80.29938059187887, "grad_norm": 16.890708923339844, "learning_rate": 9.850309704060565e-06, "loss": 0.6631, "step": 466700 }, { "epoch": 80.3079834824501, "grad_norm": 16.546457290649414, "learning_rate": 9.846008258774949e-06, "loss": 0.638, "step": 466750 }, { "epoch": 80.31658637302134, "grad_norm": 15.815997123718262, "learning_rate": 9.841706813489334e-06, "loss": 0.6811, "step": 466800 }, { "epoch": 80.32518926359256, "grad_norm": 23.08783531188965, "learning_rate": 9.837405368203717e-06, "loss": 0.6969, "step": 466850 }, { "epoch": 80.3337921541638, "grad_norm": 16.05422019958496, "learning_rate": 9.833103922918101e-06, "loss": 0.6901, "step": 466900 }, { "epoch": 80.34239504473503, "grad_norm": 17.309085845947266, "learning_rate": 9.828802477632484e-06, "loss": 0.6692, "step": 466950 }, { "epoch": 80.35099793530627, "grad_norm": 13.448179244995117, "learning_rate": 9.82450103234687e-06, "loss": 0.6088, "step": 467000 }, { "epoch": 80.3596008258775, "grad_norm": 11.753473281860352, "learning_rate": 9.820199587061253e-06, "loss": 0.6332, "step": 467050 }, { "epoch": 80.36820371644873, "grad_norm": 24.986888885498047, "learning_rate": 9.815898141775638e-06, "loss": 0.7172, "step": 467100 }, { "epoch": 80.37680660701996, "grad_norm": 15.612417221069336, "learning_rate": 9.81159669649002e-06, "loss": 0.6237, "step": 467150 }, { "epoch": 80.3854094975912, "grad_norm": 16.729887008666992, "learning_rate": 9.807295251204405e-06, "loss": 0.6704, "step": 467200 }, { "epoch": 80.39401238816242, "grad_norm": 14.431591987609863, "learning_rate": 9.80299380591879e-06, "loss": 0.6755, "step": 467250 }, { "epoch": 80.40261527873365, "grad_norm": 7.937480926513672, "learning_rate": 9.798692360633172e-06, "loss": 0.6305, "step": 467300 }, { "epoch": 80.41121816930489, "grad_norm": 9.474921226501465, "learning_rate": 9.794390915347557e-06, "loss": 0.6356, "step": 467350 }, { "epoch": 80.41982105987611, "grad_norm": 20.000900268554688, "learning_rate": 9.790089470061942e-06, "loss": 0.6086, "step": 467400 }, { "epoch": 80.42842395044735, "grad_norm": 20.742855072021484, "learning_rate": 9.785788024776326e-06, "loss": 0.6864, "step": 467450 }, { "epoch": 80.43702684101858, "grad_norm": 18.332544326782227, "learning_rate": 9.781486579490709e-06, "loss": 0.723, "step": 467500 }, { "epoch": 80.44562973158982, "grad_norm": 13.58647632598877, "learning_rate": 9.777185134205092e-06, "loss": 0.6474, "step": 467550 }, { "epoch": 80.45423262216104, "grad_norm": 21.663732528686523, "learning_rate": 9.772883688919478e-06, "loss": 0.6532, "step": 467600 }, { "epoch": 80.46283551273228, "grad_norm": 19.841487884521484, "learning_rate": 9.768582243633861e-06, "loss": 0.7324, "step": 467650 }, { "epoch": 80.47143840330351, "grad_norm": 18.674819946289062, "learning_rate": 9.764280798348246e-06, "loss": 0.6749, "step": 467700 }, { "epoch": 80.48004129387475, "grad_norm": 14.071016311645508, "learning_rate": 9.75997935306263e-06, "loss": 0.6441, "step": 467750 }, { "epoch": 80.48864418444597, "grad_norm": 17.760717391967773, "learning_rate": 9.755677907777013e-06, "loss": 0.7375, "step": 467800 }, { "epoch": 80.4972470750172, "grad_norm": 18.493467330932617, "learning_rate": 9.751376462491398e-06, "loss": 0.6828, "step": 467850 }, { "epoch": 80.50584996558844, "grad_norm": 11.320533752441406, "learning_rate": 9.74707501720578e-06, "loss": 0.6502, "step": 467900 }, { "epoch": 80.51445285615966, "grad_norm": 16.77817726135254, "learning_rate": 9.742773571920167e-06, "loss": 0.6597, "step": 467950 }, { "epoch": 80.5230557467309, "grad_norm": 20.24378204345703, "learning_rate": 9.73847212663455e-06, "loss": 0.7369, "step": 468000 }, { "epoch": 80.53165863730213, "grad_norm": 19.367494583129883, "learning_rate": 9.734170681348934e-06, "loss": 0.6841, "step": 468050 }, { "epoch": 80.54026152787337, "grad_norm": 10.611059188842773, "learning_rate": 9.729869236063317e-06, "loss": 0.6663, "step": 468100 }, { "epoch": 80.5488644184446, "grad_norm": 13.720739364624023, "learning_rate": 9.725567790777702e-06, "loss": 0.6724, "step": 468150 }, { "epoch": 80.55746730901583, "grad_norm": 12.265189170837402, "learning_rate": 9.721266345492086e-06, "loss": 0.6673, "step": 468200 }, { "epoch": 80.56607019958706, "grad_norm": 16.44211196899414, "learning_rate": 9.716964900206469e-06, "loss": 0.676, "step": 468250 }, { "epoch": 80.5746730901583, "grad_norm": 18.15322494506836, "learning_rate": 9.712663454920854e-06, "loss": 0.611, "step": 468300 }, { "epoch": 80.58327598072952, "grad_norm": 17.662357330322266, "learning_rate": 9.708362009635238e-06, "loss": 0.7311, "step": 468350 }, { "epoch": 80.59187887130075, "grad_norm": 15.227760314941406, "learning_rate": 9.704060564349623e-06, "loss": 0.6261, "step": 468400 }, { "epoch": 80.60048176187199, "grad_norm": 14.066862106323242, "learning_rate": 9.699759119064006e-06, "loss": 0.648, "step": 468450 }, { "epoch": 80.60908465244322, "grad_norm": 14.252157211303711, "learning_rate": 9.69545767377839e-06, "loss": 0.6894, "step": 468500 }, { "epoch": 80.61768754301445, "grad_norm": 16.056623458862305, "learning_rate": 9.691156228492775e-06, "loss": 0.6972, "step": 468550 }, { "epoch": 80.62629043358568, "grad_norm": 26.7744083404541, "learning_rate": 9.686854783207158e-06, "loss": 0.6341, "step": 468600 }, { "epoch": 80.63489332415692, "grad_norm": 11.686594009399414, "learning_rate": 9.682553337921542e-06, "loss": 0.5973, "step": 468650 }, { "epoch": 80.64349621472815, "grad_norm": 15.876858711242676, "learning_rate": 9.678251892635925e-06, "loss": 0.63, "step": 468700 }, { "epoch": 80.65209910529938, "grad_norm": 17.280641555786133, "learning_rate": 9.673950447350311e-06, "loss": 0.715, "step": 468750 }, { "epoch": 80.66070199587061, "grad_norm": 10.541476249694824, "learning_rate": 9.669649002064694e-06, "loss": 0.708, "step": 468800 }, { "epoch": 80.66930488644185, "grad_norm": 7.832374095916748, "learning_rate": 9.665347556779079e-06, "loss": 0.6899, "step": 468850 }, { "epoch": 80.67790777701308, "grad_norm": 16.78809928894043, "learning_rate": 9.661046111493461e-06, "loss": 0.6558, "step": 468900 }, { "epoch": 80.68651066758432, "grad_norm": 11.42977237701416, "learning_rate": 9.656744666207846e-06, "loss": 0.667, "step": 468950 }, { "epoch": 80.69511355815554, "grad_norm": 13.533136367797852, "learning_rate": 9.65244322092223e-06, "loss": 0.606, "step": 469000 }, { "epoch": 80.70371644872677, "grad_norm": 22.416168212890625, "learning_rate": 9.648141775636613e-06, "loss": 0.6454, "step": 469050 }, { "epoch": 80.712319339298, "grad_norm": 14.088825225830078, "learning_rate": 9.643840330350998e-06, "loss": 0.6822, "step": 469100 }, { "epoch": 80.72092222986923, "grad_norm": 11.689194679260254, "learning_rate": 9.639538885065383e-06, "loss": 0.6828, "step": 469150 }, { "epoch": 80.72952512044047, "grad_norm": 15.405981063842773, "learning_rate": 9.635237439779767e-06, "loss": 0.7039, "step": 469200 }, { "epoch": 80.7381280110117, "grad_norm": 10.764058113098145, "learning_rate": 9.63093599449415e-06, "loss": 0.7126, "step": 469250 }, { "epoch": 80.74673090158294, "grad_norm": 17.496356964111328, "learning_rate": 9.626634549208535e-06, "loss": 0.7726, "step": 469300 }, { "epoch": 80.75533379215416, "grad_norm": 11.913558006286621, "learning_rate": 9.622333103922919e-06, "loss": 0.6604, "step": 469350 }, { "epoch": 80.7639366827254, "grad_norm": 17.722047805786133, "learning_rate": 9.618031658637302e-06, "loss": 0.6863, "step": 469400 }, { "epoch": 80.77253957329663, "grad_norm": 20.314388275146484, "learning_rate": 9.613730213351687e-06, "loss": 0.7159, "step": 469450 }, { "epoch": 80.78114246386787, "grad_norm": 7.554251670837402, "learning_rate": 9.609428768066071e-06, "loss": 0.7244, "step": 469500 }, { "epoch": 80.78974535443909, "grad_norm": 12.826866149902344, "learning_rate": 9.605127322780456e-06, "loss": 0.6789, "step": 469550 }, { "epoch": 80.79834824501032, "grad_norm": 17.603239059448242, "learning_rate": 9.600825877494839e-06, "loss": 0.6496, "step": 469600 }, { "epoch": 80.80695113558156, "grad_norm": 15.496082305908203, "learning_rate": 9.596524432209223e-06, "loss": 0.6809, "step": 469650 }, { "epoch": 80.81555402615278, "grad_norm": 13.349013328552246, "learning_rate": 9.592222986923608e-06, "loss": 0.6123, "step": 469700 }, { "epoch": 80.82415691672402, "grad_norm": 14.32017707824707, "learning_rate": 9.58792154163799e-06, "loss": 0.6655, "step": 469750 }, { "epoch": 80.83275980729525, "grad_norm": 27.19512939453125, "learning_rate": 9.583620096352375e-06, "loss": 0.6967, "step": 469800 }, { "epoch": 80.84136269786649, "grad_norm": 22.008560180664062, "learning_rate": 9.579318651066758e-06, "loss": 0.7083, "step": 469850 }, { "epoch": 80.84996558843771, "grad_norm": 11.193188667297363, "learning_rate": 9.575017205781144e-06, "loss": 0.6408, "step": 469900 }, { "epoch": 80.85856847900895, "grad_norm": 21.006847381591797, "learning_rate": 9.570715760495527e-06, "loss": 0.7499, "step": 469950 }, { "epoch": 80.86717136958018, "grad_norm": 19.120180130004883, "learning_rate": 9.566414315209912e-06, "loss": 0.673, "step": 470000 }, { "epoch": 80.87577426015142, "grad_norm": 22.50079345703125, "learning_rate": 9.562112869924295e-06, "loss": 0.6822, "step": 470050 }, { "epoch": 80.88437715072264, "grad_norm": 12.455565452575684, "learning_rate": 9.557811424638679e-06, "loss": 0.6079, "step": 470100 }, { "epoch": 80.89298004129388, "grad_norm": 12.806296348571777, "learning_rate": 9.553509979353064e-06, "loss": 0.7049, "step": 470150 }, { "epoch": 80.90158293186511, "grad_norm": 8.215085983276367, "learning_rate": 9.549208534067447e-06, "loss": 0.6547, "step": 470200 }, { "epoch": 80.91018582243633, "grad_norm": 19.0174503326416, "learning_rate": 9.544907088781831e-06, "loss": 0.6832, "step": 470250 }, { "epoch": 80.91878871300757, "grad_norm": 15.564010620117188, "learning_rate": 9.540605643496216e-06, "loss": 0.6269, "step": 470300 }, { "epoch": 80.9273916035788, "grad_norm": 9.918898582458496, "learning_rate": 9.5363041982106e-06, "loss": 0.6714, "step": 470350 }, { "epoch": 80.93599449415004, "grad_norm": 10.589137077331543, "learning_rate": 9.532002752924983e-06, "loss": 0.6355, "step": 470400 }, { "epoch": 80.94459738472126, "grad_norm": 15.885931015014648, "learning_rate": 9.527701307639366e-06, "loss": 0.7124, "step": 470450 }, { "epoch": 80.9532002752925, "grad_norm": 18.591121673583984, "learning_rate": 9.523399862353752e-06, "loss": 0.6567, "step": 470500 }, { "epoch": 80.96180316586373, "grad_norm": 12.66530990600586, "learning_rate": 9.519098417068135e-06, "loss": 0.6959, "step": 470550 }, { "epoch": 80.97040605643497, "grad_norm": 19.14951515197754, "learning_rate": 9.51479697178252e-06, "loss": 0.6664, "step": 470600 }, { "epoch": 80.9790089470062, "grad_norm": 23.580148696899414, "learning_rate": 9.510495526496902e-06, "loss": 0.5994, "step": 470650 }, { "epoch": 80.98761183757743, "grad_norm": 11.925045013427734, "learning_rate": 9.506194081211289e-06, "loss": 0.6512, "step": 470700 }, { "epoch": 80.99621472814866, "grad_norm": 19.974876403808594, "learning_rate": 9.501892635925672e-06, "loss": 0.6304, "step": 470750 }, { "epoch": 81.0, "eval_accuracy": 0.5359662708655997, "eval_f1": 0.5114873816320975, "eval_f1_DuraRiadoRio_16x16": 0.5308661172445965, "eval_f1_Mole_16x16": 0.5920691009030231, "eval_f1_Quebrado_16x16": 0.7058139534883721, "eval_f1_RiadoRio_16x16": 0.2733333333333333, "eval_f1_RioFechado_16x16": 0.45535440319116294, "eval_loss": 2.7415475845336914, "eval_precision": 0.5735361554937818, "eval_precision_DuraRiadoRio_16x16": 0.4136176399321541, "eval_precision_Mole_16x16": 0.5405017921146953, "eval_precision_Quebrado_16x16": 0.6376050420168067, "eval_precision_RiadoRio_16x16": 0.5195852534562212, "eval_precision_RioFechado_16x16": 0.7563710499490316, "eval_recall": 0.5393864574829671, "eval_recall_DuraRiadoRio_16x16": 0.7408854166666666, "eval_recall_Mole_16x16": 0.6545138888888888, "eval_recall_Quebrado_16x16": 0.7903645833333334, "eval_recall_RiadoRio_16x16": 0.18544407894736842, "eval_recall_RioFechado_16x16": 0.3257243195785777, "eval_runtime": 46.852, "eval_samples_per_second": 248.058, "eval_steps_per_second": 15.517, "step": 470772 }, { "epoch": 81.00481761871988, "grad_norm": 17.546926498413086, "learning_rate": 9.497591190640054e-06, "loss": 0.7043, "step": 470800 }, { "epoch": 81.01342050929112, "grad_norm": 10.922755241394043, "learning_rate": 9.493289745354439e-06, "loss": 0.6627, "step": 470850 }, { "epoch": 81.02202339986235, "grad_norm": 12.395018577575684, "learning_rate": 9.488988300068824e-06, "loss": 0.6729, "step": 470900 }, { "epoch": 81.03062629043359, "grad_norm": 6.468953609466553, "learning_rate": 9.484686854783208e-06, "loss": 0.6422, "step": 470950 }, { "epoch": 81.03922918100481, "grad_norm": 11.394336700439453, "learning_rate": 9.480385409497591e-06, "loss": 0.6165, "step": 471000 }, { "epoch": 81.04783207157605, "grad_norm": 9.25304889678955, "learning_rate": 9.476083964211976e-06, "loss": 0.6611, "step": 471050 }, { "epoch": 81.05643496214728, "grad_norm": 20.636035919189453, "learning_rate": 9.47178251892636e-06, "loss": 0.6128, "step": 471100 }, { "epoch": 81.06503785271852, "grad_norm": 11.584969520568848, "learning_rate": 9.467481073640743e-06, "loss": 0.6245, "step": 471150 }, { "epoch": 81.07364074328974, "grad_norm": 15.78249454498291, "learning_rate": 9.463179628355128e-06, "loss": 0.6913, "step": 471200 }, { "epoch": 81.08224363386098, "grad_norm": 17.349002838134766, "learning_rate": 9.45887818306951e-06, "loss": 0.7269, "step": 471250 }, { "epoch": 81.09084652443221, "grad_norm": 17.560768127441406, "learning_rate": 9.454576737783897e-06, "loss": 0.6498, "step": 471300 }, { "epoch": 81.09944941500343, "grad_norm": 19.476350784301758, "learning_rate": 9.45027529249828e-06, "loss": 0.6253, "step": 471350 }, { "epoch": 81.10805230557467, "grad_norm": 17.86638641357422, "learning_rate": 9.445973847212664e-06, "loss": 0.7034, "step": 471400 }, { "epoch": 81.1166551961459, "grad_norm": 12.08938217163086, "learning_rate": 9.441672401927049e-06, "loss": 0.7049, "step": 471450 }, { "epoch": 81.12525808671714, "grad_norm": 7.91759729385376, "learning_rate": 9.437370956641432e-06, "loss": 0.6426, "step": 471500 }, { "epoch": 81.13386097728836, "grad_norm": 19.744050979614258, "learning_rate": 9.433069511355816e-06, "loss": 0.6685, "step": 471550 }, { "epoch": 81.1424638678596, "grad_norm": 17.701248168945312, "learning_rate": 9.428768066070199e-06, "loss": 0.6918, "step": 471600 }, { "epoch": 81.15106675843083, "grad_norm": 18.323101043701172, "learning_rate": 9.424466620784585e-06, "loss": 0.6897, "step": 471650 }, { "epoch": 81.15966964900207, "grad_norm": 16.411340713500977, "learning_rate": 9.420165175498968e-06, "loss": 0.6929, "step": 471700 }, { "epoch": 81.1682725395733, "grad_norm": 14.564033508300781, "learning_rate": 9.415863730213353e-06, "loss": 0.6803, "step": 471750 }, { "epoch": 81.17687543014453, "grad_norm": 12.943014144897461, "learning_rate": 9.411562284927735e-06, "loss": 0.6294, "step": 471800 }, { "epoch": 81.18547832071576, "grad_norm": 12.176604270935059, "learning_rate": 9.40726083964212e-06, "loss": 0.6674, "step": 471850 }, { "epoch": 81.194081211287, "grad_norm": 14.444488525390625, "learning_rate": 9.402959394356505e-06, "loss": 0.6599, "step": 471900 }, { "epoch": 81.20268410185822, "grad_norm": 17.80763053894043, "learning_rate": 9.398657949070887e-06, "loss": 0.6593, "step": 471950 }, { "epoch": 81.21128699242945, "grad_norm": 10.04764461517334, "learning_rate": 9.394356503785272e-06, "loss": 0.7232, "step": 472000 }, { "epoch": 81.21988988300069, "grad_norm": 24.697921752929688, "learning_rate": 9.390055058499657e-06, "loss": 0.6317, "step": 472050 }, { "epoch": 81.22849277357192, "grad_norm": 19.60358428955078, "learning_rate": 9.385753613214041e-06, "loss": 0.6598, "step": 472100 }, { "epoch": 81.23709566414315, "grad_norm": 18.877582550048828, "learning_rate": 9.381452167928424e-06, "loss": 0.7071, "step": 472150 }, { "epoch": 81.24569855471438, "grad_norm": 11.617258071899414, "learning_rate": 9.377150722642809e-06, "loss": 0.5763, "step": 472200 }, { "epoch": 81.25430144528562, "grad_norm": 12.028502464294434, "learning_rate": 9.372849277357193e-06, "loss": 0.6651, "step": 472250 }, { "epoch": 81.26290433585685, "grad_norm": 16.91242218017578, "learning_rate": 9.368547832071576e-06, "loss": 0.637, "step": 472300 }, { "epoch": 81.27150722642808, "grad_norm": 19.64677619934082, "learning_rate": 9.36424638678596e-06, "loss": 0.649, "step": 472350 }, { "epoch": 81.28011011699931, "grad_norm": 8.986295700073242, "learning_rate": 9.359944941500343e-06, "loss": 0.6893, "step": 472400 }, { "epoch": 81.28871300757055, "grad_norm": 16.62254524230957, "learning_rate": 9.35564349621473e-06, "loss": 0.6642, "step": 472450 }, { "epoch": 81.29731589814178, "grad_norm": 14.998811721801758, "learning_rate": 9.351342050929113e-06, "loss": 0.6756, "step": 472500 }, { "epoch": 81.305918788713, "grad_norm": 14.873452186584473, "learning_rate": 9.347040605643497e-06, "loss": 0.7064, "step": 472550 }, { "epoch": 81.31452167928424, "grad_norm": 15.272867202758789, "learning_rate": 9.34273916035788e-06, "loss": 0.6547, "step": 472600 }, { "epoch": 81.32312456985547, "grad_norm": 18.037782669067383, "learning_rate": 9.338437715072265e-06, "loss": 0.6691, "step": 472650 }, { "epoch": 81.3317274604267, "grad_norm": 18.310930252075195, "learning_rate": 9.334136269786649e-06, "loss": 0.6628, "step": 472700 }, { "epoch": 81.34033035099793, "grad_norm": 17.117326736450195, "learning_rate": 9.329834824501032e-06, "loss": 0.6803, "step": 472750 }, { "epoch": 81.34893324156917, "grad_norm": 18.561668395996094, "learning_rate": 9.325533379215417e-06, "loss": 0.7378, "step": 472800 }, { "epoch": 81.3575361321404, "grad_norm": 14.33683967590332, "learning_rate": 9.321231933929801e-06, "loss": 0.6977, "step": 472850 }, { "epoch": 81.36613902271164, "grad_norm": 17.014572143554688, "learning_rate": 9.316930488644186e-06, "loss": 0.6614, "step": 472900 }, { "epoch": 81.37474191328286, "grad_norm": 11.760478019714355, "learning_rate": 9.312629043358569e-06, "loss": 0.6536, "step": 472950 }, { "epoch": 81.3833448038541, "grad_norm": 12.58326530456543, "learning_rate": 9.308327598072953e-06, "loss": 0.6236, "step": 473000 }, { "epoch": 81.39194769442533, "grad_norm": 19.28775405883789, "learning_rate": 9.304026152787338e-06, "loss": 0.6553, "step": 473050 }, { "epoch": 81.40055058499657, "grad_norm": 14.226618766784668, "learning_rate": 9.29972470750172e-06, "loss": 0.6825, "step": 473100 }, { "epoch": 81.40915347556779, "grad_norm": 12.909225463867188, "learning_rate": 9.295423262216105e-06, "loss": 0.6733, "step": 473150 }, { "epoch": 81.41775636613902, "grad_norm": 20.30777931213379, "learning_rate": 9.29112181693049e-06, "loss": 0.6134, "step": 473200 }, { "epoch": 81.42635925671026, "grad_norm": 15.912545204162598, "learning_rate": 9.286820371644874e-06, "loss": 0.6314, "step": 473250 }, { "epoch": 81.43496214728148, "grad_norm": 16.211143493652344, "learning_rate": 9.282518926359257e-06, "loss": 0.6584, "step": 473300 }, { "epoch": 81.44356503785272, "grad_norm": 12.073599815368652, "learning_rate": 9.278217481073642e-06, "loss": 0.6288, "step": 473350 }, { "epoch": 81.45216792842395, "grad_norm": 20.335159301757812, "learning_rate": 9.273916035788026e-06, "loss": 0.5931, "step": 473400 }, { "epoch": 81.46077081899519, "grad_norm": 15.623811721801758, "learning_rate": 9.269614590502409e-06, "loss": 0.66, "step": 473450 }, { "epoch": 81.46937370956641, "grad_norm": 20.698509216308594, "learning_rate": 9.265313145216794e-06, "loss": 0.6582, "step": 473500 }, { "epoch": 81.47797660013765, "grad_norm": 16.715360641479492, "learning_rate": 9.261011699931176e-06, "loss": 0.74, "step": 473550 }, { "epoch": 81.48657949070888, "grad_norm": 13.908540725708008, "learning_rate": 9.256710254645563e-06, "loss": 0.7182, "step": 473600 }, { "epoch": 81.49518238128012, "grad_norm": 18.306352615356445, "learning_rate": 9.252408809359946e-06, "loss": 0.7064, "step": 473650 }, { "epoch": 81.50378527185134, "grad_norm": 14.455116271972656, "learning_rate": 9.24810736407433e-06, "loss": 0.6106, "step": 473700 }, { "epoch": 81.51238816242257, "grad_norm": 17.31832504272461, "learning_rate": 9.243805918788713e-06, "loss": 0.6363, "step": 473750 }, { "epoch": 81.5209910529938, "grad_norm": 25.91803741455078, "learning_rate": 9.239504473503098e-06, "loss": 0.686, "step": 473800 }, { "epoch": 81.52959394356503, "grad_norm": 9.495428085327148, "learning_rate": 9.235203028217482e-06, "loss": 0.7032, "step": 473850 }, { "epoch": 81.53819683413627, "grad_norm": 14.476359367370605, "learning_rate": 9.230901582931865e-06, "loss": 0.6875, "step": 473900 }, { "epoch": 81.5467997247075, "grad_norm": 13.225726127624512, "learning_rate": 9.22660013764625e-06, "loss": 0.6131, "step": 473950 }, { "epoch": 81.55540261527874, "grad_norm": 14.520237922668457, "learning_rate": 9.222298692360634e-06, "loss": 0.6356, "step": 474000 }, { "epoch": 81.56400550584996, "grad_norm": 21.87996482849121, "learning_rate": 9.217997247075017e-06, "loss": 0.6572, "step": 474050 }, { "epoch": 81.5726083964212, "grad_norm": 11.036274909973145, "learning_rate": 9.213695801789402e-06, "loss": 0.6819, "step": 474100 }, { "epoch": 81.58121128699243, "grad_norm": 15.63317584991455, "learning_rate": 9.209394356503784e-06, "loss": 0.6701, "step": 474150 }, { "epoch": 81.58981417756367, "grad_norm": 15.185318946838379, "learning_rate": 9.20509291121817e-06, "loss": 0.6403, "step": 474200 }, { "epoch": 81.59841706813489, "grad_norm": 21.605327606201172, "learning_rate": 9.200791465932554e-06, "loss": 0.6403, "step": 474250 }, { "epoch": 81.60701995870612, "grad_norm": 16.788679122924805, "learning_rate": 9.196490020646938e-06, "loss": 0.6965, "step": 474300 }, { "epoch": 81.61562284927736, "grad_norm": 11.038148880004883, "learning_rate": 9.192188575361321e-06, "loss": 0.6307, "step": 474350 }, { "epoch": 81.62422573984858, "grad_norm": 23.337446212768555, "learning_rate": 9.187887130075706e-06, "loss": 0.6986, "step": 474400 }, { "epoch": 81.63282863041982, "grad_norm": 10.120402336120605, "learning_rate": 9.18358568479009e-06, "loss": 0.6401, "step": 474450 }, { "epoch": 81.64143152099105, "grad_norm": 22.965238571166992, "learning_rate": 9.179284239504473e-06, "loss": 0.6828, "step": 474500 }, { "epoch": 81.65003441156229, "grad_norm": 17.072710037231445, "learning_rate": 9.174982794218858e-06, "loss": 0.6615, "step": 474550 }, { "epoch": 81.65863730213351, "grad_norm": 8.863936424255371, "learning_rate": 9.170681348933242e-06, "loss": 0.624, "step": 474600 }, { "epoch": 81.66724019270475, "grad_norm": 19.608997344970703, "learning_rate": 9.166379903647627e-06, "loss": 0.6885, "step": 474650 }, { "epoch": 81.67584308327598, "grad_norm": 12.217105865478516, "learning_rate": 9.16207845836201e-06, "loss": 0.6904, "step": 474700 }, { "epoch": 81.68444597384722, "grad_norm": 19.335084915161133, "learning_rate": 9.157777013076394e-06, "loss": 0.6922, "step": 474750 }, { "epoch": 81.69304886441844, "grad_norm": 19.380447387695312, "learning_rate": 9.153475567790779e-06, "loss": 0.6452, "step": 474800 }, { "epoch": 81.70165175498968, "grad_norm": 16.239055633544922, "learning_rate": 9.149174122505161e-06, "loss": 0.688, "step": 474850 }, { "epoch": 81.71025464556091, "grad_norm": 21.607465744018555, "learning_rate": 9.144872677219546e-06, "loss": 0.659, "step": 474900 }, { "epoch": 81.71885753613213, "grad_norm": 21.686691284179688, "learning_rate": 9.140571231933929e-06, "loss": 0.6346, "step": 474950 }, { "epoch": 81.72746042670337, "grad_norm": 14.299429893493652, "learning_rate": 9.136269786648315e-06, "loss": 0.637, "step": 475000 }, { "epoch": 81.7360633172746, "grad_norm": 13.319925308227539, "learning_rate": 9.131968341362698e-06, "loss": 0.7268, "step": 475050 }, { "epoch": 81.74466620784584, "grad_norm": 14.508722305297852, "learning_rate": 9.127666896077083e-06, "loss": 0.6846, "step": 475100 }, { "epoch": 81.75326909841706, "grad_norm": 29.390037536621094, "learning_rate": 9.123365450791467e-06, "loss": 0.6776, "step": 475150 }, { "epoch": 81.7618719889883, "grad_norm": 13.790307998657227, "learning_rate": 9.11906400550585e-06, "loss": 0.7139, "step": 475200 }, { "epoch": 81.77047487955953, "grad_norm": 18.03924560546875, "learning_rate": 9.114762560220235e-06, "loss": 0.705, "step": 475250 }, { "epoch": 81.77907777013077, "grad_norm": 18.172252655029297, "learning_rate": 9.110461114934617e-06, "loss": 0.6652, "step": 475300 }, { "epoch": 81.787680660702, "grad_norm": 10.736970901489258, "learning_rate": 9.106159669649004e-06, "loss": 0.6901, "step": 475350 }, { "epoch": 81.79628355127323, "grad_norm": 12.127830505371094, "learning_rate": 9.101858224363387e-06, "loss": 0.6572, "step": 475400 }, { "epoch": 81.80488644184446, "grad_norm": 13.259041786193848, "learning_rate": 9.097556779077771e-06, "loss": 0.6573, "step": 475450 }, { "epoch": 81.81348933241568, "grad_norm": 20.12798500061035, "learning_rate": 9.093255333792154e-06, "loss": 0.6462, "step": 475500 }, { "epoch": 81.82209222298692, "grad_norm": 10.155364036560059, "learning_rate": 9.088953888506539e-06, "loss": 0.6981, "step": 475550 }, { "epoch": 81.83069511355815, "grad_norm": 12.204315185546875, "learning_rate": 9.084652443220923e-06, "loss": 0.6841, "step": 475600 }, { "epoch": 81.83929800412939, "grad_norm": 9.945016860961914, "learning_rate": 9.080350997935306e-06, "loss": 0.654, "step": 475650 }, { "epoch": 81.84790089470062, "grad_norm": 24.274797439575195, "learning_rate": 9.07604955264969e-06, "loss": 0.6777, "step": 475700 }, { "epoch": 81.85650378527185, "grad_norm": 19.070938110351562, "learning_rate": 9.071748107364075e-06, "loss": 0.6402, "step": 475750 }, { "epoch": 81.86510667584308, "grad_norm": 19.86464500427246, "learning_rate": 9.06744666207846e-06, "loss": 0.7512, "step": 475800 }, { "epoch": 81.87370956641432, "grad_norm": 9.555815696716309, "learning_rate": 9.063145216792843e-06, "loss": 0.7188, "step": 475850 }, { "epoch": 81.88231245698555, "grad_norm": 15.592046737670898, "learning_rate": 9.058843771507227e-06, "loss": 0.6496, "step": 475900 }, { "epoch": 81.89091534755678, "grad_norm": 9.904273986816406, "learning_rate": 9.054542326221612e-06, "loss": 0.6531, "step": 475950 }, { "epoch": 81.89951823812801, "grad_norm": 12.209676742553711, "learning_rate": 9.050240880935995e-06, "loss": 0.6462, "step": 476000 }, { "epoch": 81.90812112869925, "grad_norm": 9.039649963378906, "learning_rate": 9.045939435650379e-06, "loss": 0.6921, "step": 476050 }, { "epoch": 81.91672401927048, "grad_norm": 12.150640487670898, "learning_rate": 9.041637990364762e-06, "loss": 0.6838, "step": 476100 }, { "epoch": 81.9253269098417, "grad_norm": 17.720125198364258, "learning_rate": 9.037336545079148e-06, "loss": 0.6753, "step": 476150 }, { "epoch": 81.93392980041294, "grad_norm": 24.798002243041992, "learning_rate": 9.033035099793531e-06, "loss": 0.6522, "step": 476200 }, { "epoch": 81.94253269098417, "grad_norm": 9.57580280303955, "learning_rate": 9.028733654507916e-06, "loss": 0.6785, "step": 476250 }, { "epoch": 81.9511355815554, "grad_norm": 10.227683067321777, "learning_rate": 9.024432209222298e-06, "loss": 0.617, "step": 476300 }, { "epoch": 81.95973847212663, "grad_norm": 24.39933204650879, "learning_rate": 9.020130763936683e-06, "loss": 0.6827, "step": 476350 }, { "epoch": 81.96834136269787, "grad_norm": 12.578754425048828, "learning_rate": 9.015829318651068e-06, "loss": 0.6849, "step": 476400 }, { "epoch": 81.9769442532691, "grad_norm": 19.488887786865234, "learning_rate": 9.01152787336545e-06, "loss": 0.6758, "step": 476450 }, { "epoch": 81.98554714384034, "grad_norm": 11.659625053405762, "learning_rate": 9.007226428079835e-06, "loss": 0.6162, "step": 476500 }, { "epoch": 81.99415003441156, "grad_norm": 12.769347190856934, "learning_rate": 9.00292498279422e-06, "loss": 0.7155, "step": 476550 }, { "epoch": 82.0, "eval_accuracy": 0.5313199105145414, "eval_f1": 0.5323353260550918, "eval_f1_DuraRiadoRio_16x16": 0.4875977653631285, "eval_f1_Mole_16x16": 0.47478406241292836, "eval_f1_Quebrado_16x16": 0.693100358422939, "eval_f1_RiadoRio_16x16": 0.43339382940108895, "eval_f1_RioFechado_16x16": 0.5728006146753746, "eval_loss": 2.7196762561798096, "eval_precision": 0.5557815318805647, "eval_precision_DuraRiadoRio_16x16": 0.5025333947489636, "eval_precision_Mole_16x16": 0.6630350194552529, "eval_precision_Quebrado_16x16": 0.7162037037037037, "eval_precision_RiadoRio_16x16": 0.38791423001949316, "eval_precision_RioFechado_16x16": 0.5092213114754098, "eval_recall": 0.5320464803818884, "eval_recall_DuraRiadoRio_16x16": 0.4735243055555556, "eval_recall_Mole_16x16": 0.3697916666666667, "eval_recall_Quebrado_16x16": 0.6714409722222222, "eval_recall_RiadoRio_16x16": 0.4909539473684211, "eval_recall_RioFechado_16x16": 0.654521510096576, "eval_runtime": 46.8294, "eval_samples_per_second": 248.178, "eval_steps_per_second": 15.524, "step": 476584 }, { "epoch": 82.0027529249828, "grad_norm": 13.428080558776855, "learning_rate": 8.998623537508604e-06, "loss": 0.6569, "step": 476600 }, { "epoch": 82.01135581555403, "grad_norm": 20.811710357666016, "learning_rate": 8.994322092222987e-06, "loss": 0.7088, "step": 476650 }, { "epoch": 82.01995870612525, "grad_norm": 16.163415908813477, "learning_rate": 8.99002064693737e-06, "loss": 0.6825, "step": 476700 }, { "epoch": 82.02856159669649, "grad_norm": 12.52689266204834, "learning_rate": 8.985719201651756e-06, "loss": 0.6502, "step": 476750 }, { "epoch": 82.03716448726772, "grad_norm": 11.276086807250977, "learning_rate": 8.981417756366139e-06, "loss": 0.5889, "step": 476800 }, { "epoch": 82.04576737783896, "grad_norm": 23.27578353881836, "learning_rate": 8.977116311080524e-06, "loss": 0.6518, "step": 476850 }, { "epoch": 82.05437026841018, "grad_norm": 13.345627784729004, "learning_rate": 8.972814865794908e-06, "loss": 0.627, "step": 476900 }, { "epoch": 82.06297315898142, "grad_norm": 21.005704879760742, "learning_rate": 8.968513420509293e-06, "loss": 0.6262, "step": 476950 }, { "epoch": 82.07157604955265, "grad_norm": 18.226638793945312, "learning_rate": 8.964211975223676e-06, "loss": 0.6784, "step": 477000 }, { "epoch": 82.08017894012389, "grad_norm": 10.697501182556152, "learning_rate": 8.959910529938058e-06, "loss": 0.6673, "step": 477050 }, { "epoch": 82.08878183069511, "grad_norm": 16.413705825805664, "learning_rate": 8.955609084652445e-06, "loss": 0.6483, "step": 477100 }, { "epoch": 82.09738472126635, "grad_norm": 13.503079414367676, "learning_rate": 8.951307639366828e-06, "loss": 0.6527, "step": 477150 }, { "epoch": 82.10598761183758, "grad_norm": 13.520648956298828, "learning_rate": 8.947006194081212e-06, "loss": 0.6714, "step": 477200 }, { "epoch": 82.1145905024088, "grad_norm": 13.843546867370605, "learning_rate": 8.942704748795595e-06, "loss": 0.6415, "step": 477250 }, { "epoch": 82.12319339298004, "grad_norm": 7.993147850036621, "learning_rate": 8.938403303509981e-06, "loss": 0.5863, "step": 477300 }, { "epoch": 82.13179628355127, "grad_norm": 15.669707298278809, "learning_rate": 8.934101858224364e-06, "loss": 0.6095, "step": 477350 }, { "epoch": 82.1403991741225, "grad_norm": 11.775097846984863, "learning_rate": 8.929800412938747e-06, "loss": 0.6718, "step": 477400 }, { "epoch": 82.14900206469373, "grad_norm": 24.482967376708984, "learning_rate": 8.925498967653132e-06, "loss": 0.698, "step": 477450 }, { "epoch": 82.15760495526497, "grad_norm": 12.626785278320312, "learning_rate": 8.921197522367516e-06, "loss": 0.6632, "step": 477500 }, { "epoch": 82.1662078458362, "grad_norm": 15.911735534667969, "learning_rate": 8.9168960770819e-06, "loss": 0.652, "step": 477550 }, { "epoch": 82.17481073640744, "grad_norm": 12.605777740478516, "learning_rate": 8.912594631796284e-06, "loss": 0.6416, "step": 477600 }, { "epoch": 82.18341362697866, "grad_norm": 10.564091682434082, "learning_rate": 8.908293186510668e-06, "loss": 0.6971, "step": 477650 }, { "epoch": 82.1920165175499, "grad_norm": 19.594545364379883, "learning_rate": 8.903991741225053e-06, "loss": 0.6861, "step": 477700 }, { "epoch": 82.20061940812113, "grad_norm": 16.825925827026367, "learning_rate": 8.899690295939436e-06, "loss": 0.6822, "step": 477750 }, { "epoch": 82.20922229869237, "grad_norm": 11.840417861938477, "learning_rate": 8.89538885065382e-06, "loss": 0.6485, "step": 477800 }, { "epoch": 82.21782518926359, "grad_norm": 9.1000337600708, "learning_rate": 8.891087405368203e-06, "loss": 0.6312, "step": 477850 }, { "epoch": 82.22642807983482, "grad_norm": 9.237177848815918, "learning_rate": 8.88678596008259e-06, "loss": 0.6599, "step": 477900 }, { "epoch": 82.23503097040606, "grad_norm": 11.40772533416748, "learning_rate": 8.882484514796972e-06, "loss": 0.652, "step": 477950 }, { "epoch": 82.24363386097728, "grad_norm": 20.517379760742188, "learning_rate": 8.878183069511357e-06, "loss": 0.6323, "step": 478000 }, { "epoch": 82.25223675154852, "grad_norm": 9.645004272460938, "learning_rate": 8.87388162422574e-06, "loss": 0.6798, "step": 478050 }, { "epoch": 82.26083964211975, "grad_norm": 13.626047134399414, "learning_rate": 8.869580178940124e-06, "loss": 0.721, "step": 478100 }, { "epoch": 82.26944253269099, "grad_norm": 20.60660743713379, "learning_rate": 8.865278733654509e-06, "loss": 0.6179, "step": 478150 }, { "epoch": 82.27804542326221, "grad_norm": 12.063943862915039, "learning_rate": 8.860977288368891e-06, "loss": 0.6614, "step": 478200 }, { "epoch": 82.28664831383345, "grad_norm": 15.330595970153809, "learning_rate": 8.856675843083276e-06, "loss": 0.6857, "step": 478250 }, { "epoch": 82.29525120440468, "grad_norm": 15.6497163772583, "learning_rate": 8.85237439779766e-06, "loss": 0.6293, "step": 478300 }, { "epoch": 82.30385409497592, "grad_norm": 16.602163314819336, "learning_rate": 8.848072952512045e-06, "loss": 0.6429, "step": 478350 }, { "epoch": 82.31245698554714, "grad_norm": 12.277408599853516, "learning_rate": 8.843771507226428e-06, "loss": 0.6985, "step": 478400 }, { "epoch": 82.32105987611837, "grad_norm": 8.355486869812012, "learning_rate": 8.839470061940813e-06, "loss": 0.6348, "step": 478450 }, { "epoch": 82.32966276668961, "grad_norm": 12.375510215759277, "learning_rate": 8.835168616655197e-06, "loss": 0.5935, "step": 478500 }, { "epoch": 82.33826565726083, "grad_norm": 13.831624031066895, "learning_rate": 8.83086717136958e-06, "loss": 0.6706, "step": 478550 }, { "epoch": 82.34686854783207, "grad_norm": 10.549190521240234, "learning_rate": 8.826565726083965e-06, "loss": 0.6763, "step": 478600 }, { "epoch": 82.3554714384033, "grad_norm": 24.139142990112305, "learning_rate": 8.822264280798347e-06, "loss": 0.734, "step": 478650 }, { "epoch": 82.36407432897454, "grad_norm": 16.44205093383789, "learning_rate": 8.817962835512734e-06, "loss": 0.6422, "step": 478700 }, { "epoch": 82.37267721954576, "grad_norm": 19.163734436035156, "learning_rate": 8.813661390227117e-06, "loss": 0.6748, "step": 478750 }, { "epoch": 82.381280110117, "grad_norm": 14.368083000183105, "learning_rate": 8.809359944941501e-06, "loss": 0.6791, "step": 478800 }, { "epoch": 82.38988300068823, "grad_norm": 21.501415252685547, "learning_rate": 8.805058499655886e-06, "loss": 0.6421, "step": 478850 }, { "epoch": 82.39848589125947, "grad_norm": 18.166345596313477, "learning_rate": 8.800757054370269e-06, "loss": 0.65, "step": 478900 }, { "epoch": 82.4070887818307, "grad_norm": 10.405637741088867, "learning_rate": 8.796455609084653e-06, "loss": 0.6693, "step": 478950 }, { "epoch": 82.41569167240193, "grad_norm": 9.832447052001953, "learning_rate": 8.792154163799036e-06, "loss": 0.7056, "step": 479000 }, { "epoch": 82.42429456297316, "grad_norm": 6.263812065124512, "learning_rate": 8.787852718513422e-06, "loss": 0.6395, "step": 479050 }, { "epoch": 82.43289745354438, "grad_norm": 11.585448265075684, "learning_rate": 8.783551273227805e-06, "loss": 0.7289, "step": 479100 }, { "epoch": 82.44150034411562, "grad_norm": 10.975445747375488, "learning_rate": 8.77924982794219e-06, "loss": 0.6847, "step": 479150 }, { "epoch": 82.45010323468685, "grad_norm": 18.216588973999023, "learning_rate": 8.774948382656573e-06, "loss": 0.6939, "step": 479200 }, { "epoch": 82.45870612525809, "grad_norm": 17.884384155273438, "learning_rate": 8.770646937370957e-06, "loss": 0.6348, "step": 479250 }, { "epoch": 82.46730901582931, "grad_norm": 25.473758697509766, "learning_rate": 8.766345492085342e-06, "loss": 0.7264, "step": 479300 }, { "epoch": 82.47591190640055, "grad_norm": 15.306251525878906, "learning_rate": 8.762044046799724e-06, "loss": 0.6303, "step": 479350 }, { "epoch": 82.48451479697178, "grad_norm": 24.09316635131836, "learning_rate": 8.757742601514109e-06, "loss": 0.684, "step": 479400 }, { "epoch": 82.49311768754302, "grad_norm": 9.776891708374023, "learning_rate": 8.753441156228494e-06, "loss": 0.6198, "step": 479450 }, { "epoch": 82.50172057811425, "grad_norm": 17.642101287841797, "learning_rate": 8.749139710942878e-06, "loss": 0.7027, "step": 479500 }, { "epoch": 82.51032346868548, "grad_norm": 14.682507514953613, "learning_rate": 8.744838265657261e-06, "loss": 0.6868, "step": 479550 }, { "epoch": 82.51892635925671, "grad_norm": 21.218746185302734, "learning_rate": 8.740536820371646e-06, "loss": 0.6219, "step": 479600 }, { "epoch": 82.52752924982794, "grad_norm": 14.650497436523438, "learning_rate": 8.73623537508603e-06, "loss": 0.6466, "step": 479650 }, { "epoch": 82.53613214039918, "grad_norm": 12.346646308898926, "learning_rate": 8.731933929800413e-06, "loss": 0.7029, "step": 479700 }, { "epoch": 82.5447350309704, "grad_norm": 13.318999290466309, "learning_rate": 8.727632484514798e-06, "loss": 0.7294, "step": 479750 }, { "epoch": 82.55333792154164, "grad_norm": 17.21027946472168, "learning_rate": 8.72333103922918e-06, "loss": 0.6473, "step": 479800 }, { "epoch": 82.56194081211287, "grad_norm": 20.12491226196289, "learning_rate": 8.719029593943567e-06, "loss": 0.6947, "step": 479850 }, { "epoch": 82.5705437026841, "grad_norm": 22.537601470947266, "learning_rate": 8.71472814865795e-06, "loss": 0.7072, "step": 479900 }, { "epoch": 82.57914659325533, "grad_norm": 14.618163108825684, "learning_rate": 8.710426703372334e-06, "loss": 0.6564, "step": 479950 }, { "epoch": 82.58774948382657, "grad_norm": 12.282538414001465, "learning_rate": 8.706125258086717e-06, "loss": 0.7257, "step": 480000 }, { "epoch": 82.5963523743978, "grad_norm": 25.69037437438965, "learning_rate": 8.701823812801102e-06, "loss": 0.647, "step": 480050 }, { "epoch": 82.60495526496904, "grad_norm": 16.990032196044922, "learning_rate": 8.697522367515486e-06, "loss": 0.6692, "step": 480100 }, { "epoch": 82.61355815554026, "grad_norm": 10.520605087280273, "learning_rate": 8.693220922229869e-06, "loss": 0.6563, "step": 480150 }, { "epoch": 82.6221610461115, "grad_norm": 14.256732940673828, "learning_rate": 8.688919476944254e-06, "loss": 0.6185, "step": 480200 }, { "epoch": 82.63076393668273, "grad_norm": 17.313154220581055, "learning_rate": 8.684618031658638e-06, "loss": 0.636, "step": 480250 }, { "epoch": 82.63936682725395, "grad_norm": 14.929203987121582, "learning_rate": 8.680316586373021e-06, "loss": 0.6504, "step": 480300 }, { "epoch": 82.64796971782519, "grad_norm": 14.96666431427002, "learning_rate": 8.676015141087406e-06, "loss": 0.6382, "step": 480350 }, { "epoch": 82.65657260839642, "grad_norm": 13.5920991897583, "learning_rate": 8.671713695801788e-06, "loss": 0.6843, "step": 480400 }, { "epoch": 82.66517549896766, "grad_norm": 15.5230712890625, "learning_rate": 8.667412250516175e-06, "loss": 0.6626, "step": 480450 }, { "epoch": 82.67377838953888, "grad_norm": 19.579334259033203, "learning_rate": 8.663110805230558e-06, "loss": 0.6959, "step": 480500 }, { "epoch": 82.68238128011012, "grad_norm": 17.978010177612305, "learning_rate": 8.658809359944942e-06, "loss": 0.6337, "step": 480550 }, { "epoch": 82.69098417068135, "grad_norm": 15.008688926696777, "learning_rate": 8.654507914659325e-06, "loss": 0.6978, "step": 480600 }, { "epoch": 82.69958706125259, "grad_norm": 16.247703552246094, "learning_rate": 8.65020646937371e-06, "loss": 0.688, "step": 480650 }, { "epoch": 82.70818995182381, "grad_norm": 20.688434600830078, "learning_rate": 8.645905024088094e-06, "loss": 0.6531, "step": 480700 }, { "epoch": 82.71679284239505, "grad_norm": 9.382669448852539, "learning_rate": 8.641603578802477e-06, "loss": 0.7017, "step": 480750 }, { "epoch": 82.72539573296628, "grad_norm": 16.745315551757812, "learning_rate": 8.637302133516863e-06, "loss": 0.6816, "step": 480800 }, { "epoch": 82.7339986235375, "grad_norm": 37.076927185058594, "learning_rate": 8.633000688231246e-06, "loss": 0.6855, "step": 480850 }, { "epoch": 82.74260151410874, "grad_norm": 17.709548950195312, "learning_rate": 8.62869924294563e-06, "loss": 0.6574, "step": 480900 }, { "epoch": 82.75120440467997, "grad_norm": 18.283000946044922, "learning_rate": 8.624397797660013e-06, "loss": 0.6274, "step": 480950 }, { "epoch": 82.7598072952512, "grad_norm": 17.592058181762695, "learning_rate": 8.620096352374398e-06, "loss": 0.6296, "step": 481000 }, { "epoch": 82.76841018582243, "grad_norm": 17.37840461730957, "learning_rate": 8.615794907088783e-06, "loss": 0.6558, "step": 481050 }, { "epoch": 82.77701307639367, "grad_norm": 16.782712936401367, "learning_rate": 8.611493461803165e-06, "loss": 0.613, "step": 481100 }, { "epoch": 82.7856159669649, "grad_norm": 12.764786720275879, "learning_rate": 8.60719201651755e-06, "loss": 0.6459, "step": 481150 }, { "epoch": 82.79421885753614, "grad_norm": 14.889739990234375, "learning_rate": 8.602890571231935e-06, "loss": 0.6679, "step": 481200 }, { "epoch": 82.80282174810736, "grad_norm": 7.6745476722717285, "learning_rate": 8.59858912594632e-06, "loss": 0.6498, "step": 481250 }, { "epoch": 82.8114246386786, "grad_norm": 10.93268871307373, "learning_rate": 8.594287680660702e-06, "loss": 0.694, "step": 481300 }, { "epoch": 82.82002752924983, "grad_norm": 26.443967819213867, "learning_rate": 8.589986235375087e-06, "loss": 0.6675, "step": 481350 }, { "epoch": 82.82863041982105, "grad_norm": 16.272668838500977, "learning_rate": 8.585684790089471e-06, "loss": 0.6776, "step": 481400 }, { "epoch": 82.83723331039229, "grad_norm": 9.14030647277832, "learning_rate": 8.581383344803854e-06, "loss": 0.6796, "step": 481450 }, { "epoch": 82.84583620096352, "grad_norm": 12.392766952514648, "learning_rate": 8.577081899518239e-06, "loss": 0.6858, "step": 481500 }, { "epoch": 82.85443909153476, "grad_norm": 20.95915412902832, "learning_rate": 8.572780454232621e-06, "loss": 0.624, "step": 481550 }, { "epoch": 82.86304198210598, "grad_norm": 15.432262420654297, "learning_rate": 8.568479008947008e-06, "loss": 0.6452, "step": 481600 }, { "epoch": 82.87164487267722, "grad_norm": 18.435972213745117, "learning_rate": 8.56417756366139e-06, "loss": 0.6851, "step": 481650 }, { "epoch": 82.88024776324845, "grad_norm": 17.26344108581543, "learning_rate": 8.559876118375775e-06, "loss": 0.6608, "step": 481700 }, { "epoch": 82.88885065381969, "grad_norm": 26.034208297729492, "learning_rate": 8.555574673090158e-06, "loss": 0.6249, "step": 481750 }, { "epoch": 82.89745354439091, "grad_norm": 28.214534759521484, "learning_rate": 8.551273227804543e-06, "loss": 0.6988, "step": 481800 }, { "epoch": 82.90605643496215, "grad_norm": 9.837757110595703, "learning_rate": 8.546971782518927e-06, "loss": 0.7032, "step": 481850 }, { "epoch": 82.91465932553338, "grad_norm": 16.579797744750977, "learning_rate": 8.54267033723331e-06, "loss": 0.619, "step": 481900 }, { "epoch": 82.92326221610462, "grad_norm": 9.0691556930542, "learning_rate": 8.538368891947695e-06, "loss": 0.6946, "step": 481950 }, { "epoch": 82.93186510667584, "grad_norm": 16.412141799926758, "learning_rate": 8.534067446662079e-06, "loss": 0.6477, "step": 482000 }, { "epoch": 82.94046799724707, "grad_norm": 17.275651931762695, "learning_rate": 8.529766001376464e-06, "loss": 0.6841, "step": 482050 }, { "epoch": 82.94907088781831, "grad_norm": 9.182684898376465, "learning_rate": 8.525464556090847e-06, "loss": 0.6816, "step": 482100 }, { "epoch": 82.95767377838953, "grad_norm": 21.615036010742188, "learning_rate": 8.521163110805231e-06, "loss": 0.689, "step": 482150 }, { "epoch": 82.96627666896077, "grad_norm": 19.816492080688477, "learning_rate": 8.516861665519616e-06, "loss": 0.6919, "step": 482200 }, { "epoch": 82.974879559532, "grad_norm": 19.969453811645508, "learning_rate": 8.512560220233999e-06, "loss": 0.6966, "step": 482250 }, { "epoch": 82.98348245010324, "grad_norm": 13.185559272766113, "learning_rate": 8.508258774948383e-06, "loss": 0.6611, "step": 482300 }, { "epoch": 82.99208534067446, "grad_norm": 18.29911994934082, "learning_rate": 8.503957329662766e-06, "loss": 0.6031, "step": 482350 }, { "epoch": 83.0, "eval_accuracy": 0.5929272070211667, "eval_f1": 0.5856165521296226, "eval_f1_DuraRiadoRio_16x16": 0.508907205530444, "eval_f1_Mole_16x16": 0.6334567485695052, "eval_f1_Quebrado_16x16": 0.6787783043707214, "eval_f1_RiadoRio_16x16": 0.4810126582278481, "eval_f1_RioFechado_16x16": 0.6259278439495943, "eval_loss": 2.014934539794922, "eval_precision": 0.6357087409923607, "eval_precision_DuraRiadoRio_16x16": 0.6568291008922443, "eval_precision_Mole_16x16": 0.5173172072567345, "eval_precision_Quebrado_16x16": 0.8627844712182061, "eval_precision_RiadoRio_16x16": 0.6258234519104084, "eval_precision_RioFechado_16x16": 0.5157894736842106, "eval_recall": 0.5956330479953176, "eval_recall_DuraRiadoRio_16x16": 0.4153645833333333, "eval_recall_Mole_16x16": 0.8168402777777778, "eval_recall_Quebrado_16x16": 0.5594618055555556, "eval_recall_RiadoRio_16x16": 0.390625, "eval_recall_RioFechado_16x16": 0.795873573309921, "eval_runtime": 46.3581, "eval_samples_per_second": 250.701, "eval_steps_per_second": 15.682, "step": 482396 }, { "epoch": 83.0006882312457, "grad_norm": 25.438232421875, "learning_rate": 8.499655884377152e-06, "loss": 0.6883, "step": 482400 }, { "epoch": 83.00929112181693, "grad_norm": 11.767084121704102, "learning_rate": 8.495354439091535e-06, "loss": 0.702, "step": 482450 }, { "epoch": 83.01789401238817, "grad_norm": 21.549436569213867, "learning_rate": 8.49105299380592e-06, "loss": 0.651, "step": 482500 }, { "epoch": 83.0264969029594, "grad_norm": 8.261144638061523, "learning_rate": 8.486751548520304e-06, "loss": 0.6421, "step": 482550 }, { "epoch": 83.03509979353062, "grad_norm": 9.542527198791504, "learning_rate": 8.482450103234687e-06, "loss": 0.7124, "step": 482600 }, { "epoch": 83.04370268410186, "grad_norm": 16.21212387084961, "learning_rate": 8.478148657949072e-06, "loss": 0.6717, "step": 482650 }, { "epoch": 83.05230557467308, "grad_norm": 10.0271577835083, "learning_rate": 8.473847212663454e-06, "loss": 0.6344, "step": 482700 }, { "epoch": 83.06090846524432, "grad_norm": 22.307449340820312, "learning_rate": 8.46954576737784e-06, "loss": 0.65, "step": 482750 }, { "epoch": 83.06951135581555, "grad_norm": 25.641122817993164, "learning_rate": 8.465244322092224e-06, "loss": 0.7015, "step": 482800 }, { "epoch": 83.07811424638679, "grad_norm": 18.04175567626953, "learning_rate": 8.460942876806608e-06, "loss": 0.6983, "step": 482850 }, { "epoch": 83.08671713695801, "grad_norm": 12.796831130981445, "learning_rate": 8.456641431520991e-06, "loss": 0.7053, "step": 482900 }, { "epoch": 83.09532002752925, "grad_norm": 10.406923294067383, "learning_rate": 8.452339986235376e-06, "loss": 0.6447, "step": 482950 }, { "epoch": 83.10392291810048, "grad_norm": 15.309198379516602, "learning_rate": 8.44803854094976e-06, "loss": 0.6525, "step": 483000 }, { "epoch": 83.11252580867172, "grad_norm": 12.240155220031738, "learning_rate": 8.443737095664143e-06, "loss": 0.6184, "step": 483050 }, { "epoch": 83.12112869924295, "grad_norm": 15.323019981384277, "learning_rate": 8.439435650378528e-06, "loss": 0.6683, "step": 483100 }, { "epoch": 83.12973158981418, "grad_norm": 7.7463765144348145, "learning_rate": 8.435134205092912e-06, "loss": 0.6803, "step": 483150 }, { "epoch": 83.13833448038541, "grad_norm": 17.14598274230957, "learning_rate": 8.430832759807297e-06, "loss": 0.6442, "step": 483200 }, { "epoch": 83.14693737095664, "grad_norm": 16.53684425354004, "learning_rate": 8.42653131452168e-06, "loss": 0.6484, "step": 483250 }, { "epoch": 83.15554026152788, "grad_norm": 12.473539352416992, "learning_rate": 8.422229869236062e-06, "loss": 0.639, "step": 483300 }, { "epoch": 83.1641431520991, "grad_norm": 13.481487274169922, "learning_rate": 8.417928423950449e-06, "loss": 0.6423, "step": 483350 }, { "epoch": 83.17274604267034, "grad_norm": 16.210458755493164, "learning_rate": 8.413626978664832e-06, "loss": 0.6264, "step": 483400 }, { "epoch": 83.18134893324157, "grad_norm": 13.08036994934082, "learning_rate": 8.409325533379216e-06, "loss": 0.6156, "step": 483450 }, { "epoch": 83.1899518238128, "grad_norm": 11.449079513549805, "learning_rate": 8.405024088093599e-06, "loss": 0.63, "step": 483500 }, { "epoch": 83.19855471438403, "grad_norm": 20.627941131591797, "learning_rate": 8.400722642807985e-06, "loss": 0.6577, "step": 483550 }, { "epoch": 83.20715760495527, "grad_norm": 27.329530715942383, "learning_rate": 8.396421197522368e-06, "loss": 0.6401, "step": 483600 }, { "epoch": 83.2157604955265, "grad_norm": 11.638360977172852, "learning_rate": 8.392119752236751e-06, "loss": 0.6099, "step": 483650 }, { "epoch": 83.22436338609774, "grad_norm": 28.3580265045166, "learning_rate": 8.387818306951136e-06, "loss": 0.6679, "step": 483700 }, { "epoch": 83.23296627666896, "grad_norm": 19.577924728393555, "learning_rate": 8.38351686166552e-06, "loss": 0.7067, "step": 483750 }, { "epoch": 83.24156916724019, "grad_norm": 14.955330848693848, "learning_rate": 8.379215416379905e-06, "loss": 0.6596, "step": 483800 }, { "epoch": 83.25017205781143, "grad_norm": 13.444372177124023, "learning_rate": 8.374913971094287e-06, "loss": 0.6468, "step": 483850 }, { "epoch": 83.25877494838265, "grad_norm": 10.234774589538574, "learning_rate": 8.370612525808672e-06, "loss": 0.6638, "step": 483900 }, { "epoch": 83.26737783895389, "grad_norm": 11.943718910217285, "learning_rate": 8.366311080523057e-06, "loss": 0.7597, "step": 483950 }, { "epoch": 83.27598072952512, "grad_norm": 15.602625846862793, "learning_rate": 8.36200963523744e-06, "loss": 0.6015, "step": 484000 }, { "epoch": 83.28458362009636, "grad_norm": 13.804827690124512, "learning_rate": 8.357708189951824e-06, "loss": 0.6643, "step": 484050 }, { "epoch": 83.29318651066758, "grad_norm": 13.564632415771484, "learning_rate": 8.353406744666207e-06, "loss": 0.7043, "step": 484100 }, { "epoch": 83.30178940123882, "grad_norm": 14.998226165771484, "learning_rate": 8.349105299380593e-06, "loss": 0.6106, "step": 484150 }, { "epoch": 83.31039229181005, "grad_norm": 18.16680335998535, "learning_rate": 8.344803854094976e-06, "loss": 0.6852, "step": 484200 }, { "epoch": 83.31899518238129, "grad_norm": 17.167016983032227, "learning_rate": 8.34050240880936e-06, "loss": 0.6967, "step": 484250 }, { "epoch": 83.32759807295251, "grad_norm": 9.232386589050293, "learning_rate": 8.336200963523743e-06, "loss": 0.6361, "step": 484300 }, { "epoch": 83.33620096352374, "grad_norm": 16.643835067749023, "learning_rate": 8.331899518238128e-06, "loss": 0.6837, "step": 484350 }, { "epoch": 83.34480385409498, "grad_norm": 9.02379322052002, "learning_rate": 8.327598072952513e-06, "loss": 0.6203, "step": 484400 }, { "epoch": 83.3534067446662, "grad_norm": 10.885017395019531, "learning_rate": 8.323296627666895e-06, "loss": 0.6465, "step": 484450 }, { "epoch": 83.36200963523744, "grad_norm": 7.198322772979736, "learning_rate": 8.318995182381282e-06, "loss": 0.662, "step": 484500 }, { "epoch": 83.37061252580867, "grad_norm": 7.684818267822266, "learning_rate": 8.314693737095665e-06, "loss": 0.626, "step": 484550 }, { "epoch": 83.3792154163799, "grad_norm": 27.516704559326172, "learning_rate": 8.310392291810049e-06, "loss": 0.6475, "step": 484600 }, { "epoch": 83.38781830695113, "grad_norm": 20.803373336791992, "learning_rate": 8.306090846524432e-06, "loss": 0.6883, "step": 484650 }, { "epoch": 83.39642119752237, "grad_norm": 20.32715606689453, "learning_rate": 8.301789401238817e-06, "loss": 0.6176, "step": 484700 }, { "epoch": 83.4050240880936, "grad_norm": 24.29396629333496, "learning_rate": 8.297487955953201e-06, "loss": 0.7281, "step": 484750 }, { "epoch": 83.41362697866484, "grad_norm": 19.79067611694336, "learning_rate": 8.293186510667584e-06, "loss": 0.7077, "step": 484800 }, { "epoch": 83.42222986923606, "grad_norm": 12.52553653717041, "learning_rate": 8.288885065381969e-06, "loss": 0.6681, "step": 484850 }, { "epoch": 83.4308327598073, "grad_norm": 16.978710174560547, "learning_rate": 8.284583620096353e-06, "loss": 0.6908, "step": 484900 }, { "epoch": 83.43943565037853, "grad_norm": 17.628137588500977, "learning_rate": 8.280282174810738e-06, "loss": 0.642, "step": 484950 }, { "epoch": 83.44803854094975, "grad_norm": 18.135900497436523, "learning_rate": 8.27598072952512e-06, "loss": 0.6885, "step": 485000 }, { "epoch": 83.45664143152099, "grad_norm": 19.771522521972656, "learning_rate": 8.271679284239505e-06, "loss": 0.654, "step": 485050 }, { "epoch": 83.46524432209222, "grad_norm": 16.76266860961914, "learning_rate": 8.26737783895389e-06, "loss": 0.6709, "step": 485100 }, { "epoch": 83.47384721266346, "grad_norm": 11.847127914428711, "learning_rate": 8.263076393668273e-06, "loss": 0.6491, "step": 485150 }, { "epoch": 83.48245010323468, "grad_norm": 12.667576789855957, "learning_rate": 8.258774948382657e-06, "loss": 0.6417, "step": 485200 }, { "epoch": 83.49105299380592, "grad_norm": 14.537628173828125, "learning_rate": 8.25447350309704e-06, "loss": 0.6626, "step": 485250 }, { "epoch": 83.49965588437715, "grad_norm": 16.05364227294922, "learning_rate": 8.250172057811426e-06, "loss": 0.6764, "step": 485300 }, { "epoch": 83.50825877494839, "grad_norm": 16.89247703552246, "learning_rate": 8.245870612525809e-06, "loss": 0.6401, "step": 485350 }, { "epoch": 83.51686166551961, "grad_norm": 12.047242164611816, "learning_rate": 8.241569167240194e-06, "loss": 0.6912, "step": 485400 }, { "epoch": 83.52546455609085, "grad_norm": 11.942200660705566, "learning_rate": 8.237267721954576e-06, "loss": 0.6634, "step": 485450 }, { "epoch": 83.53406744666208, "grad_norm": 25.75824737548828, "learning_rate": 8.232966276668961e-06, "loss": 0.6491, "step": 485500 }, { "epoch": 83.5426703372333, "grad_norm": 12.09078311920166, "learning_rate": 8.228664831383346e-06, "loss": 0.6217, "step": 485550 }, { "epoch": 83.55127322780454, "grad_norm": 7.803417682647705, "learning_rate": 8.224363386097728e-06, "loss": 0.6059, "step": 485600 }, { "epoch": 83.55987611837577, "grad_norm": 15.464713096618652, "learning_rate": 8.220061940812113e-06, "loss": 0.6412, "step": 485650 }, { "epoch": 83.56847900894701, "grad_norm": 11.608738899230957, "learning_rate": 8.215760495526498e-06, "loss": 0.6803, "step": 485700 }, { "epoch": 83.57708189951823, "grad_norm": 19.417095184326172, "learning_rate": 8.211459050240882e-06, "loss": 0.596, "step": 485750 }, { "epoch": 83.58568479008947, "grad_norm": 12.400031089782715, "learning_rate": 8.207157604955265e-06, "loss": 0.6807, "step": 485800 }, { "epoch": 83.5942876806607, "grad_norm": 13.943907737731934, "learning_rate": 8.20285615966965e-06, "loss": 0.6323, "step": 485850 }, { "epoch": 83.60289057123194, "grad_norm": 12.299248695373535, "learning_rate": 8.198554714384034e-06, "loss": 0.7125, "step": 485900 }, { "epoch": 83.61149346180316, "grad_norm": 19.75616455078125, "learning_rate": 8.194253269098417e-06, "loss": 0.6382, "step": 485950 }, { "epoch": 83.6200963523744, "grad_norm": 13.113226890563965, "learning_rate": 8.189951823812802e-06, "loss": 0.7339, "step": 486000 }, { "epoch": 83.62869924294563, "grad_norm": 19.867143630981445, "learning_rate": 8.185650378527184e-06, "loss": 0.711, "step": 486050 }, { "epoch": 83.63730213351687, "grad_norm": 14.15317440032959, "learning_rate": 8.18134893324157e-06, "loss": 0.6221, "step": 486100 }, { "epoch": 83.6459050240881, "grad_norm": 8.939547538757324, "learning_rate": 8.177047487955954e-06, "loss": 0.638, "step": 486150 }, { "epoch": 83.65450791465932, "grad_norm": 12.929239273071289, "learning_rate": 8.172746042670338e-06, "loss": 0.6691, "step": 486200 }, { "epoch": 83.66311080523056, "grad_norm": 19.554479598999023, "learning_rate": 8.168444597384723e-06, "loss": 0.7206, "step": 486250 }, { "epoch": 83.67171369580178, "grad_norm": 18.13275718688965, "learning_rate": 8.164143152099106e-06, "loss": 0.661, "step": 486300 }, { "epoch": 83.68031658637302, "grad_norm": 9.383716583251953, "learning_rate": 8.15984170681349e-06, "loss": 0.6618, "step": 486350 }, { "epoch": 83.68891947694425, "grad_norm": 20.77005386352539, "learning_rate": 8.155540261527873e-06, "loss": 0.7172, "step": 486400 }, { "epoch": 83.69752236751549, "grad_norm": 14.961509704589844, "learning_rate": 8.15123881624226e-06, "loss": 0.6855, "step": 486450 }, { "epoch": 83.70612525808671, "grad_norm": 13.638750076293945, "learning_rate": 8.146937370956642e-06, "loss": 0.6227, "step": 486500 }, { "epoch": 83.71472814865795, "grad_norm": 14.357775688171387, "learning_rate": 8.142635925671025e-06, "loss": 0.6997, "step": 486550 }, { "epoch": 83.72333103922918, "grad_norm": 18.92588996887207, "learning_rate": 8.13833448038541e-06, "loss": 0.5911, "step": 486600 }, { "epoch": 83.73193392980042, "grad_norm": 12.024611473083496, "learning_rate": 8.134033035099794e-06, "loss": 0.6251, "step": 486650 }, { "epoch": 83.74053682037164, "grad_norm": 14.042343139648438, "learning_rate": 8.129731589814179e-06, "loss": 0.6179, "step": 486700 }, { "epoch": 83.74913971094287, "grad_norm": 13.753011703491211, "learning_rate": 8.125430144528562e-06, "loss": 0.6901, "step": 486750 }, { "epoch": 83.75774260151411, "grad_norm": 20.441015243530273, "learning_rate": 8.121128699242946e-06, "loss": 0.6531, "step": 486800 }, { "epoch": 83.76634549208534, "grad_norm": 13.178189277648926, "learning_rate": 8.11682725395733e-06, "loss": 0.712, "step": 486850 }, { "epoch": 83.77494838265658, "grad_norm": 17.939517974853516, "learning_rate": 8.112525808671713e-06, "loss": 0.5857, "step": 486900 }, { "epoch": 83.7835512732278, "grad_norm": 13.060068130493164, "learning_rate": 8.108224363386098e-06, "loss": 0.6851, "step": 486950 }, { "epoch": 83.79215416379904, "grad_norm": 11.8067626953125, "learning_rate": 8.103922918100481e-06, "loss": 0.6242, "step": 487000 }, { "epoch": 83.80075705437027, "grad_norm": 15.015557289123535, "learning_rate": 8.099621472814867e-06, "loss": 0.6929, "step": 487050 }, { "epoch": 83.8093599449415, "grad_norm": 17.014225006103516, "learning_rate": 8.09532002752925e-06, "loss": 0.6848, "step": 487100 }, { "epoch": 83.81796283551273, "grad_norm": 15.466460227966309, "learning_rate": 8.091018582243635e-06, "loss": 0.6703, "step": 487150 }, { "epoch": 83.82656572608397, "grad_norm": 22.49744415283203, "learning_rate": 8.086717136958017e-06, "loss": 0.6943, "step": 487200 }, { "epoch": 83.8351686166552, "grad_norm": 24.95904541015625, "learning_rate": 8.082415691672402e-06, "loss": 0.6358, "step": 487250 }, { "epoch": 83.84377150722642, "grad_norm": 17.833066940307617, "learning_rate": 8.078114246386787e-06, "loss": 0.6475, "step": 487300 }, { "epoch": 83.85237439779766, "grad_norm": 22.96310043334961, "learning_rate": 8.07381280110117e-06, "loss": 0.7195, "step": 487350 }, { "epoch": 83.86097728836889, "grad_norm": 10.673112869262695, "learning_rate": 8.069511355815554e-06, "loss": 0.7165, "step": 487400 }, { "epoch": 83.86958017894013, "grad_norm": 15.77454948425293, "learning_rate": 8.065209910529939e-06, "loss": 0.702, "step": 487450 }, { "epoch": 83.87818306951135, "grad_norm": 23.311355590820312, "learning_rate": 8.060908465244323e-06, "loss": 0.6652, "step": 487500 }, { "epoch": 83.88678596008259, "grad_norm": 11.447202682495117, "learning_rate": 8.056607019958706e-06, "loss": 0.7288, "step": 487550 }, { "epoch": 83.89538885065382, "grad_norm": 15.166534423828125, "learning_rate": 8.05230557467309e-06, "loss": 0.6808, "step": 487600 }, { "epoch": 83.90399174122506, "grad_norm": 20.221294403076172, "learning_rate": 8.048004129387475e-06, "loss": 0.655, "step": 487650 }, { "epoch": 83.91259463179628, "grad_norm": 7.1578369140625, "learning_rate": 8.043702684101858e-06, "loss": 0.7094, "step": 487700 }, { "epoch": 83.92119752236752, "grad_norm": 9.56622314453125, "learning_rate": 8.039401238816243e-06, "loss": 0.6584, "step": 487750 }, { "epoch": 83.92980041293875, "grad_norm": 17.47549819946289, "learning_rate": 8.035099793530625e-06, "loss": 0.7114, "step": 487800 }, { "epoch": 83.93840330350999, "grad_norm": 14.549206733703613, "learning_rate": 8.030798348245012e-06, "loss": 0.6574, "step": 487850 }, { "epoch": 83.94700619408121, "grad_norm": 17.384780883789062, "learning_rate": 8.026496902959395e-06, "loss": 0.6962, "step": 487900 }, { "epoch": 83.95560908465244, "grad_norm": 15.782092094421387, "learning_rate": 8.022195457673779e-06, "loss": 0.6898, "step": 487950 }, { "epoch": 83.96421197522368, "grad_norm": 10.193105697631836, "learning_rate": 8.017894012388162e-06, "loss": 0.6846, "step": 488000 }, { "epoch": 83.9728148657949, "grad_norm": 17.99187660217285, "learning_rate": 8.013592567102547e-06, "loss": 0.6177, "step": 488050 }, { "epoch": 83.98141775636614, "grad_norm": 18.370275497436523, "learning_rate": 8.009291121816931e-06, "loss": 0.6968, "step": 488100 }, { "epoch": 83.99002064693737, "grad_norm": 6.480834484100342, "learning_rate": 8.004989676531314e-06, "loss": 0.6828, "step": 488150 }, { "epoch": 83.9986235375086, "grad_norm": 13.740110397338867, "learning_rate": 8.0006882312457e-06, "loss": 0.6239, "step": 488200 }, { "epoch": 84.0, "eval_accuracy": 0.55377731887799, "eval_f1": 0.5352265057521578, "eval_f1_DuraRiadoRio_16x16": 0.5657080451401529, "eval_f1_Mole_16x16": 0.3887020847343645, "eval_f1_Quebrado_16x16": 0.6704469646430954, "eval_f1_RiadoRio_16x16": 0.488, "eval_f1_RioFechado_16x16": 0.5632754342431762, "eval_loss": 2.432405471801758, "eval_precision": 0.6084170642359435, "eval_precision_DuraRiadoRio_16x16": 0.4871473354231975, "eval_precision_Mole_16x16": 0.8626865671641791, "eval_precision_Quebrado_16x16": 0.5444203683640303, "eval_precision_RiadoRio_16x16": 0.5, "eval_precision_RioFechado_16x16": 0.6478310502283106, "eval_recall": 0.5545099258608917, "eval_recall_DuraRiadoRio_16x16": 0.6744791666666666, "eval_recall_Mole_16x16": 0.2508680555555556, "eval_recall_Quebrado_16x16": 0.8723958333333334, "eval_recall_RiadoRio_16x16": 0.4765625, "eval_recall_RioFechado_16x16": 0.49824407374890256, "eval_runtime": 45.4902, "eval_samples_per_second": 255.483, "eval_steps_per_second": 15.981, "step": 488208 }, { "epoch": 84.00722642807983, "grad_norm": 9.641728401184082, "learning_rate": 7.996386785960083e-06, "loss": 0.6472, "step": 488250 }, { "epoch": 84.01582931865107, "grad_norm": 18.477807998657227, "learning_rate": 7.992085340674468e-06, "loss": 0.669, "step": 488300 }, { "epoch": 84.0244322092223, "grad_norm": 9.476344108581543, "learning_rate": 7.98778389538885e-06, "loss": 0.6955, "step": 488350 }, { "epoch": 84.03303509979354, "grad_norm": 10.063339233398438, "learning_rate": 7.983482450103235e-06, "loss": 0.6653, "step": 488400 }, { "epoch": 84.04163799036476, "grad_norm": 25.398719787597656, "learning_rate": 7.97918100481762e-06, "loss": 0.7258, "step": 488450 }, { "epoch": 84.05024088093599, "grad_norm": 16.803939819335938, "learning_rate": 7.974879559532002e-06, "loss": 0.645, "step": 488500 }, { "epoch": 84.05884377150723, "grad_norm": 22.41292953491211, "learning_rate": 7.970578114246387e-06, "loss": 0.6339, "step": 488550 }, { "epoch": 84.06744666207845, "grad_norm": 18.330768585205078, "learning_rate": 7.966276668960772e-06, "loss": 0.6129, "step": 488600 }, { "epoch": 84.07604955264969, "grad_norm": 20.501922607421875, "learning_rate": 7.961975223675156e-06, "loss": 0.6177, "step": 488650 }, { "epoch": 84.08465244322092, "grad_norm": 15.222174644470215, "learning_rate": 7.957673778389539e-06, "loss": 0.6774, "step": 488700 }, { "epoch": 84.09325533379216, "grad_norm": 23.073535919189453, "learning_rate": 7.953372333103924e-06, "loss": 0.6535, "step": 488750 }, { "epoch": 84.10185822436338, "grad_norm": 18.390607833862305, "learning_rate": 7.949070887818308e-06, "loss": 0.6797, "step": 488800 }, { "epoch": 84.11046111493462, "grad_norm": 16.603940963745117, "learning_rate": 7.944769442532691e-06, "loss": 0.6132, "step": 488850 }, { "epoch": 84.11906400550585, "grad_norm": 14.094219207763672, "learning_rate": 7.940467997247076e-06, "loss": 0.667, "step": 488900 }, { "epoch": 84.12766689607709, "grad_norm": 25.761869430541992, "learning_rate": 7.936166551961458e-06, "loss": 0.6541, "step": 488950 }, { "epoch": 84.13626978664831, "grad_norm": 15.92553997039795, "learning_rate": 7.931865106675845e-06, "loss": 0.6303, "step": 489000 }, { "epoch": 84.14487267721955, "grad_norm": 16.911422729492188, "learning_rate": 7.927563661390228e-06, "loss": 0.7059, "step": 489050 }, { "epoch": 84.15347556779078, "grad_norm": 16.297714233398438, "learning_rate": 7.923262216104612e-06, "loss": 0.716, "step": 489100 }, { "epoch": 84.162078458362, "grad_norm": 30.174053192138672, "learning_rate": 7.918960770818995e-06, "loss": 0.6517, "step": 489150 }, { "epoch": 84.17068134893324, "grad_norm": 18.794593811035156, "learning_rate": 7.91465932553338e-06, "loss": 0.6489, "step": 489200 }, { "epoch": 84.17928423950447, "grad_norm": 22.86301040649414, "learning_rate": 7.910357880247764e-06, "loss": 0.6814, "step": 489250 }, { "epoch": 84.18788713007571, "grad_norm": 17.589719772338867, "learning_rate": 7.906056434962147e-06, "loss": 0.6751, "step": 489300 }, { "epoch": 84.19649002064693, "grad_norm": 11.083344459533691, "learning_rate": 7.901754989676532e-06, "loss": 0.6818, "step": 489350 }, { "epoch": 84.20509291121817, "grad_norm": 10.422979354858398, "learning_rate": 7.897453544390916e-06, "loss": 0.6109, "step": 489400 }, { "epoch": 84.2136958017894, "grad_norm": 11.514827728271484, "learning_rate": 7.8931520991053e-06, "loss": 0.6222, "step": 489450 }, { "epoch": 84.22229869236064, "grad_norm": 11.386133193969727, "learning_rate": 7.888850653819684e-06, "loss": 0.6655, "step": 489500 }, { "epoch": 84.23090158293186, "grad_norm": 19.8861083984375, "learning_rate": 7.884549208534066e-06, "loss": 0.7016, "step": 489550 }, { "epoch": 84.2395044735031, "grad_norm": 20.27396011352539, "learning_rate": 7.880247763248453e-06, "loss": 0.6688, "step": 489600 }, { "epoch": 84.24810736407433, "grad_norm": 13.502131462097168, "learning_rate": 7.875946317962836e-06, "loss": 0.6759, "step": 489650 }, { "epoch": 84.25671025464555, "grad_norm": 21.868480682373047, "learning_rate": 7.87164487267722e-06, "loss": 0.7007, "step": 489700 }, { "epoch": 84.2653131452168, "grad_norm": 17.089797973632812, "learning_rate": 7.867343427391603e-06, "loss": 0.6411, "step": 489750 }, { "epoch": 84.27391603578802, "grad_norm": 17.170305252075195, "learning_rate": 7.86304198210599e-06, "loss": 0.6291, "step": 489800 }, { "epoch": 84.28251892635926, "grad_norm": 16.976259231567383, "learning_rate": 7.858740536820372e-06, "loss": 0.6014, "step": 489850 }, { "epoch": 84.29112181693048, "grad_norm": 16.137832641601562, "learning_rate": 7.854439091534755e-06, "loss": 0.6921, "step": 489900 }, { "epoch": 84.29972470750172, "grad_norm": 24.491304397583008, "learning_rate": 7.850137646249141e-06, "loss": 0.6424, "step": 489950 }, { "epoch": 84.30832759807295, "grad_norm": 11.86611270904541, "learning_rate": 7.845836200963524e-06, "loss": 0.6332, "step": 490000 }, { "epoch": 84.31693048864419, "grad_norm": 13.270759582519531, "learning_rate": 7.841534755677909e-06, "loss": 0.6705, "step": 490050 }, { "epoch": 84.32553337921541, "grad_norm": 16.906063079833984, "learning_rate": 7.837233310392291e-06, "loss": 0.7031, "step": 490100 }, { "epoch": 84.33413626978665, "grad_norm": 20.26074981689453, "learning_rate": 7.832931865106676e-06, "loss": 0.6434, "step": 490150 }, { "epoch": 84.34273916035788, "grad_norm": 16.961673736572266, "learning_rate": 7.82863041982106e-06, "loss": 0.682, "step": 490200 }, { "epoch": 84.3513420509291, "grad_norm": 16.940053939819336, "learning_rate": 7.824328974535443e-06, "loss": 0.6697, "step": 490250 }, { "epoch": 84.35994494150034, "grad_norm": 16.030658721923828, "learning_rate": 7.820027529249828e-06, "loss": 0.6508, "step": 490300 }, { "epoch": 84.36854783207157, "grad_norm": 14.03860092163086, "learning_rate": 7.815726083964213e-06, "loss": 0.6779, "step": 490350 }, { "epoch": 84.37715072264281, "grad_norm": 11.804716110229492, "learning_rate": 7.811424638678597e-06, "loss": 0.6199, "step": 490400 }, { "epoch": 84.38575361321404, "grad_norm": 19.541946411132812, "learning_rate": 7.80712319339298e-06, "loss": 0.6722, "step": 490450 }, { "epoch": 84.39435650378527, "grad_norm": 5.6333136558532715, "learning_rate": 7.802821748107365e-06, "loss": 0.6965, "step": 490500 }, { "epoch": 84.4029593943565, "grad_norm": 13.434272766113281, "learning_rate": 7.798520302821749e-06, "loss": 0.6878, "step": 490550 }, { "epoch": 84.41156228492774, "grad_norm": 12.493491172790527, "learning_rate": 7.794218857536132e-06, "loss": 0.6749, "step": 490600 }, { "epoch": 84.42016517549897, "grad_norm": 15.300950050354004, "learning_rate": 7.789917412250517e-06, "loss": 0.6421, "step": 490650 }, { "epoch": 84.4287680660702, "grad_norm": 16.57004737854004, "learning_rate": 7.7856159669649e-06, "loss": 0.6372, "step": 490700 }, { "epoch": 84.43737095664143, "grad_norm": 18.785608291625977, "learning_rate": 7.781314521679286e-06, "loss": 0.6292, "step": 490750 }, { "epoch": 84.44597384721267, "grad_norm": 13.243295669555664, "learning_rate": 7.777013076393669e-06, "loss": 0.6042, "step": 490800 }, { "epoch": 84.4545767377839, "grad_norm": 13.142218589782715, "learning_rate": 7.772711631108053e-06, "loss": 0.6663, "step": 490850 }, { "epoch": 84.46317962835512, "grad_norm": 16.66331672668457, "learning_rate": 7.768410185822436e-06, "loss": 0.6451, "step": 490900 }, { "epoch": 84.47178251892636, "grad_norm": 14.631213188171387, "learning_rate": 7.76410874053682e-06, "loss": 0.6639, "step": 490950 }, { "epoch": 84.48038540949759, "grad_norm": 13.762508392333984, "learning_rate": 7.759807295251205e-06, "loss": 0.6594, "step": 491000 }, { "epoch": 84.48898830006883, "grad_norm": 11.276511192321777, "learning_rate": 7.755505849965588e-06, "loss": 0.6198, "step": 491050 }, { "epoch": 84.49759119064005, "grad_norm": 13.425012588500977, "learning_rate": 7.751204404679973e-06, "loss": 0.5986, "step": 491100 }, { "epoch": 84.50619408121129, "grad_norm": 11.934087753295898, "learning_rate": 7.746902959394357e-06, "loss": 0.5622, "step": 491150 }, { "epoch": 84.51479697178252, "grad_norm": 13.950657844543457, "learning_rate": 7.742601514108742e-06, "loss": 0.6818, "step": 491200 }, { "epoch": 84.52339986235376, "grad_norm": 19.304401397705078, "learning_rate": 7.738300068823125e-06, "loss": 0.6282, "step": 491250 }, { "epoch": 84.53200275292498, "grad_norm": 17.461225509643555, "learning_rate": 7.733998623537509e-06, "loss": 0.6866, "step": 491300 }, { "epoch": 84.54060564349622, "grad_norm": 20.548086166381836, "learning_rate": 7.729697178251894e-06, "loss": 0.6757, "step": 491350 }, { "epoch": 84.54920853406745, "grad_norm": 12.167441368103027, "learning_rate": 7.725395732966276e-06, "loss": 0.6448, "step": 491400 }, { "epoch": 84.55781142463867, "grad_norm": 16.77322769165039, "learning_rate": 7.721094287680661e-06, "loss": 0.6419, "step": 491450 }, { "epoch": 84.56641431520991, "grad_norm": 14.834619522094727, "learning_rate": 7.716792842395044e-06, "loss": 0.7284, "step": 491500 }, { "epoch": 84.57501720578114, "grad_norm": 13.47657299041748, "learning_rate": 7.71249139710943e-06, "loss": 0.6722, "step": 491550 }, { "epoch": 84.58362009635238, "grad_norm": 19.32236671447754, "learning_rate": 7.708189951823813e-06, "loss": 0.6394, "step": 491600 }, { "epoch": 84.5922229869236, "grad_norm": 20.494047164916992, "learning_rate": 7.703888506538198e-06, "loss": 0.6719, "step": 491650 }, { "epoch": 84.60082587749484, "grad_norm": 21.023191452026367, "learning_rate": 7.69958706125258e-06, "loss": 0.6382, "step": 491700 }, { "epoch": 84.60942876806607, "grad_norm": 13.381793975830078, "learning_rate": 7.695285615966965e-06, "loss": 0.6501, "step": 491750 }, { "epoch": 84.6180316586373, "grad_norm": 22.66042709350586, "learning_rate": 7.69098417068135e-06, "loss": 0.7127, "step": 491800 }, { "epoch": 84.62663454920853, "grad_norm": 16.048118591308594, "learning_rate": 7.686682725395732e-06, "loss": 0.6537, "step": 491850 }, { "epoch": 84.63523743977977, "grad_norm": 18.620119094848633, "learning_rate": 7.682381280110119e-06, "loss": 0.6713, "step": 491900 }, { "epoch": 84.643840330351, "grad_norm": 20.548545837402344, "learning_rate": 7.678079834824502e-06, "loss": 0.6193, "step": 491950 }, { "epoch": 84.65244322092224, "grad_norm": 17.670907974243164, "learning_rate": 7.673778389538886e-06, "loss": 0.6982, "step": 492000 }, { "epoch": 84.66104611149346, "grad_norm": 11.349405288696289, "learning_rate": 7.669476944253269e-06, "loss": 0.6518, "step": 492050 }, { "epoch": 84.66964900206469, "grad_norm": 9.5938138961792, "learning_rate": 7.665175498967654e-06, "loss": 0.6865, "step": 492100 }, { "epoch": 84.67825189263593, "grad_norm": 7.580497741699219, "learning_rate": 7.660874053682038e-06, "loss": 0.6534, "step": 492150 }, { "epoch": 84.68685478320715, "grad_norm": 22.219940185546875, "learning_rate": 7.656572608396421e-06, "loss": 0.6121, "step": 492200 }, { "epoch": 84.69545767377839, "grad_norm": 13.552388191223145, "learning_rate": 7.652271163110806e-06, "loss": 0.6867, "step": 492250 }, { "epoch": 84.70406056434962, "grad_norm": 21.599626541137695, "learning_rate": 7.64796971782519e-06, "loss": 0.6671, "step": 492300 }, { "epoch": 84.71266345492086, "grad_norm": 16.971446990966797, "learning_rate": 7.643668272539575e-06, "loss": 0.6595, "step": 492350 }, { "epoch": 84.72126634549208, "grad_norm": 21.345792770385742, "learning_rate": 7.639366827253958e-06, "loss": 0.6676, "step": 492400 }, { "epoch": 84.72986923606332, "grad_norm": 15.572640419006348, "learning_rate": 7.635065381968342e-06, "loss": 0.6661, "step": 492450 }, { "epoch": 84.73847212663455, "grad_norm": 14.476476669311523, "learning_rate": 7.630763936682727e-06, "loss": 0.6634, "step": 492500 }, { "epoch": 84.74707501720579, "grad_norm": 10.929024696350098, "learning_rate": 7.62646249139711e-06, "loss": 0.6573, "step": 492550 }, { "epoch": 84.75567790777701, "grad_norm": 13.75326156616211, "learning_rate": 7.622161046111494e-06, "loss": 0.6045, "step": 492600 }, { "epoch": 84.76428079834824, "grad_norm": 15.455034255981445, "learning_rate": 7.617859600825877e-06, "loss": 0.6266, "step": 492650 }, { "epoch": 84.77288368891948, "grad_norm": 17.55058479309082, "learning_rate": 7.613558155540262e-06, "loss": 0.7138, "step": 492700 }, { "epoch": 84.7814865794907, "grad_norm": 13.530488014221191, "learning_rate": 7.609256710254646e-06, "loss": 0.6136, "step": 492750 }, { "epoch": 84.79008947006194, "grad_norm": 23.888269424438477, "learning_rate": 7.60495526496903e-06, "loss": 0.7434, "step": 492800 }, { "epoch": 84.79869236063317, "grad_norm": 11.054130554199219, "learning_rate": 7.6006538196834135e-06, "loss": 0.66, "step": 492850 }, { "epoch": 84.80729525120441, "grad_norm": 18.22342872619629, "learning_rate": 7.596352374397799e-06, "loss": 0.6386, "step": 492900 }, { "epoch": 84.81589814177563, "grad_norm": 25.08291244506836, "learning_rate": 7.592050929112182e-06, "loss": 0.6075, "step": 492950 }, { "epoch": 84.82450103234687, "grad_norm": 13.059106826782227, "learning_rate": 7.5877494838265655e-06, "loss": 0.6349, "step": 493000 }, { "epoch": 84.8331039229181, "grad_norm": 15.67974853515625, "learning_rate": 7.583448038540949e-06, "loss": 0.6268, "step": 493050 }, { "epoch": 84.84170681348934, "grad_norm": 11.798656463623047, "learning_rate": 7.579146593255335e-06, "loss": 0.6466, "step": 493100 }, { "epoch": 84.85030970406056, "grad_norm": 10.85315227508545, "learning_rate": 7.574845147969718e-06, "loss": 0.6772, "step": 493150 }, { "epoch": 84.85891259463179, "grad_norm": 22.459959030151367, "learning_rate": 7.570543702684102e-06, "loss": 0.6692, "step": 493200 }, { "epoch": 84.86751548520303, "grad_norm": 15.510790824890137, "learning_rate": 7.566242257398486e-06, "loss": 0.6572, "step": 493250 }, { "epoch": 84.87611837577425, "grad_norm": 11.928265571594238, "learning_rate": 7.56194081211287e-06, "loss": 0.6372, "step": 493300 }, { "epoch": 84.8847212663455, "grad_norm": 11.896647453308105, "learning_rate": 7.557639366827254e-06, "loss": 0.6301, "step": 493350 }, { "epoch": 84.89332415691672, "grad_norm": 20.492244720458984, "learning_rate": 7.553337921541638e-06, "loss": 0.7043, "step": 493400 }, { "epoch": 84.90192704748796, "grad_norm": 14.336495399475098, "learning_rate": 7.5490364762560215e-06, "loss": 0.6737, "step": 493450 }, { "epoch": 84.91052993805918, "grad_norm": 24.61209487915039, "learning_rate": 7.544735030970407e-06, "loss": 0.6375, "step": 493500 }, { "epoch": 84.91913282863042, "grad_norm": 14.548975944519043, "learning_rate": 7.5404335856847906e-06, "loss": 0.6672, "step": 493550 }, { "epoch": 84.92773571920165, "grad_norm": 13.557744026184082, "learning_rate": 7.536132140399174e-06, "loss": 0.6813, "step": 493600 }, { "epoch": 84.93633860977289, "grad_norm": 13.52298641204834, "learning_rate": 7.531830695113559e-06, "loss": 0.6282, "step": 493650 }, { "epoch": 84.94494150034411, "grad_norm": 16.39856719970703, "learning_rate": 7.5275292498279426e-06, "loss": 0.6545, "step": 493700 }, { "epoch": 84.95354439091535, "grad_norm": 15.494378089904785, "learning_rate": 7.523227804542326e-06, "loss": 0.6524, "step": 493750 }, { "epoch": 84.96214728148658, "grad_norm": 11.905097961425781, "learning_rate": 7.51892635925671e-06, "loss": 0.6794, "step": 493800 }, { "epoch": 84.9707501720578, "grad_norm": 14.256656646728516, "learning_rate": 7.514624913971095e-06, "loss": 0.6996, "step": 493850 }, { "epoch": 84.97935306262904, "grad_norm": 17.649730682373047, "learning_rate": 7.510323468685479e-06, "loss": 0.6479, "step": 493900 }, { "epoch": 84.98795595320027, "grad_norm": 16.09511375427246, "learning_rate": 7.506022023399863e-06, "loss": 0.6638, "step": 493950 }, { "epoch": 84.99655884377151, "grad_norm": 16.422264099121094, "learning_rate": 7.5017205781142465e-06, "loss": 0.6656, "step": 494000 }, { "epoch": 85.0, "eval_accuracy": 0.5340733092410945, "eval_f1": 0.5389419153807113, "eval_f1_DuraRiadoRio_16x16": 0.48044978277536415, "eval_f1_Mole_16x16": 0.5583611243449261, "eval_f1_Quebrado_16x16": 0.6867067468269873, "eval_f1_RiadoRio_16x16": 0.4625829347322944, "eval_f1_RioFechado_16x16": 0.5066089882239846, "eval_loss": 2.6047279834747314, "eval_precision": 0.567609079444451, "eval_precision_DuraRiadoRio_16x16": 0.584213797389683, "eval_precision_Mole_16x16": 0.6187961985216474, "eval_precision_Quebrado_16x16": 0.7050754458161865, "eval_precision_RiadoRio_16x16": 0.3702148678686095, "eval_precision_RioFechado_16x16": 0.5597450876261285, "eval_recall": 0.5329978397486252, "eval_recall_DuraRiadoRio_16x16": 0.4079861111111111, "eval_recall_Mole_16x16": 0.5086805555555556, "eval_recall_Quebrado_16x16": 0.6692708333333334, "eval_recall_RiadoRio_16x16": 0.6163651315789473, "eval_recall_RioFechado_16x16": 0.4626865671641791, "eval_runtime": 45.7504, "eval_samples_per_second": 254.031, "eval_steps_per_second": 15.891, "step": 494020 }, { "epoch": 85.00516173434274, "grad_norm": 12.121907234191895, "learning_rate": 7.497419132828631e-06, "loss": 0.6704, "step": 494050 }, { "epoch": 85.01376462491397, "grad_norm": 9.939773559570312, "learning_rate": 7.493117687543015e-06, "loss": 0.6725, "step": 494100 }, { "epoch": 85.0223675154852, "grad_norm": 21.06436538696289, "learning_rate": 7.4888162422573985e-06, "loss": 0.6048, "step": 494150 }, { "epoch": 85.03097040605644, "grad_norm": 26.1251277923584, "learning_rate": 7.484514796971782e-06, "loss": 0.686, "step": 494200 }, { "epoch": 85.03957329662767, "grad_norm": 8.658251762390137, "learning_rate": 7.480213351686168e-06, "loss": 0.63, "step": 494250 }, { "epoch": 85.0481761871989, "grad_norm": 11.36697769165039, "learning_rate": 7.475911906400551e-06, "loss": 0.6435, "step": 494300 }, { "epoch": 85.05677907777013, "grad_norm": 15.742008209228516, "learning_rate": 7.471610461114935e-06, "loss": 0.6059, "step": 494350 }, { "epoch": 85.06538196834136, "grad_norm": 23.837217330932617, "learning_rate": 7.467309015829319e-06, "loss": 0.6851, "step": 494400 }, { "epoch": 85.0739848589126, "grad_norm": 13.877946853637695, "learning_rate": 7.463007570543703e-06, "loss": 0.6699, "step": 494450 }, { "epoch": 85.08258774948382, "grad_norm": 14.717090606689453, "learning_rate": 7.458706125258087e-06, "loss": 0.6179, "step": 494500 }, { "epoch": 85.09119064005506, "grad_norm": 11.868171691894531, "learning_rate": 7.454404679972471e-06, "loss": 0.5788, "step": 494550 }, { "epoch": 85.09979353062629, "grad_norm": 19.485021591186523, "learning_rate": 7.4501032346868545e-06, "loss": 0.639, "step": 494600 }, { "epoch": 85.10839642119753, "grad_norm": 11.085288047790527, "learning_rate": 7.44580178940124e-06, "loss": 0.5759, "step": 494650 }, { "epoch": 85.11699931176875, "grad_norm": 18.27659797668457, "learning_rate": 7.441500344115624e-06, "loss": 0.6533, "step": 494700 }, { "epoch": 85.12560220233999, "grad_norm": 7.8041229248046875, "learning_rate": 7.437198898830007e-06, "loss": 0.6711, "step": 494750 }, { "epoch": 85.13420509291122, "grad_norm": 11.049503326416016, "learning_rate": 7.432897453544391e-06, "loss": 0.6448, "step": 494800 }, { "epoch": 85.14280798348246, "grad_norm": 21.67627716064453, "learning_rate": 7.428596008258776e-06, "loss": 0.6284, "step": 494850 }, { "epoch": 85.15141087405368, "grad_norm": 13.728326797485352, "learning_rate": 7.424294562973159e-06, "loss": 0.6244, "step": 494900 }, { "epoch": 85.16001376462492, "grad_norm": 24.054689407348633, "learning_rate": 7.419993117687543e-06, "loss": 0.6449, "step": 494950 }, { "epoch": 85.16861665519615, "grad_norm": 12.264347076416016, "learning_rate": 7.415691672401927e-06, "loss": 0.7576, "step": 495000 }, { "epoch": 85.17721954576737, "grad_norm": 21.835906982421875, "learning_rate": 7.411390227116312e-06, "loss": 0.6417, "step": 495050 }, { "epoch": 85.18582243633861, "grad_norm": 14.848447799682617, "learning_rate": 7.407088781830696e-06, "loss": 0.6445, "step": 495100 }, { "epoch": 85.19442532690984, "grad_norm": 17.211286544799805, "learning_rate": 7.4027873365450796e-06, "loss": 0.6341, "step": 495150 }, { "epoch": 85.20302821748108, "grad_norm": 8.344683647155762, "learning_rate": 7.398485891259463e-06, "loss": 0.651, "step": 495200 }, { "epoch": 85.2116311080523, "grad_norm": 15.045486450195312, "learning_rate": 7.394184445973848e-06, "loss": 0.6861, "step": 495250 }, { "epoch": 85.22023399862354, "grad_norm": 15.439095497131348, "learning_rate": 7.3898830006882316e-06, "loss": 0.6625, "step": 495300 }, { "epoch": 85.22883688919477, "grad_norm": 17.65044403076172, "learning_rate": 7.385581555402615e-06, "loss": 0.628, "step": 495350 }, { "epoch": 85.237439779766, "grad_norm": 19.044832229614258, "learning_rate": 7.381280110116999e-06, "loss": 0.6671, "step": 495400 }, { "epoch": 85.24604267033723, "grad_norm": 17.346649169921875, "learning_rate": 7.376978664831384e-06, "loss": 0.7533, "step": 495450 }, { "epoch": 85.25464556090847, "grad_norm": 11.23167610168457, "learning_rate": 7.372677219545768e-06, "loss": 0.6574, "step": 495500 }, { "epoch": 85.2632484514797, "grad_norm": 22.901662826538086, "learning_rate": 7.368375774260152e-06, "loss": 0.669, "step": 495550 }, { "epoch": 85.27185134205092, "grad_norm": 22.64760971069336, "learning_rate": 7.364074328974536e-06, "loss": 0.6946, "step": 495600 }, { "epoch": 85.28045423262216, "grad_norm": 19.623701095581055, "learning_rate": 7.35977288368892e-06, "loss": 0.703, "step": 495650 }, { "epoch": 85.28905712319339, "grad_norm": 23.323970794677734, "learning_rate": 7.355471438403304e-06, "loss": 0.6961, "step": 495700 }, { "epoch": 85.29766001376463, "grad_norm": 10.11562442779541, "learning_rate": 7.3511699931176875e-06, "loss": 0.6559, "step": 495750 }, { "epoch": 85.30626290433585, "grad_norm": 16.6705265045166, "learning_rate": 7.346868547832073e-06, "loss": 0.6799, "step": 495800 }, { "epoch": 85.31486579490709, "grad_norm": 20.55484390258789, "learning_rate": 7.342567102546457e-06, "loss": 0.68, "step": 495850 }, { "epoch": 85.32346868547832, "grad_norm": 14.308670043945312, "learning_rate": 7.3382656572608395e-06, "loss": 0.6709, "step": 495900 }, { "epoch": 85.33207157604956, "grad_norm": 12.137628555297852, "learning_rate": 7.333964211975223e-06, "loss": 0.6573, "step": 495950 }, { "epoch": 85.34067446662078, "grad_norm": 11.444792747497559, "learning_rate": 7.329662766689609e-06, "loss": 0.6253, "step": 496000 }, { "epoch": 85.34927735719202, "grad_norm": 11.312138557434082, "learning_rate": 7.325361321403992e-06, "loss": 0.6867, "step": 496050 }, { "epoch": 85.35788024776325, "grad_norm": 17.33177947998047, "learning_rate": 7.321059876118376e-06, "loss": 0.6676, "step": 496100 }, { "epoch": 85.36648313833447, "grad_norm": 13.071491241455078, "learning_rate": 7.31675843083276e-06, "loss": 0.6809, "step": 496150 }, { "epoch": 85.37508602890571, "grad_norm": 12.909051895141602, "learning_rate": 7.312456985547145e-06, "loss": 0.6392, "step": 496200 }, { "epoch": 85.38368891947694, "grad_norm": 21.25800323486328, "learning_rate": 7.308155540261528e-06, "loss": 0.7131, "step": 496250 }, { "epoch": 85.39229181004818, "grad_norm": 19.877464294433594, "learning_rate": 7.303854094975912e-06, "loss": 0.6808, "step": 496300 }, { "epoch": 85.4008947006194, "grad_norm": 18.205839157104492, "learning_rate": 7.2995526496902955e-06, "loss": 0.6371, "step": 496350 }, { "epoch": 85.40949759119064, "grad_norm": 13.016493797302246, "learning_rate": 7.295251204404681e-06, "loss": 0.6477, "step": 496400 }, { "epoch": 85.41810048176187, "grad_norm": 24.732053756713867, "learning_rate": 7.290949759119065e-06, "loss": 0.7032, "step": 496450 }, { "epoch": 85.42670337233311, "grad_norm": 22.062068939208984, "learning_rate": 7.286648313833448e-06, "loss": 0.6833, "step": 496500 }, { "epoch": 85.43530626290433, "grad_norm": 27.36079978942871, "learning_rate": 7.282346868547832e-06, "loss": 0.6131, "step": 496550 }, { "epoch": 85.44390915347557, "grad_norm": 16.342737197875977, "learning_rate": 7.2780454232622166e-06, "loss": 0.6251, "step": 496600 }, { "epoch": 85.4525120440468, "grad_norm": 7.5021162033081055, "learning_rate": 7.2737439779766e-06, "loss": 0.6373, "step": 496650 }, { "epoch": 85.46111493461804, "grad_norm": 13.676824569702148, "learning_rate": 7.269442532690984e-06, "loss": 0.681, "step": 496700 }, { "epoch": 85.46971782518926, "grad_norm": 16.172428131103516, "learning_rate": 7.265141087405368e-06, "loss": 0.6232, "step": 496750 }, { "epoch": 85.47832071576049, "grad_norm": 21.883005142211914, "learning_rate": 7.260839642119753e-06, "loss": 0.7139, "step": 496800 }, { "epoch": 85.48692360633173, "grad_norm": 11.89913558959961, "learning_rate": 7.256538196834137e-06, "loss": 0.6884, "step": 496850 }, { "epoch": 85.49552649690295, "grad_norm": 13.400835037231445, "learning_rate": 7.2522367515485205e-06, "loss": 0.6593, "step": 496900 }, { "epoch": 85.5041293874742, "grad_norm": 16.362415313720703, "learning_rate": 7.247935306262904e-06, "loss": 0.6372, "step": 496950 }, { "epoch": 85.51273227804542, "grad_norm": 16.456125259399414, "learning_rate": 7.243633860977289e-06, "loss": 0.6632, "step": 497000 }, { "epoch": 85.52133516861666, "grad_norm": 10.758688926696777, "learning_rate": 7.2393324156916725e-06, "loss": 0.6552, "step": 497050 }, { "epoch": 85.52993805918788, "grad_norm": 19.65382957458496, "learning_rate": 7.235030970406056e-06, "loss": 0.6445, "step": 497100 }, { "epoch": 85.53854094975912, "grad_norm": 17.061946868896484, "learning_rate": 7.23072952512044e-06, "loss": 0.6478, "step": 497150 }, { "epoch": 85.54714384033035, "grad_norm": 15.167831420898438, "learning_rate": 7.226428079834825e-06, "loss": 0.6432, "step": 497200 }, { "epoch": 85.55574673090159, "grad_norm": 14.153757095336914, "learning_rate": 7.222126634549209e-06, "loss": 0.5964, "step": 497250 }, { "epoch": 85.56434962147281, "grad_norm": 20.324230194091797, "learning_rate": 7.217825189263593e-06, "loss": 0.7052, "step": 497300 }, { "epoch": 85.57295251204404, "grad_norm": 17.208457946777344, "learning_rate": 7.2135237439779765e-06, "loss": 0.6953, "step": 497350 }, { "epoch": 85.58155540261528, "grad_norm": 17.173891067504883, "learning_rate": 7.209222298692361e-06, "loss": 0.6433, "step": 497400 }, { "epoch": 85.5901582931865, "grad_norm": 27.04961395263672, "learning_rate": 7.204920853406745e-06, "loss": 0.6741, "step": 497450 }, { "epoch": 85.59876118375774, "grad_norm": 21.048025131225586, "learning_rate": 7.2006194081211285e-06, "loss": 0.6868, "step": 497500 }, { "epoch": 85.60736407432897, "grad_norm": 17.261119842529297, "learning_rate": 7.196317962835514e-06, "loss": 0.6442, "step": 497550 }, { "epoch": 85.61596696490021, "grad_norm": 13.412262916564941, "learning_rate": 7.192016517549898e-06, "loss": 0.6891, "step": 497600 }, { "epoch": 85.62456985547144, "grad_norm": 24.829273223876953, "learning_rate": 7.187715072264281e-06, "loss": 0.6695, "step": 497650 }, { "epoch": 85.63317274604267, "grad_norm": 11.41588306427002, "learning_rate": 7.183413626978665e-06, "loss": 0.6788, "step": 497700 }, { "epoch": 85.6417756366139, "grad_norm": 12.488724708557129, "learning_rate": 7.17911218169305e-06, "loss": 0.6975, "step": 497750 }, { "epoch": 85.65037852718514, "grad_norm": 18.445688247680664, "learning_rate": 7.174810736407433e-06, "loss": 0.6107, "step": 497800 }, { "epoch": 85.65898141775637, "grad_norm": 18.521739959716797, "learning_rate": 7.170509291121817e-06, "loss": 0.6678, "step": 497850 }, { "epoch": 85.6675843083276, "grad_norm": 14.982516288757324, "learning_rate": 7.166207845836201e-06, "loss": 0.6402, "step": 497900 }, { "epoch": 85.67618719889883, "grad_norm": 14.87734317779541, "learning_rate": 7.161906400550586e-06, "loss": 0.6571, "step": 497950 }, { "epoch": 85.68479008947006, "grad_norm": 16.076461791992188, "learning_rate": 7.15760495526497e-06, "loss": 0.7023, "step": 498000 }, { "epoch": 85.6933929800413, "grad_norm": 16.430400848388672, "learning_rate": 7.1533035099793536e-06, "loss": 0.6451, "step": 498050 }, { "epoch": 85.70199587061252, "grad_norm": 15.442585945129395, "learning_rate": 7.149002064693737e-06, "loss": 0.6284, "step": 498100 }, { "epoch": 85.71059876118376, "grad_norm": 14.557232856750488, "learning_rate": 7.144700619408122e-06, "loss": 0.6728, "step": 498150 }, { "epoch": 85.71920165175499, "grad_norm": 13.648591041564941, "learning_rate": 7.1403991741225056e-06, "loss": 0.662, "step": 498200 }, { "epoch": 85.72780454232623, "grad_norm": 17.537797927856445, "learning_rate": 7.136097728836889e-06, "loss": 0.6818, "step": 498250 }, { "epoch": 85.73640743289745, "grad_norm": 13.178910255432129, "learning_rate": 7.131796283551273e-06, "loss": 0.6745, "step": 498300 }, { "epoch": 85.74501032346869, "grad_norm": 11.226384162902832, "learning_rate": 7.127494838265658e-06, "loss": 0.6952, "step": 498350 }, { "epoch": 85.75361321403992, "grad_norm": 11.534028053283691, "learning_rate": 7.123193392980042e-06, "loss": 0.6147, "step": 498400 }, { "epoch": 85.76221610461116, "grad_norm": 10.397270202636719, "learning_rate": 7.118891947694426e-06, "loss": 0.6208, "step": 498450 }, { "epoch": 85.77081899518238, "grad_norm": 9.433154106140137, "learning_rate": 7.1145905024088095e-06, "loss": 0.6794, "step": 498500 }, { "epoch": 85.7794218857536, "grad_norm": 15.385808944702148, "learning_rate": 7.110289057123194e-06, "loss": 0.7195, "step": 498550 }, { "epoch": 85.78802477632485, "grad_norm": 16.232572555541992, "learning_rate": 7.105987611837578e-06, "loss": 0.6493, "step": 498600 }, { "epoch": 85.79662766689607, "grad_norm": 6.919379234313965, "learning_rate": 7.1016861665519615e-06, "loss": 0.6539, "step": 498650 }, { "epoch": 85.80523055746731, "grad_norm": 13.761364936828613, "learning_rate": 7.097384721266345e-06, "loss": 0.6377, "step": 498700 }, { "epoch": 85.81383344803854, "grad_norm": 7.388372898101807, "learning_rate": 7.093083275980731e-06, "loss": 0.6208, "step": 498750 }, { "epoch": 85.82243633860978, "grad_norm": 12.166308403015137, "learning_rate": 7.088781830695114e-06, "loss": 0.7306, "step": 498800 }, { "epoch": 85.831039229181, "grad_norm": 12.773713111877441, "learning_rate": 7.084480385409498e-06, "loss": 0.6019, "step": 498850 }, { "epoch": 85.83964211975224, "grad_norm": 16.827171325683594, "learning_rate": 7.080178940123881e-06, "loss": 0.6113, "step": 498900 }, { "epoch": 85.84824501032347, "grad_norm": 10.46429443359375, "learning_rate": 7.075877494838266e-06, "loss": 0.6589, "step": 498950 }, { "epoch": 85.8568479008947, "grad_norm": 14.500535011291504, "learning_rate": 7.07157604955265e-06, "loss": 0.6661, "step": 499000 }, { "epoch": 85.86545079146593, "grad_norm": 38.701080322265625, "learning_rate": 7.067274604267034e-06, "loss": 0.6843, "step": 499050 }, { "epoch": 85.87405368203716, "grad_norm": 20.640729904174805, "learning_rate": 7.0629731589814175e-06, "loss": 0.6503, "step": 499100 }, { "epoch": 85.8826565726084, "grad_norm": 13.235135078430176, "learning_rate": 7.058671713695803e-06, "loss": 0.6177, "step": 499150 }, { "epoch": 85.89125946317962, "grad_norm": 14.648286819458008, "learning_rate": 7.054370268410186e-06, "loss": 0.6684, "step": 499200 }, { "epoch": 85.89986235375086, "grad_norm": 10.645027160644531, "learning_rate": 7.0500688231245695e-06, "loss": 0.6544, "step": 499250 }, { "epoch": 85.90846524432209, "grad_norm": 11.816900253295898, "learning_rate": 7.045767377838955e-06, "loss": 0.6548, "step": 499300 }, { "epoch": 85.91706813489333, "grad_norm": 13.552079200744629, "learning_rate": 7.041465932553339e-06, "loss": 0.6969, "step": 499350 }, { "epoch": 85.92567102546455, "grad_norm": 31.642078399658203, "learning_rate": 7.037164487267722e-06, "loss": 0.6309, "step": 499400 }, { "epoch": 85.93427391603579, "grad_norm": 21.909788131713867, "learning_rate": 7.032863041982106e-06, "loss": 0.657, "step": 499450 }, { "epoch": 85.94287680660702, "grad_norm": 15.933046340942383, "learning_rate": 7.028561596696491e-06, "loss": 0.6978, "step": 499500 }, { "epoch": 85.95147969717826, "grad_norm": 22.22858428955078, "learning_rate": 7.024260151410874e-06, "loss": 0.678, "step": 499550 }, { "epoch": 85.96008258774948, "grad_norm": 9.873037338256836, "learning_rate": 7.019958706125258e-06, "loss": 0.6558, "step": 499600 }, { "epoch": 85.96868547832072, "grad_norm": 17.157350540161133, "learning_rate": 7.015657260839642e-06, "loss": 0.6261, "step": 499650 }, { "epoch": 85.97728836889195, "grad_norm": 18.4902400970459, "learning_rate": 7.011355815554027e-06, "loss": 0.6806, "step": 499700 }, { "epoch": 85.98589125946317, "grad_norm": 12.531737327575684, "learning_rate": 7.007054370268411e-06, "loss": 0.6279, "step": 499750 }, { "epoch": 85.99449415003441, "grad_norm": 20.104318618774414, "learning_rate": 7.0027529249827946e-06, "loss": 0.6743, "step": 499800 }, { "epoch": 86.0, "eval_accuracy": 0.5811392187231114, "eval_f1": 0.5653454835621985, "eval_f1_DuraRiadoRio_16x16": 0.5112208258527827, "eval_f1_Mole_16x16": 0.625, "eval_f1_Quebrado_16x16": 0.7303740238388821, "eval_f1_RiadoRio_16x16": 0.3511360283269401, "eval_f1_RioFechado_16x16": 0.6089965397923875, "eval_loss": 2.3569395542144775, "eval_precision": 0.5955824177900833, "eval_precision_DuraRiadoRio_16x16": 0.529275092936803, "eval_precision_Mole_16x16": 0.6458333333333334, "eval_precision_Quebrado_16x16": 0.6935987509758001, "eval_precision_RiadoRio_16x16": 0.6217345872518286, "eval_precision_RioFechado_16x16": 0.48747032445265104, "eval_recall": 0.5853972566540364, "eval_recall_DuraRiadoRio_16x16": 0.4943576388888889, "eval_recall_Mole_16x16": 0.60546875, "eval_recall_Quebrado_16x16": 0.7712673611111112, "eval_recall_RiadoRio_16x16": 0.24465460526315788, "eval_recall_RioFechado_16x16": 0.8112379280070237, "eval_runtime": 45.4142, "eval_samples_per_second": 255.911, "eval_steps_per_second": 16.008, "step": 499832 }, { "epoch": 86.00309704060564, "grad_norm": 11.367704391479492, "learning_rate": 6.998451479697178e-06, "loss": 0.6573, "step": 499850 }, { "epoch": 86.01169993117688, "grad_norm": 19.07600212097168, "learning_rate": 6.994150034411563e-06, "loss": 0.6794, "step": 499900 }, { "epoch": 86.0203028217481, "grad_norm": 21.339250564575195, "learning_rate": 6.9898485891259465e-06, "loss": 0.6184, "step": 499950 }, { "epoch": 86.02890571231934, "grad_norm": 14.464712142944336, "learning_rate": 6.98554714384033e-06, "loss": 0.6879, "step": 500000 }, { "epoch": 86.03750860289057, "grad_norm": 20.26146697998047, "learning_rate": 6.981245698554714e-06, "loss": 0.6487, "step": 500050 }, { "epoch": 86.04611149346181, "grad_norm": 23.602930068969727, "learning_rate": 6.976944253269099e-06, "loss": 0.6798, "step": 500100 }, { "epoch": 86.05471438403303, "grad_norm": 12.821354866027832, "learning_rate": 6.972642807983483e-06, "loss": 0.648, "step": 500150 }, { "epoch": 86.06331727460427, "grad_norm": 15.891075134277344, "learning_rate": 6.968341362697867e-06, "loss": 0.6361, "step": 500200 }, { "epoch": 86.0719201651755, "grad_norm": 6.146921157836914, "learning_rate": 6.9640399174122505e-06, "loss": 0.6562, "step": 500250 }, { "epoch": 86.08052305574672, "grad_norm": 21.705991744995117, "learning_rate": 6.959738472126635e-06, "loss": 0.6335, "step": 500300 }, { "epoch": 86.08912594631796, "grad_norm": 13.26646614074707, "learning_rate": 6.955437026841019e-06, "loss": 0.6315, "step": 500350 }, { "epoch": 86.09772883688919, "grad_norm": 17.005796432495117, "learning_rate": 6.9511355815554025e-06, "loss": 0.6212, "step": 500400 }, { "epoch": 86.10633172746043, "grad_norm": 15.242807388305664, "learning_rate": 6.946834136269786e-06, "loss": 0.6971, "step": 500450 }, { "epoch": 86.11493461803165, "grad_norm": 16.591007232666016, "learning_rate": 6.942532690984172e-06, "loss": 0.6693, "step": 500500 }, { "epoch": 86.1235375086029, "grad_norm": 17.258052825927734, "learning_rate": 6.938231245698555e-06, "loss": 0.682, "step": 500550 }, { "epoch": 86.13214039917412, "grad_norm": 14.609521865844727, "learning_rate": 6.933929800412939e-06, "loss": 0.6111, "step": 500600 }, { "epoch": 86.14074328974536, "grad_norm": 9.750370979309082, "learning_rate": 6.929628355127323e-06, "loss": 0.6336, "step": 500650 }, { "epoch": 86.14934618031658, "grad_norm": 24.297367095947266, "learning_rate": 6.925326909841707e-06, "loss": 0.6206, "step": 500700 }, { "epoch": 86.15794907088782, "grad_norm": 12.549568176269531, "learning_rate": 6.921025464556091e-06, "loss": 0.6316, "step": 500750 }, { "epoch": 86.16655196145905, "grad_norm": 11.406854629516602, "learning_rate": 6.916724019270475e-06, "loss": 0.6786, "step": 500800 }, { "epoch": 86.17515485203029, "grad_norm": 16.29994010925293, "learning_rate": 6.9124225739848585e-06, "loss": 0.6814, "step": 500850 }, { "epoch": 86.18375774260151, "grad_norm": 14.845381736755371, "learning_rate": 6.908121128699244e-06, "loss": 0.681, "step": 500900 }, { "epoch": 86.19236063317274, "grad_norm": 12.40816879272461, "learning_rate": 6.903819683413628e-06, "loss": 0.6283, "step": 500950 }, { "epoch": 86.20096352374398, "grad_norm": 15.203371047973633, "learning_rate": 6.899518238128011e-06, "loss": 0.6656, "step": 501000 }, { "epoch": 86.2095664143152, "grad_norm": 13.408097267150879, "learning_rate": 6.895216792842395e-06, "loss": 0.626, "step": 501050 }, { "epoch": 86.21816930488644, "grad_norm": 9.157877922058105, "learning_rate": 6.8909153475567796e-06, "loss": 0.6738, "step": 501100 }, { "epoch": 86.22677219545767, "grad_norm": 13.456871032714844, "learning_rate": 6.886613902271163e-06, "loss": 0.683, "step": 501150 }, { "epoch": 86.23537508602891, "grad_norm": 15.169949531555176, "learning_rate": 6.882312456985547e-06, "loss": 0.6613, "step": 501200 }, { "epoch": 86.24397797660014, "grad_norm": 31.953195571899414, "learning_rate": 6.878011011699932e-06, "loss": 0.6932, "step": 501250 }, { "epoch": 86.25258086717137, "grad_norm": 22.82187843322754, "learning_rate": 6.873709566414316e-06, "loss": 0.6771, "step": 501300 }, { "epoch": 86.2611837577426, "grad_norm": 11.480945587158203, "learning_rate": 6.8694081211287e-06, "loss": 0.7013, "step": 501350 }, { "epoch": 86.26978664831384, "grad_norm": 16.15761947631836, "learning_rate": 6.8651066758430835e-06, "loss": 0.7037, "step": 501400 }, { "epoch": 86.27838953888507, "grad_norm": 14.572465896606445, "learning_rate": 6.860805230557468e-06, "loss": 0.7052, "step": 501450 }, { "epoch": 86.28699242945629, "grad_norm": 12.010329246520996, "learning_rate": 6.856503785271852e-06, "loss": 0.6395, "step": 501500 }, { "epoch": 86.29559532002753, "grad_norm": 27.602046966552734, "learning_rate": 6.8522023399862355e-06, "loss": 0.6957, "step": 501550 }, { "epoch": 86.30419821059876, "grad_norm": 18.015987396240234, "learning_rate": 6.847900894700619e-06, "loss": 0.6506, "step": 501600 }, { "epoch": 86.31280110117, "grad_norm": 21.663311004638672, "learning_rate": 6.843599449415005e-06, "loss": 0.6523, "step": 501650 }, { "epoch": 86.32140399174122, "grad_norm": 15.133079528808594, "learning_rate": 6.839298004129388e-06, "loss": 0.606, "step": 501700 }, { "epoch": 86.33000688231246, "grad_norm": 21.11598014831543, "learning_rate": 6.834996558843772e-06, "loss": 0.6486, "step": 501750 }, { "epoch": 86.33860977288369, "grad_norm": 14.089672088623047, "learning_rate": 6.830695113558156e-06, "loss": 0.5761, "step": 501800 }, { "epoch": 86.34721266345493, "grad_norm": 15.819223403930664, "learning_rate": 6.82639366827254e-06, "loss": 0.6597, "step": 501850 }, { "epoch": 86.35581555402615, "grad_norm": 14.618165016174316, "learning_rate": 6.822092222986924e-06, "loss": 0.6168, "step": 501900 }, { "epoch": 86.36441844459739, "grad_norm": 22.194107055664062, "learning_rate": 6.817790777701308e-06, "loss": 0.627, "step": 501950 }, { "epoch": 86.37302133516862, "grad_norm": 14.307670593261719, "learning_rate": 6.8134893324156915e-06, "loss": 0.6863, "step": 502000 }, { "epoch": 86.38162422573986, "grad_norm": 16.68120574951172, "learning_rate": 6.809187887130077e-06, "loss": 0.6434, "step": 502050 }, { "epoch": 86.39022711631108, "grad_norm": 15.373997688293457, "learning_rate": 6.804886441844461e-06, "loss": 0.6563, "step": 502100 }, { "epoch": 86.3988300068823, "grad_norm": 8.967508316040039, "learning_rate": 6.8005849965588435e-06, "loss": 0.6009, "step": 502150 }, { "epoch": 86.40743289745355, "grad_norm": 22.73145294189453, "learning_rate": 6.796283551273227e-06, "loss": 0.6459, "step": 502200 }, { "epoch": 86.41603578802477, "grad_norm": 12.130484580993652, "learning_rate": 6.791982105987613e-06, "loss": 0.6331, "step": 502250 }, { "epoch": 86.42463867859601, "grad_norm": 14.867904663085938, "learning_rate": 6.787680660701996e-06, "loss": 0.663, "step": 502300 }, { "epoch": 86.43324156916724, "grad_norm": 19.212265014648438, "learning_rate": 6.78337921541638e-06, "loss": 0.6759, "step": 502350 }, { "epoch": 86.44184445973848, "grad_norm": 10.059684753417969, "learning_rate": 6.779077770130764e-06, "loss": 0.667, "step": 502400 }, { "epoch": 86.4504473503097, "grad_norm": 17.156442642211914, "learning_rate": 6.774776324845149e-06, "loss": 0.6738, "step": 502450 }, { "epoch": 86.45905024088094, "grad_norm": 19.22283935546875, "learning_rate": 6.770474879559532e-06, "loss": 0.7064, "step": 502500 }, { "epoch": 86.46765313145217, "grad_norm": 11.230630874633789, "learning_rate": 6.766173434273916e-06, "loss": 0.6407, "step": 502550 }, { "epoch": 86.4762560220234, "grad_norm": 13.6318359375, "learning_rate": 6.7618719889882994e-06, "loss": 0.6947, "step": 502600 }, { "epoch": 86.48485891259463, "grad_norm": 14.928206443786621, "learning_rate": 6.757570543702685e-06, "loss": 0.6444, "step": 502650 }, { "epoch": 86.49346180316586, "grad_norm": 9.459137916564941, "learning_rate": 6.7532690984170686e-06, "loss": 0.6683, "step": 502700 }, { "epoch": 86.5020646937371, "grad_norm": 14.469294548034668, "learning_rate": 6.748967653131452e-06, "loss": 0.6769, "step": 502750 }, { "epoch": 86.51066758430832, "grad_norm": 10.79570484161377, "learning_rate": 6.744666207845836e-06, "loss": 0.6024, "step": 502800 }, { "epoch": 86.51927047487956, "grad_norm": 7.178287982940674, "learning_rate": 6.7403647625602206e-06, "loss": 0.6409, "step": 502850 }, { "epoch": 86.52787336545079, "grad_norm": 26.560056686401367, "learning_rate": 6.736063317274604e-06, "loss": 0.6712, "step": 502900 }, { "epoch": 86.53647625602203, "grad_norm": 11.014328956604004, "learning_rate": 6.731761871988988e-06, "loss": 0.6717, "step": 502950 }, { "epoch": 86.54507914659325, "grad_norm": 17.639156341552734, "learning_rate": 6.727460426703373e-06, "loss": 0.6102, "step": 503000 }, { "epoch": 86.55368203716449, "grad_norm": 15.322061538696289, "learning_rate": 6.723158981417757e-06, "loss": 0.6916, "step": 503050 }, { "epoch": 86.56228492773572, "grad_norm": 16.042877197265625, "learning_rate": 6.718857536132141e-06, "loss": 0.6694, "step": 503100 }, { "epoch": 86.57088781830696, "grad_norm": 9.498831748962402, "learning_rate": 6.7145560908465245e-06, "loss": 0.5917, "step": 503150 }, { "epoch": 86.57949070887818, "grad_norm": 15.746196746826172, "learning_rate": 6.710254645560909e-06, "loss": 0.6457, "step": 503200 }, { "epoch": 86.58809359944941, "grad_norm": 11.558852195739746, "learning_rate": 6.705953200275293e-06, "loss": 0.6727, "step": 503250 }, { "epoch": 86.59669649002065, "grad_norm": 12.171392440795898, "learning_rate": 6.7016517549896765e-06, "loss": 0.6833, "step": 503300 }, { "epoch": 86.60529938059187, "grad_norm": 25.378250122070312, "learning_rate": 6.69735030970406e-06, "loss": 0.6766, "step": 503350 }, { "epoch": 86.61390227116311, "grad_norm": 13.998470306396484, "learning_rate": 6.693048864418446e-06, "loss": 0.6303, "step": 503400 }, { "epoch": 86.62250516173434, "grad_norm": 16.736839294433594, "learning_rate": 6.688747419132829e-06, "loss": 0.6344, "step": 503450 }, { "epoch": 86.63110805230558, "grad_norm": 10.015517234802246, "learning_rate": 6.684445973847213e-06, "loss": 0.633, "step": 503500 }, { "epoch": 86.6397109428768, "grad_norm": 11.33863353729248, "learning_rate": 6.680144528561597e-06, "loss": 0.5963, "step": 503550 }, { "epoch": 86.64831383344804, "grad_norm": 16.80500030517578, "learning_rate": 6.675843083275981e-06, "loss": 0.5492, "step": 503600 }, { "epoch": 86.65691672401927, "grad_norm": 19.788938522338867, "learning_rate": 6.671541637990365e-06, "loss": 0.6334, "step": 503650 }, { "epoch": 86.66551961459051, "grad_norm": 18.708396911621094, "learning_rate": 6.667240192704749e-06, "loss": 0.6686, "step": 503700 }, { "epoch": 86.67412250516173, "grad_norm": 15.688907623291016, "learning_rate": 6.6629387474191325e-06, "loss": 0.6167, "step": 503750 }, { "epoch": 86.68272539573297, "grad_norm": 29.043485641479492, "learning_rate": 6.658637302133518e-06, "loss": 0.6653, "step": 503800 }, { "epoch": 86.6913282863042, "grad_norm": 18.231557846069336, "learning_rate": 6.654335856847902e-06, "loss": 0.6519, "step": 503850 }, { "epoch": 86.69993117687542, "grad_norm": 28.431913375854492, "learning_rate": 6.650034411562285e-06, "loss": 0.6681, "step": 503900 }, { "epoch": 86.70853406744666, "grad_norm": 16.414077758789062, "learning_rate": 6.645732966276669e-06, "loss": 0.6128, "step": 503950 }, { "epoch": 86.71713695801789, "grad_norm": 18.964454650878906, "learning_rate": 6.641431520991054e-06, "loss": 0.6625, "step": 504000 }, { "epoch": 86.72573984858913, "grad_norm": 20.85394859313965, "learning_rate": 6.637130075705437e-06, "loss": 0.6438, "step": 504050 }, { "epoch": 86.73434273916035, "grad_norm": 15.118297576904297, "learning_rate": 6.632828630419821e-06, "loss": 0.6781, "step": 504100 }, { "epoch": 86.7429456297316, "grad_norm": 8.738155364990234, "learning_rate": 6.628527185134205e-06, "loss": 0.6121, "step": 504150 }, { "epoch": 86.75154852030282, "grad_norm": 19.51012420654297, "learning_rate": 6.62422573984859e-06, "loss": 0.6178, "step": 504200 }, { "epoch": 86.76015141087406, "grad_norm": 23.82736587524414, "learning_rate": 6.619924294562974e-06, "loss": 0.6009, "step": 504250 }, { "epoch": 86.76875430144528, "grad_norm": 17.02295684814453, "learning_rate": 6.6156228492773576e-06, "loss": 0.6314, "step": 504300 }, { "epoch": 86.77735719201652, "grad_norm": 17.74040412902832, "learning_rate": 6.611321403991741e-06, "loss": 0.6244, "step": 504350 }, { "epoch": 86.78596008258775, "grad_norm": 26.099395751953125, "learning_rate": 6.607019958706126e-06, "loss": 0.6878, "step": 504400 }, { "epoch": 86.79456297315897, "grad_norm": 9.526656150817871, "learning_rate": 6.6027185134205095e-06, "loss": 0.6493, "step": 504450 }, { "epoch": 86.80316586373021, "grad_norm": 12.68276309967041, "learning_rate": 6.598417068134893e-06, "loss": 0.6573, "step": 504500 }, { "epoch": 86.81176875430144, "grad_norm": 9.962464332580566, "learning_rate": 6.594115622849277e-06, "loss": 0.6352, "step": 504550 }, { "epoch": 86.82037164487268, "grad_norm": 17.761180877685547, "learning_rate": 6.589814177563662e-06, "loss": 0.6909, "step": 504600 }, { "epoch": 86.8289745354439, "grad_norm": 17.77806854248047, "learning_rate": 6.585512732278046e-06, "loss": 0.6531, "step": 504650 }, { "epoch": 86.83757742601514, "grad_norm": 23.532995223999023, "learning_rate": 6.58121128699243e-06, "loss": 0.7352, "step": 504700 }, { "epoch": 86.84618031658637, "grad_norm": 14.864411354064941, "learning_rate": 6.5769098417068135e-06, "loss": 0.6223, "step": 504750 }, { "epoch": 86.85478320715761, "grad_norm": 26.739431381225586, "learning_rate": 6.572608396421198e-06, "loss": 0.6856, "step": 504800 }, { "epoch": 86.86338609772884, "grad_norm": 23.073440551757812, "learning_rate": 6.568306951135582e-06, "loss": 0.65, "step": 504850 }, { "epoch": 86.87198898830007, "grad_norm": 13.516006469726562, "learning_rate": 6.5640055058499655e-06, "loss": 0.6512, "step": 504900 }, { "epoch": 86.8805918788713, "grad_norm": 18.625045776367188, "learning_rate": 6.559704060564351e-06, "loss": 0.6371, "step": 504950 }, { "epoch": 86.88919476944254, "grad_norm": 14.627379417419434, "learning_rate": 6.555402615278735e-06, "loss": 0.6405, "step": 505000 }, { "epoch": 86.89779766001377, "grad_norm": 27.697750091552734, "learning_rate": 6.551101169993118e-06, "loss": 0.696, "step": 505050 }, { "epoch": 86.90640055058499, "grad_norm": 10.542112350463867, "learning_rate": 6.546799724707502e-06, "loss": 0.6627, "step": 505100 }, { "epoch": 86.91500344115623, "grad_norm": 23.14087677001953, "learning_rate": 6.542498279421887e-06, "loss": 0.6614, "step": 505150 }, { "epoch": 86.92360633172746, "grad_norm": 20.68804931640625, "learning_rate": 6.53819683413627e-06, "loss": 0.6667, "step": 505200 }, { "epoch": 86.9322092222987, "grad_norm": 21.798913955688477, "learning_rate": 6.533895388850654e-06, "loss": 0.6333, "step": 505250 }, { "epoch": 86.94081211286992, "grad_norm": 17.839996337890625, "learning_rate": 6.529593943565038e-06, "loss": 0.5916, "step": 505300 }, { "epoch": 86.94941500344116, "grad_norm": 20.623525619506836, "learning_rate": 6.525292498279423e-06, "loss": 0.6733, "step": 505350 }, { "epoch": 86.95801789401239, "grad_norm": 14.290094375610352, "learning_rate": 6.520991052993807e-06, "loss": 0.629, "step": 505400 }, { "epoch": 86.96662078458363, "grad_norm": 18.839263916015625, "learning_rate": 6.51668960770819e-06, "loss": 0.6578, "step": 505450 }, { "epoch": 86.97522367515485, "grad_norm": 11.260491371154785, "learning_rate": 6.5123881624225735e-06, "loss": 0.7046, "step": 505500 }, { "epoch": 86.98382656572609, "grad_norm": 36.90781784057617, "learning_rate": 6.508086717136959e-06, "loss": 0.6588, "step": 505550 }, { "epoch": 86.99242945629732, "grad_norm": 15.939160346984863, "learning_rate": 6.5037852718513426e-06, "loss": 0.6588, "step": 505600 }, { "epoch": 87.0, "eval_accuracy": 0.5947341249354672, "eval_f1": 0.5895927526181944, "eval_f1_DuraRiadoRio_16x16": 0.5905737704918033, "eval_f1_Mole_16x16": 0.5272108843537415, "eval_f1_Quebrado_16x16": 0.7478260869565218, "eval_f1_RiadoRio_16x16": 0.47718322909427846, "eval_f1_RioFechado_16x16": 0.6051697921946275, "eval_loss": 2.120391368865967, "eval_precision": 0.6213288359553578, "eval_precision_DuraRiadoRio_16x16": 0.5593944099378882, "eval_precision_Mole_16x16": 0.7598039215686274, "eval_precision_Quebrado_16x16": 0.7491289198606271, "eval_precision_RiadoRio_16x16": 0.5464190981432361, "eval_precision_RioFechado_16x16": 0.4918978302664103, "eval_recall": 0.5970686709319758, "eval_recall_DuraRiadoRio_16x16": 0.6254340277777778, "eval_recall_Mole_16x16": 0.4036458333333333, "eval_recall_Quebrado_16x16": 0.7465277777777778, "eval_recall_RiadoRio_16x16": 0.42351973684210525, "eval_recall_RioFechado_16x16": 0.786215978928885, "eval_runtime": 46.3215, "eval_samples_per_second": 250.898, "eval_steps_per_second": 15.695, "step": 505644 }, { "epoch": 87.00103234686854, "grad_norm": 13.706893920898438, "learning_rate": 6.499483826565726e-06, "loss": 0.6248, "step": 505650 }, { "epoch": 87.00963523743978, "grad_norm": 16.5247745513916, "learning_rate": 6.49518238128011e-06, "loss": 0.6549, "step": 505700 }, { "epoch": 87.018238128011, "grad_norm": 13.631352424621582, "learning_rate": 6.4908809359944946e-06, "loss": 0.6748, "step": 505750 }, { "epoch": 87.02684101858225, "grad_norm": 18.84281349182129, "learning_rate": 6.486579490708878e-06, "loss": 0.6046, "step": 505800 }, { "epoch": 87.03544390915347, "grad_norm": 17.099449157714844, "learning_rate": 6.482278045423262e-06, "loss": 0.6449, "step": 505850 }, { "epoch": 87.04404679972471, "grad_norm": 12.21634578704834, "learning_rate": 6.477976600137646e-06, "loss": 0.6264, "step": 505900 }, { "epoch": 87.05264969029594, "grad_norm": 23.478200912475586, "learning_rate": 6.473675154852031e-06, "loss": 0.5939, "step": 505950 }, { "epoch": 87.06125258086718, "grad_norm": 9.25395679473877, "learning_rate": 6.469373709566415e-06, "loss": 0.6316, "step": 506000 }, { "epoch": 87.0698554714384, "grad_norm": 14.077034950256348, "learning_rate": 6.4650722642807985e-06, "loss": 0.6561, "step": 506050 }, { "epoch": 87.07845836200964, "grad_norm": 12.664212226867676, "learning_rate": 6.460770818995182e-06, "loss": 0.6881, "step": 506100 }, { "epoch": 87.08706125258087, "grad_norm": 16.553783416748047, "learning_rate": 6.456469373709567e-06, "loss": 0.6776, "step": 506150 }, { "epoch": 87.09566414315209, "grad_norm": 16.327436447143555, "learning_rate": 6.4521679284239505e-06, "loss": 0.6537, "step": 506200 }, { "epoch": 87.10426703372333, "grad_norm": 17.465822219848633, "learning_rate": 6.447866483138334e-06, "loss": 0.5855, "step": 506250 }, { "epoch": 87.11286992429456, "grad_norm": 12.334312438964844, "learning_rate": 6.443565037852718e-06, "loss": 0.6431, "step": 506300 }, { "epoch": 87.1214728148658, "grad_norm": 19.59227752685547, "learning_rate": 6.439263592567103e-06, "loss": 0.6841, "step": 506350 }, { "epoch": 87.13007570543702, "grad_norm": 28.261213302612305, "learning_rate": 6.434962147281487e-06, "loss": 0.6907, "step": 506400 }, { "epoch": 87.13867859600826, "grad_norm": 22.597393035888672, "learning_rate": 6.430660701995871e-06, "loss": 0.6513, "step": 506450 }, { "epoch": 87.14728148657949, "grad_norm": 19.215415954589844, "learning_rate": 6.4263592567102545e-06, "loss": 0.6813, "step": 506500 }, { "epoch": 87.15588437715073, "grad_norm": 17.2226505279541, "learning_rate": 6.422057811424639e-06, "loss": 0.677, "step": 506550 }, { "epoch": 87.16448726772195, "grad_norm": 14.331602096557617, "learning_rate": 6.417756366139023e-06, "loss": 0.6632, "step": 506600 }, { "epoch": 87.17309015829319, "grad_norm": 16.250944137573242, "learning_rate": 6.4134549208534065e-06, "loss": 0.6674, "step": 506650 }, { "epoch": 87.18169304886442, "grad_norm": 14.23869800567627, "learning_rate": 6.409153475567792e-06, "loss": 0.6524, "step": 506700 }, { "epoch": 87.19029593943566, "grad_norm": 19.132165908813477, "learning_rate": 6.404852030282176e-06, "loss": 0.649, "step": 506750 }, { "epoch": 87.19889883000688, "grad_norm": 9.911375999450684, "learning_rate": 6.400550584996559e-06, "loss": 0.6564, "step": 506800 }, { "epoch": 87.20750172057811, "grad_norm": 15.162425994873047, "learning_rate": 6.396249139710943e-06, "loss": 0.6267, "step": 506850 }, { "epoch": 87.21610461114935, "grad_norm": 12.433995246887207, "learning_rate": 6.391947694425328e-06, "loss": 0.6013, "step": 506900 }, { "epoch": 87.22470750172057, "grad_norm": 12.003435134887695, "learning_rate": 6.387646249139711e-06, "loss": 0.6824, "step": 506950 }, { "epoch": 87.23331039229181, "grad_norm": 18.087268829345703, "learning_rate": 6.383344803854095e-06, "loss": 0.6486, "step": 507000 }, { "epoch": 87.24191328286304, "grad_norm": 16.997846603393555, "learning_rate": 6.379043358568479e-06, "loss": 0.673, "step": 507050 }, { "epoch": 87.25051617343428, "grad_norm": 15.340096473693848, "learning_rate": 6.374741913282864e-06, "loss": 0.6576, "step": 507100 }, { "epoch": 87.2591190640055, "grad_norm": 18.279537200927734, "learning_rate": 6.370440467997248e-06, "loss": 0.6319, "step": 507150 }, { "epoch": 87.26772195457674, "grad_norm": 17.06557846069336, "learning_rate": 6.3661390227116316e-06, "loss": 0.646, "step": 507200 }, { "epoch": 87.27632484514797, "grad_norm": 11.195451736450195, "learning_rate": 6.361837577426015e-06, "loss": 0.6714, "step": 507250 }, { "epoch": 87.28492773571921, "grad_norm": 24.553863525390625, "learning_rate": 6.3575361321404e-06, "loss": 0.6475, "step": 507300 }, { "epoch": 87.29353062629043, "grad_norm": 18.231801986694336, "learning_rate": 6.3532346868547836e-06, "loss": 0.651, "step": 507350 }, { "epoch": 87.30213351686166, "grad_norm": 22.40170669555664, "learning_rate": 6.348933241569167e-06, "loss": 0.6472, "step": 507400 }, { "epoch": 87.3107364074329, "grad_norm": 9.904083251953125, "learning_rate": 6.344631796283551e-06, "loss": 0.6547, "step": 507450 }, { "epoch": 87.31933929800412, "grad_norm": 17.186416625976562, "learning_rate": 6.340330350997936e-06, "loss": 0.6979, "step": 507500 }, { "epoch": 87.32794218857536, "grad_norm": 22.747119903564453, "learning_rate": 6.33602890571232e-06, "loss": 0.6851, "step": 507550 }, { "epoch": 87.33654507914659, "grad_norm": 19.55828094482422, "learning_rate": 6.331727460426704e-06, "loss": 0.6622, "step": 507600 }, { "epoch": 87.34514796971783, "grad_norm": 12.977876663208008, "learning_rate": 6.3274260151410875e-06, "loss": 0.6831, "step": 507650 }, { "epoch": 87.35375086028905, "grad_norm": 12.371395111083984, "learning_rate": 6.323124569855472e-06, "loss": 0.6746, "step": 507700 }, { "epoch": 87.3623537508603, "grad_norm": 20.75403594970703, "learning_rate": 6.318823124569856e-06, "loss": 0.669, "step": 507750 }, { "epoch": 87.37095664143152, "grad_norm": 17.954858779907227, "learning_rate": 6.3145216792842395e-06, "loss": 0.6319, "step": 507800 }, { "epoch": 87.37955953200276, "grad_norm": 10.829290390014648, "learning_rate": 6.310220233998623e-06, "loss": 0.7005, "step": 507850 }, { "epoch": 87.38816242257398, "grad_norm": 15.643950462341309, "learning_rate": 6.305918788713009e-06, "loss": 0.7194, "step": 507900 }, { "epoch": 87.39676531314522, "grad_norm": 14.70565414428711, "learning_rate": 6.301617343427392e-06, "loss": 0.6671, "step": 507950 }, { "epoch": 87.40536820371645, "grad_norm": 12.538710594177246, "learning_rate": 6.297315898141776e-06, "loss": 0.658, "step": 508000 }, { "epoch": 87.41397109428767, "grad_norm": 19.39971351623535, "learning_rate": 6.29301445285616e-06, "loss": 0.5912, "step": 508050 }, { "epoch": 87.42257398485891, "grad_norm": 12.967759132385254, "learning_rate": 6.288713007570544e-06, "loss": 0.6443, "step": 508100 }, { "epoch": 87.43117687543014, "grad_norm": 15.868030548095703, "learning_rate": 6.284411562284928e-06, "loss": 0.6177, "step": 508150 }, { "epoch": 87.43977976600138, "grad_norm": 18.55802345275879, "learning_rate": 6.280110116999312e-06, "loss": 0.6512, "step": 508200 }, { "epoch": 87.4483826565726, "grad_norm": 17.311819076538086, "learning_rate": 6.2758086717136955e-06, "loss": 0.6302, "step": 508250 }, { "epoch": 87.45698554714384, "grad_norm": 15.956097602844238, "learning_rate": 6.271507226428081e-06, "loss": 0.6645, "step": 508300 }, { "epoch": 87.46558843771507, "grad_norm": 16.49057388305664, "learning_rate": 6.267205781142465e-06, "loss": 0.6342, "step": 508350 }, { "epoch": 87.47419132828631, "grad_norm": 19.877609252929688, "learning_rate": 6.2629043358568475e-06, "loss": 0.6541, "step": 508400 }, { "epoch": 87.48279421885753, "grad_norm": 14.10483169555664, "learning_rate": 6.258602890571231e-06, "loss": 0.6337, "step": 508450 }, { "epoch": 87.49139710942877, "grad_norm": 14.575467109680176, "learning_rate": 6.254301445285617e-06, "loss": 0.6423, "step": 508500 }, { "epoch": 87.5, "grad_norm": 25.723949432373047, "learning_rate": 6.25e-06, "loss": 0.6823, "step": 508550 }, { "epoch": 87.50860289057123, "grad_norm": 11.776397705078125, "learning_rate": 6.245698554714385e-06, "loss": 0.6177, "step": 508600 }, { "epoch": 87.51720578114247, "grad_norm": 13.535714149475098, "learning_rate": 6.2413971094287686e-06, "loss": 0.7128, "step": 508650 }, { "epoch": 87.52580867171369, "grad_norm": 19.04958152770996, "learning_rate": 6.237095664143153e-06, "loss": 0.6546, "step": 508700 }, { "epoch": 87.53441156228493, "grad_norm": 21.045698165893555, "learning_rate": 6.232794218857536e-06, "loss": 0.6486, "step": 508750 }, { "epoch": 87.54301445285616, "grad_norm": 12.45191478729248, "learning_rate": 6.2284927735719206e-06, "loss": 0.6425, "step": 508800 }, { "epoch": 87.5516173434274, "grad_norm": 18.020296096801758, "learning_rate": 6.224191328286304e-06, "loss": 0.6314, "step": 508850 }, { "epoch": 87.56022023399862, "grad_norm": 16.51337432861328, "learning_rate": 6.219889883000689e-06, "loss": 0.6452, "step": 508900 }, { "epoch": 87.56882312456986, "grad_norm": 17.572803497314453, "learning_rate": 6.2155884377150725e-06, "loss": 0.6788, "step": 508950 }, { "epoch": 87.57742601514109, "grad_norm": 19.462961196899414, "learning_rate": 6.211286992429457e-06, "loss": 0.6632, "step": 509000 }, { "epoch": 87.58602890571233, "grad_norm": 22.3441219329834, "learning_rate": 6.206985547143841e-06, "loss": 0.6882, "step": 509050 }, { "epoch": 87.59463179628355, "grad_norm": 9.545377731323242, "learning_rate": 6.2026841018582245e-06, "loss": 0.6123, "step": 509100 }, { "epoch": 87.60323468685478, "grad_norm": 13.7587308883667, "learning_rate": 6.198382656572608e-06, "loss": 0.63, "step": 509150 }, { "epoch": 87.61183757742602, "grad_norm": 10.786284446716309, "learning_rate": 6.194081211286993e-06, "loss": 0.6589, "step": 509200 }, { "epoch": 87.62044046799724, "grad_norm": 16.691286087036133, "learning_rate": 6.1897797660013765e-06, "loss": 0.6682, "step": 509250 }, { "epoch": 87.62904335856848, "grad_norm": 19.100948333740234, "learning_rate": 6.185478320715761e-06, "loss": 0.6687, "step": 509300 }, { "epoch": 87.6376462491397, "grad_norm": 20.283227920532227, "learning_rate": 6.181176875430145e-06, "loss": 0.701, "step": 509350 }, { "epoch": 87.64624913971095, "grad_norm": 22.297523498535156, "learning_rate": 6.176875430144529e-06, "loss": 0.6984, "step": 509400 }, { "epoch": 87.65485203028217, "grad_norm": 17.19627571105957, "learning_rate": 6.172573984858913e-06, "loss": 0.6323, "step": 509450 }, { "epoch": 87.66345492085341, "grad_norm": 16.85948944091797, "learning_rate": 6.168272539573297e-06, "loss": 0.6154, "step": 509500 }, { "epoch": 87.67205781142464, "grad_norm": 12.153596878051758, "learning_rate": 6.1639710942876805e-06, "loss": 0.6263, "step": 509550 }, { "epoch": 87.68066070199588, "grad_norm": 10.111740112304688, "learning_rate": 6.159669649002065e-06, "loss": 0.6653, "step": 509600 }, { "epoch": 87.6892635925671, "grad_norm": 18.873844146728516, "learning_rate": 6.155368203716449e-06, "loss": 0.641, "step": 509650 }, { "epoch": 87.69786648313834, "grad_norm": 36.35639190673828, "learning_rate": 6.151066758430833e-06, "loss": 0.6092, "step": 509700 }, { "epoch": 87.70646937370957, "grad_norm": 10.031547546386719, "learning_rate": 6.146765313145217e-06, "loss": 0.6296, "step": 509750 }, { "epoch": 87.71507226428079, "grad_norm": 19.07547950744629, "learning_rate": 6.142463867859602e-06, "loss": 0.674, "step": 509800 }, { "epoch": 87.72367515485203, "grad_norm": 23.60349464416504, "learning_rate": 6.138162422573985e-06, "loss": 0.5969, "step": 509850 }, { "epoch": 87.73227804542326, "grad_norm": 13.373129844665527, "learning_rate": 6.133860977288369e-06, "loss": 0.6556, "step": 509900 }, { "epoch": 87.7408809359945, "grad_norm": 20.519094467163086, "learning_rate": 6.129559532002753e-06, "loss": 0.6602, "step": 509950 }, { "epoch": 87.74948382656572, "grad_norm": 19.303855895996094, "learning_rate": 6.125258086717137e-06, "loss": 0.7305, "step": 510000 }, { "epoch": 87.75808671713696, "grad_norm": 15.501511573791504, "learning_rate": 6.120956641431521e-06, "loss": 0.5909, "step": 510050 }, { "epoch": 87.76668960770819, "grad_norm": 7.648715019226074, "learning_rate": 6.116655196145906e-06, "loss": 0.5997, "step": 510100 }, { "epoch": 87.77529249827943, "grad_norm": 15.594297409057617, "learning_rate": 6.112353750860289e-06, "loss": 0.59, "step": 510150 }, { "epoch": 87.78389538885065, "grad_norm": 14.548832893371582, "learning_rate": 6.108052305574673e-06, "loss": 0.6474, "step": 510200 }, { "epoch": 87.79249827942189, "grad_norm": 11.04488754272461, "learning_rate": 6.103750860289057e-06, "loss": 0.6473, "step": 510250 }, { "epoch": 87.80110116999312, "grad_norm": 12.927803039550781, "learning_rate": 6.099449415003441e-06, "loss": 0.6308, "step": 510300 }, { "epoch": 87.80970406056434, "grad_norm": 10.156615257263184, "learning_rate": 6.095147969717825e-06, "loss": 0.7292, "step": 510350 }, { "epoch": 87.81830695113558, "grad_norm": 22.413427352905273, "learning_rate": 6.0908465244322096e-06, "loss": 0.6104, "step": 510400 }, { "epoch": 87.82690984170681, "grad_norm": 15.608546257019043, "learning_rate": 6.086545079146594e-06, "loss": 0.7047, "step": 510450 }, { "epoch": 87.83551273227805, "grad_norm": 14.658751487731934, "learning_rate": 6.082243633860978e-06, "loss": 0.6699, "step": 510500 }, { "epoch": 87.84411562284927, "grad_norm": 9.553267478942871, "learning_rate": 6.0779421885753615e-06, "loss": 0.6912, "step": 510550 }, { "epoch": 87.85271851342051, "grad_norm": 10.852364540100098, "learning_rate": 6.073640743289745e-06, "loss": 0.6335, "step": 510600 }, { "epoch": 87.86132140399174, "grad_norm": 13.63857650756836, "learning_rate": 6.06933929800413e-06, "loss": 0.6518, "step": 510650 }, { "epoch": 87.86992429456298, "grad_norm": 27.99633026123047, "learning_rate": 6.0650378527185135e-06, "loss": 0.6679, "step": 510700 }, { "epoch": 87.8785271851342, "grad_norm": 11.728743553161621, "learning_rate": 6.060736407432898e-06, "loss": 0.6843, "step": 510750 }, { "epoch": 87.88713007570544, "grad_norm": 11.995609283447266, "learning_rate": 6.056434962147282e-06, "loss": 0.64, "step": 510800 }, { "epoch": 87.89573296627667, "grad_norm": 24.684389114379883, "learning_rate": 6.052133516861666e-06, "loss": 0.6934, "step": 510850 }, { "epoch": 87.90433585684791, "grad_norm": 19.76558494567871, "learning_rate": 6.04783207157605e-06, "loss": 0.6928, "step": 510900 }, { "epoch": 87.91293874741913, "grad_norm": 27.51551055908203, "learning_rate": 6.043530626290434e-06, "loss": 0.6465, "step": 510950 }, { "epoch": 87.92154163799036, "grad_norm": 12.686985969543457, "learning_rate": 6.0392291810048175e-06, "loss": 0.6423, "step": 511000 }, { "epoch": 87.9301445285616, "grad_norm": 10.705570220947266, "learning_rate": 6.034927735719202e-06, "loss": 0.6585, "step": 511050 }, { "epoch": 87.93874741913282, "grad_norm": 15.536213874816895, "learning_rate": 6.030626290433586e-06, "loss": 0.6631, "step": 511100 }, { "epoch": 87.94735030970406, "grad_norm": 18.91145896911621, "learning_rate": 6.02632484514797e-06, "loss": 0.5837, "step": 511150 }, { "epoch": 87.95595320027529, "grad_norm": 9.4952392578125, "learning_rate": 6.022023399862354e-06, "loss": 0.6304, "step": 511200 }, { "epoch": 87.96455609084653, "grad_norm": 27.701845169067383, "learning_rate": 6.017721954576739e-06, "loss": 0.6506, "step": 511250 }, { "epoch": 87.97315898141775, "grad_norm": 22.727252960205078, "learning_rate": 6.013420509291122e-06, "loss": 0.6304, "step": 511300 }, { "epoch": 87.981761871989, "grad_norm": 17.343149185180664, "learning_rate": 6.009119064005506e-06, "loss": 0.7376, "step": 511350 }, { "epoch": 87.99036476256022, "grad_norm": 13.75645923614502, "learning_rate": 6.00481761871989e-06, "loss": 0.6138, "step": 511400 }, { "epoch": 87.99896765313146, "grad_norm": 6.437226295471191, "learning_rate": 6.000516173434274e-06, "loss": 0.6914, "step": 511450 }, { "epoch": 88.0, "eval_accuracy": 0.5050765789020822, "eval_f1": 0.49259593155830494, "eval_f1_DuraRiadoRio_16x16": 0.41375358166189113, "eval_f1_Mole_16x16": 0.41801242236024844, "eval_f1_Quebrado_16x16": 0.6465267727930536, "eval_f1_RiadoRio_16x16": 0.4675569520816968, "eval_f1_RioFechado_16x16": 0.5171299288946348, "eval_loss": 3.128356695175171, "eval_precision": 0.5567863407337161, "eval_precision_DuraRiadoRio_16x16": 0.6087689713322091, "eval_precision_Mole_16x16": 0.7347161572052402, "eval_precision_Quebrado_16x16": 0.5542803970223326, "eval_precision_RiadoRio_16x16": 0.37833714721586575, "eval_precision_RioFechado_16x16": 0.5078290308929327, "eval_recall": 0.5039392738962566, "eval_recall_DuraRiadoRio_16x16": 0.3133680555555556, "eval_recall_Mole_16x16": 0.2921006944444444, "eval_recall_Quebrado_16x16": 0.7756076388888888, "eval_recall_RiadoRio_16x16": 0.6118421052631579, "eval_recall_RioFechado_16x16": 0.5267778753292361, "eval_runtime": 45.9977, "eval_samples_per_second": 252.665, "eval_steps_per_second": 15.805, "step": 511456 }, { "epoch": 88.00757054370268, "grad_norm": 15.253803253173828, "learning_rate": 5.996214728148658e-06, "loss": 0.6789, "step": 511500 }, { "epoch": 88.01617343427391, "grad_norm": 19.771780014038086, "learning_rate": 5.991913282863043e-06, "loss": 0.6513, "step": 511550 }, { "epoch": 88.02477632484515, "grad_norm": 17.02337646484375, "learning_rate": 5.987611837577426e-06, "loss": 0.6633, "step": 511600 }, { "epoch": 88.03337921541637, "grad_norm": 12.729588508605957, "learning_rate": 5.983310392291811e-06, "loss": 0.6567, "step": 511650 }, { "epoch": 88.04198210598761, "grad_norm": 8.420125961303711, "learning_rate": 5.979008947006194e-06, "loss": 0.6592, "step": 511700 }, { "epoch": 88.05058499655884, "grad_norm": 19.414112091064453, "learning_rate": 5.974707501720578e-06, "loss": 0.6216, "step": 511750 }, { "epoch": 88.05918788713008, "grad_norm": 16.840009689331055, "learning_rate": 5.970406056434962e-06, "loss": 0.6629, "step": 511800 }, { "epoch": 88.0677907777013, "grad_norm": 19.236804962158203, "learning_rate": 5.9661046111493466e-06, "loss": 0.6556, "step": 511850 }, { "epoch": 88.07639366827254, "grad_norm": 12.657967567443848, "learning_rate": 5.96180316586373e-06, "loss": 0.6481, "step": 511900 }, { "epoch": 88.08499655884377, "grad_norm": 19.446945190429688, "learning_rate": 5.957501720578115e-06, "loss": 0.6307, "step": 511950 }, { "epoch": 88.09359944941501, "grad_norm": 17.7238826751709, "learning_rate": 5.9532002752924985e-06, "loss": 0.6161, "step": 512000 }, { "epoch": 88.10220233998623, "grad_norm": 15.349797248840332, "learning_rate": 5.948898830006882e-06, "loss": 0.6578, "step": 512050 }, { "epoch": 88.11080523055746, "grad_norm": 14.297823905944824, "learning_rate": 5.944597384721266e-06, "loss": 0.5814, "step": 512100 }, { "epoch": 88.1194081211287, "grad_norm": 14.342781066894531, "learning_rate": 5.9402959394356505e-06, "loss": 0.642, "step": 512150 }, { "epoch": 88.12801101169993, "grad_norm": 10.405618667602539, "learning_rate": 5.935994494150034e-06, "loss": 0.6278, "step": 512200 }, { "epoch": 88.13661390227116, "grad_norm": 9.805435180664062, "learning_rate": 5.931693048864419e-06, "loss": 0.647, "step": 512250 }, { "epoch": 88.14521679284239, "grad_norm": 14.155518531799316, "learning_rate": 5.927391603578803e-06, "loss": 0.6354, "step": 512300 }, { "epoch": 88.15381968341363, "grad_norm": 16.47942352294922, "learning_rate": 5.923090158293187e-06, "loss": 0.6581, "step": 512350 }, { "epoch": 88.16242257398486, "grad_norm": 18.72136878967285, "learning_rate": 5.918788713007571e-06, "loss": 0.6352, "step": 512400 }, { "epoch": 88.1710254645561, "grad_norm": 18.700428009033203, "learning_rate": 5.9144872677219545e-06, "loss": 0.6793, "step": 512450 }, { "epoch": 88.17962835512732, "grad_norm": 16.905542373657227, "learning_rate": 5.910185822436339e-06, "loss": 0.6545, "step": 512500 }, { "epoch": 88.18823124569856, "grad_norm": 15.524137496948242, "learning_rate": 5.905884377150723e-06, "loss": 0.6146, "step": 512550 }, { "epoch": 88.19683413626979, "grad_norm": 24.471603393554688, "learning_rate": 5.901582931865107e-06, "loss": 0.6407, "step": 512600 }, { "epoch": 88.20543702684103, "grad_norm": 17.43509864807129, "learning_rate": 5.897281486579491e-06, "loss": 0.6602, "step": 512650 }, { "epoch": 88.21403991741225, "grad_norm": 18.905488967895508, "learning_rate": 5.892980041293876e-06, "loss": 0.6647, "step": 512700 }, { "epoch": 88.22264280798348, "grad_norm": 14.609736442565918, "learning_rate": 5.888678596008259e-06, "loss": 0.6559, "step": 512750 }, { "epoch": 88.23124569855472, "grad_norm": 20.845090866088867, "learning_rate": 5.884377150722643e-06, "loss": 0.6785, "step": 512800 }, { "epoch": 88.23984858912594, "grad_norm": 15.953295707702637, "learning_rate": 5.880075705437027e-06, "loss": 0.6404, "step": 512850 }, { "epoch": 88.24845147969718, "grad_norm": 17.74320411682129, "learning_rate": 5.875774260151411e-06, "loss": 0.6215, "step": 512900 }, { "epoch": 88.2570543702684, "grad_norm": 20.850812911987305, "learning_rate": 5.871472814865795e-06, "loss": 0.6401, "step": 512950 }, { "epoch": 88.26565726083965, "grad_norm": 16.5103702545166, "learning_rate": 5.86717136958018e-06, "loss": 0.6495, "step": 513000 }, { "epoch": 88.27426015141087, "grad_norm": 13.490988731384277, "learning_rate": 5.862869924294563e-06, "loss": 0.6692, "step": 513050 }, { "epoch": 88.28286304198211, "grad_norm": 14.038399696350098, "learning_rate": 5.858568479008948e-06, "loss": 0.6901, "step": 513100 }, { "epoch": 88.29146593255334, "grad_norm": 22.578073501586914, "learning_rate": 5.8542670337233316e-06, "loss": 0.665, "step": 513150 }, { "epoch": 88.30006882312458, "grad_norm": 16.567014694213867, "learning_rate": 5.849965588437715e-06, "loss": 0.6751, "step": 513200 }, { "epoch": 88.3086717136958, "grad_norm": 17.684402465820312, "learning_rate": 5.845664143152099e-06, "loss": 0.6447, "step": 513250 }, { "epoch": 88.31727460426703, "grad_norm": 20.073965072631836, "learning_rate": 5.8413626978664836e-06, "loss": 0.5896, "step": 513300 }, { "epoch": 88.32587749483827, "grad_norm": 18.161022186279297, "learning_rate": 5.837061252580867e-06, "loss": 0.655, "step": 513350 }, { "epoch": 88.33448038540949, "grad_norm": 18.21771812438965, "learning_rate": 5.832759807295252e-06, "loss": 0.6511, "step": 513400 }, { "epoch": 88.34308327598073, "grad_norm": 19.954357147216797, "learning_rate": 5.8284583620096355e-06, "loss": 0.69, "step": 513450 }, { "epoch": 88.35168616655196, "grad_norm": 16.82762908935547, "learning_rate": 5.824156916724019e-06, "loss": 0.6604, "step": 513500 }, { "epoch": 88.3602890571232, "grad_norm": 14.733063697814941, "learning_rate": 5.819855471438403e-06, "loss": 0.6643, "step": 513550 }, { "epoch": 88.36889194769442, "grad_norm": 9.858926773071289, "learning_rate": 5.8155540261527875e-06, "loss": 0.6457, "step": 513600 }, { "epoch": 88.37749483826566, "grad_norm": 20.625484466552734, "learning_rate": 5.811252580867171e-06, "loss": 0.7339, "step": 513650 }, { "epoch": 88.38609772883689, "grad_norm": 9.513934135437012, "learning_rate": 5.806951135581556e-06, "loss": 0.6148, "step": 513700 }, { "epoch": 88.39470061940813, "grad_norm": 11.233830451965332, "learning_rate": 5.8026496902959395e-06, "loss": 0.6364, "step": 513750 }, { "epoch": 88.40330350997935, "grad_norm": 20.421276092529297, "learning_rate": 5.798348245010324e-06, "loss": 0.6189, "step": 513800 }, { "epoch": 88.41190640055059, "grad_norm": 12.850944519042969, "learning_rate": 5.794046799724708e-06, "loss": 0.616, "step": 513850 }, { "epoch": 88.42050929112182, "grad_norm": 12.603731155395508, "learning_rate": 5.7897453544390915e-06, "loss": 0.7013, "step": 513900 }, { "epoch": 88.42911218169304, "grad_norm": 11.208621978759766, "learning_rate": 5.785443909153475e-06, "loss": 0.7007, "step": 513950 }, { "epoch": 88.43771507226428, "grad_norm": 17.53771209716797, "learning_rate": 5.78114246386786e-06, "loss": 0.6183, "step": 514000 }, { "epoch": 88.44631796283551, "grad_norm": 15.975615501403809, "learning_rate": 5.7768410185822435e-06, "loss": 0.6054, "step": 514050 }, { "epoch": 88.45492085340675, "grad_norm": 13.576854705810547, "learning_rate": 5.772539573296628e-06, "loss": 0.6236, "step": 514100 }, { "epoch": 88.46352374397797, "grad_norm": 14.353033065795898, "learning_rate": 5.768238128011013e-06, "loss": 0.6821, "step": 514150 }, { "epoch": 88.47212663454921, "grad_norm": 18.20975685119629, "learning_rate": 5.763936682725396e-06, "loss": 0.6152, "step": 514200 }, { "epoch": 88.48072952512044, "grad_norm": 7.847320079803467, "learning_rate": 5.75963523743978e-06, "loss": 0.7033, "step": 514250 }, { "epoch": 88.48933241569168, "grad_norm": 16.42988395690918, "learning_rate": 5.755333792154164e-06, "loss": 0.6408, "step": 514300 }, { "epoch": 88.4979353062629, "grad_norm": 20.21946144104004, "learning_rate": 5.751032346868548e-06, "loss": 0.6798, "step": 514350 }, { "epoch": 88.50653819683414, "grad_norm": 12.55544376373291, "learning_rate": 5.746730901582932e-06, "loss": 0.6442, "step": 514400 }, { "epoch": 88.51514108740537, "grad_norm": 18.664072036743164, "learning_rate": 5.742429456297317e-06, "loss": 0.6775, "step": 514450 }, { "epoch": 88.5237439779766, "grad_norm": 13.428474426269531, "learning_rate": 5.7381280110117e-06, "loss": 0.6585, "step": 514500 }, { "epoch": 88.53234686854783, "grad_norm": 11.544197082519531, "learning_rate": 5.733826565726085e-06, "loss": 0.6706, "step": 514550 }, { "epoch": 88.54094975911906, "grad_norm": 9.319703102111816, "learning_rate": 5.729525120440469e-06, "loss": 0.6061, "step": 514600 }, { "epoch": 88.5495526496903, "grad_norm": 13.555671691894531, "learning_rate": 5.725223675154852e-06, "loss": 0.7307, "step": 514650 }, { "epoch": 88.55815554026152, "grad_norm": 19.314668655395508, "learning_rate": 5.720922229869236e-06, "loss": 0.6635, "step": 514700 }, { "epoch": 88.56675843083276, "grad_norm": 19.168685913085938, "learning_rate": 5.7166207845836206e-06, "loss": 0.6164, "step": 514750 }, { "epoch": 88.57536132140399, "grad_norm": 21.61888885498047, "learning_rate": 5.712319339298004e-06, "loss": 0.65, "step": 514800 }, { "epoch": 88.58396421197523, "grad_norm": 13.2701997756958, "learning_rate": 5.708017894012389e-06, "loss": 0.6127, "step": 514850 }, { "epoch": 88.59256710254645, "grad_norm": 21.146976470947266, "learning_rate": 5.7037164487267726e-06, "loss": 0.6694, "step": 514900 }, { "epoch": 88.6011699931177, "grad_norm": 11.03477668762207, "learning_rate": 5.699415003441157e-06, "loss": 0.6639, "step": 514950 }, { "epoch": 88.60977288368892, "grad_norm": 9.514806747436523, "learning_rate": 5.69511355815554e-06, "loss": 0.6515, "step": 515000 }, { "epoch": 88.61837577426016, "grad_norm": 21.70977020263672, "learning_rate": 5.6908121128699245e-06, "loss": 0.7034, "step": 515050 }, { "epoch": 88.62697866483138, "grad_norm": 19.126455307006836, "learning_rate": 5.686510667584308e-06, "loss": 0.6787, "step": 515100 }, { "epoch": 88.63558155540261, "grad_norm": 9.055164337158203, "learning_rate": 5.682209222298693e-06, "loss": 0.6847, "step": 515150 }, { "epoch": 88.64418444597385, "grad_norm": 19.157363891601562, "learning_rate": 5.6779077770130765e-06, "loss": 0.5953, "step": 515200 }, { "epoch": 88.65278733654507, "grad_norm": 15.672728538513184, "learning_rate": 5.673606331727461e-06, "loss": 0.6928, "step": 515250 }, { "epoch": 88.66139022711631, "grad_norm": 24.119977951049805, "learning_rate": 5.669304886441845e-06, "loss": 0.6801, "step": 515300 }, { "epoch": 88.66999311768754, "grad_norm": 15.747650146484375, "learning_rate": 5.6650034411562285e-06, "loss": 0.649, "step": 515350 }, { "epoch": 88.67859600825878, "grad_norm": 19.431459426879883, "learning_rate": 5.660701995870612e-06, "loss": 0.6677, "step": 515400 }, { "epoch": 88.68719889883, "grad_norm": 13.584957122802734, "learning_rate": 5.656400550584997e-06, "loss": 0.7153, "step": 515450 }, { "epoch": 88.69580178940124, "grad_norm": 8.954976081848145, "learning_rate": 5.6520991052993805e-06, "loss": 0.6466, "step": 515500 }, { "epoch": 88.70440467997247, "grad_norm": 7.082923412322998, "learning_rate": 5.647797660013765e-06, "loss": 0.6341, "step": 515550 }, { "epoch": 88.71300757054371, "grad_norm": 10.19987964630127, "learning_rate": 5.643496214728149e-06, "loss": 0.6922, "step": 515600 }, { "epoch": 88.72161046111493, "grad_norm": 11.981895446777344, "learning_rate": 5.639194769442533e-06, "loss": 0.6784, "step": 515650 }, { "epoch": 88.73021335168616, "grad_norm": 20.122541427612305, "learning_rate": 5.634893324156917e-06, "loss": 0.699, "step": 515700 }, { "epoch": 88.7388162422574, "grad_norm": 13.366806983947754, "learning_rate": 5.630591878871301e-06, "loss": 0.6664, "step": 515750 }, { "epoch": 88.74741913282863, "grad_norm": 10.750764846801758, "learning_rate": 5.6262904335856845e-06, "loss": 0.6468, "step": 515800 }, { "epoch": 88.75602202339986, "grad_norm": 12.647062301635742, "learning_rate": 5.621988988300069e-06, "loss": 0.6518, "step": 515850 }, { "epoch": 88.76462491397109, "grad_norm": 11.712965965270996, "learning_rate": 5.617687543014453e-06, "loss": 0.6306, "step": 515900 }, { "epoch": 88.77322780454233, "grad_norm": 13.209558486938477, "learning_rate": 5.613386097728837e-06, "loss": 0.6711, "step": 515950 }, { "epoch": 88.78183069511356, "grad_norm": 14.084070205688477, "learning_rate": 5.609084652443221e-06, "loss": 0.7257, "step": 516000 }, { "epoch": 88.7904335856848, "grad_norm": 12.866605758666992, "learning_rate": 5.604783207157606e-06, "loss": 0.6217, "step": 516050 }, { "epoch": 88.79903647625602, "grad_norm": 19.42498207092285, "learning_rate": 5.600481761871989e-06, "loss": 0.6448, "step": 516100 }, { "epoch": 88.80763936682726, "grad_norm": 11.579509735107422, "learning_rate": 5.596180316586373e-06, "loss": 0.7153, "step": 516150 }, { "epoch": 88.81624225739849, "grad_norm": 14.055475234985352, "learning_rate": 5.5918788713007576e-06, "loss": 0.6659, "step": 516200 }, { "epoch": 88.82484514796971, "grad_norm": 8.388358116149902, "learning_rate": 5.587577426015141e-06, "loss": 0.6983, "step": 516250 }, { "epoch": 88.83344803854095, "grad_norm": 27.671907424926758, "learning_rate": 5.583275980729526e-06, "loss": 0.6312, "step": 516300 }, { "epoch": 88.84205092911218, "grad_norm": 9.529219627380371, "learning_rate": 5.5789745354439096e-06, "loss": 0.6095, "step": 516350 }, { "epoch": 88.85065381968342, "grad_norm": 17.187320709228516, "learning_rate": 5.574673090158294e-06, "loss": 0.682, "step": 516400 }, { "epoch": 88.85925671025464, "grad_norm": 13.256640434265137, "learning_rate": 5.570371644872677e-06, "loss": 0.6118, "step": 516450 }, { "epoch": 88.86785960082588, "grad_norm": 10.522855758666992, "learning_rate": 5.5660701995870615e-06, "loss": 0.6629, "step": 516500 }, { "epoch": 88.8764624913971, "grad_norm": 11.89578628540039, "learning_rate": 5.561768754301445e-06, "loss": 0.6523, "step": 516550 }, { "epoch": 88.88506538196835, "grad_norm": 26.823375701904297, "learning_rate": 5.55746730901583e-06, "loss": 0.6512, "step": 516600 }, { "epoch": 88.89366827253957, "grad_norm": 11.532902717590332, "learning_rate": 5.5531658637302135e-06, "loss": 0.6614, "step": 516650 }, { "epoch": 88.90227116311081, "grad_norm": 12.341559410095215, "learning_rate": 5.548864418444598e-06, "loss": 0.646, "step": 516700 }, { "epoch": 88.91087405368204, "grad_norm": 15.966952323913574, "learning_rate": 5.544562973158982e-06, "loss": 0.6034, "step": 516750 }, { "epoch": 88.91947694425328, "grad_norm": 13.50430679321289, "learning_rate": 5.5402615278733655e-06, "loss": 0.6458, "step": 516800 }, { "epoch": 88.9280798348245, "grad_norm": 19.902938842773438, "learning_rate": 5.535960082587749e-06, "loss": 0.6555, "step": 516850 }, { "epoch": 88.93668272539573, "grad_norm": 14.649930953979492, "learning_rate": 5.531658637302134e-06, "loss": 0.6697, "step": 516900 }, { "epoch": 88.94528561596697, "grad_norm": 13.47498607635498, "learning_rate": 5.5273571920165175e-06, "loss": 0.7141, "step": 516950 }, { "epoch": 88.95388850653819, "grad_norm": 25.433231353759766, "learning_rate": 5.523055746730902e-06, "loss": 0.6507, "step": 517000 }, { "epoch": 88.96249139710943, "grad_norm": 21.59850311279297, "learning_rate": 5.518754301445286e-06, "loss": 0.6675, "step": 517050 }, { "epoch": 88.97109428768066, "grad_norm": 11.335089683532715, "learning_rate": 5.51445285615967e-06, "loss": 0.6636, "step": 517100 }, { "epoch": 88.9796971782519, "grad_norm": 20.344833374023438, "learning_rate": 5.510151410874054e-06, "loss": 0.6261, "step": 517150 }, { "epoch": 88.98830006882312, "grad_norm": 15.319243431091309, "learning_rate": 5.505849965588438e-06, "loss": 0.6805, "step": 517200 }, { "epoch": 88.99690295939436, "grad_norm": 7.786344051361084, "learning_rate": 5.5015485203028215e-06, "loss": 0.6262, "step": 517250 }, { "epoch": 89.0, "eval_accuracy": 0.5641025641025641, "eval_f1": 0.531926047280348, "eval_f1_DuraRiadoRio_16x16": 0.34365430418509696, "eval_f1_Mole_16x16": 0.6215153681200858, "eval_f1_Quebrado_16x16": 0.664321608040201, "eval_f1_RiadoRio_16x16": 0.40217687074829933, "eval_f1_RioFechado_16x16": 0.6279620853080569, "eval_loss": 2.3909754753112793, "eval_precision": 0.6059364970751654, "eval_precision_DuraRiadoRio_16x16": 0.7952755905511811, "eval_precision_Mole_16x16": 0.5282503037667071, "eval_precision_Quebrado_16x16": 0.5409165302782324, "eval_precision_RiadoRio_16x16": 0.5945293644408689, "eval_precision_RioFechado_16x16": 0.5707106963388371, "eval_recall": 0.5672962466113703, "eval_recall_DuraRiadoRio_16x16": 0.2191840277777778, "eval_recall_Mole_16x16": 0.7547743055555556, "eval_recall_Quebrado_16x16": 0.8606770833333334, "eval_recall_RiadoRio_16x16": 0.30386513157894735, "eval_recall_RioFechado_16x16": 0.6979806848112379, "eval_runtime": 46.284, "eval_samples_per_second": 251.102, "eval_steps_per_second": 15.707, "step": 517268 }, { "epoch": 89.00550584996559, "grad_norm": 17.127193450927734, "learning_rate": 5.497247075017206e-06, "loss": 0.648, "step": 517300 }, { "epoch": 89.01410874053683, "grad_norm": 23.126569747924805, "learning_rate": 5.49294562973159e-06, "loss": 0.7023, "step": 517350 }, { "epoch": 89.02271163110805, "grad_norm": 18.66118049621582, "learning_rate": 5.488644184445974e-06, "loss": 0.636, "step": 517400 }, { "epoch": 89.03131452167928, "grad_norm": 8.543896675109863, "learning_rate": 5.484342739160358e-06, "loss": 0.6359, "step": 517450 }, { "epoch": 89.03991741225052, "grad_norm": 8.737836837768555, "learning_rate": 5.480041293874743e-06, "loss": 0.6308, "step": 517500 }, { "epoch": 89.04852030282174, "grad_norm": 13.313056945800781, "learning_rate": 5.475739848589126e-06, "loss": 0.6481, "step": 517550 }, { "epoch": 89.05712319339298, "grad_norm": 9.401148796081543, "learning_rate": 5.47143840330351e-06, "loss": 0.6201, "step": 517600 }, { "epoch": 89.06572608396421, "grad_norm": 23.910491943359375, "learning_rate": 5.467136958017894e-06, "loss": 0.6607, "step": 517650 }, { "epoch": 89.07432897453545, "grad_norm": 15.517803192138672, "learning_rate": 5.462835512732278e-06, "loss": 0.6259, "step": 517700 }, { "epoch": 89.08293186510667, "grad_norm": 18.700939178466797, "learning_rate": 5.458534067446662e-06, "loss": 0.6562, "step": 517750 }, { "epoch": 89.09153475567791, "grad_norm": 15.470602035522461, "learning_rate": 5.4542326221610466e-06, "loss": 0.6019, "step": 517800 }, { "epoch": 89.10013764624914, "grad_norm": 22.98947525024414, "learning_rate": 5.44993117687543e-06, "loss": 0.6476, "step": 517850 }, { "epoch": 89.10874053682038, "grad_norm": 17.204042434692383, "learning_rate": 5.445629731589815e-06, "loss": 0.6124, "step": 517900 }, { "epoch": 89.1173434273916, "grad_norm": 15.952340126037598, "learning_rate": 5.4413282863041986e-06, "loss": 0.6121, "step": 517950 }, { "epoch": 89.12594631796284, "grad_norm": 19.49895668029785, "learning_rate": 5.437026841018582e-06, "loss": 0.6256, "step": 518000 }, { "epoch": 89.13454920853407, "grad_norm": 16.622167587280273, "learning_rate": 5.432725395732967e-06, "loss": 0.6424, "step": 518050 }, { "epoch": 89.1431520991053, "grad_norm": 16.70691680908203, "learning_rate": 5.4284239504473505e-06, "loss": 0.6541, "step": 518100 }, { "epoch": 89.15175498967653, "grad_norm": 18.5230655670166, "learning_rate": 5.424122505161735e-06, "loss": 0.6721, "step": 518150 }, { "epoch": 89.16035788024776, "grad_norm": 13.80899429321289, "learning_rate": 5.419821059876119e-06, "loss": 0.6465, "step": 518200 }, { "epoch": 89.168960770819, "grad_norm": 14.180459022521973, "learning_rate": 5.4155196145905025e-06, "loss": 0.6465, "step": 518250 }, { "epoch": 89.17756366139022, "grad_norm": 8.175030708312988, "learning_rate": 5.411218169304886e-06, "loss": 0.6941, "step": 518300 }, { "epoch": 89.18616655196146, "grad_norm": 18.475522994995117, "learning_rate": 5.406916724019271e-06, "loss": 0.5976, "step": 518350 }, { "epoch": 89.19476944253269, "grad_norm": 17.124130249023438, "learning_rate": 5.4026152787336545e-06, "loss": 0.6825, "step": 518400 }, { "epoch": 89.20337233310393, "grad_norm": 20.241735458374023, "learning_rate": 5.398313833448039e-06, "loss": 0.6868, "step": 518450 }, { "epoch": 89.21197522367515, "grad_norm": 9.920989990234375, "learning_rate": 5.394012388162423e-06, "loss": 0.6559, "step": 518500 }, { "epoch": 89.2205781142464, "grad_norm": 21.810178756713867, "learning_rate": 5.389710942876807e-06, "loss": 0.6673, "step": 518550 }, { "epoch": 89.22918100481762, "grad_norm": 13.282573699951172, "learning_rate": 5.385409497591191e-06, "loss": 0.6385, "step": 518600 }, { "epoch": 89.23778389538884, "grad_norm": 19.638147354125977, "learning_rate": 5.381108052305575e-06, "loss": 0.6502, "step": 518650 }, { "epoch": 89.24638678596008, "grad_norm": 15.624773025512695, "learning_rate": 5.3768066070199585e-06, "loss": 0.6434, "step": 518700 }, { "epoch": 89.25498967653131, "grad_norm": 13.319419860839844, "learning_rate": 5.372505161734343e-06, "loss": 0.6982, "step": 518750 }, { "epoch": 89.26359256710255, "grad_norm": 14.374815940856934, "learning_rate": 5.368203716448727e-06, "loss": 0.6551, "step": 518800 }, { "epoch": 89.27219545767377, "grad_norm": 19.747730255126953, "learning_rate": 5.363902271163111e-06, "loss": 0.6719, "step": 518850 }, { "epoch": 89.28079834824501, "grad_norm": 16.416629791259766, "learning_rate": 5.359600825877495e-06, "loss": 0.6228, "step": 518900 }, { "epoch": 89.28940123881624, "grad_norm": 18.439937591552734, "learning_rate": 5.35529938059188e-06, "loss": 0.6353, "step": 518950 }, { "epoch": 89.29800412938748, "grad_norm": 13.311431884765625, "learning_rate": 5.350997935306263e-06, "loss": 0.5948, "step": 519000 }, { "epoch": 89.3066070199587, "grad_norm": 13.56313705444336, "learning_rate": 5.346696490020647e-06, "loss": 0.6106, "step": 519050 }, { "epoch": 89.31520991052994, "grad_norm": 17.29841423034668, "learning_rate": 5.342395044735031e-06, "loss": 0.6953, "step": 519100 }, { "epoch": 89.32381280110117, "grad_norm": 13.554557800292969, "learning_rate": 5.338093599449415e-06, "loss": 0.7239, "step": 519150 }, { "epoch": 89.3324156916724, "grad_norm": 13.464914321899414, "learning_rate": 5.333792154163799e-06, "loss": 0.664, "step": 519200 }, { "epoch": 89.34101858224363, "grad_norm": 11.419644355773926, "learning_rate": 5.3294907088781836e-06, "loss": 0.6465, "step": 519250 }, { "epoch": 89.34962147281486, "grad_norm": 19.50766944885254, "learning_rate": 5.325189263592567e-06, "loss": 0.6367, "step": 519300 }, { "epoch": 89.3582243633861, "grad_norm": 18.088924407958984, "learning_rate": 5.320887818306952e-06, "loss": 0.6905, "step": 519350 }, { "epoch": 89.36682725395733, "grad_norm": 19.722129821777344, "learning_rate": 5.3165863730213356e-06, "loss": 0.6447, "step": 519400 }, { "epoch": 89.37543014452856, "grad_norm": 7.914827823638916, "learning_rate": 5.312284927735719e-06, "loss": 0.6545, "step": 519450 }, { "epoch": 89.38403303509979, "grad_norm": 14.107178688049316, "learning_rate": 5.307983482450103e-06, "loss": 0.6318, "step": 519500 }, { "epoch": 89.39263592567103, "grad_norm": 10.600886344909668, "learning_rate": 5.3036820371644875e-06, "loss": 0.6401, "step": 519550 }, { "epoch": 89.40123881624226, "grad_norm": 13.813966751098633, "learning_rate": 5.299380591878871e-06, "loss": 0.626, "step": 519600 }, { "epoch": 89.4098417068135, "grad_norm": 18.991785049438477, "learning_rate": 5.295079146593256e-06, "loss": 0.6364, "step": 519650 }, { "epoch": 89.41844459738472, "grad_norm": 25.38077163696289, "learning_rate": 5.2907777013076395e-06, "loss": 0.6447, "step": 519700 }, { "epoch": 89.42704748795596, "grad_norm": 10.24166488647461, "learning_rate": 5.286476256022023e-06, "loss": 0.6957, "step": 519750 }, { "epoch": 89.43565037852719, "grad_norm": 18.71224021911621, "learning_rate": 5.282174810736408e-06, "loss": 0.6683, "step": 519800 }, { "epoch": 89.44425326909841, "grad_norm": 25.951534271240234, "learning_rate": 5.2778733654507915e-06, "loss": 0.717, "step": 519850 }, { "epoch": 89.45285615966965, "grad_norm": 14.181407928466797, "learning_rate": 5.273571920165176e-06, "loss": 0.6606, "step": 519900 }, { "epoch": 89.46145905024088, "grad_norm": 9.562050819396973, "learning_rate": 5.26927047487956e-06, "loss": 0.6526, "step": 519950 }, { "epoch": 89.47006194081212, "grad_norm": 6.7116193771362305, "learning_rate": 5.264969029593944e-06, "loss": 0.6626, "step": 520000 }, { "epoch": 89.47866483138334, "grad_norm": 20.403274536132812, "learning_rate": 5.260667584308328e-06, "loss": 0.6478, "step": 520050 }, { "epoch": 89.48726772195458, "grad_norm": 13.5233154296875, "learning_rate": 5.256366139022712e-06, "loss": 0.6299, "step": 520100 }, { "epoch": 89.4958706125258, "grad_norm": 9.566851615905762, "learning_rate": 5.2520646937370955e-06, "loss": 0.6927, "step": 520150 }, { "epoch": 89.50447350309705, "grad_norm": 11.268318176269531, "learning_rate": 5.24776324845148e-06, "loss": 0.6293, "step": 520200 }, { "epoch": 89.51307639366827, "grad_norm": 19.26783561706543, "learning_rate": 5.243461803165864e-06, "loss": 0.6365, "step": 520250 }, { "epoch": 89.52167928423951, "grad_norm": 19.113750457763672, "learning_rate": 5.239160357880248e-06, "loss": 0.6424, "step": 520300 }, { "epoch": 89.53028217481074, "grad_norm": 18.560922622680664, "learning_rate": 5.234858912594632e-06, "loss": 0.6926, "step": 520350 }, { "epoch": 89.53888506538196, "grad_norm": 21.12982177734375, "learning_rate": 5.230557467309017e-06, "loss": 0.6291, "step": 520400 }, { "epoch": 89.5474879559532, "grad_norm": 18.565683364868164, "learning_rate": 5.2262560220234e-06, "loss": 0.6034, "step": 520450 }, { "epoch": 89.55609084652443, "grad_norm": 14.96968936920166, "learning_rate": 5.221954576737784e-06, "loss": 0.6923, "step": 520500 }, { "epoch": 89.56469373709567, "grad_norm": 15.048370361328125, "learning_rate": 5.217653131452168e-06, "loss": 0.6067, "step": 520550 }, { "epoch": 89.57329662766689, "grad_norm": 20.110376358032227, "learning_rate": 5.213351686166552e-06, "loss": 0.6488, "step": 520600 }, { "epoch": 89.58189951823813, "grad_norm": 11.16752815246582, "learning_rate": 5.209050240880936e-06, "loss": 0.5751, "step": 520650 }, { "epoch": 89.59050240880936, "grad_norm": 11.306755065917969, "learning_rate": 5.2047487955953206e-06, "loss": 0.6159, "step": 520700 }, { "epoch": 89.5991052993806, "grad_norm": 18.44753646850586, "learning_rate": 5.200447350309704e-06, "loss": 0.6115, "step": 520750 }, { "epoch": 89.60770818995182, "grad_norm": 18.912853240966797, "learning_rate": 5.196145905024089e-06, "loss": 0.6204, "step": 520800 }, { "epoch": 89.61631108052306, "grad_norm": 20.78105926513672, "learning_rate": 5.1918444597384726e-06, "loss": 0.6919, "step": 520850 }, { "epoch": 89.62491397109429, "grad_norm": 27.862009048461914, "learning_rate": 5.187543014452856e-06, "loss": 0.7329, "step": 520900 }, { "epoch": 89.63351686166553, "grad_norm": 14.395604133605957, "learning_rate": 5.18324156916724e-06, "loss": 0.642, "step": 520950 }, { "epoch": 89.64211975223675, "grad_norm": 25.44158363342285, "learning_rate": 5.1789401238816245e-06, "loss": 0.6368, "step": 521000 }, { "epoch": 89.65072264280798, "grad_norm": 16.24249839782715, "learning_rate": 5.174638678596008e-06, "loss": 0.6735, "step": 521050 }, { "epoch": 89.65932553337922, "grad_norm": 22.254228591918945, "learning_rate": 5.170337233310393e-06, "loss": 0.6478, "step": 521100 }, { "epoch": 89.66792842395044, "grad_norm": 24.555391311645508, "learning_rate": 5.1660357880247765e-06, "loss": 0.6259, "step": 521150 }, { "epoch": 89.67653131452168, "grad_norm": 8.26043701171875, "learning_rate": 5.161734342739161e-06, "loss": 0.6104, "step": 521200 }, { "epoch": 89.68513420509291, "grad_norm": 14.07037353515625, "learning_rate": 5.157432897453544e-06, "loss": 0.6506, "step": 521250 }, { "epoch": 89.69373709566415, "grad_norm": 19.633691787719727, "learning_rate": 5.1531314521679285e-06, "loss": 0.6635, "step": 521300 }, { "epoch": 89.70233998623537, "grad_norm": 26.606151580810547, "learning_rate": 5.148830006882312e-06, "loss": 0.6254, "step": 521350 }, { "epoch": 89.71094287680661, "grad_norm": 14.180624961853027, "learning_rate": 5.144528561596697e-06, "loss": 0.6311, "step": 521400 }, { "epoch": 89.71954576737784, "grad_norm": 8.504090309143066, "learning_rate": 5.1402271163110805e-06, "loss": 0.6948, "step": 521450 }, { "epoch": 89.72814865794908, "grad_norm": 17.662351608276367, "learning_rate": 5.135925671025465e-06, "loss": 0.6238, "step": 521500 }, { "epoch": 89.7367515485203, "grad_norm": 22.50717544555664, "learning_rate": 5.131624225739849e-06, "loss": 0.6201, "step": 521550 }, { "epoch": 89.74535443909153, "grad_norm": 11.823286056518555, "learning_rate": 5.1273227804542325e-06, "loss": 0.6389, "step": 521600 }, { "epoch": 89.75395732966277, "grad_norm": 13.46589183807373, "learning_rate": 5.123021335168617e-06, "loss": 0.6835, "step": 521650 }, { "epoch": 89.762560220234, "grad_norm": 12.430127143859863, "learning_rate": 5.118719889883001e-06, "loss": 0.6329, "step": 521700 }, { "epoch": 89.77116311080523, "grad_norm": 12.608955383300781, "learning_rate": 5.114418444597385e-06, "loss": 0.6198, "step": 521750 }, { "epoch": 89.77976600137646, "grad_norm": 24.46034049987793, "learning_rate": 5.110116999311769e-06, "loss": 0.6667, "step": 521800 }, { "epoch": 89.7883688919477, "grad_norm": 13.387639045715332, "learning_rate": 5.105815554026154e-06, "loss": 0.6429, "step": 521850 }, { "epoch": 89.79697178251892, "grad_norm": 17.06013298034668, "learning_rate": 5.101514108740537e-06, "loss": 0.6299, "step": 521900 }, { "epoch": 89.80557467309016, "grad_norm": 22.881376266479492, "learning_rate": 5.097212663454921e-06, "loss": 0.5678, "step": 521950 }, { "epoch": 89.81417756366139, "grad_norm": 19.90369415283203, "learning_rate": 5.092911218169305e-06, "loss": 0.6183, "step": 522000 }, { "epoch": 89.82278045423263, "grad_norm": 10.950478553771973, "learning_rate": 5.088609772883689e-06, "loss": 0.654, "step": 522050 }, { "epoch": 89.83138334480385, "grad_norm": 19.79897117614746, "learning_rate": 5.084308327598073e-06, "loss": 0.6361, "step": 522100 }, { "epoch": 89.83998623537508, "grad_norm": 10.41519832611084, "learning_rate": 5.080006882312458e-06, "loss": 0.6546, "step": 522150 }, { "epoch": 89.84858912594632, "grad_norm": 19.28386116027832, "learning_rate": 5.075705437026841e-06, "loss": 0.6013, "step": 522200 }, { "epoch": 89.85719201651754, "grad_norm": 18.53818130493164, "learning_rate": 5.071403991741226e-06, "loss": 0.6836, "step": 522250 }, { "epoch": 89.86579490708878, "grad_norm": 16.626493453979492, "learning_rate": 5.0671025464556096e-06, "loss": 0.6275, "step": 522300 }, { "epoch": 89.87439779766001, "grad_norm": 7.128108501434326, "learning_rate": 5.062801101169993e-06, "loss": 0.6547, "step": 522350 }, { "epoch": 89.88300068823125, "grad_norm": 20.615510940551758, "learning_rate": 5.058499655884377e-06, "loss": 0.6213, "step": 522400 }, { "epoch": 89.89160357880247, "grad_norm": 10.279121398925781, "learning_rate": 5.0541982105987616e-06, "loss": 0.6662, "step": 522450 }, { "epoch": 89.90020646937371, "grad_norm": 7.166886806488037, "learning_rate": 5.049896765313145e-06, "loss": 0.6415, "step": 522500 }, { "epoch": 89.90880935994494, "grad_norm": 9.241726875305176, "learning_rate": 5.04559532002753e-06, "loss": 0.6246, "step": 522550 }, { "epoch": 89.91741225051618, "grad_norm": 18.531137466430664, "learning_rate": 5.0412938747419135e-06, "loss": 0.6349, "step": 522600 }, { "epoch": 89.9260151410874, "grad_norm": 17.640666961669922, "learning_rate": 5.036992429456298e-06, "loss": 0.6012, "step": 522650 }, { "epoch": 89.93461803165864, "grad_norm": 24.293283462524414, "learning_rate": 5.032690984170681e-06, "loss": 0.7293, "step": 522700 }, { "epoch": 89.94322092222987, "grad_norm": 13.15676498413086, "learning_rate": 5.0283895388850655e-06, "loss": 0.6521, "step": 522750 }, { "epoch": 89.9518238128011, "grad_norm": 23.22406005859375, "learning_rate": 5.024088093599449e-06, "loss": 0.6578, "step": 522800 }, { "epoch": 89.96042670337233, "grad_norm": 12.175211906433105, "learning_rate": 5.019786648313834e-06, "loss": 0.6328, "step": 522850 }, { "epoch": 89.96902959394356, "grad_norm": 24.301956176757812, "learning_rate": 5.0154852030282175e-06, "loss": 0.649, "step": 522900 }, { "epoch": 89.9776324845148, "grad_norm": 15.746935844421387, "learning_rate": 5.011183757742602e-06, "loss": 0.7232, "step": 522950 }, { "epoch": 89.98623537508603, "grad_norm": 14.880098342895508, "learning_rate": 5.006882312456986e-06, "loss": 0.6094, "step": 523000 }, { "epoch": 89.99483826565726, "grad_norm": 17.67267608642578, "learning_rate": 5.0025808671713695e-06, "loss": 0.6209, "step": 523050 }, { "epoch": 90.0, "eval_accuracy": 0.4826191705386336, "eval_f1": 0.4751608635793483, "eval_f1_DuraRiadoRio_16x16": 0.4858938360743415, "eval_f1_Mole_16x16": 0.4895070766227428, "eval_f1_Quebrado_16x16": 0.5771230502599654, "eval_f1_RiadoRio_16x16": 0.26041363239149434, "eval_f1_RioFechado_16x16": 0.5628667225481978, "eval_loss": 3.3475661277770996, "eval_precision": 0.5515874163186696, "eval_precision_DuraRiadoRio_16x16": 0.3511111111111111, "eval_precision_Mole_16x16": 0.5590858416945373, "eval_precision_Quebrado_16x16": 0.8626943005181347, "eval_precision_RiadoRio_16x16": 0.44655344655344653, "eval_precision_RioFechado_16x16": 0.5384923817161187, "eval_recall": 0.48618073284891333, "eval_recall_DuraRiadoRio_16x16": 0.7886284722222222, "eval_recall_Mole_16x16": 0.4353298611111111, "eval_recall_Quebrado_16x16": 0.43359375, "eval_recall_RiadoRio_16x16": 0.18379934210526316, "eval_recall_RioFechado_16x16": 0.5895522388059702, "eval_runtime": 46.2259, "eval_samples_per_second": 251.417, "eval_steps_per_second": 15.727, "step": 523080 }, { "epoch": 90.00344115622849, "grad_norm": 17.359237670898438, "learning_rate": 4.998279421885753e-06, "loss": 0.6521, "step": 523100 }, { "epoch": 90.01204404679973, "grad_norm": 16.333799362182617, "learning_rate": 4.993977976600138e-06, "loss": 0.6368, "step": 523150 }, { "epoch": 90.02064693737096, "grad_norm": 21.275819778442383, "learning_rate": 4.9896765313145215e-06, "loss": 0.6449, "step": 523200 }, { "epoch": 90.0292498279422, "grad_norm": 16.91061019897461, "learning_rate": 4.985375086028906e-06, "loss": 0.6383, "step": 523250 }, { "epoch": 90.03785271851342, "grad_norm": 16.834848403930664, "learning_rate": 4.98107364074329e-06, "loss": 0.6182, "step": 523300 }, { "epoch": 90.04645560908465, "grad_norm": 30.559728622436523, "learning_rate": 4.976772195457674e-06, "loss": 0.6576, "step": 523350 }, { "epoch": 90.05505849965589, "grad_norm": 14.352315902709961, "learning_rate": 4.972470750172058e-06, "loss": 0.646, "step": 523400 }, { "epoch": 90.06366139022711, "grad_norm": 12.058995246887207, "learning_rate": 4.968169304886442e-06, "loss": 0.6107, "step": 523450 }, { "epoch": 90.07226428079835, "grad_norm": 16.795379638671875, "learning_rate": 4.963867859600826e-06, "loss": 0.6449, "step": 523500 }, { "epoch": 90.08086717136958, "grad_norm": 10.526067733764648, "learning_rate": 4.95956641431521e-06, "loss": 0.6598, "step": 523550 }, { "epoch": 90.08947006194082, "grad_norm": 9.531549453735352, "learning_rate": 4.955264969029595e-06, "loss": 0.6408, "step": 523600 }, { "epoch": 90.09807295251204, "grad_norm": 8.558094024658203, "learning_rate": 4.950963523743978e-06, "loss": 0.6412, "step": 523650 }, { "epoch": 90.10667584308328, "grad_norm": 18.655698776245117, "learning_rate": 4.946662078458363e-06, "loss": 0.5997, "step": 523700 }, { "epoch": 90.1152787336545, "grad_norm": 16.22745132446289, "learning_rate": 4.9423606331727466e-06, "loss": 0.6803, "step": 523750 }, { "epoch": 90.12388162422575, "grad_norm": 10.204887390136719, "learning_rate": 4.93805918788713e-06, "loss": 0.613, "step": 523800 }, { "epoch": 90.13248451479697, "grad_norm": 20.574447631835938, "learning_rate": 4.933757742601514e-06, "loss": 0.6738, "step": 523850 }, { "epoch": 90.14108740536821, "grad_norm": 24.071199417114258, "learning_rate": 4.9294562973158986e-06, "loss": 0.7091, "step": 523900 }, { "epoch": 90.14969029593944, "grad_norm": 17.750612258911133, "learning_rate": 4.925154852030282e-06, "loss": 0.6711, "step": 523950 }, { "epoch": 90.15829318651066, "grad_norm": 12.497634887695312, "learning_rate": 4.920853406744667e-06, "loss": 0.5936, "step": 524000 }, { "epoch": 90.1668960770819, "grad_norm": 8.95720386505127, "learning_rate": 4.9165519614590505e-06, "loss": 0.7045, "step": 524050 }, { "epoch": 90.17549896765313, "grad_norm": 14.632973670959473, "learning_rate": 4.912250516173435e-06, "loss": 0.6347, "step": 524100 }, { "epoch": 90.18410185822437, "grad_norm": 21.118640899658203, "learning_rate": 4.907949070887819e-06, "loss": 0.6397, "step": 524150 }, { "epoch": 90.19270474879559, "grad_norm": 9.003459930419922, "learning_rate": 4.9036476256022025e-06, "loss": 0.7056, "step": 524200 }, { "epoch": 90.20130763936683, "grad_norm": 13.693758964538574, "learning_rate": 4.899346180316586e-06, "loss": 0.6832, "step": 524250 }, { "epoch": 90.20991052993806, "grad_norm": 19.318870544433594, "learning_rate": 4.895044735030971e-06, "loss": 0.6856, "step": 524300 }, { "epoch": 90.2185134205093, "grad_norm": 15.309004783630371, "learning_rate": 4.8907432897453545e-06, "loss": 0.6385, "step": 524350 }, { "epoch": 90.22711631108052, "grad_norm": 17.24470329284668, "learning_rate": 4.886441844459739e-06, "loss": 0.5853, "step": 524400 }, { "epoch": 90.23571920165176, "grad_norm": 19.854345321655273, "learning_rate": 4.882140399174123e-06, "loss": 0.6488, "step": 524450 }, { "epoch": 90.24432209222299, "grad_norm": 10.654064178466797, "learning_rate": 4.8778389538885065e-06, "loss": 0.7131, "step": 524500 }, { "epoch": 90.25292498279421, "grad_norm": 8.675409317016602, "learning_rate": 4.87353750860289e-06, "loss": 0.6538, "step": 524550 }, { "epoch": 90.26152787336545, "grad_norm": 13.339293479919434, "learning_rate": 4.869236063317275e-06, "loss": 0.6761, "step": 524600 }, { "epoch": 90.27013076393668, "grad_norm": 12.558691024780273, "learning_rate": 4.8649346180316585e-06, "loss": 0.6737, "step": 524650 }, { "epoch": 90.27873365450792, "grad_norm": 12.70842456817627, "learning_rate": 4.860633172746043e-06, "loss": 0.63, "step": 524700 }, { "epoch": 90.28733654507914, "grad_norm": 17.160938262939453, "learning_rate": 4.856331727460427e-06, "loss": 0.6237, "step": 524750 }, { "epoch": 90.29593943565038, "grad_norm": 12.80269718170166, "learning_rate": 4.852030282174811e-06, "loss": 0.6222, "step": 524800 }, { "epoch": 90.30454232622161, "grad_norm": 8.821508407592773, "learning_rate": 4.847728836889195e-06, "loss": 0.6549, "step": 524850 }, { "epoch": 90.31314521679285, "grad_norm": 16.499786376953125, "learning_rate": 4.843427391603579e-06, "loss": 0.662, "step": 524900 }, { "epoch": 90.32174810736407, "grad_norm": 15.386077880859375, "learning_rate": 4.8391259463179625e-06, "loss": 0.6188, "step": 524950 }, { "epoch": 90.33035099793531, "grad_norm": 15.575333595275879, "learning_rate": 4.834824501032347e-06, "loss": 0.6522, "step": 525000 }, { "epoch": 90.33895388850654, "grad_norm": 13.318735122680664, "learning_rate": 4.830523055746731e-06, "loss": 0.5915, "step": 525050 }, { "epoch": 90.34755677907776, "grad_norm": 16.374019622802734, "learning_rate": 4.826221610461115e-06, "loss": 0.6433, "step": 525100 }, { "epoch": 90.356159669649, "grad_norm": 12.44859790802002, "learning_rate": 4.821920165175499e-06, "loss": 0.6242, "step": 525150 }, { "epoch": 90.36476256022023, "grad_norm": 10.295659065246582, "learning_rate": 4.817618719889884e-06, "loss": 0.6746, "step": 525200 }, { "epoch": 90.37336545079147, "grad_norm": 11.073135375976562, "learning_rate": 4.813317274604267e-06, "loss": 0.6684, "step": 525250 }, { "epoch": 90.3819683413627, "grad_norm": 15.629142761230469, "learning_rate": 4.809015829318651e-06, "loss": 0.6816, "step": 525300 }, { "epoch": 90.39057123193393, "grad_norm": 26.5439395904541, "learning_rate": 4.8047143840330356e-06, "loss": 0.6258, "step": 525350 }, { "epoch": 90.39917412250516, "grad_norm": 11.751750946044922, "learning_rate": 4.800412938747419e-06, "loss": 0.7088, "step": 525400 }, { "epoch": 90.4077770130764, "grad_norm": 9.42908763885498, "learning_rate": 4.796111493461804e-06, "loss": 0.6474, "step": 525450 }, { "epoch": 90.41637990364762, "grad_norm": 21.1853084564209, "learning_rate": 4.7918100481761876e-06, "loss": 0.6368, "step": 525500 }, { "epoch": 90.42498279421886, "grad_norm": 13.314969062805176, "learning_rate": 4.787508602890572e-06, "loss": 0.6114, "step": 525550 }, { "epoch": 90.43358568479009, "grad_norm": 15.74174690246582, "learning_rate": 4.783207157604956e-06, "loss": 0.6927, "step": 525600 }, { "epoch": 90.44218857536133, "grad_norm": 15.825530052185059, "learning_rate": 4.7789057123193395e-06, "loss": 0.6144, "step": 525650 }, { "epoch": 90.45079146593255, "grad_norm": 14.370680809020996, "learning_rate": 4.774604267033723e-06, "loss": 0.5859, "step": 525700 }, { "epoch": 90.45939435650378, "grad_norm": 15.000802040100098, "learning_rate": 4.770302821748108e-06, "loss": 0.6545, "step": 525750 }, { "epoch": 90.46799724707502, "grad_norm": 17.08260154724121, "learning_rate": 4.7660013764624915e-06, "loss": 0.6671, "step": 525800 }, { "epoch": 90.47660013764624, "grad_norm": 10.338700294494629, "learning_rate": 4.761699931176876e-06, "loss": 0.6597, "step": 525850 }, { "epoch": 90.48520302821748, "grad_norm": 19.114459991455078, "learning_rate": 4.75739848589126e-06, "loss": 0.5572, "step": 525900 }, { "epoch": 90.49380591878871, "grad_norm": 20.250837326049805, "learning_rate": 4.753097040605644e-06, "loss": 0.6504, "step": 525950 }, { "epoch": 90.50240880935995, "grad_norm": 13.405237197875977, "learning_rate": 4.748795595320027e-06, "loss": 0.6337, "step": 526000 }, { "epoch": 90.51101169993117, "grad_norm": 15.693138122558594, "learning_rate": 4.744494150034412e-06, "loss": 0.6222, "step": 526050 }, { "epoch": 90.51961459050241, "grad_norm": 23.585004806518555, "learning_rate": 4.7401927047487955e-06, "loss": 0.6454, "step": 526100 }, { "epoch": 90.52821748107364, "grad_norm": 18.137454986572266, "learning_rate": 4.73589125946318e-06, "loss": 0.5911, "step": 526150 }, { "epoch": 90.53682037164488, "grad_norm": 17.426673889160156, "learning_rate": 4.731589814177564e-06, "loss": 0.6593, "step": 526200 }, { "epoch": 90.5454232622161, "grad_norm": 15.01510238647461, "learning_rate": 4.727288368891948e-06, "loss": 0.689, "step": 526250 }, { "epoch": 90.55402615278733, "grad_norm": 22.1403865814209, "learning_rate": 4.722986923606332e-06, "loss": 0.639, "step": 526300 }, { "epoch": 90.56262904335857, "grad_norm": 16.931982040405273, "learning_rate": 4.718685478320716e-06, "loss": 0.6388, "step": 526350 }, { "epoch": 90.5712319339298, "grad_norm": 6.255841255187988, "learning_rate": 4.7143840330350995e-06, "loss": 0.6583, "step": 526400 }, { "epoch": 90.57983482450103, "grad_norm": 10.807768821716309, "learning_rate": 4.710082587749484e-06, "loss": 0.632, "step": 526450 }, { "epoch": 90.58843771507226, "grad_norm": 13.95446491241455, "learning_rate": 4.705781142463868e-06, "loss": 0.6617, "step": 526500 }, { "epoch": 90.5970406056435, "grad_norm": 10.636712074279785, "learning_rate": 4.701479697178252e-06, "loss": 0.6282, "step": 526550 }, { "epoch": 90.60564349621473, "grad_norm": 14.647695541381836, "learning_rate": 4.697178251892636e-06, "loss": 0.6165, "step": 526600 }, { "epoch": 90.61424638678596, "grad_norm": 14.735116004943848, "learning_rate": 4.692876806607021e-06, "loss": 0.6345, "step": 526650 }, { "epoch": 90.62284927735719, "grad_norm": 26.96714973449707, "learning_rate": 4.688575361321404e-06, "loss": 0.6564, "step": 526700 }, { "epoch": 90.63145216792843, "grad_norm": 16.157268524169922, "learning_rate": 4.684273916035788e-06, "loss": 0.6266, "step": 526750 }, { "epoch": 90.64005505849966, "grad_norm": 11.738672256469727, "learning_rate": 4.679972470750172e-06, "loss": 0.6443, "step": 526800 }, { "epoch": 90.6486579490709, "grad_norm": 18.685632705688477, "learning_rate": 4.675671025464556e-06, "loss": 0.6824, "step": 526850 }, { "epoch": 90.65726083964212, "grad_norm": 28.05509376525879, "learning_rate": 4.67136958017894e-06, "loss": 0.6915, "step": 526900 }, { "epoch": 90.66586373021335, "grad_norm": 8.563732147216797, "learning_rate": 4.6670681348933246e-06, "loss": 0.6844, "step": 526950 }, { "epoch": 90.67446662078459, "grad_norm": 15.093868255615234, "learning_rate": 4.662766689607708e-06, "loss": 0.6434, "step": 527000 }, { "epoch": 90.68306951135581, "grad_norm": 13.29855728149414, "learning_rate": 4.658465244322093e-06, "loss": 0.6592, "step": 527050 }, { "epoch": 90.69167240192705, "grad_norm": 15.067900657653809, "learning_rate": 4.6541637990364765e-06, "loss": 0.5919, "step": 527100 }, { "epoch": 90.70027529249828, "grad_norm": 16.663639068603516, "learning_rate": 4.64986235375086e-06, "loss": 0.606, "step": 527150 }, { "epoch": 90.70887818306952, "grad_norm": 13.032881736755371, "learning_rate": 4.645560908465245e-06, "loss": 0.5963, "step": 527200 }, { "epoch": 90.71748107364074, "grad_norm": 12.473075866699219, "learning_rate": 4.6412594631796285e-06, "loss": 0.6393, "step": 527250 }, { "epoch": 90.72608396421198, "grad_norm": 34.21964645385742, "learning_rate": 4.636958017894013e-06, "loss": 0.6312, "step": 527300 }, { "epoch": 90.7346868547832, "grad_norm": 9.040145874023438, "learning_rate": 4.632656572608397e-06, "loss": 0.6519, "step": 527350 }, { "epoch": 90.74328974535445, "grad_norm": 19.77798843383789, "learning_rate": 4.628355127322781e-06, "loss": 0.6542, "step": 527400 }, { "epoch": 90.75189263592567, "grad_norm": 28.8294734954834, "learning_rate": 4.624053682037165e-06, "loss": 0.6684, "step": 527450 }, { "epoch": 90.7604955264969, "grad_norm": 17.744747161865234, "learning_rate": 4.619752236751549e-06, "loss": 0.6858, "step": 527500 }, { "epoch": 90.76909841706814, "grad_norm": 17.237951278686523, "learning_rate": 4.6154507914659325e-06, "loss": 0.6553, "step": 527550 }, { "epoch": 90.77770130763936, "grad_norm": 14.964247703552246, "learning_rate": 4.611149346180317e-06, "loss": 0.5678, "step": 527600 }, { "epoch": 90.7863041982106, "grad_norm": 15.100150108337402, "learning_rate": 4.606847900894701e-06, "loss": 0.6608, "step": 527650 }, { "epoch": 90.79490708878183, "grad_norm": 16.663393020629883, "learning_rate": 4.602546455609085e-06, "loss": 0.6188, "step": 527700 }, { "epoch": 90.80350997935307, "grad_norm": 19.521106719970703, "learning_rate": 4.598245010323469e-06, "loss": 0.6178, "step": 527750 }, { "epoch": 90.81211286992429, "grad_norm": 12.903666496276855, "learning_rate": 4.593943565037853e-06, "loss": 0.6728, "step": 527800 }, { "epoch": 90.82071576049553, "grad_norm": 12.436702728271484, "learning_rate": 4.5896421197522365e-06, "loss": 0.6347, "step": 527850 }, { "epoch": 90.82931865106676, "grad_norm": 17.412353515625, "learning_rate": 4.585340674466621e-06, "loss": 0.6919, "step": 527900 }, { "epoch": 90.837921541638, "grad_norm": 9.257006645202637, "learning_rate": 4.581039229181005e-06, "loss": 0.7128, "step": 527950 }, { "epoch": 90.84652443220922, "grad_norm": 19.78324317932129, "learning_rate": 4.576737783895389e-06, "loss": 0.7076, "step": 528000 }, { "epoch": 90.85512732278045, "grad_norm": 21.57174301147461, "learning_rate": 4.572436338609773e-06, "loss": 0.6252, "step": 528050 }, { "epoch": 90.86373021335169, "grad_norm": 17.22938346862793, "learning_rate": 4.568134893324158e-06, "loss": 0.6859, "step": 528100 }, { "epoch": 90.87233310392291, "grad_norm": 20.37257194519043, "learning_rate": 4.563833448038541e-06, "loss": 0.6645, "step": 528150 }, { "epoch": 90.88093599449415, "grad_norm": 15.670124053955078, "learning_rate": 4.559532002752925e-06, "loss": 0.6645, "step": 528200 }, { "epoch": 90.88953888506538, "grad_norm": 10.00631332397461, "learning_rate": 4.555230557467309e-06, "loss": 0.6299, "step": 528250 }, { "epoch": 90.89814177563662, "grad_norm": 15.779656410217285, "learning_rate": 4.550929112181693e-06, "loss": 0.6977, "step": 528300 }, { "epoch": 90.90674466620784, "grad_norm": 19.3257999420166, "learning_rate": 4.546627666896077e-06, "loss": 0.6063, "step": 528350 }, { "epoch": 90.91534755677908, "grad_norm": 21.437255859375, "learning_rate": 4.5423262216104616e-06, "loss": 0.6141, "step": 528400 }, { "epoch": 90.92395044735031, "grad_norm": 12.977441787719727, "learning_rate": 4.538024776324845e-06, "loss": 0.5996, "step": 528450 }, { "epoch": 90.93255333792155, "grad_norm": 18.182331085205078, "learning_rate": 4.53372333103923e-06, "loss": 0.6363, "step": 528500 }, { "epoch": 90.94115622849277, "grad_norm": 11.298503875732422, "learning_rate": 4.5294218857536135e-06, "loss": 0.6421, "step": 528550 }, { "epoch": 90.94975911906401, "grad_norm": 8.924979209899902, "learning_rate": 4.525120440467997e-06, "loss": 0.6583, "step": 528600 }, { "epoch": 90.95836200963524, "grad_norm": 13.671965599060059, "learning_rate": 4.520818995182381e-06, "loss": 0.6877, "step": 528650 }, { "epoch": 90.96696490020646, "grad_norm": 14.2829008102417, "learning_rate": 4.5165175498967655e-06, "loss": 0.6717, "step": 528700 }, { "epoch": 90.9755677907777, "grad_norm": 11.71275520324707, "learning_rate": 4.512216104611149e-06, "loss": 0.6856, "step": 528750 }, { "epoch": 90.98417068134893, "grad_norm": 12.350844383239746, "learning_rate": 4.507914659325534e-06, "loss": 0.6448, "step": 528800 }, { "epoch": 90.99277357192017, "grad_norm": 23.85361671447754, "learning_rate": 4.5036132140399175e-06, "loss": 0.6806, "step": 528850 }, { "epoch": 91.0, "eval_accuracy": 0.525124763379797, "eval_f1": 0.5099441583376535, "eval_f1_DuraRiadoRio_16x16": 0.3702786377708978, "eval_f1_Mole_16x16": 0.5303436509870826, "eval_f1_Quebrado_16x16": 0.6641234610716144, "eval_f1_RiadoRio_16x16": 0.4165005978477481, "eval_f1_RioFechado_16x16": 0.5684744440109247, "eval_loss": 3.086801290512085, "eval_precision": 0.543848716366707, "eval_precision_DuraRiadoRio_16x16": 0.6457883369330454, "eval_precision_Mole_16x16": 0.6047804335742079, "eval_precision_Quebrado_16x16": 0.5529887380883627, "eval_precision_RiadoRio_16x16": 0.40409899458623355, "eval_precision_RioFechado_16x16": 0.5115870786516854, "eval_recall": 0.526443532948005, "eval_recall_DuraRiadoRio_16x16": 0.2595486111111111, "eval_recall_Mole_16x16": 0.4722222222222222, "eval_recall_Quebrado_16x16": 0.8311631944444444, "eval_recall_RiadoRio_16x16": 0.4296875, "eval_recall_RioFechado_16x16": 0.6395961369622476, "eval_runtime": 45.8433, "eval_samples_per_second": 253.516, "eval_steps_per_second": 15.858, "step": 528892 }, { "epoch": 91.0013764624914, "grad_norm": 10.79364013671875, "learning_rate": 4.499311768754302e-06, "loss": 0.6899, "step": 528900 }, { "epoch": 91.00997935306263, "grad_norm": 11.927083969116211, "learning_rate": 4.495010323468685e-06, "loss": 0.6784, "step": 528950 }, { "epoch": 91.01858224363386, "grad_norm": 8.84472942352295, "learning_rate": 4.4907088781830695e-06, "loss": 0.5983, "step": 529000 }, { "epoch": 91.0271851342051, "grad_norm": 20.97145652770996, "learning_rate": 4.486407432897454e-06, "loss": 0.594, "step": 529050 }, { "epoch": 91.03578802477632, "grad_norm": 16.416292190551758, "learning_rate": 4.482105987611838e-06, "loss": 0.6481, "step": 529100 }, { "epoch": 91.04439091534756, "grad_norm": 12.899629592895508, "learning_rate": 4.477804542326222e-06, "loss": 0.6263, "step": 529150 }, { "epoch": 91.05299380591879, "grad_norm": 10.82292652130127, "learning_rate": 4.473503097040606e-06, "loss": 0.6315, "step": 529200 }, { "epoch": 91.06159669649001, "grad_norm": 17.37244415283203, "learning_rate": 4.469201651754991e-06, "loss": 0.605, "step": 529250 }, { "epoch": 91.07019958706125, "grad_norm": 12.814189910888672, "learning_rate": 4.4649002064693735e-06, "loss": 0.6149, "step": 529300 }, { "epoch": 91.07880247763248, "grad_norm": 8.92855167388916, "learning_rate": 4.460598761183758e-06, "loss": 0.6548, "step": 529350 }, { "epoch": 91.08740536820372, "grad_norm": 16.3372859954834, "learning_rate": 4.456297315898142e-06, "loss": 0.6759, "step": 529400 }, { "epoch": 91.09600825877494, "grad_norm": 16.4509334564209, "learning_rate": 4.451995870612526e-06, "loss": 0.5951, "step": 529450 }, { "epoch": 91.10461114934618, "grad_norm": 23.012537002563477, "learning_rate": 4.44769442532691e-06, "loss": 0.6421, "step": 529500 }, { "epoch": 91.11321403991741, "grad_norm": 17.18221664428711, "learning_rate": 4.443392980041295e-06, "loss": 0.6575, "step": 529550 }, { "epoch": 91.12181693048865, "grad_norm": 14.391858100891113, "learning_rate": 4.439091534755678e-06, "loss": 0.6436, "step": 529600 }, { "epoch": 91.13041982105987, "grad_norm": 12.542521476745605, "learning_rate": 4.434790089470062e-06, "loss": 0.6523, "step": 529650 }, { "epoch": 91.13902271163111, "grad_norm": 21.236225128173828, "learning_rate": 4.430488644184446e-06, "loss": 0.6356, "step": 529700 }, { "epoch": 91.14762560220234, "grad_norm": 17.490989685058594, "learning_rate": 4.42618719889883e-06, "loss": 0.6503, "step": 529750 }, { "epoch": 91.15622849277358, "grad_norm": 15.475342750549316, "learning_rate": 4.421885753613214e-06, "loss": 0.6302, "step": 529800 }, { "epoch": 91.1648313833448, "grad_norm": 12.619339942932129, "learning_rate": 4.4175843083275986e-06, "loss": 0.6028, "step": 529850 }, { "epoch": 91.17343427391603, "grad_norm": 29.783164978027344, "learning_rate": 4.413282863041982e-06, "loss": 0.6829, "step": 529900 }, { "epoch": 91.18203716448727, "grad_norm": 20.432165145874023, "learning_rate": 4.408981417756367e-06, "loss": 0.6479, "step": 529950 }, { "epoch": 91.1906400550585, "grad_norm": 15.999459266662598, "learning_rate": 4.4046799724707506e-06, "loss": 0.589, "step": 530000 }, { "epoch": 91.19924294562973, "grad_norm": 25.663360595703125, "learning_rate": 4.400378527185134e-06, "loss": 0.6813, "step": 530050 }, { "epoch": 91.20784583620096, "grad_norm": 14.665979385375977, "learning_rate": 4.396077081899518e-06, "loss": 0.6494, "step": 530100 }, { "epoch": 91.2164487267722, "grad_norm": 17.721649169921875, "learning_rate": 4.3917756366139025e-06, "loss": 0.7161, "step": 530150 }, { "epoch": 91.22505161734342, "grad_norm": 18.85146141052246, "learning_rate": 4.387474191328286e-06, "loss": 0.667, "step": 530200 }, { "epoch": 91.23365450791466, "grad_norm": 25.01194953918457, "learning_rate": 4.383172746042671e-06, "loss": 0.6198, "step": 530250 }, { "epoch": 91.24225739848589, "grad_norm": 17.388134002685547, "learning_rate": 4.3788713007570545e-06, "loss": 0.6716, "step": 530300 }, { "epoch": 91.25086028905713, "grad_norm": 6.905690670013428, "learning_rate": 4.374569855471439e-06, "loss": 0.66, "step": 530350 }, { "epoch": 91.25946317962836, "grad_norm": 12.766855239868164, "learning_rate": 4.370268410185823e-06, "loss": 0.6434, "step": 530400 }, { "epoch": 91.26806607019958, "grad_norm": 22.206947326660156, "learning_rate": 4.3659669649002065e-06, "loss": 0.6867, "step": 530450 }, { "epoch": 91.27666896077082, "grad_norm": 15.024200439453125, "learning_rate": 4.36166551961459e-06, "loss": 0.6369, "step": 530500 }, { "epoch": 91.28527185134205, "grad_norm": 7.3987531661987305, "learning_rate": 4.357364074328975e-06, "loss": 0.644, "step": 530550 }, { "epoch": 91.29387474191329, "grad_norm": 10.273664474487305, "learning_rate": 4.3530626290433585e-06, "loss": 0.703, "step": 530600 }, { "epoch": 91.30247763248451, "grad_norm": 10.628952026367188, "learning_rate": 4.348761183757743e-06, "loss": 0.6343, "step": 530650 }, { "epoch": 91.31108052305575, "grad_norm": 22.117095947265625, "learning_rate": 4.344459738472127e-06, "loss": 0.6354, "step": 530700 }, { "epoch": 91.31968341362698, "grad_norm": 9.768302917480469, "learning_rate": 4.3401582931865105e-06, "loss": 0.5959, "step": 530750 }, { "epoch": 91.32828630419822, "grad_norm": 17.96776580810547, "learning_rate": 4.335856847900894e-06, "loss": 0.6628, "step": 530800 }, { "epoch": 91.33688919476944, "grad_norm": 10.253615379333496, "learning_rate": 4.331555402615279e-06, "loss": 0.6413, "step": 530850 }, { "epoch": 91.34549208534068, "grad_norm": 27.188655853271484, "learning_rate": 4.3272539573296625e-06, "loss": 0.6601, "step": 530900 }, { "epoch": 91.3540949759119, "grad_norm": 29.490827560424805, "learning_rate": 4.322952512044047e-06, "loss": 0.6418, "step": 530950 }, { "epoch": 91.36269786648313, "grad_norm": 15.188312530517578, "learning_rate": 4.318651066758432e-06, "loss": 0.6156, "step": 531000 }, { "epoch": 91.37130075705437, "grad_norm": 19.145797729492188, "learning_rate": 4.314349621472815e-06, "loss": 0.6417, "step": 531050 }, { "epoch": 91.3799036476256, "grad_norm": 15.43026351928711, "learning_rate": 4.310048176187199e-06, "loss": 0.6663, "step": 531100 }, { "epoch": 91.38850653819684, "grad_norm": 13.004819869995117, "learning_rate": 4.305746730901583e-06, "loss": 0.6041, "step": 531150 }, { "epoch": 91.39710942876806, "grad_norm": 17.322038650512695, "learning_rate": 4.301445285615967e-06, "loss": 0.6667, "step": 531200 }, { "epoch": 91.4057123193393, "grad_norm": 27.805456161499023, "learning_rate": 4.297143840330351e-06, "loss": 0.6319, "step": 531250 }, { "epoch": 91.41431520991053, "grad_norm": 18.579524993896484, "learning_rate": 4.2928423950447356e-06, "loss": 0.6141, "step": 531300 }, { "epoch": 91.42291810048177, "grad_norm": 17.87987518310547, "learning_rate": 4.288540949759119e-06, "loss": 0.5879, "step": 531350 }, { "epoch": 91.43152099105299, "grad_norm": 18.652137756347656, "learning_rate": 4.284239504473504e-06, "loss": 0.6901, "step": 531400 }, { "epoch": 91.44012388162423, "grad_norm": 14.148897171020508, "learning_rate": 4.2799380591878876e-06, "loss": 0.6424, "step": 531450 }, { "epoch": 91.44872677219546, "grad_norm": 20.6695499420166, "learning_rate": 4.275636613902271e-06, "loss": 0.624, "step": 531500 }, { "epoch": 91.4573296627667, "grad_norm": 14.146936416625977, "learning_rate": 4.271335168616655e-06, "loss": 0.6627, "step": 531550 }, { "epoch": 91.46593255333792, "grad_norm": 16.075271606445312, "learning_rate": 4.2670337233310395e-06, "loss": 0.6421, "step": 531600 }, { "epoch": 91.47453544390915, "grad_norm": 16.8681583404541, "learning_rate": 4.262732278045423e-06, "loss": 0.673, "step": 531650 }, { "epoch": 91.48313833448039, "grad_norm": 18.955556869506836, "learning_rate": 4.258430832759808e-06, "loss": 0.681, "step": 531700 }, { "epoch": 91.49174122505161, "grad_norm": 17.558759689331055, "learning_rate": 4.2541293874741915e-06, "loss": 0.6854, "step": 531750 }, { "epoch": 91.50034411562285, "grad_norm": 16.983102798461914, "learning_rate": 4.249827942188576e-06, "loss": 0.6408, "step": 531800 }, { "epoch": 91.50894700619408, "grad_norm": 17.29424285888672, "learning_rate": 4.24552649690296e-06, "loss": 0.627, "step": 531850 }, { "epoch": 91.51754989676532, "grad_norm": 9.923822402954102, "learning_rate": 4.2412250516173435e-06, "loss": 0.629, "step": 531900 }, { "epoch": 91.52615278733654, "grad_norm": 16.129379272460938, "learning_rate": 4.236923606331727e-06, "loss": 0.672, "step": 531950 }, { "epoch": 91.53475567790778, "grad_norm": 12.550492286682129, "learning_rate": 4.232622161046112e-06, "loss": 0.6506, "step": 532000 }, { "epoch": 91.54335856847901, "grad_norm": 16.956283569335938, "learning_rate": 4.2283207157604955e-06, "loss": 0.6915, "step": 532050 }, { "epoch": 91.55196145905025, "grad_norm": 16.624109268188477, "learning_rate": 4.22401927047488e-06, "loss": 0.6108, "step": 532100 }, { "epoch": 91.56056434962147, "grad_norm": 17.284515380859375, "learning_rate": 4.219717825189264e-06, "loss": 0.6655, "step": 532150 }, { "epoch": 91.5691672401927, "grad_norm": 24.25404167175293, "learning_rate": 4.215416379903648e-06, "loss": 0.7033, "step": 532200 }, { "epoch": 91.57777013076394, "grad_norm": 12.8054838180542, "learning_rate": 4.211114934618031e-06, "loss": 0.6388, "step": 532250 }, { "epoch": 91.58637302133516, "grad_norm": 24.07123374938965, "learning_rate": 4.206813489332416e-06, "loss": 0.6401, "step": 532300 }, { "epoch": 91.5949759119064, "grad_norm": 14.499510765075684, "learning_rate": 4.2025120440467995e-06, "loss": 0.6332, "step": 532350 }, { "epoch": 91.60357880247763, "grad_norm": 10.84012222290039, "learning_rate": 4.198210598761184e-06, "loss": 0.671, "step": 532400 }, { "epoch": 91.61218169304887, "grad_norm": 18.242111206054688, "learning_rate": 4.193909153475568e-06, "loss": 0.6416, "step": 532450 }, { "epoch": 91.6207845836201, "grad_norm": 12.388981819152832, "learning_rate": 4.189607708189952e-06, "loss": 0.6801, "step": 532500 }, { "epoch": 91.62938747419133, "grad_norm": 15.90573501586914, "learning_rate": 4.185306262904336e-06, "loss": 0.6161, "step": 532550 }, { "epoch": 91.63799036476256, "grad_norm": 29.72260093688965, "learning_rate": 4.18100481761872e-06, "loss": 0.6706, "step": 532600 }, { "epoch": 91.6465932553338, "grad_norm": 7.201624870300293, "learning_rate": 4.1767033723331035e-06, "loss": 0.6206, "step": 532650 }, { "epoch": 91.65519614590502, "grad_norm": 9.652589797973633, "learning_rate": 4.172401927047488e-06, "loss": 0.6613, "step": 532700 }, { "epoch": 91.66379903647626, "grad_norm": 11.060710906982422, "learning_rate": 4.168100481761872e-06, "loss": 0.6092, "step": 532750 }, { "epoch": 91.67240192704749, "grad_norm": 22.413869857788086, "learning_rate": 4.163799036476256e-06, "loss": 0.6804, "step": 532800 }, { "epoch": 91.68100481761871, "grad_norm": 12.821415901184082, "learning_rate": 4.159497591190641e-06, "loss": 0.5951, "step": 532850 }, { "epoch": 91.68960770818995, "grad_norm": 29.472434997558594, "learning_rate": 4.1551961459050246e-06, "loss": 0.6811, "step": 532900 }, { "epoch": 91.69821059876118, "grad_norm": 8.365569114685059, "learning_rate": 4.150894700619408e-06, "loss": 0.6658, "step": 532950 }, { "epoch": 91.70681348933242, "grad_norm": 18.185260772705078, "learning_rate": 4.146593255333792e-06, "loss": 0.6623, "step": 533000 }, { "epoch": 91.71541637990364, "grad_norm": 16.221420288085938, "learning_rate": 4.1422918100481766e-06, "loss": 0.6759, "step": 533050 }, { "epoch": 91.72401927047488, "grad_norm": 20.376846313476562, "learning_rate": 4.13799036476256e-06, "loss": 0.641, "step": 533100 }, { "epoch": 91.73262216104611, "grad_norm": 26.929502487182617, "learning_rate": 4.133688919476945e-06, "loss": 0.6073, "step": 533150 }, { "epoch": 91.74122505161735, "grad_norm": 23.713558197021484, "learning_rate": 4.1293874741913285e-06, "loss": 0.6804, "step": 533200 }, { "epoch": 91.74982794218857, "grad_norm": 17.45433235168457, "learning_rate": 4.125086028905713e-06, "loss": 0.649, "step": 533250 }, { "epoch": 91.75843083275981, "grad_norm": 17.010202407836914, "learning_rate": 4.120784583620097e-06, "loss": 0.6581, "step": 533300 }, { "epoch": 91.76703372333104, "grad_norm": 16.25815200805664, "learning_rate": 4.1164831383344805e-06, "loss": 0.6309, "step": 533350 }, { "epoch": 91.77563661390226, "grad_norm": 21.816394805908203, "learning_rate": 4.112181693048864e-06, "loss": 0.6681, "step": 533400 }, { "epoch": 91.7842395044735, "grad_norm": 10.684185028076172, "learning_rate": 4.107880247763249e-06, "loss": 0.6904, "step": 533450 }, { "epoch": 91.79284239504473, "grad_norm": 13.495977401733398, "learning_rate": 4.1035788024776325e-06, "loss": 0.6078, "step": 533500 }, { "epoch": 91.80144528561597, "grad_norm": 15.982829093933105, "learning_rate": 4.099277357192017e-06, "loss": 0.6882, "step": 533550 }, { "epoch": 91.8100481761872, "grad_norm": 28.39436149597168, "learning_rate": 4.094975911906401e-06, "loss": 0.6652, "step": 533600 }, { "epoch": 91.81865106675843, "grad_norm": 19.127498626708984, "learning_rate": 4.090674466620785e-06, "loss": 0.7034, "step": 533650 }, { "epoch": 91.82725395732966, "grad_norm": 17.877779006958008, "learning_rate": 4.086373021335169e-06, "loss": 0.6503, "step": 533700 }, { "epoch": 91.8358568479009, "grad_norm": 17.781864166259766, "learning_rate": 4.082071576049553e-06, "loss": 0.6819, "step": 533750 }, { "epoch": 91.84445973847212, "grad_norm": 35.09139633178711, "learning_rate": 4.0777701307639365e-06, "loss": 0.6371, "step": 533800 }, { "epoch": 91.85306262904336, "grad_norm": 19.20825958251953, "learning_rate": 4.073468685478321e-06, "loss": 0.6522, "step": 533850 }, { "epoch": 91.86166551961459, "grad_norm": 7.191163539886475, "learning_rate": 4.069167240192705e-06, "loss": 0.6743, "step": 533900 }, { "epoch": 91.87026841018582, "grad_norm": 13.25834846496582, "learning_rate": 4.064865794907089e-06, "loss": 0.6, "step": 533950 }, { "epoch": 91.87887130075705, "grad_norm": 16.630268096923828, "learning_rate": 4.060564349621473e-06, "loss": 0.6476, "step": 534000 }, { "epoch": 91.88747419132828, "grad_norm": 16.569135665893555, "learning_rate": 4.056262904335857e-06, "loss": 0.6801, "step": 534050 }, { "epoch": 91.89607708189952, "grad_norm": 20.58182144165039, "learning_rate": 4.0519614590502405e-06, "loss": 0.6364, "step": 534100 }, { "epoch": 91.90467997247075, "grad_norm": 9.484567642211914, "learning_rate": 4.047660013764625e-06, "loss": 0.6451, "step": 534150 }, { "epoch": 91.91328286304199, "grad_norm": 22.654138565063477, "learning_rate": 4.043358568479009e-06, "loss": 0.626, "step": 534200 }, { "epoch": 91.92188575361321, "grad_norm": 32.42964553833008, "learning_rate": 4.039057123193393e-06, "loss": 0.6749, "step": 534250 }, { "epoch": 91.93048864418445, "grad_norm": 23.248300552368164, "learning_rate": 4.034755677907777e-06, "loss": 0.6506, "step": 534300 }, { "epoch": 91.93909153475568, "grad_norm": 23.79932975769043, "learning_rate": 4.0304542326221616e-06, "loss": 0.6703, "step": 534350 }, { "epoch": 91.94769442532692, "grad_norm": 13.981221199035645, "learning_rate": 4.026152787336545e-06, "loss": 0.6263, "step": 534400 }, { "epoch": 91.95629731589814, "grad_norm": 19.17823600769043, "learning_rate": 4.021851342050929e-06, "loss": 0.6497, "step": 534450 }, { "epoch": 91.96490020646938, "grad_norm": 19.241506576538086, "learning_rate": 4.017549896765313e-06, "loss": 0.7093, "step": 534500 }, { "epoch": 91.9735030970406, "grad_norm": 30.303918838500977, "learning_rate": 4.013248451479697e-06, "loss": 0.6325, "step": 534550 }, { "epoch": 91.98210598761183, "grad_norm": 9.230043411254883, "learning_rate": 4.008947006194081e-06, "loss": 0.6818, "step": 534600 }, { "epoch": 91.99070887818307, "grad_norm": 17.576316833496094, "learning_rate": 4.0046455609084655e-06, "loss": 0.6049, "step": 534650 }, { "epoch": 91.9993117687543, "grad_norm": 16.243852615356445, "learning_rate": 4.00034411562285e-06, "loss": 0.6452, "step": 534700 }, { "epoch": 92.0, "eval_accuracy": 0.5531750129065566, "eval_f1": 0.536397961224258, "eval_f1_DuraRiadoRio_16x16": 0.4611832611832612, "eval_f1_Mole_16x16": 0.44203347799132053, "eval_f1_Quebrado_16x16": 0.6609695973705834, "eval_f1_RiadoRio_16x16": 0.5264241592312972, "eval_f1_RioFechado_16x16": 0.5913793103448276, "eval_loss": 2.9336140155792236, "eval_precision": 0.6051920283870733, "eval_precision_DuraRiadoRio_16x16": 0.6881998277347114, "eval_precision_Mole_16x16": 0.7733188720173536, "eval_precision_Quebrado_16x16": 0.5318698756942608, "eval_precision_RiadoRio_16x16": 0.45170789163722025, "eval_precision_RioFechado_16x16": 0.5808636748518204, "eval_recall": 0.5524238288369813, "eval_recall_DuraRiadoRio_16x16": 0.3467881944444444, "eval_recall_Mole_16x16": 0.3094618055555556, "eval_recall_Quebrado_16x16": 0.8728298611111112, "eval_recall_RiadoRio_16x16": 0.6307565789473685, "eval_recall_RioFechado_16x16": 0.6022827041264267, "eval_runtime": 45.9707, "eval_samples_per_second": 252.813, "eval_steps_per_second": 15.814, "step": 534704 }, { "epoch": 92.00791465932554, "grad_norm": 19.380552291870117, "learning_rate": 3.996042670337234e-06, "loss": 0.7639, "step": 534750 }, { "epoch": 92.01651754989676, "grad_norm": 19.293907165527344, "learning_rate": 3.9917412250516175e-06, "loss": 0.6714, "step": 534800 }, { "epoch": 92.025120440468, "grad_norm": 6.0627593994140625, "learning_rate": 3.987439779766001e-06, "loss": 0.6261, "step": 534850 }, { "epoch": 92.03372333103923, "grad_norm": 19.168664932250977, "learning_rate": 3.983138334480386e-06, "loss": 0.7111, "step": 534900 }, { "epoch": 92.04232622161047, "grad_norm": 9.332268714904785, "learning_rate": 3.9788368891947695e-06, "loss": 0.6047, "step": 534950 }, { "epoch": 92.05092911218169, "grad_norm": 10.839081764221191, "learning_rate": 3.974535443909154e-06, "loss": 0.6148, "step": 535000 }, { "epoch": 92.05953200275293, "grad_norm": 19.556066513061523, "learning_rate": 3.970233998623538e-06, "loss": 0.6214, "step": 535050 }, { "epoch": 92.06813489332416, "grad_norm": 17.430871963500977, "learning_rate": 3.965932553337922e-06, "loss": 0.6615, "step": 535100 }, { "epoch": 92.07673778389538, "grad_norm": 12.054634094238281, "learning_rate": 3.961631108052306e-06, "loss": 0.6287, "step": 535150 }, { "epoch": 92.08534067446662, "grad_norm": 15.099161148071289, "learning_rate": 3.95732966276669e-06, "loss": 0.595, "step": 535200 }, { "epoch": 92.09394356503785, "grad_norm": 11.678991317749023, "learning_rate": 3.9530282174810735e-06, "loss": 0.6434, "step": 535250 }, { "epoch": 92.10254645560909, "grad_norm": 8.879698753356934, "learning_rate": 3.948726772195458e-06, "loss": 0.6, "step": 535300 }, { "epoch": 92.11114934618031, "grad_norm": 24.55936622619629, "learning_rate": 3.944425326909842e-06, "loss": 0.643, "step": 535350 }, { "epoch": 92.11975223675155, "grad_norm": 16.104127883911133, "learning_rate": 3.940123881624226e-06, "loss": 0.6787, "step": 535400 }, { "epoch": 92.12835512732278, "grad_norm": 11.238005638122559, "learning_rate": 3.93582243633861e-06, "loss": 0.633, "step": 535450 }, { "epoch": 92.13695801789402, "grad_norm": 9.253297805786133, "learning_rate": 3.931520991052995e-06, "loss": 0.6189, "step": 535500 }, { "epoch": 92.14556090846524, "grad_norm": 14.67104721069336, "learning_rate": 3.9272195457673775e-06, "loss": 0.6048, "step": 535550 }, { "epoch": 92.15416379903648, "grad_norm": 11.426817893981934, "learning_rate": 3.922918100481762e-06, "loss": 0.5938, "step": 535600 }, { "epoch": 92.16276668960771, "grad_norm": 15.937711715698242, "learning_rate": 3.918616655196146e-06, "loss": 0.6614, "step": 535650 }, { "epoch": 92.17136958017895, "grad_norm": 16.948022842407227, "learning_rate": 3.91431520991053e-06, "loss": 0.6975, "step": 535700 }, { "epoch": 92.17997247075017, "grad_norm": 9.510089874267578, "learning_rate": 3.910013764624914e-06, "loss": 0.6581, "step": 535750 }, { "epoch": 92.1885753613214, "grad_norm": 11.57341480255127, "learning_rate": 3.9057123193392986e-06, "loss": 0.6246, "step": 535800 }, { "epoch": 92.19717825189264, "grad_norm": 18.09457015991211, "learning_rate": 3.901410874053682e-06, "loss": 0.6555, "step": 535850 }, { "epoch": 92.20578114246386, "grad_norm": 17.30645179748535, "learning_rate": 3.897109428768066e-06, "loss": 0.6321, "step": 535900 }, { "epoch": 92.2143840330351, "grad_norm": 11.748177528381348, "learning_rate": 3.89280798348245e-06, "loss": 0.6423, "step": 535950 }, { "epoch": 92.22298692360633, "grad_norm": 19.39117431640625, "learning_rate": 3.888506538196834e-06, "loss": 0.6283, "step": 536000 }, { "epoch": 92.23158981417757, "grad_norm": 21.968292236328125, "learning_rate": 3.884205092911218e-06, "loss": 0.6162, "step": 536050 }, { "epoch": 92.2401927047488, "grad_norm": 16.46332550048828, "learning_rate": 3.8799036476256025e-06, "loss": 0.6247, "step": 536100 }, { "epoch": 92.24879559532003, "grad_norm": 14.138916015625, "learning_rate": 3.875602202339986e-06, "loss": 0.6676, "step": 536150 }, { "epoch": 92.25739848589126, "grad_norm": 15.860625267028809, "learning_rate": 3.871300757054371e-06, "loss": 0.6478, "step": 536200 }, { "epoch": 92.2660013764625, "grad_norm": 24.913585662841797, "learning_rate": 3.8669993117687545e-06, "loss": 0.6292, "step": 536250 }, { "epoch": 92.27460426703372, "grad_norm": 9.32266902923584, "learning_rate": 3.862697866483138e-06, "loss": 0.6487, "step": 536300 }, { "epoch": 92.28320715760495, "grad_norm": 18.303829193115234, "learning_rate": 3.858396421197522e-06, "loss": 0.6406, "step": 536350 }, { "epoch": 92.29181004817619, "grad_norm": 20.079179763793945, "learning_rate": 3.8540949759119065e-06, "loss": 0.6404, "step": 536400 }, { "epoch": 92.30041293874741, "grad_norm": 12.631620407104492, "learning_rate": 3.84979353062629e-06, "loss": 0.654, "step": 536450 }, { "epoch": 92.30901582931865, "grad_norm": 9.691442489624023, "learning_rate": 3.845492085340675e-06, "loss": 0.6446, "step": 536500 }, { "epoch": 92.31761871988988, "grad_norm": 19.264074325561523, "learning_rate": 3.841190640055059e-06, "loss": 0.6361, "step": 536550 }, { "epoch": 92.32622161046112, "grad_norm": 15.86652946472168, "learning_rate": 3.836889194769443e-06, "loss": 0.5847, "step": 536600 }, { "epoch": 92.33482450103234, "grad_norm": 9.661989212036133, "learning_rate": 3.832587749483827e-06, "loss": 0.6525, "step": 536650 }, { "epoch": 92.34342739160358, "grad_norm": 24.333980560302734, "learning_rate": 3.8282863041982105e-06, "loss": 0.6509, "step": 536700 }, { "epoch": 92.35203028217481, "grad_norm": 15.862485885620117, "learning_rate": 3.823984858912595e-06, "loss": 0.5963, "step": 536750 }, { "epoch": 92.36063317274605, "grad_norm": 12.48328685760498, "learning_rate": 3.819683413626979e-06, "loss": 0.6668, "step": 536800 }, { "epoch": 92.36923606331727, "grad_norm": 7.039649963378906, "learning_rate": 3.815381968341363e-06, "loss": 0.6639, "step": 536850 }, { "epoch": 92.3778389538885, "grad_norm": 12.352269172668457, "learning_rate": 3.811080523055747e-06, "loss": 0.6201, "step": 536900 }, { "epoch": 92.38644184445974, "grad_norm": 12.384590148925781, "learning_rate": 3.806779077770131e-06, "loss": 0.6205, "step": 536950 }, { "epoch": 92.39504473503096, "grad_norm": 18.368377685546875, "learning_rate": 3.802477632484515e-06, "loss": 0.6707, "step": 537000 }, { "epoch": 92.4036476256022, "grad_norm": 19.537580490112305, "learning_rate": 3.7981761871988995e-06, "loss": 0.6571, "step": 537050 }, { "epoch": 92.41225051617343, "grad_norm": 22.005285263061523, "learning_rate": 3.7938747419132827e-06, "loss": 0.6532, "step": 537100 }, { "epoch": 92.42085340674467, "grad_norm": 15.990429878234863, "learning_rate": 3.7895732966276673e-06, "loss": 0.6673, "step": 537150 }, { "epoch": 92.4294562973159, "grad_norm": 13.891692161560059, "learning_rate": 3.785271851342051e-06, "loss": 0.6612, "step": 537200 }, { "epoch": 92.43805918788713, "grad_norm": 8.617033004760742, "learning_rate": 3.780970406056435e-06, "loss": 0.5708, "step": 537250 }, { "epoch": 92.44666207845836, "grad_norm": 11.10152816772461, "learning_rate": 3.776668960770819e-06, "loss": 0.6605, "step": 537300 }, { "epoch": 92.4552649690296, "grad_norm": 11.940013885498047, "learning_rate": 3.7723675154852034e-06, "loss": 0.602, "step": 537350 }, { "epoch": 92.46386785960082, "grad_norm": 11.613067626953125, "learning_rate": 3.768066070199587e-06, "loss": 0.6495, "step": 537400 }, { "epoch": 92.47247075017206, "grad_norm": 17.467750549316406, "learning_rate": 3.7637646249139713e-06, "loss": 0.6562, "step": 537450 }, { "epoch": 92.48107364074329, "grad_norm": 18.311262130737305, "learning_rate": 3.759463179628355e-06, "loss": 0.6154, "step": 537500 }, { "epoch": 92.48967653131452, "grad_norm": 19.921838760375977, "learning_rate": 3.7551617343427396e-06, "loss": 0.6755, "step": 537550 }, { "epoch": 92.49827942188575, "grad_norm": 18.39672088623047, "learning_rate": 3.7508602890571233e-06, "loss": 0.6602, "step": 537600 }, { "epoch": 92.50688231245698, "grad_norm": 27.244461059570312, "learning_rate": 3.7465588437715074e-06, "loss": 0.6423, "step": 537650 }, { "epoch": 92.51548520302822, "grad_norm": 14.636077880859375, "learning_rate": 3.742257398485891e-06, "loss": 0.6465, "step": 537700 }, { "epoch": 92.52408809359945, "grad_norm": 14.916301727294922, "learning_rate": 3.7379559532002757e-06, "loss": 0.5889, "step": 537750 }, { "epoch": 92.53269098417069, "grad_norm": 14.444426536560059, "learning_rate": 3.7336545079146594e-06, "loss": 0.6192, "step": 537800 }, { "epoch": 92.54129387474191, "grad_norm": 21.459178924560547, "learning_rate": 3.7293530626290435e-06, "loss": 0.6193, "step": 537850 }, { "epoch": 92.54989676531315, "grad_norm": 21.594390869140625, "learning_rate": 3.7250516173434272e-06, "loss": 0.6398, "step": 537900 }, { "epoch": 92.55849965588438, "grad_norm": 12.673218727111816, "learning_rate": 3.720750172057812e-06, "loss": 0.6544, "step": 537950 }, { "epoch": 92.56710254645562, "grad_norm": 16.632095336914062, "learning_rate": 3.7164487267721955e-06, "loss": 0.635, "step": 538000 }, { "epoch": 92.57570543702684, "grad_norm": 10.917662620544434, "learning_rate": 3.7121472814865797e-06, "loss": 0.7208, "step": 538050 }, { "epoch": 92.58430832759807, "grad_norm": 18.142173767089844, "learning_rate": 3.7078458362009634e-06, "loss": 0.6735, "step": 538100 }, { "epoch": 92.5929112181693, "grad_norm": 8.63364028930664, "learning_rate": 3.703544390915348e-06, "loss": 0.6456, "step": 538150 }, { "epoch": 92.60151410874053, "grad_norm": 27.331745147705078, "learning_rate": 3.6992429456297316e-06, "loss": 0.7064, "step": 538200 }, { "epoch": 92.61011699931177, "grad_norm": 15.775068283081055, "learning_rate": 3.6949415003441158e-06, "loss": 0.6194, "step": 538250 }, { "epoch": 92.618719889883, "grad_norm": 17.11605453491211, "learning_rate": 3.6906400550584995e-06, "loss": 0.6518, "step": 538300 }, { "epoch": 92.62732278045424, "grad_norm": 17.53357696533203, "learning_rate": 3.686338609772884e-06, "loss": 0.6126, "step": 538350 }, { "epoch": 92.63592567102546, "grad_norm": 17.737796783447266, "learning_rate": 3.682037164487268e-06, "loss": 0.6249, "step": 538400 }, { "epoch": 92.6445285615967, "grad_norm": 21.224641799926758, "learning_rate": 3.677735719201652e-06, "loss": 0.6562, "step": 538450 }, { "epoch": 92.65313145216793, "grad_norm": 13.487146377563477, "learning_rate": 3.6734342739160365e-06, "loss": 0.6555, "step": 538500 }, { "epoch": 92.66173434273917, "grad_norm": 20.42936897277832, "learning_rate": 3.6691328286304198e-06, "loss": 0.6489, "step": 538550 }, { "epoch": 92.67033723331039, "grad_norm": 15.493565559387207, "learning_rate": 3.6648313833448043e-06, "loss": 0.6241, "step": 538600 }, { "epoch": 92.67894012388163, "grad_norm": 17.093957901000977, "learning_rate": 3.660529938059188e-06, "loss": 0.6981, "step": 538650 }, { "epoch": 92.68754301445286, "grad_norm": 17.796796798706055, "learning_rate": 3.6562284927735726e-06, "loss": 0.6426, "step": 538700 }, { "epoch": 92.69614590502408, "grad_norm": 26.906766891479492, "learning_rate": 3.651927047487956e-06, "loss": 0.6185, "step": 538750 }, { "epoch": 92.70474879559532, "grad_norm": 10.614823341369629, "learning_rate": 3.6476256022023404e-06, "loss": 0.6222, "step": 538800 }, { "epoch": 92.71335168616655, "grad_norm": 9.241232872009277, "learning_rate": 3.643324156916724e-06, "loss": 0.6551, "step": 538850 }, { "epoch": 92.72195457673779, "grad_norm": 15.781522750854492, "learning_rate": 3.6390227116311083e-06, "loss": 0.6428, "step": 538900 }, { "epoch": 92.73055746730901, "grad_norm": 15.423216819763184, "learning_rate": 3.634721266345492e-06, "loss": 0.6644, "step": 538950 }, { "epoch": 92.73916035788025, "grad_norm": 16.140052795410156, "learning_rate": 3.6304198210598766e-06, "loss": 0.6413, "step": 539000 }, { "epoch": 92.74776324845148, "grad_norm": 9.273163795471191, "learning_rate": 3.6261183757742603e-06, "loss": 0.6635, "step": 539050 }, { "epoch": 92.75636613902272, "grad_norm": 15.396299362182617, "learning_rate": 3.6218169304886444e-06, "loss": 0.6143, "step": 539100 }, { "epoch": 92.76496902959394, "grad_norm": 12.494102478027344, "learning_rate": 3.617515485203028e-06, "loss": 0.6282, "step": 539150 }, { "epoch": 92.77357192016518, "grad_norm": 8.45889949798584, "learning_rate": 3.6132140399174127e-06, "loss": 0.6825, "step": 539200 }, { "epoch": 92.78217481073641, "grad_norm": 9.105408668518066, "learning_rate": 3.6089125946317964e-06, "loss": 0.5845, "step": 539250 }, { "epoch": 92.79077770130763, "grad_norm": 19.966703414916992, "learning_rate": 3.6046111493461805e-06, "loss": 0.6181, "step": 539300 }, { "epoch": 92.79938059187887, "grad_norm": 17.85556411743164, "learning_rate": 3.6003097040605642e-06, "loss": 0.6538, "step": 539350 }, { "epoch": 92.8079834824501, "grad_norm": 25.503507614135742, "learning_rate": 3.596008258774949e-06, "loss": 0.7052, "step": 539400 }, { "epoch": 92.81658637302134, "grad_norm": 20.55242347717285, "learning_rate": 3.5917068134893325e-06, "loss": 0.6377, "step": 539450 }, { "epoch": 92.82518926359256, "grad_norm": 16.191146850585938, "learning_rate": 3.5874053682037167e-06, "loss": 0.6193, "step": 539500 }, { "epoch": 92.8337921541638, "grad_norm": 18.711353302001953, "learning_rate": 3.5831039229181004e-06, "loss": 0.5609, "step": 539550 }, { "epoch": 92.84239504473503, "grad_norm": 10.583699226379395, "learning_rate": 3.578802477632485e-06, "loss": 0.5917, "step": 539600 }, { "epoch": 92.85099793530627, "grad_norm": 13.096092224121094, "learning_rate": 3.5745010323468686e-06, "loss": 0.62, "step": 539650 }, { "epoch": 92.8596008258775, "grad_norm": 14.110672950744629, "learning_rate": 3.5701995870612528e-06, "loss": 0.7186, "step": 539700 }, { "epoch": 92.86820371644873, "grad_norm": 19.457279205322266, "learning_rate": 3.5658981417756365e-06, "loss": 0.5718, "step": 539750 }, { "epoch": 92.87680660701996, "grad_norm": 8.563591957092285, "learning_rate": 3.561596696490021e-06, "loss": 0.5898, "step": 539800 }, { "epoch": 92.88540949759118, "grad_norm": 22.79060173034668, "learning_rate": 3.5572952512044048e-06, "loss": 0.6445, "step": 539850 }, { "epoch": 92.89401238816242, "grad_norm": 10.802861213684082, "learning_rate": 3.552993805918789e-06, "loss": 0.6316, "step": 539900 }, { "epoch": 92.90261527873365, "grad_norm": 13.664207458496094, "learning_rate": 3.5486923606331726e-06, "loss": 0.6239, "step": 539950 }, { "epoch": 92.91121816930489, "grad_norm": 24.426206588745117, "learning_rate": 3.544390915347557e-06, "loss": 0.6621, "step": 540000 }, { "epoch": 92.91982105987611, "grad_norm": 18.76677131652832, "learning_rate": 3.5400894700619405e-06, "loss": 0.641, "step": 540050 }, { "epoch": 92.92842395044735, "grad_norm": 19.123043060302734, "learning_rate": 3.535788024776325e-06, "loss": 0.6516, "step": 540100 }, { "epoch": 92.93702684101858, "grad_norm": 16.1373291015625, "learning_rate": 3.5314865794907087e-06, "loss": 0.6863, "step": 540150 }, { "epoch": 92.94562973158982, "grad_norm": 15.74080753326416, "learning_rate": 3.527185134205093e-06, "loss": 0.6644, "step": 540200 }, { "epoch": 92.95423262216104, "grad_norm": 12.215340614318848, "learning_rate": 3.5228836889194774e-06, "loss": 0.6486, "step": 540250 }, { "epoch": 92.96283551273228, "grad_norm": 14.349640846252441, "learning_rate": 3.518582243633861e-06, "loss": 0.6542, "step": 540300 }, { "epoch": 92.97143840330351, "grad_norm": 15.08072280883789, "learning_rate": 3.5142807983482453e-06, "loss": 0.6589, "step": 540350 }, { "epoch": 92.98004129387475, "grad_norm": 8.355720520019531, "learning_rate": 3.509979353062629e-06, "loss": 0.6648, "step": 540400 }, { "epoch": 92.98864418444597, "grad_norm": 19.40382957458496, "learning_rate": 3.5056779077770136e-06, "loss": 0.6852, "step": 540450 }, { "epoch": 92.9972470750172, "grad_norm": 14.068017959594727, "learning_rate": 3.5013764624913973e-06, "loss": 0.7151, "step": 540500 }, { "epoch": 93.0, "eval_accuracy": 0.5535191877473756, "eval_f1": 0.5359742213598062, "eval_f1_DuraRiadoRio_16x16": 0.5263926052055461, "eval_f1_Mole_16x16": 0.38008745375042047, "eval_f1_Quebrado_16x16": 0.6758597208035411, "eval_f1_RiadoRio_16x16": 0.5148995148995149, "eval_f1_RioFechado_16x16": 0.5826318121400088, "eval_loss": 2.466514825820923, "eval_precision": 0.6064725906858921, "eval_precision_DuraRiadoRio_16x16": 0.5987825124515772, "eval_precision_Mole_16x16": 0.8445440956651719, "eval_precision_Quebrado_16x16": 0.5560224089635855, "eval_precision_RiadoRio_16x16": 0.44491017964071855, "eval_precision_RioFechado_16x16": 0.5881037567084079, "eval_recall": 0.5529338761558564, "eval_recall_DuraRiadoRio_16x16": 0.4696180555555556, "eval_recall_Mole_16x16": 0.24522569444444445, "eval_recall_Quebrado_16x16": 0.8615451388888888, "eval_recall_RiadoRio_16x16": 0.6110197368421053, "eval_recall_RioFechado_16x16": 0.577260755048288, "eval_runtime": 45.5064, "eval_samples_per_second": 255.392, "eval_steps_per_second": 15.976, "step": 540516 }, { "epoch": 93.00584996558844, "grad_norm": 15.683414459228516, "learning_rate": 3.4970750172057814e-06, "loss": 0.7082, "step": 540550 }, { "epoch": 93.01445285615966, "grad_norm": 13.511077880859375, "learning_rate": 3.492773571920165e-06, "loss": 0.6702, "step": 540600 }, { "epoch": 93.0230557467309, "grad_norm": 20.36753273010254, "learning_rate": 3.4884721266345497e-06, "loss": 0.6144, "step": 540650 }, { "epoch": 93.03165863730213, "grad_norm": 15.776017189025879, "learning_rate": 3.4841706813489334e-06, "loss": 0.6551, "step": 540700 }, { "epoch": 93.04026152787337, "grad_norm": 11.212902069091797, "learning_rate": 3.4798692360633175e-06, "loss": 0.6884, "step": 540750 }, { "epoch": 93.0488644184446, "grad_norm": 19.716697692871094, "learning_rate": 3.4755677907777013e-06, "loss": 0.6545, "step": 540800 }, { "epoch": 93.05746730901583, "grad_norm": 12.892132759094238, "learning_rate": 3.471266345492086e-06, "loss": 0.6198, "step": 540850 }, { "epoch": 93.06607019958706, "grad_norm": 11.229418754577637, "learning_rate": 3.4669649002064695e-06, "loss": 0.6355, "step": 540900 }, { "epoch": 93.0746730901583, "grad_norm": 23.0851993560791, "learning_rate": 3.4626634549208537e-06, "loss": 0.6486, "step": 540950 }, { "epoch": 93.08327598072952, "grad_norm": 18.092453002929688, "learning_rate": 3.4583620096352374e-06, "loss": 0.6657, "step": 541000 }, { "epoch": 93.09187887130075, "grad_norm": 10.7114839553833, "learning_rate": 3.454060564349622e-06, "loss": 0.5959, "step": 541050 }, { "epoch": 93.10048176187199, "grad_norm": 9.313252449035645, "learning_rate": 3.4497591190640056e-06, "loss": 0.6643, "step": 541100 }, { "epoch": 93.10908465244322, "grad_norm": 15.114627838134766, "learning_rate": 3.4454576737783898e-06, "loss": 0.6271, "step": 541150 }, { "epoch": 93.11768754301445, "grad_norm": 12.258750915527344, "learning_rate": 3.4411562284927735e-06, "loss": 0.6261, "step": 541200 }, { "epoch": 93.12629043358568, "grad_norm": 22.06733512878418, "learning_rate": 3.436854783207158e-06, "loss": 0.6857, "step": 541250 }, { "epoch": 93.13489332415692, "grad_norm": 17.17879295349121, "learning_rate": 3.4325533379215418e-06, "loss": 0.6383, "step": 541300 }, { "epoch": 93.14349621472815, "grad_norm": 17.4126033782959, "learning_rate": 3.428251892635926e-06, "loss": 0.631, "step": 541350 }, { "epoch": 93.15209910529938, "grad_norm": 17.369644165039062, "learning_rate": 3.4239504473503096e-06, "loss": 0.6101, "step": 541400 }, { "epoch": 93.16070199587061, "grad_norm": 20.69442367553711, "learning_rate": 3.419649002064694e-06, "loss": 0.6586, "step": 541450 }, { "epoch": 93.16930488644185, "grad_norm": 16.92311668395996, "learning_rate": 3.415347556779078e-06, "loss": 0.6564, "step": 541500 }, { "epoch": 93.17790777701308, "grad_norm": 6.7577643394470215, "learning_rate": 3.411046111493462e-06, "loss": 0.623, "step": 541550 }, { "epoch": 93.18651066758432, "grad_norm": 10.069173812866211, "learning_rate": 3.4067446662078457e-06, "loss": 0.6077, "step": 541600 }, { "epoch": 93.19511355815554, "grad_norm": 10.859237670898438, "learning_rate": 3.4024432209222303e-06, "loss": 0.6541, "step": 541650 }, { "epoch": 93.20371644872677, "grad_norm": 10.830582618713379, "learning_rate": 3.3981417756366136e-06, "loss": 0.6747, "step": 541700 }, { "epoch": 93.212319339298, "grad_norm": 12.778170585632324, "learning_rate": 3.393840330350998e-06, "loss": 0.6436, "step": 541750 }, { "epoch": 93.22092222986923, "grad_norm": 25.691370010375977, "learning_rate": 3.389538885065382e-06, "loss": 0.6766, "step": 541800 }, { "epoch": 93.22952512044047, "grad_norm": 14.253006935119629, "learning_rate": 3.385237439779766e-06, "loss": 0.662, "step": 541850 }, { "epoch": 93.2381280110117, "grad_norm": 13.798110961914062, "learning_rate": 3.3809359944941497e-06, "loss": 0.6174, "step": 541900 }, { "epoch": 93.24673090158294, "grad_norm": 14.35859203338623, "learning_rate": 3.3766345492085343e-06, "loss": 0.6076, "step": 541950 }, { "epoch": 93.25533379215416, "grad_norm": 13.870186805725098, "learning_rate": 3.372333103922918e-06, "loss": 0.6111, "step": 542000 }, { "epoch": 93.2639366827254, "grad_norm": 14.342181205749512, "learning_rate": 3.368031658637302e-06, "loss": 0.6147, "step": 542050 }, { "epoch": 93.27253957329663, "grad_norm": 24.88372802734375, "learning_rate": 3.3637302133516867e-06, "loss": 0.6239, "step": 542100 }, { "epoch": 93.28114246386787, "grad_norm": 11.534753799438477, "learning_rate": 3.3594287680660704e-06, "loss": 0.6225, "step": 542150 }, { "epoch": 93.28974535443909, "grad_norm": 16.38965606689453, "learning_rate": 3.3551273227804545e-06, "loss": 0.637, "step": 542200 }, { "epoch": 93.29834824501032, "grad_norm": 15.169517517089844, "learning_rate": 3.3508258774948383e-06, "loss": 0.6131, "step": 542250 }, { "epoch": 93.30695113558156, "grad_norm": 9.139866828918457, "learning_rate": 3.346524432209223e-06, "loss": 0.6401, "step": 542300 }, { "epoch": 93.31555402615278, "grad_norm": 15.49238395690918, "learning_rate": 3.3422229869236065e-06, "loss": 0.601, "step": 542350 }, { "epoch": 93.32415691672402, "grad_norm": 23.124160766601562, "learning_rate": 3.3379215416379907e-06, "loss": 0.6119, "step": 542400 }, { "epoch": 93.33275980729525, "grad_norm": 11.748249053955078, "learning_rate": 3.3336200963523744e-06, "loss": 0.6901, "step": 542450 }, { "epoch": 93.34136269786649, "grad_norm": 27.170808792114258, "learning_rate": 3.329318651066759e-06, "loss": 0.6712, "step": 542500 }, { "epoch": 93.34996558843771, "grad_norm": 14.693628311157227, "learning_rate": 3.3250172057811427e-06, "loss": 0.6215, "step": 542550 }, { "epoch": 93.35856847900895, "grad_norm": 15.125642776489258, "learning_rate": 3.320715760495527e-06, "loss": 0.6355, "step": 542600 }, { "epoch": 93.36717136958018, "grad_norm": 19.362293243408203, "learning_rate": 3.3164143152099105e-06, "loss": 0.6737, "step": 542650 }, { "epoch": 93.37577426015142, "grad_norm": 9.907008171081543, "learning_rate": 3.312112869924295e-06, "loss": 0.5917, "step": 542700 }, { "epoch": 93.38437715072264, "grad_norm": 14.799338340759277, "learning_rate": 3.3078114246386788e-06, "loss": 0.5887, "step": 542750 }, { "epoch": 93.39298004129388, "grad_norm": 17.584318161010742, "learning_rate": 3.303509979353063e-06, "loss": 0.6448, "step": 542800 }, { "epoch": 93.40158293186511, "grad_norm": 19.220380783081055, "learning_rate": 3.2992085340674466e-06, "loss": 0.7379, "step": 542850 }, { "epoch": 93.41018582243633, "grad_norm": 14.23182487487793, "learning_rate": 3.294907088781831e-06, "loss": 0.6529, "step": 542900 }, { "epoch": 93.41878871300757, "grad_norm": 12.237110137939453, "learning_rate": 3.290605643496215e-06, "loss": 0.6454, "step": 542950 }, { "epoch": 93.4273916035788, "grad_norm": 15.09355354309082, "learning_rate": 3.286304198210599e-06, "loss": 0.6469, "step": 543000 }, { "epoch": 93.43599449415004, "grad_norm": 10.908658981323242, "learning_rate": 3.2820027529249828e-06, "loss": 0.6186, "step": 543050 }, { "epoch": 93.44459738472126, "grad_norm": 13.02155590057373, "learning_rate": 3.2777013076393673e-06, "loss": 0.6592, "step": 543100 }, { "epoch": 93.4532002752925, "grad_norm": 8.489781379699707, "learning_rate": 3.273399862353751e-06, "loss": 0.6656, "step": 543150 }, { "epoch": 93.46180316586373, "grad_norm": 17.19764518737793, "learning_rate": 3.269098417068135e-06, "loss": 0.5845, "step": 543200 }, { "epoch": 93.47040605643497, "grad_norm": 14.347383499145508, "learning_rate": 3.264796971782519e-06, "loss": 0.6464, "step": 543250 }, { "epoch": 93.4790089470062, "grad_norm": 16.985157012939453, "learning_rate": 3.2604955264969034e-06, "loss": 0.653, "step": 543300 }, { "epoch": 93.48761183757743, "grad_norm": 12.716592788696289, "learning_rate": 3.2561940812112867e-06, "loss": 0.6564, "step": 543350 }, { "epoch": 93.49621472814866, "grad_norm": 7.255436897277832, "learning_rate": 3.2518926359256713e-06, "loss": 0.626, "step": 543400 }, { "epoch": 93.50481761871988, "grad_norm": 13.888875007629395, "learning_rate": 3.247591190640055e-06, "loss": 0.6872, "step": 543450 }, { "epoch": 93.51342050929112, "grad_norm": 21.698896408081055, "learning_rate": 3.243289745354439e-06, "loss": 0.6463, "step": 543500 }, { "epoch": 93.52202339986235, "grad_norm": 11.888068199157715, "learning_rate": 3.238988300068823e-06, "loss": 0.6865, "step": 543550 }, { "epoch": 93.53062629043359, "grad_norm": 28.491558074951172, "learning_rate": 3.2346868547832074e-06, "loss": 0.6185, "step": 543600 }, { "epoch": 93.53922918100481, "grad_norm": 17.97162437438965, "learning_rate": 3.230385409497591e-06, "loss": 0.6435, "step": 543650 }, { "epoch": 93.54783207157605, "grad_norm": 21.673452377319336, "learning_rate": 3.2260839642119753e-06, "loss": 0.6165, "step": 543700 }, { "epoch": 93.55643496214728, "grad_norm": 12.669404029846191, "learning_rate": 3.221782518926359e-06, "loss": 0.6889, "step": 543750 }, { "epoch": 93.56503785271852, "grad_norm": 25.377016067504883, "learning_rate": 3.2174810736407435e-06, "loss": 0.6903, "step": 543800 }, { "epoch": 93.57364074328974, "grad_norm": 19.434263229370117, "learning_rate": 3.2131796283551272e-06, "loss": 0.6612, "step": 543850 }, { "epoch": 93.58224363386098, "grad_norm": 17.019868850708008, "learning_rate": 3.2088781830695114e-06, "loss": 0.6619, "step": 543900 }, { "epoch": 93.59084652443221, "grad_norm": 10.427120208740234, "learning_rate": 3.204576737783896e-06, "loss": 0.6837, "step": 543950 }, { "epoch": 93.59944941500343, "grad_norm": 16.622106552124023, "learning_rate": 3.2002752924982797e-06, "loss": 0.6817, "step": 544000 }, { "epoch": 93.60805230557467, "grad_norm": 22.598600387573242, "learning_rate": 3.195973847212664e-06, "loss": 0.6512, "step": 544050 }, { "epoch": 93.6166551961459, "grad_norm": 9.251970291137695, "learning_rate": 3.1916724019270475e-06, "loss": 0.6288, "step": 544100 }, { "epoch": 93.62525808671714, "grad_norm": 20.99315071105957, "learning_rate": 3.187370956641432e-06, "loss": 0.6461, "step": 544150 }, { "epoch": 93.63386097728836, "grad_norm": 8.450881958007812, "learning_rate": 3.1830695113558158e-06, "loss": 0.6235, "step": 544200 }, { "epoch": 93.6424638678596, "grad_norm": 22.308616638183594, "learning_rate": 3.1787680660702e-06, "loss": 0.6629, "step": 544250 }, { "epoch": 93.65106675843083, "grad_norm": 12.062960624694824, "learning_rate": 3.1744666207845836e-06, "loss": 0.6533, "step": 544300 }, { "epoch": 93.65966964900207, "grad_norm": 10.80909252166748, "learning_rate": 3.170165175498968e-06, "loss": 0.6087, "step": 544350 }, { "epoch": 93.6682725395733, "grad_norm": 12.680150032043457, "learning_rate": 3.165863730213352e-06, "loss": 0.5693, "step": 544400 }, { "epoch": 93.67687543014453, "grad_norm": 13.016845703125, "learning_rate": 3.161562284927736e-06, "loss": 0.6783, "step": 544450 }, { "epoch": 93.68547832071576, "grad_norm": 7.860245704650879, "learning_rate": 3.1572608396421198e-06, "loss": 0.6348, "step": 544500 }, { "epoch": 93.694081211287, "grad_norm": 15.817825317382812, "learning_rate": 3.1529593943565043e-06, "loss": 0.6919, "step": 544550 }, { "epoch": 93.70268410185822, "grad_norm": 15.705666542053223, "learning_rate": 3.148657949070888e-06, "loss": 0.6094, "step": 544600 }, { "epoch": 93.71128699242945, "grad_norm": 12.156822204589844, "learning_rate": 3.144356503785272e-06, "loss": 0.6341, "step": 544650 }, { "epoch": 93.71988988300069, "grad_norm": 12.075591087341309, "learning_rate": 3.140055058499656e-06, "loss": 0.6045, "step": 544700 }, { "epoch": 93.72849277357192, "grad_norm": 15.054741859436035, "learning_rate": 3.1357536132140404e-06, "loss": 0.704, "step": 544750 }, { "epoch": 93.73709566414315, "grad_norm": 21.048789978027344, "learning_rate": 3.1314521679284237e-06, "loss": 0.667, "step": 544800 }, { "epoch": 93.74569855471438, "grad_norm": 11.805648803710938, "learning_rate": 3.1271507226428083e-06, "loss": 0.645, "step": 544850 }, { "epoch": 93.75430144528562, "grad_norm": 7.389939308166504, "learning_rate": 3.1228492773571924e-06, "loss": 0.6636, "step": 544900 }, { "epoch": 93.76290433585685, "grad_norm": 11.455891609191895, "learning_rate": 3.1185478320715766e-06, "loss": 0.6438, "step": 544950 }, { "epoch": 93.77150722642808, "grad_norm": 10.937999725341797, "learning_rate": 3.1142463867859603e-06, "loss": 0.6542, "step": 545000 }, { "epoch": 93.78011011699931, "grad_norm": 12.51047134399414, "learning_rate": 3.1099449415003444e-06, "loss": 0.6114, "step": 545050 }, { "epoch": 93.78871300757055, "grad_norm": 13.279121398925781, "learning_rate": 3.1056434962147286e-06, "loss": 0.6637, "step": 545100 }, { "epoch": 93.79731589814178, "grad_norm": 23.008039474487305, "learning_rate": 3.1013420509291123e-06, "loss": 0.6335, "step": 545150 }, { "epoch": 93.805918788713, "grad_norm": 19.43250274658203, "learning_rate": 3.0970406056434964e-06, "loss": 0.6421, "step": 545200 }, { "epoch": 93.81452167928424, "grad_norm": 16.91084098815918, "learning_rate": 3.0927391603578805e-06, "loss": 0.6215, "step": 545250 }, { "epoch": 93.82312456985547, "grad_norm": 13.830048561096191, "learning_rate": 3.0884377150722647e-06, "loss": 0.6791, "step": 545300 }, { "epoch": 93.8317274604267, "grad_norm": 14.841375350952148, "learning_rate": 3.0841362697866484e-06, "loss": 0.6115, "step": 545350 }, { "epoch": 93.84033035099793, "grad_norm": 19.20914077758789, "learning_rate": 3.0798348245010325e-06, "loss": 0.5725, "step": 545400 }, { "epoch": 93.84893324156917, "grad_norm": 32.843666076660156, "learning_rate": 3.0755333792154167e-06, "loss": 0.6358, "step": 545450 }, { "epoch": 93.8575361321404, "grad_norm": 13.338712692260742, "learning_rate": 3.071231933929801e-06, "loss": 0.6378, "step": 545500 }, { "epoch": 93.86613902271164, "grad_norm": 15.66506290435791, "learning_rate": 3.0669304886441845e-06, "loss": 0.6145, "step": 545550 }, { "epoch": 93.87474191328286, "grad_norm": 19.694419860839844, "learning_rate": 3.0626290433585687e-06, "loss": 0.6854, "step": 545600 }, { "epoch": 93.8833448038541, "grad_norm": 10.83318042755127, "learning_rate": 3.058327598072953e-06, "loss": 0.6832, "step": 545650 }, { "epoch": 93.89194769442533, "grad_norm": 22.489524841308594, "learning_rate": 3.0540261527873365e-06, "loss": 0.629, "step": 545700 }, { "epoch": 93.90055058499657, "grad_norm": 15.026604652404785, "learning_rate": 3.0497247075017206e-06, "loss": 0.6094, "step": 545750 }, { "epoch": 93.90915347556779, "grad_norm": 17.437541961669922, "learning_rate": 3.0454232622161048e-06, "loss": 0.6431, "step": 545800 }, { "epoch": 93.91775636613902, "grad_norm": 13.585321426391602, "learning_rate": 3.041121816930489e-06, "loss": 0.582, "step": 545850 }, { "epoch": 93.92635925671026, "grad_norm": 8.876351356506348, "learning_rate": 3.0368203716448726e-06, "loss": 0.6129, "step": 545900 }, { "epoch": 93.93496214728148, "grad_norm": 10.829166412353516, "learning_rate": 3.0325189263592568e-06, "loss": 0.6582, "step": 545950 }, { "epoch": 93.94356503785272, "grad_norm": 18.44012451171875, "learning_rate": 3.028217481073641e-06, "loss": 0.6126, "step": 546000 }, { "epoch": 93.95216792842395, "grad_norm": 16.50237274169922, "learning_rate": 3.023916035788025e-06, "loss": 0.623, "step": 546050 }, { "epoch": 93.96077081899519, "grad_norm": 28.541906356811523, "learning_rate": 3.0196145905024088e-06, "loss": 0.6549, "step": 546100 }, { "epoch": 93.96937370956641, "grad_norm": 15.886334419250488, "learning_rate": 3.015313145216793e-06, "loss": 0.7131, "step": 546150 }, { "epoch": 93.97797660013765, "grad_norm": 21.134916305541992, "learning_rate": 3.011011699931177e-06, "loss": 0.6707, "step": 546200 }, { "epoch": 93.98657949070888, "grad_norm": 14.420744895935059, "learning_rate": 3.006710254645561e-06, "loss": 0.6534, "step": 546250 }, { "epoch": 93.99518238128012, "grad_norm": 22.194894790649414, "learning_rate": 3.002408809359945e-06, "loss": 0.6112, "step": 546300 }, { "epoch": 94.0, "eval_accuracy": 0.5655653071760455, "eval_f1": 0.5434197561672415, "eval_f1_DuraRiadoRio_16x16": 0.3346430910281598, "eval_f1_Mole_16x16": 0.5918271488961954, "eval_f1_Quebrado_16x16": 0.6963685272360457, "eval_f1_RiadoRio_16x16": 0.5243804956035172, "eval_f1_RioFechado_16x16": 0.5698795180722892, "eval_loss": 2.541344404220581, "eval_precision": 0.5859542721666884, "eval_precision_DuraRiadoRio_16x16": 0.6813333333333333, "eval_precision_Mole_16x16": 0.6448311156601843, "eval_precision_Quebrado_16x16": 0.568331503841932, "eval_precision_RiadoRio_16x16": 0.5101088646967341, "eval_precision_RioFechado_16x16": 0.5251665433012583, "eval_recall": 0.5659846488019141, "eval_recall_DuraRiadoRio_16x16": 0.22178819444444445, "eval_recall_Mole_16x16": 0.546875, "eval_recall_Quebrado_16x16": 0.8988715277777778, "eval_recall_RiadoRio_16x16": 0.5394736842105263, "eval_recall_RioFechado_16x16": 0.6229148375768218, "eval_runtime": 46.29, "eval_samples_per_second": 251.069, "eval_steps_per_second": 15.705, "step": 546328 }, { "epoch": 94.00378527185134, "grad_norm": 15.661688804626465, "learning_rate": 2.998107364074329e-06, "loss": 0.6285, "step": 546350 }, { "epoch": 94.01238816242257, "grad_norm": 21.248933792114258, "learning_rate": 2.993805918788713e-06, "loss": 0.654, "step": 546400 }, { "epoch": 94.0209910529938, "grad_norm": 12.604482650756836, "learning_rate": 2.989504473503097e-06, "loss": 0.6206, "step": 546450 }, { "epoch": 94.02959394356503, "grad_norm": 28.430742263793945, "learning_rate": 2.985203028217481e-06, "loss": 0.6209, "step": 546500 }, { "epoch": 94.03819683413627, "grad_norm": 12.578839302062988, "learning_rate": 2.980901582931865e-06, "loss": 0.6341, "step": 546550 }, { "epoch": 94.0467997247075, "grad_norm": 23.583038330078125, "learning_rate": 2.9766001376462493e-06, "loss": 0.6739, "step": 546600 }, { "epoch": 94.05540261527874, "grad_norm": 11.2398042678833, "learning_rate": 2.972298692360633e-06, "loss": 0.6231, "step": 546650 }, { "epoch": 94.06400550584996, "grad_norm": 10.613985061645508, "learning_rate": 2.967997247075017e-06, "loss": 0.6698, "step": 546700 }, { "epoch": 94.0726083964212, "grad_norm": 18.90979766845703, "learning_rate": 2.9636958017894017e-06, "loss": 0.5613, "step": 546750 }, { "epoch": 94.08121128699243, "grad_norm": 8.980939865112305, "learning_rate": 2.9593943565037854e-06, "loss": 0.6332, "step": 546800 }, { "epoch": 94.08981417756367, "grad_norm": 23.671546936035156, "learning_rate": 2.9550929112181695e-06, "loss": 0.6916, "step": 546850 }, { "epoch": 94.09841706813489, "grad_norm": 20.040700912475586, "learning_rate": 2.9507914659325537e-06, "loss": 0.6502, "step": 546900 }, { "epoch": 94.10701995870612, "grad_norm": 14.086402893066406, "learning_rate": 2.946490020646938e-06, "loss": 0.6464, "step": 546950 }, { "epoch": 94.11562284927736, "grad_norm": 21.834592819213867, "learning_rate": 2.9421885753613215e-06, "loss": 0.6174, "step": 547000 }, { "epoch": 94.12422573984858, "grad_norm": 13.30579948425293, "learning_rate": 2.9378871300757057e-06, "loss": 0.6043, "step": 547050 }, { "epoch": 94.13282863041982, "grad_norm": 12.617623329162598, "learning_rate": 2.93358568479009e-06, "loss": 0.6198, "step": 547100 }, { "epoch": 94.14143152099105, "grad_norm": 24.41472816467285, "learning_rate": 2.929284239504474e-06, "loss": 0.6597, "step": 547150 }, { "epoch": 94.15003441156229, "grad_norm": 14.088244438171387, "learning_rate": 2.9249827942188576e-06, "loss": 0.6253, "step": 547200 }, { "epoch": 94.15863730213351, "grad_norm": 18.477060317993164, "learning_rate": 2.9206813489332418e-06, "loss": 0.6236, "step": 547250 }, { "epoch": 94.16724019270475, "grad_norm": 24.929357528686523, "learning_rate": 2.916379903647626e-06, "loss": 0.6458, "step": 547300 }, { "epoch": 94.17584308327598, "grad_norm": 15.701444625854492, "learning_rate": 2.9120784583620096e-06, "loss": 0.6321, "step": 547350 }, { "epoch": 94.18444597384722, "grad_norm": 25.911426544189453, "learning_rate": 2.9077770130763938e-06, "loss": 0.6503, "step": 547400 }, { "epoch": 94.19304886441844, "grad_norm": 13.107559204101562, "learning_rate": 2.903475567790778e-06, "loss": 0.6739, "step": 547450 }, { "epoch": 94.20165175498968, "grad_norm": 26.700468063354492, "learning_rate": 2.899174122505162e-06, "loss": 0.6643, "step": 547500 }, { "epoch": 94.21025464556091, "grad_norm": 10.31753158569336, "learning_rate": 2.8948726772195458e-06, "loss": 0.6072, "step": 547550 }, { "epoch": 94.21885753613213, "grad_norm": 16.2934627532959, "learning_rate": 2.89057123193393e-06, "loss": 0.6524, "step": 547600 }, { "epoch": 94.22746042670337, "grad_norm": 15.567931175231934, "learning_rate": 2.886269786648314e-06, "loss": 0.7115, "step": 547650 }, { "epoch": 94.2360633172746, "grad_norm": 23.331375122070312, "learning_rate": 2.881968341362698e-06, "loss": 0.6964, "step": 547700 }, { "epoch": 94.24466620784584, "grad_norm": 17.591299057006836, "learning_rate": 2.877666896077082e-06, "loss": 0.6194, "step": 547750 }, { "epoch": 94.25326909841706, "grad_norm": 16.782196044921875, "learning_rate": 2.873365450791466e-06, "loss": 0.6955, "step": 547800 }, { "epoch": 94.2618719889883, "grad_norm": 15.606372833251953, "learning_rate": 2.86906400550585e-06, "loss": 0.626, "step": 547850 }, { "epoch": 94.27047487955953, "grad_norm": 8.343709945678711, "learning_rate": 2.8647625602202343e-06, "loss": 0.6557, "step": 547900 }, { "epoch": 94.27907777013077, "grad_norm": 16.6112117767334, "learning_rate": 2.860461114934618e-06, "loss": 0.6561, "step": 547950 }, { "epoch": 94.287680660702, "grad_norm": 19.701770782470703, "learning_rate": 2.856159669649002e-06, "loss": 0.6131, "step": 548000 }, { "epoch": 94.29628355127323, "grad_norm": 15.580122947692871, "learning_rate": 2.8518582243633863e-06, "loss": 0.6144, "step": 548050 }, { "epoch": 94.30488644184446, "grad_norm": 17.278600692749023, "learning_rate": 2.84755677907777e-06, "loss": 0.6066, "step": 548100 }, { "epoch": 94.31348933241568, "grad_norm": 20.59518051147461, "learning_rate": 2.843255333792154e-06, "loss": 0.6676, "step": 548150 }, { "epoch": 94.32209222298692, "grad_norm": 19.27459716796875, "learning_rate": 2.8389538885065383e-06, "loss": 0.678, "step": 548200 }, { "epoch": 94.33069511355815, "grad_norm": 14.12773609161377, "learning_rate": 2.8346524432209224e-06, "loss": 0.6495, "step": 548250 }, { "epoch": 94.33929800412939, "grad_norm": 11.16890811920166, "learning_rate": 2.830350997935306e-06, "loss": 0.6205, "step": 548300 }, { "epoch": 94.34790089470062, "grad_norm": 20.857553482055664, "learning_rate": 2.8260495526496903e-06, "loss": 0.6012, "step": 548350 }, { "epoch": 94.35650378527185, "grad_norm": 23.03809356689453, "learning_rate": 2.8217481073640744e-06, "loss": 0.6671, "step": 548400 }, { "epoch": 94.36510667584308, "grad_norm": 22.172149658203125, "learning_rate": 2.8174466620784585e-06, "loss": 0.6309, "step": 548450 }, { "epoch": 94.37370956641432, "grad_norm": 9.987625122070312, "learning_rate": 2.8131452167928422e-06, "loss": 0.6832, "step": 548500 }, { "epoch": 94.38231245698555, "grad_norm": 23.024433135986328, "learning_rate": 2.8088437715072264e-06, "loss": 0.6764, "step": 548550 }, { "epoch": 94.39091534755678, "grad_norm": 25.095943450927734, "learning_rate": 2.8045423262216105e-06, "loss": 0.6608, "step": 548600 }, { "epoch": 94.39951823812801, "grad_norm": 17.53130340576172, "learning_rate": 2.8002408809359946e-06, "loss": 0.6859, "step": 548650 }, { "epoch": 94.40812112869925, "grad_norm": 19.27217674255371, "learning_rate": 2.7959394356503788e-06, "loss": 0.6065, "step": 548700 }, { "epoch": 94.41672401927048, "grad_norm": 24.620086669921875, "learning_rate": 2.791637990364763e-06, "loss": 0.6418, "step": 548750 }, { "epoch": 94.4253269098417, "grad_norm": 19.65086555480957, "learning_rate": 2.787336545079147e-06, "loss": 0.6327, "step": 548800 }, { "epoch": 94.43392980041294, "grad_norm": 21.155380249023438, "learning_rate": 2.7830350997935308e-06, "loss": 0.6841, "step": 548850 }, { "epoch": 94.44253269098417, "grad_norm": 27.17923355102539, "learning_rate": 2.778733654507915e-06, "loss": 0.6755, "step": 548900 }, { "epoch": 94.4511355815554, "grad_norm": 15.33849811553955, "learning_rate": 2.774432209222299e-06, "loss": 0.6204, "step": 548950 }, { "epoch": 94.45973847212663, "grad_norm": 19.613750457763672, "learning_rate": 2.7701307639366828e-06, "loss": 0.6649, "step": 549000 }, { "epoch": 94.46834136269787, "grad_norm": 16.025352478027344, "learning_rate": 2.765829318651067e-06, "loss": 0.6348, "step": 549050 }, { "epoch": 94.4769442532691, "grad_norm": 20.743122100830078, "learning_rate": 2.761527873365451e-06, "loss": 0.6113, "step": 549100 }, { "epoch": 94.48554714384034, "grad_norm": 9.471858024597168, "learning_rate": 2.757226428079835e-06, "loss": 0.6417, "step": 549150 }, { "epoch": 94.49415003441156, "grad_norm": 12.296422958374023, "learning_rate": 2.752924982794219e-06, "loss": 0.6487, "step": 549200 }, { "epoch": 94.5027529249828, "grad_norm": 17.40047264099121, "learning_rate": 2.748623537508603e-06, "loss": 0.5952, "step": 549250 }, { "epoch": 94.51135581555403, "grad_norm": 18.86263084411621, "learning_rate": 2.744322092222987e-06, "loss": 0.6939, "step": 549300 }, { "epoch": 94.51995870612525, "grad_norm": 7.575594902038574, "learning_rate": 2.7400206469373713e-06, "loss": 0.6446, "step": 549350 }, { "epoch": 94.52856159669649, "grad_norm": 27.296934127807617, "learning_rate": 2.735719201651755e-06, "loss": 0.6371, "step": 549400 }, { "epoch": 94.53716448726772, "grad_norm": 20.021337509155273, "learning_rate": 2.731417756366139e-06, "loss": 0.6454, "step": 549450 }, { "epoch": 94.54576737783896, "grad_norm": 27.959077835083008, "learning_rate": 2.7271163110805233e-06, "loss": 0.6046, "step": 549500 }, { "epoch": 94.55437026841018, "grad_norm": 8.801491737365723, "learning_rate": 2.7228148657949074e-06, "loss": 0.6343, "step": 549550 }, { "epoch": 94.56297315898142, "grad_norm": 16.668018341064453, "learning_rate": 2.718513420509291e-06, "loss": 0.6484, "step": 549600 }, { "epoch": 94.57157604955265, "grad_norm": 28.877708435058594, "learning_rate": 2.7142119752236753e-06, "loss": 0.6406, "step": 549650 }, { "epoch": 94.58017894012389, "grad_norm": 26.632902145385742, "learning_rate": 2.7099105299380594e-06, "loss": 0.6344, "step": 549700 }, { "epoch": 94.58878183069511, "grad_norm": 9.668696403503418, "learning_rate": 2.705609084652443e-06, "loss": 0.6644, "step": 549750 }, { "epoch": 94.59738472126635, "grad_norm": 14.560320854187012, "learning_rate": 2.7013076393668273e-06, "loss": 0.6428, "step": 549800 }, { "epoch": 94.60598761183758, "grad_norm": 12.682247161865234, "learning_rate": 2.6970061940812114e-06, "loss": 0.6874, "step": 549850 }, { "epoch": 94.61459050240882, "grad_norm": 15.363481521606445, "learning_rate": 2.6927047487955955e-06, "loss": 0.6226, "step": 549900 }, { "epoch": 94.62319339298004, "grad_norm": 13.234254837036133, "learning_rate": 2.6884033035099792e-06, "loss": 0.6236, "step": 549950 }, { "epoch": 94.63179628355127, "grad_norm": 9.551309585571289, "learning_rate": 2.6841018582243634e-06, "loss": 0.6092, "step": 550000 }, { "epoch": 94.6403991741225, "grad_norm": 21.962312698364258, "learning_rate": 2.6798004129387475e-06, "loss": 0.6287, "step": 550050 }, { "epoch": 94.64900206469373, "grad_norm": 7.8691534996032715, "learning_rate": 2.6754989676531317e-06, "loss": 0.6289, "step": 550100 }, { "epoch": 94.65760495526497, "grad_norm": 14.665496826171875, "learning_rate": 2.6711975223675154e-06, "loss": 0.6305, "step": 550150 }, { "epoch": 94.6662078458362, "grad_norm": 16.959543228149414, "learning_rate": 2.6668960770818995e-06, "loss": 0.6151, "step": 550200 }, { "epoch": 94.67481073640744, "grad_norm": 15.515735626220703, "learning_rate": 2.6625946317962836e-06, "loss": 0.6455, "step": 550250 }, { "epoch": 94.68341362697866, "grad_norm": 19.496803283691406, "learning_rate": 2.6582931865106678e-06, "loss": 0.6479, "step": 550300 }, { "epoch": 94.6920165175499, "grad_norm": 12.756437301635742, "learning_rate": 2.6539917412250515e-06, "loss": 0.6236, "step": 550350 }, { "epoch": 94.70061940812113, "grad_norm": 13.199463844299316, "learning_rate": 2.6496902959394356e-06, "loss": 0.6263, "step": 550400 }, { "epoch": 94.70922229869237, "grad_norm": 21.088951110839844, "learning_rate": 2.6453888506538198e-06, "loss": 0.6056, "step": 550450 }, { "epoch": 94.71782518926359, "grad_norm": 20.674686431884766, "learning_rate": 2.641087405368204e-06, "loss": 0.6516, "step": 550500 }, { "epoch": 94.72642807983482, "grad_norm": 15.973745346069336, "learning_rate": 2.636785960082588e-06, "loss": 0.6419, "step": 550550 }, { "epoch": 94.73503097040606, "grad_norm": 9.522205352783203, "learning_rate": 2.632484514796972e-06, "loss": 0.6419, "step": 550600 }, { "epoch": 94.74363386097728, "grad_norm": 12.826592445373535, "learning_rate": 2.628183069511356e-06, "loss": 0.6316, "step": 550650 }, { "epoch": 94.75223675154852, "grad_norm": 14.75760269165039, "learning_rate": 2.62388162422574e-06, "loss": 0.6295, "step": 550700 }, { "epoch": 94.76083964211975, "grad_norm": 17.927520751953125, "learning_rate": 2.619580178940124e-06, "loss": 0.6435, "step": 550750 }, { "epoch": 94.76944253269099, "grad_norm": 25.289440155029297, "learning_rate": 2.6152787336545083e-06, "loss": 0.6786, "step": 550800 }, { "epoch": 94.77804542326221, "grad_norm": 16.789846420288086, "learning_rate": 2.610977288368892e-06, "loss": 0.615, "step": 550850 }, { "epoch": 94.78664831383345, "grad_norm": 17.92622184753418, "learning_rate": 2.606675843083276e-06, "loss": 0.6575, "step": 550900 }, { "epoch": 94.79525120440468, "grad_norm": 14.223265647888184, "learning_rate": 2.6023743977976603e-06, "loss": 0.6329, "step": 550950 }, { "epoch": 94.80385409497592, "grad_norm": 9.07654094696045, "learning_rate": 2.5980729525120444e-06, "loss": 0.6119, "step": 551000 }, { "epoch": 94.81245698554714, "grad_norm": 16.28241729736328, "learning_rate": 2.593771507226428e-06, "loss": 0.6051, "step": 551050 }, { "epoch": 94.82105987611837, "grad_norm": 14.85804271697998, "learning_rate": 2.5894700619408123e-06, "loss": 0.6839, "step": 551100 }, { "epoch": 94.82966276668961, "grad_norm": 7.88606071472168, "learning_rate": 2.5851686166551964e-06, "loss": 0.6, "step": 551150 }, { "epoch": 94.83826565726083, "grad_norm": 9.88338565826416, "learning_rate": 2.5808671713695805e-06, "loss": 0.5787, "step": 551200 }, { "epoch": 94.84686854783207, "grad_norm": 18.193038940429688, "learning_rate": 2.5765657260839643e-06, "loss": 0.6533, "step": 551250 }, { "epoch": 94.8554714384033, "grad_norm": 22.489059448242188, "learning_rate": 2.5722642807983484e-06, "loss": 0.6627, "step": 551300 }, { "epoch": 94.86407432897454, "grad_norm": 18.973716735839844, "learning_rate": 2.5679628355127325e-06, "loss": 0.6728, "step": 551350 }, { "epoch": 94.87267721954576, "grad_norm": 18.310190200805664, "learning_rate": 2.5636613902271162e-06, "loss": 0.6764, "step": 551400 }, { "epoch": 94.881280110117, "grad_norm": 13.671536445617676, "learning_rate": 2.5593599449415004e-06, "loss": 0.683, "step": 551450 }, { "epoch": 94.88988300068823, "grad_norm": 16.589990615844727, "learning_rate": 2.5550584996558845e-06, "loss": 0.6388, "step": 551500 }, { "epoch": 94.89848589125947, "grad_norm": 14.735958099365234, "learning_rate": 2.5507570543702687e-06, "loss": 0.6796, "step": 551550 }, { "epoch": 94.9070887818307, "grad_norm": 14.682092666625977, "learning_rate": 2.5464556090846524e-06, "loss": 0.6633, "step": 551600 }, { "epoch": 94.91569167240193, "grad_norm": 12.613260269165039, "learning_rate": 2.5421541637990365e-06, "loss": 0.6209, "step": 551650 }, { "epoch": 94.92429456297316, "grad_norm": 13.86113452911377, "learning_rate": 2.5378527185134206e-06, "loss": 0.6362, "step": 551700 }, { "epoch": 94.93289745354438, "grad_norm": 16.919113159179688, "learning_rate": 2.5335512732278048e-06, "loss": 0.6329, "step": 551750 }, { "epoch": 94.94150034411562, "grad_norm": 27.2540225982666, "learning_rate": 2.5292498279421885e-06, "loss": 0.6655, "step": 551800 }, { "epoch": 94.95010323468685, "grad_norm": 17.688684463500977, "learning_rate": 2.5249483826565726e-06, "loss": 0.6311, "step": 551850 }, { "epoch": 94.95870612525809, "grad_norm": 14.214184761047363, "learning_rate": 2.5206469373709568e-06, "loss": 0.6164, "step": 551900 }, { "epoch": 94.96730901582931, "grad_norm": 23.394084930419922, "learning_rate": 2.5163454920853405e-06, "loss": 0.6942, "step": 551950 }, { "epoch": 94.97591190640055, "grad_norm": 9.648112297058105, "learning_rate": 2.5120440467997246e-06, "loss": 0.6251, "step": 552000 }, { "epoch": 94.98451479697178, "grad_norm": 9.040225982666016, "learning_rate": 2.5077426015141088e-06, "loss": 0.643, "step": 552050 }, { "epoch": 94.99311768754302, "grad_norm": 14.617473602294922, "learning_rate": 2.503441156228493e-06, "loss": 0.6329, "step": 552100 }, { "epoch": 95.0, "eval_accuracy": 0.574083634486319, "eval_f1": 0.5658924970677942, "eval_f1_DuraRiadoRio_16x16": 0.5447914292227034, "eval_f1_Mole_16x16": 0.550013777900248, "eval_f1_Quebrado_16x16": 0.6711299804049641, "eval_f1_RiadoRio_16x16": 0.49584922593672875, "eval_f1_RioFechado_16x16": 0.5676780718743275, "eval_loss": 2.482473611831665, "eval_precision": 0.5934622951876445, "eval_precision_DuraRiadoRio_16x16": 0.5736917906865099, "eval_precision_Mole_16x16": 0.7532075471698113, "eval_precision_Quebrado_16x16": 0.5379581151832461, "eval_precision_RiadoRio_16x16": 0.5456790123456791, "eval_precision_RioFechado_16x16": 0.556775010552976, "eval_recall": 0.5754250467861929, "eval_recall_DuraRiadoRio_16x16": 0.5186631944444444, "eval_recall_Mole_16x16": 0.4331597222222222, "eval_recall_Quebrado_16x16": 0.8919270833333334, "eval_recall_RiadoRio_16x16": 0.4543585526315789, "eval_recall_RioFechado_16x16": 0.5790166812993854, "eval_runtime": 46.1816, "eval_samples_per_second": 251.659, "eval_steps_per_second": 15.742, "step": 552140 }, { "epoch": 95.00172057811425, "grad_norm": 12.270920753479004, "learning_rate": 2.4991397109428766e-06, "loss": 0.6307, "step": 552150 }, { "epoch": 95.01032346868548, "grad_norm": 20.795867919921875, "learning_rate": 2.4948382656572607e-06, "loss": 0.6654, "step": 552200 }, { "epoch": 95.01892635925671, "grad_norm": 20.302532196044922, "learning_rate": 2.490536820371645e-06, "loss": 0.6956, "step": 552250 }, { "epoch": 95.02752924982794, "grad_norm": 16.396270751953125, "learning_rate": 2.486235375086029e-06, "loss": 0.572, "step": 552300 }, { "epoch": 95.03613214039918, "grad_norm": 27.255064010620117, "learning_rate": 2.481933929800413e-06, "loss": 0.6356, "step": 552350 }, { "epoch": 95.0447350309704, "grad_norm": 20.74248504638672, "learning_rate": 2.4776324845147973e-06, "loss": 0.6203, "step": 552400 }, { "epoch": 95.05333792154164, "grad_norm": 10.792712211608887, "learning_rate": 2.4733310392291814e-06, "loss": 0.6641, "step": 552450 }, { "epoch": 95.06194081211287, "grad_norm": 11.253053665161133, "learning_rate": 2.469029593943565e-06, "loss": 0.6423, "step": 552500 }, { "epoch": 95.0705437026841, "grad_norm": 20.994140625, "learning_rate": 2.4647281486579493e-06, "loss": 0.6531, "step": 552550 }, { "epoch": 95.07914659325533, "grad_norm": 23.351390838623047, "learning_rate": 2.4604267033723334e-06, "loss": 0.661, "step": 552600 }, { "epoch": 95.08774948382657, "grad_norm": 10.845829010009766, "learning_rate": 2.4561252580867176e-06, "loss": 0.5831, "step": 552650 }, { "epoch": 95.0963523743978, "grad_norm": 17.953969955444336, "learning_rate": 2.4518238128011013e-06, "loss": 0.6435, "step": 552700 }, { "epoch": 95.10495526496904, "grad_norm": 18.899084091186523, "learning_rate": 2.4475223675154854e-06, "loss": 0.6403, "step": 552750 }, { "epoch": 95.11355815554026, "grad_norm": 12.728574752807617, "learning_rate": 2.4432209222298695e-06, "loss": 0.6718, "step": 552800 }, { "epoch": 95.12216104611149, "grad_norm": 18.12833595275879, "learning_rate": 2.4389194769442533e-06, "loss": 0.6591, "step": 552850 }, { "epoch": 95.13076393668273, "grad_norm": 15.17020320892334, "learning_rate": 2.4346180316586374e-06, "loss": 0.6625, "step": 552900 }, { "epoch": 95.13936682725395, "grad_norm": 13.595468521118164, "learning_rate": 2.4303165863730215e-06, "loss": 0.5873, "step": 552950 }, { "epoch": 95.14796971782519, "grad_norm": 10.928699493408203, "learning_rate": 2.4260151410874057e-06, "loss": 0.6024, "step": 553000 }, { "epoch": 95.15657260839642, "grad_norm": 15.679282188415527, "learning_rate": 2.4217136958017894e-06, "loss": 0.6442, "step": 553050 }, { "epoch": 95.16517549896766, "grad_norm": 8.980783462524414, "learning_rate": 2.4174122505161735e-06, "loss": 0.6253, "step": 553100 }, { "epoch": 95.17377838953888, "grad_norm": 19.590856552124023, "learning_rate": 2.4131108052305577e-06, "loss": 0.6287, "step": 553150 }, { "epoch": 95.18238128011012, "grad_norm": 15.999622344970703, "learning_rate": 2.408809359944942e-06, "loss": 0.6288, "step": 553200 }, { "epoch": 95.19098417068135, "grad_norm": 19.56664276123047, "learning_rate": 2.4045079146593255e-06, "loss": 0.6615, "step": 553250 }, { "epoch": 95.19958706125259, "grad_norm": 18.633882522583008, "learning_rate": 2.4002064693737096e-06, "loss": 0.6414, "step": 553300 }, { "epoch": 95.20818995182381, "grad_norm": 18.80942153930664, "learning_rate": 2.3959050240880938e-06, "loss": 0.6595, "step": 553350 }, { "epoch": 95.21679284239505, "grad_norm": 13.080646514892578, "learning_rate": 2.391603578802478e-06, "loss": 0.6532, "step": 553400 }, { "epoch": 95.22539573296628, "grad_norm": 9.403877258300781, "learning_rate": 2.3873021335168616e-06, "loss": 0.6351, "step": 553450 }, { "epoch": 95.2339986235375, "grad_norm": 18.678909301757812, "learning_rate": 2.3830006882312458e-06, "loss": 0.641, "step": 553500 }, { "epoch": 95.24260151410874, "grad_norm": 18.780717849731445, "learning_rate": 2.37869924294563e-06, "loss": 0.6556, "step": 553550 }, { "epoch": 95.25120440467997, "grad_norm": 23.374631881713867, "learning_rate": 2.3743977976600136e-06, "loss": 0.6636, "step": 553600 }, { "epoch": 95.2598072952512, "grad_norm": 21.58243179321289, "learning_rate": 2.3700963523743978e-06, "loss": 0.6995, "step": 553650 }, { "epoch": 95.26841018582243, "grad_norm": 7.491250038146973, "learning_rate": 2.365794907088782e-06, "loss": 0.6476, "step": 553700 }, { "epoch": 95.27701307639367, "grad_norm": 18.815654754638672, "learning_rate": 2.361493461803166e-06, "loss": 0.6261, "step": 553750 }, { "epoch": 95.2856159669649, "grad_norm": 11.39062786102295, "learning_rate": 2.3571920165175497e-06, "loss": 0.6748, "step": 553800 }, { "epoch": 95.29421885753614, "grad_norm": 11.908299446105957, "learning_rate": 2.352890571231934e-06, "loss": 0.6353, "step": 553850 }, { "epoch": 95.30282174810736, "grad_norm": 13.691070556640625, "learning_rate": 2.348589125946318e-06, "loss": 0.617, "step": 553900 }, { "epoch": 95.3114246386786, "grad_norm": 27.943357467651367, "learning_rate": 2.344287680660702e-06, "loss": 0.5737, "step": 553950 }, { "epoch": 95.32002752924983, "grad_norm": 18.57220458984375, "learning_rate": 2.339986235375086e-06, "loss": 0.5695, "step": 554000 }, { "epoch": 95.32863041982105, "grad_norm": 9.468082427978516, "learning_rate": 2.33568479008947e-06, "loss": 0.6273, "step": 554050 }, { "epoch": 95.33723331039229, "grad_norm": 18.537763595581055, "learning_rate": 2.331383344803854e-06, "loss": 0.6362, "step": 554100 }, { "epoch": 95.34583620096352, "grad_norm": 13.962305068969727, "learning_rate": 2.3270818995182383e-06, "loss": 0.6739, "step": 554150 }, { "epoch": 95.35443909153476, "grad_norm": 16.10093116760254, "learning_rate": 2.3227804542326224e-06, "loss": 0.6534, "step": 554200 }, { "epoch": 95.36304198210598, "grad_norm": 16.123149871826172, "learning_rate": 2.3184790089470065e-06, "loss": 0.578, "step": 554250 }, { "epoch": 95.37164487267722, "grad_norm": 10.9259614944458, "learning_rate": 2.3141775636613907e-06, "loss": 0.626, "step": 554300 }, { "epoch": 95.38024776324845, "grad_norm": 26.358707427978516, "learning_rate": 2.3098761183757744e-06, "loss": 0.6169, "step": 554350 }, { "epoch": 95.38885065381969, "grad_norm": 12.369949340820312, "learning_rate": 2.3055746730901585e-06, "loss": 0.6357, "step": 554400 }, { "epoch": 95.39745354439091, "grad_norm": 11.57968807220459, "learning_rate": 2.3012732278045427e-06, "loss": 0.6234, "step": 554450 }, { "epoch": 95.40605643496215, "grad_norm": 9.976631164550781, "learning_rate": 2.2969717825189264e-06, "loss": 0.6327, "step": 554500 }, { "epoch": 95.41465932553338, "grad_norm": 17.85121726989746, "learning_rate": 2.2926703372333105e-06, "loss": 0.6496, "step": 554550 }, { "epoch": 95.42326221610462, "grad_norm": 14.75669002532959, "learning_rate": 2.2883688919476947e-06, "loss": 0.6455, "step": 554600 }, { "epoch": 95.43186510667584, "grad_norm": 11.036243438720703, "learning_rate": 2.284067446662079e-06, "loss": 0.6046, "step": 554650 }, { "epoch": 95.44046799724707, "grad_norm": 13.325764656066895, "learning_rate": 2.2797660013764625e-06, "loss": 0.642, "step": 554700 }, { "epoch": 95.44907088781831, "grad_norm": 15.45727825164795, "learning_rate": 2.2754645560908466e-06, "loss": 0.637, "step": 554750 }, { "epoch": 95.45767377838953, "grad_norm": 12.41168212890625, "learning_rate": 2.2711631108052308e-06, "loss": 0.591, "step": 554800 }, { "epoch": 95.46627666896077, "grad_norm": 18.90811538696289, "learning_rate": 2.266861665519615e-06, "loss": 0.6817, "step": 554850 }, { "epoch": 95.474879559532, "grad_norm": 12.141444206237793, "learning_rate": 2.2625602202339986e-06, "loss": 0.657, "step": 554900 }, { "epoch": 95.48348245010324, "grad_norm": 14.941337585449219, "learning_rate": 2.2582587749483828e-06, "loss": 0.6197, "step": 554950 }, { "epoch": 95.49208534067446, "grad_norm": 9.355320930480957, "learning_rate": 2.253957329662767e-06, "loss": 0.6155, "step": 555000 }, { "epoch": 95.5006882312457, "grad_norm": 30.097551345825195, "learning_rate": 2.249655884377151e-06, "loss": 0.6522, "step": 555050 }, { "epoch": 95.50929112181693, "grad_norm": 9.56309986114502, "learning_rate": 2.2453544390915348e-06, "loss": 0.6014, "step": 555100 }, { "epoch": 95.51789401238817, "grad_norm": 14.629178047180176, "learning_rate": 2.241052993805919e-06, "loss": 0.6881, "step": 555150 }, { "epoch": 95.5264969029594, "grad_norm": 22.990379333496094, "learning_rate": 2.236751548520303e-06, "loss": 0.6425, "step": 555200 }, { "epoch": 95.53509979353062, "grad_norm": 14.603804588317871, "learning_rate": 2.2324501032346867e-06, "loss": 0.6405, "step": 555250 }, { "epoch": 95.54370268410186, "grad_norm": 11.970803260803223, "learning_rate": 2.228148657949071e-06, "loss": 0.6075, "step": 555300 }, { "epoch": 95.55230557467308, "grad_norm": 20.5654296875, "learning_rate": 2.223847212663455e-06, "loss": 0.5979, "step": 555350 }, { "epoch": 95.56090846524432, "grad_norm": 12.543355941772461, "learning_rate": 2.219545767377839e-06, "loss": 0.6007, "step": 555400 }, { "epoch": 95.56951135581555, "grad_norm": 27.193748474121094, "learning_rate": 2.215244322092223e-06, "loss": 0.5929, "step": 555450 }, { "epoch": 95.57811424638679, "grad_norm": 20.12854766845703, "learning_rate": 2.210942876806607e-06, "loss": 0.6499, "step": 555500 }, { "epoch": 95.58671713695801, "grad_norm": 22.474946975708008, "learning_rate": 2.206641431520991e-06, "loss": 0.6439, "step": 555550 }, { "epoch": 95.59532002752925, "grad_norm": 18.248716354370117, "learning_rate": 2.2023399862353753e-06, "loss": 0.6343, "step": 555600 }, { "epoch": 95.60392291810048, "grad_norm": 31.540340423583984, "learning_rate": 2.198038540949759e-06, "loss": 0.5668, "step": 555650 }, { "epoch": 95.61252580867172, "grad_norm": 14.38753890991211, "learning_rate": 2.193737095664143e-06, "loss": 0.6766, "step": 555700 }, { "epoch": 95.62112869924295, "grad_norm": 13.949178695678711, "learning_rate": 2.1894356503785273e-06, "loss": 0.689, "step": 555750 }, { "epoch": 95.62973158981418, "grad_norm": 14.8200101852417, "learning_rate": 2.1851342050929114e-06, "loss": 0.6558, "step": 555800 }, { "epoch": 95.63833448038541, "grad_norm": 20.2263240814209, "learning_rate": 2.180832759807295e-06, "loss": 0.6273, "step": 555850 }, { "epoch": 95.64693737095664, "grad_norm": 23.678543090820312, "learning_rate": 2.1765313145216793e-06, "loss": 0.6522, "step": 555900 }, { "epoch": 95.65554026152788, "grad_norm": 19.00031852722168, "learning_rate": 2.1722298692360634e-06, "loss": 0.6117, "step": 555950 }, { "epoch": 95.6641431520991, "grad_norm": 17.135848999023438, "learning_rate": 2.167928423950447e-06, "loss": 0.6166, "step": 556000 }, { "epoch": 95.67274604267034, "grad_norm": 16.569746017456055, "learning_rate": 2.1636269786648312e-06, "loss": 0.6852, "step": 556050 }, { "epoch": 95.68134893324157, "grad_norm": 22.6848201751709, "learning_rate": 2.159325533379216e-06, "loss": 0.5585, "step": 556100 }, { "epoch": 95.6899518238128, "grad_norm": 13.697364807128906, "learning_rate": 2.1550240880935995e-06, "loss": 0.632, "step": 556150 }, { "epoch": 95.69855471438403, "grad_norm": 16.276456832885742, "learning_rate": 2.1507226428079836e-06, "loss": 0.6779, "step": 556200 }, { "epoch": 95.70715760495527, "grad_norm": 14.517303466796875, "learning_rate": 2.1464211975223678e-06, "loss": 0.6615, "step": 556250 }, { "epoch": 95.7157604955265, "grad_norm": 13.043216705322266, "learning_rate": 2.142119752236752e-06, "loss": 0.6702, "step": 556300 }, { "epoch": 95.72436338609774, "grad_norm": 17.924299240112305, "learning_rate": 2.1378183069511356e-06, "loss": 0.7541, "step": 556350 }, { "epoch": 95.73296627666896, "grad_norm": 16.72734832763672, "learning_rate": 2.1335168616655198e-06, "loss": 0.6355, "step": 556400 }, { "epoch": 95.74156916724019, "grad_norm": 30.031986236572266, "learning_rate": 2.129215416379904e-06, "loss": 0.6327, "step": 556450 }, { "epoch": 95.75017205781143, "grad_norm": 15.08935832977295, "learning_rate": 2.124913971094288e-06, "loss": 0.6303, "step": 556500 }, { "epoch": 95.75877494838265, "grad_norm": 15.052163124084473, "learning_rate": 2.1206125258086718e-06, "loss": 0.6504, "step": 556550 }, { "epoch": 95.76737783895389, "grad_norm": 14.00178337097168, "learning_rate": 2.116311080523056e-06, "loss": 0.584, "step": 556600 }, { "epoch": 95.77598072952512, "grad_norm": 16.90911865234375, "learning_rate": 2.11200963523744e-06, "loss": 0.6494, "step": 556650 }, { "epoch": 95.78458362009636, "grad_norm": 14.752416610717773, "learning_rate": 2.107708189951824e-06, "loss": 0.6268, "step": 556700 }, { "epoch": 95.79318651066758, "grad_norm": 14.48753833770752, "learning_rate": 2.103406744666208e-06, "loss": 0.6631, "step": 556750 }, { "epoch": 95.80178940123882, "grad_norm": 17.751453399658203, "learning_rate": 2.099105299380592e-06, "loss": 0.6533, "step": 556800 }, { "epoch": 95.81039229181005, "grad_norm": 15.802138328552246, "learning_rate": 2.094803854094976e-06, "loss": 0.6777, "step": 556850 }, { "epoch": 95.81899518238129, "grad_norm": 13.617415428161621, "learning_rate": 2.09050240880936e-06, "loss": 0.5745, "step": 556900 }, { "epoch": 95.82759807295251, "grad_norm": 13.942891120910645, "learning_rate": 2.086200963523744e-06, "loss": 0.6403, "step": 556950 }, { "epoch": 95.83620096352374, "grad_norm": 12.950194358825684, "learning_rate": 2.081899518238128e-06, "loss": 0.566, "step": 557000 }, { "epoch": 95.84480385409498, "grad_norm": 11.455955505371094, "learning_rate": 2.0775980729525123e-06, "loss": 0.6666, "step": 557050 }, { "epoch": 95.8534067446662, "grad_norm": 9.864497184753418, "learning_rate": 2.073296627666896e-06, "loss": 0.6052, "step": 557100 }, { "epoch": 95.86200963523744, "grad_norm": 21.733152389526367, "learning_rate": 2.06899518238128e-06, "loss": 0.6696, "step": 557150 }, { "epoch": 95.87061252580867, "grad_norm": 7.209300994873047, "learning_rate": 2.0646937370956643e-06, "loss": 0.654, "step": 557200 }, { "epoch": 95.8792154163799, "grad_norm": 15.630556106567383, "learning_rate": 2.0603922918100484e-06, "loss": 0.6096, "step": 557250 }, { "epoch": 95.88781830695113, "grad_norm": 13.882407188415527, "learning_rate": 2.056090846524432e-06, "loss": 0.6524, "step": 557300 }, { "epoch": 95.89642119752237, "grad_norm": 16.821186065673828, "learning_rate": 2.0517894012388163e-06, "loss": 0.6504, "step": 557350 }, { "epoch": 95.9050240880936, "grad_norm": 24.290014266967773, "learning_rate": 2.0474879559532004e-06, "loss": 0.605, "step": 557400 }, { "epoch": 95.91362697866484, "grad_norm": 16.593708038330078, "learning_rate": 2.0431865106675845e-06, "loss": 0.6755, "step": 557450 }, { "epoch": 95.92222986923606, "grad_norm": 9.933677673339844, "learning_rate": 2.0388850653819682e-06, "loss": 0.6545, "step": 557500 }, { "epoch": 95.9308327598073, "grad_norm": 20.485902786254883, "learning_rate": 2.0345836200963524e-06, "loss": 0.6219, "step": 557550 }, { "epoch": 95.93943565037853, "grad_norm": 16.605968475341797, "learning_rate": 2.0302821748107365e-06, "loss": 0.6683, "step": 557600 }, { "epoch": 95.94803854094975, "grad_norm": 24.31841468811035, "learning_rate": 2.0259807295251202e-06, "loss": 0.6213, "step": 557650 }, { "epoch": 95.95664143152099, "grad_norm": 11.510846138000488, "learning_rate": 2.0216792842395044e-06, "loss": 0.5876, "step": 557700 }, { "epoch": 95.96524432209222, "grad_norm": 8.451972007751465, "learning_rate": 2.0173778389538885e-06, "loss": 0.6534, "step": 557750 }, { "epoch": 95.97384721266346, "grad_norm": 23.353158950805664, "learning_rate": 2.0130763936682726e-06, "loss": 0.6394, "step": 557800 }, { "epoch": 95.98245010323468, "grad_norm": 11.098348617553711, "learning_rate": 2.0087749483826564e-06, "loss": 0.6669, "step": 557850 }, { "epoch": 95.99105299380592, "grad_norm": 9.428887367248535, "learning_rate": 2.0044735030970405e-06, "loss": 0.5985, "step": 557900 }, { "epoch": 95.99965588437715, "grad_norm": 6.899609565734863, "learning_rate": 2.000172057811425e-06, "loss": 0.6281, "step": 557950 }, { "epoch": 96.0, "eval_accuracy": 0.5998107038375495, "eval_f1": 0.5875728573891685, "eval_f1_DuraRiadoRio_16x16": 0.5784905660377359, "eval_f1_Mole_16x16": 0.6216798277099784, "eval_f1_Quebrado_16x16": 0.7425949103045474, "eval_f1_RiadoRio_16x16": 0.36021354313009274, "eval_f1_RioFechado_16x16": 0.6348854397634885, "eval_loss": 1.9397656917572021, "eval_precision": 0.6072579111731475, "eval_precision_DuraRiadoRio_16x16": 0.5116822429906542, "eval_precision_Mole_16x16": 0.6928, "eval_precision_Quebrado_16x16": 0.714859437751004, "eval_precision_RiadoRio_16x16": 0.5687666370896185, "eval_precision_RioFechado_16x16": 0.5481812380344607, "eval_recall": 0.6038951029809672, "eval_recall_DuraRiadoRio_16x16": 0.6653645833333334, "eval_recall_Mole_16x16": 0.5638020833333334, "eval_recall_Quebrado_16x16": 0.7725694444444444, "eval_recall_RiadoRio_16x16": 0.2635690789473684, "eval_recall_RioFechado_16x16": 0.7541703248463565, "eval_runtime": 46.0389, "eval_samples_per_second": 252.439, "eval_steps_per_second": 15.791, "step": 557952 }, { "epoch": 96.00825877494839, "grad_norm": 21.11766242980957, "learning_rate": 1.9958706125258088e-06, "loss": 0.639, "step": 558000 }, { "epoch": 96.01686166551961, "grad_norm": 19.64596176147461, "learning_rate": 1.991569167240193e-06, "loss": 0.6111, "step": 558050 }, { "epoch": 96.02546455609085, "grad_norm": 24.48341941833496, "learning_rate": 1.987267721954577e-06, "loss": 0.6551, "step": 558100 }, { "epoch": 96.03406744666208, "grad_norm": 10.607918739318848, "learning_rate": 1.982966276668961e-06, "loss": 0.6376, "step": 558150 }, { "epoch": 96.0426703372333, "grad_norm": 18.231935501098633, "learning_rate": 1.978664831383345e-06, "loss": 0.5893, "step": 558200 }, { "epoch": 96.05127322780454, "grad_norm": 9.948173522949219, "learning_rate": 1.974363386097729e-06, "loss": 0.6459, "step": 558250 }, { "epoch": 96.05987611837577, "grad_norm": 9.47885513305664, "learning_rate": 1.970061940812113e-06, "loss": 0.609, "step": 558300 }, { "epoch": 96.06847900894701, "grad_norm": 14.96304988861084, "learning_rate": 1.9657604955264973e-06, "loss": 0.6027, "step": 558350 }, { "epoch": 96.07708189951823, "grad_norm": 16.562828063964844, "learning_rate": 1.961459050240881e-06, "loss": 0.6584, "step": 558400 }, { "epoch": 96.08568479008947, "grad_norm": 23.24325942993164, "learning_rate": 1.957157604955265e-06, "loss": 0.6868, "step": 558450 }, { "epoch": 96.0942876806607, "grad_norm": 16.472482681274414, "learning_rate": 1.9528561596696493e-06, "loss": 0.6363, "step": 558500 }, { "epoch": 96.10289057123194, "grad_norm": 13.012544631958008, "learning_rate": 1.948554714384033e-06, "loss": 0.6745, "step": 558550 }, { "epoch": 96.11149346180316, "grad_norm": 18.417266845703125, "learning_rate": 1.944253269098417e-06, "loss": 0.6664, "step": 558600 }, { "epoch": 96.1200963523744, "grad_norm": 21.725866317749023, "learning_rate": 1.9399518238128013e-06, "loss": 0.6746, "step": 558650 }, { "epoch": 96.12869924294563, "grad_norm": 27.073169708251953, "learning_rate": 1.9356503785271854e-06, "loss": 0.6353, "step": 558700 }, { "epoch": 96.13730213351687, "grad_norm": 18.23621940612793, "learning_rate": 1.931348933241569e-06, "loss": 0.6512, "step": 558750 }, { "epoch": 96.1459050240881, "grad_norm": 12.223143577575684, "learning_rate": 1.9270474879559533e-06, "loss": 0.6335, "step": 558800 }, { "epoch": 96.15450791465932, "grad_norm": 15.262873649597168, "learning_rate": 1.9227460426703374e-06, "loss": 0.6553, "step": 558850 }, { "epoch": 96.16311080523056, "grad_norm": 18.125242233276367, "learning_rate": 1.9184445973847215e-06, "loss": 0.6803, "step": 558900 }, { "epoch": 96.17171369580178, "grad_norm": 16.24423599243164, "learning_rate": 1.9141431520991052e-06, "loss": 0.6399, "step": 558950 }, { "epoch": 96.18031658637302, "grad_norm": 15.257929801940918, "learning_rate": 1.9098417068134894e-06, "loss": 0.5945, "step": 559000 }, { "epoch": 96.18891947694425, "grad_norm": 14.961328506469727, "learning_rate": 1.9055402615278735e-06, "loss": 0.6563, "step": 559050 }, { "epoch": 96.19752236751549, "grad_norm": 13.461302757263184, "learning_rate": 1.9012388162422574e-06, "loss": 0.6522, "step": 559100 }, { "epoch": 96.20612525808671, "grad_norm": 18.440629959106445, "learning_rate": 1.8969373709566414e-06, "loss": 0.6919, "step": 559150 }, { "epoch": 96.21472814865795, "grad_norm": 19.969762802124023, "learning_rate": 1.8926359256710255e-06, "loss": 0.6255, "step": 559200 }, { "epoch": 96.22333103922918, "grad_norm": 16.606136322021484, "learning_rate": 1.8883344803854094e-06, "loss": 0.6482, "step": 559250 }, { "epoch": 96.23193392980042, "grad_norm": 24.184810638427734, "learning_rate": 1.8840330350997936e-06, "loss": 0.622, "step": 559300 }, { "epoch": 96.24053682037164, "grad_norm": 16.100296020507812, "learning_rate": 1.8797315898141775e-06, "loss": 0.5919, "step": 559350 }, { "epoch": 96.24913971094287, "grad_norm": 19.2967472076416, "learning_rate": 1.8754301445285616e-06, "loss": 0.6124, "step": 559400 }, { "epoch": 96.25774260151411, "grad_norm": 18.047374725341797, "learning_rate": 1.8711286992429456e-06, "loss": 0.6305, "step": 559450 }, { "epoch": 96.26634549208534, "grad_norm": 8.601137161254883, "learning_rate": 1.8668272539573297e-06, "loss": 0.5923, "step": 559500 }, { "epoch": 96.27494838265658, "grad_norm": 22.994943618774414, "learning_rate": 1.8625258086717136e-06, "loss": 0.6272, "step": 559550 }, { "epoch": 96.2835512732278, "grad_norm": 11.0675687789917, "learning_rate": 1.8582243633860978e-06, "loss": 0.6281, "step": 559600 }, { "epoch": 96.29215416379904, "grad_norm": 15.295648574829102, "learning_rate": 1.8539229181004817e-06, "loss": 0.6147, "step": 559650 }, { "epoch": 96.30075705437027, "grad_norm": 23.949220657348633, "learning_rate": 1.8496214728148658e-06, "loss": 0.6279, "step": 559700 }, { "epoch": 96.3093599449415, "grad_norm": 10.693904876708984, "learning_rate": 1.8453200275292497e-06, "loss": 0.6897, "step": 559750 }, { "epoch": 96.31796283551273, "grad_norm": 9.806870460510254, "learning_rate": 1.841018582243634e-06, "loss": 0.6108, "step": 559800 }, { "epoch": 96.32656572608397, "grad_norm": 9.489972114562988, "learning_rate": 1.8367171369580182e-06, "loss": 0.5696, "step": 559850 }, { "epoch": 96.3351686166552, "grad_norm": 20.892549514770508, "learning_rate": 1.8324156916724022e-06, "loss": 0.6233, "step": 559900 }, { "epoch": 96.34377150722642, "grad_norm": 10.62155532836914, "learning_rate": 1.8281142463867863e-06, "loss": 0.6329, "step": 559950 }, { "epoch": 96.35237439779766, "grad_norm": 21.239233016967773, "learning_rate": 1.8238128011011702e-06, "loss": 0.6693, "step": 560000 }, { "epoch": 96.36097728836889, "grad_norm": 16.887048721313477, "learning_rate": 1.8195113558155541e-06, "loss": 0.6098, "step": 560050 }, { "epoch": 96.36958017894013, "grad_norm": 15.765379905700684, "learning_rate": 1.8152099105299383e-06, "loss": 0.6226, "step": 560100 }, { "epoch": 96.37818306951135, "grad_norm": 22.482572555541992, "learning_rate": 1.8109084652443222e-06, "loss": 0.6626, "step": 560150 }, { "epoch": 96.38678596008259, "grad_norm": 15.68970775604248, "learning_rate": 1.8066070199587063e-06, "loss": 0.6741, "step": 560200 }, { "epoch": 96.39538885065382, "grad_norm": 12.195517539978027, "learning_rate": 1.8023055746730903e-06, "loss": 0.6322, "step": 560250 }, { "epoch": 96.40399174122506, "grad_norm": 25.902315139770508, "learning_rate": 1.7980041293874744e-06, "loss": 0.6815, "step": 560300 }, { "epoch": 96.41259463179628, "grad_norm": 20.946155548095703, "learning_rate": 1.7937026841018583e-06, "loss": 0.6451, "step": 560350 }, { "epoch": 96.42119752236752, "grad_norm": 15.948908805847168, "learning_rate": 1.7894012388162425e-06, "loss": 0.6355, "step": 560400 }, { "epoch": 96.42980041293875, "grad_norm": 13.001435279846191, "learning_rate": 1.7850997935306264e-06, "loss": 0.6319, "step": 560450 }, { "epoch": 96.43840330350999, "grad_norm": 23.6208553314209, "learning_rate": 1.7807983482450105e-06, "loss": 0.683, "step": 560500 }, { "epoch": 96.44700619408121, "grad_norm": 14.155139923095703, "learning_rate": 1.7764969029593945e-06, "loss": 0.5919, "step": 560550 }, { "epoch": 96.45560908465244, "grad_norm": 19.603260040283203, "learning_rate": 1.7721954576737786e-06, "loss": 0.661, "step": 560600 }, { "epoch": 96.46421197522368, "grad_norm": 16.904964447021484, "learning_rate": 1.7678940123881625e-06, "loss": 0.6522, "step": 560650 }, { "epoch": 96.4728148657949, "grad_norm": 9.543476104736328, "learning_rate": 1.7635925671025464e-06, "loss": 0.6416, "step": 560700 }, { "epoch": 96.48141775636614, "grad_norm": 9.265984535217285, "learning_rate": 1.7592911218169306e-06, "loss": 0.6187, "step": 560750 }, { "epoch": 96.49002064693737, "grad_norm": 26.439334869384766, "learning_rate": 1.7549896765313145e-06, "loss": 0.6346, "step": 560800 }, { "epoch": 96.4986235375086, "grad_norm": 17.66620635986328, "learning_rate": 1.7506882312456986e-06, "loss": 0.6794, "step": 560850 }, { "epoch": 96.50722642807983, "grad_norm": 17.052093505859375, "learning_rate": 1.7463867859600826e-06, "loss": 0.6565, "step": 560900 }, { "epoch": 96.51582931865107, "grad_norm": 22.13176155090332, "learning_rate": 1.7420853406744667e-06, "loss": 0.6472, "step": 560950 }, { "epoch": 96.5244322092223, "grad_norm": 18.989643096923828, "learning_rate": 1.7377838953888506e-06, "loss": 0.605, "step": 561000 }, { "epoch": 96.53303509979354, "grad_norm": 19.501739501953125, "learning_rate": 1.7334824501032348e-06, "loss": 0.637, "step": 561050 }, { "epoch": 96.54163799036476, "grad_norm": 24.411205291748047, "learning_rate": 1.7291810048176187e-06, "loss": 0.6253, "step": 561100 }, { "epoch": 96.55024088093599, "grad_norm": 13.541902542114258, "learning_rate": 1.7248795595320028e-06, "loss": 0.6492, "step": 561150 }, { "epoch": 96.55884377150723, "grad_norm": 11.47234058380127, "learning_rate": 1.7205781142463868e-06, "loss": 0.5994, "step": 561200 }, { "epoch": 96.56744666207845, "grad_norm": 18.017087936401367, "learning_rate": 1.7162766689607709e-06, "loss": 0.6271, "step": 561250 }, { "epoch": 96.57604955264969, "grad_norm": 17.74639320373535, "learning_rate": 1.7119752236751548e-06, "loss": 0.6562, "step": 561300 }, { "epoch": 96.58465244322092, "grad_norm": 17.182601928710938, "learning_rate": 1.707673778389539e-06, "loss": 0.604, "step": 561350 }, { "epoch": 96.59325533379216, "grad_norm": 14.401815414428711, "learning_rate": 1.7033723331039229e-06, "loss": 0.6085, "step": 561400 }, { "epoch": 96.60185822436338, "grad_norm": 21.008108139038086, "learning_rate": 1.6990708878183068e-06, "loss": 0.6404, "step": 561450 }, { "epoch": 96.61046111493462, "grad_norm": 21.146114349365234, "learning_rate": 1.694769442532691e-06, "loss": 0.6363, "step": 561500 }, { "epoch": 96.61906400550585, "grad_norm": 12.127013206481934, "learning_rate": 1.6904679972470749e-06, "loss": 0.6725, "step": 561550 }, { "epoch": 96.62766689607709, "grad_norm": 7.791915416717529, "learning_rate": 1.686166551961459e-06, "loss": 0.6574, "step": 561600 }, { "epoch": 96.63626978664831, "grad_norm": 16.032278060913086, "learning_rate": 1.6818651066758433e-06, "loss": 0.6483, "step": 561650 }, { "epoch": 96.64487267721955, "grad_norm": 11.334680557250977, "learning_rate": 1.6775636613902273e-06, "loss": 0.6244, "step": 561700 }, { "epoch": 96.65347556779078, "grad_norm": 19.841552734375, "learning_rate": 1.6732622161046114e-06, "loss": 0.6734, "step": 561750 }, { "epoch": 96.662078458362, "grad_norm": 14.089622497558594, "learning_rate": 1.6689607708189953e-06, "loss": 0.6345, "step": 561800 }, { "epoch": 96.67068134893324, "grad_norm": 27.02939796447754, "learning_rate": 1.6646593255333795e-06, "loss": 0.6765, "step": 561850 }, { "epoch": 96.67928423950447, "grad_norm": 17.556529998779297, "learning_rate": 1.6603578802477634e-06, "loss": 0.6266, "step": 561900 }, { "epoch": 96.68788713007571, "grad_norm": 15.644983291625977, "learning_rate": 1.6560564349621475e-06, "loss": 0.6777, "step": 561950 }, { "epoch": 96.69649002064693, "grad_norm": 20.934560775756836, "learning_rate": 1.6517549896765315e-06, "loss": 0.6721, "step": 562000 }, { "epoch": 96.70509291121817, "grad_norm": 21.863880157470703, "learning_rate": 1.6474535443909156e-06, "loss": 0.6257, "step": 562050 }, { "epoch": 96.7136958017894, "grad_norm": 14.889857292175293, "learning_rate": 1.6431520991052995e-06, "loss": 0.634, "step": 562100 }, { "epoch": 96.72229869236064, "grad_norm": 22.2916316986084, "learning_rate": 1.6388506538196837e-06, "loss": 0.6204, "step": 562150 }, { "epoch": 96.73090158293186, "grad_norm": 20.469148635864258, "learning_rate": 1.6345492085340676e-06, "loss": 0.6436, "step": 562200 }, { "epoch": 96.7395044735031, "grad_norm": 12.377115249633789, "learning_rate": 1.6302477632484517e-06, "loss": 0.6292, "step": 562250 }, { "epoch": 96.74810736407433, "grad_norm": 17.269418716430664, "learning_rate": 1.6259463179628356e-06, "loss": 0.6435, "step": 562300 }, { "epoch": 96.75671025464555, "grad_norm": 23.8367919921875, "learning_rate": 1.6216448726772196e-06, "loss": 0.6021, "step": 562350 }, { "epoch": 96.7653131452168, "grad_norm": 15.376815795898438, "learning_rate": 1.6173434273916037e-06, "loss": 0.603, "step": 562400 }, { "epoch": 96.77391603578802, "grad_norm": 13.208040237426758, "learning_rate": 1.6130419821059876e-06, "loss": 0.6485, "step": 562450 }, { "epoch": 96.78251892635926, "grad_norm": 25.78182029724121, "learning_rate": 1.6087405368203718e-06, "loss": 0.6067, "step": 562500 }, { "epoch": 96.79112181693048, "grad_norm": 18.571189880371094, "learning_rate": 1.6044390915347557e-06, "loss": 0.6491, "step": 562550 }, { "epoch": 96.79972470750172, "grad_norm": 19.44525146484375, "learning_rate": 1.6001376462491398e-06, "loss": 0.7042, "step": 562600 }, { "epoch": 96.80832759807295, "grad_norm": 17.312828063964844, "learning_rate": 1.5958362009635238e-06, "loss": 0.6529, "step": 562650 }, { "epoch": 96.81693048864419, "grad_norm": 26.472444534301758, "learning_rate": 1.5915347556779079e-06, "loss": 0.6513, "step": 562700 }, { "epoch": 96.82553337921541, "grad_norm": 20.706188201904297, "learning_rate": 1.5872333103922918e-06, "loss": 0.6261, "step": 562750 }, { "epoch": 96.83413626978665, "grad_norm": 24.103050231933594, "learning_rate": 1.582931865106676e-06, "loss": 0.6657, "step": 562800 }, { "epoch": 96.84273916035788, "grad_norm": 17.73017692565918, "learning_rate": 1.5786304198210599e-06, "loss": 0.6464, "step": 562850 }, { "epoch": 96.8513420509291, "grad_norm": 29.766380310058594, "learning_rate": 1.574328974535444e-06, "loss": 0.6244, "step": 562900 }, { "epoch": 96.85994494150034, "grad_norm": 12.228777885437012, "learning_rate": 1.570027529249828e-06, "loss": 0.6578, "step": 562950 }, { "epoch": 96.86854783207157, "grad_norm": 13.682926177978516, "learning_rate": 1.5657260839642119e-06, "loss": 0.6395, "step": 563000 }, { "epoch": 96.87715072264281, "grad_norm": 17.993179321289062, "learning_rate": 1.5614246386785962e-06, "loss": 0.642, "step": 563050 }, { "epoch": 96.88575361321404, "grad_norm": 19.522314071655273, "learning_rate": 1.5571231933929801e-06, "loss": 0.5971, "step": 563100 }, { "epoch": 96.89435650378527, "grad_norm": 22.113452911376953, "learning_rate": 1.5528217481073643e-06, "loss": 0.5839, "step": 563150 }, { "epoch": 96.9029593943565, "grad_norm": 13.74272346496582, "learning_rate": 1.5485203028217482e-06, "loss": 0.6988, "step": 563200 }, { "epoch": 96.91156228492774, "grad_norm": 11.920940399169922, "learning_rate": 1.5442188575361323e-06, "loss": 0.5929, "step": 563250 }, { "epoch": 96.92016517549897, "grad_norm": 19.857744216918945, "learning_rate": 1.5399174122505163e-06, "loss": 0.6578, "step": 563300 }, { "epoch": 96.9287680660702, "grad_norm": 13.346370697021484, "learning_rate": 1.5356159669649004e-06, "loss": 0.6256, "step": 563350 }, { "epoch": 96.93737095664143, "grad_norm": 25.34156036376953, "learning_rate": 1.5313145216792843e-06, "loss": 0.6545, "step": 563400 }, { "epoch": 96.94597384721267, "grad_norm": 21.17723846435547, "learning_rate": 1.5270130763936683e-06, "loss": 0.6572, "step": 563450 }, { "epoch": 96.9545767377839, "grad_norm": 18.05360221862793, "learning_rate": 1.5227116311080524e-06, "loss": 0.6272, "step": 563500 }, { "epoch": 96.96317962835512, "grad_norm": 16.15801239013672, "learning_rate": 1.5184101858224363e-06, "loss": 0.6726, "step": 563550 }, { "epoch": 96.97178251892636, "grad_norm": 15.875999450683594, "learning_rate": 1.5141087405368205e-06, "loss": 0.6701, "step": 563600 }, { "epoch": 96.98038540949759, "grad_norm": 20.00021743774414, "learning_rate": 1.5098072952512044e-06, "loss": 0.6462, "step": 563650 }, { "epoch": 96.98898830006883, "grad_norm": 10.793068885803223, "learning_rate": 1.5055058499655885e-06, "loss": 0.6689, "step": 563700 }, { "epoch": 96.99759119064005, "grad_norm": 20.43362045288086, "learning_rate": 1.5012044046799724e-06, "loss": 0.6549, "step": 563750 }, { "epoch": 97.0, "eval_accuracy": 0.5368267079676475, "eval_f1": 0.5172506846910101, "eval_f1_DuraRiadoRio_16x16": 0.44432040189785094, "eval_f1_Mole_16x16": 0.6055658627087198, "eval_f1_Quebrado_16x16": 0.6460360504067635, "eval_f1_RiadoRio_16x16": 0.45434494927216584, "eval_f1_RioFechado_16x16": 0.4359861591695502, "eval_loss": 2.9523065090179443, "eval_precision": 0.5574447127469095, "eval_precision_DuraRiadoRio_16x16": 0.6223612197028929, "eval_precision_Mole_16x16": 0.5288399222294232, "eval_precision_Quebrado_16x16": 0.510718789407314, "eval_precision_RiadoRio_16x16": 0.4900095147478592, "eval_precision_RioFechado_16x16": 0.6352941176470588, "eval_recall": 0.5376230985487936, "eval_recall_DuraRiadoRio_16x16": 0.3454861111111111, "eval_recall_Mole_16x16": 0.7083333333333334, "eval_recall_Quebrado_16x16": 0.87890625, "eval_recall_RiadoRio_16x16": 0.42351973684210525, "eval_recall_RioFechado_16x16": 0.33187006145741876, "eval_runtime": 45.9012, "eval_samples_per_second": 253.196, "eval_steps_per_second": 15.838, "step": 563764 }, { "epoch": 97.00619408121129, "grad_norm": 13.141412734985352, "learning_rate": 1.4969029593943566e-06, "loss": 0.6568, "step": 563800 }, { "epoch": 97.01479697178252, "grad_norm": 23.57411003112793, "learning_rate": 1.4926015141087405e-06, "loss": 0.6402, "step": 563850 }, { "epoch": 97.02339986235376, "grad_norm": 6.86679744720459, "learning_rate": 1.4883000688231246e-06, "loss": 0.6572, "step": 563900 }, { "epoch": 97.03200275292498, "grad_norm": 22.409948348999023, "learning_rate": 1.4839986235375086e-06, "loss": 0.6378, "step": 563950 }, { "epoch": 97.04060564349622, "grad_norm": 11.28157901763916, "learning_rate": 1.4796971782518927e-06, "loss": 0.6107, "step": 564000 }, { "epoch": 97.04920853406745, "grad_norm": 18.164655685424805, "learning_rate": 1.4753957329662768e-06, "loss": 0.6615, "step": 564050 }, { "epoch": 97.05781142463867, "grad_norm": 8.996477127075195, "learning_rate": 1.4710942876806608e-06, "loss": 0.6213, "step": 564100 }, { "epoch": 97.06641431520991, "grad_norm": 15.611969947814941, "learning_rate": 1.466792842395045e-06, "loss": 0.6261, "step": 564150 }, { "epoch": 97.07501720578114, "grad_norm": 19.747543334960938, "learning_rate": 1.4624913971094288e-06, "loss": 0.6589, "step": 564200 }, { "epoch": 97.08362009635238, "grad_norm": 17.852785110473633, "learning_rate": 1.458189951823813e-06, "loss": 0.6027, "step": 564250 }, { "epoch": 97.0922229869236, "grad_norm": 20.429241180419922, "learning_rate": 1.4538885065381969e-06, "loss": 0.689, "step": 564300 }, { "epoch": 97.10082587749484, "grad_norm": 18.13984489440918, "learning_rate": 1.449587061252581e-06, "loss": 0.6328, "step": 564350 }, { "epoch": 97.10942876806607, "grad_norm": 11.87940502166748, "learning_rate": 1.445285615966965e-06, "loss": 0.6585, "step": 564400 }, { "epoch": 97.1180316586373, "grad_norm": 20.996999740600586, "learning_rate": 1.440984170681349e-06, "loss": 0.6797, "step": 564450 }, { "epoch": 97.12663454920853, "grad_norm": 9.471452713012695, "learning_rate": 1.436682725395733e-06, "loss": 0.6838, "step": 564500 }, { "epoch": 97.13523743977977, "grad_norm": 11.675865173339844, "learning_rate": 1.4323812801101171e-06, "loss": 0.6255, "step": 564550 }, { "epoch": 97.143840330351, "grad_norm": 14.999674797058105, "learning_rate": 1.428079834824501e-06, "loss": 0.6878, "step": 564600 }, { "epoch": 97.15244322092224, "grad_norm": 20.578609466552734, "learning_rate": 1.423778389538885e-06, "loss": 0.6374, "step": 564650 }, { "epoch": 97.16104611149346, "grad_norm": 16.052814483642578, "learning_rate": 1.4194769442532691e-06, "loss": 0.6539, "step": 564700 }, { "epoch": 97.16964900206469, "grad_norm": 14.7485933303833, "learning_rate": 1.415175498967653e-06, "loss": 0.6453, "step": 564750 }, { "epoch": 97.17825189263593, "grad_norm": 18.070697784423828, "learning_rate": 1.4108740536820372e-06, "loss": 0.6444, "step": 564800 }, { "epoch": 97.18685478320715, "grad_norm": 19.401296615600586, "learning_rate": 1.4065726083964211e-06, "loss": 0.6474, "step": 564850 }, { "epoch": 97.19545767377839, "grad_norm": 24.01054573059082, "learning_rate": 1.4022711631108053e-06, "loss": 0.6584, "step": 564900 }, { "epoch": 97.20406056434962, "grad_norm": 21.60502052307129, "learning_rate": 1.3979697178251894e-06, "loss": 0.643, "step": 564950 }, { "epoch": 97.21266345492086, "grad_norm": 16.405733108520508, "learning_rate": 1.3936682725395735e-06, "loss": 0.6899, "step": 565000 }, { "epoch": 97.22126634549208, "grad_norm": 12.276189804077148, "learning_rate": 1.3893668272539575e-06, "loss": 0.6054, "step": 565050 }, { "epoch": 97.22986923606332, "grad_norm": 10.305635452270508, "learning_rate": 1.3850653819683414e-06, "loss": 0.6481, "step": 565100 }, { "epoch": 97.23847212663455, "grad_norm": 17.556501388549805, "learning_rate": 1.3807639366827255e-06, "loss": 0.569, "step": 565150 }, { "epoch": 97.24707501720579, "grad_norm": 15.958812713623047, "learning_rate": 1.3764624913971094e-06, "loss": 0.5998, "step": 565200 }, { "epoch": 97.25567790777701, "grad_norm": 21.18624496459961, "learning_rate": 1.3721610461114936e-06, "loss": 0.6362, "step": 565250 }, { "epoch": 97.26428079834824, "grad_norm": 17.397422790527344, "learning_rate": 1.3678596008258775e-06, "loss": 0.6101, "step": 565300 }, { "epoch": 97.27288368891948, "grad_norm": 8.899341583251953, "learning_rate": 1.3635581555402616e-06, "loss": 0.6646, "step": 565350 }, { "epoch": 97.2814865794907, "grad_norm": 21.753623962402344, "learning_rate": 1.3592567102546456e-06, "loss": 0.6363, "step": 565400 }, { "epoch": 97.29008947006194, "grad_norm": 21.64961051940918, "learning_rate": 1.3549552649690297e-06, "loss": 0.6174, "step": 565450 }, { "epoch": 97.29869236063317, "grad_norm": 19.29924201965332, "learning_rate": 1.3506538196834136e-06, "loss": 0.6667, "step": 565500 }, { "epoch": 97.30729525120441, "grad_norm": 12.42965316772461, "learning_rate": 1.3463523743977978e-06, "loss": 0.6328, "step": 565550 }, { "epoch": 97.31589814177563, "grad_norm": 12.960898399353027, "learning_rate": 1.3420509291121817e-06, "loss": 0.6219, "step": 565600 }, { "epoch": 97.32450103234687, "grad_norm": 7.256200790405273, "learning_rate": 1.3377494838265658e-06, "loss": 0.675, "step": 565650 }, { "epoch": 97.3331039229181, "grad_norm": 19.34781837463379, "learning_rate": 1.3334480385409498e-06, "loss": 0.6301, "step": 565700 }, { "epoch": 97.34170681348934, "grad_norm": 16.091289520263672, "learning_rate": 1.3291465932553339e-06, "loss": 0.5764, "step": 565750 }, { "epoch": 97.35030970406056, "grad_norm": 21.537511825561523, "learning_rate": 1.3248451479697178e-06, "loss": 0.6477, "step": 565800 }, { "epoch": 97.35891259463179, "grad_norm": 18.95317268371582, "learning_rate": 1.320543702684102e-06, "loss": 0.7002, "step": 565850 }, { "epoch": 97.36751548520303, "grad_norm": 18.38950538635254, "learning_rate": 1.316242257398486e-06, "loss": 0.6006, "step": 565900 }, { "epoch": 97.37611837577425, "grad_norm": 23.412511825561523, "learning_rate": 1.31194081211287e-06, "loss": 0.675, "step": 565950 }, { "epoch": 97.3847212663455, "grad_norm": 13.599705696105957, "learning_rate": 1.3076393668272542e-06, "loss": 0.6289, "step": 566000 }, { "epoch": 97.39332415691672, "grad_norm": 17.148107528686523, "learning_rate": 1.303337921541638e-06, "loss": 0.5963, "step": 566050 }, { "epoch": 97.40192704748796, "grad_norm": 13.956710815429688, "learning_rate": 1.2990364762560222e-06, "loss": 0.608, "step": 566100 }, { "epoch": 97.41052993805918, "grad_norm": 9.44296646118164, "learning_rate": 1.2947350309704061e-06, "loss": 0.6258, "step": 566150 }, { "epoch": 97.41913282863042, "grad_norm": 19.212453842163086, "learning_rate": 1.2904335856847903e-06, "loss": 0.6469, "step": 566200 }, { "epoch": 97.42773571920165, "grad_norm": 19.181591033935547, "learning_rate": 1.2861321403991742e-06, "loss": 0.6492, "step": 566250 }, { "epoch": 97.43633860977289, "grad_norm": 29.21280288696289, "learning_rate": 1.2818306951135581e-06, "loss": 0.6193, "step": 566300 }, { "epoch": 97.44494150034411, "grad_norm": 32.723388671875, "learning_rate": 1.2775292498279423e-06, "loss": 0.6107, "step": 566350 }, { "epoch": 97.45354439091535, "grad_norm": 17.35598373413086, "learning_rate": 1.2732278045423262e-06, "loss": 0.6241, "step": 566400 }, { "epoch": 97.46214728148658, "grad_norm": 13.432461738586426, "learning_rate": 1.2689263592567103e-06, "loss": 0.6268, "step": 566450 }, { "epoch": 97.4707501720578, "grad_norm": 9.2783203125, "learning_rate": 1.2646249139710942e-06, "loss": 0.5992, "step": 566500 }, { "epoch": 97.47935306262904, "grad_norm": 11.454887390136719, "learning_rate": 1.2603234686854784e-06, "loss": 0.6375, "step": 566550 }, { "epoch": 97.48795595320027, "grad_norm": 21.98680305480957, "learning_rate": 1.2560220233998623e-06, "loss": 0.6133, "step": 566600 }, { "epoch": 97.49655884377151, "grad_norm": 20.38108253479004, "learning_rate": 1.2517205781142464e-06, "loss": 0.5729, "step": 566650 }, { "epoch": 97.50516173434274, "grad_norm": 22.34868049621582, "learning_rate": 1.2474191328286304e-06, "loss": 0.6466, "step": 566700 }, { "epoch": 97.51376462491397, "grad_norm": 11.243112564086914, "learning_rate": 1.2431176875430145e-06, "loss": 0.6008, "step": 566750 }, { "epoch": 97.5223675154852, "grad_norm": 24.33563995361328, "learning_rate": 1.2388162422573986e-06, "loss": 0.653, "step": 566800 }, { "epoch": 97.53097040605644, "grad_norm": 18.376108169555664, "learning_rate": 1.2345147969717826e-06, "loss": 0.63, "step": 566850 }, { "epoch": 97.53957329662767, "grad_norm": 11.234420776367188, "learning_rate": 1.2302133516861667e-06, "loss": 0.6069, "step": 566900 }, { "epoch": 97.5481761871989, "grad_norm": 14.480706214904785, "learning_rate": 1.2259119064005506e-06, "loss": 0.7401, "step": 566950 }, { "epoch": 97.55677907777013, "grad_norm": 13.688298225402832, "learning_rate": 1.2216104611149348e-06, "loss": 0.6267, "step": 567000 }, { "epoch": 97.56538196834136, "grad_norm": 21.45121955871582, "learning_rate": 1.2173090158293187e-06, "loss": 0.6368, "step": 567050 }, { "epoch": 97.5739848589126, "grad_norm": 9.717896461486816, "learning_rate": 1.2130075705437028e-06, "loss": 0.6223, "step": 567100 }, { "epoch": 97.58258774948382, "grad_norm": 20.03841781616211, "learning_rate": 1.2087061252580868e-06, "loss": 0.6459, "step": 567150 }, { "epoch": 97.59119064005506, "grad_norm": 23.156030654907227, "learning_rate": 1.204404679972471e-06, "loss": 0.666, "step": 567200 }, { "epoch": 97.59979353062629, "grad_norm": 15.738905906677246, "learning_rate": 1.2001032346868548e-06, "loss": 0.6312, "step": 567250 }, { "epoch": 97.60839642119753, "grad_norm": 10.420814514160156, "learning_rate": 1.195801789401239e-06, "loss": 0.6553, "step": 567300 }, { "epoch": 97.61699931176875, "grad_norm": 17.591449737548828, "learning_rate": 1.1915003441156229e-06, "loss": 0.6512, "step": 567350 }, { "epoch": 97.62560220233999, "grad_norm": 18.106237411499023, "learning_rate": 1.1871988988300068e-06, "loss": 0.6694, "step": 567400 }, { "epoch": 97.63420509291122, "grad_norm": 11.92471981048584, "learning_rate": 1.182897453544391e-06, "loss": 0.6391, "step": 567450 }, { "epoch": 97.64280798348246, "grad_norm": 17.39653968811035, "learning_rate": 1.1785960082587749e-06, "loss": 0.6726, "step": 567500 }, { "epoch": 97.65141087405368, "grad_norm": 21.19171142578125, "learning_rate": 1.174294562973159e-06, "loss": 0.6374, "step": 567550 }, { "epoch": 97.66001376462492, "grad_norm": 12.157723426818848, "learning_rate": 1.169993117687543e-06, "loss": 0.6715, "step": 567600 }, { "epoch": 97.66861665519615, "grad_norm": 13.978261947631836, "learning_rate": 1.165691672401927e-06, "loss": 0.6169, "step": 567650 }, { "epoch": 97.67721954576737, "grad_norm": 14.163848876953125, "learning_rate": 1.1613902271163112e-06, "loss": 0.6526, "step": 567700 }, { "epoch": 97.68582243633861, "grad_norm": 14.469045639038086, "learning_rate": 1.1570887818306953e-06, "loss": 0.6231, "step": 567750 }, { "epoch": 97.69442532690984, "grad_norm": 26.683719635009766, "learning_rate": 1.1527873365450793e-06, "loss": 0.6254, "step": 567800 }, { "epoch": 97.70302821748108, "grad_norm": 24.61820411682129, "learning_rate": 1.1484858912594632e-06, "loss": 0.6245, "step": 567850 }, { "epoch": 97.7116311080523, "grad_norm": 23.387210845947266, "learning_rate": 1.1441844459738473e-06, "loss": 0.6642, "step": 567900 }, { "epoch": 97.72023399862354, "grad_norm": 23.606287002563477, "learning_rate": 1.1398830006882313e-06, "loss": 0.6789, "step": 567950 }, { "epoch": 97.72883688919477, "grad_norm": 24.334745407104492, "learning_rate": 1.1355815554026154e-06, "loss": 0.6691, "step": 568000 }, { "epoch": 97.737439779766, "grad_norm": 15.031010627746582, "learning_rate": 1.1312801101169993e-06, "loss": 0.608, "step": 568050 }, { "epoch": 97.74604267033723, "grad_norm": 15.414190292358398, "learning_rate": 1.1269786648313835e-06, "loss": 0.6802, "step": 568100 }, { "epoch": 97.75464556090847, "grad_norm": 13.842873573303223, "learning_rate": 1.1226772195457674e-06, "loss": 0.6527, "step": 568150 }, { "epoch": 97.7632484514797, "grad_norm": 14.449563026428223, "learning_rate": 1.1183757742601515e-06, "loss": 0.6504, "step": 568200 }, { "epoch": 97.77185134205092, "grad_norm": 18.77463150024414, "learning_rate": 1.1140743289745354e-06, "loss": 0.5796, "step": 568250 }, { "epoch": 97.78045423262216, "grad_norm": 25.781232833862305, "learning_rate": 1.1097728836889196e-06, "loss": 0.6855, "step": 568300 }, { "epoch": 97.78905712319339, "grad_norm": 16.80483627319336, "learning_rate": 1.1054714384033035e-06, "loss": 0.7557, "step": 568350 }, { "epoch": 97.79766001376463, "grad_norm": 10.571518898010254, "learning_rate": 1.1011699931176876e-06, "loss": 0.7033, "step": 568400 }, { "epoch": 97.80626290433585, "grad_norm": 24.58553123474121, "learning_rate": 1.0968685478320716e-06, "loss": 0.6289, "step": 568450 }, { "epoch": 97.81486579490709, "grad_norm": 25.537586212158203, "learning_rate": 1.0925671025464557e-06, "loss": 0.6077, "step": 568500 }, { "epoch": 97.82346868547832, "grad_norm": 23.131746292114258, "learning_rate": 1.0882656572608396e-06, "loss": 0.6599, "step": 568550 }, { "epoch": 97.83207157604956, "grad_norm": 18.08833122253418, "learning_rate": 1.0839642119752236e-06, "loss": 0.6509, "step": 568600 }, { "epoch": 97.84067446662078, "grad_norm": 14.714912414550781, "learning_rate": 1.079662766689608e-06, "loss": 0.5965, "step": 568650 }, { "epoch": 97.84927735719202, "grad_norm": 19.37051010131836, "learning_rate": 1.0753613214039918e-06, "loss": 0.6361, "step": 568700 }, { "epoch": 97.85788024776325, "grad_norm": 15.894518852233887, "learning_rate": 1.071059876118376e-06, "loss": 0.5574, "step": 568750 }, { "epoch": 97.86648313833447, "grad_norm": 9.394820213317871, "learning_rate": 1.0667584308327599e-06, "loss": 0.6459, "step": 568800 }, { "epoch": 97.87508602890571, "grad_norm": 15.678838729858398, "learning_rate": 1.062456985547144e-06, "loss": 0.6334, "step": 568850 }, { "epoch": 97.88368891947694, "grad_norm": 9.741395950317383, "learning_rate": 1.058155540261528e-06, "loss": 0.6544, "step": 568900 }, { "epoch": 97.89229181004818, "grad_norm": 21.802576065063477, "learning_rate": 1.053854094975912e-06, "loss": 0.5797, "step": 568950 }, { "epoch": 97.9008947006194, "grad_norm": 17.14535140991211, "learning_rate": 1.049552649690296e-06, "loss": 0.6409, "step": 569000 }, { "epoch": 97.90949759119064, "grad_norm": 8.215913772583008, "learning_rate": 1.04525120440468e-06, "loss": 0.6367, "step": 569050 }, { "epoch": 97.91810048176187, "grad_norm": 13.350854873657227, "learning_rate": 1.040949759119064e-06, "loss": 0.645, "step": 569100 }, { "epoch": 97.92670337233311, "grad_norm": 22.056259155273438, "learning_rate": 1.036648313833448e-06, "loss": 0.615, "step": 569150 }, { "epoch": 97.93530626290433, "grad_norm": 11.541362762451172, "learning_rate": 1.0323468685478321e-06, "loss": 0.6134, "step": 569200 }, { "epoch": 97.94390915347557, "grad_norm": 13.361388206481934, "learning_rate": 1.028045423262216e-06, "loss": 0.6203, "step": 569250 }, { "epoch": 97.9525120440468, "grad_norm": 13.07082462310791, "learning_rate": 1.0237439779766002e-06, "loss": 0.6465, "step": 569300 }, { "epoch": 97.96111493461804, "grad_norm": 19.487030029296875, "learning_rate": 1.0194425326909841e-06, "loss": 0.6662, "step": 569350 }, { "epoch": 97.96971782518926, "grad_norm": 20.74216079711914, "learning_rate": 1.0151410874053683e-06, "loss": 0.5877, "step": 569400 }, { "epoch": 97.97832071576049, "grad_norm": 15.004919052124023, "learning_rate": 1.0108396421197522e-06, "loss": 0.6252, "step": 569450 }, { "epoch": 97.98692360633173, "grad_norm": 14.185888290405273, "learning_rate": 1.0065381968341363e-06, "loss": 0.6432, "step": 569500 }, { "epoch": 97.99552649690295, "grad_norm": 19.507516860961914, "learning_rate": 1.0022367515485202e-06, "loss": 0.6468, "step": 569550 }, { "epoch": 98.0, "eval_accuracy": 0.5503355704697986, "eval_f1": 0.5437209062592141, "eval_f1_DuraRiadoRio_16x16": 0.49038461538461536, "eval_f1_Mole_16x16": 0.608983451536643, "eval_f1_Quebrado_16x16": 0.6561972306653158, "eval_f1_RiadoRio_16x16": 0.47640282422891117, "eval_f1_RioFechado_16x16": 0.48663640948058495, "eval_loss": 2.790480136871338, "eval_precision": 0.5699078565086919, "eval_precision_DuraRiadoRio_16x16": 0.6375, "eval_precision_Mole_16x16": 0.6687435098650052, "eval_precision_Quebrado_16x16": 0.5370370370370371, "eval_precision_RiadoRio_16x16": 0.43457627118644065, "eval_precision_RioFechado_16x16": 0.5716824644549763, "eval_recall": 0.5503073231943996, "eval_recall_DuraRiadoRio_16x16": 0.3984375, "eval_recall_Mole_16x16": 0.5590277777777778, "eval_recall_Quebrado_16x16": 0.8433159722222222, "eval_recall_RiadoRio_16x16": 0.5271381578947368, "eval_recall_RioFechado_16x16": 0.4236172080772608, "eval_runtime": 46.4268, "eval_samples_per_second": 250.329, "eval_steps_per_second": 15.659, "step": 569576 }, { "epoch": 98.0041293874742, "grad_norm": 16.22036361694336, "learning_rate": 9.979353062629044e-07, "loss": 0.6714, "step": 569600 }, { "epoch": 98.01273227804542, "grad_norm": 10.134373664855957, "learning_rate": 9.936338609772885e-07, "loss": 0.6789, "step": 569650 }, { "epoch": 98.02133516861666, "grad_norm": 15.088767051696777, "learning_rate": 9.893324156916724e-07, "loss": 0.6362, "step": 569700 }, { "epoch": 98.02993805918788, "grad_norm": 17.47422218322754, "learning_rate": 9.850309704060566e-07, "loss": 0.6728, "step": 569750 }, { "epoch": 98.03854094975912, "grad_norm": 7.591610908508301, "learning_rate": 9.807295251204405e-07, "loss": 0.5843, "step": 569800 }, { "epoch": 98.04714384033035, "grad_norm": 18.518272399902344, "learning_rate": 9.764280798348246e-07, "loss": 0.6415, "step": 569850 }, { "epoch": 98.05574673090159, "grad_norm": 22.641332626342773, "learning_rate": 9.721266345492086e-07, "loss": 0.657, "step": 569900 }, { "epoch": 98.06434962147281, "grad_norm": 10.207098960876465, "learning_rate": 9.678251892635927e-07, "loss": 0.6357, "step": 569950 }, { "epoch": 98.07295251204404, "grad_norm": 19.400142669677734, "learning_rate": 9.635237439779766e-07, "loss": 0.6464, "step": 570000 }, { "epoch": 98.08155540261528, "grad_norm": 11.250284194946289, "learning_rate": 9.592222986923608e-07, "loss": 0.6755, "step": 570050 }, { "epoch": 98.0901582931865, "grad_norm": 14.391164779663086, "learning_rate": 9.549208534067447e-07, "loss": 0.6187, "step": 570100 }, { "epoch": 98.09876118375774, "grad_norm": 14.083440780639648, "learning_rate": 9.506194081211287e-07, "loss": 0.6578, "step": 570150 }, { "epoch": 98.10736407432897, "grad_norm": 13.93995189666748, "learning_rate": 9.463179628355128e-07, "loss": 0.6551, "step": 570200 }, { "epoch": 98.11596696490021, "grad_norm": 11.66760540008545, "learning_rate": 9.420165175498968e-07, "loss": 0.5944, "step": 570250 }, { "epoch": 98.12456985547144, "grad_norm": 13.525476455688477, "learning_rate": 9.377150722642808e-07, "loss": 0.6805, "step": 570300 }, { "epoch": 98.13317274604267, "grad_norm": 14.240143775939941, "learning_rate": 9.334136269786648e-07, "loss": 0.6648, "step": 570350 }, { "epoch": 98.1417756366139, "grad_norm": 9.21216869354248, "learning_rate": 9.291121816930489e-07, "loss": 0.6754, "step": 570400 }, { "epoch": 98.15037852718514, "grad_norm": 14.795700073242188, "learning_rate": 9.248107364074329e-07, "loss": 0.611, "step": 570450 }, { "epoch": 98.15898141775637, "grad_norm": 19.66262435913086, "learning_rate": 9.20509291121817e-07, "loss": 0.6616, "step": 570500 }, { "epoch": 98.1675843083276, "grad_norm": 15.123143196105957, "learning_rate": 9.162078458362011e-07, "loss": 0.6276, "step": 570550 }, { "epoch": 98.17618719889883, "grad_norm": 18.91570281982422, "learning_rate": 9.119064005505851e-07, "loss": 0.5973, "step": 570600 }, { "epoch": 98.18479008947006, "grad_norm": 15.383556365966797, "learning_rate": 9.076049552649691e-07, "loss": 0.6128, "step": 570650 }, { "epoch": 98.1933929800413, "grad_norm": 12.50133991241455, "learning_rate": 9.033035099793532e-07, "loss": 0.6561, "step": 570700 }, { "epoch": 98.20199587061252, "grad_norm": 17.751935958862305, "learning_rate": 8.990020646937372e-07, "loss": 0.6331, "step": 570750 }, { "epoch": 98.21059876118376, "grad_norm": 11.809656143188477, "learning_rate": 8.947006194081212e-07, "loss": 0.6539, "step": 570800 }, { "epoch": 98.21920165175499, "grad_norm": 11.778143882751465, "learning_rate": 8.903991741225053e-07, "loss": 0.6484, "step": 570850 }, { "epoch": 98.22780454232623, "grad_norm": 12.381441116333008, "learning_rate": 8.860977288368893e-07, "loss": 0.6454, "step": 570900 }, { "epoch": 98.23640743289745, "grad_norm": 16.110116958618164, "learning_rate": 8.817962835512732e-07, "loss": 0.7284, "step": 570950 }, { "epoch": 98.24501032346869, "grad_norm": 12.199095726013184, "learning_rate": 8.774948382656573e-07, "loss": 0.6542, "step": 571000 }, { "epoch": 98.25361321403992, "grad_norm": 17.845556259155273, "learning_rate": 8.731933929800413e-07, "loss": 0.6983, "step": 571050 }, { "epoch": 98.26221610461116, "grad_norm": 22.392898559570312, "learning_rate": 8.688919476944253e-07, "loss": 0.6139, "step": 571100 }, { "epoch": 98.27081899518238, "grad_norm": 15.182225227355957, "learning_rate": 8.645905024088093e-07, "loss": 0.5851, "step": 571150 }, { "epoch": 98.2794218857536, "grad_norm": 13.779279708862305, "learning_rate": 8.602890571231934e-07, "loss": 0.6206, "step": 571200 }, { "epoch": 98.28802477632485, "grad_norm": 9.197880744934082, "learning_rate": 8.559876118375774e-07, "loss": 0.5811, "step": 571250 }, { "epoch": 98.29662766689607, "grad_norm": 17.368846893310547, "learning_rate": 8.516861665519614e-07, "loss": 0.6002, "step": 571300 }, { "epoch": 98.30523055746731, "grad_norm": 24.131620407104492, "learning_rate": 8.473847212663455e-07, "loss": 0.6284, "step": 571350 }, { "epoch": 98.31383344803854, "grad_norm": 24.327608108520508, "learning_rate": 8.430832759807295e-07, "loss": 0.6497, "step": 571400 }, { "epoch": 98.32243633860978, "grad_norm": 24.27882957458496, "learning_rate": 8.387818306951136e-07, "loss": 0.6542, "step": 571450 }, { "epoch": 98.331039229181, "grad_norm": 12.824979782104492, "learning_rate": 8.344803854094977e-07, "loss": 0.685, "step": 571500 }, { "epoch": 98.33964211975224, "grad_norm": 10.894647598266602, "learning_rate": 8.301789401238817e-07, "loss": 0.6099, "step": 571550 }, { "epoch": 98.34824501032347, "grad_norm": 20.16236114501953, "learning_rate": 8.258774948382657e-07, "loss": 0.6272, "step": 571600 }, { "epoch": 98.3568479008947, "grad_norm": 28.8655948638916, "learning_rate": 8.215760495526498e-07, "loss": 0.6035, "step": 571650 }, { "epoch": 98.36545079146593, "grad_norm": 11.105042457580566, "learning_rate": 8.172746042670338e-07, "loss": 0.6619, "step": 571700 }, { "epoch": 98.37405368203716, "grad_norm": 9.244367599487305, "learning_rate": 8.129731589814178e-07, "loss": 0.6508, "step": 571750 }, { "epoch": 98.3826565726084, "grad_norm": 10.723464965820312, "learning_rate": 8.086717136958019e-07, "loss": 0.6091, "step": 571800 }, { "epoch": 98.39125946317962, "grad_norm": 11.696352005004883, "learning_rate": 8.043702684101859e-07, "loss": 0.6387, "step": 571850 }, { "epoch": 98.39986235375086, "grad_norm": 14.653553009033203, "learning_rate": 8.000688231245699e-07, "loss": 0.6332, "step": 571900 }, { "epoch": 98.40846524432209, "grad_norm": 10.719120025634766, "learning_rate": 7.957673778389539e-07, "loss": 0.6047, "step": 571950 }, { "epoch": 98.41706813489333, "grad_norm": 14.129646301269531, "learning_rate": 7.91465932553338e-07, "loss": 0.6409, "step": 572000 }, { "epoch": 98.42567102546455, "grad_norm": 13.477071762084961, "learning_rate": 7.87164487267722e-07, "loss": 0.6348, "step": 572050 }, { "epoch": 98.43427391603579, "grad_norm": 16.93089485168457, "learning_rate": 7.828630419821059e-07, "loss": 0.6326, "step": 572100 }, { "epoch": 98.44287680660702, "grad_norm": 20.67863655090332, "learning_rate": 7.785615966964901e-07, "loss": 0.688, "step": 572150 }, { "epoch": 98.45147969717826, "grad_norm": 13.632122993469238, "learning_rate": 7.742601514108741e-07, "loss": 0.6349, "step": 572200 }, { "epoch": 98.46008258774948, "grad_norm": 10.026862144470215, "learning_rate": 7.699587061252581e-07, "loss": 0.6114, "step": 572250 }, { "epoch": 98.46868547832072, "grad_norm": 30.1480770111084, "learning_rate": 7.656572608396422e-07, "loss": 0.6039, "step": 572300 }, { "epoch": 98.47728836889195, "grad_norm": 16.156017303466797, "learning_rate": 7.613558155540262e-07, "loss": 0.6354, "step": 572350 }, { "epoch": 98.48589125946317, "grad_norm": 21.684616088867188, "learning_rate": 7.570543702684102e-07, "loss": 0.671, "step": 572400 }, { "epoch": 98.49449415003441, "grad_norm": 18.41983413696289, "learning_rate": 7.527529249827943e-07, "loss": 0.6029, "step": 572450 }, { "epoch": 98.50309704060564, "grad_norm": 24.766342163085938, "learning_rate": 7.484514796971783e-07, "loss": 0.6288, "step": 572500 }, { "epoch": 98.51169993117688, "grad_norm": 15.522348403930664, "learning_rate": 7.441500344115623e-07, "loss": 0.6826, "step": 572550 }, { "epoch": 98.5203028217481, "grad_norm": 16.61366081237793, "learning_rate": 7.398485891259463e-07, "loss": 0.5839, "step": 572600 }, { "epoch": 98.52890571231934, "grad_norm": 9.534927368164062, "learning_rate": 7.355471438403304e-07, "loss": 0.6232, "step": 572650 }, { "epoch": 98.53750860289057, "grad_norm": 20.366100311279297, "learning_rate": 7.312456985547144e-07, "loss": 0.6072, "step": 572700 }, { "epoch": 98.54611149346181, "grad_norm": 10.824377059936523, "learning_rate": 7.269442532690984e-07, "loss": 0.6397, "step": 572750 }, { "epoch": 98.55471438403303, "grad_norm": 15.825360298156738, "learning_rate": 7.226428079834825e-07, "loss": 0.5956, "step": 572800 }, { "epoch": 98.56331727460427, "grad_norm": 8.237814903259277, "learning_rate": 7.183413626978665e-07, "loss": 0.603, "step": 572850 }, { "epoch": 98.5719201651755, "grad_norm": 18.652997970581055, "learning_rate": 7.140399174122505e-07, "loss": 0.6948, "step": 572900 }, { "epoch": 98.58052305574672, "grad_norm": 21.239784240722656, "learning_rate": 7.097384721266346e-07, "loss": 0.6175, "step": 572950 }, { "epoch": 98.58912594631796, "grad_norm": 15.484299659729004, "learning_rate": 7.054370268410186e-07, "loss": 0.6162, "step": 573000 }, { "epoch": 98.59772883688919, "grad_norm": 14.89274787902832, "learning_rate": 7.011355815554026e-07, "loss": 0.5778, "step": 573050 }, { "epoch": 98.60633172746043, "grad_norm": 12.425783157348633, "learning_rate": 6.968341362697868e-07, "loss": 0.6948, "step": 573100 }, { "epoch": 98.61493461803165, "grad_norm": 26.113723754882812, "learning_rate": 6.925326909841707e-07, "loss": 0.6311, "step": 573150 }, { "epoch": 98.6235375086029, "grad_norm": 19.807666778564453, "learning_rate": 6.882312456985547e-07, "loss": 0.6053, "step": 573200 }, { "epoch": 98.63214039917412, "grad_norm": 13.955669403076172, "learning_rate": 6.839298004129388e-07, "loss": 0.6689, "step": 573250 }, { "epoch": 98.64074328974536, "grad_norm": 17.59453773498535, "learning_rate": 6.796283551273228e-07, "loss": 0.6106, "step": 573300 }, { "epoch": 98.64934618031658, "grad_norm": 11.899919509887695, "learning_rate": 6.753269098417068e-07, "loss": 0.6089, "step": 573350 }, { "epoch": 98.65794907088782, "grad_norm": 11.79401969909668, "learning_rate": 6.710254645560908e-07, "loss": 0.6522, "step": 573400 }, { "epoch": 98.66655196145905, "grad_norm": 16.250158309936523, "learning_rate": 6.667240192704749e-07, "loss": 0.6319, "step": 573450 }, { "epoch": 98.67515485203029, "grad_norm": 12.241318702697754, "learning_rate": 6.624225739848589e-07, "loss": 0.6633, "step": 573500 }, { "epoch": 98.68375774260151, "grad_norm": 17.616682052612305, "learning_rate": 6.58121128699243e-07, "loss": 0.6461, "step": 573550 }, { "epoch": 98.69236063317274, "grad_norm": 17.107439041137695, "learning_rate": 6.538196834136271e-07, "loss": 0.6235, "step": 573600 }, { "epoch": 98.70096352374398, "grad_norm": 12.243489265441895, "learning_rate": 6.495182381280111e-07, "loss": 0.6304, "step": 573650 }, { "epoch": 98.7095664143152, "grad_norm": 17.555137634277344, "learning_rate": 6.452167928423951e-07, "loss": 0.6184, "step": 573700 }, { "epoch": 98.71816930488644, "grad_norm": 5.917167663574219, "learning_rate": 6.409153475567791e-07, "loss": 0.592, "step": 573750 }, { "epoch": 98.72677219545767, "grad_norm": 25.11557960510254, "learning_rate": 6.366139022711631e-07, "loss": 0.6995, "step": 573800 }, { "epoch": 98.73537508602891, "grad_norm": 10.21005630493164, "learning_rate": 6.323124569855471e-07, "loss": 0.6311, "step": 573850 }, { "epoch": 98.74397797660014, "grad_norm": 11.712712287902832, "learning_rate": 6.280110116999312e-07, "loss": 0.636, "step": 573900 }, { "epoch": 98.75258086717137, "grad_norm": 25.022563934326172, "learning_rate": 6.237095664143152e-07, "loss": 0.6673, "step": 573950 }, { "epoch": 98.7611837577426, "grad_norm": 15.88394832611084, "learning_rate": 6.194081211286993e-07, "loss": 0.5888, "step": 574000 }, { "epoch": 98.76978664831384, "grad_norm": 29.505672454833984, "learning_rate": 6.151066758430834e-07, "loss": 0.576, "step": 574050 }, { "epoch": 98.77838953888507, "grad_norm": 18.68693733215332, "learning_rate": 6.108052305574674e-07, "loss": 0.6208, "step": 574100 }, { "epoch": 98.78699242945629, "grad_norm": 15.735933303833008, "learning_rate": 6.065037852718514e-07, "loss": 0.5924, "step": 574150 }, { "epoch": 98.79559532002753, "grad_norm": 15.306607246398926, "learning_rate": 6.022023399862354e-07, "loss": 0.6207, "step": 574200 }, { "epoch": 98.80419821059876, "grad_norm": 28.18427085876465, "learning_rate": 5.979008947006195e-07, "loss": 0.6526, "step": 574250 }, { "epoch": 98.81280110117, "grad_norm": 22.294782638549805, "learning_rate": 5.935994494150034e-07, "loss": 0.6463, "step": 574300 }, { "epoch": 98.82140399174122, "grad_norm": 24.855514526367188, "learning_rate": 5.892980041293874e-07, "loss": 0.7002, "step": 574350 }, { "epoch": 98.83000688231246, "grad_norm": 11.62291145324707, "learning_rate": 5.849965588437715e-07, "loss": 0.6158, "step": 574400 }, { "epoch": 98.83860977288369, "grad_norm": 10.788324356079102, "learning_rate": 5.806951135581556e-07, "loss": 0.6666, "step": 574450 }, { "epoch": 98.84721266345493, "grad_norm": 25.30550193786621, "learning_rate": 5.763936682725396e-07, "loss": 0.6053, "step": 574500 }, { "epoch": 98.85581555402615, "grad_norm": 13.2006196975708, "learning_rate": 5.720922229869237e-07, "loss": 0.6218, "step": 574550 }, { "epoch": 98.86441844459739, "grad_norm": 12.129538536071777, "learning_rate": 5.677907777013077e-07, "loss": 0.6038, "step": 574600 }, { "epoch": 98.87302133516862, "grad_norm": 15.511380195617676, "learning_rate": 5.634893324156917e-07, "loss": 0.6099, "step": 574650 }, { "epoch": 98.88162422573984, "grad_norm": 15.465303421020508, "learning_rate": 5.591878871300758e-07, "loss": 0.6972, "step": 574700 }, { "epoch": 98.89022711631108, "grad_norm": 14.81236457824707, "learning_rate": 5.548864418444598e-07, "loss": 0.6533, "step": 574750 }, { "epoch": 98.8988300068823, "grad_norm": 21.72563362121582, "learning_rate": 5.505849965588438e-07, "loss": 0.596, "step": 574800 }, { "epoch": 98.90743289745355, "grad_norm": 15.661117553710938, "learning_rate": 5.462835512732279e-07, "loss": 0.6161, "step": 574850 }, { "epoch": 98.91603578802477, "grad_norm": 22.628087997436523, "learning_rate": 5.419821059876118e-07, "loss": 0.6094, "step": 574900 }, { "epoch": 98.92463867859601, "grad_norm": 18.385820388793945, "learning_rate": 5.376806607019959e-07, "loss": 0.6708, "step": 574950 }, { "epoch": 98.93324156916724, "grad_norm": 14.045741081237793, "learning_rate": 5.333792154163799e-07, "loss": 0.6673, "step": 575000 }, { "epoch": 98.94184445973848, "grad_norm": 11.349858283996582, "learning_rate": 5.29077770130764e-07, "loss": 0.6132, "step": 575050 }, { "epoch": 98.9504473503097, "grad_norm": 13.043968200683594, "learning_rate": 5.24776324845148e-07, "loss": 0.6272, "step": 575100 }, { "epoch": 98.95905024088094, "grad_norm": 19.070497512817383, "learning_rate": 5.20474879559532e-07, "loss": 0.653, "step": 575150 }, { "epoch": 98.96765313145217, "grad_norm": 20.346359252929688, "learning_rate": 5.161734342739161e-07, "loss": 0.6673, "step": 575200 }, { "epoch": 98.9762560220234, "grad_norm": 14.34564208984375, "learning_rate": 5.118719889883001e-07, "loss": 0.6558, "step": 575250 }, { "epoch": 98.98485891259463, "grad_norm": 5.714820384979248, "learning_rate": 5.075705437026841e-07, "loss": 0.6439, "step": 575300 }, { "epoch": 98.99346180316586, "grad_norm": 9.55800724029541, "learning_rate": 5.032690984170682e-07, "loss": 0.6366, "step": 575350 }, { "epoch": 99.0, "eval_accuracy": 0.5736534159352952, "eval_f1": 0.5666849915288468, "eval_f1_DuraRiadoRio_16x16": 0.5020576131687243, "eval_f1_Mole_16x16": 0.5183898059658268, "eval_f1_Quebrado_16x16": 0.746745065098698, "eval_f1_RiadoRio_16x16": 0.4826805524111388, "eval_f1_RioFechado_16x16": 0.5835519209998457, "eval_loss": 2.2962472438812256, "eval_precision": 0.6113664227711496, "eval_precision_DuraRiadoRio_16x16": 0.5675971538040504, "eval_precision_Mole_16x16": 0.7789382071366405, "eval_precision_Quebrado_16x16": 0.7233523189585028, "eval_precision_RiadoRio_16x16": 0.5370277078085642, "eval_precision_RioFechado_16x16": 0.44991672614798955, "eval_recall": 0.5757359118365859, "eval_recall_DuraRiadoRio_16x16": 0.4500868055555556, "eval_recall_Mole_16x16": 0.3884548611111111, "eval_recall_Quebrado_16x16": 0.7717013888888888, "eval_recall_RiadoRio_16x16": 0.43832236842105265, "eval_recall_RioFechado_16x16": 0.8301141352063214, "eval_runtime": 46.8977, "eval_samples_per_second": 247.816, "eval_steps_per_second": 15.502, "step": 575388 }, { "epoch": 99.0020646937371, "grad_norm": 5.060451030731201, "learning_rate": 4.989676531314522e-07, "loss": 0.6386, "step": 575400 }, { "epoch": 99.01066758430832, "grad_norm": 24.734567642211914, "learning_rate": 4.946662078458362e-07, "loss": 0.6141, "step": 575450 }, { "epoch": 99.01927047487956, "grad_norm": 11.09949779510498, "learning_rate": 4.903647625602203e-07, "loss": 0.6155, "step": 575500 }, { "epoch": 99.02787336545079, "grad_norm": 19.241863250732422, "learning_rate": 4.860633172746043e-07, "loss": 0.6107, "step": 575550 }, { "epoch": 99.03647625602203, "grad_norm": 18.038888931274414, "learning_rate": 4.817618719889883e-07, "loss": 0.5891, "step": 575600 }, { "epoch": 99.04507914659325, "grad_norm": 21.434415817260742, "learning_rate": 4.774604267033723e-07, "loss": 0.6678, "step": 575650 }, { "epoch": 99.05368203716449, "grad_norm": 17.06410789489746, "learning_rate": 4.731589814177564e-07, "loss": 0.6913, "step": 575700 }, { "epoch": 99.06228492773572, "grad_norm": 26.295061111450195, "learning_rate": 4.688575361321404e-07, "loss": 0.6334, "step": 575750 }, { "epoch": 99.07088781830696, "grad_norm": 18.294584274291992, "learning_rate": 4.6455609084652444e-07, "loss": 0.5731, "step": 575800 }, { "epoch": 99.07949070887818, "grad_norm": 12.44128131866455, "learning_rate": 4.602546455609085e-07, "loss": 0.6518, "step": 575850 }, { "epoch": 99.08809359944941, "grad_norm": 9.12630558013916, "learning_rate": 4.5595320027529255e-07, "loss": 0.5836, "step": 575900 }, { "epoch": 99.09669649002065, "grad_norm": 13.484378814697266, "learning_rate": 4.516517549896766e-07, "loss": 0.6724, "step": 575950 }, { "epoch": 99.10529938059187, "grad_norm": 12.666686058044434, "learning_rate": 4.473503097040606e-07, "loss": 0.5697, "step": 576000 }, { "epoch": 99.11390227116311, "grad_norm": 16.177650451660156, "learning_rate": 4.4304886441844465e-07, "loss": 0.6418, "step": 576050 }, { "epoch": 99.12250516173434, "grad_norm": 19.508695602416992, "learning_rate": 4.387474191328286e-07, "loss": 0.6412, "step": 576100 }, { "epoch": 99.13110805230558, "grad_norm": 14.248987197875977, "learning_rate": 4.3444597384721266e-07, "loss": 0.6608, "step": 576150 }, { "epoch": 99.1397109428768, "grad_norm": 17.335437774658203, "learning_rate": 4.301445285615967e-07, "loss": 0.6512, "step": 576200 }, { "epoch": 99.14831383344804, "grad_norm": 14.918718338012695, "learning_rate": 4.258430832759807e-07, "loss": 0.6698, "step": 576250 }, { "epoch": 99.15691672401927, "grad_norm": 24.749237060546875, "learning_rate": 4.2154163799036475e-07, "loss": 0.6503, "step": 576300 }, { "epoch": 99.16551961459051, "grad_norm": 10.805778503417969, "learning_rate": 4.1724019270474883e-07, "loss": 0.6352, "step": 576350 }, { "epoch": 99.17412250516173, "grad_norm": 27.153362274169922, "learning_rate": 4.1293874741913286e-07, "loss": 0.6412, "step": 576400 }, { "epoch": 99.18272539573297, "grad_norm": 18.840566635131836, "learning_rate": 4.086373021335169e-07, "loss": 0.6537, "step": 576450 }, { "epoch": 99.1913282863042, "grad_norm": 27.46942710876465, "learning_rate": 4.0433585684790093e-07, "loss": 0.6383, "step": 576500 }, { "epoch": 99.19993117687542, "grad_norm": 20.2578125, "learning_rate": 4.0003441156228496e-07, "loss": 0.6423, "step": 576550 }, { "epoch": 99.20853406744666, "grad_norm": 16.45766830444336, "learning_rate": 3.95732966276669e-07, "loss": 0.6748, "step": 576600 }, { "epoch": 99.21713695801789, "grad_norm": 14.214873313903809, "learning_rate": 3.9143152099105297e-07, "loss": 0.6355, "step": 576650 }, { "epoch": 99.22573984858913, "grad_norm": 12.47038459777832, "learning_rate": 3.8713007570543705e-07, "loss": 0.6487, "step": 576700 }, { "epoch": 99.23434273916035, "grad_norm": 18.214344024658203, "learning_rate": 3.828286304198211e-07, "loss": 0.6747, "step": 576750 }, { "epoch": 99.2429456297316, "grad_norm": 13.166083335876465, "learning_rate": 3.785271851342051e-07, "loss": 0.6445, "step": 576800 }, { "epoch": 99.25154852030282, "grad_norm": 18.072980880737305, "learning_rate": 3.7422573984858914e-07, "loss": 0.6416, "step": 576850 }, { "epoch": 99.26015141087406, "grad_norm": 23.46883773803711, "learning_rate": 3.699242945629732e-07, "loss": 0.6429, "step": 576900 }, { "epoch": 99.26875430144528, "grad_norm": 17.822538375854492, "learning_rate": 3.656228492773572e-07, "loss": 0.6517, "step": 576950 }, { "epoch": 99.27735719201652, "grad_norm": 15.118852615356445, "learning_rate": 3.6132140399174124e-07, "loss": 0.6186, "step": 577000 }, { "epoch": 99.28596008258775, "grad_norm": 16.585311889648438, "learning_rate": 3.5701995870612527e-07, "loss": 0.623, "step": 577050 }, { "epoch": 99.29456297315897, "grad_norm": 21.726423263549805, "learning_rate": 3.527185134205093e-07, "loss": 0.6887, "step": 577100 }, { "epoch": 99.30316586373021, "grad_norm": 13.763529777526855, "learning_rate": 3.484170681348934e-07, "loss": 0.5876, "step": 577150 }, { "epoch": 99.31176875430144, "grad_norm": 12.460923194885254, "learning_rate": 3.4411562284927736e-07, "loss": 0.633, "step": 577200 }, { "epoch": 99.32037164487268, "grad_norm": 13.614339828491211, "learning_rate": 3.398141775636614e-07, "loss": 0.6161, "step": 577250 }, { "epoch": 99.3289745354439, "grad_norm": 15.412359237670898, "learning_rate": 3.355127322780454e-07, "loss": 0.6228, "step": 577300 }, { "epoch": 99.33757742601514, "grad_norm": 16.82701873779297, "learning_rate": 3.3121128699242945e-07, "loss": 0.6072, "step": 577350 }, { "epoch": 99.34618031658637, "grad_norm": 10.047409057617188, "learning_rate": 3.2690984170681354e-07, "loss": 0.5685, "step": 577400 }, { "epoch": 99.35478320715761, "grad_norm": 22.484268188476562, "learning_rate": 3.2260839642119757e-07, "loss": 0.6951, "step": 577450 }, { "epoch": 99.36338609772884, "grad_norm": 22.99928092956543, "learning_rate": 3.1830695113558155e-07, "loss": 0.6491, "step": 577500 }, { "epoch": 99.37198898830007, "grad_norm": 12.289894104003906, "learning_rate": 3.140055058499656e-07, "loss": 0.6979, "step": 577550 }, { "epoch": 99.3805918788713, "grad_norm": 21.53241729736328, "learning_rate": 3.0970406056434966e-07, "loss": 0.6632, "step": 577600 }, { "epoch": 99.38919476944254, "grad_norm": 37.986270904541016, "learning_rate": 3.054026152787337e-07, "loss": 0.6797, "step": 577650 }, { "epoch": 99.39779766001377, "grad_norm": 23.381155014038086, "learning_rate": 3.011011699931177e-07, "loss": 0.6383, "step": 577700 }, { "epoch": 99.40640055058499, "grad_norm": 23.128965377807617, "learning_rate": 2.967997247075017e-07, "loss": 0.6153, "step": 577750 }, { "epoch": 99.41500344115623, "grad_norm": 11.650248527526855, "learning_rate": 2.9249827942188573e-07, "loss": 0.6226, "step": 577800 }, { "epoch": 99.42360633172746, "grad_norm": 7.3599324226379395, "learning_rate": 2.881968341362698e-07, "loss": 0.5781, "step": 577850 }, { "epoch": 99.4322092222987, "grad_norm": 15.004666328430176, "learning_rate": 2.8389538885065385e-07, "loss": 0.6376, "step": 577900 }, { "epoch": 99.44081211286992, "grad_norm": 13.30484390258789, "learning_rate": 2.795939435650379e-07, "loss": 0.6608, "step": 577950 }, { "epoch": 99.44941500344116, "grad_norm": 10.512864112854004, "learning_rate": 2.752924982794219e-07, "loss": 0.6096, "step": 578000 }, { "epoch": 99.45801789401239, "grad_norm": 10.017942428588867, "learning_rate": 2.709910529938059e-07, "loss": 0.6073, "step": 578050 }, { "epoch": 99.46662078458363, "grad_norm": 13.629181861877441, "learning_rate": 2.6668960770818997e-07, "loss": 0.6356, "step": 578100 }, { "epoch": 99.47522367515485, "grad_norm": 11.909382820129395, "learning_rate": 2.62388162422574e-07, "loss": 0.6855, "step": 578150 }, { "epoch": 99.48382656572609, "grad_norm": 25.590389251708984, "learning_rate": 2.5808671713695803e-07, "loss": 0.6446, "step": 578200 }, { "epoch": 99.49242945629732, "grad_norm": 12.031659126281738, "learning_rate": 2.5378527185134206e-07, "loss": 0.6442, "step": 578250 }, { "epoch": 99.50103234686854, "grad_norm": 14.19442081451416, "learning_rate": 2.494838265657261e-07, "loss": 0.6439, "step": 578300 }, { "epoch": 99.50963523743978, "grad_norm": 10.8143310546875, "learning_rate": 2.4518238128011013e-07, "loss": 0.6061, "step": 578350 }, { "epoch": 99.518238128011, "grad_norm": 10.05512523651123, "learning_rate": 2.4088093599449416e-07, "loss": 0.6271, "step": 578400 }, { "epoch": 99.52684101858225, "grad_norm": 20.9061279296875, "learning_rate": 2.365794907088782e-07, "loss": 0.6257, "step": 578450 }, { "epoch": 99.53544390915347, "grad_norm": 12.349711418151855, "learning_rate": 2.3227804542326222e-07, "loss": 0.6161, "step": 578500 }, { "epoch": 99.54404679972471, "grad_norm": 15.096070289611816, "learning_rate": 2.2797660013764628e-07, "loss": 0.6525, "step": 578550 }, { "epoch": 99.55264969029594, "grad_norm": 17.069271087646484, "learning_rate": 2.236751548520303e-07, "loss": 0.63, "step": 578600 }, { "epoch": 99.56125258086718, "grad_norm": 18.378677368164062, "learning_rate": 2.193737095664143e-07, "loss": 0.6066, "step": 578650 }, { "epoch": 99.5698554714384, "grad_norm": 11.349108695983887, "learning_rate": 2.1507226428079834e-07, "loss": 0.6402, "step": 578700 }, { "epoch": 99.57845836200964, "grad_norm": 18.446048736572266, "learning_rate": 2.1077081899518237e-07, "loss": 0.6562, "step": 578750 }, { "epoch": 99.58706125258087, "grad_norm": 9.431604385375977, "learning_rate": 2.0646937370956643e-07, "loss": 0.6406, "step": 578800 }, { "epoch": 99.59566414315209, "grad_norm": 17.49921417236328, "learning_rate": 2.0216792842395046e-07, "loss": 0.6769, "step": 578850 }, { "epoch": 99.60426703372333, "grad_norm": 28.0980224609375, "learning_rate": 1.978664831383345e-07, "loss": 0.6455, "step": 578900 }, { "epoch": 99.61286992429456, "grad_norm": 14.163009643554688, "learning_rate": 1.9356503785271853e-07, "loss": 0.6384, "step": 578950 }, { "epoch": 99.6214728148658, "grad_norm": 15.026597023010254, "learning_rate": 1.8926359256710256e-07, "loss": 0.6113, "step": 579000 }, { "epoch": 99.63007570543702, "grad_norm": 22.417470932006836, "learning_rate": 1.849621472814866e-07, "loss": 0.6263, "step": 579050 }, { "epoch": 99.63867859600826, "grad_norm": 10.57209300994873, "learning_rate": 1.8066070199587062e-07, "loss": 0.65, "step": 579100 }, { "epoch": 99.64728148657949, "grad_norm": 16.51609992980957, "learning_rate": 1.7635925671025465e-07, "loss": 0.6591, "step": 579150 }, { "epoch": 99.65588437715073, "grad_norm": 14.146124839782715, "learning_rate": 1.7205781142463868e-07, "loss": 0.6655, "step": 579200 }, { "epoch": 99.66448726772195, "grad_norm": 17.13487434387207, "learning_rate": 1.677563661390227e-07, "loss": 0.6006, "step": 579250 }, { "epoch": 99.67309015829319, "grad_norm": 12.445178031921387, "learning_rate": 1.6345492085340677e-07, "loss": 0.6578, "step": 579300 }, { "epoch": 99.68169304886442, "grad_norm": 11.269501686096191, "learning_rate": 1.5915347556779077e-07, "loss": 0.6148, "step": 579350 }, { "epoch": 99.69029593943566, "grad_norm": 14.284321784973145, "learning_rate": 1.5485203028217483e-07, "loss": 0.6419, "step": 579400 }, { "epoch": 99.69889883000688, "grad_norm": 16.655933380126953, "learning_rate": 1.5055058499655886e-07, "loss": 0.5976, "step": 579450 }, { "epoch": 99.70750172057811, "grad_norm": 13.524371147155762, "learning_rate": 1.4624913971094287e-07, "loss": 0.5706, "step": 579500 }, { "epoch": 99.71610461114935, "grad_norm": 11.970343589782715, "learning_rate": 1.4194769442532692e-07, "loss": 0.6502, "step": 579550 }, { "epoch": 99.72470750172057, "grad_norm": 13.95887279510498, "learning_rate": 1.3764624913971095e-07, "loss": 0.6422, "step": 579600 }, { "epoch": 99.73331039229181, "grad_norm": 18.95633888244629, "learning_rate": 1.3334480385409499e-07, "loss": 0.641, "step": 579650 }, { "epoch": 99.74191328286304, "grad_norm": 12.565098762512207, "learning_rate": 1.2904335856847902e-07, "loss": 0.5999, "step": 579700 }, { "epoch": 99.75051617343428, "grad_norm": 20.62059783935547, "learning_rate": 1.2474191328286305e-07, "loss": 0.6723, "step": 579750 }, { "epoch": 99.7591190640055, "grad_norm": 22.85693359375, "learning_rate": 1.2044046799724708e-07, "loss": 0.6789, "step": 579800 }, { "epoch": 99.76772195457674, "grad_norm": 12.018401145935059, "learning_rate": 1.1613902271163111e-07, "loss": 0.6197, "step": 579850 }, { "epoch": 99.77632484514797, "grad_norm": 12.805892944335938, "learning_rate": 1.1183757742601515e-07, "loss": 0.6372, "step": 579900 }, { "epoch": 99.78492773571921, "grad_norm": 16.218595504760742, "learning_rate": 1.0753613214039917e-07, "loss": 0.6427, "step": 579950 }, { "epoch": 99.79353062629043, "grad_norm": 14.787798881530762, "learning_rate": 1.0323468685478322e-07, "loss": 0.5878, "step": 580000 }, { "epoch": 99.80213351686166, "grad_norm": 11.281668663024902, "learning_rate": 9.893324156916725e-08, "loss": 0.6546, "step": 580050 }, { "epoch": 99.8107364074329, "grad_norm": 11.750572204589844, "learning_rate": 9.463179628355128e-08, "loss": 0.6571, "step": 580100 }, { "epoch": 99.81933929800412, "grad_norm": 11.950610160827637, "learning_rate": 9.033035099793531e-08, "loss": 0.6313, "step": 580150 }, { "epoch": 99.82794218857536, "grad_norm": 18.463136672973633, "learning_rate": 8.602890571231934e-08, "loss": 0.6598, "step": 580200 }, { "epoch": 99.83654507914659, "grad_norm": 12.253210067749023, "learning_rate": 8.172746042670338e-08, "loss": 0.6406, "step": 580250 }, { "epoch": 99.84514796971783, "grad_norm": 15.614834785461426, "learning_rate": 7.742601514108742e-08, "loss": 0.6214, "step": 580300 }, { "epoch": 99.85375086028905, "grad_norm": 14.795942306518555, "learning_rate": 7.312456985547143e-08, "loss": 0.5975, "step": 580350 }, { "epoch": 99.8623537508603, "grad_norm": 27.562490463256836, "learning_rate": 6.882312456985548e-08, "loss": 0.6352, "step": 580400 }, { "epoch": 99.87095664143152, "grad_norm": 13.931257247924805, "learning_rate": 6.452167928423951e-08, "loss": 0.6128, "step": 580450 }, { "epoch": 99.87955953200276, "grad_norm": 10.359968185424805, "learning_rate": 6.022023399862354e-08, "loss": 0.6178, "step": 580500 }, { "epoch": 99.88816242257398, "grad_norm": 21.26670265197754, "learning_rate": 5.591878871300758e-08, "loss": 0.6132, "step": 580550 }, { "epoch": 99.89676531314522, "grad_norm": 20.848350524902344, "learning_rate": 5.161734342739161e-08, "loss": 0.6332, "step": 580600 }, { "epoch": 99.90536820371645, "grad_norm": 29.012853622436523, "learning_rate": 4.731589814177564e-08, "loss": 0.628, "step": 580650 }, { "epoch": 99.91397109428767, "grad_norm": 13.92016315460205, "learning_rate": 4.301445285615967e-08, "loss": 0.638, "step": 580700 }, { "epoch": 99.92257398485891, "grad_norm": 16.870155334472656, "learning_rate": 3.871300757054371e-08, "loss": 0.6399, "step": 580750 }, { "epoch": 99.93117687543014, "grad_norm": 18.60354232788086, "learning_rate": 3.441156228492774e-08, "loss": 0.6558, "step": 580800 }, { "epoch": 99.93977976600138, "grad_norm": 19.31717300415039, "learning_rate": 3.011011699931177e-08, "loss": 0.6262, "step": 580850 }, { "epoch": 99.9483826565726, "grad_norm": 16.700529098510742, "learning_rate": 2.5808671713695804e-08, "loss": 0.5896, "step": 580900 }, { "epoch": 99.95698554714384, "grad_norm": 25.59731101989746, "learning_rate": 2.1507226428079835e-08, "loss": 0.6134, "step": 580950 }, { "epoch": 99.96558843771507, "grad_norm": 11.31701374053955, "learning_rate": 1.720578114246387e-08, "loss": 0.6714, "step": 581000 }, { "epoch": 99.97419132828631, "grad_norm": 12.44153118133545, "learning_rate": 1.2904335856847902e-08, "loss": 0.6156, "step": 581050 }, { "epoch": 99.98279421885753, "grad_norm": 8.95052719116211, "learning_rate": 8.602890571231935e-09, "loss": 0.6286, "step": 581100 }, { "epoch": 99.99139710942877, "grad_norm": 18.24790382385254, "learning_rate": 4.301445285615967e-09, "loss": 0.6107, "step": 581150 }, { "epoch": 100.0, "grad_norm": 41.09204864501953, "learning_rate": 0.0, "loss": 0.6384, "step": 581200 }, { "epoch": 100.0, "eval_accuracy": 0.6183961452417829, "eval_f1": 0.6203678815285474, "eval_f1_DuraRiadoRio_16x16": 0.5736224028906955, "eval_f1_Mole_16x16": 0.6530239356842682, "eval_f1_Quebrado_16x16": 0.710446247464503, "eval_f1_RiadoRio_16x16": 0.5321252059308073, "eval_f1_RioFechado_16x16": 0.6326216156724631, "eval_loss": 1.8485805988311768, "eval_precision": 0.6367074800036849, "eval_precision_DuraRiadoRio_16x16": 0.5979284369114878, "eval_precision_Mole_16x16": 0.5639002840012622, "eval_precision_Quebrado_16x16": 0.8542682926829268, "eval_precision_RiadoRio_16x16": 0.533003300330033, "eval_precision_RioFechado_16x16": 0.6344370860927152, "eval_recall": 0.6193924678080187, "eval_recall_DuraRiadoRio_16x16": 0.5512152777777778, "eval_recall_Mole_16x16": 0.7756076388888888, "eval_recall_Quebrado_16x16": 0.6080729166666666, "eval_recall_RiadoRio_16x16": 0.53125, "eval_recall_RioFechado_16x16": 0.6308165057067603, "eval_runtime": 46.2688, "eval_samples_per_second": 251.184, "eval_steps_per_second": 15.713, "step": 581200 }, { "epoch": 100.0, "step": 581200, "total_flos": 2.446223378542756e+19, "train_loss": 0.8178866318153334, "train_runtime": 92994.6535, "train_samples_per_second": 99.983, "train_steps_per_second": 6.25 } ], "logging_steps": 50, "max_steps": 581200, "num_input_tokens_seen": 0, "num_train_epochs": 100, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.446223378542756e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }