diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4299 @@ +{ + "best_metric": 0.540250366102693, + "best_model_checkpoint": "/data0/checkpoints/Qwen2.5-Math-7B-ScalePRM-v3.0/checkpoint-600", + "epoch": 0.757934628138323, + "eval_steps": 100, + "global_step": 600, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0012632243802305385, + "grad_norm": 2.429877281188965, + "learning_rate": 7e-06, + "loss": 1.1465, + "step": 1 + }, + { + "epoch": 0.002526448760461077, + "grad_norm": 2.2121567726135254, + "learning_rate": 7e-06, + "loss": 1.0451, + "step": 2 + }, + { + "epoch": 0.0037896731406916154, + "grad_norm": 1.3855836391448975, + "learning_rate": 7e-06, + "loss": 0.735, + "step": 3 + }, + { + "epoch": 0.005052897520922154, + "grad_norm": 0.3500981330871582, + "learning_rate": 7e-06, + "loss": 0.5603, + "step": 4 + }, + { + "epoch": 0.0063161219011526925, + "grad_norm": 0.40845438838005066, + "learning_rate": 7e-06, + "loss": 1.0627, + "step": 5 + }, + { + "epoch": 0.007579346281383231, + "grad_norm": 0.9035907983779907, + "learning_rate": 7e-06, + "loss": 5.0569, + "step": 6 + }, + { + "epoch": 0.00884257066161377, + "grad_norm": 0.6223624348640442, + "learning_rate": 7e-06, + "loss": 9.4706, + "step": 7 + }, + { + "epoch": 0.010105795041844308, + "grad_norm": 0.5229220390319824, + "learning_rate": 7e-06, + "loss": 5.6685, + "step": 8 + }, + { + "epoch": 0.011369019422074847, + "grad_norm": 0.5141741633415222, + "learning_rate": 7e-06, + "loss": 5.8831, + "step": 9 + }, + { + "epoch": 0.012632243802305385, + "grad_norm": 0.6420879364013672, + "learning_rate": 7e-06, + "loss": 4.6915, + "step": 10 + }, + { + "epoch": 0.013895468182535923, + "grad_norm": 0.47964179515838623, + "learning_rate": 7e-06, + "loss": 3.9531, + "step": 11 + }, + { + "epoch": 0.015158692562766462, + "grad_norm": 0.40398040413856506, + "learning_rate": 7e-06, + "loss": 1.9501, + "step": 12 + }, + { + "epoch": 0.016421916942996998, + "grad_norm": 0.5141711235046387, + "learning_rate": 7e-06, + "loss": 1.9378, + "step": 13 + }, + { + "epoch": 0.01768514132322754, + "grad_norm": 0.24602794647216797, + "learning_rate": 7e-06, + "loss": 1.535, + "step": 14 + }, + { + "epoch": 0.018948365703458078, + "grad_norm": 0.21208855509757996, + "learning_rate": 7e-06, + "loss": 1.5177, + "step": 15 + }, + { + "epoch": 0.020211590083688617, + "grad_norm": 0.6067216992378235, + "learning_rate": 7e-06, + "loss": 0.9104, + "step": 16 + }, + { + "epoch": 0.021474814463919155, + "grad_norm": 0.422442227602005, + "learning_rate": 7e-06, + "loss": 0.9266, + "step": 17 + }, + { + "epoch": 0.022738038844149693, + "grad_norm": 0.659572958946228, + "learning_rate": 7e-06, + "loss": 0.7499, + "step": 18 + }, + { + "epoch": 0.02400126322438023, + "grad_norm": 0.4817348122596741, + "learning_rate": 7e-06, + "loss": 0.7436, + "step": 19 + }, + { + "epoch": 0.02526448760461077, + "grad_norm": 0.20682591199874878, + "learning_rate": 7e-06, + "loss": 0.4786, + "step": 20 + }, + { + "epoch": 0.026527711984841308, + "grad_norm": 0.2613360583782196, + "learning_rate": 7e-06, + "loss": 0.8464, + "step": 21 + }, + { + "epoch": 0.027790936365071846, + "grad_norm": 0.2720305621623993, + "learning_rate": 7e-06, + "loss": 0.7255, + "step": 22 + }, + { + "epoch": 0.029054160745302385, + "grad_norm": 0.25043392181396484, + "learning_rate": 7e-06, + "loss": 0.9661, + "step": 23 + }, + { + "epoch": 0.030317385125532923, + "grad_norm": 0.2801963686943054, + "learning_rate": 7e-06, + "loss": 0.9602, + "step": 24 + }, + { + "epoch": 0.03158060950576346, + "grad_norm": 0.2137051522731781, + "learning_rate": 7e-06, + "loss": 0.8132, + "step": 25 + }, + { + "epoch": 0.032843833885993996, + "grad_norm": 0.13553065061569214, + "learning_rate": 7e-06, + "loss": 0.4873, + "step": 26 + }, + { + "epoch": 0.03410705826622454, + "grad_norm": 0.1350618302822113, + "learning_rate": 7e-06, + "loss": 0.5186, + "step": 27 + }, + { + "epoch": 0.03537028264645508, + "grad_norm": 0.1236298605799675, + "learning_rate": 7e-06, + "loss": 0.3401, + "step": 28 + }, + { + "epoch": 0.036633507026685615, + "grad_norm": 0.15515856444835663, + "learning_rate": 7e-06, + "loss": 1.2493, + "step": 29 + }, + { + "epoch": 0.037896731406916156, + "grad_norm": 0.09012973308563232, + "learning_rate": 7e-06, + "loss": 0.7651, + "step": 30 + }, + { + "epoch": 0.03915995578714669, + "grad_norm": 0.14378102123737335, + "learning_rate": 7e-06, + "loss": 0.4005, + "step": 31 + }, + { + "epoch": 0.04042318016737723, + "grad_norm": 0.2546883523464203, + "learning_rate": 7e-06, + "loss": 0.5304, + "step": 32 + }, + { + "epoch": 0.04168640454760777, + "grad_norm": 0.1023496687412262, + "learning_rate": 7e-06, + "loss": 0.493, + "step": 33 + }, + { + "epoch": 0.04294962892783831, + "grad_norm": 0.1719491183757782, + "learning_rate": 7e-06, + "loss": 0.3707, + "step": 34 + }, + { + "epoch": 0.044212853308068845, + "grad_norm": 0.08337250351905823, + "learning_rate": 7e-06, + "loss": 0.3143, + "step": 35 + }, + { + "epoch": 0.045476077688299386, + "grad_norm": 0.09040359407663345, + "learning_rate": 7e-06, + "loss": 0.3942, + "step": 36 + }, + { + "epoch": 0.04673930206852992, + "grad_norm": 0.10850965231657028, + "learning_rate": 7e-06, + "loss": 0.4453, + "step": 37 + }, + { + "epoch": 0.04800252644876046, + "grad_norm": 0.08887636661529541, + "learning_rate": 7e-06, + "loss": 1.2015, + "step": 38 + }, + { + "epoch": 0.049265750828991, + "grad_norm": 0.1864442229270935, + "learning_rate": 7e-06, + "loss": 0.3895, + "step": 39 + }, + { + "epoch": 0.05052897520922154, + "grad_norm": 0.175123393535614, + "learning_rate": 7e-06, + "loss": 0.3626, + "step": 40 + }, + { + "epoch": 0.051792199589452075, + "grad_norm": 0.10572918504476547, + "learning_rate": 7e-06, + "loss": 0.3335, + "step": 41 + }, + { + "epoch": 0.053055423969682616, + "grad_norm": 0.09624486416578293, + "learning_rate": 7e-06, + "loss": 0.3065, + "step": 42 + }, + { + "epoch": 0.05431864834991315, + "grad_norm": 0.13604743778705597, + "learning_rate": 7e-06, + "loss": 0.317, + "step": 43 + }, + { + "epoch": 0.05558187273014369, + "grad_norm": 0.15408551692962646, + "learning_rate": 7e-06, + "loss": 0.7709, + "step": 44 + }, + { + "epoch": 0.05684509711037423, + "grad_norm": 0.09676961600780487, + "learning_rate": 7e-06, + "loss": 0.5114, + "step": 45 + }, + { + "epoch": 0.05810832149060477, + "grad_norm": 0.11936207115650177, + "learning_rate": 7e-06, + "loss": 0.2785, + "step": 46 + }, + { + "epoch": 0.059371545870835304, + "grad_norm": 0.1744876503944397, + "learning_rate": 7e-06, + "loss": 0.2689, + "step": 47 + }, + { + "epoch": 0.060634770251065846, + "grad_norm": 0.17397810518741608, + "learning_rate": 7e-06, + "loss": 0.3316, + "step": 48 + }, + { + "epoch": 0.06189799463129638, + "grad_norm": 0.1329212635755539, + "learning_rate": 7e-06, + "loss": 0.2853, + "step": 49 + }, + { + "epoch": 0.06316121901152692, + "grad_norm": 0.09555013477802277, + "learning_rate": 7e-06, + "loss": 0.3182, + "step": 50 + }, + { + "epoch": 0.06442444339175746, + "grad_norm": 0.15529152750968933, + "learning_rate": 7e-06, + "loss": 0.5497, + "step": 51 + }, + { + "epoch": 0.06568766777198799, + "grad_norm": 0.09599810838699341, + "learning_rate": 7e-06, + "loss": 0.3102, + "step": 52 + }, + { + "epoch": 0.06695089215221854, + "grad_norm": 0.12325876951217651, + "learning_rate": 7e-06, + "loss": 0.2788, + "step": 53 + }, + { + "epoch": 0.06821411653244908, + "grad_norm": 0.2820286154747009, + "learning_rate": 7e-06, + "loss": 0.3934, + "step": 54 + }, + { + "epoch": 0.06947734091267961, + "grad_norm": 0.17912541329860687, + "learning_rate": 7e-06, + "loss": 0.3709, + "step": 55 + }, + { + "epoch": 0.07074056529291016, + "grad_norm": 0.14083553850650787, + "learning_rate": 7e-06, + "loss": 0.4105, + "step": 56 + }, + { + "epoch": 0.0720037896731407, + "grad_norm": 0.09743569046258926, + "learning_rate": 7e-06, + "loss": 0.2376, + "step": 57 + }, + { + "epoch": 0.07326701405337123, + "grad_norm": 0.10704771429300308, + "learning_rate": 7e-06, + "loss": 0.2714, + "step": 58 + }, + { + "epoch": 0.07453023843360176, + "grad_norm": 0.11463718861341476, + "learning_rate": 7e-06, + "loss": 0.456, + "step": 59 + }, + { + "epoch": 0.07579346281383231, + "grad_norm": 0.12085901200771332, + "learning_rate": 7e-06, + "loss": 0.3099, + "step": 60 + }, + { + "epoch": 0.07705668719406285, + "grad_norm": 0.10744248330593109, + "learning_rate": 7e-06, + "loss": 0.2312, + "step": 61 + }, + { + "epoch": 0.07831991157429338, + "grad_norm": 0.08374691009521484, + "learning_rate": 7e-06, + "loss": 0.2685, + "step": 62 + }, + { + "epoch": 0.07958313595452392, + "grad_norm": 0.10826320946216583, + "learning_rate": 7e-06, + "loss": 0.3069, + "step": 63 + }, + { + "epoch": 0.08084636033475447, + "grad_norm": 0.17864489555358887, + "learning_rate": 7e-06, + "loss": 0.5024, + "step": 64 + }, + { + "epoch": 0.082109584714985, + "grad_norm": 0.11988472938537598, + "learning_rate": 7e-06, + "loss": 0.3164, + "step": 65 + }, + { + "epoch": 0.08337280909521554, + "grad_norm": 0.1612488180398941, + "learning_rate": 7e-06, + "loss": 0.3295, + "step": 66 + }, + { + "epoch": 0.08463603347544607, + "grad_norm": 0.13754408061504364, + "learning_rate": 7e-06, + "loss": 0.2406, + "step": 67 + }, + { + "epoch": 0.08589925785567662, + "grad_norm": 0.11351214349269867, + "learning_rate": 7e-06, + "loss": 0.3149, + "step": 68 + }, + { + "epoch": 0.08716248223590715, + "grad_norm": 0.07585523277521133, + "learning_rate": 7e-06, + "loss": 0.2706, + "step": 69 + }, + { + "epoch": 0.08842570661613769, + "grad_norm": 0.0744984969496727, + "learning_rate": 7e-06, + "loss": 0.214, + "step": 70 + }, + { + "epoch": 0.08968893099636822, + "grad_norm": 0.2244742512702942, + "learning_rate": 7e-06, + "loss": 0.4242, + "step": 71 + }, + { + "epoch": 0.09095215537659877, + "grad_norm": 0.08662209659814835, + "learning_rate": 7e-06, + "loss": 0.2691, + "step": 72 + }, + { + "epoch": 0.09221537975682931, + "grad_norm": 0.10564761608839035, + "learning_rate": 7e-06, + "loss": 0.3228, + "step": 73 + }, + { + "epoch": 0.09347860413705984, + "grad_norm": 0.18067984282970428, + "learning_rate": 7e-06, + "loss": 0.3384, + "step": 74 + }, + { + "epoch": 0.09474182851729038, + "grad_norm": 0.07762212306261063, + "learning_rate": 7e-06, + "loss": 0.2377, + "step": 75 + }, + { + "epoch": 0.09600505289752093, + "grad_norm": 0.07793518900871277, + "learning_rate": 7e-06, + "loss": 0.248, + "step": 76 + }, + { + "epoch": 0.09726827727775146, + "grad_norm": 0.1307854801416397, + "learning_rate": 7e-06, + "loss": 0.2667, + "step": 77 + }, + { + "epoch": 0.098531501657982, + "grad_norm": 0.09771443158388138, + "learning_rate": 7e-06, + "loss": 0.318, + "step": 78 + }, + { + "epoch": 0.09979472603821253, + "grad_norm": 0.10437527298927307, + "learning_rate": 7e-06, + "loss": 0.3303, + "step": 79 + }, + { + "epoch": 0.10105795041844308, + "grad_norm": 0.11160580813884735, + "learning_rate": 7e-06, + "loss": 0.2845, + "step": 80 + }, + { + "epoch": 0.10232117479867361, + "grad_norm": 0.0809980109333992, + "learning_rate": 7e-06, + "loss": 0.2736, + "step": 81 + }, + { + "epoch": 0.10358439917890415, + "grad_norm": 0.10574865341186523, + "learning_rate": 7e-06, + "loss": 0.3012, + "step": 82 + }, + { + "epoch": 0.1048476235591347, + "grad_norm": 0.07807318866252899, + "learning_rate": 7e-06, + "loss": 0.2283, + "step": 83 + }, + { + "epoch": 0.10611084793936523, + "grad_norm": 0.10281991213560104, + "learning_rate": 7e-06, + "loss": 0.2544, + "step": 84 + }, + { + "epoch": 0.10737407231959577, + "grad_norm": 0.12749870121479034, + "learning_rate": 7e-06, + "loss": 0.2973, + "step": 85 + }, + { + "epoch": 0.1086372966998263, + "grad_norm": 0.13138003647327423, + "learning_rate": 7e-06, + "loss": 0.3399, + "step": 86 + }, + { + "epoch": 0.10990052108005685, + "grad_norm": 0.10815514624118805, + "learning_rate": 7e-06, + "loss": 0.3221, + "step": 87 + }, + { + "epoch": 0.11116374546028739, + "grad_norm": 0.13537508249282837, + "learning_rate": 7e-06, + "loss": 0.308, + "step": 88 + }, + { + "epoch": 0.11242696984051792, + "grad_norm": 0.09689060598611832, + "learning_rate": 7e-06, + "loss": 0.2511, + "step": 89 + }, + { + "epoch": 0.11369019422074846, + "grad_norm": 0.08782925456762314, + "learning_rate": 7e-06, + "loss": 0.2936, + "step": 90 + }, + { + "epoch": 0.114953418600979, + "grad_norm": 0.12655287981033325, + "learning_rate": 7e-06, + "loss": 0.4158, + "step": 91 + }, + { + "epoch": 0.11621664298120954, + "grad_norm": 0.11866717785596848, + "learning_rate": 7e-06, + "loss": 0.4059, + "step": 92 + }, + { + "epoch": 0.11747986736144007, + "grad_norm": 0.12691305577754974, + "learning_rate": 7e-06, + "loss": 0.2453, + "step": 93 + }, + { + "epoch": 0.11874309174167061, + "grad_norm": 0.11844722181558609, + "learning_rate": 7e-06, + "loss": 0.2249, + "step": 94 + }, + { + "epoch": 0.12000631612190116, + "grad_norm": 0.07606595754623413, + "learning_rate": 7e-06, + "loss": 0.2789, + "step": 95 + }, + { + "epoch": 0.12126954050213169, + "grad_norm": 0.11529266834259033, + "learning_rate": 7e-06, + "loss": 0.2654, + "step": 96 + }, + { + "epoch": 0.12253276488236223, + "grad_norm": 0.12648285925388336, + "learning_rate": 7e-06, + "loss": 0.2279, + "step": 97 + }, + { + "epoch": 0.12379598926259276, + "grad_norm": 0.1504458636045456, + "learning_rate": 7e-06, + "loss": 0.4048, + "step": 98 + }, + { + "epoch": 0.1250592136428233, + "grad_norm": 0.09578829258680344, + "learning_rate": 7e-06, + "loss": 0.2912, + "step": 99 + }, + { + "epoch": 0.12632243802305385, + "grad_norm": 0.10936733335256577, + "learning_rate": 7e-06, + "loss": 0.2644, + "step": 100 + }, + { + "epoch": 0.12632243802305385, + "eval_correct_accuracy": 0.5708227311280747, + "eval_error_accuracy": 0.45610085547050877, + "eval_f1": 0.5070537660000148, + "eval_loss": 0.43133699893951416, + "eval_runtime": 35.3366, + "eval_samples_per_second": 96.217, + "eval_steps_per_second": 6.028, + "step": 100 + }, + { + "epoch": 0.1275856624032844, + "grad_norm": 0.1165054589509964, + "learning_rate": 7e-06, + "loss": 0.2693, + "step": 101 + }, + { + "epoch": 0.12884888678351492, + "grad_norm": 0.08343573659658432, + "learning_rate": 7e-06, + "loss": 0.2388, + "step": 102 + }, + { + "epoch": 0.13011211116374546, + "grad_norm": 0.10629656910896301, + "learning_rate": 7e-06, + "loss": 0.2603, + "step": 103 + }, + { + "epoch": 0.13137533554397599, + "grad_norm": 0.07509850710630417, + "learning_rate": 7e-06, + "loss": 0.253, + "step": 104 + }, + { + "epoch": 0.13263855992420653, + "grad_norm": 0.08039335906505585, + "learning_rate": 7e-06, + "loss": 0.224, + "step": 105 + }, + { + "epoch": 0.13390178430443708, + "grad_norm": 0.10666981339454651, + "learning_rate": 7e-06, + "loss": 0.3945, + "step": 106 + }, + { + "epoch": 0.1351650086846676, + "grad_norm": 0.16490086913108826, + "learning_rate": 7e-06, + "loss": 0.3087, + "step": 107 + }, + { + "epoch": 0.13642823306489815, + "grad_norm": 0.09013114124536514, + "learning_rate": 7e-06, + "loss": 0.3355, + "step": 108 + }, + { + "epoch": 0.1376914574451287, + "grad_norm": 0.1580226719379425, + "learning_rate": 7e-06, + "loss": 0.2433, + "step": 109 + }, + { + "epoch": 0.13895468182535922, + "grad_norm": 0.09130299836397171, + "learning_rate": 7e-06, + "loss": 0.1928, + "step": 110 + }, + { + "epoch": 0.14021790620558977, + "grad_norm": 0.07702811807394028, + "learning_rate": 7e-06, + "loss": 0.2319, + "step": 111 + }, + { + "epoch": 0.14148113058582032, + "grad_norm": 0.14257381856441498, + "learning_rate": 7e-06, + "loss": 0.2496, + "step": 112 + }, + { + "epoch": 0.14274435496605084, + "grad_norm": 0.11546823382377625, + "learning_rate": 7e-06, + "loss": 0.2592, + "step": 113 + }, + { + "epoch": 0.1440075793462814, + "grad_norm": 0.12595829367637634, + "learning_rate": 7e-06, + "loss": 0.2539, + "step": 114 + }, + { + "epoch": 0.1452708037265119, + "grad_norm": 0.10172153264284134, + "learning_rate": 7e-06, + "loss": 0.2728, + "step": 115 + }, + { + "epoch": 0.14653402810674246, + "grad_norm": 0.10145121812820435, + "learning_rate": 7e-06, + "loss": 0.163, + "step": 116 + }, + { + "epoch": 0.147797252486973, + "grad_norm": 0.15631917119026184, + "learning_rate": 7e-06, + "loss": 0.2526, + "step": 117 + }, + { + "epoch": 0.14906047686720353, + "grad_norm": 0.13442394137382507, + "learning_rate": 7e-06, + "loss": 0.2591, + "step": 118 + }, + { + "epoch": 0.15032370124743408, + "grad_norm": 0.08642445504665375, + "learning_rate": 7e-06, + "loss": 0.2505, + "step": 119 + }, + { + "epoch": 0.15158692562766463, + "grad_norm": 0.13054709136486053, + "learning_rate": 7e-06, + "loss": 0.2704, + "step": 120 + }, + { + "epoch": 0.15285015000789515, + "grad_norm": 0.19653519988059998, + "learning_rate": 7e-06, + "loss": 0.2384, + "step": 121 + }, + { + "epoch": 0.1541133743881257, + "grad_norm": 0.20973946154117584, + "learning_rate": 7e-06, + "loss": 0.2385, + "step": 122 + }, + { + "epoch": 0.15537659876835622, + "grad_norm": 0.096860371530056, + "learning_rate": 7e-06, + "loss": 0.241, + "step": 123 + }, + { + "epoch": 0.15663982314858677, + "grad_norm": 0.10356521606445312, + "learning_rate": 7e-06, + "loss": 0.4737, + "step": 124 + }, + { + "epoch": 0.15790304752881731, + "grad_norm": 0.17340725660324097, + "learning_rate": 7e-06, + "loss": 0.2673, + "step": 125 + }, + { + "epoch": 0.15916627190904784, + "grad_norm": 0.3000679612159729, + "learning_rate": 7e-06, + "loss": 0.431, + "step": 126 + }, + { + "epoch": 0.16042949628927838, + "grad_norm": 0.11215244233608246, + "learning_rate": 7e-06, + "loss": 0.206, + "step": 127 + }, + { + "epoch": 0.16169272066950893, + "grad_norm": 0.07078877836465836, + "learning_rate": 7e-06, + "loss": 0.201, + "step": 128 + }, + { + "epoch": 0.16295594504973945, + "grad_norm": 0.16037459671497345, + "learning_rate": 7e-06, + "loss": 0.268, + "step": 129 + }, + { + "epoch": 0.16421916942997, + "grad_norm": 0.20243118703365326, + "learning_rate": 7e-06, + "loss": 0.2503, + "step": 130 + }, + { + "epoch": 0.16548239381020052, + "grad_norm": 0.1389663964509964, + "learning_rate": 7e-06, + "loss": 0.2517, + "step": 131 + }, + { + "epoch": 0.16674561819043107, + "grad_norm": 0.12263572961091995, + "learning_rate": 7e-06, + "loss": 0.2359, + "step": 132 + }, + { + "epoch": 0.16800884257066162, + "grad_norm": 0.14491412043571472, + "learning_rate": 7e-06, + "loss": 0.3347, + "step": 133 + }, + { + "epoch": 0.16927206695089214, + "grad_norm": 0.1378932148218155, + "learning_rate": 7e-06, + "loss": 0.2638, + "step": 134 + }, + { + "epoch": 0.1705352913311227, + "grad_norm": 0.07053989171981812, + "learning_rate": 7e-06, + "loss": 0.2299, + "step": 135 + }, + { + "epoch": 0.17179851571135324, + "grad_norm": 0.19610151648521423, + "learning_rate": 7e-06, + "loss": 0.2789, + "step": 136 + }, + { + "epoch": 0.17306174009158376, + "grad_norm": 0.1290581375360489, + "learning_rate": 7e-06, + "loss": 0.209, + "step": 137 + }, + { + "epoch": 0.1743249644718143, + "grad_norm": 0.1481819599866867, + "learning_rate": 7e-06, + "loss": 0.2723, + "step": 138 + }, + { + "epoch": 0.17558818885204486, + "grad_norm": 0.1427401453256607, + "learning_rate": 7e-06, + "loss": 0.2778, + "step": 139 + }, + { + "epoch": 0.17685141323227538, + "grad_norm": 0.0666273981332779, + "learning_rate": 7e-06, + "loss": 0.2008, + "step": 140 + }, + { + "epoch": 0.17811463761250593, + "grad_norm": 0.13182522356510162, + "learning_rate": 7e-06, + "loss": 0.2441, + "step": 141 + }, + { + "epoch": 0.17937786199273645, + "grad_norm": 0.08374546468257904, + "learning_rate": 7e-06, + "loss": 0.2603, + "step": 142 + }, + { + "epoch": 0.180641086372967, + "grad_norm": 0.10638394951820374, + "learning_rate": 7e-06, + "loss": 0.2354, + "step": 143 + }, + { + "epoch": 0.18190431075319755, + "grad_norm": 0.10801179707050323, + "learning_rate": 7e-06, + "loss": 0.2875, + "step": 144 + }, + { + "epoch": 0.18316753513342807, + "grad_norm": 0.13121351599693298, + "learning_rate": 7e-06, + "loss": 0.2304, + "step": 145 + }, + { + "epoch": 0.18443075951365862, + "grad_norm": 0.10176476836204529, + "learning_rate": 7e-06, + "loss": 0.2311, + "step": 146 + }, + { + "epoch": 0.18569398389388916, + "grad_norm": 0.10199464112520218, + "learning_rate": 7e-06, + "loss": 0.2522, + "step": 147 + }, + { + "epoch": 0.18695720827411969, + "grad_norm": 0.09650130569934845, + "learning_rate": 7e-06, + "loss": 0.2351, + "step": 148 + }, + { + "epoch": 0.18822043265435023, + "grad_norm": 0.12842021882534027, + "learning_rate": 7e-06, + "loss": 0.2244, + "step": 149 + }, + { + "epoch": 0.18948365703458075, + "grad_norm": 0.1237226277589798, + "learning_rate": 7e-06, + "loss": 0.2706, + "step": 150 + }, + { + "epoch": 0.1907468814148113, + "grad_norm": 0.12939125299453735, + "learning_rate": 7e-06, + "loss": 0.2445, + "step": 151 + }, + { + "epoch": 0.19201010579504185, + "grad_norm": 0.11460690945386887, + "learning_rate": 7e-06, + "loss": 0.2601, + "step": 152 + }, + { + "epoch": 0.19327333017527237, + "grad_norm": 0.18108275532722473, + "learning_rate": 7e-06, + "loss": 0.3465, + "step": 153 + }, + { + "epoch": 0.19453655455550292, + "grad_norm": 0.0727877989411354, + "learning_rate": 7e-06, + "loss": 0.1878, + "step": 154 + }, + { + "epoch": 0.19579977893573347, + "grad_norm": 0.12313497066497803, + "learning_rate": 7e-06, + "loss": 0.2311, + "step": 155 + }, + { + "epoch": 0.197063003315964, + "grad_norm": 0.1377153992652893, + "learning_rate": 7e-06, + "loss": 0.2573, + "step": 156 + }, + { + "epoch": 0.19832622769619454, + "grad_norm": 0.08758647739887238, + "learning_rate": 7e-06, + "loss": 0.2156, + "step": 157 + }, + { + "epoch": 0.19958945207642506, + "grad_norm": 0.11441980302333832, + "learning_rate": 7e-06, + "loss": 0.2801, + "step": 158 + }, + { + "epoch": 0.2008526764566556, + "grad_norm": 0.12151770293712616, + "learning_rate": 7e-06, + "loss": 0.242, + "step": 159 + }, + { + "epoch": 0.20211590083688616, + "grad_norm": 0.159256711602211, + "learning_rate": 7e-06, + "loss": 0.2612, + "step": 160 + }, + { + "epoch": 0.20337912521711668, + "grad_norm": 0.08577941358089447, + "learning_rate": 7e-06, + "loss": 0.2115, + "step": 161 + }, + { + "epoch": 0.20464234959734723, + "grad_norm": 0.1190810427069664, + "learning_rate": 7e-06, + "loss": 0.2434, + "step": 162 + }, + { + "epoch": 0.20590557397757778, + "grad_norm": 0.09624910354614258, + "learning_rate": 7e-06, + "loss": 0.2438, + "step": 163 + }, + { + "epoch": 0.2071687983578083, + "grad_norm": 0.16024184226989746, + "learning_rate": 7e-06, + "loss": 0.2088, + "step": 164 + }, + { + "epoch": 0.20843202273803885, + "grad_norm": 0.1891951858997345, + "learning_rate": 7e-06, + "loss": 0.2751, + "step": 165 + }, + { + "epoch": 0.2096952471182694, + "grad_norm": 0.08837898820638657, + "learning_rate": 7e-06, + "loss": 0.212, + "step": 166 + }, + { + "epoch": 0.21095847149849992, + "grad_norm": 0.0905027762055397, + "learning_rate": 7e-06, + "loss": 0.2189, + "step": 167 + }, + { + "epoch": 0.21222169587873047, + "grad_norm": 0.07917249947786331, + "learning_rate": 7e-06, + "loss": 0.2324, + "step": 168 + }, + { + "epoch": 0.213484920258961, + "grad_norm": 0.13524577021598816, + "learning_rate": 7e-06, + "loss": 0.2143, + "step": 169 + }, + { + "epoch": 0.21474814463919153, + "grad_norm": 0.13222923874855042, + "learning_rate": 7e-06, + "loss": 0.2983, + "step": 170 + }, + { + "epoch": 0.21601136901942208, + "grad_norm": 0.1525893360376358, + "learning_rate": 7e-06, + "loss": 0.2408, + "step": 171 + }, + { + "epoch": 0.2172745933996526, + "grad_norm": 0.08309401571750641, + "learning_rate": 7e-06, + "loss": 0.1722, + "step": 172 + }, + { + "epoch": 0.21853781777988315, + "grad_norm": 0.08370368182659149, + "learning_rate": 7e-06, + "loss": 0.1981, + "step": 173 + }, + { + "epoch": 0.2198010421601137, + "grad_norm": 0.11228370666503906, + "learning_rate": 7e-06, + "loss": 0.2336, + "step": 174 + }, + { + "epoch": 0.22106426654034422, + "grad_norm": 0.19010692834854126, + "learning_rate": 7e-06, + "loss": 0.3069, + "step": 175 + }, + { + "epoch": 0.22232749092057477, + "grad_norm": 0.08182361721992493, + "learning_rate": 7e-06, + "loss": 0.2549, + "step": 176 + }, + { + "epoch": 0.2235907153008053, + "grad_norm": 0.1046992763876915, + "learning_rate": 7e-06, + "loss": 0.2458, + "step": 177 + }, + { + "epoch": 0.22485393968103584, + "grad_norm": 0.11583778262138367, + "learning_rate": 7e-06, + "loss": 0.269, + "step": 178 + }, + { + "epoch": 0.2261171640612664, + "grad_norm": 0.07805290818214417, + "learning_rate": 7e-06, + "loss": 0.1784, + "step": 179 + }, + { + "epoch": 0.2273803884414969, + "grad_norm": 0.11022092401981354, + "learning_rate": 7e-06, + "loss": 0.232, + "step": 180 + }, + { + "epoch": 0.22864361282172746, + "grad_norm": 0.1311209499835968, + "learning_rate": 7e-06, + "loss": 0.2603, + "step": 181 + }, + { + "epoch": 0.229906837201958, + "grad_norm": 0.08558022975921631, + "learning_rate": 7e-06, + "loss": 0.2524, + "step": 182 + }, + { + "epoch": 0.23117006158218853, + "grad_norm": 0.0957944467663765, + "learning_rate": 7e-06, + "loss": 0.281, + "step": 183 + }, + { + "epoch": 0.23243328596241908, + "grad_norm": 0.086683489382267, + "learning_rate": 7e-06, + "loss": 0.2112, + "step": 184 + }, + { + "epoch": 0.2336965103426496, + "grad_norm": 0.09485982358455658, + "learning_rate": 7e-06, + "loss": 0.2146, + "step": 185 + }, + { + "epoch": 0.23495973472288015, + "grad_norm": 0.14843790233135223, + "learning_rate": 7e-06, + "loss": 0.2036, + "step": 186 + }, + { + "epoch": 0.2362229591031107, + "grad_norm": 0.09375383704900742, + "learning_rate": 7e-06, + "loss": 0.2386, + "step": 187 + }, + { + "epoch": 0.23748618348334122, + "grad_norm": 0.10639740526676178, + "learning_rate": 7e-06, + "loss": 0.2202, + "step": 188 + }, + { + "epoch": 0.23874940786357177, + "grad_norm": 0.10205169022083282, + "learning_rate": 7e-06, + "loss": 0.2297, + "step": 189 + }, + { + "epoch": 0.24001263224380232, + "grad_norm": 0.1138874888420105, + "learning_rate": 7e-06, + "loss": 0.2511, + "step": 190 + }, + { + "epoch": 0.24127585662403284, + "grad_norm": 0.12742598354816437, + "learning_rate": 7e-06, + "loss": 0.2247, + "step": 191 + }, + { + "epoch": 0.24253908100426338, + "grad_norm": 0.14605408906936646, + "learning_rate": 7e-06, + "loss": 0.2366, + "step": 192 + }, + { + "epoch": 0.24380230538449393, + "grad_norm": 0.10053393989801407, + "learning_rate": 7e-06, + "loss": 0.4711, + "step": 193 + }, + { + "epoch": 0.24506552976472445, + "grad_norm": 0.08829181641340256, + "learning_rate": 7e-06, + "loss": 0.1501, + "step": 194 + }, + { + "epoch": 0.246328754144955, + "grad_norm": 0.1484231799840927, + "learning_rate": 7e-06, + "loss": 0.2063, + "step": 195 + }, + { + "epoch": 0.24759197852518552, + "grad_norm": 0.17242765426635742, + "learning_rate": 7e-06, + "loss": 0.2317, + "step": 196 + }, + { + "epoch": 0.24885520290541607, + "grad_norm": 0.12016981095075607, + "learning_rate": 7e-06, + "loss": 0.2272, + "step": 197 + }, + { + "epoch": 0.2501184272856466, + "grad_norm": 0.1021333634853363, + "learning_rate": 7e-06, + "loss": 0.2402, + "step": 198 + }, + { + "epoch": 0.25138165166587717, + "grad_norm": 0.11179149895906448, + "learning_rate": 7e-06, + "loss": 0.246, + "step": 199 + }, + { + "epoch": 0.2526448760461077, + "grad_norm": 0.10811345279216766, + "learning_rate": 7e-06, + "loss": 0.2125, + "step": 200 + }, + { + "epoch": 0.2526448760461077, + "eval_correct_accuracy": 0.5988125530110263, + "eval_error_accuracy": 0.4524988743809095, + "eval_f1": 0.5154742907624302, + "eval_loss": 0.4177984297275543, + "eval_runtime": 35.0506, + "eval_samples_per_second": 97.003, + "eval_steps_per_second": 6.077, + "step": 200 + }, + { + "epoch": 0.2539081004263382, + "grad_norm": 0.12190552800893784, + "learning_rate": 7e-06, + "loss": 0.2372, + "step": 201 + }, + { + "epoch": 0.2551713248065688, + "grad_norm": 0.07629604637622833, + "learning_rate": 7e-06, + "loss": 0.1976, + "step": 202 + }, + { + "epoch": 0.2564345491867993, + "grad_norm": 0.10825781524181366, + "learning_rate": 7e-06, + "loss": 0.2169, + "step": 203 + }, + { + "epoch": 0.25769777356702983, + "grad_norm": 0.09181591868400574, + "learning_rate": 7e-06, + "loss": 0.2225, + "step": 204 + }, + { + "epoch": 0.2589609979472604, + "grad_norm": 0.1266108900308609, + "learning_rate": 7e-06, + "loss": 0.1858, + "step": 205 + }, + { + "epoch": 0.26022422232749093, + "grad_norm": 0.11106186360120773, + "learning_rate": 7e-06, + "loss": 0.2443, + "step": 206 + }, + { + "epoch": 0.26148744670772145, + "grad_norm": 0.11874532699584961, + "learning_rate": 7e-06, + "loss": 0.2224, + "step": 207 + }, + { + "epoch": 0.26275067108795197, + "grad_norm": 0.06901393085718155, + "learning_rate": 7e-06, + "loss": 0.1683, + "step": 208 + }, + { + "epoch": 0.26401389546818255, + "grad_norm": 0.1774539351463318, + "learning_rate": 7e-06, + "loss": 0.2588, + "step": 209 + }, + { + "epoch": 0.26527711984841307, + "grad_norm": 0.06564710289239883, + "learning_rate": 7e-06, + "loss": 0.1966, + "step": 210 + }, + { + "epoch": 0.2665403442286436, + "grad_norm": 0.1348266899585724, + "learning_rate": 7e-06, + "loss": 0.2094, + "step": 211 + }, + { + "epoch": 0.26780356860887417, + "grad_norm": 0.10280844569206238, + "learning_rate": 7e-06, + "loss": 0.2208, + "step": 212 + }, + { + "epoch": 0.2690667929891047, + "grad_norm": 0.09777519851922989, + "learning_rate": 7e-06, + "loss": 0.2259, + "step": 213 + }, + { + "epoch": 0.2703300173693352, + "grad_norm": 0.11480893194675446, + "learning_rate": 7e-06, + "loss": 0.2402, + "step": 214 + }, + { + "epoch": 0.2715932417495658, + "grad_norm": 0.17719541490077972, + "learning_rate": 7e-06, + "loss": 0.2692, + "step": 215 + }, + { + "epoch": 0.2728564661297963, + "grad_norm": 0.07069459557533264, + "learning_rate": 7e-06, + "loss": 0.1781, + "step": 216 + }, + { + "epoch": 0.2741196905100268, + "grad_norm": 0.06251855194568634, + "learning_rate": 7e-06, + "loss": 0.1819, + "step": 217 + }, + { + "epoch": 0.2753829148902574, + "grad_norm": 0.1753867119550705, + "learning_rate": 7e-06, + "loss": 0.2362, + "step": 218 + }, + { + "epoch": 0.2766461392704879, + "grad_norm": 0.1843274086713791, + "learning_rate": 7e-06, + "loss": 0.2638, + "step": 219 + }, + { + "epoch": 0.27790936365071844, + "grad_norm": 0.18026292324066162, + "learning_rate": 7e-06, + "loss": 0.2274, + "step": 220 + }, + { + "epoch": 0.279172588030949, + "grad_norm": 0.0640600174665451, + "learning_rate": 7e-06, + "loss": 0.3739, + "step": 221 + }, + { + "epoch": 0.28043581241117954, + "grad_norm": 0.091743104159832, + "learning_rate": 7e-06, + "loss": 0.2274, + "step": 222 + }, + { + "epoch": 0.28169903679141006, + "grad_norm": 0.10185891389846802, + "learning_rate": 7e-06, + "loss": 0.471, + "step": 223 + }, + { + "epoch": 0.28296226117164064, + "grad_norm": 0.08672218769788742, + "learning_rate": 7e-06, + "loss": 0.2171, + "step": 224 + }, + { + "epoch": 0.28422548555187116, + "grad_norm": 0.11758771538734436, + "learning_rate": 7e-06, + "loss": 0.2211, + "step": 225 + }, + { + "epoch": 0.2854887099321017, + "grad_norm": 0.07176447659730911, + "learning_rate": 7e-06, + "loss": 0.1967, + "step": 226 + }, + { + "epoch": 0.2867519343123322, + "grad_norm": 0.1037454828619957, + "learning_rate": 7e-06, + "loss": 0.2457, + "step": 227 + }, + { + "epoch": 0.2880151586925628, + "grad_norm": 0.07262658327817917, + "learning_rate": 7e-06, + "loss": 0.2026, + "step": 228 + }, + { + "epoch": 0.2892783830727933, + "grad_norm": 0.13171784579753876, + "learning_rate": 7e-06, + "loss": 0.209, + "step": 229 + }, + { + "epoch": 0.2905416074530238, + "grad_norm": 0.08208411931991577, + "learning_rate": 7e-06, + "loss": 0.1964, + "step": 230 + }, + { + "epoch": 0.2918048318332544, + "grad_norm": 0.10370495170354843, + "learning_rate": 7e-06, + "loss": 0.2202, + "step": 231 + }, + { + "epoch": 0.2930680562134849, + "grad_norm": 0.26831239461898804, + "learning_rate": 7e-06, + "loss": 0.2651, + "step": 232 + }, + { + "epoch": 0.29433128059371544, + "grad_norm": 0.12230344116687775, + "learning_rate": 7e-06, + "loss": 0.2265, + "step": 233 + }, + { + "epoch": 0.295594504973946, + "grad_norm": 0.08064734190702438, + "learning_rate": 7e-06, + "loss": 0.1711, + "step": 234 + }, + { + "epoch": 0.29685772935417654, + "grad_norm": 0.10691053420305252, + "learning_rate": 7e-06, + "loss": 0.1753, + "step": 235 + }, + { + "epoch": 0.29812095373440706, + "grad_norm": 0.08961788564920425, + "learning_rate": 7e-06, + "loss": 0.2682, + "step": 236 + }, + { + "epoch": 0.29938417811463763, + "grad_norm": 0.2417578548192978, + "learning_rate": 7e-06, + "loss": 0.2622, + "step": 237 + }, + { + "epoch": 0.30064740249486815, + "grad_norm": 0.09739197045564651, + "learning_rate": 7e-06, + "loss": 0.1747, + "step": 238 + }, + { + "epoch": 0.3019106268750987, + "grad_norm": 0.15415729582309723, + "learning_rate": 7e-06, + "loss": 0.2289, + "step": 239 + }, + { + "epoch": 0.30317385125532925, + "grad_norm": 0.08798956125974655, + "learning_rate": 7e-06, + "loss": 0.2076, + "step": 240 + }, + { + "epoch": 0.3044370756355598, + "grad_norm": 0.09532306343317032, + "learning_rate": 7e-06, + "loss": 0.3761, + "step": 241 + }, + { + "epoch": 0.3057003000157903, + "grad_norm": 0.06419141590595245, + "learning_rate": 7e-06, + "loss": 0.2308, + "step": 242 + }, + { + "epoch": 0.30696352439602087, + "grad_norm": 0.13766047358512878, + "learning_rate": 7e-06, + "loss": 0.2203, + "step": 243 + }, + { + "epoch": 0.3082267487762514, + "grad_norm": 0.09225375950336456, + "learning_rate": 7e-06, + "loss": 0.2023, + "step": 244 + }, + { + "epoch": 0.3094899731564819, + "grad_norm": 0.1266135275363922, + "learning_rate": 7e-06, + "loss": 0.2823, + "step": 245 + }, + { + "epoch": 0.31075319753671243, + "grad_norm": 0.17997467517852783, + "learning_rate": 7e-06, + "loss": 0.225, + "step": 246 + }, + { + "epoch": 0.312016421916943, + "grad_norm": 0.12776713073253632, + "learning_rate": 7e-06, + "loss": 0.1906, + "step": 247 + }, + { + "epoch": 0.31327964629717353, + "grad_norm": 0.14866380393505096, + "learning_rate": 7e-06, + "loss": 0.2119, + "step": 248 + }, + { + "epoch": 0.31454287067740405, + "grad_norm": 0.11824511736631393, + "learning_rate": 7e-06, + "loss": 0.2219, + "step": 249 + }, + { + "epoch": 0.31580609505763463, + "grad_norm": 0.14409460127353668, + "learning_rate": 7e-06, + "loss": 0.2116, + "step": 250 + }, + { + "epoch": 0.31706931943786515, + "grad_norm": 0.10304541140794754, + "learning_rate": 7e-06, + "loss": 0.2073, + "step": 251 + }, + { + "epoch": 0.31833254381809567, + "grad_norm": 0.09163326770067215, + "learning_rate": 7e-06, + "loss": 0.1882, + "step": 252 + }, + { + "epoch": 0.31959576819832625, + "grad_norm": 0.12692378461360931, + "learning_rate": 7e-06, + "loss": 0.2386, + "step": 253 + }, + { + "epoch": 0.32085899257855677, + "grad_norm": 0.1747879534959793, + "learning_rate": 7e-06, + "loss": 0.2054, + "step": 254 + }, + { + "epoch": 0.3221222169587873, + "grad_norm": 0.12346009910106659, + "learning_rate": 7e-06, + "loss": 0.2397, + "step": 255 + }, + { + "epoch": 0.32338544133901787, + "grad_norm": 0.1731298863887787, + "learning_rate": 7e-06, + "loss": 0.2575, + "step": 256 + }, + { + "epoch": 0.3246486657192484, + "grad_norm": 0.08011125028133392, + "learning_rate": 7e-06, + "loss": 0.215, + "step": 257 + }, + { + "epoch": 0.3259118900994789, + "grad_norm": 0.13160613179206848, + "learning_rate": 7e-06, + "loss": 0.222, + "step": 258 + }, + { + "epoch": 0.3271751144797095, + "grad_norm": 0.18522977828979492, + "learning_rate": 7e-06, + "loss": 0.2548, + "step": 259 + }, + { + "epoch": 0.32843833885994, + "grad_norm": 0.14212659001350403, + "learning_rate": 7e-06, + "loss": 0.3002, + "step": 260 + }, + { + "epoch": 0.3297015632401705, + "grad_norm": 0.13445697724819183, + "learning_rate": 7e-06, + "loss": 0.2351, + "step": 261 + }, + { + "epoch": 0.33096478762040105, + "grad_norm": 0.11636935919523239, + "learning_rate": 7e-06, + "loss": 0.2106, + "step": 262 + }, + { + "epoch": 0.3322280120006316, + "grad_norm": 0.14159604907035828, + "learning_rate": 7e-06, + "loss": 0.2531, + "step": 263 + }, + { + "epoch": 0.33349123638086214, + "grad_norm": 0.10319356620311737, + "learning_rate": 7e-06, + "loss": 0.2346, + "step": 264 + }, + { + "epoch": 0.33475446076109266, + "grad_norm": 0.09796885401010513, + "learning_rate": 7e-06, + "loss": 0.2059, + "step": 265 + }, + { + "epoch": 0.33601768514132324, + "grad_norm": 0.1082499697804451, + "learning_rate": 7e-06, + "loss": 0.1863, + "step": 266 + }, + { + "epoch": 0.33728090952155376, + "grad_norm": 0.12492396682500839, + "learning_rate": 7e-06, + "loss": 0.215, + "step": 267 + }, + { + "epoch": 0.3385441339017843, + "grad_norm": 0.06617411971092224, + "learning_rate": 7e-06, + "loss": 0.1782, + "step": 268 + }, + { + "epoch": 0.33980735828201486, + "grad_norm": 0.15060101449489594, + "learning_rate": 7e-06, + "loss": 0.3509, + "step": 269 + }, + { + "epoch": 0.3410705826622454, + "grad_norm": 0.11944282054901123, + "learning_rate": 7e-06, + "loss": 0.1862, + "step": 270 + }, + { + "epoch": 0.3423338070424759, + "grad_norm": 0.12389136850833893, + "learning_rate": 7e-06, + "loss": 0.2409, + "step": 271 + }, + { + "epoch": 0.3435970314227065, + "grad_norm": 0.11000983417034149, + "learning_rate": 7e-06, + "loss": 0.223, + "step": 272 + }, + { + "epoch": 0.344860255802937, + "grad_norm": 0.09012436866760254, + "learning_rate": 7e-06, + "loss": 0.2068, + "step": 273 + }, + { + "epoch": 0.3461234801831675, + "grad_norm": 0.27014490962028503, + "learning_rate": 7e-06, + "loss": 0.2087, + "step": 274 + }, + { + "epoch": 0.3473867045633981, + "grad_norm": 0.08035814762115479, + "learning_rate": 7e-06, + "loss": 0.262, + "step": 275 + }, + { + "epoch": 0.3486499289436286, + "grad_norm": 0.09129905700683594, + "learning_rate": 7e-06, + "loss": 0.2082, + "step": 276 + }, + { + "epoch": 0.34991315332385914, + "grad_norm": 0.11665099114179611, + "learning_rate": 7e-06, + "loss": 0.1841, + "step": 277 + }, + { + "epoch": 0.3511763777040897, + "grad_norm": 0.08812276273965836, + "learning_rate": 7e-06, + "loss": 0.2164, + "step": 278 + }, + { + "epoch": 0.35243960208432024, + "grad_norm": 0.1272403746843338, + "learning_rate": 7e-06, + "loss": 0.1927, + "step": 279 + }, + { + "epoch": 0.35370282646455076, + "grad_norm": 0.11256379634141922, + "learning_rate": 7e-06, + "loss": 0.2991, + "step": 280 + }, + { + "epoch": 0.3549660508447813, + "grad_norm": 0.15795424580574036, + "learning_rate": 7e-06, + "loss": 0.2363, + "step": 281 + }, + { + "epoch": 0.35622927522501185, + "grad_norm": 0.22632326185703278, + "learning_rate": 7e-06, + "loss": 0.2088, + "step": 282 + }, + { + "epoch": 0.3574924996052424, + "grad_norm": 0.18535131216049194, + "learning_rate": 7e-06, + "loss": 0.2746, + "step": 283 + }, + { + "epoch": 0.3587557239854729, + "grad_norm": 0.08579732477664948, + "learning_rate": 7e-06, + "loss": 0.1899, + "step": 284 + }, + { + "epoch": 0.3600189483657035, + "grad_norm": 0.10859379917383194, + "learning_rate": 7e-06, + "loss": 0.2067, + "step": 285 + }, + { + "epoch": 0.361282172745934, + "grad_norm": 0.07765299826860428, + "learning_rate": 7e-06, + "loss": 0.1761, + "step": 286 + }, + { + "epoch": 0.3625453971261645, + "grad_norm": 0.17053595185279846, + "learning_rate": 7e-06, + "loss": 0.2373, + "step": 287 + }, + { + "epoch": 0.3638086215063951, + "grad_norm": 0.09873699396848679, + "learning_rate": 7e-06, + "loss": 0.2176, + "step": 288 + }, + { + "epoch": 0.3650718458866256, + "grad_norm": 0.07418286055326462, + "learning_rate": 7e-06, + "loss": 0.1797, + "step": 289 + }, + { + "epoch": 0.36633507026685613, + "grad_norm": 0.11981359124183655, + "learning_rate": 7e-06, + "loss": 0.1988, + "step": 290 + }, + { + "epoch": 0.3675982946470867, + "grad_norm": 0.06424502283334732, + "learning_rate": 7e-06, + "loss": 0.2121, + "step": 291 + }, + { + "epoch": 0.36886151902731723, + "grad_norm": 0.09006607532501221, + "learning_rate": 7e-06, + "loss": 0.1945, + "step": 292 + }, + { + "epoch": 0.37012474340754775, + "grad_norm": 0.10973497480154037, + "learning_rate": 7e-06, + "loss": 0.2046, + "step": 293 + }, + { + "epoch": 0.37138796778777833, + "grad_norm": 0.09228470921516418, + "learning_rate": 7e-06, + "loss": 0.207, + "step": 294 + }, + { + "epoch": 0.37265119216800885, + "grad_norm": 0.10961271822452545, + "learning_rate": 7e-06, + "loss": 0.2128, + "step": 295 + }, + { + "epoch": 0.37391441654823937, + "grad_norm": 0.09072300046682358, + "learning_rate": 7e-06, + "loss": 0.4585, + "step": 296 + }, + { + "epoch": 0.37517764092846995, + "grad_norm": 0.08374742418527603, + "learning_rate": 7e-06, + "loss": 0.2178, + "step": 297 + }, + { + "epoch": 0.37644086530870047, + "grad_norm": 0.05344458296895027, + "learning_rate": 7e-06, + "loss": 0.1595, + "step": 298 + }, + { + "epoch": 0.377704089688931, + "grad_norm": 0.07841549813747406, + "learning_rate": 7e-06, + "loss": 0.2306, + "step": 299 + }, + { + "epoch": 0.3789673140691615, + "grad_norm": 0.09865035116672516, + "learning_rate": 7e-06, + "loss": 0.2274, + "step": 300 + }, + { + "epoch": 0.3789673140691615, + "eval_correct_accuracy": 0.5818490245971162, + "eval_error_accuracy": 0.494822152183701, + "eval_f1": 0.5348184158843582, + "eval_loss": 0.41273096203804016, + "eval_runtime": 35.0595, + "eval_samples_per_second": 96.978, + "eval_steps_per_second": 6.075, + "step": 300 + }, + { + "epoch": 0.3802305384493921, + "grad_norm": 0.11520479619503021, + "learning_rate": 7e-06, + "loss": 0.194, + "step": 301 + }, + { + "epoch": 0.3814937628296226, + "grad_norm": 0.17391149699687958, + "learning_rate": 7e-06, + "loss": 0.2053, + "step": 302 + }, + { + "epoch": 0.38275698720985313, + "grad_norm": 0.08927040547132492, + "learning_rate": 7e-06, + "loss": 0.1885, + "step": 303 + }, + { + "epoch": 0.3840202115900837, + "grad_norm": 0.10747874528169632, + "learning_rate": 7e-06, + "loss": 0.2357, + "step": 304 + }, + { + "epoch": 0.3852834359703142, + "grad_norm": 0.0821816474199295, + "learning_rate": 7e-06, + "loss": 0.2017, + "step": 305 + }, + { + "epoch": 0.38654666035054475, + "grad_norm": 0.08718965202569962, + "learning_rate": 7e-06, + "loss": 0.2333, + "step": 306 + }, + { + "epoch": 0.3878098847307753, + "grad_norm": 0.14753767848014832, + "learning_rate": 7e-06, + "loss": 0.2501, + "step": 307 + }, + { + "epoch": 0.38907310911100584, + "grad_norm": 0.12474358081817627, + "learning_rate": 7e-06, + "loss": 0.1895, + "step": 308 + }, + { + "epoch": 0.39033633349123636, + "grad_norm": 0.14409278333187103, + "learning_rate": 7e-06, + "loss": 0.208, + "step": 309 + }, + { + "epoch": 0.39159955787146694, + "grad_norm": 0.06918184459209442, + "learning_rate": 7e-06, + "loss": 0.1817, + "step": 310 + }, + { + "epoch": 0.39286278225169746, + "grad_norm": 0.08502199500799179, + "learning_rate": 7e-06, + "loss": 0.1832, + "step": 311 + }, + { + "epoch": 0.394126006631928, + "grad_norm": 0.06989938765764236, + "learning_rate": 7e-06, + "loss": 0.1741, + "step": 312 + }, + { + "epoch": 0.39538923101215856, + "grad_norm": 0.08131398260593414, + "learning_rate": 7e-06, + "loss": 0.226, + "step": 313 + }, + { + "epoch": 0.3966524553923891, + "grad_norm": 0.16150841116905212, + "learning_rate": 7e-06, + "loss": 0.2081, + "step": 314 + }, + { + "epoch": 0.3979156797726196, + "grad_norm": 0.10033854097127914, + "learning_rate": 7e-06, + "loss": 0.1757, + "step": 315 + }, + { + "epoch": 0.3991789041528501, + "grad_norm": 0.2944275438785553, + "learning_rate": 7e-06, + "loss": 0.2039, + "step": 316 + }, + { + "epoch": 0.4004421285330807, + "grad_norm": 0.09300543367862701, + "learning_rate": 7e-06, + "loss": 0.2403, + "step": 317 + }, + { + "epoch": 0.4017053529133112, + "grad_norm": 0.089630626142025, + "learning_rate": 7e-06, + "loss": 0.2457, + "step": 318 + }, + { + "epoch": 0.40296857729354174, + "grad_norm": 0.06648046523332596, + "learning_rate": 7e-06, + "loss": 0.2155, + "step": 319 + }, + { + "epoch": 0.4042318016737723, + "grad_norm": 0.18262338638305664, + "learning_rate": 7e-06, + "loss": 0.2087, + "step": 320 + }, + { + "epoch": 0.40549502605400284, + "grad_norm": 0.0919061154127121, + "learning_rate": 7e-06, + "loss": 0.2062, + "step": 321 + }, + { + "epoch": 0.40675825043423336, + "grad_norm": 0.113703154027462, + "learning_rate": 7e-06, + "loss": 0.1859, + "step": 322 + }, + { + "epoch": 0.40802147481446394, + "grad_norm": 0.20705194771289825, + "learning_rate": 7e-06, + "loss": 0.1769, + "step": 323 + }, + { + "epoch": 0.40928469919469446, + "grad_norm": 0.11209185421466827, + "learning_rate": 7e-06, + "loss": 0.1819, + "step": 324 + }, + { + "epoch": 0.410547923574925, + "grad_norm": 0.05803574621677399, + "learning_rate": 7e-06, + "loss": 0.1852, + "step": 325 + }, + { + "epoch": 0.41181114795515555, + "grad_norm": 0.16077323257923126, + "learning_rate": 7e-06, + "loss": 0.2117, + "step": 326 + }, + { + "epoch": 0.4130743723353861, + "grad_norm": 0.10078177601099014, + "learning_rate": 7e-06, + "loss": 0.193, + "step": 327 + }, + { + "epoch": 0.4143375967156166, + "grad_norm": 0.09989168494939804, + "learning_rate": 7e-06, + "loss": 0.2053, + "step": 328 + }, + { + "epoch": 0.4156008210958472, + "grad_norm": 0.13987579941749573, + "learning_rate": 7e-06, + "loss": 0.2678, + "step": 329 + }, + { + "epoch": 0.4168640454760777, + "grad_norm": 0.13039669394493103, + "learning_rate": 7e-06, + "loss": 0.1998, + "step": 330 + }, + { + "epoch": 0.4181272698563082, + "grad_norm": 0.1029522716999054, + "learning_rate": 7e-06, + "loss": 0.2337, + "step": 331 + }, + { + "epoch": 0.4193904942365388, + "grad_norm": 0.08752740174531937, + "learning_rate": 7e-06, + "loss": 0.1854, + "step": 332 + }, + { + "epoch": 0.4206537186167693, + "grad_norm": 0.07876112312078476, + "learning_rate": 7e-06, + "loss": 0.1909, + "step": 333 + }, + { + "epoch": 0.42191694299699983, + "grad_norm": 0.2126246988773346, + "learning_rate": 7e-06, + "loss": 0.2535, + "step": 334 + }, + { + "epoch": 0.42318016737723035, + "grad_norm": 0.11913909763097763, + "learning_rate": 7e-06, + "loss": 0.2184, + "step": 335 + }, + { + "epoch": 0.42444339175746093, + "grad_norm": 0.1513642817735672, + "learning_rate": 7e-06, + "loss": 0.1994, + "step": 336 + }, + { + "epoch": 0.42570661613769145, + "grad_norm": 0.1306588500738144, + "learning_rate": 7e-06, + "loss": 0.2126, + "step": 337 + }, + { + "epoch": 0.426969840517922, + "grad_norm": 0.1171175092458725, + "learning_rate": 7e-06, + "loss": 0.1994, + "step": 338 + }, + { + "epoch": 0.42823306489815255, + "grad_norm": 0.05895727127790451, + "learning_rate": 7e-06, + "loss": 0.155, + "step": 339 + }, + { + "epoch": 0.42949628927838307, + "grad_norm": 0.08570288121700287, + "learning_rate": 7e-06, + "loss": 0.1986, + "step": 340 + }, + { + "epoch": 0.4307595136586136, + "grad_norm": 0.0765470489859581, + "learning_rate": 7e-06, + "loss": 0.1696, + "step": 341 + }, + { + "epoch": 0.43202273803884417, + "grad_norm": 0.08286664634943008, + "learning_rate": 7e-06, + "loss": 0.1626, + "step": 342 + }, + { + "epoch": 0.4332859624190747, + "grad_norm": 0.2282284051179886, + "learning_rate": 7e-06, + "loss": 0.1877, + "step": 343 + }, + { + "epoch": 0.4345491867993052, + "grad_norm": 0.11943413317203522, + "learning_rate": 7e-06, + "loss": 0.1897, + "step": 344 + }, + { + "epoch": 0.4358124111795358, + "grad_norm": 0.10935524851083755, + "learning_rate": 7e-06, + "loss": 0.1828, + "step": 345 + }, + { + "epoch": 0.4370756355597663, + "grad_norm": 0.07996437698602676, + "learning_rate": 7e-06, + "loss": 0.1991, + "step": 346 + }, + { + "epoch": 0.4383388599399968, + "grad_norm": 0.13088780641555786, + "learning_rate": 7e-06, + "loss": 0.3873, + "step": 347 + }, + { + "epoch": 0.4396020843202274, + "grad_norm": 0.15082432329654694, + "learning_rate": 7e-06, + "loss": 0.2091, + "step": 348 + }, + { + "epoch": 0.4408653087004579, + "grad_norm": 0.132376566529274, + "learning_rate": 7e-06, + "loss": 0.2471, + "step": 349 + }, + { + "epoch": 0.44212853308068845, + "grad_norm": 0.07796452194452286, + "learning_rate": 7e-06, + "loss": 0.1751, + "step": 350 + }, + { + "epoch": 0.443391757460919, + "grad_norm": 0.12849055230617523, + "learning_rate": 7e-06, + "loss": 0.3155, + "step": 351 + }, + { + "epoch": 0.44465498184114954, + "grad_norm": 0.06422396749258041, + "learning_rate": 7e-06, + "loss": 0.1486, + "step": 352 + }, + { + "epoch": 0.44591820622138006, + "grad_norm": 0.1800646334886551, + "learning_rate": 7e-06, + "loss": 0.1993, + "step": 353 + }, + { + "epoch": 0.4471814306016106, + "grad_norm": 0.15747664868831635, + "learning_rate": 7e-06, + "loss": 0.209, + "step": 354 + }, + { + "epoch": 0.44844465498184116, + "grad_norm": 0.11023043096065521, + "learning_rate": 7e-06, + "loss": 0.21, + "step": 355 + }, + { + "epoch": 0.4497078793620717, + "grad_norm": 0.0927424430847168, + "learning_rate": 7e-06, + "loss": 0.1745, + "step": 356 + }, + { + "epoch": 0.4509711037423022, + "grad_norm": 0.08278126269578934, + "learning_rate": 7e-06, + "loss": 0.3105, + "step": 357 + }, + { + "epoch": 0.4522343281225328, + "grad_norm": 0.08794251829385757, + "learning_rate": 7e-06, + "loss": 0.1979, + "step": 358 + }, + { + "epoch": 0.4534975525027633, + "grad_norm": 0.11653570830821991, + "learning_rate": 7e-06, + "loss": 0.1828, + "step": 359 + }, + { + "epoch": 0.4547607768829938, + "grad_norm": 0.11114069074392319, + "learning_rate": 7e-06, + "loss": 0.1826, + "step": 360 + }, + { + "epoch": 0.4560240012632244, + "grad_norm": 0.2608173191547394, + "learning_rate": 7e-06, + "loss": 0.2304, + "step": 361 + }, + { + "epoch": 0.4572872256434549, + "grad_norm": 0.08441725373268127, + "learning_rate": 7e-06, + "loss": 0.1757, + "step": 362 + }, + { + "epoch": 0.45855045002368544, + "grad_norm": 0.10891429334878922, + "learning_rate": 7e-06, + "loss": 0.2122, + "step": 363 + }, + { + "epoch": 0.459813674403916, + "grad_norm": 0.07106776535511017, + "learning_rate": 7e-06, + "loss": 0.1721, + "step": 364 + }, + { + "epoch": 0.46107689878414654, + "grad_norm": 0.08842181414365768, + "learning_rate": 7e-06, + "loss": 0.2226, + "step": 365 + }, + { + "epoch": 0.46234012316437706, + "grad_norm": 0.0870131179690361, + "learning_rate": 7e-06, + "loss": 0.2474, + "step": 366 + }, + { + "epoch": 0.46360334754460764, + "grad_norm": 0.14521507918834686, + "learning_rate": 7e-06, + "loss": 0.2615, + "step": 367 + }, + { + "epoch": 0.46486657192483816, + "grad_norm": 0.09553767740726471, + "learning_rate": 7e-06, + "loss": 0.1791, + "step": 368 + }, + { + "epoch": 0.4661297963050687, + "grad_norm": 0.11010967195034027, + "learning_rate": 7e-06, + "loss": 0.1874, + "step": 369 + }, + { + "epoch": 0.4673930206852992, + "grad_norm": 0.09533923864364624, + "learning_rate": 7e-06, + "loss": 0.228, + "step": 370 + }, + { + "epoch": 0.4686562450655298, + "grad_norm": 0.0890774354338646, + "learning_rate": 7e-06, + "loss": 0.2345, + "step": 371 + }, + { + "epoch": 0.4699194694457603, + "grad_norm": 0.12173017859458923, + "learning_rate": 7e-06, + "loss": 0.2082, + "step": 372 + }, + { + "epoch": 0.4711826938259908, + "grad_norm": 0.0602993369102478, + "learning_rate": 7e-06, + "loss": 0.1893, + "step": 373 + }, + { + "epoch": 0.4724459182062214, + "grad_norm": 0.13122287392616272, + "learning_rate": 7e-06, + "loss": 0.2178, + "step": 374 + }, + { + "epoch": 0.4737091425864519, + "grad_norm": 0.07299527525901794, + "learning_rate": 7e-06, + "loss": 0.1888, + "step": 375 + }, + { + "epoch": 0.47497236696668244, + "grad_norm": 0.08244926482439041, + "learning_rate": 7e-06, + "loss": 0.174, + "step": 376 + }, + { + "epoch": 0.476235591346913, + "grad_norm": 0.08397851884365082, + "learning_rate": 7e-06, + "loss": 0.2108, + "step": 377 + }, + { + "epoch": 0.47749881572714353, + "grad_norm": 0.07320383936166763, + "learning_rate": 7e-06, + "loss": 0.1804, + "step": 378 + }, + { + "epoch": 0.47876204010737405, + "grad_norm": 0.0849589854478836, + "learning_rate": 7e-06, + "loss": 0.1829, + "step": 379 + }, + { + "epoch": 0.48002526448760463, + "grad_norm": 0.10207744687795639, + "learning_rate": 7e-06, + "loss": 0.2174, + "step": 380 + }, + { + "epoch": 0.48128848886783515, + "grad_norm": 0.07175120711326599, + "learning_rate": 7e-06, + "loss": 0.1847, + "step": 381 + }, + { + "epoch": 0.4825517132480657, + "grad_norm": 0.10446271300315857, + "learning_rate": 7e-06, + "loss": 0.216, + "step": 382 + }, + { + "epoch": 0.48381493762829625, + "grad_norm": 0.23799718916416168, + "learning_rate": 7e-06, + "loss": 0.2136, + "step": 383 + }, + { + "epoch": 0.48507816200852677, + "grad_norm": 0.11531874537467957, + "learning_rate": 7e-06, + "loss": 0.2357, + "step": 384 + }, + { + "epoch": 0.4863413863887573, + "grad_norm": 0.10034700483083725, + "learning_rate": 7e-06, + "loss": 0.2258, + "step": 385 + }, + { + "epoch": 0.48760461076898787, + "grad_norm": 0.0934348776936531, + "learning_rate": 7e-06, + "loss": 0.284, + "step": 386 + }, + { + "epoch": 0.4888678351492184, + "grad_norm": 0.165315181016922, + "learning_rate": 7e-06, + "loss": 0.2264, + "step": 387 + }, + { + "epoch": 0.4901310595294489, + "grad_norm": 0.1086471751332283, + "learning_rate": 7e-06, + "loss": 0.2028, + "step": 388 + }, + { + "epoch": 0.49139428390967943, + "grad_norm": 0.14764176309108734, + "learning_rate": 7e-06, + "loss": 0.1775, + "step": 389 + }, + { + "epoch": 0.49265750828991, + "grad_norm": 0.21734580397605896, + "learning_rate": 7e-06, + "loss": 0.1924, + "step": 390 + }, + { + "epoch": 0.4939207326701405, + "grad_norm": 0.0923137441277504, + "learning_rate": 7e-06, + "loss": 0.2031, + "step": 391 + }, + { + "epoch": 0.49518395705037105, + "grad_norm": 0.06933951377868652, + "learning_rate": 7e-06, + "loss": 0.1739, + "step": 392 + }, + { + "epoch": 0.4964471814306016, + "grad_norm": 0.0930216833949089, + "learning_rate": 7e-06, + "loss": 0.2083, + "step": 393 + }, + { + "epoch": 0.49771040581083215, + "grad_norm": 0.08797884732484818, + "learning_rate": 7e-06, + "loss": 0.2636, + "step": 394 + }, + { + "epoch": 0.49897363019106267, + "grad_norm": 0.0919070690870285, + "learning_rate": 7e-06, + "loss": 0.2154, + "step": 395 + }, + { + "epoch": 0.5002368545712932, + "grad_norm": 0.07787168025970459, + "learning_rate": 7e-06, + "loss": 0.2207, + "step": 396 + }, + { + "epoch": 0.5015000789515237, + "grad_norm": 0.11572758853435516, + "learning_rate": 7e-06, + "loss": 0.2002, + "step": 397 + }, + { + "epoch": 0.5027633033317543, + "grad_norm": 0.08295108377933502, + "learning_rate": 7e-06, + "loss": 0.2614, + "step": 398 + }, + { + "epoch": 0.5040265277119849, + "grad_norm": 0.0625801831483841, + "learning_rate": 7e-06, + "loss": 0.1644, + "step": 399 + }, + { + "epoch": 0.5052897520922154, + "grad_norm": 0.07405094802379608, + "learning_rate": 7e-06, + "loss": 0.2234, + "step": 400 + }, + { + "epoch": 0.5052897520922154, + "eval_correct_accuracy": 0.5852417302798982, + "eval_error_accuracy": 0.4709590274651058, + "eval_f1": 0.5219175883059916, + "eval_loss": 0.4229665994644165, + "eval_runtime": 35.2003, + "eval_samples_per_second": 96.59, + "eval_steps_per_second": 6.051, + "step": 400 + }, + { + "epoch": 0.5065529764724459, + "grad_norm": 0.09175197780132294, + "learning_rate": 7e-06, + "loss": 0.2027, + "step": 401 + }, + { + "epoch": 0.5078162008526764, + "grad_norm": 0.1550239622592926, + "learning_rate": 7e-06, + "loss": 0.2559, + "step": 402 + }, + { + "epoch": 0.509079425232907, + "grad_norm": 0.139438658952713, + "learning_rate": 7e-06, + "loss": 0.1762, + "step": 403 + }, + { + "epoch": 0.5103426496131376, + "grad_norm": 0.11481575667858124, + "learning_rate": 7e-06, + "loss": 0.1623, + "step": 404 + }, + { + "epoch": 0.5116058739933681, + "grad_norm": 0.05404340475797653, + "learning_rate": 7e-06, + "loss": 0.1961, + "step": 405 + }, + { + "epoch": 0.5128690983735986, + "grad_norm": 0.14743672311306, + "learning_rate": 7e-06, + "loss": 0.2279, + "step": 406 + }, + { + "epoch": 0.5141323227538291, + "grad_norm": 0.11647465080022812, + "learning_rate": 7e-06, + "loss": 0.2001, + "step": 407 + }, + { + "epoch": 0.5153955471340597, + "grad_norm": 0.08203577995300293, + "learning_rate": 7e-06, + "loss": 0.1752, + "step": 408 + }, + { + "epoch": 0.5166587715142902, + "grad_norm": 0.11073414981365204, + "learning_rate": 7e-06, + "loss": 0.3686, + "step": 409 + }, + { + "epoch": 0.5179219958945208, + "grad_norm": 0.11331301182508469, + "learning_rate": 7e-06, + "loss": 0.3378, + "step": 410 + }, + { + "epoch": 0.5191852202747513, + "grad_norm": 0.09435959905385971, + "learning_rate": 7e-06, + "loss": 0.1562, + "step": 411 + }, + { + "epoch": 0.5204484446549819, + "grad_norm": 0.08365237712860107, + "learning_rate": 7e-06, + "loss": 0.1926, + "step": 412 + }, + { + "epoch": 0.5217116690352124, + "grad_norm": 0.08092326670885086, + "learning_rate": 7e-06, + "loss": 0.1931, + "step": 413 + }, + { + "epoch": 0.5229748934154429, + "grad_norm": 0.07763849943876266, + "learning_rate": 7e-06, + "loss": 0.2038, + "step": 414 + }, + { + "epoch": 0.5242381177956734, + "grad_norm": 0.1350603550672531, + "learning_rate": 7e-06, + "loss": 0.2392, + "step": 415 + }, + { + "epoch": 0.5255013421759039, + "grad_norm": 0.10287491232156754, + "learning_rate": 7e-06, + "loss": 0.196, + "step": 416 + }, + { + "epoch": 0.5267645665561346, + "grad_norm": 0.0719987079501152, + "learning_rate": 7e-06, + "loss": 0.2016, + "step": 417 + }, + { + "epoch": 0.5280277909363651, + "grad_norm": 0.22227227687835693, + "learning_rate": 7e-06, + "loss": 0.2131, + "step": 418 + }, + { + "epoch": 0.5292910153165956, + "grad_norm": 0.06136275455355644, + "learning_rate": 7e-06, + "loss": 0.2044, + "step": 419 + }, + { + "epoch": 0.5305542396968261, + "grad_norm": 0.0627446323633194, + "learning_rate": 7e-06, + "loss": 0.1793, + "step": 420 + }, + { + "epoch": 0.5318174640770567, + "grad_norm": 0.20960237085819244, + "learning_rate": 7e-06, + "loss": 0.1982, + "step": 421 + }, + { + "epoch": 0.5330806884572872, + "grad_norm": 0.11971580237150192, + "learning_rate": 7e-06, + "loss": 0.2587, + "step": 422 + }, + { + "epoch": 0.5343439128375178, + "grad_norm": 0.0932474359869957, + "learning_rate": 7e-06, + "loss": 0.1947, + "step": 423 + }, + { + "epoch": 0.5356071372177483, + "grad_norm": 0.09686949849128723, + "learning_rate": 7e-06, + "loss": 0.1907, + "step": 424 + }, + { + "epoch": 0.5368703615979789, + "grad_norm": 0.07940957695245743, + "learning_rate": 7e-06, + "loss": 0.2535, + "step": 425 + }, + { + "epoch": 0.5381335859782094, + "grad_norm": 0.09676375240087509, + "learning_rate": 7e-06, + "loss": 0.2038, + "step": 426 + }, + { + "epoch": 0.5393968103584399, + "grad_norm": 0.09989267587661743, + "learning_rate": 7e-06, + "loss": 0.1976, + "step": 427 + }, + { + "epoch": 0.5406600347386704, + "grad_norm": 0.0823327898979187, + "learning_rate": 7e-06, + "loss": 0.1708, + "step": 428 + }, + { + "epoch": 0.541923259118901, + "grad_norm": 0.10633084177970886, + "learning_rate": 7e-06, + "loss": 0.1619, + "step": 429 + }, + { + "epoch": 0.5431864834991316, + "grad_norm": 0.08448205143213272, + "learning_rate": 7e-06, + "loss": 0.1854, + "step": 430 + }, + { + "epoch": 0.5444497078793621, + "grad_norm": 0.07697522640228271, + "learning_rate": 7e-06, + "loss": 0.188, + "step": 431 + }, + { + "epoch": 0.5457129322595926, + "grad_norm": 0.1970750391483307, + "learning_rate": 7e-06, + "loss": 0.2479, + "step": 432 + }, + { + "epoch": 0.5469761566398231, + "grad_norm": 0.08660274744033813, + "learning_rate": 7e-06, + "loss": 0.1536, + "step": 433 + }, + { + "epoch": 0.5482393810200537, + "grad_norm": 0.0931171253323555, + "learning_rate": 7e-06, + "loss": 0.2169, + "step": 434 + }, + { + "epoch": 0.5495026054002842, + "grad_norm": 0.07317376136779785, + "learning_rate": 7e-06, + "loss": 0.1941, + "step": 435 + }, + { + "epoch": 0.5507658297805148, + "grad_norm": 0.07506151497364044, + "learning_rate": 7e-06, + "loss": 0.1761, + "step": 436 + }, + { + "epoch": 0.5520290541607453, + "grad_norm": 0.059854380786418915, + "learning_rate": 7e-06, + "loss": 0.3068, + "step": 437 + }, + { + "epoch": 0.5532922785409758, + "grad_norm": 0.1609865128993988, + "learning_rate": 7e-06, + "loss": 0.1857, + "step": 438 + }, + { + "epoch": 0.5545555029212064, + "grad_norm": 0.08996118605136871, + "learning_rate": 7e-06, + "loss": 0.1805, + "step": 439 + }, + { + "epoch": 0.5558187273014369, + "grad_norm": 0.12089324742555618, + "learning_rate": 7e-06, + "loss": 0.1826, + "step": 440 + }, + { + "epoch": 0.5570819516816674, + "grad_norm": 0.08772964775562286, + "learning_rate": 7e-06, + "loss": 0.1791, + "step": 441 + }, + { + "epoch": 0.558345176061898, + "grad_norm": 0.10977458208799362, + "learning_rate": 7e-06, + "loss": 0.2084, + "step": 442 + }, + { + "epoch": 0.5596084004421286, + "grad_norm": 0.09188458323478699, + "learning_rate": 7e-06, + "loss": 0.3045, + "step": 443 + }, + { + "epoch": 0.5608716248223591, + "grad_norm": 0.07033522427082062, + "learning_rate": 7e-06, + "loss": 0.1926, + "step": 444 + }, + { + "epoch": 0.5621348492025896, + "grad_norm": 0.0652671530842781, + "learning_rate": 7e-06, + "loss": 0.1998, + "step": 445 + }, + { + "epoch": 0.5633980735828201, + "grad_norm": 0.07860173285007477, + "learning_rate": 7e-06, + "loss": 0.2172, + "step": 446 + }, + { + "epoch": 0.5646612979630506, + "grad_norm": 0.0679745227098465, + "learning_rate": 7e-06, + "loss": 0.179, + "step": 447 + }, + { + "epoch": 0.5659245223432813, + "grad_norm": 0.10545714199542999, + "learning_rate": 7e-06, + "loss": 0.1547, + "step": 448 + }, + { + "epoch": 0.5671877467235118, + "grad_norm": 0.07516340911388397, + "learning_rate": 7e-06, + "loss": 0.1643, + "step": 449 + }, + { + "epoch": 0.5684509711037423, + "grad_norm": 0.06046690791845322, + "learning_rate": 7e-06, + "loss": 0.1624, + "step": 450 + }, + { + "epoch": 0.5697141954839728, + "grad_norm": 0.07889428734779358, + "learning_rate": 7e-06, + "loss": 0.1808, + "step": 451 + }, + { + "epoch": 0.5709774198642034, + "grad_norm": 0.08698045462369919, + "learning_rate": 7e-06, + "loss": 0.2257, + "step": 452 + }, + { + "epoch": 0.5722406442444339, + "grad_norm": 0.08498376607894897, + "learning_rate": 7e-06, + "loss": 0.1773, + "step": 453 + }, + { + "epoch": 0.5735038686246644, + "grad_norm": 0.12781842052936554, + "learning_rate": 7e-06, + "loss": 0.258, + "step": 454 + }, + { + "epoch": 0.574767093004895, + "grad_norm": 0.11314232647418976, + "learning_rate": 7e-06, + "loss": 0.1608, + "step": 455 + }, + { + "epoch": 0.5760303173851256, + "grad_norm": 0.3507859408855438, + "learning_rate": 7e-06, + "loss": 0.1656, + "step": 456 + }, + { + "epoch": 0.5772935417653561, + "grad_norm": 0.08430968970060349, + "learning_rate": 7e-06, + "loss": 0.201, + "step": 457 + }, + { + "epoch": 0.5785567661455866, + "grad_norm": 0.09361864626407623, + "learning_rate": 7e-06, + "loss": 0.2034, + "step": 458 + }, + { + "epoch": 0.5798199905258171, + "grad_norm": 0.09698746353387833, + "learning_rate": 7e-06, + "loss": 0.2444, + "step": 459 + }, + { + "epoch": 0.5810832149060476, + "grad_norm": 0.16023226082324982, + "learning_rate": 7e-06, + "loss": 0.19, + "step": 460 + }, + { + "epoch": 0.5823464392862783, + "grad_norm": 0.08157742023468018, + "learning_rate": 7e-06, + "loss": 0.2149, + "step": 461 + }, + { + "epoch": 0.5836096636665088, + "grad_norm": 0.09342104941606522, + "learning_rate": 7e-06, + "loss": 0.2022, + "step": 462 + }, + { + "epoch": 0.5848728880467393, + "grad_norm": 0.07538167387247086, + "learning_rate": 7e-06, + "loss": 0.2516, + "step": 463 + }, + { + "epoch": 0.5861361124269698, + "grad_norm": 0.12720584869384766, + "learning_rate": 7e-06, + "loss": 0.2208, + "step": 464 + }, + { + "epoch": 0.5873993368072004, + "grad_norm": 0.08613109588623047, + "learning_rate": 7e-06, + "loss": 0.2193, + "step": 465 + }, + { + "epoch": 0.5886625611874309, + "grad_norm": 0.08249358087778091, + "learning_rate": 7e-06, + "loss": 0.1676, + "step": 466 + }, + { + "epoch": 0.5899257855676615, + "grad_norm": 0.1288759857416153, + "learning_rate": 7e-06, + "loss": 0.3238, + "step": 467 + }, + { + "epoch": 0.591189009947892, + "grad_norm": 0.05164247751235962, + "learning_rate": 7e-06, + "loss": 0.1418, + "step": 468 + }, + { + "epoch": 0.5924522343281226, + "grad_norm": 0.0994580090045929, + "learning_rate": 7e-06, + "loss": 0.2238, + "step": 469 + }, + { + "epoch": 0.5937154587083531, + "grad_norm": 0.09253129363059998, + "learning_rate": 7e-06, + "loss": 0.2006, + "step": 470 + }, + { + "epoch": 0.5949786830885836, + "grad_norm": 0.12927457690238953, + "learning_rate": 7e-06, + "loss": 0.2027, + "step": 471 + }, + { + "epoch": 0.5962419074688141, + "grad_norm": 0.11283280700445175, + "learning_rate": 7e-06, + "loss": 0.2169, + "step": 472 + }, + { + "epoch": 0.5975051318490446, + "grad_norm": 0.174880713224411, + "learning_rate": 7e-06, + "loss": 0.1746, + "step": 473 + }, + { + "epoch": 0.5987683562292753, + "grad_norm": 0.11614017933607101, + "learning_rate": 7e-06, + "loss": 0.2968, + "step": 474 + }, + { + "epoch": 0.6000315806095058, + "grad_norm": 0.07818127423524857, + "learning_rate": 7e-06, + "loss": 0.1643, + "step": 475 + }, + { + "epoch": 0.6012948049897363, + "grad_norm": 0.08300397545099258, + "learning_rate": 7e-06, + "loss": 0.1712, + "step": 476 + }, + { + "epoch": 0.6025580293699668, + "grad_norm": 0.1014489009976387, + "learning_rate": 7e-06, + "loss": 0.2089, + "step": 477 + }, + { + "epoch": 0.6038212537501974, + "grad_norm": 0.11591055244207382, + "learning_rate": 7e-06, + "loss": 0.2231, + "step": 478 + }, + { + "epoch": 0.6050844781304279, + "grad_norm": 0.13137224316596985, + "learning_rate": 7e-06, + "loss": 0.1844, + "step": 479 + }, + { + "epoch": 0.6063477025106585, + "grad_norm": 0.09693000465631485, + "learning_rate": 7e-06, + "loss": 0.2033, + "step": 480 + }, + { + "epoch": 0.607610926890889, + "grad_norm": 0.1250012218952179, + "learning_rate": 7e-06, + "loss": 0.2219, + "step": 481 + }, + { + "epoch": 0.6088741512711195, + "grad_norm": 0.18828216195106506, + "learning_rate": 7e-06, + "loss": 0.2427, + "step": 482 + }, + { + "epoch": 0.6101373756513501, + "grad_norm": 0.11057613044977188, + "learning_rate": 7e-06, + "loss": 0.2193, + "step": 483 + }, + { + "epoch": 0.6114006000315806, + "grad_norm": 0.15523040294647217, + "learning_rate": 7e-06, + "loss": 0.1883, + "step": 484 + }, + { + "epoch": 0.6126638244118111, + "grad_norm": 0.16174635291099548, + "learning_rate": 7e-06, + "loss": 0.2067, + "step": 485 + }, + { + "epoch": 0.6139270487920417, + "grad_norm": 0.2738276422023773, + "learning_rate": 7e-06, + "loss": 0.2292, + "step": 486 + }, + { + "epoch": 0.6151902731722723, + "grad_norm": 0.05995164066553116, + "learning_rate": 7e-06, + "loss": 0.1956, + "step": 487 + }, + { + "epoch": 0.6164534975525028, + "grad_norm": 0.05519471690058708, + "learning_rate": 7e-06, + "loss": 0.1501, + "step": 488 + }, + { + "epoch": 0.6177167219327333, + "grad_norm": 0.08133929967880249, + "learning_rate": 7e-06, + "loss": 0.2224, + "step": 489 + }, + { + "epoch": 0.6189799463129638, + "grad_norm": 0.12239203602075577, + "learning_rate": 7e-06, + "loss": 0.2503, + "step": 490 + }, + { + "epoch": 0.6202431706931943, + "grad_norm": 0.11004896461963654, + "learning_rate": 7e-06, + "loss": 0.2796, + "step": 491 + }, + { + "epoch": 0.6215063950734249, + "grad_norm": 0.1722228229045868, + "learning_rate": 7e-06, + "loss": 0.1807, + "step": 492 + }, + { + "epoch": 0.6227696194536555, + "grad_norm": 0.10695190727710724, + "learning_rate": 7e-06, + "loss": 0.1782, + "step": 493 + }, + { + "epoch": 0.624032843833886, + "grad_norm": 0.08578750491142273, + "learning_rate": 7e-06, + "loss": 0.2389, + "step": 494 + }, + { + "epoch": 0.6252960682141165, + "grad_norm": 0.10670057684183121, + "learning_rate": 7e-06, + "loss": 0.1801, + "step": 495 + }, + { + "epoch": 0.6265592925943471, + "grad_norm": 0.04314388707280159, + "learning_rate": 7e-06, + "loss": 0.1366, + "step": 496 + }, + { + "epoch": 0.6278225169745776, + "grad_norm": 0.11937737464904785, + "learning_rate": 7e-06, + "loss": 0.2017, + "step": 497 + }, + { + "epoch": 0.6290857413548081, + "grad_norm": 0.08274619281291962, + "learning_rate": 7e-06, + "loss": 0.1721, + "step": 498 + }, + { + "epoch": 0.6303489657350387, + "grad_norm": 0.07380262762308121, + "learning_rate": 7e-06, + "loss": 0.1962, + "step": 499 + }, + { + "epoch": 0.6316121901152693, + "grad_norm": 0.04727354645729065, + "learning_rate": 7e-06, + "loss": 0.1509, + "step": 500 + }, + { + "epoch": 0.6316121901152693, + "eval_correct_accuracy": 0.7056827820186599, + "eval_error_accuracy": 0.4376407023863125, + "eval_f1": 0.5402416946684214, + "eval_loss": 0.362473726272583, + "eval_runtime": 35.262, + "eval_samples_per_second": 96.421, + "eval_steps_per_second": 6.04, + "step": 500 + }, + { + "epoch": 0.6328754144954998, + "grad_norm": 0.08614058047533035, + "learning_rate": 7e-06, + "loss": 0.1706, + "step": 501 + }, + { + "epoch": 0.6341386388757303, + "grad_norm": 0.06968270987272263, + "learning_rate": 7e-06, + "loss": 0.2028, + "step": 502 + }, + { + "epoch": 0.6354018632559608, + "grad_norm": 0.13758571445941925, + "learning_rate": 7e-06, + "loss": 0.2296, + "step": 503 + }, + { + "epoch": 0.6366650876361913, + "grad_norm": 0.07083171606063843, + "learning_rate": 7e-06, + "loss": 0.1807, + "step": 504 + }, + { + "epoch": 0.637928312016422, + "grad_norm": 0.06689167022705078, + "learning_rate": 7e-06, + "loss": 0.1997, + "step": 505 + }, + { + "epoch": 0.6391915363966525, + "grad_norm": 0.07969733327627182, + "learning_rate": 7e-06, + "loss": 0.1803, + "step": 506 + }, + { + "epoch": 0.640454760776883, + "grad_norm": 0.095677949488163, + "learning_rate": 7e-06, + "loss": 0.1874, + "step": 507 + }, + { + "epoch": 0.6417179851571135, + "grad_norm": 0.10759231448173523, + "learning_rate": 7e-06, + "loss": 0.1656, + "step": 508 + }, + { + "epoch": 0.6429812095373441, + "grad_norm": 0.13282425701618195, + "learning_rate": 7e-06, + "loss": 0.2538, + "step": 509 + }, + { + "epoch": 0.6442444339175746, + "grad_norm": 0.09664168953895569, + "learning_rate": 7e-06, + "loss": 0.1768, + "step": 510 + }, + { + "epoch": 0.6455076582978051, + "grad_norm": 0.11897934973239899, + "learning_rate": 7e-06, + "loss": 0.2023, + "step": 511 + }, + { + "epoch": 0.6467708826780357, + "grad_norm": 0.05450622737407684, + "learning_rate": 7e-06, + "loss": 0.1277, + "step": 512 + }, + { + "epoch": 0.6480341070582663, + "grad_norm": 0.07977665215730667, + "learning_rate": 7e-06, + "loss": 0.2231, + "step": 513 + }, + { + "epoch": 0.6492973314384968, + "grad_norm": 0.19492259621620178, + "learning_rate": 7e-06, + "loss": 0.2253, + "step": 514 + }, + { + "epoch": 0.6505605558187273, + "grad_norm": 0.09466379135847092, + "learning_rate": 7e-06, + "loss": 0.3611, + "step": 515 + }, + { + "epoch": 0.6518237801989578, + "grad_norm": 0.17244236171245575, + "learning_rate": 7e-06, + "loss": 0.2149, + "step": 516 + }, + { + "epoch": 0.6530870045791883, + "grad_norm": 0.08291974663734436, + "learning_rate": 7e-06, + "loss": 0.1848, + "step": 517 + }, + { + "epoch": 0.654350228959419, + "grad_norm": 0.06109621748328209, + "learning_rate": 7e-06, + "loss": 0.1435, + "step": 518 + }, + { + "epoch": 0.6556134533396495, + "grad_norm": 0.06171726807951927, + "learning_rate": 7e-06, + "loss": 0.1749, + "step": 519 + }, + { + "epoch": 0.65687667771988, + "grad_norm": 0.09645943343639374, + "learning_rate": 7e-06, + "loss": 0.197, + "step": 520 + }, + { + "epoch": 0.6581399021001105, + "grad_norm": 0.09050124883651733, + "learning_rate": 7e-06, + "loss": 0.1609, + "step": 521 + }, + { + "epoch": 0.659403126480341, + "grad_norm": 0.09600576758384705, + "learning_rate": 7e-06, + "loss": 0.1826, + "step": 522 + }, + { + "epoch": 0.6606663508605716, + "grad_norm": 0.1261880248785019, + "learning_rate": 7e-06, + "loss": 0.1875, + "step": 523 + }, + { + "epoch": 0.6619295752408021, + "grad_norm": 0.13587896525859833, + "learning_rate": 7e-06, + "loss": 0.1905, + "step": 524 + }, + { + "epoch": 0.6631927996210327, + "grad_norm": 0.12359704077243805, + "learning_rate": 7e-06, + "loss": 0.2087, + "step": 525 + }, + { + "epoch": 0.6644560240012632, + "grad_norm": 0.10092345625162125, + "learning_rate": 7e-06, + "loss": 0.2041, + "step": 526 + }, + { + "epoch": 0.6657192483814938, + "grad_norm": 0.12595926225185394, + "learning_rate": 7e-06, + "loss": 0.1928, + "step": 527 + }, + { + "epoch": 0.6669824727617243, + "grad_norm": 0.08753985911607742, + "learning_rate": 7e-06, + "loss": 0.1444, + "step": 528 + }, + { + "epoch": 0.6682456971419548, + "grad_norm": 0.08193645626306534, + "learning_rate": 7e-06, + "loss": 0.1545, + "step": 529 + }, + { + "epoch": 0.6695089215221853, + "grad_norm": 0.07170840352773666, + "learning_rate": 7e-06, + "loss": 0.1652, + "step": 530 + }, + { + "epoch": 0.670772145902416, + "grad_norm": 0.18759992718696594, + "learning_rate": 7e-06, + "loss": 0.232, + "step": 531 + }, + { + "epoch": 0.6720353702826465, + "grad_norm": 0.13691110908985138, + "learning_rate": 7e-06, + "loss": 0.1905, + "step": 532 + }, + { + "epoch": 0.673298594662877, + "grad_norm": 0.06453829258680344, + "learning_rate": 7e-06, + "loss": 0.2283, + "step": 533 + }, + { + "epoch": 0.6745618190431075, + "grad_norm": 0.12694236636161804, + "learning_rate": 7e-06, + "loss": 0.2013, + "step": 534 + }, + { + "epoch": 0.675825043423338, + "grad_norm": 0.06403839588165283, + "learning_rate": 7e-06, + "loss": 0.1585, + "step": 535 + }, + { + "epoch": 0.6770882678035686, + "grad_norm": 0.13636727631092072, + "learning_rate": 7e-06, + "loss": 0.2156, + "step": 536 + }, + { + "epoch": 0.6783514921837992, + "grad_norm": 0.12285730242729187, + "learning_rate": 7e-06, + "loss": 0.1967, + "step": 537 + }, + { + "epoch": 0.6796147165640297, + "grad_norm": 0.0780211091041565, + "learning_rate": 7e-06, + "loss": 0.1751, + "step": 538 + }, + { + "epoch": 0.6808779409442602, + "grad_norm": 0.09688100218772888, + "learning_rate": 7e-06, + "loss": 0.2141, + "step": 539 + }, + { + "epoch": 0.6821411653244908, + "grad_norm": 0.07864505052566528, + "learning_rate": 7e-06, + "loss": 0.2138, + "step": 540 + }, + { + "epoch": 0.6834043897047213, + "grad_norm": 0.060981281101703644, + "learning_rate": 7e-06, + "loss": 0.187, + "step": 541 + }, + { + "epoch": 0.6846676140849518, + "grad_norm": 0.06510937958955765, + "learning_rate": 7e-06, + "loss": 0.1771, + "step": 542 + }, + { + "epoch": 0.6859308384651823, + "grad_norm": 0.07638704031705856, + "learning_rate": 7e-06, + "loss": 0.2129, + "step": 543 + }, + { + "epoch": 0.687194062845413, + "grad_norm": 0.11518476903438568, + "learning_rate": 7e-06, + "loss": 0.1854, + "step": 544 + }, + { + "epoch": 0.6884572872256435, + "grad_norm": 0.06868738681077957, + "learning_rate": 7e-06, + "loss": 0.1581, + "step": 545 + }, + { + "epoch": 0.689720511605874, + "grad_norm": 0.09059899300336838, + "learning_rate": 7e-06, + "loss": 0.2245, + "step": 546 + }, + { + "epoch": 0.6909837359861045, + "grad_norm": 0.06422233581542969, + "learning_rate": 7e-06, + "loss": 0.1559, + "step": 547 + }, + { + "epoch": 0.692246960366335, + "grad_norm": 0.10189103335142136, + "learning_rate": 7e-06, + "loss": 0.193, + "step": 548 + }, + { + "epoch": 0.6935101847465656, + "grad_norm": 0.08199501782655716, + "learning_rate": 7e-06, + "loss": 0.1908, + "step": 549 + }, + { + "epoch": 0.6947734091267962, + "grad_norm": 0.20546898245811462, + "learning_rate": 7e-06, + "loss": 0.2011, + "step": 550 + }, + { + "epoch": 0.6960366335070267, + "grad_norm": 0.14664340019226074, + "learning_rate": 7e-06, + "loss": 0.1892, + "step": 551 + }, + { + "epoch": 0.6972998578872572, + "grad_norm": 0.08695843815803528, + "learning_rate": 7e-06, + "loss": 0.1871, + "step": 552 + }, + { + "epoch": 0.6985630822674878, + "grad_norm": 0.08112246543169022, + "learning_rate": 7e-06, + "loss": 0.1616, + "step": 553 + }, + { + "epoch": 0.6998263066477183, + "grad_norm": 0.08381661772727966, + "learning_rate": 7e-06, + "loss": 0.2231, + "step": 554 + }, + { + "epoch": 0.7010895310279488, + "grad_norm": 0.09177428483963013, + "learning_rate": 7e-06, + "loss": 0.1956, + "step": 555 + }, + { + "epoch": 0.7023527554081794, + "grad_norm": 0.08766631782054901, + "learning_rate": 7e-06, + "loss": 0.1871, + "step": 556 + }, + { + "epoch": 0.70361597978841, + "grad_norm": 0.07755694538354874, + "learning_rate": 7e-06, + "loss": 0.2718, + "step": 557 + }, + { + "epoch": 0.7048792041686405, + "grad_norm": 0.08710070699453354, + "learning_rate": 7e-06, + "loss": 0.1909, + "step": 558 + }, + { + "epoch": 0.706142428548871, + "grad_norm": 0.07648595422506332, + "learning_rate": 7e-06, + "loss": 0.1948, + "step": 559 + }, + { + "epoch": 0.7074056529291015, + "grad_norm": 0.10871299356222153, + "learning_rate": 7e-06, + "loss": 0.2093, + "step": 560 + }, + { + "epoch": 0.708668877309332, + "grad_norm": 0.07032714784145355, + "learning_rate": 7e-06, + "loss": 0.1699, + "step": 561 + }, + { + "epoch": 0.7099321016895626, + "grad_norm": 0.0873897522687912, + "learning_rate": 7e-06, + "loss": 0.1372, + "step": 562 + }, + { + "epoch": 0.7111953260697932, + "grad_norm": 0.07188841700553894, + "learning_rate": 7e-06, + "loss": 0.1794, + "step": 563 + }, + { + "epoch": 0.7124585504500237, + "grad_norm": 0.07733464986085892, + "learning_rate": 7e-06, + "loss": 0.2043, + "step": 564 + }, + { + "epoch": 0.7137217748302542, + "grad_norm": 0.07270821928977966, + "learning_rate": 7e-06, + "loss": 0.177, + "step": 565 + }, + { + "epoch": 0.7149849992104848, + "grad_norm": 0.1570441722869873, + "learning_rate": 7e-06, + "loss": 0.2411, + "step": 566 + }, + { + "epoch": 0.7162482235907153, + "grad_norm": 0.2707260549068451, + "learning_rate": 7e-06, + "loss": 0.2307, + "step": 567 + }, + { + "epoch": 0.7175114479709458, + "grad_norm": 0.07656281441450119, + "learning_rate": 7e-06, + "loss": 0.1759, + "step": 568 + }, + { + "epoch": 0.7187746723511764, + "grad_norm": 0.09973770380020142, + "learning_rate": 7e-06, + "loss": 0.1965, + "step": 569 + }, + { + "epoch": 0.720037896731407, + "grad_norm": 0.06791306287050247, + "learning_rate": 7e-06, + "loss": 0.1749, + "step": 570 + }, + { + "epoch": 0.7213011211116375, + "grad_norm": 0.17801041901111603, + "learning_rate": 7e-06, + "loss": 0.1941, + "step": 571 + }, + { + "epoch": 0.722564345491868, + "grad_norm": 0.18452543020248413, + "learning_rate": 7e-06, + "loss": 0.2446, + "step": 572 + }, + { + "epoch": 0.7238275698720985, + "grad_norm": 0.12178942561149597, + "learning_rate": 7e-06, + "loss": 0.1583, + "step": 573 + }, + { + "epoch": 0.725090794252329, + "grad_norm": 0.13167473673820496, + "learning_rate": 7e-06, + "loss": 0.2324, + "step": 574 + }, + { + "epoch": 0.7263540186325597, + "grad_norm": 0.05255408585071564, + "learning_rate": 7e-06, + "loss": 0.195, + "step": 575 + }, + { + "epoch": 0.7276172430127902, + "grad_norm": 0.09154222905635834, + "learning_rate": 7e-06, + "loss": 0.1878, + "step": 576 + }, + { + "epoch": 0.7288804673930207, + "grad_norm": 0.0887879729270935, + "learning_rate": 7e-06, + "loss": 0.2052, + "step": 577 + }, + { + "epoch": 0.7301436917732512, + "grad_norm": 0.1336040198802948, + "learning_rate": 7e-06, + "loss": 0.1952, + "step": 578 + }, + { + "epoch": 0.7314069161534817, + "grad_norm": 0.08207479119300842, + "learning_rate": 7e-06, + "loss": 0.185, + "step": 579 + }, + { + "epoch": 0.7326701405337123, + "grad_norm": 0.05941140279173851, + "learning_rate": 7e-06, + "loss": 0.204, + "step": 580 + }, + { + "epoch": 0.7339333649139428, + "grad_norm": 0.06899949908256531, + "learning_rate": 7e-06, + "loss": 0.1408, + "step": 581 + }, + { + "epoch": 0.7351965892941734, + "grad_norm": 0.09259360283613205, + "learning_rate": 7e-06, + "loss": 0.1823, + "step": 582 + }, + { + "epoch": 0.7364598136744039, + "grad_norm": 0.1346062421798706, + "learning_rate": 7e-06, + "loss": 0.2174, + "step": 583 + }, + { + "epoch": 0.7377230380546345, + "grad_norm": 0.1547420769929886, + "learning_rate": 7e-06, + "loss": 0.1864, + "step": 584 + }, + { + "epoch": 0.738986262434865, + "grad_norm": 0.10551164299249649, + "learning_rate": 7e-06, + "loss": 0.1554, + "step": 585 + }, + { + "epoch": 0.7402494868150955, + "grad_norm": 0.08826129138469696, + "learning_rate": 7e-06, + "loss": 0.2044, + "step": 586 + }, + { + "epoch": 0.741512711195326, + "grad_norm": 0.07170785963535309, + "learning_rate": 7e-06, + "loss": 0.1777, + "step": 587 + }, + { + "epoch": 0.7427759355755567, + "grad_norm": 0.1085812896490097, + "learning_rate": 7e-06, + "loss": 0.2822, + "step": 588 + }, + { + "epoch": 0.7440391599557872, + "grad_norm": 0.08545360714197159, + "learning_rate": 7e-06, + "loss": 0.19, + "step": 589 + }, + { + "epoch": 0.7453023843360177, + "grad_norm": 0.05576294660568237, + "learning_rate": 7e-06, + "loss": 0.1826, + "step": 590 + }, + { + "epoch": 0.7465656087162482, + "grad_norm": 0.056626636534929276, + "learning_rate": 7e-06, + "loss": 0.2191, + "step": 591 + }, + { + "epoch": 0.7478288330964787, + "grad_norm": 0.06961087882518768, + "learning_rate": 7e-06, + "loss": 0.2004, + "step": 592 + }, + { + "epoch": 0.7490920574767093, + "grad_norm": 0.09317582845687866, + "learning_rate": 7e-06, + "loss": 0.1465, + "step": 593 + }, + { + "epoch": 0.7503552818569399, + "grad_norm": 0.13993658125400543, + "learning_rate": 7e-06, + "loss": 0.1876, + "step": 594 + }, + { + "epoch": 0.7516185062371704, + "grad_norm": 0.06080286204814911, + "learning_rate": 7e-06, + "loss": 0.2012, + "step": 595 + }, + { + "epoch": 0.7528817306174009, + "grad_norm": 0.060514189302921295, + "learning_rate": 7e-06, + "loss": 0.1658, + "step": 596 + }, + { + "epoch": 0.7541449549976315, + "grad_norm": 0.09004813432693481, + "learning_rate": 7e-06, + "loss": 0.3195, + "step": 597 + }, + { + "epoch": 0.755408179377862, + "grad_norm": 0.07283802330493927, + "learning_rate": 7e-06, + "loss": 0.2021, + "step": 598 + }, + { + "epoch": 0.7566714037580925, + "grad_norm": 0.08824078738689423, + "learning_rate": 7e-06, + "loss": 0.1941, + "step": 599 + }, + { + "epoch": 0.757934628138323, + "grad_norm": 0.09339006245136261, + "learning_rate": 7e-06, + "loss": 0.4697, + "step": 600 + }, + { + "epoch": 0.757934628138323, + "eval_correct_accuracy": 0.6098388464800678, + "eval_error_accuracy": 0.48491670418730304, + "eval_f1": 0.540250366102693, + "eval_loss": 0.42047378420829773, + "eval_runtime": 35.1541, + "eval_samples_per_second": 96.717, + "eval_steps_per_second": 6.059, + "step": 600 + } + ], + "logging_steps": 1.0, + "max_steps": 791, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.7774891537268736e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}