{ "best_metric": null, "best_model_checkpoint": null, "epoch": 20.0, "eval_steps": 500, "global_step": 5060, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003952569169960474, "grad_norm": 6.42405891418457, "learning_rate": 5e-06, "loss": 3.0263, "step": 1 }, { "epoch": 0.007905138339920948, "grad_norm": 5.687948226928711, "learning_rate": 1e-05, "loss": 2.9982, "step": 2 }, { "epoch": 0.011857707509881422, "grad_norm": 8.810317993164062, "learning_rate": 1.5e-05, "loss": 3.2613, "step": 3 }, { "epoch": 0.015810276679841896, "grad_norm": 3.873013496398926, "learning_rate": 2e-05, "loss": 2.8407, "step": 4 }, { "epoch": 0.019762845849802372, "grad_norm": 3.863713264465332, "learning_rate": 2.5e-05, "loss": 2.5842, "step": 5 }, { "epoch": 0.023715415019762844, "grad_norm": 3.931561231613159, "learning_rate": 3e-05, "loss": 2.8448, "step": 6 }, { "epoch": 0.02766798418972332, "grad_norm": 3.4672374725341797, "learning_rate": 3.5e-05, "loss": 2.507, "step": 7 }, { "epoch": 0.03162055335968379, "grad_norm": 3.668203830718994, "learning_rate": 4e-05, "loss": 2.3853, "step": 8 }, { "epoch": 0.03557312252964427, "grad_norm": 3.338822603225708, "learning_rate": 4.5e-05, "loss": 2.3706, "step": 9 }, { "epoch": 0.039525691699604744, "grad_norm": 3.353273391723633, "learning_rate": 5e-05, "loss": 2.5131, "step": 10 }, { "epoch": 0.043478260869565216, "grad_norm": 3.8022966384887695, "learning_rate": 4.9990099009900993e-05, "loss": 2.4314, "step": 11 }, { "epoch": 0.04743083003952569, "grad_norm": 2.1579272747039795, "learning_rate": 4.9980198019801985e-05, "loss": 2.3252, "step": 12 }, { "epoch": 0.05138339920948617, "grad_norm": 2.2012343406677246, "learning_rate": 4.9970297029702976e-05, "loss": 2.5036, "step": 13 }, { "epoch": 0.05533596837944664, "grad_norm": 2.029061794281006, "learning_rate": 4.996039603960396e-05, "loss": 2.5409, "step": 14 }, { "epoch": 0.05928853754940711, "grad_norm": 2.6418192386627197, "learning_rate": 4.995049504950495e-05, "loss": 2.5395, "step": 15 }, { "epoch": 0.06324110671936758, "grad_norm": 2.637308120727539, "learning_rate": 4.994059405940594e-05, "loss": 2.2763, "step": 16 }, { "epoch": 0.06719367588932806, "grad_norm": 1.806755781173706, "learning_rate": 4.993069306930693e-05, "loss": 2.3699, "step": 17 }, { "epoch": 0.07114624505928854, "grad_norm": 2.0135483741760254, "learning_rate": 4.9920792079207924e-05, "loss": 2.411, "step": 18 }, { "epoch": 0.07509881422924901, "grad_norm": 1.7150806188583374, "learning_rate": 4.9910891089108915e-05, "loss": 2.2789, "step": 19 }, { "epoch": 0.07905138339920949, "grad_norm": 2.1692118644714355, "learning_rate": 4.9900990099009906e-05, "loss": 2.2766, "step": 20 }, { "epoch": 0.08300395256916997, "grad_norm": 1.9280359745025635, "learning_rate": 4.989108910891089e-05, "loss": 2.5211, "step": 21 }, { "epoch": 0.08695652173913043, "grad_norm": 2.562676191329956, "learning_rate": 4.988118811881188e-05, "loss": 2.5064, "step": 22 }, { "epoch": 0.09090909090909091, "grad_norm": 2.229430913925171, "learning_rate": 4.987128712871287e-05, "loss": 2.2737, "step": 23 }, { "epoch": 0.09486166007905138, "grad_norm": 1.9107154607772827, "learning_rate": 4.9861386138613864e-05, "loss": 2.2041, "step": 24 }, { "epoch": 0.09881422924901186, "grad_norm": 5.523063659667969, "learning_rate": 4.9851485148514855e-05, "loss": 2.3345, "step": 25 }, { "epoch": 0.10276679841897234, "grad_norm": 2.5889744758605957, "learning_rate": 4.9841584158415846e-05, "loss": 2.38, "step": 26 }, { "epoch": 0.1067193675889328, "grad_norm": 2.3176913261413574, "learning_rate": 4.983168316831684e-05, "loss": 2.4227, "step": 27 }, { "epoch": 0.11067193675889328, "grad_norm": 2.0367422103881836, "learning_rate": 4.982178217821782e-05, "loss": 2.1927, "step": 28 }, { "epoch": 0.11462450592885376, "grad_norm": 2.2616076469421387, "learning_rate": 4.981188118811881e-05, "loss": 2.0993, "step": 29 }, { "epoch": 0.11857707509881422, "grad_norm": 2.0256075859069824, "learning_rate": 4.98019801980198e-05, "loss": 2.0463, "step": 30 }, { "epoch": 0.1225296442687747, "grad_norm": 2.696810483932495, "learning_rate": 4.9792079207920794e-05, "loss": 2.0585, "step": 31 }, { "epoch": 0.12648221343873517, "grad_norm": 2.203913450241089, "learning_rate": 4.9782178217821786e-05, "loss": 2.0411, "step": 32 }, { "epoch": 0.13043478260869565, "grad_norm": 2.674555540084839, "learning_rate": 4.9772277227722777e-05, "loss": 1.9486, "step": 33 }, { "epoch": 0.13438735177865613, "grad_norm": 2.1000349521636963, "learning_rate": 4.976237623762377e-05, "loss": 2.3206, "step": 34 }, { "epoch": 0.1383399209486166, "grad_norm": 2.147045135498047, "learning_rate": 4.975247524752475e-05, "loss": 2.0522, "step": 35 }, { "epoch": 0.1422924901185771, "grad_norm": 2.577526330947876, "learning_rate": 4.974257425742574e-05, "loss": 2.1733, "step": 36 }, { "epoch": 0.14624505928853754, "grad_norm": 1.8023768663406372, "learning_rate": 4.9732673267326734e-05, "loss": 2.1481, "step": 37 }, { "epoch": 0.15019762845849802, "grad_norm": 1.824354887008667, "learning_rate": 4.9722772277227725e-05, "loss": 2.3413, "step": 38 }, { "epoch": 0.1541501976284585, "grad_norm": 1.9494798183441162, "learning_rate": 4.9712871287128716e-05, "loss": 1.965, "step": 39 }, { "epoch": 0.15810276679841898, "grad_norm": 2.411818265914917, "learning_rate": 4.970297029702971e-05, "loss": 2.4337, "step": 40 }, { "epoch": 0.16205533596837945, "grad_norm": 2.390172243118286, "learning_rate": 4.96930693069307e-05, "loss": 1.9862, "step": 41 }, { "epoch": 0.16600790513833993, "grad_norm": 2.2208991050720215, "learning_rate": 4.968316831683168e-05, "loss": 2.0316, "step": 42 }, { "epoch": 0.16996047430830039, "grad_norm": 1.8159395456314087, "learning_rate": 4.9673267326732674e-05, "loss": 2.0531, "step": 43 }, { "epoch": 0.17391304347826086, "grad_norm": 2.0326318740844727, "learning_rate": 4.9663366336633665e-05, "loss": 2.3082, "step": 44 }, { "epoch": 0.17786561264822134, "grad_norm": 1.9229705333709717, "learning_rate": 4.9653465346534656e-05, "loss": 2.0981, "step": 45 }, { "epoch": 0.18181818181818182, "grad_norm": 1.871531367301941, "learning_rate": 4.964356435643565e-05, "loss": 2.128, "step": 46 }, { "epoch": 0.1857707509881423, "grad_norm": 3.107696771621704, "learning_rate": 4.963366336633664e-05, "loss": 2.1601, "step": 47 }, { "epoch": 0.18972332015810275, "grad_norm": 1.942887544631958, "learning_rate": 4.962376237623763e-05, "loss": 2.0642, "step": 48 }, { "epoch": 0.19367588932806323, "grad_norm": 1.752061128616333, "learning_rate": 4.961386138613861e-05, "loss": 2.2405, "step": 49 }, { "epoch": 0.1976284584980237, "grad_norm": 2.037332057952881, "learning_rate": 4.9603960396039604e-05, "loss": 2.0954, "step": 50 }, { "epoch": 0.2015810276679842, "grad_norm": 2.211359977722168, "learning_rate": 4.9594059405940595e-05, "loss": 1.8162, "step": 51 }, { "epoch": 0.20553359683794467, "grad_norm": 2.0617964267730713, "learning_rate": 4.9584158415841587e-05, "loss": 2.0327, "step": 52 }, { "epoch": 0.20948616600790515, "grad_norm": 2.295633316040039, "learning_rate": 4.957425742574258e-05, "loss": 2.1788, "step": 53 }, { "epoch": 0.2134387351778656, "grad_norm": 1.6934067010879517, "learning_rate": 4.956435643564357e-05, "loss": 2.1971, "step": 54 }, { "epoch": 0.21739130434782608, "grad_norm": 1.859398603439331, "learning_rate": 4.955445544554456e-05, "loss": 2.0339, "step": 55 }, { "epoch": 0.22134387351778656, "grad_norm": 2.26220965385437, "learning_rate": 4.9544554455445544e-05, "loss": 1.8617, "step": 56 }, { "epoch": 0.22529644268774704, "grad_norm": 2.23677921295166, "learning_rate": 4.9534653465346535e-05, "loss": 2.0758, "step": 57 }, { "epoch": 0.22924901185770752, "grad_norm": 1.780326247215271, "learning_rate": 4.9524752475247526e-05, "loss": 2.2487, "step": 58 }, { "epoch": 0.233201581027668, "grad_norm": 1.9278745651245117, "learning_rate": 4.951485148514852e-05, "loss": 2.3116, "step": 59 }, { "epoch": 0.23715415019762845, "grad_norm": 1.9197441339492798, "learning_rate": 4.950495049504951e-05, "loss": 1.9265, "step": 60 }, { "epoch": 0.24110671936758893, "grad_norm": 2.0942647457122803, "learning_rate": 4.94950495049505e-05, "loss": 1.8434, "step": 61 }, { "epoch": 0.2450592885375494, "grad_norm": 2.295727491378784, "learning_rate": 4.948514851485149e-05, "loss": 2.1543, "step": 62 }, { "epoch": 0.2490118577075099, "grad_norm": 2.050112247467041, "learning_rate": 4.9475247524752475e-05, "loss": 1.6333, "step": 63 }, { "epoch": 0.25296442687747034, "grad_norm": 1.726313591003418, "learning_rate": 4.9465346534653466e-05, "loss": 1.7324, "step": 64 }, { "epoch": 0.25691699604743085, "grad_norm": 2.1288046836853027, "learning_rate": 4.945544554455446e-05, "loss": 1.9901, "step": 65 }, { "epoch": 0.2608695652173913, "grad_norm": 2.044255018234253, "learning_rate": 4.944554455445545e-05, "loss": 1.89, "step": 66 }, { "epoch": 0.2648221343873518, "grad_norm": 2.056349039077759, "learning_rate": 4.943564356435644e-05, "loss": 2.1743, "step": 67 }, { "epoch": 0.26877470355731226, "grad_norm": 2.0013623237609863, "learning_rate": 4.942574257425743e-05, "loss": 2.0334, "step": 68 }, { "epoch": 0.2727272727272727, "grad_norm": 2.5802996158599854, "learning_rate": 4.9415841584158414e-05, "loss": 1.7817, "step": 69 }, { "epoch": 0.2766798418972332, "grad_norm": 1.9824113845825195, "learning_rate": 4.9405940594059405e-05, "loss": 1.8911, "step": 70 }, { "epoch": 0.28063241106719367, "grad_norm": 1.8202821016311646, "learning_rate": 4.9396039603960396e-05, "loss": 1.9153, "step": 71 }, { "epoch": 0.2845849802371542, "grad_norm": 2.5200858116149902, "learning_rate": 4.938613861386139e-05, "loss": 1.8586, "step": 72 }, { "epoch": 0.2885375494071146, "grad_norm": 2.2708740234375, "learning_rate": 4.937623762376238e-05, "loss": 2.051, "step": 73 }, { "epoch": 0.2924901185770751, "grad_norm": 2.0685415267944336, "learning_rate": 4.936633663366337e-05, "loss": 2.0847, "step": 74 }, { "epoch": 0.2964426877470356, "grad_norm": 1.9750455617904663, "learning_rate": 4.935643564356436e-05, "loss": 1.8515, "step": 75 }, { "epoch": 0.30039525691699603, "grad_norm": 2.1509861946105957, "learning_rate": 4.9346534653465345e-05, "loss": 2.5348, "step": 76 }, { "epoch": 0.30434782608695654, "grad_norm": 1.5361539125442505, "learning_rate": 4.9336633663366336e-05, "loss": 1.9332, "step": 77 }, { "epoch": 0.308300395256917, "grad_norm": 2.3222951889038086, "learning_rate": 4.932673267326733e-05, "loss": 2.166, "step": 78 }, { "epoch": 0.31225296442687744, "grad_norm": 2.2165160179138184, "learning_rate": 4.931683168316832e-05, "loss": 1.6141, "step": 79 }, { "epoch": 0.31620553359683795, "grad_norm": 1.765684962272644, "learning_rate": 4.930693069306931e-05, "loss": 2.03, "step": 80 }, { "epoch": 0.3201581027667984, "grad_norm": 2.062169075012207, "learning_rate": 4.92970297029703e-05, "loss": 1.9594, "step": 81 }, { "epoch": 0.3241106719367589, "grad_norm": 2.2629172801971436, "learning_rate": 4.928712871287129e-05, "loss": 2.0889, "step": 82 }, { "epoch": 0.32806324110671936, "grad_norm": 2.35288667678833, "learning_rate": 4.9277227722772276e-05, "loss": 1.9283, "step": 83 }, { "epoch": 0.33201581027667987, "grad_norm": 2.1537673473358154, "learning_rate": 4.926732673267327e-05, "loss": 2.1634, "step": 84 }, { "epoch": 0.3359683794466403, "grad_norm": 2.549635410308838, "learning_rate": 4.925742574257426e-05, "loss": 1.7051, "step": 85 }, { "epoch": 0.33992094861660077, "grad_norm": 2.118133544921875, "learning_rate": 4.924752475247525e-05, "loss": 1.8949, "step": 86 }, { "epoch": 0.3438735177865613, "grad_norm": 2.322223424911499, "learning_rate": 4.923762376237624e-05, "loss": 2.2646, "step": 87 }, { "epoch": 0.34782608695652173, "grad_norm": 2.027369260787964, "learning_rate": 4.922772277227723e-05, "loss": 1.7717, "step": 88 }, { "epoch": 0.35177865612648224, "grad_norm": 2.0829505920410156, "learning_rate": 4.921782178217822e-05, "loss": 1.7726, "step": 89 }, { "epoch": 0.3557312252964427, "grad_norm": 1.9348795413970947, "learning_rate": 4.9207920792079206e-05, "loss": 1.8098, "step": 90 }, { "epoch": 0.35968379446640314, "grad_norm": 1.9080532789230347, "learning_rate": 4.91980198019802e-05, "loss": 1.9261, "step": 91 }, { "epoch": 0.36363636363636365, "grad_norm": 2.1999659538269043, "learning_rate": 4.918811881188119e-05, "loss": 1.7742, "step": 92 }, { "epoch": 0.3675889328063241, "grad_norm": 1.7426645755767822, "learning_rate": 4.917821782178218e-05, "loss": 1.8598, "step": 93 }, { "epoch": 0.3715415019762846, "grad_norm": 1.4845582246780396, "learning_rate": 4.916831683168317e-05, "loss": 2.1001, "step": 94 }, { "epoch": 0.37549407114624506, "grad_norm": 1.7559888362884521, "learning_rate": 4.915841584158416e-05, "loss": 2.4173, "step": 95 }, { "epoch": 0.3794466403162055, "grad_norm": 1.8770387172698975, "learning_rate": 4.914851485148515e-05, "loss": 2.022, "step": 96 }, { "epoch": 0.383399209486166, "grad_norm": 1.7226651906967163, "learning_rate": 4.913861386138614e-05, "loss": 2.0759, "step": 97 }, { "epoch": 0.38735177865612647, "grad_norm": 3.655583143234253, "learning_rate": 4.912871287128713e-05, "loss": 1.9627, "step": 98 }, { "epoch": 0.391304347826087, "grad_norm": 1.892067790031433, "learning_rate": 4.911881188118812e-05, "loss": 1.8974, "step": 99 }, { "epoch": 0.3952569169960474, "grad_norm": 2.1055660247802734, "learning_rate": 4.910891089108911e-05, "loss": 1.9498, "step": 100 }, { "epoch": 0.39920948616600793, "grad_norm": 2.046591281890869, "learning_rate": 4.90990099009901e-05, "loss": 1.8152, "step": 101 }, { "epoch": 0.4031620553359684, "grad_norm": 2.086322784423828, "learning_rate": 4.908910891089109e-05, "loss": 1.7347, "step": 102 }, { "epoch": 0.40711462450592883, "grad_norm": 2.5637149810791016, "learning_rate": 4.9079207920792083e-05, "loss": 1.7118, "step": 103 }, { "epoch": 0.41106719367588934, "grad_norm": 1.9537923336029053, "learning_rate": 4.906930693069307e-05, "loss": 1.9423, "step": 104 }, { "epoch": 0.4150197628458498, "grad_norm": 1.8360991477966309, "learning_rate": 4.905940594059406e-05, "loss": 1.9806, "step": 105 }, { "epoch": 0.4189723320158103, "grad_norm": 1.7105575799942017, "learning_rate": 4.904950495049505e-05, "loss": 2.0889, "step": 106 }, { "epoch": 0.42292490118577075, "grad_norm": 2.4374871253967285, "learning_rate": 4.903960396039604e-05, "loss": 1.6598, "step": 107 }, { "epoch": 0.4268774703557312, "grad_norm": 1.699758529663086, "learning_rate": 4.902970297029703e-05, "loss": 1.9162, "step": 108 }, { "epoch": 0.4308300395256917, "grad_norm": 1.9177852869033813, "learning_rate": 4.901980198019802e-05, "loss": 2.1056, "step": 109 }, { "epoch": 0.43478260869565216, "grad_norm": 1.6028285026550293, "learning_rate": 4.9009900990099014e-05, "loss": 2.1706, "step": 110 }, { "epoch": 0.43873517786561267, "grad_norm": 3.0957303047180176, "learning_rate": 4.9e-05, "loss": 1.833, "step": 111 }, { "epoch": 0.4426877470355731, "grad_norm": 2.0074870586395264, "learning_rate": 4.899009900990099e-05, "loss": 2.0696, "step": 112 }, { "epoch": 0.44664031620553357, "grad_norm": 1.4162400960922241, "learning_rate": 4.898019801980198e-05, "loss": 2.0065, "step": 113 }, { "epoch": 0.4505928853754941, "grad_norm": 1.8403061628341675, "learning_rate": 4.897029702970297e-05, "loss": 2.309, "step": 114 }, { "epoch": 0.45454545454545453, "grad_norm": 1.9115809202194214, "learning_rate": 4.896039603960396e-05, "loss": 2.1382, "step": 115 }, { "epoch": 0.45849802371541504, "grad_norm": 1.8471789360046387, "learning_rate": 4.8950495049504954e-05, "loss": 1.9083, "step": 116 }, { "epoch": 0.4624505928853755, "grad_norm": 2.064030647277832, "learning_rate": 4.8940594059405945e-05, "loss": 1.8866, "step": 117 }, { "epoch": 0.466403162055336, "grad_norm": 1.8261295557022095, "learning_rate": 4.893069306930693e-05, "loss": 1.953, "step": 118 }, { "epoch": 0.47035573122529645, "grad_norm": 2.176332712173462, "learning_rate": 4.892079207920792e-05, "loss": 1.813, "step": 119 }, { "epoch": 0.4743083003952569, "grad_norm": 1.7104378938674927, "learning_rate": 4.891089108910891e-05, "loss": 1.9231, "step": 120 }, { "epoch": 0.4782608695652174, "grad_norm": 1.7965821027755737, "learning_rate": 4.89009900990099e-05, "loss": 1.9432, "step": 121 }, { "epoch": 0.48221343873517786, "grad_norm": 1.4174447059631348, "learning_rate": 4.889108910891089e-05, "loss": 1.928, "step": 122 }, { "epoch": 0.48616600790513836, "grad_norm": 1.685515284538269, "learning_rate": 4.8881188118811884e-05, "loss": 1.5771, "step": 123 }, { "epoch": 0.4901185770750988, "grad_norm": 1.905871033668518, "learning_rate": 4.8871287128712875e-05, "loss": 1.9173, "step": 124 }, { "epoch": 0.49407114624505927, "grad_norm": 1.7542794942855835, "learning_rate": 4.886138613861386e-05, "loss": 1.9286, "step": 125 }, { "epoch": 0.4980237154150198, "grad_norm": 2.1139333248138428, "learning_rate": 4.885148514851485e-05, "loss": 1.6032, "step": 126 }, { "epoch": 0.5019762845849802, "grad_norm": 2.06842303276062, "learning_rate": 4.884158415841584e-05, "loss": 2.0378, "step": 127 }, { "epoch": 0.5059288537549407, "grad_norm": 2.1950454711914062, "learning_rate": 4.883168316831683e-05, "loss": 1.5897, "step": 128 }, { "epoch": 0.5098814229249012, "grad_norm": 5.622288703918457, "learning_rate": 4.8821782178217824e-05, "loss": 2.1312, "step": 129 }, { "epoch": 0.5138339920948617, "grad_norm": 1.97512948513031, "learning_rate": 4.8811881188118815e-05, "loss": 1.9469, "step": 130 }, { "epoch": 0.5177865612648221, "grad_norm": 2.136324405670166, "learning_rate": 4.8801980198019806e-05, "loss": 1.7378, "step": 131 }, { "epoch": 0.5217391304347826, "grad_norm": 1.6107057332992554, "learning_rate": 4.879207920792079e-05, "loss": 1.9374, "step": 132 }, { "epoch": 0.525691699604743, "grad_norm": 1.8602632284164429, "learning_rate": 4.878217821782178e-05, "loss": 2.0179, "step": 133 }, { "epoch": 0.5296442687747036, "grad_norm": 1.6500134468078613, "learning_rate": 4.877227722772277e-05, "loss": 1.7701, "step": 134 }, { "epoch": 0.5335968379446641, "grad_norm": 1.7294130325317383, "learning_rate": 4.8762376237623764e-05, "loss": 1.8758, "step": 135 }, { "epoch": 0.5375494071146245, "grad_norm": 1.8450415134429932, "learning_rate": 4.8752475247524755e-05, "loss": 1.8356, "step": 136 }, { "epoch": 0.541501976284585, "grad_norm": 2.019686460494995, "learning_rate": 4.8742574257425746e-05, "loss": 1.9206, "step": 137 }, { "epoch": 0.5454545454545454, "grad_norm": 1.969763994216919, "learning_rate": 4.873267326732674e-05, "loss": 1.5955, "step": 138 }, { "epoch": 0.549407114624506, "grad_norm": 2.2199866771698, "learning_rate": 4.872277227722772e-05, "loss": 1.6791, "step": 139 }, { "epoch": 0.5533596837944664, "grad_norm": 1.751599907875061, "learning_rate": 4.871287128712871e-05, "loss": 1.9097, "step": 140 }, { "epoch": 0.5573122529644269, "grad_norm": 1.666996955871582, "learning_rate": 4.87029702970297e-05, "loss": 1.6382, "step": 141 }, { "epoch": 0.5612648221343873, "grad_norm": 2.3530306816101074, "learning_rate": 4.8693069306930694e-05, "loss": 1.6568, "step": 142 }, { "epoch": 0.5652173913043478, "grad_norm": 1.7121120691299438, "learning_rate": 4.8683168316831685e-05, "loss": 2.0137, "step": 143 }, { "epoch": 0.5691699604743083, "grad_norm": 2.0414509773254395, "learning_rate": 4.8673267326732676e-05, "loss": 1.9287, "step": 144 }, { "epoch": 0.5731225296442688, "grad_norm": 1.821754813194275, "learning_rate": 4.866336633663367e-05, "loss": 1.9478, "step": 145 }, { "epoch": 0.5770750988142292, "grad_norm": 1.4749019145965576, "learning_rate": 4.865346534653465e-05, "loss": 2.009, "step": 146 }, { "epoch": 0.5810276679841897, "grad_norm": 1.8353281021118164, "learning_rate": 4.864356435643564e-05, "loss": 1.9419, "step": 147 }, { "epoch": 0.5849802371541502, "grad_norm": 1.9659324884414673, "learning_rate": 4.8633663366336634e-05, "loss": 2.2665, "step": 148 }, { "epoch": 0.5889328063241107, "grad_norm": 1.934365153312683, "learning_rate": 4.8623762376237625e-05, "loss": 1.776, "step": 149 }, { "epoch": 0.5928853754940712, "grad_norm": 1.6378803253173828, "learning_rate": 4.8613861386138616e-05, "loss": 1.7791, "step": 150 }, { "epoch": 0.5968379446640316, "grad_norm": 1.7188308238983154, "learning_rate": 4.860396039603961e-05, "loss": 1.8526, "step": 151 }, { "epoch": 0.6007905138339921, "grad_norm": 1.9700491428375244, "learning_rate": 4.85940594059406e-05, "loss": 1.5954, "step": 152 }, { "epoch": 0.6047430830039525, "grad_norm": 1.6589596271514893, "learning_rate": 4.858415841584158e-05, "loss": 1.7704, "step": 153 }, { "epoch": 0.6086956521739131, "grad_norm": 1.9736006259918213, "learning_rate": 4.8574257425742574e-05, "loss": 1.7296, "step": 154 }, { "epoch": 0.6126482213438735, "grad_norm": 1.7853156328201294, "learning_rate": 4.8564356435643565e-05, "loss": 1.5596, "step": 155 }, { "epoch": 0.616600790513834, "grad_norm": 1.9807164669036865, "learning_rate": 4.8554455445544556e-05, "loss": 1.6601, "step": 156 }, { "epoch": 0.6205533596837944, "grad_norm": 1.4205448627471924, "learning_rate": 4.854455445544555e-05, "loss": 1.8244, "step": 157 }, { "epoch": 0.6245059288537549, "grad_norm": 1.7971817255020142, "learning_rate": 4.853465346534654e-05, "loss": 1.7217, "step": 158 }, { "epoch": 0.6284584980237155, "grad_norm": 1.9977649450302124, "learning_rate": 4.852475247524753e-05, "loss": 1.8663, "step": 159 }, { "epoch": 0.6324110671936759, "grad_norm": 1.616550326347351, "learning_rate": 4.851485148514851e-05, "loss": 1.8783, "step": 160 }, { "epoch": 0.6363636363636364, "grad_norm": 1.6368743181228638, "learning_rate": 4.8504950495049504e-05, "loss": 2.2251, "step": 161 }, { "epoch": 0.6403162055335968, "grad_norm": 1.9426034688949585, "learning_rate": 4.8495049504950495e-05, "loss": 2.0544, "step": 162 }, { "epoch": 0.6442687747035574, "grad_norm": 1.8865758180618286, "learning_rate": 4.8485148514851486e-05, "loss": 1.6952, "step": 163 }, { "epoch": 0.6482213438735178, "grad_norm": 2.111178159713745, "learning_rate": 4.847524752475248e-05, "loss": 1.888, "step": 164 }, { "epoch": 0.6521739130434783, "grad_norm": 1.9977130889892578, "learning_rate": 4.846534653465347e-05, "loss": 1.583, "step": 165 }, { "epoch": 0.6561264822134387, "grad_norm": 1.732893705368042, "learning_rate": 4.845544554455446e-05, "loss": 1.9761, "step": 166 }, { "epoch": 0.6600790513833992, "grad_norm": 1.585363745689392, "learning_rate": 4.8445544554455444e-05, "loss": 1.8327, "step": 167 }, { "epoch": 0.6640316205533597, "grad_norm": 1.8743613958358765, "learning_rate": 4.8435643564356435e-05, "loss": 1.7943, "step": 168 }, { "epoch": 0.6679841897233202, "grad_norm": 1.9972338676452637, "learning_rate": 4.8425742574257426e-05, "loss": 1.9723, "step": 169 }, { "epoch": 0.6719367588932806, "grad_norm": 1.495535135269165, "learning_rate": 4.841584158415842e-05, "loss": 1.5312, "step": 170 }, { "epoch": 0.6758893280632411, "grad_norm": 2.111295223236084, "learning_rate": 4.840594059405941e-05, "loss": 1.6446, "step": 171 }, { "epoch": 0.6798418972332015, "grad_norm": 2.6367835998535156, "learning_rate": 4.83960396039604e-05, "loss": 1.3652, "step": 172 }, { "epoch": 0.6837944664031621, "grad_norm": 2.19242525100708, "learning_rate": 4.838613861386139e-05, "loss": 1.6927, "step": 173 }, { "epoch": 0.6877470355731226, "grad_norm": 1.8796133995056152, "learning_rate": 4.8376237623762375e-05, "loss": 1.9665, "step": 174 }, { "epoch": 0.691699604743083, "grad_norm": 1.9940913915634155, "learning_rate": 4.8366336633663366e-05, "loss": 1.9458, "step": 175 }, { "epoch": 0.6956521739130435, "grad_norm": 1.6950244903564453, "learning_rate": 4.835643564356436e-05, "loss": 1.8733, "step": 176 }, { "epoch": 0.6996047430830039, "grad_norm": 1.4987918138504028, "learning_rate": 4.834653465346535e-05, "loss": 2.0609, "step": 177 }, { "epoch": 0.7035573122529645, "grad_norm": 2.1115901470184326, "learning_rate": 4.833663366336634e-05, "loss": 1.8782, "step": 178 }, { "epoch": 0.7075098814229249, "grad_norm": 1.6857377290725708, "learning_rate": 4.832673267326733e-05, "loss": 1.9104, "step": 179 }, { "epoch": 0.7114624505928854, "grad_norm": 1.8482292890548706, "learning_rate": 4.831683168316832e-05, "loss": 1.7571, "step": 180 }, { "epoch": 0.7154150197628458, "grad_norm": 1.901819109916687, "learning_rate": 4.8306930693069305e-05, "loss": 2.0543, "step": 181 }, { "epoch": 0.7193675889328063, "grad_norm": 1.9455138444900513, "learning_rate": 4.8297029702970296e-05, "loss": 1.876, "step": 182 }, { "epoch": 0.7233201581027668, "grad_norm": 1.9429447650909424, "learning_rate": 4.828712871287129e-05, "loss": 1.8792, "step": 183 }, { "epoch": 0.7272727272727273, "grad_norm": 1.6803079843521118, "learning_rate": 4.827722772277228e-05, "loss": 1.7666, "step": 184 }, { "epoch": 0.7312252964426877, "grad_norm": 1.6958234310150146, "learning_rate": 4.826732673267327e-05, "loss": 1.95, "step": 185 }, { "epoch": 0.7351778656126482, "grad_norm": 1.9184436798095703, "learning_rate": 4.825742574257426e-05, "loss": 1.8282, "step": 186 }, { "epoch": 0.7391304347826086, "grad_norm": 1.8601891994476318, "learning_rate": 4.8247524752475245e-05, "loss": 1.6713, "step": 187 }, { "epoch": 0.7430830039525692, "grad_norm": 1.7922300100326538, "learning_rate": 4.8237623762376236e-05, "loss": 1.5923, "step": 188 }, { "epoch": 0.7470355731225297, "grad_norm": 2.041184902191162, "learning_rate": 4.822772277227723e-05, "loss": 1.8726, "step": 189 }, { "epoch": 0.7509881422924901, "grad_norm": 2.1383960247039795, "learning_rate": 4.821782178217822e-05, "loss": 1.9707, "step": 190 }, { "epoch": 0.7549407114624506, "grad_norm": 1.906326413154602, "learning_rate": 4.820792079207921e-05, "loss": 2.0391, "step": 191 }, { "epoch": 0.758893280632411, "grad_norm": 1.7709237337112427, "learning_rate": 4.81980198019802e-05, "loss": 1.6286, "step": 192 }, { "epoch": 0.7628458498023716, "grad_norm": 1.889207124710083, "learning_rate": 4.818811881188119e-05, "loss": 1.8482, "step": 193 }, { "epoch": 0.766798418972332, "grad_norm": 1.7642382383346558, "learning_rate": 4.8178217821782176e-05, "loss": 1.7327, "step": 194 }, { "epoch": 0.7707509881422925, "grad_norm": 1.6484183073043823, "learning_rate": 4.816831683168317e-05, "loss": 1.9291, "step": 195 }, { "epoch": 0.7747035573122529, "grad_norm": 1.4432159662246704, "learning_rate": 4.815841584158416e-05, "loss": 1.5317, "step": 196 }, { "epoch": 0.7786561264822134, "grad_norm": 1.6716856956481934, "learning_rate": 4.814851485148515e-05, "loss": 1.832, "step": 197 }, { "epoch": 0.782608695652174, "grad_norm": 1.8640812635421753, "learning_rate": 4.813861386138614e-05, "loss": 1.3399, "step": 198 }, { "epoch": 0.7865612648221344, "grad_norm": 2.9547786712646484, "learning_rate": 4.812871287128713e-05, "loss": 1.503, "step": 199 }, { "epoch": 0.7905138339920948, "grad_norm": 2.1680080890655518, "learning_rate": 4.811881188118812e-05, "loss": 1.7396, "step": 200 }, { "epoch": 0.7944664031620553, "grad_norm": 1.6820303201675415, "learning_rate": 4.8108910891089106e-05, "loss": 1.6711, "step": 201 }, { "epoch": 0.7984189723320159, "grad_norm": 1.5495200157165527, "learning_rate": 4.80990099009901e-05, "loss": 1.7934, "step": 202 }, { "epoch": 0.8023715415019763, "grad_norm": 1.8983807563781738, "learning_rate": 4.808910891089109e-05, "loss": 1.7568, "step": 203 }, { "epoch": 0.8063241106719368, "grad_norm": 2.478468894958496, "learning_rate": 4.807920792079208e-05, "loss": 1.4284, "step": 204 }, { "epoch": 0.8102766798418972, "grad_norm": 2.0719857215881348, "learning_rate": 4.806930693069307e-05, "loss": 1.7449, "step": 205 }, { "epoch": 0.8142292490118577, "grad_norm": 1.534789800643921, "learning_rate": 4.805940594059406e-05, "loss": 1.9735, "step": 206 }, { "epoch": 0.8181818181818182, "grad_norm": 1.5846573114395142, "learning_rate": 4.804950495049505e-05, "loss": 1.5419, "step": 207 }, { "epoch": 0.8221343873517787, "grad_norm": 2.10806941986084, "learning_rate": 4.803960396039604e-05, "loss": 1.7536, "step": 208 }, { "epoch": 0.8260869565217391, "grad_norm": 1.8637593984603882, "learning_rate": 4.802970297029703e-05, "loss": 1.9423, "step": 209 }, { "epoch": 0.8300395256916996, "grad_norm": 2.160545825958252, "learning_rate": 4.801980198019802e-05, "loss": 1.9304, "step": 210 }, { "epoch": 0.83399209486166, "grad_norm": 1.5259414911270142, "learning_rate": 4.800990099009901e-05, "loss": 1.8618, "step": 211 }, { "epoch": 0.8379446640316206, "grad_norm": 2.065152883529663, "learning_rate": 4.8e-05, "loss": 1.8356, "step": 212 }, { "epoch": 0.841897233201581, "grad_norm": 6.894260883331299, "learning_rate": 4.799009900990099e-05, "loss": 1.4841, "step": 213 }, { "epoch": 0.8458498023715415, "grad_norm": 1.8995124101638794, "learning_rate": 4.798019801980198e-05, "loss": 1.854, "step": 214 }, { "epoch": 0.849802371541502, "grad_norm": 1.677310824394226, "learning_rate": 4.797029702970297e-05, "loss": 1.8507, "step": 215 }, { "epoch": 0.8537549407114624, "grad_norm": 1.968108892440796, "learning_rate": 4.796039603960396e-05, "loss": 2.0363, "step": 216 }, { "epoch": 0.857707509881423, "grad_norm": 1.6589280366897583, "learning_rate": 4.795049504950495e-05, "loss": 1.6847, "step": 217 }, { "epoch": 0.8616600790513834, "grad_norm": 2.0190234184265137, "learning_rate": 4.794059405940594e-05, "loss": 1.4633, "step": 218 }, { "epoch": 0.8656126482213439, "grad_norm": 2.177816152572632, "learning_rate": 4.793069306930693e-05, "loss": 1.7488, "step": 219 }, { "epoch": 0.8695652173913043, "grad_norm": 1.5932660102844238, "learning_rate": 4.792079207920792e-05, "loss": 1.8604, "step": 220 }, { "epoch": 0.8735177865612648, "grad_norm": 1.424700379371643, "learning_rate": 4.7910891089108914e-05, "loss": 1.7068, "step": 221 }, { "epoch": 0.8774703557312253, "grad_norm": 2.0248913764953613, "learning_rate": 4.79009900990099e-05, "loss": 1.5631, "step": 222 }, { "epoch": 0.8814229249011858, "grad_norm": 1.7615950107574463, "learning_rate": 4.789108910891089e-05, "loss": 1.7083, "step": 223 }, { "epoch": 0.8853754940711462, "grad_norm": 1.7386078834533691, "learning_rate": 4.788118811881188e-05, "loss": 1.5821, "step": 224 }, { "epoch": 0.8893280632411067, "grad_norm": 1.7770156860351562, "learning_rate": 4.787128712871287e-05, "loss": 1.6365, "step": 225 }, { "epoch": 0.8932806324110671, "grad_norm": 1.6331318616867065, "learning_rate": 4.786138613861386e-05, "loss": 1.8557, "step": 226 }, { "epoch": 0.8972332015810277, "grad_norm": 1.6917909383773804, "learning_rate": 4.7851485148514854e-05, "loss": 1.69, "step": 227 }, { "epoch": 0.9011857707509882, "grad_norm": 1.9439878463745117, "learning_rate": 4.7841584158415845e-05, "loss": 1.6285, "step": 228 }, { "epoch": 0.9051383399209486, "grad_norm": 1.9159228801727295, "learning_rate": 4.783168316831683e-05, "loss": 1.568, "step": 229 }, { "epoch": 0.9090909090909091, "grad_norm": 1.4213659763336182, "learning_rate": 4.782178217821782e-05, "loss": 1.8279, "step": 230 }, { "epoch": 0.9130434782608695, "grad_norm": 3.2380495071411133, "learning_rate": 4.781188118811881e-05, "loss": 1.289, "step": 231 }, { "epoch": 0.9169960474308301, "grad_norm": 1.791998028755188, "learning_rate": 4.78019801980198e-05, "loss": 1.8407, "step": 232 }, { "epoch": 0.9209486166007905, "grad_norm": 2.08573055267334, "learning_rate": 4.779207920792079e-05, "loss": 1.8537, "step": 233 }, { "epoch": 0.924901185770751, "grad_norm": 1.8268327713012695, "learning_rate": 4.7782178217821784e-05, "loss": 1.7117, "step": 234 }, { "epoch": 0.9288537549407114, "grad_norm": 1.7242017984390259, "learning_rate": 4.7772277227722775e-05, "loss": 1.7021, "step": 235 }, { "epoch": 0.932806324110672, "grad_norm": 1.9589897394180298, "learning_rate": 4.776237623762376e-05, "loss": 2.0264, "step": 236 }, { "epoch": 0.9367588932806324, "grad_norm": 1.5871520042419434, "learning_rate": 4.775247524752475e-05, "loss": 1.9194, "step": 237 }, { "epoch": 0.9407114624505929, "grad_norm": 1.6635723114013672, "learning_rate": 4.774257425742574e-05, "loss": 1.5483, "step": 238 }, { "epoch": 0.9446640316205533, "grad_norm": 1.4513344764709473, "learning_rate": 4.773267326732673e-05, "loss": 1.5104, "step": 239 }, { "epoch": 0.9486166007905138, "grad_norm": 2.5077614784240723, "learning_rate": 4.7722772277227724e-05, "loss": 1.5918, "step": 240 }, { "epoch": 0.9525691699604744, "grad_norm": 1.8712127208709717, "learning_rate": 4.7712871287128715e-05, "loss": 1.9318, "step": 241 }, { "epoch": 0.9565217391304348, "grad_norm": 1.7041776180267334, "learning_rate": 4.7702970297029706e-05, "loss": 1.7548, "step": 242 }, { "epoch": 0.9604743083003953, "grad_norm": 2.150559663772583, "learning_rate": 4.769306930693069e-05, "loss": 1.5461, "step": 243 }, { "epoch": 0.9644268774703557, "grad_norm": 1.8353697061538696, "learning_rate": 4.768316831683168e-05, "loss": 1.7414, "step": 244 }, { "epoch": 0.9683794466403162, "grad_norm": 2.2976202964782715, "learning_rate": 4.767326732673267e-05, "loss": 1.5618, "step": 245 }, { "epoch": 0.9723320158102767, "grad_norm": 1.4848116636276245, "learning_rate": 4.7663366336633664e-05, "loss": 1.8281, "step": 246 }, { "epoch": 0.9762845849802372, "grad_norm": 1.5551823377609253, "learning_rate": 4.7653465346534655e-05, "loss": 1.6624, "step": 247 }, { "epoch": 0.9802371541501976, "grad_norm": 1.6832767724990845, "learning_rate": 4.7643564356435646e-05, "loss": 1.5533, "step": 248 }, { "epoch": 0.9841897233201581, "grad_norm": 1.600754737854004, "learning_rate": 4.763366336633664e-05, "loss": 1.5478, "step": 249 }, { "epoch": 0.9881422924901185, "grad_norm": 1.3893166780471802, "learning_rate": 4.762376237623762e-05, "loss": 1.7481, "step": 250 }, { "epoch": 0.9920948616600791, "grad_norm": 2.0788445472717285, "learning_rate": 4.761386138613861e-05, "loss": 1.7457, "step": 251 }, { "epoch": 0.9960474308300395, "grad_norm": 1.9635781049728394, "learning_rate": 4.76039603960396e-05, "loss": 1.726, "step": 252 }, { "epoch": 1.0, "grad_norm": 7.22420072555542, "learning_rate": 4.7594059405940594e-05, "loss": 1.6875, "step": 253 }, { "epoch": 1.0039525691699605, "grad_norm": 1.6005382537841797, "learning_rate": 4.7584158415841585e-05, "loss": 1.4686, "step": 254 }, { "epoch": 1.007905138339921, "grad_norm": 2.0420384407043457, "learning_rate": 4.7574257425742576e-05, "loss": 1.6077, "step": 255 }, { "epoch": 1.0118577075098814, "grad_norm": 1.517318844795227, "learning_rate": 4.756435643564357e-05, "loss": 1.7143, "step": 256 }, { "epoch": 1.0158102766798418, "grad_norm": 1.6022824048995972, "learning_rate": 4.755445544554455e-05, "loss": 1.1845, "step": 257 }, { "epoch": 1.0197628458498025, "grad_norm": 2.2452311515808105, "learning_rate": 4.754455445544554e-05, "loss": 1.4135, "step": 258 }, { "epoch": 1.023715415019763, "grad_norm": 2.4582247734069824, "learning_rate": 4.7534653465346534e-05, "loss": 1.401, "step": 259 }, { "epoch": 1.0276679841897234, "grad_norm": 1.5721155405044556, "learning_rate": 4.7524752475247525e-05, "loss": 1.3633, "step": 260 }, { "epoch": 1.0316205533596838, "grad_norm": 2.115096092224121, "learning_rate": 4.7514851485148516e-05, "loss": 1.6125, "step": 261 }, { "epoch": 1.0355731225296443, "grad_norm": 2.092665433883667, "learning_rate": 4.750495049504951e-05, "loss": 1.4099, "step": 262 }, { "epoch": 1.0395256916996047, "grad_norm": 3.7029614448547363, "learning_rate": 4.74950495049505e-05, "loss": 1.158, "step": 263 }, { "epoch": 1.0434782608695652, "grad_norm": 1.939917802810669, "learning_rate": 4.748514851485148e-05, "loss": 1.4732, "step": 264 }, { "epoch": 1.0474308300395256, "grad_norm": 1.7452563047409058, "learning_rate": 4.7475247524752474e-05, "loss": 1.2295, "step": 265 }, { "epoch": 1.051383399209486, "grad_norm": 1.7349773645401, "learning_rate": 4.7465346534653465e-05, "loss": 1.5561, "step": 266 }, { "epoch": 1.0553359683794465, "grad_norm": 2.0192389488220215, "learning_rate": 4.7455445544554456e-05, "loss": 1.5066, "step": 267 }, { "epoch": 1.0592885375494072, "grad_norm": 1.7572988271713257, "learning_rate": 4.744554455445545e-05, "loss": 1.4367, "step": 268 }, { "epoch": 1.0632411067193677, "grad_norm": 1.7771142721176147, "learning_rate": 4.743564356435644e-05, "loss": 1.5148, "step": 269 }, { "epoch": 1.0671936758893281, "grad_norm": 1.5750019550323486, "learning_rate": 4.742574257425743e-05, "loss": 1.3297, "step": 270 }, { "epoch": 1.0711462450592886, "grad_norm": 2.3078713417053223, "learning_rate": 4.741584158415841e-05, "loss": 1.1715, "step": 271 }, { "epoch": 1.075098814229249, "grad_norm": 2.1549389362335205, "learning_rate": 4.7405940594059404e-05, "loss": 1.3559, "step": 272 }, { "epoch": 1.0790513833992095, "grad_norm": 2.1807212829589844, "learning_rate": 4.7396039603960395e-05, "loss": 1.1357, "step": 273 }, { "epoch": 1.08300395256917, "grad_norm": 1.934232234954834, "learning_rate": 4.7386138613861386e-05, "loss": 1.3438, "step": 274 }, { "epoch": 1.0869565217391304, "grad_norm": 4.971157550811768, "learning_rate": 4.737623762376238e-05, "loss": 1.738, "step": 275 }, { "epoch": 1.0909090909090908, "grad_norm": 1.7957723140716553, "learning_rate": 4.736633663366337e-05, "loss": 1.3101, "step": 276 }, { "epoch": 1.0948616600790513, "grad_norm": 1.902584433555603, "learning_rate": 4.735643564356436e-05, "loss": 1.2114, "step": 277 }, { "epoch": 1.098814229249012, "grad_norm": 1.9963431358337402, "learning_rate": 4.7346534653465344e-05, "loss": 1.4809, "step": 278 }, { "epoch": 1.1027667984189724, "grad_norm": 2.336531162261963, "learning_rate": 4.7336633663366335e-05, "loss": 1.5755, "step": 279 }, { "epoch": 1.1067193675889329, "grad_norm": 1.9655275344848633, "learning_rate": 4.7326732673267326e-05, "loss": 1.6481, "step": 280 }, { "epoch": 1.1106719367588933, "grad_norm": 1.7847837209701538, "learning_rate": 4.731683168316832e-05, "loss": 1.364, "step": 281 }, { "epoch": 1.1146245059288538, "grad_norm": 2.3622517585754395, "learning_rate": 4.730693069306931e-05, "loss": 1.2871, "step": 282 }, { "epoch": 1.1185770750988142, "grad_norm": 1.8175315856933594, "learning_rate": 4.72970297029703e-05, "loss": 1.3615, "step": 283 }, { "epoch": 1.1225296442687747, "grad_norm": 1.6893761157989502, "learning_rate": 4.728712871287129e-05, "loss": 1.4906, "step": 284 }, { "epoch": 1.1264822134387351, "grad_norm": 1.6157723665237427, "learning_rate": 4.7277227722772274e-05, "loss": 1.0983, "step": 285 }, { "epoch": 1.1304347826086956, "grad_norm": 3.089219570159912, "learning_rate": 4.7267326732673266e-05, "loss": 1.6502, "step": 286 }, { "epoch": 1.1343873517786562, "grad_norm": 1.8247320652008057, "learning_rate": 4.725742574257426e-05, "loss": 1.6806, "step": 287 }, { "epoch": 1.1383399209486167, "grad_norm": 2.0166752338409424, "learning_rate": 4.724752475247525e-05, "loss": 1.2487, "step": 288 }, { "epoch": 1.1422924901185771, "grad_norm": 2.132845878601074, "learning_rate": 4.723762376237624e-05, "loss": 1.3088, "step": 289 }, { "epoch": 1.1462450592885376, "grad_norm": 2.269001007080078, "learning_rate": 4.722772277227723e-05, "loss": 1.2635, "step": 290 }, { "epoch": 1.150197628458498, "grad_norm": 2.1111152172088623, "learning_rate": 4.721782178217822e-05, "loss": 1.5949, "step": 291 }, { "epoch": 1.1541501976284585, "grad_norm": 1.4548749923706055, "learning_rate": 4.7207920792079205e-05, "loss": 1.4707, "step": 292 }, { "epoch": 1.158102766798419, "grad_norm": 2.211790084838867, "learning_rate": 4.7198019801980196e-05, "loss": 1.3766, "step": 293 }, { "epoch": 1.1620553359683794, "grad_norm": 1.576881766319275, "learning_rate": 4.718811881188119e-05, "loss": 1.4683, "step": 294 }, { "epoch": 1.1660079051383399, "grad_norm": 1.2804337739944458, "learning_rate": 4.717821782178218e-05, "loss": 1.3973, "step": 295 }, { "epoch": 1.1699604743083003, "grad_norm": 1.6789839267730713, "learning_rate": 4.716831683168317e-05, "loss": 1.2494, "step": 296 }, { "epoch": 1.1739130434782608, "grad_norm": 1.6707631349563599, "learning_rate": 4.715841584158416e-05, "loss": 0.916, "step": 297 }, { "epoch": 1.1778656126482214, "grad_norm": 2.0857536792755127, "learning_rate": 4.714851485148515e-05, "loss": 1.4103, "step": 298 }, { "epoch": 1.1818181818181819, "grad_norm": 1.9525047540664673, "learning_rate": 4.7138613861386136e-05, "loss": 1.3881, "step": 299 }, { "epoch": 1.1857707509881423, "grad_norm": 2.110231399536133, "learning_rate": 4.712871287128713e-05, "loss": 1.5059, "step": 300 }, { "epoch": 1.1897233201581028, "grad_norm": 2.0692641735076904, "learning_rate": 4.711881188118812e-05, "loss": 1.4967, "step": 301 }, { "epoch": 1.1936758893280632, "grad_norm": 1.7283746004104614, "learning_rate": 4.710891089108911e-05, "loss": 1.2775, "step": 302 }, { "epoch": 1.1976284584980237, "grad_norm": 1.5576273202896118, "learning_rate": 4.70990099009901e-05, "loss": 1.487, "step": 303 }, { "epoch": 1.2015810276679841, "grad_norm": 2.0909595489501953, "learning_rate": 4.708910891089109e-05, "loss": 1.2885, "step": 304 }, { "epoch": 1.2055335968379446, "grad_norm": 1.5556882619857788, "learning_rate": 4.7079207920792075e-05, "loss": 1.3502, "step": 305 }, { "epoch": 1.2094861660079053, "grad_norm": 1.9144208431243896, "learning_rate": 4.7069306930693067e-05, "loss": 1.3107, "step": 306 }, { "epoch": 1.2134387351778657, "grad_norm": 2.0427002906799316, "learning_rate": 4.705940594059406e-05, "loss": 1.5829, "step": 307 }, { "epoch": 1.2173913043478262, "grad_norm": 2.344679117202759, "learning_rate": 4.704950495049505e-05, "loss": 1.3832, "step": 308 }, { "epoch": 1.2213438735177866, "grad_norm": 1.676110029220581, "learning_rate": 4.703960396039604e-05, "loss": 1.4398, "step": 309 }, { "epoch": 1.225296442687747, "grad_norm": 1.6690326929092407, "learning_rate": 4.702970297029703e-05, "loss": 1.2756, "step": 310 }, { "epoch": 1.2292490118577075, "grad_norm": 1.9874050617218018, "learning_rate": 4.701980198019802e-05, "loss": 1.5199, "step": 311 }, { "epoch": 1.233201581027668, "grad_norm": 1.9189642667770386, "learning_rate": 4.7009900990099006e-05, "loss": 1.508, "step": 312 }, { "epoch": 1.2371541501976284, "grad_norm": 2.53155517578125, "learning_rate": 4.7e-05, "loss": 1.4133, "step": 313 }, { "epoch": 1.2411067193675889, "grad_norm": 1.950171947479248, "learning_rate": 4.699009900990099e-05, "loss": 1.2193, "step": 314 }, { "epoch": 1.2450592885375493, "grad_norm": 2.155329704284668, "learning_rate": 4.698019801980198e-05, "loss": 1.3372, "step": 315 }, { "epoch": 1.2490118577075098, "grad_norm": 1.5941109657287598, "learning_rate": 4.697029702970297e-05, "loss": 1.3854, "step": 316 }, { "epoch": 1.2529644268774702, "grad_norm": 2.200648784637451, "learning_rate": 4.696039603960396e-05, "loss": 1.3279, "step": 317 }, { "epoch": 1.256916996047431, "grad_norm": 2.5741448402404785, "learning_rate": 4.695049504950495e-05, "loss": 1.4066, "step": 318 }, { "epoch": 1.2608695652173914, "grad_norm": 2.06169056892395, "learning_rate": 4.694059405940594e-05, "loss": 1.2926, "step": 319 }, { "epoch": 1.2648221343873518, "grad_norm": 1.40762197971344, "learning_rate": 4.693069306930693e-05, "loss": 1.2594, "step": 320 }, { "epoch": 1.2687747035573123, "grad_norm": 2.086052179336548, "learning_rate": 4.692079207920792e-05, "loss": 1.6162, "step": 321 }, { "epoch": 1.2727272727272727, "grad_norm": 2.0672683715820312, "learning_rate": 4.691089108910891e-05, "loss": 1.147, "step": 322 }, { "epoch": 1.2766798418972332, "grad_norm": 1.7091121673583984, "learning_rate": 4.69009900990099e-05, "loss": 1.5007, "step": 323 }, { "epoch": 1.2806324110671936, "grad_norm": 2.085710287094116, "learning_rate": 4.689108910891089e-05, "loss": 1.5608, "step": 324 }, { "epoch": 1.2845849802371543, "grad_norm": 1.903905987739563, "learning_rate": 4.688118811881188e-05, "loss": 1.3325, "step": 325 }, { "epoch": 1.2885375494071147, "grad_norm": 1.621509075164795, "learning_rate": 4.6871287128712874e-05, "loss": 1.4383, "step": 326 }, { "epoch": 1.2924901185770752, "grad_norm": 1.8243176937103271, "learning_rate": 4.6861386138613865e-05, "loss": 1.3309, "step": 327 }, { "epoch": 1.2964426877470356, "grad_norm": 1.8314933776855469, "learning_rate": 4.6851485148514856e-05, "loss": 1.5152, "step": 328 }, { "epoch": 1.300395256916996, "grad_norm": 1.84877610206604, "learning_rate": 4.684158415841585e-05, "loss": 1.5485, "step": 329 }, { "epoch": 1.3043478260869565, "grad_norm": 3.0376999378204346, "learning_rate": 4.683168316831684e-05, "loss": 1.3884, "step": 330 }, { "epoch": 1.308300395256917, "grad_norm": 1.826781153678894, "learning_rate": 4.682178217821783e-05, "loss": 1.3867, "step": 331 }, { "epoch": 1.3122529644268774, "grad_norm": 2.367933511734009, "learning_rate": 4.6811881188118814e-05, "loss": 1.2829, "step": 332 }, { "epoch": 1.316205533596838, "grad_norm": 2.2275338172912598, "learning_rate": 4.6801980198019805e-05, "loss": 1.3365, "step": 333 }, { "epoch": 1.3201581027667983, "grad_norm": 1.9872102737426758, "learning_rate": 4.6792079207920796e-05, "loss": 1.3518, "step": 334 }, { "epoch": 1.3241106719367588, "grad_norm": 1.675817847251892, "learning_rate": 4.678217821782179e-05, "loss": 1.4485, "step": 335 }, { "epoch": 1.3280632411067192, "grad_norm": 1.8062676191329956, "learning_rate": 4.677227722772278e-05, "loss": 1.1217, "step": 336 }, { "epoch": 1.33201581027668, "grad_norm": 1.6064897775650024, "learning_rate": 4.676237623762377e-05, "loss": 1.5592, "step": 337 }, { "epoch": 1.3359683794466404, "grad_norm": 2.206401824951172, "learning_rate": 4.675247524752476e-05, "loss": 1.4308, "step": 338 }, { "epoch": 1.3399209486166008, "grad_norm": 1.3740999698638916, "learning_rate": 4.6742574257425745e-05, "loss": 1.5133, "step": 339 }, { "epoch": 1.3438735177865613, "grad_norm": 1.599977731704712, "learning_rate": 4.6732673267326736e-05, "loss": 1.16, "step": 340 }, { "epoch": 1.3478260869565217, "grad_norm": 2.25478196144104, "learning_rate": 4.672277227722773e-05, "loss": 1.2871, "step": 341 }, { "epoch": 1.3517786561264822, "grad_norm": 2.2603561878204346, "learning_rate": 4.671287128712872e-05, "loss": 1.3089, "step": 342 }, { "epoch": 1.3557312252964426, "grad_norm": 2.7896456718444824, "learning_rate": 4.670297029702971e-05, "loss": 1.2485, "step": 343 }, { "epoch": 1.359683794466403, "grad_norm": 2.2158701419830322, "learning_rate": 4.66930693069307e-05, "loss": 1.2645, "step": 344 }, { "epoch": 1.3636363636363638, "grad_norm": 2.152134418487549, "learning_rate": 4.668316831683169e-05, "loss": 1.4099, "step": 345 }, { "epoch": 1.3675889328063242, "grad_norm": 1.6516609191894531, "learning_rate": 4.6673267326732675e-05, "loss": 1.3783, "step": 346 }, { "epoch": 1.3715415019762847, "grad_norm": 2.1257731914520264, "learning_rate": 4.6663366336633666e-05, "loss": 1.2942, "step": 347 }, { "epoch": 1.3754940711462451, "grad_norm": 2.157472610473633, "learning_rate": 4.665346534653466e-05, "loss": 1.5212, "step": 348 }, { "epoch": 1.3794466403162056, "grad_norm": 1.7937648296356201, "learning_rate": 4.664356435643565e-05, "loss": 1.1783, "step": 349 }, { "epoch": 1.383399209486166, "grad_norm": 1.5477609634399414, "learning_rate": 4.663366336633664e-05, "loss": 1.4729, "step": 350 }, { "epoch": 1.3873517786561265, "grad_norm": 2.1787993907928467, "learning_rate": 4.662376237623763e-05, "loss": 1.2097, "step": 351 }, { "epoch": 1.391304347826087, "grad_norm": 1.8767383098602295, "learning_rate": 4.661386138613862e-05, "loss": 1.3436, "step": 352 }, { "epoch": 1.3952569169960474, "grad_norm": 2.041501045227051, "learning_rate": 4.6603960396039606e-05, "loss": 0.9086, "step": 353 }, { "epoch": 1.3992094861660078, "grad_norm": 2.3226981163024902, "learning_rate": 4.65940594059406e-05, "loss": 1.4897, "step": 354 }, { "epoch": 1.4031620553359683, "grad_norm": 2.1870439052581787, "learning_rate": 4.658415841584159e-05, "loss": 1.3072, "step": 355 }, { "epoch": 1.4071146245059287, "grad_norm": 1.7698314189910889, "learning_rate": 4.657425742574258e-05, "loss": 1.5381, "step": 356 }, { "epoch": 1.4110671936758894, "grad_norm": 2.035149097442627, "learning_rate": 4.656435643564357e-05, "loss": 1.6832, "step": 357 }, { "epoch": 1.4150197628458498, "grad_norm": 1.6357266902923584, "learning_rate": 4.655445544554456e-05, "loss": 1.6517, "step": 358 }, { "epoch": 1.4189723320158103, "grad_norm": 1.7072739601135254, "learning_rate": 4.654455445544555e-05, "loss": 1.2053, "step": 359 }, { "epoch": 1.4229249011857708, "grad_norm": 2.007456064224243, "learning_rate": 4.653465346534654e-05, "loss": 1.3499, "step": 360 }, { "epoch": 1.4268774703557312, "grad_norm": 1.8351974487304688, "learning_rate": 4.652475247524753e-05, "loss": 1.5207, "step": 361 }, { "epoch": 1.4308300395256917, "grad_norm": 1.8796175718307495, "learning_rate": 4.651485148514852e-05, "loss": 1.3183, "step": 362 }, { "epoch": 1.434782608695652, "grad_norm": 2.100463390350342, "learning_rate": 4.650495049504951e-05, "loss": 1.3909, "step": 363 }, { "epoch": 1.4387351778656128, "grad_norm": 1.6312026977539062, "learning_rate": 4.64950495049505e-05, "loss": 1.4801, "step": 364 }, { "epoch": 1.4426877470355732, "grad_norm": 2.014965295791626, "learning_rate": 4.648514851485149e-05, "loss": 1.3134, "step": 365 }, { "epoch": 1.4466403162055337, "grad_norm": 1.7578192949295044, "learning_rate": 4.647524752475248e-05, "loss": 1.3006, "step": 366 }, { "epoch": 1.4505928853754941, "grad_norm": 2.0413758754730225, "learning_rate": 4.646534653465347e-05, "loss": 1.7193, "step": 367 }, { "epoch": 1.4545454545454546, "grad_norm": 2.005418539047241, "learning_rate": 4.645544554455446e-05, "loss": 1.3437, "step": 368 }, { "epoch": 1.458498023715415, "grad_norm": 1.766155481338501, "learning_rate": 4.644554455445545e-05, "loss": 1.5195, "step": 369 }, { "epoch": 1.4624505928853755, "grad_norm": 1.669294834136963, "learning_rate": 4.643564356435644e-05, "loss": 1.2493, "step": 370 }, { "epoch": 1.466403162055336, "grad_norm": 1.814496397972107, "learning_rate": 4.642574257425743e-05, "loss": 1.2513, "step": 371 }, { "epoch": 1.4703557312252964, "grad_norm": 1.6811078786849976, "learning_rate": 4.641584158415842e-05, "loss": 1.2867, "step": 372 }, { "epoch": 1.4743083003952568, "grad_norm": 1.6480709314346313, "learning_rate": 4.6405940594059414e-05, "loss": 1.1799, "step": 373 }, { "epoch": 1.4782608695652173, "grad_norm": 1.842932939529419, "learning_rate": 4.63960396039604e-05, "loss": 1.4516, "step": 374 }, { "epoch": 1.4822134387351777, "grad_norm": 1.7719305753707886, "learning_rate": 4.638613861386139e-05, "loss": 1.2765, "step": 375 }, { "epoch": 1.4861660079051384, "grad_norm": 1.9938666820526123, "learning_rate": 4.637623762376238e-05, "loss": 1.355, "step": 376 }, { "epoch": 1.4901185770750989, "grad_norm": 1.5959348678588867, "learning_rate": 4.636633663366337e-05, "loss": 1.553, "step": 377 }, { "epoch": 1.4940711462450593, "grad_norm": 1.7426046133041382, "learning_rate": 4.635643564356436e-05, "loss": 1.43, "step": 378 }, { "epoch": 1.4980237154150198, "grad_norm": 2.4649877548217773, "learning_rate": 4.634653465346535e-05, "loss": 1.3998, "step": 379 }, { "epoch": 1.5019762845849802, "grad_norm": 1.7895821332931519, "learning_rate": 4.6336633663366344e-05, "loss": 1.0499, "step": 380 }, { "epoch": 1.5059288537549407, "grad_norm": 1.9846243858337402, "learning_rate": 4.632673267326733e-05, "loss": 1.5214, "step": 381 }, { "epoch": 1.5098814229249014, "grad_norm": 2.552523136138916, "learning_rate": 4.631683168316832e-05, "loss": 1.5568, "step": 382 }, { "epoch": 1.5138339920948618, "grad_norm": 2.194394111633301, "learning_rate": 4.630693069306931e-05, "loss": 1.2123, "step": 383 }, { "epoch": 1.5177865612648223, "grad_norm": 2.573134660720825, "learning_rate": 4.62970297029703e-05, "loss": 1.5278, "step": 384 }, { "epoch": 1.5217391304347827, "grad_norm": 1.716528058052063, "learning_rate": 4.628712871287129e-05, "loss": 1.4175, "step": 385 }, { "epoch": 1.5256916996047432, "grad_norm": 2.3797268867492676, "learning_rate": 4.6277227722772284e-05, "loss": 1.9483, "step": 386 }, { "epoch": 1.5296442687747036, "grad_norm": 2.731419801712036, "learning_rate": 4.6267326732673275e-05, "loss": 1.287, "step": 387 }, { "epoch": 1.533596837944664, "grad_norm": 1.919443130493164, "learning_rate": 4.625742574257426e-05, "loss": 1.7387, "step": 388 }, { "epoch": 1.5375494071146245, "grad_norm": 1.6352767944335938, "learning_rate": 4.624752475247525e-05, "loss": 1.4139, "step": 389 }, { "epoch": 1.541501976284585, "grad_norm": 2.0181920528411865, "learning_rate": 4.623762376237624e-05, "loss": 1.4922, "step": 390 }, { "epoch": 1.5454545454545454, "grad_norm": 1.6219663619995117, "learning_rate": 4.622772277227723e-05, "loss": 1.506, "step": 391 }, { "epoch": 1.5494071146245059, "grad_norm": 1.6920379400253296, "learning_rate": 4.6217821782178224e-05, "loss": 1.3612, "step": 392 }, { "epoch": 1.5533596837944663, "grad_norm": 1.5363556146621704, "learning_rate": 4.6207920792079215e-05, "loss": 1.401, "step": 393 }, { "epoch": 1.5573122529644268, "grad_norm": 1.5093425512313843, "learning_rate": 4.61980198019802e-05, "loss": 1.3071, "step": 394 }, { "epoch": 1.5612648221343872, "grad_norm": 1.8421881198883057, "learning_rate": 4.618811881188119e-05, "loss": 1.2572, "step": 395 }, { "epoch": 1.5652173913043477, "grad_norm": 2.371893882751465, "learning_rate": 4.617821782178218e-05, "loss": 1.6744, "step": 396 }, { "epoch": 1.5691699604743083, "grad_norm": 1.7512977123260498, "learning_rate": 4.616831683168317e-05, "loss": 1.4221, "step": 397 }, { "epoch": 1.5731225296442688, "grad_norm": 2.2294204235076904, "learning_rate": 4.615841584158416e-05, "loss": 1.3428, "step": 398 }, { "epoch": 1.5770750988142292, "grad_norm": 2.5290157794952393, "learning_rate": 4.6148514851485154e-05, "loss": 1.5037, "step": 399 }, { "epoch": 1.5810276679841897, "grad_norm": 1.7091094255447388, "learning_rate": 4.6138613861386145e-05, "loss": 1.3447, "step": 400 }, { "epoch": 1.5849802371541502, "grad_norm": 1.9178377389907837, "learning_rate": 4.612871287128713e-05, "loss": 1.6564, "step": 401 }, { "epoch": 1.5889328063241108, "grad_norm": 2.1111249923706055, "learning_rate": 4.611881188118812e-05, "loss": 1.3789, "step": 402 }, { "epoch": 1.5928853754940713, "grad_norm": 1.8857834339141846, "learning_rate": 4.610891089108911e-05, "loss": 1.2806, "step": 403 }, { "epoch": 1.5968379446640317, "grad_norm": 2.637451410293579, "learning_rate": 4.60990099009901e-05, "loss": 1.4099, "step": 404 }, { "epoch": 1.6007905138339922, "grad_norm": 1.9618099927902222, "learning_rate": 4.6089108910891094e-05, "loss": 1.2571, "step": 405 }, { "epoch": 1.6047430830039526, "grad_norm": 2.03883695602417, "learning_rate": 4.6079207920792085e-05, "loss": 1.5109, "step": 406 }, { "epoch": 1.608695652173913, "grad_norm": 1.5719876289367676, "learning_rate": 4.6069306930693076e-05, "loss": 1.5353, "step": 407 }, { "epoch": 1.6126482213438735, "grad_norm": 1.6818652153015137, "learning_rate": 4.605940594059406e-05, "loss": 1.2373, "step": 408 }, { "epoch": 1.616600790513834, "grad_norm": 1.4927862882614136, "learning_rate": 4.604950495049505e-05, "loss": 1.3111, "step": 409 }, { "epoch": 1.6205533596837944, "grad_norm": 2.346480131149292, "learning_rate": 4.603960396039604e-05, "loss": 1.3175, "step": 410 }, { "epoch": 1.6245059288537549, "grad_norm": 1.755994200706482, "learning_rate": 4.6029702970297034e-05, "loss": 1.361, "step": 411 }, { "epoch": 1.6284584980237153, "grad_norm": 1.8060498237609863, "learning_rate": 4.6019801980198025e-05, "loss": 1.5327, "step": 412 }, { "epoch": 1.6324110671936758, "grad_norm": 1.9229106903076172, "learning_rate": 4.6009900990099016e-05, "loss": 1.2883, "step": 413 }, { "epoch": 1.6363636363636362, "grad_norm": 1.8613306283950806, "learning_rate": 4.600000000000001e-05, "loss": 1.1791, "step": 414 }, { "epoch": 1.6403162055335967, "grad_norm": 2.2563979625701904, "learning_rate": 4.599009900990099e-05, "loss": 1.4539, "step": 415 }, { "epoch": 1.6442687747035574, "grad_norm": 1.8289929628372192, "learning_rate": 4.598019801980198e-05, "loss": 1.3568, "step": 416 }, { "epoch": 1.6482213438735178, "grad_norm": 1.6632519960403442, "learning_rate": 4.597029702970297e-05, "loss": 1.6013, "step": 417 }, { "epoch": 1.6521739130434783, "grad_norm": 1.6696910858154297, "learning_rate": 4.5960396039603964e-05, "loss": 1.31, "step": 418 }, { "epoch": 1.6561264822134387, "grad_norm": 2.149606227874756, "learning_rate": 4.5950495049504955e-05, "loss": 1.3549, "step": 419 }, { "epoch": 1.6600790513833992, "grad_norm": 1.774701476097107, "learning_rate": 4.5940594059405946e-05, "loss": 1.4967, "step": 420 }, { "epoch": 1.6640316205533598, "grad_norm": 9.298806190490723, "learning_rate": 4.593069306930694e-05, "loss": 1.4333, "step": 421 }, { "epoch": 1.6679841897233203, "grad_norm": 2.544452428817749, "learning_rate": 4.592079207920792e-05, "loss": 1.2452, "step": 422 }, { "epoch": 1.6719367588932808, "grad_norm": 1.6767677068710327, "learning_rate": 4.591089108910891e-05, "loss": 1.3579, "step": 423 }, { "epoch": 1.6758893280632412, "grad_norm": 2.0297420024871826, "learning_rate": 4.5900990099009904e-05, "loss": 0.9547, "step": 424 }, { "epoch": 1.6798418972332017, "grad_norm": 2.2107417583465576, "learning_rate": 4.5891089108910895e-05, "loss": 1.2621, "step": 425 }, { "epoch": 1.683794466403162, "grad_norm": 1.9431469440460205, "learning_rate": 4.5881188118811886e-05, "loss": 1.1582, "step": 426 }, { "epoch": 1.6877470355731226, "grad_norm": 2.2594752311706543, "learning_rate": 4.587128712871288e-05, "loss": 1.2714, "step": 427 }, { "epoch": 1.691699604743083, "grad_norm": 1.7576017379760742, "learning_rate": 4.586138613861387e-05, "loss": 1.2844, "step": 428 }, { "epoch": 1.6956521739130435, "grad_norm": 2.163604497909546, "learning_rate": 4.585148514851485e-05, "loss": 1.3588, "step": 429 }, { "epoch": 1.699604743083004, "grad_norm": 2.29953932762146, "learning_rate": 4.5841584158415844e-05, "loss": 1.415, "step": 430 }, { "epoch": 1.7035573122529644, "grad_norm": 2.103315830230713, "learning_rate": 4.5831683168316835e-05, "loss": 1.3986, "step": 431 }, { "epoch": 1.7075098814229248, "grad_norm": 2.5195417404174805, "learning_rate": 4.5821782178217826e-05, "loss": 1.2951, "step": 432 }, { "epoch": 1.7114624505928853, "grad_norm": 2.234647750854492, "learning_rate": 4.581188118811882e-05, "loss": 1.4059, "step": 433 }, { "epoch": 1.7154150197628457, "grad_norm": 1.8694403171539307, "learning_rate": 4.580198019801981e-05, "loss": 1.516, "step": 434 }, { "epoch": 1.7193675889328062, "grad_norm": 2.334695339202881, "learning_rate": 4.57920792079208e-05, "loss": 1.2044, "step": 435 }, { "epoch": 1.7233201581027668, "grad_norm": 1.8155462741851807, "learning_rate": 4.578217821782178e-05, "loss": 1.6631, "step": 436 }, { "epoch": 1.7272727272727273, "grad_norm": 1.8584500551223755, "learning_rate": 4.5772277227722774e-05, "loss": 1.1225, "step": 437 }, { "epoch": 1.7312252964426877, "grad_norm": 1.9015271663665771, "learning_rate": 4.5762376237623765e-05, "loss": 1.3277, "step": 438 }, { "epoch": 1.7351778656126482, "grad_norm": 2.047449827194214, "learning_rate": 4.5752475247524756e-05, "loss": 1.3, "step": 439 }, { "epoch": 1.7391304347826086, "grad_norm": 1.9489177465438843, "learning_rate": 4.574257425742575e-05, "loss": 1.4589, "step": 440 }, { "epoch": 1.7430830039525693, "grad_norm": 2.0734496116638184, "learning_rate": 4.573267326732674e-05, "loss": 1.2396, "step": 441 }, { "epoch": 1.7470355731225298, "grad_norm": 1.774956464767456, "learning_rate": 4.572277227722773e-05, "loss": 1.573, "step": 442 }, { "epoch": 1.7509881422924902, "grad_norm": 2.037916421890259, "learning_rate": 4.5712871287128714e-05, "loss": 1.2748, "step": 443 }, { "epoch": 1.7549407114624507, "grad_norm": 1.7255709171295166, "learning_rate": 4.5702970297029705e-05, "loss": 1.4542, "step": 444 }, { "epoch": 1.7588932806324111, "grad_norm": 2.3671793937683105, "learning_rate": 4.5693069306930696e-05, "loss": 1.5334, "step": 445 }, { "epoch": 1.7628458498023716, "grad_norm": 1.7773637771606445, "learning_rate": 4.568316831683169e-05, "loss": 1.2174, "step": 446 }, { "epoch": 1.766798418972332, "grad_norm": 1.6897904872894287, "learning_rate": 4.567326732673268e-05, "loss": 1.8058, "step": 447 }, { "epoch": 1.7707509881422925, "grad_norm": 1.8180333375930786, "learning_rate": 4.566336633663367e-05, "loss": 1.3808, "step": 448 }, { "epoch": 1.774703557312253, "grad_norm": 1.7750916481018066, "learning_rate": 4.565346534653466e-05, "loss": 1.2635, "step": 449 }, { "epoch": 1.7786561264822134, "grad_norm": 1.878635287284851, "learning_rate": 4.5643564356435645e-05, "loss": 1.3163, "step": 450 }, { "epoch": 1.7826086956521738, "grad_norm": 2.408804416656494, "learning_rate": 4.5633663366336636e-05, "loss": 1.3988, "step": 451 }, { "epoch": 1.7865612648221343, "grad_norm": 1.867692470550537, "learning_rate": 4.562376237623763e-05, "loss": 1.1118, "step": 452 }, { "epoch": 1.7905138339920947, "grad_norm": 1.925492763519287, "learning_rate": 4.561386138613862e-05, "loss": 1.1434, "step": 453 }, { "epoch": 1.7944664031620552, "grad_norm": 2.450352668762207, "learning_rate": 4.560396039603961e-05, "loss": 1.0641, "step": 454 }, { "epoch": 1.7984189723320159, "grad_norm": 2.1818490028381348, "learning_rate": 4.55940594059406e-05, "loss": 1.2692, "step": 455 }, { "epoch": 1.8023715415019763, "grad_norm": 1.7538120746612549, "learning_rate": 4.558415841584159e-05, "loss": 1.532, "step": 456 }, { "epoch": 1.8063241106719368, "grad_norm": 1.8396257162094116, "learning_rate": 4.5574257425742575e-05, "loss": 1.5914, "step": 457 }, { "epoch": 1.8102766798418972, "grad_norm": 2.263643741607666, "learning_rate": 4.5564356435643566e-05, "loss": 1.1817, "step": 458 }, { "epoch": 1.8142292490118577, "grad_norm": 2.4400599002838135, "learning_rate": 4.555445544554456e-05, "loss": 1.116, "step": 459 }, { "epoch": 1.8181818181818183, "grad_norm": 1.5119582414627075, "learning_rate": 4.554455445544555e-05, "loss": 1.2407, "step": 460 }, { "epoch": 1.8221343873517788, "grad_norm": 1.5930187702178955, "learning_rate": 4.553465346534654e-05, "loss": 1.5179, "step": 461 }, { "epoch": 1.8260869565217392, "grad_norm": 1.5701802968978882, "learning_rate": 4.552475247524753e-05, "loss": 1.274, "step": 462 }, { "epoch": 1.8300395256916997, "grad_norm": 1.9805805683135986, "learning_rate": 4.551485148514852e-05, "loss": 1.1487, "step": 463 }, { "epoch": 1.8339920948616601, "grad_norm": 2.153428554534912, "learning_rate": 4.5504950495049506e-05, "loss": 1.1947, "step": 464 }, { "epoch": 1.8379446640316206, "grad_norm": 1.7646703720092773, "learning_rate": 4.54950495049505e-05, "loss": 1.2369, "step": 465 }, { "epoch": 1.841897233201581, "grad_norm": 1.950351357460022, "learning_rate": 4.548514851485149e-05, "loss": 1.4005, "step": 466 }, { "epoch": 1.8458498023715415, "grad_norm": 2.204277276992798, "learning_rate": 4.547524752475248e-05, "loss": 1.2623, "step": 467 }, { "epoch": 1.849802371541502, "grad_norm": 2.1276376247406006, "learning_rate": 4.546534653465347e-05, "loss": 1.3835, "step": 468 }, { "epoch": 1.8537549407114624, "grad_norm": 2.4032175540924072, "learning_rate": 4.545544554455446e-05, "loss": 1.6914, "step": 469 }, { "epoch": 1.8577075098814229, "grad_norm": 2.2454400062561035, "learning_rate": 4.544554455445545e-05, "loss": 1.3167, "step": 470 }, { "epoch": 1.8616600790513833, "grad_norm": 2.1577837467193604, "learning_rate": 4.5435643564356437e-05, "loss": 1.2607, "step": 471 }, { "epoch": 1.8656126482213438, "grad_norm": 1.3870981931686401, "learning_rate": 4.542574257425743e-05, "loss": 1.352, "step": 472 }, { "epoch": 1.8695652173913042, "grad_norm": 1.6090911626815796, "learning_rate": 4.541584158415842e-05, "loss": 1.2107, "step": 473 }, { "epoch": 1.8735177865612647, "grad_norm": 1.6216145753860474, "learning_rate": 4.540594059405941e-05, "loss": 1.5291, "step": 474 }, { "epoch": 1.8774703557312253, "grad_norm": 1.8958593606948853, "learning_rate": 4.53960396039604e-05, "loss": 1.3123, "step": 475 }, { "epoch": 1.8814229249011858, "grad_norm": 1.8213366270065308, "learning_rate": 4.538613861386139e-05, "loss": 1.401, "step": 476 }, { "epoch": 1.8853754940711462, "grad_norm": 2.5192534923553467, "learning_rate": 4.537623762376238e-05, "loss": 1.5532, "step": 477 }, { "epoch": 1.8893280632411067, "grad_norm": 2.3187332153320312, "learning_rate": 4.536633663366337e-05, "loss": 1.4063, "step": 478 }, { "epoch": 1.8932806324110671, "grad_norm": 2.1068081855773926, "learning_rate": 4.535643564356436e-05, "loss": 1.4429, "step": 479 }, { "epoch": 1.8972332015810278, "grad_norm": 2.338728427886963, "learning_rate": 4.534653465346535e-05, "loss": 1.5375, "step": 480 }, { "epoch": 1.9011857707509883, "grad_norm": 1.9374494552612305, "learning_rate": 4.533663366336634e-05, "loss": 1.1688, "step": 481 }, { "epoch": 1.9051383399209487, "grad_norm": 2.142967462539673, "learning_rate": 4.532673267326733e-05, "loss": 1.4018, "step": 482 }, { "epoch": 1.9090909090909092, "grad_norm": 1.7175180912017822, "learning_rate": 4.531683168316832e-05, "loss": 1.4475, "step": 483 }, { "epoch": 1.9130434782608696, "grad_norm": 2.1896677017211914, "learning_rate": 4.5306930693069314e-05, "loss": 1.5983, "step": 484 }, { "epoch": 1.91699604743083, "grad_norm": 1.9315211772918701, "learning_rate": 4.52970297029703e-05, "loss": 1.3842, "step": 485 }, { "epoch": 1.9209486166007905, "grad_norm": 1.895024061203003, "learning_rate": 4.528712871287129e-05, "loss": 1.4591, "step": 486 }, { "epoch": 1.924901185770751, "grad_norm": 1.6127315759658813, "learning_rate": 4.527722772277228e-05, "loss": 1.4084, "step": 487 }, { "epoch": 1.9288537549407114, "grad_norm": 2.004408121109009, "learning_rate": 4.526732673267327e-05, "loss": 1.1735, "step": 488 }, { "epoch": 1.9328063241106719, "grad_norm": 1.909749150276184, "learning_rate": 4.525742574257426e-05, "loss": 1.4131, "step": 489 }, { "epoch": 1.9367588932806323, "grad_norm": 2.3290045261383057, "learning_rate": 4.524752475247525e-05, "loss": 1.3486, "step": 490 }, { "epoch": 1.9407114624505928, "grad_norm": 1.6117582321166992, "learning_rate": 4.5237623762376244e-05, "loss": 1.3085, "step": 491 }, { "epoch": 1.9446640316205532, "grad_norm": 1.4751336574554443, "learning_rate": 4.522772277227723e-05, "loss": 1.1964, "step": 492 }, { "epoch": 1.9486166007905137, "grad_norm": 1.6752618551254272, "learning_rate": 4.521782178217822e-05, "loss": 1.3035, "step": 493 }, { "epoch": 1.9525691699604744, "grad_norm": 1.6711186170578003, "learning_rate": 4.520792079207921e-05, "loss": 1.328, "step": 494 }, { "epoch": 1.9565217391304348, "grad_norm": 2.0698461532592773, "learning_rate": 4.51980198019802e-05, "loss": 1.3439, "step": 495 }, { "epoch": 1.9604743083003953, "grad_norm": 2.2718327045440674, "learning_rate": 4.518811881188119e-05, "loss": 1.2004, "step": 496 }, { "epoch": 1.9644268774703557, "grad_norm": 1.431687593460083, "learning_rate": 4.5178217821782184e-05, "loss": 1.4495, "step": 497 }, { "epoch": 1.9683794466403162, "grad_norm": 1.991818904876709, "learning_rate": 4.5168316831683175e-05, "loss": 1.296, "step": 498 }, { "epoch": 1.9723320158102768, "grad_norm": 2.0060036182403564, "learning_rate": 4.515841584158416e-05, "loss": 1.257, "step": 499 }, { "epoch": 1.9762845849802373, "grad_norm": 1.9507633447647095, "learning_rate": 4.514851485148515e-05, "loss": 1.1858, "step": 500 }, { "epoch": 1.9802371541501977, "grad_norm": 1.9538617134094238, "learning_rate": 4.513861386138614e-05, "loss": 1.3933, "step": 501 }, { "epoch": 1.9841897233201582, "grad_norm": 1.5946288108825684, "learning_rate": 4.512871287128713e-05, "loss": 1.4485, "step": 502 }, { "epoch": 1.9881422924901186, "grad_norm": 2.8496248722076416, "learning_rate": 4.5118811881188124e-05, "loss": 1.3571, "step": 503 }, { "epoch": 1.992094861660079, "grad_norm": 2.015263795852661, "learning_rate": 4.5108910891089115e-05, "loss": 1.1728, "step": 504 }, { "epoch": 1.9960474308300395, "grad_norm": 1.5661966800689697, "learning_rate": 4.5099009900990106e-05, "loss": 1.2938, "step": 505 }, { "epoch": 2.0, "grad_norm": 1.962953805923462, "learning_rate": 4.508910891089109e-05, "loss": 1.3947, "step": 506 }, { "epoch": 2.0039525691699605, "grad_norm": 1.5289802551269531, "learning_rate": 4.507920792079208e-05, "loss": 0.9196, "step": 507 }, { "epoch": 2.007905138339921, "grad_norm": 1.9140394926071167, "learning_rate": 4.506930693069307e-05, "loss": 0.8456, "step": 508 }, { "epoch": 2.0118577075098814, "grad_norm": 2.3913121223449707, "learning_rate": 4.505940594059406e-05, "loss": 1.0999, "step": 509 }, { "epoch": 2.015810276679842, "grad_norm": 2.737964630126953, "learning_rate": 4.5049504950495054e-05, "loss": 0.937, "step": 510 }, { "epoch": 2.0197628458498023, "grad_norm": 3.1837165355682373, "learning_rate": 4.5039603960396045e-05, "loss": 1.0765, "step": 511 }, { "epoch": 2.0237154150197627, "grad_norm": 2.5629496574401855, "learning_rate": 4.502970297029703e-05, "loss": 0.8658, "step": 512 }, { "epoch": 2.027667984189723, "grad_norm": 3.653461456298828, "learning_rate": 4.501980198019802e-05, "loss": 0.936, "step": 513 }, { "epoch": 2.0316205533596836, "grad_norm": 2.618825674057007, "learning_rate": 4.500990099009901e-05, "loss": 1.1728, "step": 514 }, { "epoch": 2.035573122529644, "grad_norm": 2.738436460494995, "learning_rate": 4.5e-05, "loss": 1.0414, "step": 515 }, { "epoch": 2.039525691699605, "grad_norm": 2.4087257385253906, "learning_rate": 4.4990099009900994e-05, "loss": 1.2158, "step": 516 }, { "epoch": 2.0434782608695654, "grad_norm": 2.185291051864624, "learning_rate": 4.4980198019801985e-05, "loss": 0.9252, "step": 517 }, { "epoch": 2.047430830039526, "grad_norm": 1.9906668663024902, "learning_rate": 4.4970297029702976e-05, "loss": 0.8715, "step": 518 }, { "epoch": 2.0513833992094863, "grad_norm": 2.337731122970581, "learning_rate": 4.496039603960396e-05, "loss": 0.9453, "step": 519 }, { "epoch": 2.0553359683794468, "grad_norm": 2.298905372619629, "learning_rate": 4.495049504950495e-05, "loss": 0.9059, "step": 520 }, { "epoch": 2.059288537549407, "grad_norm": 2.6414473056793213, "learning_rate": 4.494059405940594e-05, "loss": 1.1508, "step": 521 }, { "epoch": 2.0632411067193677, "grad_norm": 2.558394432067871, "learning_rate": 4.4930693069306934e-05, "loss": 1.103, "step": 522 }, { "epoch": 2.067193675889328, "grad_norm": 2.5661203861236572, "learning_rate": 4.4920792079207925e-05, "loss": 0.9934, "step": 523 }, { "epoch": 2.0711462450592886, "grad_norm": 2.458836317062378, "learning_rate": 4.4910891089108916e-05, "loss": 0.9717, "step": 524 }, { "epoch": 2.075098814229249, "grad_norm": 2.6270315647125244, "learning_rate": 4.490099009900991e-05, "loss": 1.1437, "step": 525 }, { "epoch": 2.0790513833992095, "grad_norm": 2.641350030899048, "learning_rate": 4.489108910891089e-05, "loss": 0.9379, "step": 526 }, { "epoch": 2.08300395256917, "grad_norm": 2.517012596130371, "learning_rate": 4.488118811881188e-05, "loss": 0.9568, "step": 527 }, { "epoch": 2.0869565217391304, "grad_norm": 1.8660167455673218, "learning_rate": 4.487128712871287e-05, "loss": 0.883, "step": 528 }, { "epoch": 2.090909090909091, "grad_norm": 2.767169713973999, "learning_rate": 4.4861386138613864e-05, "loss": 0.6253, "step": 529 }, { "epoch": 2.0948616600790513, "grad_norm": 2.228656768798828, "learning_rate": 4.4851485148514855e-05, "loss": 1.0468, "step": 530 }, { "epoch": 2.0988142292490117, "grad_norm": 2.4214541912078857, "learning_rate": 4.4841584158415846e-05, "loss": 1.166, "step": 531 }, { "epoch": 2.102766798418972, "grad_norm": 2.0022764205932617, "learning_rate": 4.483168316831684e-05, "loss": 0.8676, "step": 532 }, { "epoch": 2.1067193675889326, "grad_norm": 3.2772185802459717, "learning_rate": 4.482178217821782e-05, "loss": 0.9662, "step": 533 }, { "epoch": 2.110671936758893, "grad_norm": 1.924759030342102, "learning_rate": 4.481188118811881e-05, "loss": 1.006, "step": 534 }, { "epoch": 2.1146245059288535, "grad_norm": 1.97169029712677, "learning_rate": 4.4801980198019804e-05, "loss": 0.8718, "step": 535 }, { "epoch": 2.1185770750988144, "grad_norm": 2.0296850204467773, "learning_rate": 4.4792079207920795e-05, "loss": 0.9817, "step": 536 }, { "epoch": 2.122529644268775, "grad_norm": 2.1732065677642822, "learning_rate": 4.4782178217821786e-05, "loss": 0.8081, "step": 537 }, { "epoch": 2.1264822134387353, "grad_norm": 2.146390914916992, "learning_rate": 4.477227722772278e-05, "loss": 0.9508, "step": 538 }, { "epoch": 2.130434782608696, "grad_norm": 2.7521779537200928, "learning_rate": 4.476237623762377e-05, "loss": 1.0425, "step": 539 }, { "epoch": 2.1343873517786562, "grad_norm": 2.5437231063842773, "learning_rate": 4.475247524752475e-05, "loss": 1.0022, "step": 540 }, { "epoch": 2.1383399209486167, "grad_norm": 1.9837696552276611, "learning_rate": 4.4742574257425743e-05, "loss": 0.9399, "step": 541 }, { "epoch": 2.142292490118577, "grad_norm": 2.9563467502593994, "learning_rate": 4.4732673267326735e-05, "loss": 0.8627, "step": 542 }, { "epoch": 2.1462450592885376, "grad_norm": 2.4109456539154053, "learning_rate": 4.4722772277227726e-05, "loss": 0.9322, "step": 543 }, { "epoch": 2.150197628458498, "grad_norm": 2.616680145263672, "learning_rate": 4.471287128712872e-05, "loss": 1.0157, "step": 544 }, { "epoch": 2.1541501976284585, "grad_norm": 2.143428087234497, "learning_rate": 4.470297029702971e-05, "loss": 0.8327, "step": 545 }, { "epoch": 2.158102766798419, "grad_norm": 2.4112675189971924, "learning_rate": 4.46930693069307e-05, "loss": 0.6814, "step": 546 }, { "epoch": 2.1620553359683794, "grad_norm": 2.6342999935150146, "learning_rate": 4.468316831683168e-05, "loss": 0.8047, "step": 547 }, { "epoch": 2.16600790513834, "grad_norm": 2.10105562210083, "learning_rate": 4.4673267326732674e-05, "loss": 0.8877, "step": 548 }, { "epoch": 2.1699604743083003, "grad_norm": 1.8829654455184937, "learning_rate": 4.4663366336633665e-05, "loss": 0.8548, "step": 549 }, { "epoch": 2.1739130434782608, "grad_norm": 2.355497360229492, "learning_rate": 4.4653465346534656e-05, "loss": 1.1907, "step": 550 }, { "epoch": 2.177865612648221, "grad_norm": 2.614638566970825, "learning_rate": 4.464356435643565e-05, "loss": 1.0278, "step": 551 }, { "epoch": 2.1818181818181817, "grad_norm": 1.9568201303482056, "learning_rate": 4.463366336633664e-05, "loss": 0.9075, "step": 552 }, { "epoch": 2.185770750988142, "grad_norm": 2.7377095222473145, "learning_rate": 4.462376237623763e-05, "loss": 0.825, "step": 553 }, { "epoch": 2.1897233201581026, "grad_norm": 2.07133412361145, "learning_rate": 4.4613861386138614e-05, "loss": 0.9024, "step": 554 }, { "epoch": 2.1936758893280635, "grad_norm": 2.981724262237549, "learning_rate": 4.4603960396039605e-05, "loss": 1.0011, "step": 555 }, { "epoch": 2.197628458498024, "grad_norm": 2.2700235843658447, "learning_rate": 4.4594059405940596e-05, "loss": 1.0149, "step": 556 }, { "epoch": 2.2015810276679844, "grad_norm": 2.921574115753174, "learning_rate": 4.458415841584159e-05, "loss": 1.0805, "step": 557 }, { "epoch": 2.205533596837945, "grad_norm": 2.062511444091797, "learning_rate": 4.457425742574258e-05, "loss": 0.876, "step": 558 }, { "epoch": 2.2094861660079053, "grad_norm": 2.197080612182617, "learning_rate": 4.456435643564357e-05, "loss": 0.9358, "step": 559 }, { "epoch": 2.2134387351778657, "grad_norm": 2.5579307079315186, "learning_rate": 4.455445544554456e-05, "loss": 0.8697, "step": 560 }, { "epoch": 2.217391304347826, "grad_norm": 2.3534817695617676, "learning_rate": 4.4544554455445544e-05, "loss": 0.7327, "step": 561 }, { "epoch": 2.2213438735177866, "grad_norm": 2.183434009552002, "learning_rate": 4.4534653465346536e-05, "loss": 0.9254, "step": 562 }, { "epoch": 2.225296442687747, "grad_norm": 2.156205415725708, "learning_rate": 4.4524752475247527e-05, "loss": 1.0098, "step": 563 }, { "epoch": 2.2292490118577075, "grad_norm": 2.701768636703491, "learning_rate": 4.451485148514852e-05, "loss": 0.8127, "step": 564 }, { "epoch": 2.233201581027668, "grad_norm": 1.9933669567108154, "learning_rate": 4.450495049504951e-05, "loss": 1.0148, "step": 565 }, { "epoch": 2.2371541501976284, "grad_norm": 2.4481515884399414, "learning_rate": 4.44950495049505e-05, "loss": 0.9259, "step": 566 }, { "epoch": 2.241106719367589, "grad_norm": 2.540609359741211, "learning_rate": 4.448514851485149e-05, "loss": 0.81, "step": 567 }, { "epoch": 2.2450592885375493, "grad_norm": 1.9956486225128174, "learning_rate": 4.4475247524752475e-05, "loss": 0.8255, "step": 568 }, { "epoch": 2.2490118577075098, "grad_norm": 2.4563422203063965, "learning_rate": 4.4465346534653466e-05, "loss": 0.9576, "step": 569 }, { "epoch": 2.2529644268774702, "grad_norm": 3.125199794769287, "learning_rate": 4.445544554455446e-05, "loss": 1.0371, "step": 570 }, { "epoch": 2.2569169960474307, "grad_norm": 2.674996852874756, "learning_rate": 4.444554455445545e-05, "loss": 1.0372, "step": 571 }, { "epoch": 2.260869565217391, "grad_norm": 2.516646146774292, "learning_rate": 4.443564356435644e-05, "loss": 0.7121, "step": 572 }, { "epoch": 2.2648221343873516, "grad_norm": 2.1260321140289307, "learning_rate": 4.442574257425743e-05, "loss": 0.8071, "step": 573 }, { "epoch": 2.2687747035573125, "grad_norm": 1.982050895690918, "learning_rate": 4.441584158415842e-05, "loss": 0.9094, "step": 574 }, { "epoch": 2.2727272727272725, "grad_norm": 2.090538263320923, "learning_rate": 4.4405940594059406e-05, "loss": 1.0343, "step": 575 }, { "epoch": 2.2766798418972334, "grad_norm": 2.2843616008758545, "learning_rate": 4.43960396039604e-05, "loss": 1.0341, "step": 576 }, { "epoch": 2.280632411067194, "grad_norm": 2.686245918273926, "learning_rate": 4.438613861386139e-05, "loss": 0.9317, "step": 577 }, { "epoch": 2.2845849802371543, "grad_norm": 2.6296350955963135, "learning_rate": 4.437623762376238e-05, "loss": 0.9902, "step": 578 }, { "epoch": 2.2885375494071147, "grad_norm": 2.013939142227173, "learning_rate": 4.436633663366337e-05, "loss": 0.8567, "step": 579 }, { "epoch": 2.292490118577075, "grad_norm": 1.8960076570510864, "learning_rate": 4.435643564356436e-05, "loss": 0.9057, "step": 580 }, { "epoch": 2.2964426877470356, "grad_norm": 2.0619733333587646, "learning_rate": 4.434653465346535e-05, "loss": 1.0566, "step": 581 }, { "epoch": 2.300395256916996, "grad_norm": 2.267512321472168, "learning_rate": 4.4336633663366336e-05, "loss": 1.1003, "step": 582 }, { "epoch": 2.3043478260869565, "grad_norm": 2.7851717472076416, "learning_rate": 4.432673267326733e-05, "loss": 0.8922, "step": 583 }, { "epoch": 2.308300395256917, "grad_norm": 2.3739750385284424, "learning_rate": 4.431683168316832e-05, "loss": 0.7786, "step": 584 }, { "epoch": 2.3122529644268774, "grad_norm": 3.0041046142578125, "learning_rate": 4.430693069306931e-05, "loss": 1.2571, "step": 585 }, { "epoch": 2.316205533596838, "grad_norm": 2.265202283859253, "learning_rate": 4.42970297029703e-05, "loss": 1.05, "step": 586 }, { "epoch": 2.3201581027667983, "grad_norm": 1.930483341217041, "learning_rate": 4.428712871287129e-05, "loss": 0.8879, "step": 587 }, { "epoch": 2.324110671936759, "grad_norm": 2.2211341857910156, "learning_rate": 4.427722772277228e-05, "loss": 0.9002, "step": 588 }, { "epoch": 2.3280632411067192, "grad_norm": 2.468590021133423, "learning_rate": 4.426732673267327e-05, "loss": 0.8709, "step": 589 }, { "epoch": 2.3320158102766797, "grad_norm": 2.0975170135498047, "learning_rate": 4.425742574257426e-05, "loss": 0.8501, "step": 590 }, { "epoch": 2.33596837944664, "grad_norm": 3.9977915287017822, "learning_rate": 4.424752475247525e-05, "loss": 0.9317, "step": 591 }, { "epoch": 2.3399209486166006, "grad_norm": 2.26987886428833, "learning_rate": 4.423762376237624e-05, "loss": 0.8817, "step": 592 }, { "epoch": 2.3438735177865615, "grad_norm": 2.413039445877075, "learning_rate": 4.422772277227723e-05, "loss": 0.7044, "step": 593 }, { "epoch": 2.3478260869565215, "grad_norm": 2.449648857116699, "learning_rate": 4.421782178217822e-05, "loss": 1.1089, "step": 594 }, { "epoch": 2.3517786561264824, "grad_norm": 2.1894166469573975, "learning_rate": 4.4207920792079214e-05, "loss": 0.9863, "step": 595 }, { "epoch": 2.355731225296443, "grad_norm": 1.9356539249420166, "learning_rate": 4.41980198019802e-05, "loss": 0.9989, "step": 596 }, { "epoch": 2.3596837944664033, "grad_norm": 2.5594429969787598, "learning_rate": 4.418811881188119e-05, "loss": 1.1583, "step": 597 }, { "epoch": 2.3636363636363638, "grad_norm": 1.7523540258407593, "learning_rate": 4.417821782178218e-05, "loss": 0.7049, "step": 598 }, { "epoch": 2.367588932806324, "grad_norm": 2.0250518321990967, "learning_rate": 4.416831683168317e-05, "loss": 0.8876, "step": 599 }, { "epoch": 2.3715415019762847, "grad_norm": 2.218608856201172, "learning_rate": 4.415841584158416e-05, "loss": 0.7953, "step": 600 }, { "epoch": 2.375494071146245, "grad_norm": 2.6930036544799805, "learning_rate": 4.414851485148515e-05, "loss": 1.0615, "step": 601 }, { "epoch": 2.3794466403162056, "grad_norm": 2.345737934112549, "learning_rate": 4.4138613861386144e-05, "loss": 1.1687, "step": 602 }, { "epoch": 2.383399209486166, "grad_norm": 2.5030248165130615, "learning_rate": 4.412871287128713e-05, "loss": 0.9805, "step": 603 }, { "epoch": 2.3873517786561265, "grad_norm": 2.123748302459717, "learning_rate": 4.411881188118812e-05, "loss": 0.9928, "step": 604 }, { "epoch": 2.391304347826087, "grad_norm": 2.076786994934082, "learning_rate": 4.410891089108911e-05, "loss": 0.9201, "step": 605 }, { "epoch": 2.3952569169960474, "grad_norm": 2.1941967010498047, "learning_rate": 4.40990099009901e-05, "loss": 0.97, "step": 606 }, { "epoch": 2.399209486166008, "grad_norm": 2.5832958221435547, "learning_rate": 4.408910891089109e-05, "loss": 1.0131, "step": 607 }, { "epoch": 2.4031620553359683, "grad_norm": 3.0189125537872314, "learning_rate": 4.4079207920792084e-05, "loss": 1.0661, "step": 608 }, { "epoch": 2.4071146245059287, "grad_norm": 2.2533063888549805, "learning_rate": 4.4069306930693075e-05, "loss": 0.8428, "step": 609 }, { "epoch": 2.411067193675889, "grad_norm": 2.353600263595581, "learning_rate": 4.405940594059406e-05, "loss": 0.9634, "step": 610 }, { "epoch": 2.4150197628458496, "grad_norm": 2.320209503173828, "learning_rate": 4.404950495049505e-05, "loss": 0.9109, "step": 611 }, { "epoch": 2.4189723320158105, "grad_norm": 2.3635990619659424, "learning_rate": 4.403960396039604e-05, "loss": 0.9215, "step": 612 }, { "epoch": 2.4229249011857705, "grad_norm": 2.695899724960327, "learning_rate": 4.402970297029703e-05, "loss": 0.9744, "step": 613 }, { "epoch": 2.4268774703557314, "grad_norm": 2.3702011108398438, "learning_rate": 4.4019801980198024e-05, "loss": 0.7991, "step": 614 }, { "epoch": 2.430830039525692, "grad_norm": 2.5832464694976807, "learning_rate": 4.4009900990099015e-05, "loss": 0.8881, "step": 615 }, { "epoch": 2.4347826086956523, "grad_norm": 2.076077461242676, "learning_rate": 4.4000000000000006e-05, "loss": 1.1643, "step": 616 }, { "epoch": 2.438735177865613, "grad_norm": 2.1855406761169434, "learning_rate": 4.399009900990099e-05, "loss": 0.8196, "step": 617 }, { "epoch": 2.4426877470355732, "grad_norm": 2.2200517654418945, "learning_rate": 4.398019801980198e-05, "loss": 1.0685, "step": 618 }, { "epoch": 2.4466403162055337, "grad_norm": 3.2805979251861572, "learning_rate": 4.397029702970297e-05, "loss": 0.7403, "step": 619 }, { "epoch": 2.450592885375494, "grad_norm": 2.392289876937866, "learning_rate": 4.396039603960396e-05, "loss": 0.9324, "step": 620 }, { "epoch": 2.4545454545454546, "grad_norm": 2.4792675971984863, "learning_rate": 4.3950495049504954e-05, "loss": 0.9203, "step": 621 }, { "epoch": 2.458498023715415, "grad_norm": 2.440727710723877, "learning_rate": 4.3940594059405945e-05, "loss": 1.1504, "step": 622 }, { "epoch": 2.4624505928853755, "grad_norm": 2.4961142539978027, "learning_rate": 4.3930693069306936e-05, "loss": 0.8575, "step": 623 }, { "epoch": 2.466403162055336, "grad_norm": 2.6662189960479736, "learning_rate": 4.392079207920792e-05, "loss": 1.1297, "step": 624 }, { "epoch": 2.4703557312252964, "grad_norm": 1.9681288003921509, "learning_rate": 4.391089108910891e-05, "loss": 0.8209, "step": 625 }, { "epoch": 2.474308300395257, "grad_norm": 2.3691046237945557, "learning_rate": 4.39009900990099e-05, "loss": 0.9866, "step": 626 }, { "epoch": 2.4782608695652173, "grad_norm": 2.217207670211792, "learning_rate": 4.3891089108910894e-05, "loss": 0.8097, "step": 627 }, { "epoch": 2.4822134387351777, "grad_norm": 2.163919448852539, "learning_rate": 4.3881188118811885e-05, "loss": 0.8636, "step": 628 }, { "epoch": 2.486166007905138, "grad_norm": 2.5630953311920166, "learning_rate": 4.3871287128712876e-05, "loss": 0.8747, "step": 629 }, { "epoch": 2.4901185770750986, "grad_norm": 2.313148260116577, "learning_rate": 4.386138613861386e-05, "loss": 0.9534, "step": 630 }, { "epoch": 2.494071146245059, "grad_norm": 2.6939351558685303, "learning_rate": 4.385148514851485e-05, "loss": 0.9823, "step": 631 }, { "epoch": 2.4980237154150196, "grad_norm": 2.1868441104888916, "learning_rate": 4.384158415841584e-05, "loss": 0.9308, "step": 632 }, { "epoch": 2.5019762845849804, "grad_norm": 1.9445887804031372, "learning_rate": 4.3831683168316833e-05, "loss": 0.853, "step": 633 }, { "epoch": 2.5059288537549405, "grad_norm": 2.340928077697754, "learning_rate": 4.3821782178217824e-05, "loss": 0.8946, "step": 634 }, { "epoch": 2.5098814229249014, "grad_norm": 2.856757402420044, "learning_rate": 4.3811881188118816e-05, "loss": 0.7781, "step": 635 }, { "epoch": 2.513833992094862, "grad_norm": 2.6104776859283447, "learning_rate": 4.380198019801981e-05, "loss": 0.7235, "step": 636 }, { "epoch": 2.5177865612648223, "grad_norm": 2.776491403579712, "learning_rate": 4.379207920792079e-05, "loss": 0.9209, "step": 637 }, { "epoch": 2.5217391304347827, "grad_norm": 2.5769119262695312, "learning_rate": 4.378217821782178e-05, "loss": 0.8636, "step": 638 }, { "epoch": 2.525691699604743, "grad_norm": 1.977169156074524, "learning_rate": 4.377227722772277e-05, "loss": 0.9887, "step": 639 }, { "epoch": 2.5296442687747036, "grad_norm": 2.3627588748931885, "learning_rate": 4.3762376237623764e-05, "loss": 0.8129, "step": 640 }, { "epoch": 2.533596837944664, "grad_norm": 1.8942148685455322, "learning_rate": 4.3752475247524755e-05, "loss": 0.9455, "step": 641 }, { "epoch": 2.5375494071146245, "grad_norm": 2.17875075340271, "learning_rate": 4.3742574257425746e-05, "loss": 0.947, "step": 642 }, { "epoch": 2.541501976284585, "grad_norm": 1.8505522012710571, "learning_rate": 4.373267326732674e-05, "loss": 0.8869, "step": 643 }, { "epoch": 2.5454545454545454, "grad_norm": 2.101501941680908, "learning_rate": 4.372277227722772e-05, "loss": 0.8892, "step": 644 }, { "epoch": 2.549407114624506, "grad_norm": 3.1386516094207764, "learning_rate": 4.371287128712871e-05, "loss": 1.0419, "step": 645 }, { "epoch": 2.5533596837944663, "grad_norm": 2.5623793601989746, "learning_rate": 4.3702970297029704e-05, "loss": 0.7749, "step": 646 }, { "epoch": 2.5573122529644268, "grad_norm": 2.369387626647949, "learning_rate": 4.3693069306930695e-05, "loss": 0.9459, "step": 647 }, { "epoch": 2.561264822134387, "grad_norm": 2.3831751346588135, "learning_rate": 4.3683168316831686e-05, "loss": 0.9535, "step": 648 }, { "epoch": 2.5652173913043477, "grad_norm": 2.5698206424713135, "learning_rate": 4.367326732673268e-05, "loss": 0.9413, "step": 649 }, { "epoch": 2.5691699604743086, "grad_norm": 2.4377827644348145, "learning_rate": 4.366336633663367e-05, "loss": 0.7042, "step": 650 }, { "epoch": 2.5731225296442686, "grad_norm": 2.5129380226135254, "learning_rate": 4.365346534653465e-05, "loss": 0.7327, "step": 651 }, { "epoch": 2.5770750988142295, "grad_norm": 2.4875426292419434, "learning_rate": 4.364356435643564e-05, "loss": 0.8791, "step": 652 }, { "epoch": 2.5810276679841895, "grad_norm": 2.3245186805725098, "learning_rate": 4.3633663366336634e-05, "loss": 1.0295, "step": 653 }, { "epoch": 2.5849802371541504, "grad_norm": 2.2866084575653076, "learning_rate": 4.3623762376237625e-05, "loss": 0.8515, "step": 654 }, { "epoch": 2.588932806324111, "grad_norm": 2.086591958999634, "learning_rate": 4.3613861386138617e-05, "loss": 0.7583, "step": 655 }, { "epoch": 2.5928853754940713, "grad_norm": 2.5532939434051514, "learning_rate": 4.360396039603961e-05, "loss": 0.81, "step": 656 }, { "epoch": 2.5968379446640317, "grad_norm": 2.86692476272583, "learning_rate": 4.35940594059406e-05, "loss": 0.8981, "step": 657 }, { "epoch": 2.600790513833992, "grad_norm": 2.8362505435943604, "learning_rate": 4.358415841584158e-05, "loss": 0.7082, "step": 658 }, { "epoch": 2.6047430830039526, "grad_norm": 2.603944778442383, "learning_rate": 4.3574257425742574e-05, "loss": 1.0281, "step": 659 }, { "epoch": 2.608695652173913, "grad_norm": 2.2189369201660156, "learning_rate": 4.3564356435643565e-05, "loss": 0.8436, "step": 660 }, { "epoch": 2.6126482213438735, "grad_norm": 2.7203996181488037, "learning_rate": 4.3554455445544556e-05, "loss": 1.383, "step": 661 }, { "epoch": 2.616600790513834, "grad_norm": 1.9215599298477173, "learning_rate": 4.354455445544555e-05, "loss": 0.7793, "step": 662 }, { "epoch": 2.6205533596837944, "grad_norm": 2.0869908332824707, "learning_rate": 4.353465346534654e-05, "loss": 0.9128, "step": 663 }, { "epoch": 2.624505928853755, "grad_norm": 2.3292601108551025, "learning_rate": 4.352475247524753e-05, "loss": 1.0712, "step": 664 }, { "epoch": 2.6284584980237153, "grad_norm": 2.7547478675842285, "learning_rate": 4.3514851485148514e-05, "loss": 0.9882, "step": 665 }, { "epoch": 2.632411067193676, "grad_norm": 2.8896234035491943, "learning_rate": 4.3504950495049505e-05, "loss": 1.0682, "step": 666 }, { "epoch": 2.6363636363636362, "grad_norm": 2.0239450931549072, "learning_rate": 4.3495049504950496e-05, "loss": 0.9413, "step": 667 }, { "epoch": 2.6403162055335967, "grad_norm": 3.359004020690918, "learning_rate": 4.348514851485149e-05, "loss": 0.6784, "step": 668 }, { "epoch": 2.6442687747035576, "grad_norm": 2.7798683643341064, "learning_rate": 4.347524752475248e-05, "loss": 0.9408, "step": 669 }, { "epoch": 2.6482213438735176, "grad_norm": 1.9975160360336304, "learning_rate": 4.346534653465347e-05, "loss": 0.953, "step": 670 }, { "epoch": 2.6521739130434785, "grad_norm": 2.2007336616516113, "learning_rate": 4.345544554455446e-05, "loss": 0.9499, "step": 671 }, { "epoch": 2.6561264822134385, "grad_norm": 3.028135299682617, "learning_rate": 4.3445544554455444e-05, "loss": 0.8387, "step": 672 }, { "epoch": 2.6600790513833994, "grad_norm": 3.0381453037261963, "learning_rate": 4.3435643564356435e-05, "loss": 1.077, "step": 673 }, { "epoch": 2.66403162055336, "grad_norm": 2.4370057582855225, "learning_rate": 4.3425742574257426e-05, "loss": 0.8711, "step": 674 }, { "epoch": 2.6679841897233203, "grad_norm": 2.2608211040496826, "learning_rate": 4.341584158415842e-05, "loss": 0.6876, "step": 675 }, { "epoch": 2.6719367588932808, "grad_norm": 5.736217975616455, "learning_rate": 4.340594059405941e-05, "loss": 0.9116, "step": 676 }, { "epoch": 2.675889328063241, "grad_norm": 2.4403767585754395, "learning_rate": 4.33960396039604e-05, "loss": 0.788, "step": 677 }, { "epoch": 2.6798418972332017, "grad_norm": 2.404578924179077, "learning_rate": 4.338613861386139e-05, "loss": 0.7621, "step": 678 }, { "epoch": 2.683794466403162, "grad_norm": 3.339783191680908, "learning_rate": 4.3376237623762375e-05, "loss": 1.0514, "step": 679 }, { "epoch": 2.6877470355731226, "grad_norm": 2.9524576663970947, "learning_rate": 4.3366336633663366e-05, "loss": 0.8482, "step": 680 }, { "epoch": 2.691699604743083, "grad_norm": 2.888458013534546, "learning_rate": 4.335643564356436e-05, "loss": 0.9797, "step": 681 }, { "epoch": 2.6956521739130435, "grad_norm": 2.987743854522705, "learning_rate": 4.334653465346535e-05, "loss": 1.0822, "step": 682 }, { "epoch": 2.699604743083004, "grad_norm": 2.1577627658843994, "learning_rate": 4.333663366336634e-05, "loss": 1.0429, "step": 683 }, { "epoch": 2.7035573122529644, "grad_norm": 2.7446835041046143, "learning_rate": 4.332673267326733e-05, "loss": 1.312, "step": 684 }, { "epoch": 2.707509881422925, "grad_norm": 2.3941147327423096, "learning_rate": 4.331683168316832e-05, "loss": 1.0169, "step": 685 }, { "epoch": 2.7114624505928853, "grad_norm": 2.1516363620758057, "learning_rate": 4.3306930693069306e-05, "loss": 0.8445, "step": 686 }, { "epoch": 2.7154150197628457, "grad_norm": 1.9346096515655518, "learning_rate": 4.32970297029703e-05, "loss": 0.8786, "step": 687 }, { "epoch": 2.719367588932806, "grad_norm": 1.7370835542678833, "learning_rate": 4.328712871287129e-05, "loss": 1.0397, "step": 688 }, { "epoch": 2.7233201581027666, "grad_norm": 2.3953020572662354, "learning_rate": 4.327722772277228e-05, "loss": 0.8304, "step": 689 }, { "epoch": 2.7272727272727275, "grad_norm": 3.269364356994629, "learning_rate": 4.326732673267327e-05, "loss": 1.0926, "step": 690 }, { "epoch": 2.7312252964426875, "grad_norm": 2.2393388748168945, "learning_rate": 4.325742574257426e-05, "loss": 0.9695, "step": 691 }, { "epoch": 2.7351778656126484, "grad_norm": 2.2462520599365234, "learning_rate": 4.324752475247525e-05, "loss": 1.0486, "step": 692 }, { "epoch": 2.7391304347826084, "grad_norm": 2.1283669471740723, "learning_rate": 4.3237623762376236e-05, "loss": 0.9946, "step": 693 }, { "epoch": 2.7430830039525693, "grad_norm": 2.211625099182129, "learning_rate": 4.322772277227723e-05, "loss": 0.8475, "step": 694 }, { "epoch": 2.7470355731225298, "grad_norm": 2.0631792545318604, "learning_rate": 4.321782178217822e-05, "loss": 1.0292, "step": 695 }, { "epoch": 2.7509881422924902, "grad_norm": 2.638406276702881, "learning_rate": 4.320792079207921e-05, "loss": 0.8373, "step": 696 }, { "epoch": 2.7549407114624507, "grad_norm": 2.304776430130005, "learning_rate": 4.31980198019802e-05, "loss": 0.7593, "step": 697 }, { "epoch": 2.758893280632411, "grad_norm": 2.801795244216919, "learning_rate": 4.318811881188119e-05, "loss": 0.9151, "step": 698 }, { "epoch": 2.7628458498023716, "grad_norm": 2.3793420791625977, "learning_rate": 4.317821782178218e-05, "loss": 0.7194, "step": 699 }, { "epoch": 2.766798418972332, "grad_norm": 2.7071168422698975, "learning_rate": 4.316831683168317e-05, "loss": 1.0741, "step": 700 }, { "epoch": 2.7707509881422925, "grad_norm": 2.374814987182617, "learning_rate": 4.315841584158416e-05, "loss": 0.8991, "step": 701 }, { "epoch": 2.774703557312253, "grad_norm": 2.462678909301758, "learning_rate": 4.314851485148515e-05, "loss": 0.9469, "step": 702 }, { "epoch": 2.7786561264822134, "grad_norm": 2.4586679935455322, "learning_rate": 4.313861386138614e-05, "loss": 0.9867, "step": 703 }, { "epoch": 2.782608695652174, "grad_norm": 2.261760950088501, "learning_rate": 4.312871287128713e-05, "loss": 0.9321, "step": 704 }, { "epoch": 2.7865612648221343, "grad_norm": 2.5425806045532227, "learning_rate": 4.311881188118812e-05, "loss": 1.0651, "step": 705 }, { "epoch": 2.7905138339920947, "grad_norm": 2.6827871799468994, "learning_rate": 4.3108910891089113e-05, "loss": 1.0182, "step": 706 }, { "epoch": 2.794466403162055, "grad_norm": 1.9536657333374023, "learning_rate": 4.30990099009901e-05, "loss": 1.0023, "step": 707 }, { "epoch": 2.7984189723320156, "grad_norm": 2.4319167137145996, "learning_rate": 4.308910891089109e-05, "loss": 0.9775, "step": 708 }, { "epoch": 2.8023715415019765, "grad_norm": 2.3153910636901855, "learning_rate": 4.307920792079208e-05, "loss": 0.8241, "step": 709 }, { "epoch": 2.8063241106719365, "grad_norm": 1.9590847492218018, "learning_rate": 4.306930693069307e-05, "loss": 1.1188, "step": 710 }, { "epoch": 2.8102766798418974, "grad_norm": 2.610215902328491, "learning_rate": 4.305940594059406e-05, "loss": 1.12, "step": 711 }, { "epoch": 2.8142292490118574, "grad_norm": 2.841136932373047, "learning_rate": 4.304950495049505e-05, "loss": 1.2067, "step": 712 }, { "epoch": 2.8181818181818183, "grad_norm": 2.2919881343841553, "learning_rate": 4.3039603960396044e-05, "loss": 0.9164, "step": 713 }, { "epoch": 2.822134387351779, "grad_norm": 2.4537272453308105, "learning_rate": 4.302970297029703e-05, "loss": 1.0815, "step": 714 }, { "epoch": 2.8260869565217392, "grad_norm": 2.629290819168091, "learning_rate": 4.301980198019802e-05, "loss": 1.0119, "step": 715 }, { "epoch": 2.8300395256916997, "grad_norm": 2.261491537094116, "learning_rate": 4.300990099009901e-05, "loss": 0.951, "step": 716 }, { "epoch": 2.83399209486166, "grad_norm": 2.092386245727539, "learning_rate": 4.3e-05, "loss": 0.9175, "step": 717 }, { "epoch": 2.8379446640316206, "grad_norm": 2.174060344696045, "learning_rate": 4.299009900990099e-05, "loss": 1.1296, "step": 718 }, { "epoch": 2.841897233201581, "grad_norm": 2.556680202484131, "learning_rate": 4.2980198019801984e-05, "loss": 1.0698, "step": 719 }, { "epoch": 2.8458498023715415, "grad_norm": 2.519099235534668, "learning_rate": 4.2970297029702975e-05, "loss": 1.013, "step": 720 }, { "epoch": 2.849802371541502, "grad_norm": 3.334442615509033, "learning_rate": 4.296039603960396e-05, "loss": 0.5598, "step": 721 }, { "epoch": 2.8537549407114624, "grad_norm": 3.269895076751709, "learning_rate": 4.295049504950495e-05, "loss": 1.1371, "step": 722 }, { "epoch": 2.857707509881423, "grad_norm": 2.406057357788086, "learning_rate": 4.294059405940594e-05, "loss": 0.7926, "step": 723 }, { "epoch": 2.8616600790513833, "grad_norm": 2.0798585414886475, "learning_rate": 4.293069306930693e-05, "loss": 0.9634, "step": 724 }, { "epoch": 2.8656126482213438, "grad_norm": 2.6829497814178467, "learning_rate": 4.2920792079207923e-05, "loss": 0.6884, "step": 725 }, { "epoch": 2.869565217391304, "grad_norm": 3.107074499130249, "learning_rate": 4.2910891089108914e-05, "loss": 0.9665, "step": 726 }, { "epoch": 2.8735177865612647, "grad_norm": 2.970815896987915, "learning_rate": 4.2900990099009906e-05, "loss": 1.3717, "step": 727 }, { "epoch": 2.8774703557312256, "grad_norm": 3.4182064533233643, "learning_rate": 4.289108910891089e-05, "loss": 0.736, "step": 728 }, { "epoch": 2.8814229249011856, "grad_norm": 2.1826579570770264, "learning_rate": 4.288118811881188e-05, "loss": 1.0718, "step": 729 }, { "epoch": 2.8853754940711465, "grad_norm": 2.1282830238342285, "learning_rate": 4.287128712871287e-05, "loss": 0.8882, "step": 730 }, { "epoch": 2.8893280632411065, "grad_norm": 2.3411455154418945, "learning_rate": 4.286138613861386e-05, "loss": 1.0242, "step": 731 }, { "epoch": 2.8932806324110674, "grad_norm": 2.0753939151763916, "learning_rate": 4.2851485148514854e-05, "loss": 1.0974, "step": 732 }, { "epoch": 2.897233201581028, "grad_norm": 2.0541632175445557, "learning_rate": 4.2841584158415845e-05, "loss": 0.9515, "step": 733 }, { "epoch": 2.9011857707509883, "grad_norm": 1.9826310873031616, "learning_rate": 4.2831683168316836e-05, "loss": 0.9416, "step": 734 }, { "epoch": 2.9051383399209487, "grad_norm": 2.041292428970337, "learning_rate": 4.282178217821782e-05, "loss": 1.0467, "step": 735 }, { "epoch": 2.909090909090909, "grad_norm": 2.023148775100708, "learning_rate": 4.281188118811881e-05, "loss": 0.7527, "step": 736 }, { "epoch": 2.9130434782608696, "grad_norm": 1.9297372102737427, "learning_rate": 4.28019801980198e-05, "loss": 0.8833, "step": 737 }, { "epoch": 2.91699604743083, "grad_norm": 2.2003679275512695, "learning_rate": 4.2792079207920794e-05, "loss": 1.0502, "step": 738 }, { "epoch": 2.9209486166007905, "grad_norm": 2.891721487045288, "learning_rate": 4.2782178217821785e-05, "loss": 0.6356, "step": 739 }, { "epoch": 2.924901185770751, "grad_norm": 2.4546079635620117, "learning_rate": 4.2772277227722776e-05, "loss": 0.8593, "step": 740 }, { "epoch": 2.9288537549407114, "grad_norm": 2.4899215698242188, "learning_rate": 4.276237623762377e-05, "loss": 1.0728, "step": 741 }, { "epoch": 2.932806324110672, "grad_norm": 3.4617972373962402, "learning_rate": 4.275247524752475e-05, "loss": 0.6087, "step": 742 }, { "epoch": 2.9367588932806323, "grad_norm": 2.6773808002471924, "learning_rate": 4.274257425742574e-05, "loss": 1.0489, "step": 743 }, { "epoch": 2.940711462450593, "grad_norm": 2.0755105018615723, "learning_rate": 4.273267326732673e-05, "loss": 0.8829, "step": 744 }, { "epoch": 2.9446640316205532, "grad_norm": 2.201080322265625, "learning_rate": 4.2722772277227724e-05, "loss": 1.0814, "step": 745 }, { "epoch": 2.9486166007905137, "grad_norm": 1.9096810817718506, "learning_rate": 4.2712871287128715e-05, "loss": 0.8172, "step": 746 }, { "epoch": 2.9525691699604746, "grad_norm": 1.8115899562835693, "learning_rate": 4.2702970297029707e-05, "loss": 0.665, "step": 747 }, { "epoch": 2.9565217391304346, "grad_norm": 2.9084932804107666, "learning_rate": 4.269306930693069e-05, "loss": 0.637, "step": 748 }, { "epoch": 2.9604743083003955, "grad_norm": 2.9316818714141846, "learning_rate": 4.268316831683168e-05, "loss": 0.5777, "step": 749 }, { "epoch": 2.9644268774703555, "grad_norm": 2.852191209793091, "learning_rate": 4.267326732673267e-05, "loss": 0.9825, "step": 750 }, { "epoch": 2.9683794466403164, "grad_norm": 2.142752170562744, "learning_rate": 4.2663366336633664e-05, "loss": 0.8901, "step": 751 }, { "epoch": 2.972332015810277, "grad_norm": 2.532355308532715, "learning_rate": 4.2653465346534655e-05, "loss": 1.2216, "step": 752 }, { "epoch": 2.9762845849802373, "grad_norm": 2.832993268966675, "learning_rate": 4.2643564356435646e-05, "loss": 1.2194, "step": 753 }, { "epoch": 2.9802371541501977, "grad_norm": 2.2944655418395996, "learning_rate": 4.263366336633664e-05, "loss": 1.0729, "step": 754 }, { "epoch": 2.984189723320158, "grad_norm": 2.4382858276367188, "learning_rate": 4.262376237623762e-05, "loss": 1.0537, "step": 755 }, { "epoch": 2.9881422924901186, "grad_norm": 2.0621492862701416, "learning_rate": 4.261386138613861e-05, "loss": 0.899, "step": 756 }, { "epoch": 2.992094861660079, "grad_norm": 2.4617247581481934, "learning_rate": 4.2603960396039604e-05, "loss": 1.0344, "step": 757 }, { "epoch": 2.9960474308300395, "grad_norm": 2.4091153144836426, "learning_rate": 4.2594059405940595e-05, "loss": 0.9987, "step": 758 }, { "epoch": 3.0, "grad_norm": 2.821359395980835, "learning_rate": 4.2584158415841586e-05, "loss": 0.8814, "step": 759 }, { "epoch": 3.0039525691699605, "grad_norm": 2.201995372772217, "learning_rate": 4.257425742574258e-05, "loss": 0.619, "step": 760 }, { "epoch": 3.007905138339921, "grad_norm": 2.3793888092041016, "learning_rate": 4.256435643564357e-05, "loss": 0.4936, "step": 761 }, { "epoch": 3.0118577075098814, "grad_norm": 2.6037724018096924, "learning_rate": 4.255445544554455e-05, "loss": 0.6159, "step": 762 }, { "epoch": 3.015810276679842, "grad_norm": 2.355687141418457, "learning_rate": 4.254455445544554e-05, "loss": 0.5372, "step": 763 }, { "epoch": 3.0197628458498023, "grad_norm": 3.2323098182678223, "learning_rate": 4.2534653465346534e-05, "loss": 0.5334, "step": 764 }, { "epoch": 3.0237154150197627, "grad_norm": 3.8572168350219727, "learning_rate": 4.2524752475247525e-05, "loss": 0.5006, "step": 765 }, { "epoch": 3.027667984189723, "grad_norm": 3.9515655040740967, "learning_rate": 4.2514851485148516e-05, "loss": 0.3603, "step": 766 }, { "epoch": 3.0316205533596836, "grad_norm": 3.514148235321045, "learning_rate": 4.250495049504951e-05, "loss": 0.5465, "step": 767 }, { "epoch": 3.035573122529644, "grad_norm": 3.420142412185669, "learning_rate": 4.24950495049505e-05, "loss": 0.6551, "step": 768 }, { "epoch": 3.039525691699605, "grad_norm": 3.0441970825195312, "learning_rate": 4.248514851485148e-05, "loss": 0.5537, "step": 769 }, { "epoch": 3.0434782608695654, "grad_norm": 2.3210487365722656, "learning_rate": 4.2475247524752474e-05, "loss": 0.4853, "step": 770 }, { "epoch": 3.047430830039526, "grad_norm": 2.848736524581909, "learning_rate": 4.2465346534653465e-05, "loss": 0.4816, "step": 771 }, { "epoch": 3.0513833992094863, "grad_norm": 3.252121925354004, "learning_rate": 4.2455445544554456e-05, "loss": 0.4677, "step": 772 }, { "epoch": 3.0553359683794468, "grad_norm": 2.5904879570007324, "learning_rate": 4.244554455445545e-05, "loss": 0.5243, "step": 773 }, { "epoch": 3.059288537549407, "grad_norm": 2.8725666999816895, "learning_rate": 4.243564356435644e-05, "loss": 0.6084, "step": 774 }, { "epoch": 3.0632411067193677, "grad_norm": 2.661243438720703, "learning_rate": 4.242574257425743e-05, "loss": 0.3559, "step": 775 }, { "epoch": 3.067193675889328, "grad_norm": 2.562589645385742, "learning_rate": 4.2415841584158414e-05, "loss": 0.5622, "step": 776 }, { "epoch": 3.0711462450592886, "grad_norm": 3.9093880653381348, "learning_rate": 4.2405940594059405e-05, "loss": 0.3917, "step": 777 }, { "epoch": 3.075098814229249, "grad_norm": 2.8739657402038574, "learning_rate": 4.2396039603960396e-05, "loss": 0.633, "step": 778 }, { "epoch": 3.0790513833992095, "grad_norm": 2.4997589588165283, "learning_rate": 4.238613861386139e-05, "loss": 0.4124, "step": 779 }, { "epoch": 3.08300395256917, "grad_norm": 2.7031702995300293, "learning_rate": 4.237623762376238e-05, "loss": 0.4767, "step": 780 }, { "epoch": 3.0869565217391304, "grad_norm": 2.741298198699951, "learning_rate": 4.236633663366337e-05, "loss": 0.5227, "step": 781 }, { "epoch": 3.090909090909091, "grad_norm": 3.2926793098449707, "learning_rate": 4.235643564356436e-05, "loss": 0.4507, "step": 782 }, { "epoch": 3.0948616600790513, "grad_norm": 2.5228497982025146, "learning_rate": 4.2346534653465344e-05, "loss": 0.6158, "step": 783 }, { "epoch": 3.0988142292490117, "grad_norm": 3.2704927921295166, "learning_rate": 4.2336633663366335e-05, "loss": 0.5905, "step": 784 }, { "epoch": 3.102766798418972, "grad_norm": 3.2316770553588867, "learning_rate": 4.2326732673267326e-05, "loss": 0.7208, "step": 785 }, { "epoch": 3.1067193675889326, "grad_norm": 2.5651204586029053, "learning_rate": 4.231683168316832e-05, "loss": 0.5822, "step": 786 }, { "epoch": 3.110671936758893, "grad_norm": 2.446336030960083, "learning_rate": 4.230693069306931e-05, "loss": 0.5758, "step": 787 }, { "epoch": 3.1146245059288535, "grad_norm": 2.7452664375305176, "learning_rate": 4.22970297029703e-05, "loss": 0.5077, "step": 788 }, { "epoch": 3.1185770750988144, "grad_norm": 2.698697090148926, "learning_rate": 4.228712871287129e-05, "loss": 0.7094, "step": 789 }, { "epoch": 3.122529644268775, "grad_norm": 3.45510196685791, "learning_rate": 4.2277227722772275e-05, "loss": 0.6012, "step": 790 }, { "epoch": 3.1264822134387353, "grad_norm": 2.7735273838043213, "learning_rate": 4.2267326732673266e-05, "loss": 0.5123, "step": 791 }, { "epoch": 3.130434782608696, "grad_norm": 3.4790964126586914, "learning_rate": 4.225742574257426e-05, "loss": 0.6116, "step": 792 }, { "epoch": 3.1343873517786562, "grad_norm": 3.6229889392852783, "learning_rate": 4.224752475247525e-05, "loss": 0.5183, "step": 793 }, { "epoch": 3.1383399209486167, "grad_norm": 2.5832364559173584, "learning_rate": 4.223762376237624e-05, "loss": 0.4605, "step": 794 }, { "epoch": 3.142292490118577, "grad_norm": 2.941274881362915, "learning_rate": 4.222772277227723e-05, "loss": 0.6244, "step": 795 }, { "epoch": 3.1462450592885376, "grad_norm": 3.1703367233276367, "learning_rate": 4.221782178217822e-05, "loss": 0.6857, "step": 796 }, { "epoch": 3.150197628458498, "grad_norm": 2.6985394954681396, "learning_rate": 4.2207920792079206e-05, "loss": 0.5385, "step": 797 }, { "epoch": 3.1541501976284585, "grad_norm": 2.5677566528320312, "learning_rate": 4.21980198019802e-05, "loss": 0.4475, "step": 798 }, { "epoch": 3.158102766798419, "grad_norm": 4.0725836753845215, "learning_rate": 4.218811881188119e-05, "loss": 0.5295, "step": 799 }, { "epoch": 3.1620553359683794, "grad_norm": 3.270092725753784, "learning_rate": 4.217821782178218e-05, "loss": 0.4589, "step": 800 }, { "epoch": 3.16600790513834, "grad_norm": 2.684762954711914, "learning_rate": 4.216831683168317e-05, "loss": 0.5889, "step": 801 }, { "epoch": 3.1699604743083003, "grad_norm": 3.2751545906066895, "learning_rate": 4.215841584158416e-05, "loss": 0.4955, "step": 802 }, { "epoch": 3.1739130434782608, "grad_norm": 2.9258337020874023, "learning_rate": 4.214851485148515e-05, "loss": 0.4308, "step": 803 }, { "epoch": 3.177865612648221, "grad_norm": 3.0552611351013184, "learning_rate": 4.2138613861386136e-05, "loss": 0.4312, "step": 804 }, { "epoch": 3.1818181818181817, "grad_norm": 2.953078508377075, "learning_rate": 4.212871287128713e-05, "loss": 0.6631, "step": 805 }, { "epoch": 3.185770750988142, "grad_norm": 3.5374467372894287, "learning_rate": 4.211881188118812e-05, "loss": 0.5807, "step": 806 }, { "epoch": 3.1897233201581026, "grad_norm": 3.0554163455963135, "learning_rate": 4.210891089108911e-05, "loss": 0.4394, "step": 807 }, { "epoch": 3.1936758893280635, "grad_norm": 2.729764699935913, "learning_rate": 4.20990099009901e-05, "loss": 0.5139, "step": 808 }, { "epoch": 3.197628458498024, "grad_norm": 3.4895825386047363, "learning_rate": 4.208910891089109e-05, "loss": 0.623, "step": 809 }, { "epoch": 3.2015810276679844, "grad_norm": 2.509068489074707, "learning_rate": 4.207920792079208e-05, "loss": 0.4185, "step": 810 }, { "epoch": 3.205533596837945, "grad_norm": 3.95471453666687, "learning_rate": 4.206930693069307e-05, "loss": 0.4664, "step": 811 }, { "epoch": 3.2094861660079053, "grad_norm": 2.712590456008911, "learning_rate": 4.205940594059406e-05, "loss": 0.5487, "step": 812 }, { "epoch": 3.2134387351778657, "grad_norm": 4.558450222015381, "learning_rate": 4.204950495049505e-05, "loss": 0.408, "step": 813 }, { "epoch": 3.217391304347826, "grad_norm": 3.0374345779418945, "learning_rate": 4.203960396039604e-05, "loss": 0.4516, "step": 814 }, { "epoch": 3.2213438735177866, "grad_norm": 2.663700580596924, "learning_rate": 4.202970297029703e-05, "loss": 0.6279, "step": 815 }, { "epoch": 3.225296442687747, "grad_norm": 3.279304027557373, "learning_rate": 4.201980198019802e-05, "loss": 0.4955, "step": 816 }, { "epoch": 3.2292490118577075, "grad_norm": 2.603487968444824, "learning_rate": 4.200990099009901e-05, "loss": 0.5595, "step": 817 }, { "epoch": 3.233201581027668, "grad_norm": 2.752476453781128, "learning_rate": 4.2e-05, "loss": 0.7953, "step": 818 }, { "epoch": 3.2371541501976284, "grad_norm": 2.8399481773376465, "learning_rate": 4.199009900990099e-05, "loss": 0.5066, "step": 819 }, { "epoch": 3.241106719367589, "grad_norm": 2.615265369415283, "learning_rate": 4.198019801980198e-05, "loss": 0.6043, "step": 820 }, { "epoch": 3.2450592885375493, "grad_norm": 2.15763258934021, "learning_rate": 4.197029702970297e-05, "loss": 0.514, "step": 821 }, { "epoch": 3.2490118577075098, "grad_norm": 3.0518510341644287, "learning_rate": 4.196039603960396e-05, "loss": 0.5913, "step": 822 }, { "epoch": 3.2529644268774702, "grad_norm": 2.7348077297210693, "learning_rate": 4.195049504950495e-05, "loss": 0.6107, "step": 823 }, { "epoch": 3.2569169960474307, "grad_norm": 3.2246909141540527, "learning_rate": 4.1940594059405944e-05, "loss": 0.5814, "step": 824 }, { "epoch": 3.260869565217391, "grad_norm": 3.5306169986724854, "learning_rate": 4.193069306930693e-05, "loss": 0.6782, "step": 825 }, { "epoch": 3.2648221343873516, "grad_norm": 2.4492549896240234, "learning_rate": 4.192079207920792e-05, "loss": 0.5462, "step": 826 }, { "epoch": 3.2687747035573125, "grad_norm": 2.658447504043579, "learning_rate": 4.191089108910891e-05, "loss": 0.4848, "step": 827 }, { "epoch": 3.2727272727272725, "grad_norm": 2.767321825027466, "learning_rate": 4.19009900990099e-05, "loss": 0.6307, "step": 828 }, { "epoch": 3.2766798418972334, "grad_norm": 4.131479263305664, "learning_rate": 4.189108910891089e-05, "loss": 0.805, "step": 829 }, { "epoch": 3.280632411067194, "grad_norm": 2.982330322265625, "learning_rate": 4.1881188118811884e-05, "loss": 0.6569, "step": 830 }, { "epoch": 3.2845849802371543, "grad_norm": 3.369488000869751, "learning_rate": 4.1871287128712875e-05, "loss": 0.383, "step": 831 }, { "epoch": 3.2885375494071147, "grad_norm": 2.395831823348999, "learning_rate": 4.186138613861386e-05, "loss": 0.4348, "step": 832 }, { "epoch": 3.292490118577075, "grad_norm": 2.479269027709961, "learning_rate": 4.185148514851485e-05, "loss": 0.5081, "step": 833 }, { "epoch": 3.2964426877470356, "grad_norm": 2.9438669681549072, "learning_rate": 4.184158415841584e-05, "loss": 0.5037, "step": 834 }, { "epoch": 3.300395256916996, "grad_norm": 2.944010019302368, "learning_rate": 4.183168316831683e-05, "loss": 0.378, "step": 835 }, { "epoch": 3.3043478260869565, "grad_norm": 2.717684745788574, "learning_rate": 4.182178217821782e-05, "loss": 0.4445, "step": 836 }, { "epoch": 3.308300395256917, "grad_norm": 3.649308919906616, "learning_rate": 4.1811881188118814e-05, "loss": 0.504, "step": 837 }, { "epoch": 3.3122529644268774, "grad_norm": 3.165383815765381, "learning_rate": 4.1801980198019805e-05, "loss": 0.5418, "step": 838 }, { "epoch": 3.316205533596838, "grad_norm": 2.940809726715088, "learning_rate": 4.179207920792079e-05, "loss": 0.6626, "step": 839 }, { "epoch": 3.3201581027667983, "grad_norm": 2.885688304901123, "learning_rate": 4.178217821782178e-05, "loss": 0.4955, "step": 840 }, { "epoch": 3.324110671936759, "grad_norm": 3.498290777206421, "learning_rate": 4.177227722772277e-05, "loss": 0.5551, "step": 841 }, { "epoch": 3.3280632411067192, "grad_norm": 2.507458448410034, "learning_rate": 4.176237623762376e-05, "loss": 0.6288, "step": 842 }, { "epoch": 3.3320158102766797, "grad_norm": 3.661512613296509, "learning_rate": 4.1752475247524754e-05, "loss": 0.6459, "step": 843 }, { "epoch": 3.33596837944664, "grad_norm": 2.372563362121582, "learning_rate": 4.1742574257425745e-05, "loss": 0.4602, "step": 844 }, { "epoch": 3.3399209486166006, "grad_norm": 2.741536855697632, "learning_rate": 4.1732673267326736e-05, "loss": 0.5543, "step": 845 }, { "epoch": 3.3438735177865615, "grad_norm": 2.965439796447754, "learning_rate": 4.172277227722772e-05, "loss": 0.6412, "step": 846 }, { "epoch": 3.3478260869565215, "grad_norm": 2.271393299102783, "learning_rate": 4.171287128712871e-05, "loss": 0.518, "step": 847 }, { "epoch": 3.3517786561264824, "grad_norm": 4.426976203918457, "learning_rate": 4.17029702970297e-05, "loss": 0.4079, "step": 848 }, { "epoch": 3.355731225296443, "grad_norm": 2.6228911876678467, "learning_rate": 4.1693069306930694e-05, "loss": 0.5639, "step": 849 }, { "epoch": 3.3596837944664033, "grad_norm": 2.967618465423584, "learning_rate": 4.1683168316831685e-05, "loss": 0.5172, "step": 850 }, { "epoch": 3.3636363636363638, "grad_norm": 3.776729106903076, "learning_rate": 4.1673267326732676e-05, "loss": 0.5001, "step": 851 }, { "epoch": 3.367588932806324, "grad_norm": 2.9853498935699463, "learning_rate": 4.166336633663367e-05, "loss": 0.4628, "step": 852 }, { "epoch": 3.3715415019762847, "grad_norm": 2.3895092010498047, "learning_rate": 4.165346534653465e-05, "loss": 0.4623, "step": 853 }, { "epoch": 3.375494071146245, "grad_norm": 3.3096632957458496, "learning_rate": 4.164356435643564e-05, "loss": 0.6771, "step": 854 }, { "epoch": 3.3794466403162056, "grad_norm": 3.46929669380188, "learning_rate": 4.163366336633663e-05, "loss": 0.6849, "step": 855 }, { "epoch": 3.383399209486166, "grad_norm": 3.367293357849121, "learning_rate": 4.1623762376237624e-05, "loss": 0.548, "step": 856 }, { "epoch": 3.3873517786561265, "grad_norm": 2.884000301361084, "learning_rate": 4.1613861386138615e-05, "loss": 0.5114, "step": 857 }, { "epoch": 3.391304347826087, "grad_norm": 2.5927958488464355, "learning_rate": 4.1603960396039606e-05, "loss": 0.3663, "step": 858 }, { "epoch": 3.3952569169960474, "grad_norm": 4.04662561416626, "learning_rate": 4.15940594059406e-05, "loss": 0.6962, "step": 859 }, { "epoch": 3.399209486166008, "grad_norm": 2.2035534381866455, "learning_rate": 4.158415841584158e-05, "loss": 0.4553, "step": 860 }, { "epoch": 3.4031620553359683, "grad_norm": 2.4361939430236816, "learning_rate": 4.157425742574257e-05, "loss": 0.5728, "step": 861 }, { "epoch": 3.4071146245059287, "grad_norm": 2.7696943283081055, "learning_rate": 4.1564356435643564e-05, "loss": 0.5646, "step": 862 }, { "epoch": 3.411067193675889, "grad_norm": 3.8909192085266113, "learning_rate": 4.1554455445544555e-05, "loss": 0.4507, "step": 863 }, { "epoch": 3.4150197628458496, "grad_norm": 2.7542779445648193, "learning_rate": 4.1544554455445546e-05, "loss": 0.4375, "step": 864 }, { "epoch": 3.4189723320158105, "grad_norm": 3.5081889629364014, "learning_rate": 4.153465346534654e-05, "loss": 0.6612, "step": 865 }, { "epoch": 3.4229249011857705, "grad_norm": 3.533311128616333, "learning_rate": 4.152475247524752e-05, "loss": 0.6637, "step": 866 }, { "epoch": 3.4268774703557314, "grad_norm": 3.761533737182617, "learning_rate": 4.151485148514851e-05, "loss": 0.651, "step": 867 }, { "epoch": 3.430830039525692, "grad_norm": 2.9692132472991943, "learning_rate": 4.1504950495049504e-05, "loss": 0.5799, "step": 868 }, { "epoch": 3.4347826086956523, "grad_norm": 2.8406553268432617, "learning_rate": 4.1495049504950495e-05, "loss": 0.5498, "step": 869 }, { "epoch": 3.438735177865613, "grad_norm": 2.8397724628448486, "learning_rate": 4.1485148514851486e-05, "loss": 0.5073, "step": 870 }, { "epoch": 3.4426877470355732, "grad_norm": 2.4545493125915527, "learning_rate": 4.147524752475248e-05, "loss": 0.4933, "step": 871 }, { "epoch": 3.4466403162055337, "grad_norm": 2.1091456413269043, "learning_rate": 4.146534653465347e-05, "loss": 0.3791, "step": 872 }, { "epoch": 3.450592885375494, "grad_norm": 3.008024215698242, "learning_rate": 4.145544554455445e-05, "loss": 0.4171, "step": 873 }, { "epoch": 3.4545454545454546, "grad_norm": 2.5331947803497314, "learning_rate": 4.144554455445544e-05, "loss": 0.5177, "step": 874 }, { "epoch": 3.458498023715415, "grad_norm": 2.629890203475952, "learning_rate": 4.1435643564356434e-05, "loss": 0.6951, "step": 875 }, { "epoch": 3.4624505928853755, "grad_norm": 2.940838575363159, "learning_rate": 4.1425742574257425e-05, "loss": 0.6245, "step": 876 }, { "epoch": 3.466403162055336, "grad_norm": 2.5807888507843018, "learning_rate": 4.1415841584158416e-05, "loss": 0.6271, "step": 877 }, { "epoch": 3.4703557312252964, "grad_norm": 2.6359646320343018, "learning_rate": 4.140594059405941e-05, "loss": 0.68, "step": 878 }, { "epoch": 3.474308300395257, "grad_norm": 3.50066876411438, "learning_rate": 4.13960396039604e-05, "loss": 0.5463, "step": 879 }, { "epoch": 3.4782608695652173, "grad_norm": 2.375187397003174, "learning_rate": 4.138613861386138e-05, "loss": 0.5015, "step": 880 }, { "epoch": 3.4822134387351777, "grad_norm": 2.886279344558716, "learning_rate": 4.1376237623762374e-05, "loss": 0.68, "step": 881 }, { "epoch": 3.486166007905138, "grad_norm": 3.1308727264404297, "learning_rate": 4.1366336633663365e-05, "loss": 0.531, "step": 882 }, { "epoch": 3.4901185770750986, "grad_norm": 3.4054250717163086, "learning_rate": 4.1356435643564356e-05, "loss": 0.5636, "step": 883 }, { "epoch": 3.494071146245059, "grad_norm": 2.622769355773926, "learning_rate": 4.134653465346535e-05, "loss": 0.5597, "step": 884 }, { "epoch": 3.4980237154150196, "grad_norm": 4.360702037811279, "learning_rate": 4.133663366336634e-05, "loss": 0.596, "step": 885 }, { "epoch": 3.5019762845849804, "grad_norm": 2.7605228424072266, "learning_rate": 4.132673267326733e-05, "loss": 0.5489, "step": 886 }, { "epoch": 3.5059288537549405, "grad_norm": 2.9401211738586426, "learning_rate": 4.1316831683168313e-05, "loss": 0.6354, "step": 887 }, { "epoch": 3.5098814229249014, "grad_norm": 2.7344653606414795, "learning_rate": 4.1306930693069305e-05, "loss": 0.6027, "step": 888 }, { "epoch": 3.513833992094862, "grad_norm": 2.526149034500122, "learning_rate": 4.1297029702970296e-05, "loss": 0.6401, "step": 889 }, { "epoch": 3.5177865612648223, "grad_norm": 2.551302194595337, "learning_rate": 4.128712871287129e-05, "loss": 0.5951, "step": 890 }, { "epoch": 3.5217391304347827, "grad_norm": 4.107852935791016, "learning_rate": 4.127722772277228e-05, "loss": 0.484, "step": 891 }, { "epoch": 3.525691699604743, "grad_norm": 2.8571081161499023, "learning_rate": 4.126732673267327e-05, "loss": 0.6979, "step": 892 }, { "epoch": 3.5296442687747036, "grad_norm": 3.0081305503845215, "learning_rate": 4.125742574257426e-05, "loss": 0.5591, "step": 893 }, { "epoch": 3.533596837944664, "grad_norm": 2.868626594543457, "learning_rate": 4.1247524752475244e-05, "loss": 0.4649, "step": 894 }, { "epoch": 3.5375494071146245, "grad_norm": 3.8529651165008545, "learning_rate": 4.1237623762376235e-05, "loss": 0.5322, "step": 895 }, { "epoch": 3.541501976284585, "grad_norm": 2.836620330810547, "learning_rate": 4.1227722772277226e-05, "loss": 0.4957, "step": 896 }, { "epoch": 3.5454545454545454, "grad_norm": 2.6184310913085938, "learning_rate": 4.121782178217822e-05, "loss": 0.7287, "step": 897 }, { "epoch": 3.549407114624506, "grad_norm": 3.2553951740264893, "learning_rate": 4.120792079207921e-05, "loss": 0.6716, "step": 898 }, { "epoch": 3.5533596837944663, "grad_norm": 3.1503522396087646, "learning_rate": 4.11980198019802e-05, "loss": 0.7217, "step": 899 }, { "epoch": 3.5573122529644268, "grad_norm": 2.660388946533203, "learning_rate": 4.118811881188119e-05, "loss": 0.5023, "step": 900 }, { "epoch": 3.561264822134387, "grad_norm": 2.940899610519409, "learning_rate": 4.1178217821782175e-05, "loss": 0.4583, "step": 901 }, { "epoch": 3.5652173913043477, "grad_norm": 3.1982879638671875, "learning_rate": 4.1168316831683166e-05, "loss": 0.7917, "step": 902 }, { "epoch": 3.5691699604743086, "grad_norm": 3.424924612045288, "learning_rate": 4.115841584158416e-05, "loss": 0.5326, "step": 903 }, { "epoch": 3.5731225296442686, "grad_norm": 3.1696841716766357, "learning_rate": 4.114851485148515e-05, "loss": 0.6626, "step": 904 }, { "epoch": 3.5770750988142295, "grad_norm": 2.140401601791382, "learning_rate": 4.113861386138614e-05, "loss": 0.5067, "step": 905 }, { "epoch": 3.5810276679841895, "grad_norm": 2.8751392364501953, "learning_rate": 4.112871287128713e-05, "loss": 0.7869, "step": 906 }, { "epoch": 3.5849802371541504, "grad_norm": 2.5835886001586914, "learning_rate": 4.111881188118812e-05, "loss": 0.5752, "step": 907 }, { "epoch": 3.588932806324111, "grad_norm": 3.2149412631988525, "learning_rate": 4.1108910891089106e-05, "loss": 0.7117, "step": 908 }, { "epoch": 3.5928853754940713, "grad_norm": 2.543912887573242, "learning_rate": 4.1099009900990097e-05, "loss": 0.5567, "step": 909 }, { "epoch": 3.5968379446640317, "grad_norm": 2.917189121246338, "learning_rate": 4.108910891089109e-05, "loss": 0.6339, "step": 910 }, { "epoch": 3.600790513833992, "grad_norm": 2.416347026824951, "learning_rate": 4.107920792079208e-05, "loss": 0.5568, "step": 911 }, { "epoch": 3.6047430830039526, "grad_norm": 2.908367395401001, "learning_rate": 4.106930693069307e-05, "loss": 0.6781, "step": 912 }, { "epoch": 3.608695652173913, "grad_norm": 3.0680525302886963, "learning_rate": 4.105940594059406e-05, "loss": 0.4941, "step": 913 }, { "epoch": 3.6126482213438735, "grad_norm": 2.8340303897857666, "learning_rate": 4.104950495049505e-05, "loss": 0.5825, "step": 914 }, { "epoch": 3.616600790513834, "grad_norm": 2.956763982772827, "learning_rate": 4.1039603960396036e-05, "loss": 0.5819, "step": 915 }, { "epoch": 3.6205533596837944, "grad_norm": 3.3780250549316406, "learning_rate": 4.102970297029703e-05, "loss": 0.6364, "step": 916 }, { "epoch": 3.624505928853755, "grad_norm": 2.825469970703125, "learning_rate": 4.101980198019802e-05, "loss": 0.5237, "step": 917 }, { "epoch": 3.6284584980237153, "grad_norm": 2.7121708393096924, "learning_rate": 4.100990099009901e-05, "loss": 0.6835, "step": 918 }, { "epoch": 3.632411067193676, "grad_norm": 3.362112045288086, "learning_rate": 4.1e-05, "loss": 0.6003, "step": 919 }, { "epoch": 3.6363636363636362, "grad_norm": 2.794055223464966, "learning_rate": 4.099009900990099e-05, "loss": 0.5824, "step": 920 }, { "epoch": 3.6403162055335967, "grad_norm": 3.757965564727783, "learning_rate": 4.098019801980198e-05, "loss": 0.59, "step": 921 }, { "epoch": 3.6442687747035576, "grad_norm": 2.931434392929077, "learning_rate": 4.097029702970297e-05, "loss": 0.5918, "step": 922 }, { "epoch": 3.6482213438735176, "grad_norm": 2.6133038997650146, "learning_rate": 4.096039603960396e-05, "loss": 0.657, "step": 923 }, { "epoch": 3.6521739130434785, "grad_norm": 2.8601505756378174, "learning_rate": 4.095049504950495e-05, "loss": 0.6662, "step": 924 }, { "epoch": 3.6561264822134385, "grad_norm": 3.618299722671509, "learning_rate": 4.094059405940594e-05, "loss": 0.7431, "step": 925 }, { "epoch": 3.6600790513833994, "grad_norm": 3.0958340167999268, "learning_rate": 4.093069306930693e-05, "loss": 0.4895, "step": 926 }, { "epoch": 3.66403162055336, "grad_norm": 2.906301975250244, "learning_rate": 4.092079207920792e-05, "loss": 0.5355, "step": 927 }, { "epoch": 3.6679841897233203, "grad_norm": 3.1649231910705566, "learning_rate": 4.091089108910891e-05, "loss": 0.6275, "step": 928 }, { "epoch": 3.6719367588932808, "grad_norm": 3.420549154281616, "learning_rate": 4.09009900990099e-05, "loss": 0.4037, "step": 929 }, { "epoch": 3.675889328063241, "grad_norm": 3.499302625656128, "learning_rate": 4.089108910891089e-05, "loss": 0.6551, "step": 930 }, { "epoch": 3.6798418972332017, "grad_norm": 2.679558038711548, "learning_rate": 4.088118811881188e-05, "loss": 0.653, "step": 931 }, { "epoch": 3.683794466403162, "grad_norm": 2.739297389984131, "learning_rate": 4.087128712871287e-05, "loss": 0.557, "step": 932 }, { "epoch": 3.6877470355731226, "grad_norm": 2.774057626724243, "learning_rate": 4.086138613861386e-05, "loss": 0.7046, "step": 933 }, { "epoch": 3.691699604743083, "grad_norm": 3.202867031097412, "learning_rate": 4.085148514851485e-05, "loss": 0.5698, "step": 934 }, { "epoch": 3.6956521739130435, "grad_norm": 3.0566177368164062, "learning_rate": 4.0841584158415844e-05, "loss": 0.5205, "step": 935 }, { "epoch": 3.699604743083004, "grad_norm": 2.7728805541992188, "learning_rate": 4.083168316831683e-05, "loss": 0.5917, "step": 936 }, { "epoch": 3.7035573122529644, "grad_norm": 2.042433738708496, "learning_rate": 4.082178217821782e-05, "loss": 0.5676, "step": 937 }, { "epoch": 3.707509881422925, "grad_norm": 4.46372652053833, "learning_rate": 4.081188118811881e-05, "loss": 0.6228, "step": 938 }, { "epoch": 3.7114624505928853, "grad_norm": 2.4231038093566895, "learning_rate": 4.08019801980198e-05, "loss": 0.4575, "step": 939 }, { "epoch": 3.7154150197628457, "grad_norm": 3.103843927383423, "learning_rate": 4.079207920792079e-05, "loss": 0.5996, "step": 940 }, { "epoch": 3.719367588932806, "grad_norm": 3.0180294513702393, "learning_rate": 4.0782178217821784e-05, "loss": 0.4573, "step": 941 }, { "epoch": 3.7233201581027666, "grad_norm": 2.6009271144866943, "learning_rate": 4.0772277227722775e-05, "loss": 0.5362, "step": 942 }, { "epoch": 3.7272727272727275, "grad_norm": 3.0107502937316895, "learning_rate": 4.076237623762376e-05, "loss": 0.5191, "step": 943 }, { "epoch": 3.7312252964426875, "grad_norm": 3.0619142055511475, "learning_rate": 4.075247524752475e-05, "loss": 0.5754, "step": 944 }, { "epoch": 3.7351778656126484, "grad_norm": 3.2865827083587646, "learning_rate": 4.074257425742574e-05, "loss": 0.6806, "step": 945 }, { "epoch": 3.7391304347826084, "grad_norm": 2.5132248401641846, "learning_rate": 4.073267326732673e-05, "loss": 0.4921, "step": 946 }, { "epoch": 3.7430830039525693, "grad_norm": 2.5334227085113525, "learning_rate": 4.072277227722772e-05, "loss": 0.4583, "step": 947 }, { "epoch": 3.7470355731225298, "grad_norm": 2.3966052532196045, "learning_rate": 4.0712871287128714e-05, "loss": 0.6599, "step": 948 }, { "epoch": 3.7509881422924902, "grad_norm": 3.6752395629882812, "learning_rate": 4.0702970297029705e-05, "loss": 0.5247, "step": 949 }, { "epoch": 3.7549407114624507, "grad_norm": 2.1566295623779297, "learning_rate": 4.069306930693069e-05, "loss": 0.4263, "step": 950 }, { "epoch": 3.758893280632411, "grad_norm": 2.5106146335601807, "learning_rate": 4.068316831683168e-05, "loss": 0.5276, "step": 951 }, { "epoch": 3.7628458498023716, "grad_norm": 3.1821742057800293, "learning_rate": 4.067326732673267e-05, "loss": 0.6226, "step": 952 }, { "epoch": 3.766798418972332, "grad_norm": 2.9978811740875244, "learning_rate": 4.066336633663366e-05, "loss": 0.7531, "step": 953 }, { "epoch": 3.7707509881422925, "grad_norm": 3.94193172454834, "learning_rate": 4.0653465346534654e-05, "loss": 0.5647, "step": 954 }, { "epoch": 3.774703557312253, "grad_norm": 2.4297850131988525, "learning_rate": 4.0643564356435645e-05, "loss": 0.5039, "step": 955 }, { "epoch": 3.7786561264822134, "grad_norm": 3.2057433128356934, "learning_rate": 4.0633663366336636e-05, "loss": 0.6409, "step": 956 }, { "epoch": 3.782608695652174, "grad_norm": 3.364464521408081, "learning_rate": 4.062376237623763e-05, "loss": 0.5243, "step": 957 }, { "epoch": 3.7865612648221343, "grad_norm": 2.832167387008667, "learning_rate": 4.061386138613862e-05, "loss": 0.5233, "step": 958 }, { "epoch": 3.7905138339920947, "grad_norm": 2.821153163909912, "learning_rate": 4.060396039603961e-05, "loss": 0.4615, "step": 959 }, { "epoch": 3.794466403162055, "grad_norm": 3.158823251724243, "learning_rate": 4.05940594059406e-05, "loss": 0.5656, "step": 960 }, { "epoch": 3.7984189723320156, "grad_norm": 2.486264705657959, "learning_rate": 4.058415841584159e-05, "loss": 0.4843, "step": 961 }, { "epoch": 3.8023715415019765, "grad_norm": 3.0150299072265625, "learning_rate": 4.0574257425742576e-05, "loss": 0.5336, "step": 962 }, { "epoch": 3.8063241106719365, "grad_norm": 2.730518102645874, "learning_rate": 4.056435643564357e-05, "loss": 0.5472, "step": 963 }, { "epoch": 3.8102766798418974, "grad_norm": 2.5651566982269287, "learning_rate": 4.055445544554456e-05, "loss": 0.533, "step": 964 }, { "epoch": 3.8142292490118574, "grad_norm": 3.067702054977417, "learning_rate": 4.054455445544555e-05, "loss": 0.7059, "step": 965 }, { "epoch": 3.8181818181818183, "grad_norm": 3.3268630504608154, "learning_rate": 4.053465346534654e-05, "loss": 0.7443, "step": 966 }, { "epoch": 3.822134387351779, "grad_norm": 3.1133692264556885, "learning_rate": 4.052475247524753e-05, "loss": 0.7004, "step": 967 }, { "epoch": 3.8260869565217392, "grad_norm": 3.232699394226074, "learning_rate": 4.051485148514852e-05, "loss": 0.4959, "step": 968 }, { "epoch": 3.8300395256916997, "grad_norm": 2.9157958030700684, "learning_rate": 4.0504950495049506e-05, "loss": 0.6793, "step": 969 }, { "epoch": 3.83399209486166, "grad_norm": 3.12813401222229, "learning_rate": 4.04950495049505e-05, "loss": 0.5193, "step": 970 }, { "epoch": 3.8379446640316206, "grad_norm": 3.4619762897491455, "learning_rate": 4.048514851485149e-05, "loss": 0.5339, "step": 971 }, { "epoch": 3.841897233201581, "grad_norm": 3.0920298099517822, "learning_rate": 4.047524752475248e-05, "loss": 0.6523, "step": 972 }, { "epoch": 3.8458498023715415, "grad_norm": 2.911980628967285, "learning_rate": 4.046534653465347e-05, "loss": 0.6158, "step": 973 }, { "epoch": 3.849802371541502, "grad_norm": 3.645336866378784, "learning_rate": 4.045544554455446e-05, "loss": 0.5859, "step": 974 }, { "epoch": 3.8537549407114624, "grad_norm": 3.207629680633545, "learning_rate": 4.044554455445545e-05, "loss": 0.6653, "step": 975 }, { "epoch": 3.857707509881423, "grad_norm": 2.347885847091675, "learning_rate": 4.043564356435644e-05, "loss": 0.6219, "step": 976 }, { "epoch": 3.8616600790513833, "grad_norm": 2.431614637374878, "learning_rate": 4.042574257425743e-05, "loss": 0.4773, "step": 977 }, { "epoch": 3.8656126482213438, "grad_norm": 3.4384522438049316, "learning_rate": 4.041584158415842e-05, "loss": 0.6076, "step": 978 }, { "epoch": 3.869565217391304, "grad_norm": 2.9552202224731445, "learning_rate": 4.040594059405941e-05, "loss": 0.6872, "step": 979 }, { "epoch": 3.8735177865612647, "grad_norm": 3.165083169937134, "learning_rate": 4.03960396039604e-05, "loss": 0.5997, "step": 980 }, { "epoch": 3.8774703557312256, "grad_norm": 2.9601211547851562, "learning_rate": 4.038613861386139e-05, "loss": 0.5066, "step": 981 }, { "epoch": 3.8814229249011856, "grad_norm": 3.6214566230773926, "learning_rate": 4.0376237623762383e-05, "loss": 0.5293, "step": 982 }, { "epoch": 3.8853754940711465, "grad_norm": 2.7425880432128906, "learning_rate": 4.036633663366337e-05, "loss": 0.4746, "step": 983 }, { "epoch": 3.8893280632411065, "grad_norm": 2.756862163543701, "learning_rate": 4.035643564356436e-05, "loss": 0.4498, "step": 984 }, { "epoch": 3.8932806324110674, "grad_norm": 2.8461694717407227, "learning_rate": 4.034653465346535e-05, "loss": 0.5385, "step": 985 }, { "epoch": 3.897233201581028, "grad_norm": 4.07880163192749, "learning_rate": 4.033663366336634e-05, "loss": 0.5259, "step": 986 }, { "epoch": 3.9011857707509883, "grad_norm": 2.9115302562713623, "learning_rate": 4.032673267326733e-05, "loss": 0.5042, "step": 987 }, { "epoch": 3.9051383399209487, "grad_norm": 2.868607997894287, "learning_rate": 4.031683168316832e-05, "loss": 0.5695, "step": 988 }, { "epoch": 3.909090909090909, "grad_norm": 2.6620211601257324, "learning_rate": 4.0306930693069314e-05, "loss": 0.6854, "step": 989 }, { "epoch": 3.9130434782608696, "grad_norm": 3.4102489948272705, "learning_rate": 4.02970297029703e-05, "loss": 0.4618, "step": 990 }, { "epoch": 3.91699604743083, "grad_norm": 2.9352076053619385, "learning_rate": 4.028712871287129e-05, "loss": 0.492, "step": 991 }, { "epoch": 3.9209486166007905, "grad_norm": 2.773503065109253, "learning_rate": 4.027722772277228e-05, "loss": 0.49, "step": 992 }, { "epoch": 3.924901185770751, "grad_norm": 3.285865306854248, "learning_rate": 4.026732673267327e-05, "loss": 0.7214, "step": 993 }, { "epoch": 3.9288537549407114, "grad_norm": 3.705916404724121, "learning_rate": 4.025742574257426e-05, "loss": 0.5419, "step": 994 }, { "epoch": 3.932806324110672, "grad_norm": 2.818164825439453, "learning_rate": 4.0247524752475254e-05, "loss": 0.5098, "step": 995 }, { "epoch": 3.9367588932806323, "grad_norm": 3.379474639892578, "learning_rate": 4.0237623762376245e-05, "loss": 0.6025, "step": 996 }, { "epoch": 3.940711462450593, "grad_norm": 3.4393768310546875, "learning_rate": 4.022772277227723e-05, "loss": 0.4622, "step": 997 }, { "epoch": 3.9446640316205532, "grad_norm": 2.926403760910034, "learning_rate": 4.021782178217822e-05, "loss": 0.599, "step": 998 }, { "epoch": 3.9486166007905137, "grad_norm": 2.77158260345459, "learning_rate": 4.020792079207921e-05, "loss": 0.5384, "step": 999 }, { "epoch": 3.9525691699604746, "grad_norm": 2.9309792518615723, "learning_rate": 4.01980198019802e-05, "loss": 0.4774, "step": 1000 }, { "epoch": 3.9565217391304346, "grad_norm": 2.857879638671875, "learning_rate": 4.018811881188119e-05, "loss": 0.5314, "step": 1001 }, { "epoch": 3.9604743083003955, "grad_norm": 3.0472517013549805, "learning_rate": 4.0178217821782184e-05, "loss": 0.6871, "step": 1002 }, { "epoch": 3.9644268774703555, "grad_norm": 3.3048479557037354, "learning_rate": 4.0168316831683175e-05, "loss": 0.6245, "step": 1003 }, { "epoch": 3.9683794466403164, "grad_norm": 3.289996862411499, "learning_rate": 4.015841584158416e-05, "loss": 0.6824, "step": 1004 }, { "epoch": 3.972332015810277, "grad_norm": 3.0692379474639893, "learning_rate": 4.014851485148515e-05, "loss": 0.4947, "step": 1005 }, { "epoch": 3.9762845849802373, "grad_norm": 3.646455764770508, "learning_rate": 4.013861386138614e-05, "loss": 0.5847, "step": 1006 }, { "epoch": 3.9802371541501977, "grad_norm": 2.8091559410095215, "learning_rate": 4.012871287128713e-05, "loss": 0.5868, "step": 1007 }, { "epoch": 3.984189723320158, "grad_norm": 2.8593790531158447, "learning_rate": 4.0118811881188124e-05, "loss": 0.7904, "step": 1008 }, { "epoch": 3.9881422924901186, "grad_norm": 3.0017247200012207, "learning_rate": 4.0108910891089115e-05, "loss": 0.6615, "step": 1009 }, { "epoch": 3.992094861660079, "grad_norm": 4.116451263427734, "learning_rate": 4.0099009900990106e-05, "loss": 0.6194, "step": 1010 }, { "epoch": 3.9960474308300395, "grad_norm": 2.4350996017456055, "learning_rate": 4.008910891089109e-05, "loss": 0.5497, "step": 1011 }, { "epoch": 4.0, "grad_norm": 2.397753953933716, "learning_rate": 4.007920792079208e-05, "loss": 0.5037, "step": 1012 }, { "epoch": 4.003952569169961, "grad_norm": 1.9421703815460205, "learning_rate": 4.006930693069307e-05, "loss": 0.2973, "step": 1013 }, { "epoch": 4.007905138339921, "grad_norm": 2.6778204441070557, "learning_rate": 4.0059405940594064e-05, "loss": 0.3646, "step": 1014 }, { "epoch": 4.011857707509882, "grad_norm": 1.9245054721832275, "learning_rate": 4.0049504950495055e-05, "loss": 0.2344, "step": 1015 }, { "epoch": 4.015810276679842, "grad_norm": 2.8454062938690186, "learning_rate": 4.0039603960396046e-05, "loss": 0.2451, "step": 1016 }, { "epoch": 4.019762845849803, "grad_norm": 3.259328603744507, "learning_rate": 4.002970297029704e-05, "loss": 0.2696, "step": 1017 }, { "epoch": 4.023715415019763, "grad_norm": 4.139449596405029, "learning_rate": 4.001980198019802e-05, "loss": 0.356, "step": 1018 }, { "epoch": 4.027667984189724, "grad_norm": 4.438423156738281, "learning_rate": 4.000990099009901e-05, "loss": 0.2983, "step": 1019 }, { "epoch": 4.031620553359684, "grad_norm": 5.987069606781006, "learning_rate": 4e-05, "loss": 0.5465, "step": 1020 }, { "epoch": 4.0355731225296445, "grad_norm": 3.62335467338562, "learning_rate": 3.9990099009900994e-05, "loss": 0.1989, "step": 1021 }, { "epoch": 4.0395256916996045, "grad_norm": 6.2495574951171875, "learning_rate": 3.9980198019801985e-05, "loss": 0.2863, "step": 1022 }, { "epoch": 4.043478260869565, "grad_norm": 3.1103017330169678, "learning_rate": 3.9970297029702976e-05, "loss": 0.2105, "step": 1023 }, { "epoch": 4.047430830039525, "grad_norm": 3.0796120166778564, "learning_rate": 3.996039603960397e-05, "loss": 0.2538, "step": 1024 }, { "epoch": 4.051383399209486, "grad_norm": 2.44278883934021, "learning_rate": 3.995049504950495e-05, "loss": 0.2783, "step": 1025 }, { "epoch": 4.055335968379446, "grad_norm": 2.8014047145843506, "learning_rate": 3.994059405940594e-05, "loss": 0.2138, "step": 1026 }, { "epoch": 4.059288537549407, "grad_norm": 3.1606693267822266, "learning_rate": 3.9930693069306934e-05, "loss": 0.2371, "step": 1027 }, { "epoch": 4.063241106719367, "grad_norm": 2.735130786895752, "learning_rate": 3.9920792079207925e-05, "loss": 0.2003, "step": 1028 }, { "epoch": 4.067193675889328, "grad_norm": 2.967221260070801, "learning_rate": 3.9910891089108916e-05, "loss": 0.3614, "step": 1029 }, { "epoch": 4.071146245059288, "grad_norm": 3.15625, "learning_rate": 3.990099009900991e-05, "loss": 0.2298, "step": 1030 }, { "epoch": 4.075098814229249, "grad_norm": 2.598991632461548, "learning_rate": 3.98910891089109e-05, "loss": 0.224, "step": 1031 }, { "epoch": 4.07905138339921, "grad_norm": 2.4389264583587646, "learning_rate": 3.988118811881188e-05, "loss": 0.2456, "step": 1032 }, { "epoch": 4.08300395256917, "grad_norm": 2.7602131366729736, "learning_rate": 3.9871287128712874e-05, "loss": 0.2912, "step": 1033 }, { "epoch": 4.086956521739131, "grad_norm": 2.517242670059204, "learning_rate": 3.9861386138613865e-05, "loss": 0.209, "step": 1034 }, { "epoch": 4.090909090909091, "grad_norm": 2.962911605834961, "learning_rate": 3.9851485148514856e-05, "loss": 0.234, "step": 1035 }, { "epoch": 4.094861660079052, "grad_norm": 3.6362602710723877, "learning_rate": 3.984158415841585e-05, "loss": 0.4125, "step": 1036 }, { "epoch": 4.098814229249012, "grad_norm": 3.4367194175720215, "learning_rate": 3.983168316831684e-05, "loss": 0.3119, "step": 1037 }, { "epoch": 4.102766798418973, "grad_norm": 3.1447670459747314, "learning_rate": 3.982178217821783e-05, "loss": 0.2189, "step": 1038 }, { "epoch": 4.106719367588933, "grad_norm": 3.5581486225128174, "learning_rate": 3.981188118811881e-05, "loss": 0.2536, "step": 1039 }, { "epoch": 4.1106719367588935, "grad_norm": 4.582309722900391, "learning_rate": 3.9801980198019804e-05, "loss": 0.2457, "step": 1040 }, { "epoch": 4.1146245059288535, "grad_norm": 4.721108436584473, "learning_rate": 3.9792079207920795e-05, "loss": 0.3284, "step": 1041 }, { "epoch": 4.118577075098814, "grad_norm": 2.535555124282837, "learning_rate": 3.9782178217821786e-05, "loss": 0.2277, "step": 1042 }, { "epoch": 4.122529644268774, "grad_norm": 3.330543279647827, "learning_rate": 3.977227722772278e-05, "loss": 0.2869, "step": 1043 }, { "epoch": 4.126482213438735, "grad_norm": 3.411245584487915, "learning_rate": 3.976237623762377e-05, "loss": 0.228, "step": 1044 }, { "epoch": 4.130434782608695, "grad_norm": 2.254465341567993, "learning_rate": 3.975247524752476e-05, "loss": 0.3133, "step": 1045 }, { "epoch": 4.134387351778656, "grad_norm": 3.502152681350708, "learning_rate": 3.9742574257425744e-05, "loss": 0.3405, "step": 1046 }, { "epoch": 4.138339920948616, "grad_norm": 4.128763198852539, "learning_rate": 3.9732673267326735e-05, "loss": 0.3423, "step": 1047 }, { "epoch": 4.142292490118577, "grad_norm": 2.646292209625244, "learning_rate": 3.9722772277227726e-05, "loss": 0.2267, "step": 1048 }, { "epoch": 4.146245059288537, "grad_norm": 3.120715379714966, "learning_rate": 3.971287128712872e-05, "loss": 0.2795, "step": 1049 }, { "epoch": 4.150197628458498, "grad_norm": 2.9044315814971924, "learning_rate": 3.970297029702971e-05, "loss": 0.2982, "step": 1050 }, { "epoch": 4.154150197628459, "grad_norm": 2.7479300498962402, "learning_rate": 3.96930693069307e-05, "loss": 0.2862, "step": 1051 }, { "epoch": 4.158102766798419, "grad_norm": 3.183039903640747, "learning_rate": 3.968316831683169e-05, "loss": 0.2748, "step": 1052 }, { "epoch": 4.16205533596838, "grad_norm": 2.5643458366394043, "learning_rate": 3.9673267326732675e-05, "loss": 0.2931, "step": 1053 }, { "epoch": 4.16600790513834, "grad_norm": 2.8417370319366455, "learning_rate": 3.9663366336633666e-05, "loss": 0.2656, "step": 1054 }, { "epoch": 4.169960474308301, "grad_norm": 2.7871973514556885, "learning_rate": 3.965346534653466e-05, "loss": 0.2616, "step": 1055 }, { "epoch": 4.173913043478261, "grad_norm": 2.8472018241882324, "learning_rate": 3.964356435643565e-05, "loss": 0.2622, "step": 1056 }, { "epoch": 4.177865612648222, "grad_norm": 3.1438567638397217, "learning_rate": 3.963366336633664e-05, "loss": 0.3674, "step": 1057 }, { "epoch": 4.181818181818182, "grad_norm": 3.6160640716552734, "learning_rate": 3.962376237623763e-05, "loss": 0.2141, "step": 1058 }, { "epoch": 4.1857707509881426, "grad_norm": 3.9000306129455566, "learning_rate": 3.961386138613862e-05, "loss": 0.3071, "step": 1059 }, { "epoch": 4.189723320158103, "grad_norm": 2.6272668838500977, "learning_rate": 3.9603960396039605e-05, "loss": 0.2373, "step": 1060 }, { "epoch": 4.1936758893280635, "grad_norm": 2.835028648376465, "learning_rate": 3.9594059405940596e-05, "loss": 0.3216, "step": 1061 }, { "epoch": 4.1976284584980235, "grad_norm": 3.362684965133667, "learning_rate": 3.958415841584159e-05, "loss": 0.3313, "step": 1062 }, { "epoch": 4.201581027667984, "grad_norm": 4.075572490692139, "learning_rate": 3.957425742574258e-05, "loss": 0.2771, "step": 1063 }, { "epoch": 4.205533596837944, "grad_norm": 3.4788613319396973, "learning_rate": 3.956435643564357e-05, "loss": 0.3259, "step": 1064 }, { "epoch": 4.209486166007905, "grad_norm": 3.1352312564849854, "learning_rate": 3.955445544554456e-05, "loss": 0.2872, "step": 1065 }, { "epoch": 4.213438735177865, "grad_norm": 2.9339613914489746, "learning_rate": 3.954455445544555e-05, "loss": 0.3514, "step": 1066 }, { "epoch": 4.217391304347826, "grad_norm": 2.9395394325256348, "learning_rate": 3.9534653465346536e-05, "loss": 0.2243, "step": 1067 }, { "epoch": 4.221343873517786, "grad_norm": 3.043640375137329, "learning_rate": 3.952475247524753e-05, "loss": 0.3206, "step": 1068 }, { "epoch": 4.225296442687747, "grad_norm": 3.1784284114837646, "learning_rate": 3.951485148514852e-05, "loss": 0.3169, "step": 1069 }, { "epoch": 4.229249011857707, "grad_norm": 3.617060661315918, "learning_rate": 3.950495049504951e-05, "loss": 0.3255, "step": 1070 }, { "epoch": 4.233201581027668, "grad_norm": 2.6469147205352783, "learning_rate": 3.94950495049505e-05, "loss": 0.2952, "step": 1071 }, { "epoch": 4.237154150197629, "grad_norm": 2.7838127613067627, "learning_rate": 3.948514851485149e-05, "loss": 0.2942, "step": 1072 }, { "epoch": 4.241106719367589, "grad_norm": 2.3599328994750977, "learning_rate": 3.9475247524752476e-05, "loss": 0.2602, "step": 1073 }, { "epoch": 4.24505928853755, "grad_norm": 2.5975756645202637, "learning_rate": 3.946534653465347e-05, "loss": 0.2688, "step": 1074 }, { "epoch": 4.24901185770751, "grad_norm": 3.0292296409606934, "learning_rate": 3.945544554455446e-05, "loss": 0.2027, "step": 1075 }, { "epoch": 4.252964426877471, "grad_norm": 3.343336343765259, "learning_rate": 3.944554455445545e-05, "loss": 0.3436, "step": 1076 }, { "epoch": 4.256916996047431, "grad_norm": 2.526005983352661, "learning_rate": 3.943564356435644e-05, "loss": 0.2968, "step": 1077 }, { "epoch": 4.260869565217392, "grad_norm": 2.622889518737793, "learning_rate": 3.942574257425743e-05, "loss": 0.3147, "step": 1078 }, { "epoch": 4.264822134387352, "grad_norm": 3.7573342323303223, "learning_rate": 3.941584158415842e-05, "loss": 0.3947, "step": 1079 }, { "epoch": 4.2687747035573125, "grad_norm": 3.176950454711914, "learning_rate": 3.9405940594059406e-05, "loss": 0.2732, "step": 1080 }, { "epoch": 4.2727272727272725, "grad_norm": 3.5214250087738037, "learning_rate": 3.93960396039604e-05, "loss": 0.2361, "step": 1081 }, { "epoch": 4.276679841897233, "grad_norm": 3.19514799118042, "learning_rate": 3.938613861386139e-05, "loss": 0.3016, "step": 1082 }, { "epoch": 4.280632411067193, "grad_norm": 2.7928874492645264, "learning_rate": 3.937623762376238e-05, "loss": 0.3056, "step": 1083 }, { "epoch": 4.284584980237154, "grad_norm": 3.1483709812164307, "learning_rate": 3.936633663366337e-05, "loss": 0.3889, "step": 1084 }, { "epoch": 4.288537549407114, "grad_norm": 3.531400203704834, "learning_rate": 3.935643564356436e-05, "loss": 0.2643, "step": 1085 }, { "epoch": 4.292490118577075, "grad_norm": 3.666123151779175, "learning_rate": 3.934653465346535e-05, "loss": 0.3231, "step": 1086 }, { "epoch": 4.296442687747035, "grad_norm": 2.9479598999023438, "learning_rate": 3.933663366336634e-05, "loss": 0.288, "step": 1087 }, { "epoch": 4.300395256916996, "grad_norm": 3.3740549087524414, "learning_rate": 3.932673267326733e-05, "loss": 0.2135, "step": 1088 }, { "epoch": 4.304347826086957, "grad_norm": 3.6782562732696533, "learning_rate": 3.931683168316832e-05, "loss": 0.3292, "step": 1089 }, { "epoch": 4.308300395256917, "grad_norm": 2.2794337272644043, "learning_rate": 3.930693069306931e-05, "loss": 0.2824, "step": 1090 }, { "epoch": 4.312252964426877, "grad_norm": 3.520951271057129, "learning_rate": 3.92970297029703e-05, "loss": 0.2546, "step": 1091 }, { "epoch": 4.316205533596838, "grad_norm": 2.4747276306152344, "learning_rate": 3.928712871287129e-05, "loss": 0.2613, "step": 1092 }, { "epoch": 4.320158102766799, "grad_norm": 3.4397242069244385, "learning_rate": 3.927722772277228e-05, "loss": 0.2976, "step": 1093 }, { "epoch": 4.324110671936759, "grad_norm": 2.4284393787384033, "learning_rate": 3.926732673267327e-05, "loss": 0.2442, "step": 1094 }, { "epoch": 4.32806324110672, "grad_norm": 3.1947402954101562, "learning_rate": 3.925742574257426e-05, "loss": 0.3182, "step": 1095 }, { "epoch": 4.33201581027668, "grad_norm": 2.8851232528686523, "learning_rate": 3.924752475247525e-05, "loss": 0.3965, "step": 1096 }, { "epoch": 4.335968379446641, "grad_norm": 3.1202361583709717, "learning_rate": 3.923762376237624e-05, "loss": 0.4283, "step": 1097 }, { "epoch": 4.339920948616601, "grad_norm": 2.5294647216796875, "learning_rate": 3.922772277227723e-05, "loss": 0.2697, "step": 1098 }, { "epoch": 4.3438735177865615, "grad_norm": 3.0588459968566895, "learning_rate": 3.921782178217822e-05, "loss": 0.2572, "step": 1099 }, { "epoch": 4.3478260869565215, "grad_norm": 2.895456314086914, "learning_rate": 3.9207920792079214e-05, "loss": 0.3205, "step": 1100 }, { "epoch": 4.351778656126482, "grad_norm": 3.1435513496398926, "learning_rate": 3.91980198019802e-05, "loss": 0.3668, "step": 1101 }, { "epoch": 4.355731225296442, "grad_norm": 3.387343406677246, "learning_rate": 3.918811881188119e-05, "loss": 0.3463, "step": 1102 }, { "epoch": 4.359683794466403, "grad_norm": 3.2940165996551514, "learning_rate": 3.917821782178218e-05, "loss": 0.2982, "step": 1103 }, { "epoch": 4.363636363636363, "grad_norm": 3.1170601844787598, "learning_rate": 3.916831683168317e-05, "loss": 0.3403, "step": 1104 }, { "epoch": 4.367588932806324, "grad_norm": 2.646620512008667, "learning_rate": 3.915841584158416e-05, "loss": 0.2153, "step": 1105 }, { "epoch": 4.371541501976284, "grad_norm": 2.436699628829956, "learning_rate": 3.9148514851485154e-05, "loss": 0.1858, "step": 1106 }, { "epoch": 4.375494071146245, "grad_norm": 2.92559814453125, "learning_rate": 3.9138613861386145e-05, "loss": 0.3681, "step": 1107 }, { "epoch": 4.379446640316205, "grad_norm": 3.7100296020507812, "learning_rate": 3.912871287128713e-05, "loss": 0.3384, "step": 1108 }, { "epoch": 4.383399209486166, "grad_norm": 2.757444143295288, "learning_rate": 3.911881188118812e-05, "loss": 0.3502, "step": 1109 }, { "epoch": 4.387351778656127, "grad_norm": 3.7051267623901367, "learning_rate": 3.910891089108911e-05, "loss": 0.2856, "step": 1110 }, { "epoch": 4.391304347826087, "grad_norm": 4.636383533477783, "learning_rate": 3.90990099009901e-05, "loss": 0.3207, "step": 1111 }, { "epoch": 4.395256916996048, "grad_norm": 3.6344308853149414, "learning_rate": 3.908910891089109e-05, "loss": 0.4691, "step": 1112 }, { "epoch": 4.399209486166008, "grad_norm": 3.4184653759002686, "learning_rate": 3.9079207920792084e-05, "loss": 0.3024, "step": 1113 }, { "epoch": 4.403162055335969, "grad_norm": 3.087573528289795, "learning_rate": 3.9069306930693075e-05, "loss": 0.4391, "step": 1114 }, { "epoch": 4.407114624505929, "grad_norm": 3.7416021823883057, "learning_rate": 3.905940594059406e-05, "loss": 0.3149, "step": 1115 }, { "epoch": 4.41106719367589, "grad_norm": 3.678798198699951, "learning_rate": 3.904950495049505e-05, "loss": 0.2522, "step": 1116 }, { "epoch": 4.41501976284585, "grad_norm": 4.014713764190674, "learning_rate": 3.903960396039604e-05, "loss": 0.3337, "step": 1117 }, { "epoch": 4.4189723320158105, "grad_norm": 4.237569332122803, "learning_rate": 3.902970297029703e-05, "loss": 0.3188, "step": 1118 }, { "epoch": 4.4229249011857705, "grad_norm": 3.0236361026763916, "learning_rate": 3.9019801980198024e-05, "loss": 0.3129, "step": 1119 }, { "epoch": 4.426877470355731, "grad_norm": 3.9568967819213867, "learning_rate": 3.9009900990099015e-05, "loss": 0.2943, "step": 1120 }, { "epoch": 4.430830039525691, "grad_norm": 3.001380205154419, "learning_rate": 3.9000000000000006e-05, "loss": 0.313, "step": 1121 }, { "epoch": 4.434782608695652, "grad_norm": 2.898489475250244, "learning_rate": 3.899009900990099e-05, "loss": 0.3746, "step": 1122 }, { "epoch": 4.438735177865612, "grad_norm": 2.7782986164093018, "learning_rate": 3.898019801980198e-05, "loss": 0.3044, "step": 1123 }, { "epoch": 4.442687747035573, "grad_norm": 2.9859437942504883, "learning_rate": 3.897029702970297e-05, "loss": 0.2657, "step": 1124 }, { "epoch": 4.446640316205533, "grad_norm": 2.6165802478790283, "learning_rate": 3.8960396039603964e-05, "loss": 0.2295, "step": 1125 }, { "epoch": 4.450592885375494, "grad_norm": 2.9132537841796875, "learning_rate": 3.8950495049504955e-05, "loss": 0.2654, "step": 1126 }, { "epoch": 4.454545454545454, "grad_norm": 4.040834903717041, "learning_rate": 3.8940594059405946e-05, "loss": 0.3709, "step": 1127 }, { "epoch": 4.458498023715415, "grad_norm": 4.119834899902344, "learning_rate": 3.893069306930694e-05, "loss": 0.3154, "step": 1128 }, { "epoch": 4.462450592885375, "grad_norm": 2.7986698150634766, "learning_rate": 3.892079207920792e-05, "loss": 0.3002, "step": 1129 }, { "epoch": 4.466403162055336, "grad_norm": 3.298546075820923, "learning_rate": 3.891089108910891e-05, "loss": 0.2518, "step": 1130 }, { "epoch": 4.470355731225297, "grad_norm": 3.5031206607818604, "learning_rate": 3.89009900990099e-05, "loss": 0.2819, "step": 1131 }, { "epoch": 4.474308300395257, "grad_norm": 3.535099506378174, "learning_rate": 3.8891089108910894e-05, "loss": 0.2792, "step": 1132 }, { "epoch": 4.478260869565218, "grad_norm": 3.8441696166992188, "learning_rate": 3.8881188118811885e-05, "loss": 0.2276, "step": 1133 }, { "epoch": 4.482213438735178, "grad_norm": 3.6314444541931152, "learning_rate": 3.8871287128712876e-05, "loss": 0.2851, "step": 1134 }, { "epoch": 4.486166007905139, "grad_norm": 3.17325496673584, "learning_rate": 3.886138613861387e-05, "loss": 0.3465, "step": 1135 }, { "epoch": 4.490118577075099, "grad_norm": 3.2869083881378174, "learning_rate": 3.885148514851485e-05, "loss": 0.3633, "step": 1136 }, { "epoch": 4.4940711462450595, "grad_norm": 2.795565128326416, "learning_rate": 3.884158415841584e-05, "loss": 0.2885, "step": 1137 }, { "epoch": 4.4980237154150196, "grad_norm": 3.389631509780884, "learning_rate": 3.8831683168316834e-05, "loss": 0.4021, "step": 1138 }, { "epoch": 4.5019762845849804, "grad_norm": 3.4602138996124268, "learning_rate": 3.8821782178217825e-05, "loss": 0.2749, "step": 1139 }, { "epoch": 4.5059288537549405, "grad_norm": 3.0726306438446045, "learning_rate": 3.8811881188118816e-05, "loss": 0.1724, "step": 1140 }, { "epoch": 4.509881422924901, "grad_norm": 2.6234934329986572, "learning_rate": 3.880198019801981e-05, "loss": 0.2414, "step": 1141 }, { "epoch": 4.513833992094861, "grad_norm": 2.885284423828125, "learning_rate": 3.87920792079208e-05, "loss": 0.3924, "step": 1142 }, { "epoch": 4.517786561264822, "grad_norm": 2.526918888092041, "learning_rate": 3.878217821782178e-05, "loss": 0.254, "step": 1143 }, { "epoch": 4.521739130434782, "grad_norm": 2.743669271469116, "learning_rate": 3.8772277227722773e-05, "loss": 0.3348, "step": 1144 }, { "epoch": 4.525691699604743, "grad_norm": 3.6247944831848145, "learning_rate": 3.8762376237623765e-05, "loss": 0.2626, "step": 1145 }, { "epoch": 4.529644268774703, "grad_norm": 4.413015365600586, "learning_rate": 3.8752475247524756e-05, "loss": 0.3215, "step": 1146 }, { "epoch": 4.533596837944664, "grad_norm": 2.8632495403289795, "learning_rate": 3.874257425742575e-05, "loss": 0.3583, "step": 1147 }, { "epoch": 4.537549407114625, "grad_norm": 3.381992816925049, "learning_rate": 3.873267326732674e-05, "loss": 0.3555, "step": 1148 }, { "epoch": 4.541501976284585, "grad_norm": 3.2068960666656494, "learning_rate": 3.872277227722773e-05, "loss": 0.3275, "step": 1149 }, { "epoch": 4.545454545454545, "grad_norm": 3.2019689083099365, "learning_rate": 3.871287128712871e-05, "loss": 0.3037, "step": 1150 }, { "epoch": 4.549407114624506, "grad_norm": 3.315410614013672, "learning_rate": 3.8702970297029704e-05, "loss": 0.2505, "step": 1151 }, { "epoch": 4.553359683794467, "grad_norm": 3.393764019012451, "learning_rate": 3.8693069306930695e-05, "loss": 0.3835, "step": 1152 }, { "epoch": 4.557312252964427, "grad_norm": 3.5340983867645264, "learning_rate": 3.8683168316831686e-05, "loss": 0.2464, "step": 1153 }, { "epoch": 4.561264822134388, "grad_norm": 3.2615342140197754, "learning_rate": 3.867326732673268e-05, "loss": 0.3031, "step": 1154 }, { "epoch": 4.565217391304348, "grad_norm": 3.7444982528686523, "learning_rate": 3.866336633663367e-05, "loss": 0.2695, "step": 1155 }, { "epoch": 4.569169960474309, "grad_norm": 2.7645037174224854, "learning_rate": 3.865346534653466e-05, "loss": 0.3032, "step": 1156 }, { "epoch": 4.573122529644269, "grad_norm": 2.9634296894073486, "learning_rate": 3.8643564356435644e-05, "loss": 0.3322, "step": 1157 }, { "epoch": 4.5770750988142295, "grad_norm": 3.4946000576019287, "learning_rate": 3.8633663366336635e-05, "loss": 0.4699, "step": 1158 }, { "epoch": 4.5810276679841895, "grad_norm": 3.1742959022521973, "learning_rate": 3.8623762376237626e-05, "loss": 0.375, "step": 1159 }, { "epoch": 4.58498023715415, "grad_norm": 2.8936591148376465, "learning_rate": 3.861386138613862e-05, "loss": 0.2676, "step": 1160 }, { "epoch": 4.58893280632411, "grad_norm": 2.777315616607666, "learning_rate": 3.860396039603961e-05, "loss": 0.3019, "step": 1161 }, { "epoch": 4.592885375494071, "grad_norm": 3.59289813041687, "learning_rate": 3.85940594059406e-05, "loss": 0.3288, "step": 1162 }, { "epoch": 4.596837944664031, "grad_norm": 3.9896490573883057, "learning_rate": 3.858415841584159e-05, "loss": 0.306, "step": 1163 }, { "epoch": 4.600790513833992, "grad_norm": 2.968564987182617, "learning_rate": 3.8574257425742574e-05, "loss": 0.3646, "step": 1164 }, { "epoch": 4.604743083003952, "grad_norm": 2.1824212074279785, "learning_rate": 3.8564356435643566e-05, "loss": 0.2926, "step": 1165 }, { "epoch": 4.608695652173913, "grad_norm": 2.7332520484924316, "learning_rate": 3.8554455445544557e-05, "loss": 0.2795, "step": 1166 }, { "epoch": 4.612648221343873, "grad_norm": 3.9980952739715576, "learning_rate": 3.854455445544555e-05, "loss": 0.3083, "step": 1167 }, { "epoch": 4.616600790513834, "grad_norm": 3.009532928466797, "learning_rate": 3.853465346534654e-05, "loss": 0.2849, "step": 1168 }, { "epoch": 4.620553359683795, "grad_norm": 2.513786554336548, "learning_rate": 3.852475247524753e-05, "loss": 0.2736, "step": 1169 }, { "epoch": 4.624505928853755, "grad_norm": 2.6872434616088867, "learning_rate": 3.851485148514852e-05, "loss": 0.252, "step": 1170 }, { "epoch": 4.628458498023716, "grad_norm": 3.0371580123901367, "learning_rate": 3.8504950495049505e-05, "loss": 0.3071, "step": 1171 }, { "epoch": 4.632411067193676, "grad_norm": 2.7029552459716797, "learning_rate": 3.8495049504950496e-05, "loss": 0.2571, "step": 1172 }, { "epoch": 4.636363636363637, "grad_norm": 3.263827085494995, "learning_rate": 3.848514851485149e-05, "loss": 0.2399, "step": 1173 }, { "epoch": 4.640316205533597, "grad_norm": 3.00852370262146, "learning_rate": 3.847524752475248e-05, "loss": 0.3077, "step": 1174 }, { "epoch": 4.644268774703558, "grad_norm": 3.6045734882354736, "learning_rate": 3.846534653465347e-05, "loss": 0.3389, "step": 1175 }, { "epoch": 4.648221343873518, "grad_norm": 3.082163095474243, "learning_rate": 3.845544554455446e-05, "loss": 0.262, "step": 1176 }, { "epoch": 4.6521739130434785, "grad_norm": 4.1113104820251465, "learning_rate": 3.844554455445545e-05, "loss": 0.2851, "step": 1177 }, { "epoch": 4.6561264822134385, "grad_norm": 3.4462544918060303, "learning_rate": 3.8435643564356436e-05, "loss": 0.2472, "step": 1178 }, { "epoch": 4.660079051383399, "grad_norm": 2.5143446922302246, "learning_rate": 3.842574257425743e-05, "loss": 0.2844, "step": 1179 }, { "epoch": 4.664031620553359, "grad_norm": 2.848457098007202, "learning_rate": 3.841584158415842e-05, "loss": 0.2939, "step": 1180 }, { "epoch": 4.66798418972332, "grad_norm": 3.542649030685425, "learning_rate": 3.840594059405941e-05, "loss": 0.3926, "step": 1181 }, { "epoch": 4.67193675889328, "grad_norm": 2.890378713607788, "learning_rate": 3.83960396039604e-05, "loss": 0.2373, "step": 1182 }, { "epoch": 4.675889328063241, "grad_norm": 3.413198709487915, "learning_rate": 3.838613861386139e-05, "loss": 0.3062, "step": 1183 }, { "epoch": 4.679841897233201, "grad_norm": 3.105306625366211, "learning_rate": 3.837623762376238e-05, "loss": 0.3879, "step": 1184 }, { "epoch": 4.683794466403162, "grad_norm": 2.8441550731658936, "learning_rate": 3.8366336633663367e-05, "loss": 0.2111, "step": 1185 }, { "epoch": 4.687747035573123, "grad_norm": 2.471747875213623, "learning_rate": 3.835643564356436e-05, "loss": 0.3034, "step": 1186 }, { "epoch": 4.691699604743083, "grad_norm": 3.592402935028076, "learning_rate": 3.834653465346535e-05, "loss": 0.3171, "step": 1187 }, { "epoch": 4.695652173913043, "grad_norm": 2.562258243560791, "learning_rate": 3.833663366336634e-05, "loss": 0.1892, "step": 1188 }, { "epoch": 4.699604743083004, "grad_norm": 3.334111452102661, "learning_rate": 3.832673267326733e-05, "loss": 0.2704, "step": 1189 }, { "epoch": 4.703557312252965, "grad_norm": 2.784811496734619, "learning_rate": 3.831683168316832e-05, "loss": 0.2415, "step": 1190 }, { "epoch": 4.707509881422925, "grad_norm": 3.621302843093872, "learning_rate": 3.8306930693069306e-05, "loss": 0.3353, "step": 1191 }, { "epoch": 4.711462450592886, "grad_norm": 3.1611618995666504, "learning_rate": 3.82970297029703e-05, "loss": 0.2752, "step": 1192 }, { "epoch": 4.715415019762846, "grad_norm": 2.8270256519317627, "learning_rate": 3.828712871287129e-05, "loss": 0.2154, "step": 1193 }, { "epoch": 4.719367588932807, "grad_norm": 2.8247456550598145, "learning_rate": 3.827722772277228e-05, "loss": 0.2993, "step": 1194 }, { "epoch": 4.723320158102767, "grad_norm": 2.929565191268921, "learning_rate": 3.826732673267327e-05, "loss": 0.3625, "step": 1195 }, { "epoch": 4.7272727272727275, "grad_norm": 3.114823579788208, "learning_rate": 3.825742574257426e-05, "loss": 0.3635, "step": 1196 }, { "epoch": 4.7312252964426875, "grad_norm": 3.9383468627929688, "learning_rate": 3.824752475247525e-05, "loss": 0.2552, "step": 1197 }, { "epoch": 4.735177865612648, "grad_norm": 3.435030698776245, "learning_rate": 3.823762376237624e-05, "loss": 0.3069, "step": 1198 }, { "epoch": 4.739130434782608, "grad_norm": 2.925692558288574, "learning_rate": 3.822772277227723e-05, "loss": 0.3484, "step": 1199 }, { "epoch": 4.743083003952569, "grad_norm": 3.410959482192993, "learning_rate": 3.821782178217822e-05, "loss": 0.2744, "step": 1200 }, { "epoch": 4.747035573122529, "grad_norm": 3.5181620121002197, "learning_rate": 3.820792079207921e-05, "loss": 0.2643, "step": 1201 }, { "epoch": 4.75098814229249, "grad_norm": 2.835466146469116, "learning_rate": 3.81980198019802e-05, "loss": 0.3439, "step": 1202 }, { "epoch": 4.75494071146245, "grad_norm": 3.6465229988098145, "learning_rate": 3.818811881188119e-05, "loss": 0.3605, "step": 1203 }, { "epoch": 4.758893280632411, "grad_norm": 3.286221742630005, "learning_rate": 3.817821782178218e-05, "loss": 0.3849, "step": 1204 }, { "epoch": 4.762845849802371, "grad_norm": 3.071850538253784, "learning_rate": 3.816831683168317e-05, "loss": 0.3617, "step": 1205 }, { "epoch": 4.766798418972332, "grad_norm": 2.6553611755371094, "learning_rate": 3.815841584158416e-05, "loss": 0.2509, "step": 1206 }, { "epoch": 4.770750988142293, "grad_norm": 2.8507890701293945, "learning_rate": 3.814851485148515e-05, "loss": 0.3405, "step": 1207 }, { "epoch": 4.774703557312253, "grad_norm": 2.98594331741333, "learning_rate": 3.813861386138614e-05, "loss": 0.2618, "step": 1208 }, { "epoch": 4.778656126482213, "grad_norm": 2.5334155559539795, "learning_rate": 3.812871287128713e-05, "loss": 0.2759, "step": 1209 }, { "epoch": 4.782608695652174, "grad_norm": 3.0966250896453857, "learning_rate": 3.811881188118812e-05, "loss": 0.2415, "step": 1210 }, { "epoch": 4.786561264822135, "grad_norm": 3.011528730392456, "learning_rate": 3.8108910891089114e-05, "loss": 0.283, "step": 1211 }, { "epoch": 4.790513833992095, "grad_norm": 3.094393491744995, "learning_rate": 3.80990099009901e-05, "loss": 0.298, "step": 1212 }, { "epoch": 4.794466403162056, "grad_norm": 3.0847909450531006, "learning_rate": 3.808910891089109e-05, "loss": 0.2873, "step": 1213 }, { "epoch": 4.798418972332016, "grad_norm": 3.5377469062805176, "learning_rate": 3.807920792079208e-05, "loss": 0.2533, "step": 1214 }, { "epoch": 4.8023715415019765, "grad_norm": 3.969494104385376, "learning_rate": 3.806930693069307e-05, "loss": 0.3994, "step": 1215 }, { "epoch": 4.8063241106719365, "grad_norm": 2.6385841369628906, "learning_rate": 3.805940594059406e-05, "loss": 0.3196, "step": 1216 }, { "epoch": 4.810276679841897, "grad_norm": 3.7770674228668213, "learning_rate": 3.8049504950495054e-05, "loss": 0.3314, "step": 1217 }, { "epoch": 4.8142292490118574, "grad_norm": 3.088270902633667, "learning_rate": 3.8039603960396045e-05, "loss": 0.3285, "step": 1218 }, { "epoch": 4.818181818181818, "grad_norm": 2.0955703258514404, "learning_rate": 3.802970297029703e-05, "loss": 0.1921, "step": 1219 }, { "epoch": 4.822134387351778, "grad_norm": 2.8468704223632812, "learning_rate": 3.801980198019802e-05, "loss": 0.2684, "step": 1220 }, { "epoch": 4.826086956521739, "grad_norm": 2.674147844314575, "learning_rate": 3.800990099009901e-05, "loss": 0.3122, "step": 1221 }, { "epoch": 4.830039525691699, "grad_norm": 4.038686275482178, "learning_rate": 3.8e-05, "loss": 0.3708, "step": 1222 }, { "epoch": 4.83399209486166, "grad_norm": 5.218181610107422, "learning_rate": 3.799009900990099e-05, "loss": 0.3658, "step": 1223 }, { "epoch": 4.837944664031621, "grad_norm": 3.9841506481170654, "learning_rate": 3.7980198019801984e-05, "loss": 0.3691, "step": 1224 }, { "epoch": 4.841897233201581, "grad_norm": 4.225796699523926, "learning_rate": 3.7970297029702975e-05, "loss": 0.3076, "step": 1225 }, { "epoch": 4.845849802371541, "grad_norm": 2.942944288253784, "learning_rate": 3.796039603960396e-05, "loss": 0.3842, "step": 1226 }, { "epoch": 4.849802371541502, "grad_norm": 3.26554536819458, "learning_rate": 3.795049504950495e-05, "loss": 0.3213, "step": 1227 }, { "epoch": 4.853754940711463, "grad_norm": 3.363886833190918, "learning_rate": 3.794059405940594e-05, "loss": 0.3569, "step": 1228 }, { "epoch": 4.857707509881423, "grad_norm": 2.8089678287506104, "learning_rate": 3.793069306930693e-05, "loss": 0.3344, "step": 1229 }, { "epoch": 4.861660079051384, "grad_norm": 3.1895647048950195, "learning_rate": 3.7920792079207924e-05, "loss": 0.2466, "step": 1230 }, { "epoch": 4.865612648221344, "grad_norm": 2.4909961223602295, "learning_rate": 3.7910891089108915e-05, "loss": 0.3256, "step": 1231 }, { "epoch": 4.869565217391305, "grad_norm": 2.876573324203491, "learning_rate": 3.7900990099009906e-05, "loss": 0.2069, "step": 1232 }, { "epoch": 4.873517786561265, "grad_norm": 2.887436866760254, "learning_rate": 3.789108910891089e-05, "loss": 0.2419, "step": 1233 }, { "epoch": 4.877470355731226, "grad_norm": 2.743701696395874, "learning_rate": 3.788118811881188e-05, "loss": 0.3012, "step": 1234 }, { "epoch": 4.881422924901186, "grad_norm": 3.561992883682251, "learning_rate": 3.787128712871287e-05, "loss": 0.2446, "step": 1235 }, { "epoch": 4.8853754940711465, "grad_norm": 2.829221248626709, "learning_rate": 3.7861386138613863e-05, "loss": 0.2551, "step": 1236 }, { "epoch": 4.8893280632411065, "grad_norm": 3.6672847270965576, "learning_rate": 3.7851485148514855e-05, "loss": 0.4051, "step": 1237 }, { "epoch": 4.893280632411067, "grad_norm": 3.223508596420288, "learning_rate": 3.7841584158415846e-05, "loss": 0.2284, "step": 1238 }, { "epoch": 4.897233201581027, "grad_norm": 3.297940492630005, "learning_rate": 3.783168316831684e-05, "loss": 0.3977, "step": 1239 }, { "epoch": 4.901185770750988, "grad_norm": 3.2640323638916016, "learning_rate": 3.782178217821782e-05, "loss": 0.3088, "step": 1240 }, { "epoch": 4.905138339920948, "grad_norm": 3.330292224884033, "learning_rate": 3.781188118811881e-05, "loss": 0.317, "step": 1241 }, { "epoch": 4.909090909090909, "grad_norm": 3.064439296722412, "learning_rate": 3.78019801980198e-05, "loss": 0.3333, "step": 1242 }, { "epoch": 4.913043478260869, "grad_norm": 3.634432315826416, "learning_rate": 3.7792079207920794e-05, "loss": 0.3578, "step": 1243 }, { "epoch": 4.91699604743083, "grad_norm": 3.4017584323883057, "learning_rate": 3.7782178217821785e-05, "loss": 0.2972, "step": 1244 }, { "epoch": 4.920948616600791, "grad_norm": 3.1678974628448486, "learning_rate": 3.7772277227722776e-05, "loss": 0.3947, "step": 1245 }, { "epoch": 4.924901185770751, "grad_norm": 2.8716158866882324, "learning_rate": 3.776237623762377e-05, "loss": 0.3211, "step": 1246 }, { "epoch": 4.928853754940711, "grad_norm": 2.707505702972412, "learning_rate": 3.775247524752475e-05, "loss": 0.2497, "step": 1247 }, { "epoch": 4.932806324110672, "grad_norm": 3.1655285358428955, "learning_rate": 3.774257425742574e-05, "loss": 0.3144, "step": 1248 }, { "epoch": 4.936758893280633, "grad_norm": 3.042510747909546, "learning_rate": 3.7732673267326734e-05, "loss": 0.214, "step": 1249 }, { "epoch": 4.940711462450593, "grad_norm": 2.892810106277466, "learning_rate": 3.7722772277227725e-05, "loss": 0.2975, "step": 1250 }, { "epoch": 4.944664031620554, "grad_norm": 2.640181541442871, "learning_rate": 3.7712871287128716e-05, "loss": 0.2593, "step": 1251 }, { "epoch": 4.948616600790514, "grad_norm": 2.1869876384735107, "learning_rate": 3.770297029702971e-05, "loss": 0.1683, "step": 1252 }, { "epoch": 4.952569169960475, "grad_norm": 3.976351737976074, "learning_rate": 3.76930693069307e-05, "loss": 0.3279, "step": 1253 }, { "epoch": 4.956521739130435, "grad_norm": 3.501063823699951, "learning_rate": 3.768316831683168e-05, "loss": 0.2915, "step": 1254 }, { "epoch": 4.9604743083003955, "grad_norm": 2.9900991916656494, "learning_rate": 3.7673267326732673e-05, "loss": 0.3044, "step": 1255 }, { "epoch": 4.9644268774703555, "grad_norm": 2.6596691608428955, "learning_rate": 3.7663366336633664e-05, "loss": 0.3466, "step": 1256 }, { "epoch": 4.968379446640316, "grad_norm": 2.8459532260894775, "learning_rate": 3.7653465346534656e-05, "loss": 0.2542, "step": 1257 }, { "epoch": 4.972332015810276, "grad_norm": 3.122265338897705, "learning_rate": 3.7643564356435647e-05, "loss": 0.3141, "step": 1258 }, { "epoch": 4.976284584980237, "grad_norm": 2.841768980026245, "learning_rate": 3.763366336633664e-05, "loss": 0.3216, "step": 1259 }, { "epoch": 4.980237154150197, "grad_norm": 2.995814800262451, "learning_rate": 3.762376237623763e-05, "loss": 0.2583, "step": 1260 }, { "epoch": 4.984189723320158, "grad_norm": 3.094081163406372, "learning_rate": 3.761386138613861e-05, "loss": 0.1912, "step": 1261 }, { "epoch": 4.988142292490118, "grad_norm": 3.135359525680542, "learning_rate": 3.7603960396039604e-05, "loss": 0.3337, "step": 1262 }, { "epoch": 4.992094861660079, "grad_norm": 3.263326644897461, "learning_rate": 3.7594059405940595e-05, "loss": 0.3374, "step": 1263 }, { "epoch": 4.996047430830039, "grad_norm": 3.6752424240112305, "learning_rate": 3.7584158415841586e-05, "loss": 0.2584, "step": 1264 }, { "epoch": 5.0, "grad_norm": 3.198829174041748, "learning_rate": 3.757425742574258e-05, "loss": 0.3346, "step": 1265 }, { "epoch": 5.003952569169961, "grad_norm": 2.3619422912597656, "learning_rate": 3.756435643564357e-05, "loss": 0.1637, "step": 1266 }, { "epoch": 5.007905138339921, "grad_norm": 2.2762975692749023, "learning_rate": 3.755445544554456e-05, "loss": 0.1245, "step": 1267 }, { "epoch": 5.011857707509882, "grad_norm": 1.6592072248458862, "learning_rate": 3.7544554455445544e-05, "loss": 0.1566, "step": 1268 }, { "epoch": 5.015810276679842, "grad_norm": 2.09881854057312, "learning_rate": 3.7534653465346535e-05, "loss": 0.1181, "step": 1269 }, { "epoch": 5.019762845849803, "grad_norm": 2.657099723815918, "learning_rate": 3.7524752475247526e-05, "loss": 0.1504, "step": 1270 }, { "epoch": 5.023715415019763, "grad_norm": 2.7582905292510986, "learning_rate": 3.751485148514852e-05, "loss": 0.1089, "step": 1271 }, { "epoch": 5.027667984189724, "grad_norm": 2.7810490131378174, "learning_rate": 3.750495049504951e-05, "loss": 0.1148, "step": 1272 }, { "epoch": 5.031620553359684, "grad_norm": 3.0729897022247314, "learning_rate": 3.74950495049505e-05, "loss": 0.1461, "step": 1273 }, { "epoch": 5.0355731225296445, "grad_norm": 3.383241891860962, "learning_rate": 3.748514851485149e-05, "loss": 0.167, "step": 1274 }, { "epoch": 5.0395256916996045, "grad_norm": 2.645568609237671, "learning_rate": 3.7475247524752474e-05, "loss": 0.1705, "step": 1275 }, { "epoch": 5.043478260869565, "grad_norm": 3.4041147232055664, "learning_rate": 3.7465346534653465e-05, "loss": 0.1787, "step": 1276 }, { "epoch": 5.047430830039525, "grad_norm": 3.2558748722076416, "learning_rate": 3.7455445544554457e-05, "loss": 0.1501, "step": 1277 }, { "epoch": 5.051383399209486, "grad_norm": 2.6506659984588623, "learning_rate": 3.744554455445545e-05, "loss": 0.1397, "step": 1278 }, { "epoch": 5.055335968379446, "grad_norm": 2.59626841545105, "learning_rate": 3.743564356435644e-05, "loss": 0.1422, "step": 1279 }, { "epoch": 5.059288537549407, "grad_norm": 3.618309497833252, "learning_rate": 3.742574257425743e-05, "loss": 0.1522, "step": 1280 }, { "epoch": 5.063241106719367, "grad_norm": 2.8686888217926025, "learning_rate": 3.741584158415842e-05, "loss": 0.1885, "step": 1281 }, { "epoch": 5.067193675889328, "grad_norm": 3.024733304977417, "learning_rate": 3.7405940594059405e-05, "loss": 0.1424, "step": 1282 }, { "epoch": 5.071146245059288, "grad_norm": 2.0807433128356934, "learning_rate": 3.7396039603960396e-05, "loss": 0.1347, "step": 1283 }, { "epoch": 5.075098814229249, "grad_norm": 2.5679502487182617, "learning_rate": 3.738613861386139e-05, "loss": 0.1444, "step": 1284 }, { "epoch": 5.07905138339921, "grad_norm": 2.2008724212646484, "learning_rate": 3.737623762376238e-05, "loss": 0.1441, "step": 1285 }, { "epoch": 5.08300395256917, "grad_norm": 2.4898695945739746, "learning_rate": 3.736633663366337e-05, "loss": 0.1521, "step": 1286 }, { "epoch": 5.086956521739131, "grad_norm": 2.8401341438293457, "learning_rate": 3.735643564356436e-05, "loss": 0.1621, "step": 1287 }, { "epoch": 5.090909090909091, "grad_norm": 2.1253950595855713, "learning_rate": 3.734653465346535e-05, "loss": 0.1116, "step": 1288 }, { "epoch": 5.094861660079052, "grad_norm": 2.806594133377075, "learning_rate": 3.7336633663366336e-05, "loss": 0.1832, "step": 1289 }, { "epoch": 5.098814229249012, "grad_norm": 2.6861610412597656, "learning_rate": 3.732673267326733e-05, "loss": 0.1647, "step": 1290 }, { "epoch": 5.102766798418973, "grad_norm": 3.216548204421997, "learning_rate": 3.731683168316832e-05, "loss": 0.1724, "step": 1291 }, { "epoch": 5.106719367588933, "grad_norm": 3.619692325592041, "learning_rate": 3.730693069306931e-05, "loss": 0.1307, "step": 1292 }, { "epoch": 5.1106719367588935, "grad_norm": 2.3871982097625732, "learning_rate": 3.72970297029703e-05, "loss": 0.1368, "step": 1293 }, { "epoch": 5.1146245059288535, "grad_norm": 3.380765438079834, "learning_rate": 3.728712871287129e-05, "loss": 0.1454, "step": 1294 }, { "epoch": 5.118577075098814, "grad_norm": 2.503779411315918, "learning_rate": 3.727722772277228e-05, "loss": 0.1259, "step": 1295 }, { "epoch": 5.122529644268774, "grad_norm": 2.8413643836975098, "learning_rate": 3.7267326732673266e-05, "loss": 0.1627, "step": 1296 }, { "epoch": 5.126482213438735, "grad_norm": 3.02528977394104, "learning_rate": 3.725742574257426e-05, "loss": 0.1528, "step": 1297 }, { "epoch": 5.130434782608695, "grad_norm": 2.6431171894073486, "learning_rate": 3.724752475247525e-05, "loss": 0.1522, "step": 1298 }, { "epoch": 5.134387351778656, "grad_norm": 2.7816686630249023, "learning_rate": 3.723762376237624e-05, "loss": 0.1549, "step": 1299 }, { "epoch": 5.138339920948616, "grad_norm": 3.3987019062042236, "learning_rate": 3.722772277227723e-05, "loss": 0.184, "step": 1300 }, { "epoch": 5.142292490118577, "grad_norm": 2.720964193344116, "learning_rate": 3.721782178217822e-05, "loss": 0.1607, "step": 1301 }, { "epoch": 5.146245059288537, "grad_norm": 3.4037866592407227, "learning_rate": 3.7207920792079206e-05, "loss": 0.1729, "step": 1302 }, { "epoch": 5.150197628458498, "grad_norm": 3.188176155090332, "learning_rate": 3.71980198019802e-05, "loss": 0.1528, "step": 1303 }, { "epoch": 5.154150197628459, "grad_norm": 2.412764310836792, "learning_rate": 3.718811881188119e-05, "loss": 0.1504, "step": 1304 }, { "epoch": 5.158102766798419, "grad_norm": 2.447441339492798, "learning_rate": 3.717821782178218e-05, "loss": 0.1105, "step": 1305 }, { "epoch": 5.16205533596838, "grad_norm": 2.423706531524658, "learning_rate": 3.716831683168317e-05, "loss": 0.1177, "step": 1306 }, { "epoch": 5.16600790513834, "grad_norm": 2.367359161376953, "learning_rate": 3.715841584158416e-05, "loss": 0.1574, "step": 1307 }, { "epoch": 5.169960474308301, "grad_norm": 2.643998622894287, "learning_rate": 3.714851485148515e-05, "loss": 0.1783, "step": 1308 }, { "epoch": 5.173913043478261, "grad_norm": 2.352839231491089, "learning_rate": 3.713861386138614e-05, "loss": 0.139, "step": 1309 }, { "epoch": 5.177865612648222, "grad_norm": 3.0182619094848633, "learning_rate": 3.712871287128713e-05, "loss": 0.1478, "step": 1310 }, { "epoch": 5.181818181818182, "grad_norm": 2.8196358680725098, "learning_rate": 3.711881188118812e-05, "loss": 0.1234, "step": 1311 }, { "epoch": 5.1857707509881426, "grad_norm": 3.4850592613220215, "learning_rate": 3.710891089108911e-05, "loss": 0.147, "step": 1312 }, { "epoch": 5.189723320158103, "grad_norm": 2.3942713737487793, "learning_rate": 3.70990099009901e-05, "loss": 0.1341, "step": 1313 }, { "epoch": 5.1936758893280635, "grad_norm": 2.9597983360290527, "learning_rate": 3.708910891089109e-05, "loss": 0.1481, "step": 1314 }, { "epoch": 5.1976284584980235, "grad_norm": 2.4172310829162598, "learning_rate": 3.707920792079208e-05, "loss": 0.1423, "step": 1315 }, { "epoch": 5.201581027667984, "grad_norm": 2.686640501022339, "learning_rate": 3.706930693069307e-05, "loss": 0.157, "step": 1316 }, { "epoch": 5.205533596837944, "grad_norm": 2.884676933288574, "learning_rate": 3.705940594059406e-05, "loss": 0.1734, "step": 1317 }, { "epoch": 5.209486166007905, "grad_norm": 3.323762893676758, "learning_rate": 3.704950495049505e-05, "loss": 0.164, "step": 1318 }, { "epoch": 5.213438735177865, "grad_norm": 3.041482925415039, "learning_rate": 3.703960396039604e-05, "loss": 0.1248, "step": 1319 }, { "epoch": 5.217391304347826, "grad_norm": 2.333055019378662, "learning_rate": 3.702970297029703e-05, "loss": 0.1381, "step": 1320 }, { "epoch": 5.221343873517786, "grad_norm": 2.632110834121704, "learning_rate": 3.701980198019802e-05, "loss": 0.1551, "step": 1321 }, { "epoch": 5.225296442687747, "grad_norm": 2.7117760181427, "learning_rate": 3.7009900990099014e-05, "loss": 0.1502, "step": 1322 }, { "epoch": 5.229249011857707, "grad_norm": 2.3433663845062256, "learning_rate": 3.7e-05, "loss": 0.1708, "step": 1323 }, { "epoch": 5.233201581027668, "grad_norm": 2.763758897781372, "learning_rate": 3.699009900990099e-05, "loss": 0.1506, "step": 1324 }, { "epoch": 5.237154150197629, "grad_norm": 2.8548266887664795, "learning_rate": 3.698019801980198e-05, "loss": 0.1472, "step": 1325 }, { "epoch": 5.241106719367589, "grad_norm": 3.0606746673583984, "learning_rate": 3.697029702970297e-05, "loss": 0.1664, "step": 1326 }, { "epoch": 5.24505928853755, "grad_norm": 6.298571586608887, "learning_rate": 3.696039603960396e-05, "loss": 0.1979, "step": 1327 }, { "epoch": 5.24901185770751, "grad_norm": 3.163912057876587, "learning_rate": 3.6950495049504953e-05, "loss": 0.1539, "step": 1328 }, { "epoch": 5.252964426877471, "grad_norm": 2.8823766708374023, "learning_rate": 3.6940594059405945e-05, "loss": 0.1784, "step": 1329 }, { "epoch": 5.256916996047431, "grad_norm": 2.7282190322875977, "learning_rate": 3.693069306930693e-05, "loss": 0.1587, "step": 1330 }, { "epoch": 5.260869565217392, "grad_norm": 2.9934959411621094, "learning_rate": 3.692079207920792e-05, "loss": 0.2401, "step": 1331 }, { "epoch": 5.264822134387352, "grad_norm": 2.8303115367889404, "learning_rate": 3.691089108910891e-05, "loss": 0.1774, "step": 1332 }, { "epoch": 5.2687747035573125, "grad_norm": 2.1939761638641357, "learning_rate": 3.69009900990099e-05, "loss": 0.1219, "step": 1333 }, { "epoch": 5.2727272727272725, "grad_norm": 3.203801155090332, "learning_rate": 3.689108910891089e-05, "loss": 0.1906, "step": 1334 }, { "epoch": 5.276679841897233, "grad_norm": 2.317878484725952, "learning_rate": 3.6881188118811884e-05, "loss": 0.157, "step": 1335 }, { "epoch": 5.280632411067193, "grad_norm": 2.9838225841522217, "learning_rate": 3.6871287128712875e-05, "loss": 0.1644, "step": 1336 }, { "epoch": 5.284584980237154, "grad_norm": 2.5923445224761963, "learning_rate": 3.686138613861386e-05, "loss": 0.1388, "step": 1337 }, { "epoch": 5.288537549407114, "grad_norm": 2.182795286178589, "learning_rate": 3.685148514851485e-05, "loss": 0.1131, "step": 1338 }, { "epoch": 5.292490118577075, "grad_norm": 2.2978084087371826, "learning_rate": 3.684158415841584e-05, "loss": 0.141, "step": 1339 }, { "epoch": 5.296442687747035, "grad_norm": 3.249652624130249, "learning_rate": 3.683168316831683e-05, "loss": 0.1883, "step": 1340 }, { "epoch": 5.300395256916996, "grad_norm": 3.2128071784973145, "learning_rate": 3.6821782178217824e-05, "loss": 0.1537, "step": 1341 }, { "epoch": 5.304347826086957, "grad_norm": 3.789611339569092, "learning_rate": 3.6811881188118815e-05, "loss": 0.1634, "step": 1342 }, { "epoch": 5.308300395256917, "grad_norm": 2.9401609897613525, "learning_rate": 3.6801980198019806e-05, "loss": 0.1572, "step": 1343 }, { "epoch": 5.312252964426877, "grad_norm": 3.4480156898498535, "learning_rate": 3.679207920792079e-05, "loss": 0.1674, "step": 1344 }, { "epoch": 5.316205533596838, "grad_norm": 3.6320273876190186, "learning_rate": 3.678217821782178e-05, "loss": 0.1578, "step": 1345 }, { "epoch": 5.320158102766799, "grad_norm": 2.356677770614624, "learning_rate": 3.677227722772277e-05, "loss": 0.1257, "step": 1346 }, { "epoch": 5.324110671936759, "grad_norm": 2.692344903945923, "learning_rate": 3.676237623762376e-05, "loss": 0.1725, "step": 1347 }, { "epoch": 5.32806324110672, "grad_norm": 3.4734365940093994, "learning_rate": 3.6752475247524754e-05, "loss": 0.2149, "step": 1348 }, { "epoch": 5.33201581027668, "grad_norm": 3.12172269821167, "learning_rate": 3.6742574257425746e-05, "loss": 0.1963, "step": 1349 }, { "epoch": 5.335968379446641, "grad_norm": 2.732015609741211, "learning_rate": 3.6732673267326737e-05, "loss": 0.1479, "step": 1350 }, { "epoch": 5.339920948616601, "grad_norm": 2.513662815093994, "learning_rate": 3.672277227722772e-05, "loss": 0.1076, "step": 1351 }, { "epoch": 5.3438735177865615, "grad_norm": 2.9846608638763428, "learning_rate": 3.671287128712871e-05, "loss": 0.1758, "step": 1352 }, { "epoch": 5.3478260869565215, "grad_norm": 3.14731764793396, "learning_rate": 3.67029702970297e-05, "loss": 0.1848, "step": 1353 }, { "epoch": 5.351778656126482, "grad_norm": 2.4576733112335205, "learning_rate": 3.6693069306930694e-05, "loss": 0.2002, "step": 1354 }, { "epoch": 5.355731225296442, "grad_norm": 3.271109104156494, "learning_rate": 3.6683168316831685e-05, "loss": 0.1861, "step": 1355 }, { "epoch": 5.359683794466403, "grad_norm": 2.0204384326934814, "learning_rate": 3.6673267326732676e-05, "loss": 0.093, "step": 1356 }, { "epoch": 5.363636363636363, "grad_norm": 2.653103828430176, "learning_rate": 3.666336633663367e-05, "loss": 0.1903, "step": 1357 }, { "epoch": 5.367588932806324, "grad_norm": 3.2146623134613037, "learning_rate": 3.665346534653465e-05, "loss": 0.1512, "step": 1358 }, { "epoch": 5.371541501976284, "grad_norm": 2.7990360260009766, "learning_rate": 3.664356435643564e-05, "loss": 0.1447, "step": 1359 }, { "epoch": 5.375494071146245, "grad_norm": 2.7689239978790283, "learning_rate": 3.6633663366336634e-05, "loss": 0.1691, "step": 1360 }, { "epoch": 5.379446640316205, "grad_norm": 3.0795767307281494, "learning_rate": 3.6623762376237625e-05, "loss": 0.195, "step": 1361 }, { "epoch": 5.383399209486166, "grad_norm": 2.3812732696533203, "learning_rate": 3.6613861386138616e-05, "loss": 0.1161, "step": 1362 }, { "epoch": 5.387351778656127, "grad_norm": 3.29968523979187, "learning_rate": 3.660396039603961e-05, "loss": 0.1723, "step": 1363 }, { "epoch": 5.391304347826087, "grad_norm": 3.314619779586792, "learning_rate": 3.65940594059406e-05, "loss": 0.217, "step": 1364 }, { "epoch": 5.395256916996048, "grad_norm": 2.7420742511749268, "learning_rate": 3.658415841584158e-05, "loss": 0.1517, "step": 1365 }, { "epoch": 5.399209486166008, "grad_norm": 2.5389604568481445, "learning_rate": 3.657425742574257e-05, "loss": 0.1748, "step": 1366 }, { "epoch": 5.403162055335969, "grad_norm": 2.7112059593200684, "learning_rate": 3.6564356435643564e-05, "loss": 0.1815, "step": 1367 }, { "epoch": 5.407114624505929, "grad_norm": 2.87528657913208, "learning_rate": 3.6554455445544555e-05, "loss": 0.2173, "step": 1368 }, { "epoch": 5.41106719367589, "grad_norm": 2.787585973739624, "learning_rate": 3.6544554455445546e-05, "loss": 0.2012, "step": 1369 }, { "epoch": 5.41501976284585, "grad_norm": 2.6649298667907715, "learning_rate": 3.653465346534654e-05, "loss": 0.1236, "step": 1370 }, { "epoch": 5.4189723320158105, "grad_norm": 2.5338611602783203, "learning_rate": 3.652475247524753e-05, "loss": 0.1384, "step": 1371 }, { "epoch": 5.4229249011857705, "grad_norm": 2.7742929458618164, "learning_rate": 3.651485148514851e-05, "loss": 0.1661, "step": 1372 }, { "epoch": 5.426877470355731, "grad_norm": 2.9334542751312256, "learning_rate": 3.6504950495049504e-05, "loss": 0.1759, "step": 1373 }, { "epoch": 5.430830039525691, "grad_norm": 2.516188383102417, "learning_rate": 3.6495049504950495e-05, "loss": 0.1975, "step": 1374 }, { "epoch": 5.434782608695652, "grad_norm": 2.773871421813965, "learning_rate": 3.6485148514851486e-05, "loss": 0.1566, "step": 1375 }, { "epoch": 5.438735177865612, "grad_norm": 2.88891863822937, "learning_rate": 3.647524752475248e-05, "loss": 0.1732, "step": 1376 }, { "epoch": 5.442687747035573, "grad_norm": 2.6928138732910156, "learning_rate": 3.646534653465347e-05, "loss": 0.1551, "step": 1377 }, { "epoch": 5.446640316205533, "grad_norm": 4.594505310058594, "learning_rate": 3.645544554455446e-05, "loss": 0.2182, "step": 1378 }, { "epoch": 5.450592885375494, "grad_norm": 2.4643757343292236, "learning_rate": 3.6445544554455444e-05, "loss": 0.1426, "step": 1379 }, { "epoch": 5.454545454545454, "grad_norm": 3.460437297821045, "learning_rate": 3.6435643564356435e-05, "loss": 0.1999, "step": 1380 }, { "epoch": 5.458498023715415, "grad_norm": 3.014364242553711, "learning_rate": 3.6425742574257426e-05, "loss": 0.1747, "step": 1381 }, { "epoch": 5.462450592885375, "grad_norm": 2.887953758239746, "learning_rate": 3.641584158415842e-05, "loss": 0.1862, "step": 1382 }, { "epoch": 5.466403162055336, "grad_norm": 2.8970792293548584, "learning_rate": 3.640594059405941e-05, "loss": 0.1346, "step": 1383 }, { "epoch": 5.470355731225297, "grad_norm": 3.414642333984375, "learning_rate": 3.63960396039604e-05, "loss": 0.1478, "step": 1384 }, { "epoch": 5.474308300395257, "grad_norm": 2.37733793258667, "learning_rate": 3.638613861386139e-05, "loss": 0.164, "step": 1385 }, { "epoch": 5.478260869565218, "grad_norm": 2.3274192810058594, "learning_rate": 3.6376237623762374e-05, "loss": 0.1398, "step": 1386 }, { "epoch": 5.482213438735178, "grad_norm": 2.8411664962768555, "learning_rate": 3.6366336633663365e-05, "loss": 0.1437, "step": 1387 }, { "epoch": 5.486166007905139, "grad_norm": 3.263052225112915, "learning_rate": 3.6356435643564356e-05, "loss": 0.1644, "step": 1388 }, { "epoch": 5.490118577075099, "grad_norm": 2.4976754188537598, "learning_rate": 3.634653465346535e-05, "loss": 0.1416, "step": 1389 }, { "epoch": 5.4940711462450595, "grad_norm": 2.6434507369995117, "learning_rate": 3.633663366336634e-05, "loss": 0.1523, "step": 1390 }, { "epoch": 5.4980237154150196, "grad_norm": 3.6213502883911133, "learning_rate": 3.632673267326733e-05, "loss": 0.1958, "step": 1391 }, { "epoch": 5.5019762845849804, "grad_norm": 2.7617595195770264, "learning_rate": 3.631683168316832e-05, "loss": 0.1745, "step": 1392 }, { "epoch": 5.5059288537549405, "grad_norm": 2.665123224258423, "learning_rate": 3.6306930693069305e-05, "loss": 0.1391, "step": 1393 }, { "epoch": 5.509881422924901, "grad_norm": 2.847140073776245, "learning_rate": 3.6297029702970296e-05, "loss": 0.143, "step": 1394 }, { "epoch": 5.513833992094861, "grad_norm": 3.610077142715454, "learning_rate": 3.628712871287129e-05, "loss": 0.1723, "step": 1395 }, { "epoch": 5.517786561264822, "grad_norm": 2.880545139312744, "learning_rate": 3.627722772277228e-05, "loss": 0.1352, "step": 1396 }, { "epoch": 5.521739130434782, "grad_norm": 2.7175097465515137, "learning_rate": 3.626732673267327e-05, "loss": 0.1517, "step": 1397 }, { "epoch": 5.525691699604743, "grad_norm": 2.4527981281280518, "learning_rate": 3.625742574257426e-05, "loss": 0.1689, "step": 1398 }, { "epoch": 5.529644268774703, "grad_norm": 2.473822593688965, "learning_rate": 3.624752475247525e-05, "loss": 0.1854, "step": 1399 }, { "epoch": 5.533596837944664, "grad_norm": 3.462496757507324, "learning_rate": 3.6237623762376236e-05, "loss": 0.1814, "step": 1400 }, { "epoch": 5.537549407114625, "grad_norm": 3.3680777549743652, "learning_rate": 3.622772277227723e-05, "loss": 0.1906, "step": 1401 }, { "epoch": 5.541501976284585, "grad_norm": 2.8276915550231934, "learning_rate": 3.621782178217822e-05, "loss": 0.1456, "step": 1402 }, { "epoch": 5.545454545454545, "grad_norm": 2.5889062881469727, "learning_rate": 3.620792079207921e-05, "loss": 0.1248, "step": 1403 }, { "epoch": 5.549407114624506, "grad_norm": 3.017193078994751, "learning_rate": 3.61980198019802e-05, "loss": 0.2061, "step": 1404 }, { "epoch": 5.553359683794467, "grad_norm": 2.7323596477508545, "learning_rate": 3.618811881188119e-05, "loss": 0.174, "step": 1405 }, { "epoch": 5.557312252964427, "grad_norm": 2.952479600906372, "learning_rate": 3.617821782178218e-05, "loss": 0.1971, "step": 1406 }, { "epoch": 5.561264822134388, "grad_norm": 3.2245371341705322, "learning_rate": 3.6168316831683166e-05, "loss": 0.1476, "step": 1407 }, { "epoch": 5.565217391304348, "grad_norm": 2.861159086227417, "learning_rate": 3.615841584158416e-05, "loss": 0.1312, "step": 1408 }, { "epoch": 5.569169960474309, "grad_norm": 2.777350902557373, "learning_rate": 3.614851485148515e-05, "loss": 0.1493, "step": 1409 }, { "epoch": 5.573122529644269, "grad_norm": 2.572765588760376, "learning_rate": 3.613861386138614e-05, "loss": 0.1763, "step": 1410 }, { "epoch": 5.5770750988142295, "grad_norm": 2.7786428928375244, "learning_rate": 3.612871287128713e-05, "loss": 0.1569, "step": 1411 }, { "epoch": 5.5810276679841895, "grad_norm": 2.5611584186553955, "learning_rate": 3.611881188118812e-05, "loss": 0.1696, "step": 1412 }, { "epoch": 5.58498023715415, "grad_norm": 3.1623475551605225, "learning_rate": 3.610891089108911e-05, "loss": 0.1838, "step": 1413 }, { "epoch": 5.58893280632411, "grad_norm": 3.590118408203125, "learning_rate": 3.60990099009901e-05, "loss": 0.1767, "step": 1414 }, { "epoch": 5.592885375494071, "grad_norm": 2.8701720237731934, "learning_rate": 3.608910891089109e-05, "loss": 0.1597, "step": 1415 }, { "epoch": 5.596837944664031, "grad_norm": 2.9249684810638428, "learning_rate": 3.607920792079208e-05, "loss": 0.188, "step": 1416 }, { "epoch": 5.600790513833992, "grad_norm": 2.418109655380249, "learning_rate": 3.606930693069307e-05, "loss": 0.1196, "step": 1417 }, { "epoch": 5.604743083003952, "grad_norm": 2.385404586791992, "learning_rate": 3.605940594059406e-05, "loss": 0.1808, "step": 1418 }, { "epoch": 5.608695652173913, "grad_norm": 3.4137980937957764, "learning_rate": 3.604950495049505e-05, "loss": 0.2017, "step": 1419 }, { "epoch": 5.612648221343873, "grad_norm": 2.5052273273468018, "learning_rate": 3.603960396039604e-05, "loss": 0.1791, "step": 1420 }, { "epoch": 5.616600790513834, "grad_norm": 2.9823522567749023, "learning_rate": 3.602970297029703e-05, "loss": 0.207, "step": 1421 }, { "epoch": 5.620553359683795, "grad_norm": 2.6294872760772705, "learning_rate": 3.601980198019802e-05, "loss": 0.1517, "step": 1422 }, { "epoch": 5.624505928853755, "grad_norm": 2.7472147941589355, "learning_rate": 3.600990099009901e-05, "loss": 0.1661, "step": 1423 }, { "epoch": 5.628458498023716, "grad_norm": 3.14546275138855, "learning_rate": 3.6e-05, "loss": 0.1715, "step": 1424 }, { "epoch": 5.632411067193676, "grad_norm": 2.601322650909424, "learning_rate": 3.599009900990099e-05, "loss": 0.1444, "step": 1425 }, { "epoch": 5.636363636363637, "grad_norm": 3.20993709564209, "learning_rate": 3.598019801980198e-05, "loss": 0.1578, "step": 1426 }, { "epoch": 5.640316205533597, "grad_norm": 2.9610371589660645, "learning_rate": 3.597029702970297e-05, "loss": 0.1815, "step": 1427 }, { "epoch": 5.644268774703558, "grad_norm": 2.997885227203369, "learning_rate": 3.596039603960396e-05, "loss": 0.1337, "step": 1428 }, { "epoch": 5.648221343873518, "grad_norm": 2.457934856414795, "learning_rate": 3.595049504950495e-05, "loss": 0.1456, "step": 1429 }, { "epoch": 5.6521739130434785, "grad_norm": 2.9209437370300293, "learning_rate": 3.594059405940594e-05, "loss": 0.1581, "step": 1430 }, { "epoch": 5.6561264822134385, "grad_norm": 2.8055078983306885, "learning_rate": 3.593069306930693e-05, "loss": 0.1479, "step": 1431 }, { "epoch": 5.660079051383399, "grad_norm": 3.623997449874878, "learning_rate": 3.592079207920792e-05, "loss": 0.19, "step": 1432 }, { "epoch": 5.664031620553359, "grad_norm": 2.497096538543701, "learning_rate": 3.5910891089108914e-05, "loss": 0.12, "step": 1433 }, { "epoch": 5.66798418972332, "grad_norm": 3.3020169734954834, "learning_rate": 3.59009900990099e-05, "loss": 0.1527, "step": 1434 }, { "epoch": 5.67193675889328, "grad_norm": 3.5407721996307373, "learning_rate": 3.589108910891089e-05, "loss": 0.1938, "step": 1435 }, { "epoch": 5.675889328063241, "grad_norm": 3.9320855140686035, "learning_rate": 3.588118811881188e-05, "loss": 0.1772, "step": 1436 }, { "epoch": 5.679841897233201, "grad_norm": 3.2791812419891357, "learning_rate": 3.587128712871287e-05, "loss": 0.1953, "step": 1437 }, { "epoch": 5.683794466403162, "grad_norm": 2.5159287452697754, "learning_rate": 3.586138613861386e-05, "loss": 0.179, "step": 1438 }, { "epoch": 5.687747035573123, "grad_norm": 3.480724334716797, "learning_rate": 3.585148514851485e-05, "loss": 0.1488, "step": 1439 }, { "epoch": 5.691699604743083, "grad_norm": 2.489060878753662, "learning_rate": 3.5841584158415844e-05, "loss": 0.1601, "step": 1440 }, { "epoch": 5.695652173913043, "grad_norm": 2.8392274379730225, "learning_rate": 3.583168316831683e-05, "loss": 0.1646, "step": 1441 }, { "epoch": 5.699604743083004, "grad_norm": 2.5121779441833496, "learning_rate": 3.582178217821782e-05, "loss": 0.1771, "step": 1442 }, { "epoch": 5.703557312252965, "grad_norm": 2.8975906372070312, "learning_rate": 3.581188118811881e-05, "loss": 0.1623, "step": 1443 }, { "epoch": 5.707509881422925, "grad_norm": 3.010417938232422, "learning_rate": 3.58019801980198e-05, "loss": 0.1818, "step": 1444 }, { "epoch": 5.711462450592886, "grad_norm": 3.4154233932495117, "learning_rate": 3.579207920792079e-05, "loss": 0.1935, "step": 1445 }, { "epoch": 5.715415019762846, "grad_norm": 2.9787628650665283, "learning_rate": 3.5782178217821784e-05, "loss": 0.2024, "step": 1446 }, { "epoch": 5.719367588932807, "grad_norm": 3.305229663848877, "learning_rate": 3.5772277227722775e-05, "loss": 0.1989, "step": 1447 }, { "epoch": 5.723320158102767, "grad_norm": 2.840703248977661, "learning_rate": 3.576237623762376e-05, "loss": 0.1541, "step": 1448 }, { "epoch": 5.7272727272727275, "grad_norm": 3.0812299251556396, "learning_rate": 3.575247524752475e-05, "loss": 0.217, "step": 1449 }, { "epoch": 5.7312252964426875, "grad_norm": 3.107963800430298, "learning_rate": 3.574257425742574e-05, "loss": 0.1829, "step": 1450 }, { "epoch": 5.735177865612648, "grad_norm": 2.6069860458374023, "learning_rate": 3.573267326732673e-05, "loss": 0.1791, "step": 1451 }, { "epoch": 5.739130434782608, "grad_norm": 3.0525293350219727, "learning_rate": 3.5722772277227724e-05, "loss": 0.166, "step": 1452 }, { "epoch": 5.743083003952569, "grad_norm": 2.616919994354248, "learning_rate": 3.5712871287128715e-05, "loss": 0.1783, "step": 1453 }, { "epoch": 5.747035573122529, "grad_norm": 2.2654547691345215, "learning_rate": 3.5702970297029706e-05, "loss": 0.1735, "step": 1454 }, { "epoch": 5.75098814229249, "grad_norm": 3.1299993991851807, "learning_rate": 3.569306930693069e-05, "loss": 0.1737, "step": 1455 }, { "epoch": 5.75494071146245, "grad_norm": 2.5582950115203857, "learning_rate": 3.568316831683168e-05, "loss": 0.2032, "step": 1456 }, { "epoch": 5.758893280632411, "grad_norm": 3.449897527694702, "learning_rate": 3.567326732673267e-05, "loss": 0.2077, "step": 1457 }, { "epoch": 5.762845849802371, "grad_norm": 3.2316360473632812, "learning_rate": 3.566336633663366e-05, "loss": 0.1877, "step": 1458 }, { "epoch": 5.766798418972332, "grad_norm": 2.669506549835205, "learning_rate": 3.5653465346534654e-05, "loss": 0.1182, "step": 1459 }, { "epoch": 5.770750988142293, "grad_norm": 2.440112829208374, "learning_rate": 3.5643564356435645e-05, "loss": 0.1916, "step": 1460 }, { "epoch": 5.774703557312253, "grad_norm": 3.927710771560669, "learning_rate": 3.5633663366336636e-05, "loss": 0.1887, "step": 1461 }, { "epoch": 5.778656126482213, "grad_norm": 3.1733522415161133, "learning_rate": 3.562376237623762e-05, "loss": 0.1656, "step": 1462 }, { "epoch": 5.782608695652174, "grad_norm": 3.30419921875, "learning_rate": 3.561386138613861e-05, "loss": 0.221, "step": 1463 }, { "epoch": 5.786561264822135, "grad_norm": 14.460869789123535, "learning_rate": 3.56039603960396e-05, "loss": 0.159, "step": 1464 }, { "epoch": 5.790513833992095, "grad_norm": 3.1814379692077637, "learning_rate": 3.5594059405940594e-05, "loss": 0.1615, "step": 1465 }, { "epoch": 5.794466403162056, "grad_norm": 2.7740213871002197, "learning_rate": 3.5584158415841585e-05, "loss": 0.2069, "step": 1466 }, { "epoch": 5.798418972332016, "grad_norm": 3.195957660675049, "learning_rate": 3.5574257425742576e-05, "loss": 0.1374, "step": 1467 }, { "epoch": 5.8023715415019765, "grad_norm": 2.9545092582702637, "learning_rate": 3.556435643564357e-05, "loss": 0.192, "step": 1468 }, { "epoch": 5.8063241106719365, "grad_norm": 2.4795610904693604, "learning_rate": 3.555445544554455e-05, "loss": 0.1626, "step": 1469 }, { "epoch": 5.810276679841897, "grad_norm": 3.1961071491241455, "learning_rate": 3.554455445544554e-05, "loss": 0.1634, "step": 1470 }, { "epoch": 5.8142292490118574, "grad_norm": 2.8738620281219482, "learning_rate": 3.5534653465346534e-05, "loss": 0.1675, "step": 1471 }, { "epoch": 5.818181818181818, "grad_norm": 2.8217270374298096, "learning_rate": 3.5524752475247525e-05, "loss": 0.1799, "step": 1472 }, { "epoch": 5.822134387351778, "grad_norm": 3.2957963943481445, "learning_rate": 3.5514851485148516e-05, "loss": 0.2284, "step": 1473 }, { "epoch": 5.826086956521739, "grad_norm": 2.979456901550293, "learning_rate": 3.550495049504951e-05, "loss": 0.1307, "step": 1474 }, { "epoch": 5.830039525691699, "grad_norm": 2.4884164333343506, "learning_rate": 3.54950495049505e-05, "loss": 0.116, "step": 1475 }, { "epoch": 5.83399209486166, "grad_norm": 2.7013206481933594, "learning_rate": 3.548514851485148e-05, "loss": 0.1241, "step": 1476 }, { "epoch": 5.837944664031621, "grad_norm": 2.856135845184326, "learning_rate": 3.547524752475247e-05, "loss": 0.1864, "step": 1477 }, { "epoch": 5.841897233201581, "grad_norm": 2.1065268516540527, "learning_rate": 3.5465346534653464e-05, "loss": 0.1492, "step": 1478 }, { "epoch": 5.845849802371541, "grad_norm": 2.6963982582092285, "learning_rate": 3.5455445544554455e-05, "loss": 0.1953, "step": 1479 }, { "epoch": 5.849802371541502, "grad_norm": 3.0823781490325928, "learning_rate": 3.5445544554455446e-05, "loss": 0.1606, "step": 1480 }, { "epoch": 5.853754940711463, "grad_norm": 3.403700113296509, "learning_rate": 3.543564356435644e-05, "loss": 0.1857, "step": 1481 }, { "epoch": 5.857707509881423, "grad_norm": 2.9198403358459473, "learning_rate": 3.542574257425743e-05, "loss": 0.1861, "step": 1482 }, { "epoch": 5.861660079051384, "grad_norm": 3.3985471725463867, "learning_rate": 3.541584158415841e-05, "loss": 0.175, "step": 1483 }, { "epoch": 5.865612648221344, "grad_norm": 2.855233907699585, "learning_rate": 3.5405940594059404e-05, "loss": 0.1815, "step": 1484 }, { "epoch": 5.869565217391305, "grad_norm": 2.2637157440185547, "learning_rate": 3.5396039603960395e-05, "loss": 0.1474, "step": 1485 }, { "epoch": 5.873517786561265, "grad_norm": 2.23648738861084, "learning_rate": 3.5386138613861386e-05, "loss": 0.0971, "step": 1486 }, { "epoch": 5.877470355731226, "grad_norm": 3.580448627471924, "learning_rate": 3.537623762376238e-05, "loss": 0.1908, "step": 1487 }, { "epoch": 5.881422924901186, "grad_norm": 2.678616523742676, "learning_rate": 3.536633663366337e-05, "loss": 0.1606, "step": 1488 }, { "epoch": 5.8853754940711465, "grad_norm": 2.534902811050415, "learning_rate": 3.535643564356436e-05, "loss": 0.1259, "step": 1489 }, { "epoch": 5.8893280632411065, "grad_norm": 3.0793991088867188, "learning_rate": 3.5346534653465344e-05, "loss": 0.2324, "step": 1490 }, { "epoch": 5.893280632411067, "grad_norm": 3.5079846382141113, "learning_rate": 3.5336633663366335e-05, "loss": 0.1949, "step": 1491 }, { "epoch": 5.897233201581027, "grad_norm": 2.8772213459014893, "learning_rate": 3.5326732673267326e-05, "loss": 0.1148, "step": 1492 }, { "epoch": 5.901185770750988, "grad_norm": 3.403609275817871, "learning_rate": 3.531683168316832e-05, "loss": 0.2782, "step": 1493 }, { "epoch": 5.905138339920948, "grad_norm": 2.9689254760742188, "learning_rate": 3.530693069306931e-05, "loss": 0.1856, "step": 1494 }, { "epoch": 5.909090909090909, "grad_norm": 2.576106309890747, "learning_rate": 3.52970297029703e-05, "loss": 0.1446, "step": 1495 }, { "epoch": 5.913043478260869, "grad_norm": 3.5456080436706543, "learning_rate": 3.528712871287129e-05, "loss": 0.225, "step": 1496 }, { "epoch": 5.91699604743083, "grad_norm": 3.1134891510009766, "learning_rate": 3.5277227722772274e-05, "loss": 0.184, "step": 1497 }, { "epoch": 5.920948616600791, "grad_norm": 3.199864625930786, "learning_rate": 3.5267326732673265e-05, "loss": 0.1893, "step": 1498 }, { "epoch": 5.924901185770751, "grad_norm": 3.165670394897461, "learning_rate": 3.5257425742574256e-05, "loss": 0.2042, "step": 1499 }, { "epoch": 5.928853754940711, "grad_norm": 2.5231733322143555, "learning_rate": 3.524752475247525e-05, "loss": 0.1348, "step": 1500 }, { "epoch": 5.932806324110672, "grad_norm": 2.532444477081299, "learning_rate": 3.523762376237624e-05, "loss": 0.1758, "step": 1501 }, { "epoch": 5.936758893280633, "grad_norm": 3.249006748199463, "learning_rate": 3.522772277227723e-05, "loss": 0.1738, "step": 1502 }, { "epoch": 5.940711462450593, "grad_norm": 3.5972561836242676, "learning_rate": 3.521782178217822e-05, "loss": 0.1836, "step": 1503 }, { "epoch": 5.944664031620554, "grad_norm": 2.652209520339966, "learning_rate": 3.5207920792079205e-05, "loss": 0.1696, "step": 1504 }, { "epoch": 5.948616600790514, "grad_norm": 2.995123863220215, "learning_rate": 3.5198019801980196e-05, "loss": 0.1924, "step": 1505 }, { "epoch": 5.952569169960475, "grad_norm": 2.9341468811035156, "learning_rate": 3.518811881188119e-05, "loss": 0.1749, "step": 1506 }, { "epoch": 5.956521739130435, "grad_norm": 2.9498095512390137, "learning_rate": 3.517821782178218e-05, "loss": 0.1734, "step": 1507 }, { "epoch": 5.9604743083003955, "grad_norm": 3.252932071685791, "learning_rate": 3.516831683168317e-05, "loss": 0.2015, "step": 1508 }, { "epoch": 5.9644268774703555, "grad_norm": 3.53458571434021, "learning_rate": 3.515841584158416e-05, "loss": 0.2507, "step": 1509 }, { "epoch": 5.968379446640316, "grad_norm": 2.6905107498168945, "learning_rate": 3.514851485148515e-05, "loss": 0.1565, "step": 1510 }, { "epoch": 5.972332015810276, "grad_norm": 3.8605406284332275, "learning_rate": 3.5138613861386136e-05, "loss": 0.2295, "step": 1511 }, { "epoch": 5.976284584980237, "grad_norm": 2.7002289295196533, "learning_rate": 3.512871287128713e-05, "loss": 0.1837, "step": 1512 }, { "epoch": 5.980237154150197, "grad_norm": 4.297340393066406, "learning_rate": 3.511881188118812e-05, "loss": 0.2617, "step": 1513 }, { "epoch": 5.984189723320158, "grad_norm": 2.8572418689727783, "learning_rate": 3.510891089108911e-05, "loss": 0.1538, "step": 1514 }, { "epoch": 5.988142292490118, "grad_norm": 2.819495677947998, "learning_rate": 3.50990099009901e-05, "loss": 0.1689, "step": 1515 }, { "epoch": 5.992094861660079, "grad_norm": 2.732334613800049, "learning_rate": 3.508910891089109e-05, "loss": 0.1849, "step": 1516 }, { "epoch": 5.996047430830039, "grad_norm": 2.6902246475219727, "learning_rate": 3.507920792079208e-05, "loss": 0.1694, "step": 1517 }, { "epoch": 6.0, "grad_norm": 2.8910892009735107, "learning_rate": 3.5069306930693066e-05, "loss": 0.168, "step": 1518 }, { "epoch": 6.003952569169961, "grad_norm": 1.6254279613494873, "learning_rate": 3.505940594059406e-05, "loss": 0.0834, "step": 1519 }, { "epoch": 6.007905138339921, "grad_norm": 2.5636308193206787, "learning_rate": 3.504950495049505e-05, "loss": 0.0921, "step": 1520 }, { "epoch": 6.011857707509882, "grad_norm": 2.822171449661255, "learning_rate": 3.503960396039604e-05, "loss": 0.1528, "step": 1521 }, { "epoch": 6.015810276679842, "grad_norm": 2.1642751693725586, "learning_rate": 3.502970297029703e-05, "loss": 0.1027, "step": 1522 }, { "epoch": 6.019762845849803, "grad_norm": 1.7622253894805908, "learning_rate": 3.501980198019802e-05, "loss": 0.078, "step": 1523 }, { "epoch": 6.023715415019763, "grad_norm": 2.4607322216033936, "learning_rate": 3.500990099009901e-05, "loss": 0.0989, "step": 1524 }, { "epoch": 6.027667984189724, "grad_norm": 2.216251850128174, "learning_rate": 3.5e-05, "loss": 0.0799, "step": 1525 }, { "epoch": 6.031620553359684, "grad_norm": 2.016751766204834, "learning_rate": 3.499009900990099e-05, "loss": 0.0983, "step": 1526 }, { "epoch": 6.0355731225296445, "grad_norm": 2.0572965145111084, "learning_rate": 3.498019801980198e-05, "loss": 0.0752, "step": 1527 }, { "epoch": 6.0395256916996045, "grad_norm": 2.622939109802246, "learning_rate": 3.497029702970297e-05, "loss": 0.1035, "step": 1528 }, { "epoch": 6.043478260869565, "grad_norm": 2.342886209487915, "learning_rate": 3.496039603960396e-05, "loss": 0.0945, "step": 1529 }, { "epoch": 6.047430830039525, "grad_norm": 2.695631265640259, "learning_rate": 3.495049504950495e-05, "loss": 0.1079, "step": 1530 }, { "epoch": 6.051383399209486, "grad_norm": 2.1968178749084473, "learning_rate": 3.494059405940594e-05, "loss": 0.0971, "step": 1531 }, { "epoch": 6.055335968379446, "grad_norm": 1.9653151035308838, "learning_rate": 3.493069306930693e-05, "loss": 0.0749, "step": 1532 }, { "epoch": 6.059288537549407, "grad_norm": 3.012938976287842, "learning_rate": 3.492079207920792e-05, "loss": 0.1268, "step": 1533 }, { "epoch": 6.063241106719367, "grad_norm": 2.6731314659118652, "learning_rate": 3.491089108910891e-05, "loss": 0.1261, "step": 1534 }, { "epoch": 6.067193675889328, "grad_norm": 2.999439001083374, "learning_rate": 3.49009900990099e-05, "loss": 0.1255, "step": 1535 }, { "epoch": 6.071146245059288, "grad_norm": 2.3211820125579834, "learning_rate": 3.489108910891089e-05, "loss": 0.1297, "step": 1536 }, { "epoch": 6.075098814229249, "grad_norm": 1.980760097503662, "learning_rate": 3.488118811881188e-05, "loss": 0.086, "step": 1537 }, { "epoch": 6.07905138339921, "grad_norm": 3.4899535179138184, "learning_rate": 3.487128712871287e-05, "loss": 0.1375, "step": 1538 }, { "epoch": 6.08300395256917, "grad_norm": 3.1304643154144287, "learning_rate": 3.486138613861386e-05, "loss": 0.1024, "step": 1539 }, { "epoch": 6.086956521739131, "grad_norm": 2.2567250728607178, "learning_rate": 3.485148514851485e-05, "loss": 0.0743, "step": 1540 }, { "epoch": 6.090909090909091, "grad_norm": 2.572547197341919, "learning_rate": 3.484158415841584e-05, "loss": 0.1054, "step": 1541 }, { "epoch": 6.094861660079052, "grad_norm": 1.8252859115600586, "learning_rate": 3.483168316831683e-05, "loss": 0.0848, "step": 1542 }, { "epoch": 6.098814229249012, "grad_norm": 1.675033688545227, "learning_rate": 3.482178217821782e-05, "loss": 0.0799, "step": 1543 }, { "epoch": 6.102766798418973, "grad_norm": 2.18462872505188, "learning_rate": 3.4811881188118814e-05, "loss": 0.0978, "step": 1544 }, { "epoch": 6.106719367588933, "grad_norm": 2.493809461593628, "learning_rate": 3.48019801980198e-05, "loss": 0.1004, "step": 1545 }, { "epoch": 6.1106719367588935, "grad_norm": 2.9898531436920166, "learning_rate": 3.479207920792079e-05, "loss": 0.1208, "step": 1546 }, { "epoch": 6.1146245059288535, "grad_norm": 2.414863348007202, "learning_rate": 3.478217821782178e-05, "loss": 0.0975, "step": 1547 }, { "epoch": 6.118577075098814, "grad_norm": 2.8248300552368164, "learning_rate": 3.477227722772277e-05, "loss": 0.1006, "step": 1548 }, { "epoch": 6.122529644268774, "grad_norm": 3.01359224319458, "learning_rate": 3.476237623762376e-05, "loss": 0.1284, "step": 1549 }, { "epoch": 6.126482213438735, "grad_norm": 2.620307445526123, "learning_rate": 3.475247524752475e-05, "loss": 0.1033, "step": 1550 }, { "epoch": 6.130434782608695, "grad_norm": 2.1864633560180664, "learning_rate": 3.4742574257425744e-05, "loss": 0.1094, "step": 1551 }, { "epoch": 6.134387351778656, "grad_norm": 2.6389899253845215, "learning_rate": 3.473267326732673e-05, "loss": 0.1011, "step": 1552 }, { "epoch": 6.138339920948616, "grad_norm": 2.0675501823425293, "learning_rate": 3.472277227722772e-05, "loss": 0.0782, "step": 1553 }, { "epoch": 6.142292490118577, "grad_norm": 2.3925721645355225, "learning_rate": 3.471287128712871e-05, "loss": 0.0892, "step": 1554 }, { "epoch": 6.146245059288537, "grad_norm": 2.299866199493408, "learning_rate": 3.47029702970297e-05, "loss": 0.0931, "step": 1555 }, { "epoch": 6.150197628458498, "grad_norm": 2.4922142028808594, "learning_rate": 3.469306930693069e-05, "loss": 0.0884, "step": 1556 }, { "epoch": 6.154150197628459, "grad_norm": 2.30753231048584, "learning_rate": 3.4683168316831684e-05, "loss": 0.1068, "step": 1557 }, { "epoch": 6.158102766798419, "grad_norm": 3.6126813888549805, "learning_rate": 3.4673267326732675e-05, "loss": 0.1343, "step": 1558 }, { "epoch": 6.16205533596838, "grad_norm": 2.8587729930877686, "learning_rate": 3.466336633663366e-05, "loss": 0.1087, "step": 1559 }, { "epoch": 6.16600790513834, "grad_norm": 2.120866537094116, "learning_rate": 3.465346534653465e-05, "loss": 0.0873, "step": 1560 }, { "epoch": 6.169960474308301, "grad_norm": 2.901534080505371, "learning_rate": 3.464356435643564e-05, "loss": 0.1162, "step": 1561 }, { "epoch": 6.173913043478261, "grad_norm": 4.595358371734619, "learning_rate": 3.463366336633663e-05, "loss": 0.1127, "step": 1562 }, { "epoch": 6.177865612648222, "grad_norm": 2.7407712936401367, "learning_rate": 3.4623762376237624e-05, "loss": 0.1136, "step": 1563 }, { "epoch": 6.181818181818182, "grad_norm": 2.5797765254974365, "learning_rate": 3.4613861386138615e-05, "loss": 0.1132, "step": 1564 }, { "epoch": 6.1857707509881426, "grad_norm": 2.3501198291778564, "learning_rate": 3.4603960396039606e-05, "loss": 0.1079, "step": 1565 }, { "epoch": 6.189723320158103, "grad_norm": 2.5906755924224854, "learning_rate": 3.459405940594059e-05, "loss": 0.0967, "step": 1566 }, { "epoch": 6.1936758893280635, "grad_norm": 2.1571664810180664, "learning_rate": 3.458415841584158e-05, "loss": 0.1027, "step": 1567 }, { "epoch": 6.1976284584980235, "grad_norm": 2.892108201980591, "learning_rate": 3.457425742574257e-05, "loss": 0.1155, "step": 1568 }, { "epoch": 6.201581027667984, "grad_norm": 2.0727620124816895, "learning_rate": 3.456435643564356e-05, "loss": 0.1086, "step": 1569 }, { "epoch": 6.205533596837944, "grad_norm": 2.2134854793548584, "learning_rate": 3.4554455445544554e-05, "loss": 0.1233, "step": 1570 }, { "epoch": 6.209486166007905, "grad_norm": 2.7549948692321777, "learning_rate": 3.4544554455445545e-05, "loss": 0.0964, "step": 1571 }, { "epoch": 6.213438735177865, "grad_norm": 2.1496188640594482, "learning_rate": 3.4534653465346536e-05, "loss": 0.0784, "step": 1572 }, { "epoch": 6.217391304347826, "grad_norm": 2.2960667610168457, "learning_rate": 3.452475247524752e-05, "loss": 0.1125, "step": 1573 }, { "epoch": 6.221343873517786, "grad_norm": 2.2926414012908936, "learning_rate": 3.451485148514851e-05, "loss": 0.0967, "step": 1574 }, { "epoch": 6.225296442687747, "grad_norm": 1.9400876760482788, "learning_rate": 3.45049504950495e-05, "loss": 0.0886, "step": 1575 }, { "epoch": 6.229249011857707, "grad_norm": 2.8527143001556396, "learning_rate": 3.4495049504950494e-05, "loss": 0.0989, "step": 1576 }, { "epoch": 6.233201581027668, "grad_norm": 2.9145009517669678, "learning_rate": 3.4485148514851485e-05, "loss": 0.1221, "step": 1577 }, { "epoch": 6.237154150197629, "grad_norm": 3.128875970840454, "learning_rate": 3.4475247524752476e-05, "loss": 0.1315, "step": 1578 }, { "epoch": 6.241106719367589, "grad_norm": 2.1546571254730225, "learning_rate": 3.446534653465347e-05, "loss": 0.0861, "step": 1579 }, { "epoch": 6.24505928853755, "grad_norm": 2.939331531524658, "learning_rate": 3.445544554455445e-05, "loss": 0.1238, "step": 1580 }, { "epoch": 6.24901185770751, "grad_norm": 2.978647470474243, "learning_rate": 3.444554455445544e-05, "loss": 0.1397, "step": 1581 }, { "epoch": 6.252964426877471, "grad_norm": 2.1712610721588135, "learning_rate": 3.4435643564356433e-05, "loss": 0.1142, "step": 1582 }, { "epoch": 6.256916996047431, "grad_norm": 3.2149412631988525, "learning_rate": 3.4425742574257425e-05, "loss": 0.1241, "step": 1583 }, { "epoch": 6.260869565217392, "grad_norm": 3.148458480834961, "learning_rate": 3.4415841584158416e-05, "loss": 0.1223, "step": 1584 }, { "epoch": 6.264822134387352, "grad_norm": 2.366495370864868, "learning_rate": 3.440594059405941e-05, "loss": 0.0979, "step": 1585 }, { "epoch": 6.2687747035573125, "grad_norm": 2.655197858810425, "learning_rate": 3.43960396039604e-05, "loss": 0.1116, "step": 1586 }, { "epoch": 6.2727272727272725, "grad_norm": 2.574202537536621, "learning_rate": 3.438613861386138e-05, "loss": 0.1067, "step": 1587 }, { "epoch": 6.276679841897233, "grad_norm": 2.6245052814483643, "learning_rate": 3.437623762376237e-05, "loss": 0.1068, "step": 1588 }, { "epoch": 6.280632411067193, "grad_norm": 1.791400671005249, "learning_rate": 3.436633663366337e-05, "loss": 0.0851, "step": 1589 }, { "epoch": 6.284584980237154, "grad_norm": 1.9702666997909546, "learning_rate": 3.435643564356436e-05, "loss": 0.0988, "step": 1590 }, { "epoch": 6.288537549407114, "grad_norm": 2.76668381690979, "learning_rate": 3.434653465346535e-05, "loss": 0.1051, "step": 1591 }, { "epoch": 6.292490118577075, "grad_norm": 2.382469892501831, "learning_rate": 3.4336633663366344e-05, "loss": 0.101, "step": 1592 }, { "epoch": 6.296442687747035, "grad_norm": 3.672960042953491, "learning_rate": 3.432673267326733e-05, "loss": 0.1885, "step": 1593 }, { "epoch": 6.300395256916996, "grad_norm": 2.379645347595215, "learning_rate": 3.431683168316832e-05, "loss": 0.1046, "step": 1594 }, { "epoch": 6.304347826086957, "grad_norm": 2.7646734714508057, "learning_rate": 3.430693069306931e-05, "loss": 0.095, "step": 1595 }, { "epoch": 6.308300395256917, "grad_norm": 2.2863471508026123, "learning_rate": 3.42970297029703e-05, "loss": 0.1121, "step": 1596 }, { "epoch": 6.312252964426877, "grad_norm": 4.073939800262451, "learning_rate": 3.428712871287129e-05, "loss": 0.133, "step": 1597 }, { "epoch": 6.316205533596838, "grad_norm": 3.0593435764312744, "learning_rate": 3.4277227722772284e-05, "loss": 0.1153, "step": 1598 }, { "epoch": 6.320158102766799, "grad_norm": 3.0961148738861084, "learning_rate": 3.4267326732673275e-05, "loss": 0.1355, "step": 1599 }, { "epoch": 6.324110671936759, "grad_norm": 2.7696783542633057, "learning_rate": 3.425742574257426e-05, "loss": 0.1139, "step": 1600 }, { "epoch": 6.32806324110672, "grad_norm": 2.125610828399658, "learning_rate": 3.424752475247525e-05, "loss": 0.107, "step": 1601 }, { "epoch": 6.33201581027668, "grad_norm": 2.425043821334839, "learning_rate": 3.423762376237624e-05, "loss": 0.1068, "step": 1602 }, { "epoch": 6.335968379446641, "grad_norm": 2.0851330757141113, "learning_rate": 3.422772277227723e-05, "loss": 0.0923, "step": 1603 }, { "epoch": 6.339920948616601, "grad_norm": 2.608879327774048, "learning_rate": 3.421782178217822e-05, "loss": 0.1266, "step": 1604 }, { "epoch": 6.3438735177865615, "grad_norm": 2.51102352142334, "learning_rate": 3.4207920792079214e-05, "loss": 0.1157, "step": 1605 }, { "epoch": 6.3478260869565215, "grad_norm": 2.527832269668579, "learning_rate": 3.4198019801980206e-05, "loss": 0.1144, "step": 1606 }, { "epoch": 6.351778656126482, "grad_norm": 3.133955478668213, "learning_rate": 3.418811881188119e-05, "loss": 0.1268, "step": 1607 }, { "epoch": 6.355731225296442, "grad_norm": 2.4142909049987793, "learning_rate": 3.417821782178218e-05, "loss": 0.1074, "step": 1608 }, { "epoch": 6.359683794466403, "grad_norm": 2.29095196723938, "learning_rate": 3.416831683168317e-05, "loss": 0.1111, "step": 1609 }, { "epoch": 6.363636363636363, "grad_norm": 2.282353162765503, "learning_rate": 3.415841584158416e-05, "loss": 0.0908, "step": 1610 }, { "epoch": 6.367588932806324, "grad_norm": 2.4689390659332275, "learning_rate": 3.4148514851485154e-05, "loss": 0.0835, "step": 1611 }, { "epoch": 6.371541501976284, "grad_norm": 1.8772506713867188, "learning_rate": 3.4138613861386145e-05, "loss": 0.0878, "step": 1612 }, { "epoch": 6.375494071146245, "grad_norm": 2.1557672023773193, "learning_rate": 3.4128712871287136e-05, "loss": 0.1107, "step": 1613 }, { "epoch": 6.379446640316205, "grad_norm": 2.402582883834839, "learning_rate": 3.411881188118812e-05, "loss": 0.096, "step": 1614 }, { "epoch": 6.383399209486166, "grad_norm": 2.556732177734375, "learning_rate": 3.410891089108911e-05, "loss": 0.0911, "step": 1615 }, { "epoch": 6.387351778656127, "grad_norm": 2.0217947959899902, "learning_rate": 3.40990099009901e-05, "loss": 0.0826, "step": 1616 }, { "epoch": 6.391304347826087, "grad_norm": 2.6689555644989014, "learning_rate": 3.4089108910891094e-05, "loss": 0.1319, "step": 1617 }, { "epoch": 6.395256916996048, "grad_norm": 2.336923599243164, "learning_rate": 3.4079207920792085e-05, "loss": 0.1038, "step": 1618 }, { "epoch": 6.399209486166008, "grad_norm": 2.7645740509033203, "learning_rate": 3.4069306930693076e-05, "loss": 0.1204, "step": 1619 }, { "epoch": 6.403162055335969, "grad_norm": 2.8250434398651123, "learning_rate": 3.405940594059407e-05, "loss": 0.138, "step": 1620 }, { "epoch": 6.407114624505929, "grad_norm": 3.0159430503845215, "learning_rate": 3.404950495049505e-05, "loss": 0.1257, "step": 1621 }, { "epoch": 6.41106719367589, "grad_norm": 2.9421117305755615, "learning_rate": 3.403960396039604e-05, "loss": 0.1172, "step": 1622 }, { "epoch": 6.41501976284585, "grad_norm": 2.8791861534118652, "learning_rate": 3.402970297029703e-05, "loss": 0.1065, "step": 1623 }, { "epoch": 6.4189723320158105, "grad_norm": 2.3747315406799316, "learning_rate": 3.4019801980198024e-05, "loss": 0.0831, "step": 1624 }, { "epoch": 6.4229249011857705, "grad_norm": 3.15038800239563, "learning_rate": 3.4009900990099015e-05, "loss": 0.147, "step": 1625 }, { "epoch": 6.426877470355731, "grad_norm": 2.734018087387085, "learning_rate": 3.4000000000000007e-05, "loss": 0.1181, "step": 1626 }, { "epoch": 6.430830039525691, "grad_norm": 2.4690792560577393, "learning_rate": 3.399009900990099e-05, "loss": 0.0999, "step": 1627 }, { "epoch": 6.434782608695652, "grad_norm": 2.2538809776306152, "learning_rate": 3.398019801980198e-05, "loss": 0.113, "step": 1628 }, { "epoch": 6.438735177865612, "grad_norm": 2.7664871215820312, "learning_rate": 3.397029702970297e-05, "loss": 0.1236, "step": 1629 }, { "epoch": 6.442687747035573, "grad_norm": 2.957289934158325, "learning_rate": 3.3960396039603964e-05, "loss": 0.1383, "step": 1630 }, { "epoch": 6.446640316205533, "grad_norm": 2.5278539657592773, "learning_rate": 3.3950495049504955e-05, "loss": 0.123, "step": 1631 }, { "epoch": 6.450592885375494, "grad_norm": 3.2642111778259277, "learning_rate": 3.3940594059405946e-05, "loss": 0.1258, "step": 1632 }, { "epoch": 6.454545454545454, "grad_norm": 2.727259397506714, "learning_rate": 3.393069306930694e-05, "loss": 0.1118, "step": 1633 }, { "epoch": 6.458498023715415, "grad_norm": 2.410440683364868, "learning_rate": 3.392079207920792e-05, "loss": 0.11, "step": 1634 }, { "epoch": 6.462450592885375, "grad_norm": 3.1247360706329346, "learning_rate": 3.391089108910891e-05, "loss": 0.1474, "step": 1635 }, { "epoch": 6.466403162055336, "grad_norm": 2.0408570766448975, "learning_rate": 3.3900990099009904e-05, "loss": 0.1082, "step": 1636 }, { "epoch": 6.470355731225297, "grad_norm": 2.232130765914917, "learning_rate": 3.3891089108910895e-05, "loss": 0.102, "step": 1637 }, { "epoch": 6.474308300395257, "grad_norm": 2.211778163909912, "learning_rate": 3.3881188118811886e-05, "loss": 0.1141, "step": 1638 }, { "epoch": 6.478260869565218, "grad_norm": 2.6561810970306396, "learning_rate": 3.387128712871288e-05, "loss": 0.1293, "step": 1639 }, { "epoch": 6.482213438735178, "grad_norm": 2.591306686401367, "learning_rate": 3.386138613861387e-05, "loss": 0.1134, "step": 1640 }, { "epoch": 6.486166007905139, "grad_norm": 1.7929493188858032, "learning_rate": 3.385148514851485e-05, "loss": 0.0914, "step": 1641 }, { "epoch": 6.490118577075099, "grad_norm": 2.537294864654541, "learning_rate": 3.384158415841584e-05, "loss": 0.1203, "step": 1642 }, { "epoch": 6.4940711462450595, "grad_norm": 2.432285785675049, "learning_rate": 3.3831683168316834e-05, "loss": 0.0904, "step": 1643 }, { "epoch": 6.4980237154150196, "grad_norm": 3.1463253498077393, "learning_rate": 3.3821782178217825e-05, "loss": 0.1396, "step": 1644 }, { "epoch": 6.5019762845849804, "grad_norm": 2.973484754562378, "learning_rate": 3.3811881188118816e-05, "loss": 0.1224, "step": 1645 }, { "epoch": 6.5059288537549405, "grad_norm": 2.889249324798584, "learning_rate": 3.380198019801981e-05, "loss": 0.1173, "step": 1646 }, { "epoch": 6.509881422924901, "grad_norm": 2.27713680267334, "learning_rate": 3.37920792079208e-05, "loss": 0.1004, "step": 1647 }, { "epoch": 6.513833992094861, "grad_norm": 2.6936967372894287, "learning_rate": 3.378217821782178e-05, "loss": 0.1051, "step": 1648 }, { "epoch": 6.517786561264822, "grad_norm": 2.417397975921631, "learning_rate": 3.3772277227722774e-05, "loss": 0.1071, "step": 1649 }, { "epoch": 6.521739130434782, "grad_norm": 2.268353223800659, "learning_rate": 3.3762376237623765e-05, "loss": 0.1115, "step": 1650 }, { "epoch": 6.525691699604743, "grad_norm": 2.6652984619140625, "learning_rate": 3.3752475247524756e-05, "loss": 0.103, "step": 1651 }, { "epoch": 6.529644268774703, "grad_norm": 2.631671667098999, "learning_rate": 3.374257425742575e-05, "loss": 0.1247, "step": 1652 }, { "epoch": 6.533596837944664, "grad_norm": 2.637877941131592, "learning_rate": 3.373267326732674e-05, "loss": 0.1249, "step": 1653 }, { "epoch": 6.537549407114625, "grad_norm": 3.064127206802368, "learning_rate": 3.372277227722773e-05, "loss": 0.1171, "step": 1654 }, { "epoch": 6.541501976284585, "grad_norm": 2.760484457015991, "learning_rate": 3.3712871287128714e-05, "loss": 0.1216, "step": 1655 }, { "epoch": 6.545454545454545, "grad_norm": 2.2011611461639404, "learning_rate": 3.3702970297029705e-05, "loss": 0.0974, "step": 1656 }, { "epoch": 6.549407114624506, "grad_norm": 2.6852211952209473, "learning_rate": 3.3693069306930696e-05, "loss": 0.1239, "step": 1657 }, { "epoch": 6.553359683794467, "grad_norm": 3.0692222118377686, "learning_rate": 3.368316831683169e-05, "loss": 0.1264, "step": 1658 }, { "epoch": 6.557312252964427, "grad_norm": 2.7208423614501953, "learning_rate": 3.367326732673268e-05, "loss": 0.1368, "step": 1659 }, { "epoch": 6.561264822134388, "grad_norm": 2.864258289337158, "learning_rate": 3.366336633663367e-05, "loss": 0.115, "step": 1660 }, { "epoch": 6.565217391304348, "grad_norm": 2.5937209129333496, "learning_rate": 3.365346534653466e-05, "loss": 0.11, "step": 1661 }, { "epoch": 6.569169960474309, "grad_norm": 2.5636918544769287, "learning_rate": 3.3643564356435644e-05, "loss": 0.1098, "step": 1662 }, { "epoch": 6.573122529644269, "grad_norm": 3.7972843647003174, "learning_rate": 3.3633663366336635e-05, "loss": 0.1773, "step": 1663 }, { "epoch": 6.5770750988142295, "grad_norm": 2.0668087005615234, "learning_rate": 3.3623762376237626e-05, "loss": 0.0977, "step": 1664 }, { "epoch": 6.5810276679841895, "grad_norm": 3.2972183227539062, "learning_rate": 3.361386138613862e-05, "loss": 0.1292, "step": 1665 }, { "epoch": 6.58498023715415, "grad_norm": 2.0768086910247803, "learning_rate": 3.360396039603961e-05, "loss": 0.0927, "step": 1666 }, { "epoch": 6.58893280632411, "grad_norm": 2.2571561336517334, "learning_rate": 3.35940594059406e-05, "loss": 0.0962, "step": 1667 }, { "epoch": 6.592885375494071, "grad_norm": 2.4021155834198, "learning_rate": 3.358415841584159e-05, "loss": 0.116, "step": 1668 }, { "epoch": 6.596837944664031, "grad_norm": 2.67392897605896, "learning_rate": 3.3574257425742575e-05, "loss": 0.1294, "step": 1669 }, { "epoch": 6.600790513833992, "grad_norm": 2.8035335540771484, "learning_rate": 3.3564356435643566e-05, "loss": 0.1254, "step": 1670 }, { "epoch": 6.604743083003952, "grad_norm": 2.0258965492248535, "learning_rate": 3.355445544554456e-05, "loss": 0.1088, "step": 1671 }, { "epoch": 6.608695652173913, "grad_norm": 2.70961594581604, "learning_rate": 3.354455445544555e-05, "loss": 0.1206, "step": 1672 }, { "epoch": 6.612648221343873, "grad_norm": 2.441798448562622, "learning_rate": 3.353465346534654e-05, "loss": 0.1123, "step": 1673 }, { "epoch": 6.616600790513834, "grad_norm": 1.9946775436401367, "learning_rate": 3.352475247524753e-05, "loss": 0.1034, "step": 1674 }, { "epoch": 6.620553359683795, "grad_norm": 2.610639810562134, "learning_rate": 3.351485148514852e-05, "loss": 0.1175, "step": 1675 }, { "epoch": 6.624505928853755, "grad_norm": 1.9242851734161377, "learning_rate": 3.3504950495049506e-05, "loss": 0.0816, "step": 1676 }, { "epoch": 6.628458498023716, "grad_norm": 3.6842591762542725, "learning_rate": 3.34950495049505e-05, "loss": 0.1378, "step": 1677 }, { "epoch": 6.632411067193676, "grad_norm": 3.208308458328247, "learning_rate": 3.348514851485149e-05, "loss": 0.113, "step": 1678 }, { "epoch": 6.636363636363637, "grad_norm": 2.657059907913208, "learning_rate": 3.347524752475248e-05, "loss": 0.1024, "step": 1679 }, { "epoch": 6.640316205533597, "grad_norm": 2.755056142807007, "learning_rate": 3.346534653465347e-05, "loss": 0.109, "step": 1680 }, { "epoch": 6.644268774703558, "grad_norm": 2.546976089477539, "learning_rate": 3.345544554455446e-05, "loss": 0.1211, "step": 1681 }, { "epoch": 6.648221343873518, "grad_norm": 2.356703281402588, "learning_rate": 3.344554455445545e-05, "loss": 0.1115, "step": 1682 }, { "epoch": 6.6521739130434785, "grad_norm": 2.3479928970336914, "learning_rate": 3.3435643564356436e-05, "loss": 0.0924, "step": 1683 }, { "epoch": 6.6561264822134385, "grad_norm": 2.0064361095428467, "learning_rate": 3.342574257425743e-05, "loss": 0.0969, "step": 1684 }, { "epoch": 6.660079051383399, "grad_norm": 2.345896005630493, "learning_rate": 3.341584158415842e-05, "loss": 0.1006, "step": 1685 }, { "epoch": 6.664031620553359, "grad_norm": 2.5013794898986816, "learning_rate": 3.340594059405941e-05, "loss": 0.1156, "step": 1686 }, { "epoch": 6.66798418972332, "grad_norm": 2.683227300643921, "learning_rate": 3.33960396039604e-05, "loss": 0.1359, "step": 1687 }, { "epoch": 6.67193675889328, "grad_norm": 2.4562432765960693, "learning_rate": 3.338613861386139e-05, "loss": 0.1045, "step": 1688 }, { "epoch": 6.675889328063241, "grad_norm": 2.399925947189331, "learning_rate": 3.337623762376238e-05, "loss": 0.1072, "step": 1689 }, { "epoch": 6.679841897233201, "grad_norm": 2.8622021675109863, "learning_rate": 3.336633663366337e-05, "loss": 0.1345, "step": 1690 }, { "epoch": 6.683794466403162, "grad_norm": 2.395766258239746, "learning_rate": 3.335643564356436e-05, "loss": 0.1231, "step": 1691 }, { "epoch": 6.687747035573123, "grad_norm": 1.9716548919677734, "learning_rate": 3.334653465346535e-05, "loss": 0.0879, "step": 1692 }, { "epoch": 6.691699604743083, "grad_norm": 2.2412006855010986, "learning_rate": 3.333663366336634e-05, "loss": 0.1129, "step": 1693 }, { "epoch": 6.695652173913043, "grad_norm": 2.5873920917510986, "learning_rate": 3.332673267326733e-05, "loss": 0.116, "step": 1694 }, { "epoch": 6.699604743083004, "grad_norm": 2.6831729412078857, "learning_rate": 3.331683168316832e-05, "loss": 0.1196, "step": 1695 }, { "epoch": 6.703557312252965, "grad_norm": 2.2726802825927734, "learning_rate": 3.330693069306931e-05, "loss": 0.1357, "step": 1696 }, { "epoch": 6.707509881422925, "grad_norm": 2.7359910011291504, "learning_rate": 3.32970297029703e-05, "loss": 0.1426, "step": 1697 }, { "epoch": 6.711462450592886, "grad_norm": 2.9230923652648926, "learning_rate": 3.328712871287129e-05, "loss": 0.1272, "step": 1698 }, { "epoch": 6.715415019762846, "grad_norm": 2.454810857772827, "learning_rate": 3.327722772277228e-05, "loss": 0.1169, "step": 1699 }, { "epoch": 6.719367588932807, "grad_norm": 2.1496260166168213, "learning_rate": 3.326732673267327e-05, "loss": 0.1071, "step": 1700 }, { "epoch": 6.723320158102767, "grad_norm": 3.3024604320526123, "learning_rate": 3.325742574257426e-05, "loss": 0.1164, "step": 1701 }, { "epoch": 6.7272727272727275, "grad_norm": 2.7445926666259766, "learning_rate": 3.324752475247525e-05, "loss": 0.1023, "step": 1702 }, { "epoch": 6.7312252964426875, "grad_norm": 2.2595391273498535, "learning_rate": 3.3237623762376244e-05, "loss": 0.0882, "step": 1703 }, { "epoch": 6.735177865612648, "grad_norm": 2.9820079803466797, "learning_rate": 3.322772277227723e-05, "loss": 0.1106, "step": 1704 }, { "epoch": 6.739130434782608, "grad_norm": 2.316274881362915, "learning_rate": 3.321782178217822e-05, "loss": 0.1109, "step": 1705 }, { "epoch": 6.743083003952569, "grad_norm": 3.428133964538574, "learning_rate": 3.320792079207921e-05, "loss": 0.149, "step": 1706 }, { "epoch": 6.747035573122529, "grad_norm": 3.0487117767333984, "learning_rate": 3.31980198019802e-05, "loss": 0.0973, "step": 1707 }, { "epoch": 6.75098814229249, "grad_norm": 2.4415457248687744, "learning_rate": 3.318811881188119e-05, "loss": 0.1092, "step": 1708 }, { "epoch": 6.75494071146245, "grad_norm": 2.3210558891296387, "learning_rate": 3.3178217821782184e-05, "loss": 0.1118, "step": 1709 }, { "epoch": 6.758893280632411, "grad_norm": 2.9945900440216064, "learning_rate": 3.3168316831683175e-05, "loss": 0.133, "step": 1710 }, { "epoch": 6.762845849802371, "grad_norm": 3.0359463691711426, "learning_rate": 3.315841584158416e-05, "loss": 0.1075, "step": 1711 }, { "epoch": 6.766798418972332, "grad_norm": 2.585736036300659, "learning_rate": 3.314851485148515e-05, "loss": 0.1066, "step": 1712 }, { "epoch": 6.770750988142293, "grad_norm": 1.9474372863769531, "learning_rate": 3.313861386138614e-05, "loss": 0.091, "step": 1713 }, { "epoch": 6.774703557312253, "grad_norm": 3.075435161590576, "learning_rate": 3.312871287128713e-05, "loss": 0.1368, "step": 1714 }, { "epoch": 6.778656126482213, "grad_norm": 2.2801930904388428, "learning_rate": 3.311881188118812e-05, "loss": 0.1009, "step": 1715 }, { "epoch": 6.782608695652174, "grad_norm": 2.20497465133667, "learning_rate": 3.3108910891089114e-05, "loss": 0.11, "step": 1716 }, { "epoch": 6.786561264822135, "grad_norm": 2.2685208320617676, "learning_rate": 3.3099009900990105e-05, "loss": 0.1357, "step": 1717 }, { "epoch": 6.790513833992095, "grad_norm": 2.506948232650757, "learning_rate": 3.308910891089109e-05, "loss": 0.0955, "step": 1718 }, { "epoch": 6.794466403162056, "grad_norm": 2.657688617706299, "learning_rate": 3.307920792079208e-05, "loss": 0.1091, "step": 1719 }, { "epoch": 6.798418972332016, "grad_norm": 2.7187798023223877, "learning_rate": 3.306930693069307e-05, "loss": 0.113, "step": 1720 }, { "epoch": 6.8023715415019765, "grad_norm": 2.741548538208008, "learning_rate": 3.305940594059406e-05, "loss": 0.1202, "step": 1721 }, { "epoch": 6.8063241106719365, "grad_norm": 2.275174856185913, "learning_rate": 3.3049504950495054e-05, "loss": 0.1083, "step": 1722 }, { "epoch": 6.810276679841897, "grad_norm": 2.6430840492248535, "learning_rate": 3.3039603960396045e-05, "loss": 0.1171, "step": 1723 }, { "epoch": 6.8142292490118574, "grad_norm": 3.116791009902954, "learning_rate": 3.3029702970297036e-05, "loss": 0.1165, "step": 1724 }, { "epoch": 6.818181818181818, "grad_norm": 2.6044692993164062, "learning_rate": 3.301980198019802e-05, "loss": 0.1271, "step": 1725 }, { "epoch": 6.822134387351778, "grad_norm": 2.319330930709839, "learning_rate": 3.300990099009901e-05, "loss": 0.1143, "step": 1726 }, { "epoch": 6.826086956521739, "grad_norm": 2.5941109657287598, "learning_rate": 3.3e-05, "loss": 0.1155, "step": 1727 }, { "epoch": 6.830039525691699, "grad_norm": 2.8676633834838867, "learning_rate": 3.2990099009900994e-05, "loss": 0.1204, "step": 1728 }, { "epoch": 6.83399209486166, "grad_norm": 2.1284077167510986, "learning_rate": 3.2980198019801985e-05, "loss": 0.0888, "step": 1729 }, { "epoch": 6.837944664031621, "grad_norm": 2.328036308288574, "learning_rate": 3.2970297029702976e-05, "loss": 0.1209, "step": 1730 }, { "epoch": 6.841897233201581, "grad_norm": 2.587294340133667, "learning_rate": 3.296039603960397e-05, "loss": 0.1106, "step": 1731 }, { "epoch": 6.845849802371541, "grad_norm": 2.4901909828186035, "learning_rate": 3.295049504950495e-05, "loss": 0.0981, "step": 1732 }, { "epoch": 6.849802371541502, "grad_norm": 2.1550238132476807, "learning_rate": 3.294059405940594e-05, "loss": 0.0991, "step": 1733 }, { "epoch": 6.853754940711463, "grad_norm": 2.3638665676116943, "learning_rate": 3.293069306930693e-05, "loss": 0.1133, "step": 1734 }, { "epoch": 6.857707509881423, "grad_norm": 2.3816263675689697, "learning_rate": 3.2920792079207924e-05, "loss": 0.1162, "step": 1735 }, { "epoch": 6.861660079051384, "grad_norm": 2.593641996383667, "learning_rate": 3.2910891089108915e-05, "loss": 0.1046, "step": 1736 }, { "epoch": 6.865612648221344, "grad_norm": 2.4345366954803467, "learning_rate": 3.2900990099009906e-05, "loss": 0.1255, "step": 1737 }, { "epoch": 6.869565217391305, "grad_norm": 2.703321695327759, "learning_rate": 3.28910891089109e-05, "loss": 0.1226, "step": 1738 }, { "epoch": 6.873517786561265, "grad_norm": 2.7190582752227783, "learning_rate": 3.288118811881188e-05, "loss": 0.092, "step": 1739 }, { "epoch": 6.877470355731226, "grad_norm": 3.1326441764831543, "learning_rate": 3.287128712871287e-05, "loss": 0.1246, "step": 1740 }, { "epoch": 6.881422924901186, "grad_norm": 2.4389705657958984, "learning_rate": 3.2861386138613864e-05, "loss": 0.1014, "step": 1741 }, { "epoch": 6.8853754940711465, "grad_norm": 2.894204616546631, "learning_rate": 3.2851485148514855e-05, "loss": 0.1172, "step": 1742 }, { "epoch": 6.8893280632411065, "grad_norm": 1.991278886795044, "learning_rate": 3.2841584158415846e-05, "loss": 0.1093, "step": 1743 }, { "epoch": 6.893280632411067, "grad_norm": 2.2269797325134277, "learning_rate": 3.283168316831684e-05, "loss": 0.0983, "step": 1744 }, { "epoch": 6.897233201581027, "grad_norm": 2.7112834453582764, "learning_rate": 3.282178217821782e-05, "loss": 0.1176, "step": 1745 }, { "epoch": 6.901185770750988, "grad_norm": 3.0221238136291504, "learning_rate": 3.281188118811881e-05, "loss": 0.1059, "step": 1746 }, { "epoch": 6.905138339920948, "grad_norm": 2.450326681137085, "learning_rate": 3.2801980198019804e-05, "loss": 0.1162, "step": 1747 }, { "epoch": 6.909090909090909, "grad_norm": 3.3068504333496094, "learning_rate": 3.2792079207920795e-05, "loss": 0.1144, "step": 1748 }, { "epoch": 6.913043478260869, "grad_norm": 2.634577751159668, "learning_rate": 3.2782178217821786e-05, "loss": 0.1257, "step": 1749 }, { "epoch": 6.91699604743083, "grad_norm": 3.3611230850219727, "learning_rate": 3.277227722772278e-05, "loss": 0.1428, "step": 1750 }, { "epoch": 6.920948616600791, "grad_norm": 2.3791511058807373, "learning_rate": 3.276237623762377e-05, "loss": 0.1159, "step": 1751 }, { "epoch": 6.924901185770751, "grad_norm": 2.642995834350586, "learning_rate": 3.275247524752475e-05, "loss": 0.1169, "step": 1752 }, { "epoch": 6.928853754940711, "grad_norm": 1.963688611984253, "learning_rate": 3.274257425742574e-05, "loss": 0.1177, "step": 1753 }, { "epoch": 6.932806324110672, "grad_norm": 2.6772475242614746, "learning_rate": 3.2732673267326734e-05, "loss": 0.12, "step": 1754 }, { "epoch": 6.936758893280633, "grad_norm": 2.767031192779541, "learning_rate": 3.2722772277227725e-05, "loss": 0.1271, "step": 1755 }, { "epoch": 6.940711462450593, "grad_norm": 2.322265863418579, "learning_rate": 3.2712871287128716e-05, "loss": 0.131, "step": 1756 }, { "epoch": 6.944664031620554, "grad_norm": 2.6788172721862793, "learning_rate": 3.270297029702971e-05, "loss": 0.1279, "step": 1757 }, { "epoch": 6.948616600790514, "grad_norm": 2.371429443359375, "learning_rate": 3.26930693069307e-05, "loss": 0.1313, "step": 1758 }, { "epoch": 6.952569169960475, "grad_norm": 2.4885566234588623, "learning_rate": 3.268316831683168e-05, "loss": 0.0899, "step": 1759 }, { "epoch": 6.956521739130435, "grad_norm": 2.667280912399292, "learning_rate": 3.2673267326732674e-05, "loss": 0.1069, "step": 1760 }, { "epoch": 6.9604743083003955, "grad_norm": 2.7198739051818848, "learning_rate": 3.2663366336633665e-05, "loss": 0.1197, "step": 1761 }, { "epoch": 6.9644268774703555, "grad_norm": 3.055718183517456, "learning_rate": 3.2653465346534656e-05, "loss": 0.1519, "step": 1762 }, { "epoch": 6.968379446640316, "grad_norm": 2.126429557800293, "learning_rate": 3.264356435643565e-05, "loss": 0.0822, "step": 1763 }, { "epoch": 6.972332015810276, "grad_norm": 2.841500997543335, "learning_rate": 3.263366336633664e-05, "loss": 0.1156, "step": 1764 }, { "epoch": 6.976284584980237, "grad_norm": 2.5107874870300293, "learning_rate": 3.262376237623763e-05, "loss": 0.1327, "step": 1765 }, { "epoch": 6.980237154150197, "grad_norm": 2.721013307571411, "learning_rate": 3.2613861386138613e-05, "loss": 0.1207, "step": 1766 }, { "epoch": 6.984189723320158, "grad_norm": 2.9162023067474365, "learning_rate": 3.2603960396039605e-05, "loss": 0.122, "step": 1767 }, { "epoch": 6.988142292490118, "grad_norm": 2.427996873855591, "learning_rate": 3.2594059405940596e-05, "loss": 0.1211, "step": 1768 }, { "epoch": 6.992094861660079, "grad_norm": 2.732961654663086, "learning_rate": 3.258415841584159e-05, "loss": 0.1036, "step": 1769 }, { "epoch": 6.996047430830039, "grad_norm": 2.6545753479003906, "learning_rate": 3.257425742574258e-05, "loss": 0.1249, "step": 1770 }, { "epoch": 7.0, "grad_norm": 2.494842052459717, "learning_rate": 3.256435643564357e-05, "loss": 0.1303, "step": 1771 }, { "epoch": 7.003952569169961, "grad_norm": 1.361709713935852, "learning_rate": 3.255445544554456e-05, "loss": 0.0505, "step": 1772 }, { "epoch": 7.007905138339921, "grad_norm": 2.154360294342041, "learning_rate": 3.2544554455445544e-05, "loss": 0.08, "step": 1773 }, { "epoch": 7.011857707509882, "grad_norm": 1.9486173391342163, "learning_rate": 3.2534653465346535e-05, "loss": 0.0687, "step": 1774 }, { "epoch": 7.015810276679842, "grad_norm": 1.7919670343399048, "learning_rate": 3.2524752475247526e-05, "loss": 0.0919, "step": 1775 }, { "epoch": 7.019762845849803, "grad_norm": 2.030191659927368, "learning_rate": 3.251485148514852e-05, "loss": 0.0822, "step": 1776 }, { "epoch": 7.023715415019763, "grad_norm": 1.7546439170837402, "learning_rate": 3.250495049504951e-05, "loss": 0.0645, "step": 1777 }, { "epoch": 7.027667984189724, "grad_norm": 2.5918946266174316, "learning_rate": 3.24950495049505e-05, "loss": 0.0819, "step": 1778 }, { "epoch": 7.031620553359684, "grad_norm": 1.8037742376327515, "learning_rate": 3.248514851485149e-05, "loss": 0.0597, "step": 1779 }, { "epoch": 7.0355731225296445, "grad_norm": 2.4337825775146484, "learning_rate": 3.2475247524752475e-05, "loss": 0.0688, "step": 1780 }, { "epoch": 7.0395256916996045, "grad_norm": 1.6845300197601318, "learning_rate": 3.2465346534653466e-05, "loss": 0.0573, "step": 1781 }, { "epoch": 7.043478260869565, "grad_norm": 2.113631248474121, "learning_rate": 3.245544554455446e-05, "loss": 0.0785, "step": 1782 }, { "epoch": 7.047430830039525, "grad_norm": 1.749152421951294, "learning_rate": 3.244554455445545e-05, "loss": 0.0647, "step": 1783 }, { "epoch": 7.051383399209486, "grad_norm": 1.7756210565567017, "learning_rate": 3.243564356435644e-05, "loss": 0.0667, "step": 1784 }, { "epoch": 7.055335968379446, "grad_norm": 1.8201833963394165, "learning_rate": 3.242574257425743e-05, "loss": 0.067, "step": 1785 }, { "epoch": 7.059288537549407, "grad_norm": 1.4718295335769653, "learning_rate": 3.241584158415842e-05, "loss": 0.0509, "step": 1786 }, { "epoch": 7.063241106719367, "grad_norm": 2.529611349105835, "learning_rate": 3.2405940594059406e-05, "loss": 0.0842, "step": 1787 }, { "epoch": 7.067193675889328, "grad_norm": 2.0399839878082275, "learning_rate": 3.2396039603960397e-05, "loss": 0.0792, "step": 1788 }, { "epoch": 7.071146245059288, "grad_norm": 2.477346897125244, "learning_rate": 3.238613861386139e-05, "loss": 0.0849, "step": 1789 }, { "epoch": 7.075098814229249, "grad_norm": 1.4424556493759155, "learning_rate": 3.237623762376238e-05, "loss": 0.051, "step": 1790 }, { "epoch": 7.07905138339921, "grad_norm": 2.3111801147460938, "learning_rate": 3.236633663366337e-05, "loss": 0.0678, "step": 1791 }, { "epoch": 7.08300395256917, "grad_norm": 1.8182194232940674, "learning_rate": 3.235643564356436e-05, "loss": 0.0692, "step": 1792 }, { "epoch": 7.086956521739131, "grad_norm": 2.277977466583252, "learning_rate": 3.234653465346535e-05, "loss": 0.0667, "step": 1793 }, { "epoch": 7.090909090909091, "grad_norm": 1.9708952903747559, "learning_rate": 3.2336633663366336e-05, "loss": 0.0583, "step": 1794 }, { "epoch": 7.094861660079052, "grad_norm": 1.9571771621704102, "learning_rate": 3.232673267326733e-05, "loss": 0.0723, "step": 1795 }, { "epoch": 7.098814229249012, "grad_norm": 2.7242064476013184, "learning_rate": 3.231683168316832e-05, "loss": 0.0865, "step": 1796 }, { "epoch": 7.102766798418973, "grad_norm": 2.1025993824005127, "learning_rate": 3.230693069306931e-05, "loss": 0.0855, "step": 1797 }, { "epoch": 7.106719367588933, "grad_norm": 2.2446298599243164, "learning_rate": 3.22970297029703e-05, "loss": 0.0753, "step": 1798 }, { "epoch": 7.1106719367588935, "grad_norm": 2.1202940940856934, "learning_rate": 3.228712871287129e-05, "loss": 0.0629, "step": 1799 }, { "epoch": 7.1146245059288535, "grad_norm": 2.2542715072631836, "learning_rate": 3.227722772277228e-05, "loss": 0.0802, "step": 1800 }, { "epoch": 7.118577075098814, "grad_norm": 2.85803484916687, "learning_rate": 3.226732673267327e-05, "loss": 0.0986, "step": 1801 }, { "epoch": 7.122529644268774, "grad_norm": 1.9185051918029785, "learning_rate": 3.225742574257426e-05, "loss": 0.072, "step": 1802 }, { "epoch": 7.126482213438735, "grad_norm": 2.088524580001831, "learning_rate": 3.224752475247525e-05, "loss": 0.0828, "step": 1803 }, { "epoch": 7.130434782608695, "grad_norm": 2.130568027496338, "learning_rate": 3.223762376237624e-05, "loss": 0.0826, "step": 1804 }, { "epoch": 7.134387351778656, "grad_norm": 1.8413723707199097, "learning_rate": 3.222772277227723e-05, "loss": 0.0716, "step": 1805 }, { "epoch": 7.138339920948616, "grad_norm": 1.9733790159225464, "learning_rate": 3.221782178217822e-05, "loss": 0.0789, "step": 1806 }, { "epoch": 7.142292490118577, "grad_norm": 1.8133304119110107, "learning_rate": 3.220792079207921e-05, "loss": 0.0665, "step": 1807 }, { "epoch": 7.146245059288537, "grad_norm": 2.530758857727051, "learning_rate": 3.21980198019802e-05, "loss": 0.0897, "step": 1808 }, { "epoch": 7.150197628458498, "grad_norm": 2.503737211227417, "learning_rate": 3.218811881188119e-05, "loss": 0.0997, "step": 1809 }, { "epoch": 7.154150197628459, "grad_norm": 1.7963011264801025, "learning_rate": 3.217821782178218e-05, "loss": 0.0697, "step": 1810 }, { "epoch": 7.158102766798419, "grad_norm": 1.9213933944702148, "learning_rate": 3.216831683168317e-05, "loss": 0.0681, "step": 1811 }, { "epoch": 7.16205533596838, "grad_norm": 1.7468209266662598, "learning_rate": 3.215841584158416e-05, "loss": 0.0568, "step": 1812 }, { "epoch": 7.16600790513834, "grad_norm": 1.8300637006759644, "learning_rate": 3.214851485148515e-05, "loss": 0.0655, "step": 1813 }, { "epoch": 7.169960474308301, "grad_norm": 2.3735764026641846, "learning_rate": 3.2138613861386144e-05, "loss": 0.0875, "step": 1814 }, { "epoch": 7.173913043478261, "grad_norm": 2.8960280418395996, "learning_rate": 3.212871287128713e-05, "loss": 0.0892, "step": 1815 }, { "epoch": 7.177865612648222, "grad_norm": 1.9711147546768188, "learning_rate": 3.211881188118812e-05, "loss": 0.0749, "step": 1816 }, { "epoch": 7.181818181818182, "grad_norm": 2.2650413513183594, "learning_rate": 3.210891089108911e-05, "loss": 0.0805, "step": 1817 }, { "epoch": 7.1857707509881426, "grad_norm": 1.9703688621520996, "learning_rate": 3.20990099009901e-05, "loss": 0.0673, "step": 1818 }, { "epoch": 7.189723320158103, "grad_norm": 2.0809366703033447, "learning_rate": 3.208910891089109e-05, "loss": 0.0821, "step": 1819 }, { "epoch": 7.1936758893280635, "grad_norm": 2.30963134765625, "learning_rate": 3.2079207920792084e-05, "loss": 0.0819, "step": 1820 }, { "epoch": 7.1976284584980235, "grad_norm": 1.8199840784072876, "learning_rate": 3.2069306930693075e-05, "loss": 0.0654, "step": 1821 }, { "epoch": 7.201581027667984, "grad_norm": 1.740988850593567, "learning_rate": 3.205940594059406e-05, "loss": 0.0721, "step": 1822 }, { "epoch": 7.205533596837944, "grad_norm": 1.639424443244934, "learning_rate": 3.204950495049505e-05, "loss": 0.0585, "step": 1823 }, { "epoch": 7.209486166007905, "grad_norm": 3.0076167583465576, "learning_rate": 3.203960396039604e-05, "loss": 0.0877, "step": 1824 }, { "epoch": 7.213438735177865, "grad_norm": 2.3534812927246094, "learning_rate": 3.202970297029703e-05, "loss": 0.077, "step": 1825 }, { "epoch": 7.217391304347826, "grad_norm": 2.1637375354766846, "learning_rate": 3.201980198019802e-05, "loss": 0.0798, "step": 1826 }, { "epoch": 7.221343873517786, "grad_norm": 2.1557679176330566, "learning_rate": 3.2009900990099014e-05, "loss": 0.0724, "step": 1827 }, { "epoch": 7.225296442687747, "grad_norm": 2.326730966567993, "learning_rate": 3.2000000000000005e-05, "loss": 0.0793, "step": 1828 }, { "epoch": 7.229249011857707, "grad_norm": 1.8394434452056885, "learning_rate": 3.199009900990099e-05, "loss": 0.0719, "step": 1829 }, { "epoch": 7.233201581027668, "grad_norm": 1.7972646951675415, "learning_rate": 3.198019801980198e-05, "loss": 0.0778, "step": 1830 }, { "epoch": 7.237154150197629, "grad_norm": 2.5534863471984863, "learning_rate": 3.197029702970297e-05, "loss": 0.0865, "step": 1831 }, { "epoch": 7.241106719367589, "grad_norm": 1.6211029291152954, "learning_rate": 3.196039603960396e-05, "loss": 0.0645, "step": 1832 }, { "epoch": 7.24505928853755, "grad_norm": 2.4950435161590576, "learning_rate": 3.1950495049504954e-05, "loss": 0.0894, "step": 1833 }, { "epoch": 7.24901185770751, "grad_norm": 2.2166330814361572, "learning_rate": 3.1940594059405945e-05, "loss": 0.0729, "step": 1834 }, { "epoch": 7.252964426877471, "grad_norm": 2.333777904510498, "learning_rate": 3.1930693069306936e-05, "loss": 0.0899, "step": 1835 }, { "epoch": 7.256916996047431, "grad_norm": 1.9162453413009644, "learning_rate": 3.192079207920792e-05, "loss": 0.0698, "step": 1836 }, { "epoch": 7.260869565217392, "grad_norm": 2.456912040710449, "learning_rate": 3.191089108910891e-05, "loss": 0.0835, "step": 1837 }, { "epoch": 7.264822134387352, "grad_norm": 2.1436233520507812, "learning_rate": 3.19009900990099e-05, "loss": 0.082, "step": 1838 }, { "epoch": 7.2687747035573125, "grad_norm": 2.360877275466919, "learning_rate": 3.1891089108910894e-05, "loss": 0.0751, "step": 1839 }, { "epoch": 7.2727272727272725, "grad_norm": 2.1815385818481445, "learning_rate": 3.1881188118811885e-05, "loss": 0.082, "step": 1840 }, { "epoch": 7.276679841897233, "grad_norm": 1.846612811088562, "learning_rate": 3.1871287128712876e-05, "loss": 0.066, "step": 1841 }, { "epoch": 7.280632411067193, "grad_norm": 3.6257145404815674, "learning_rate": 3.186138613861387e-05, "loss": 0.12, "step": 1842 }, { "epoch": 7.284584980237154, "grad_norm": 2.0456714630126953, "learning_rate": 3.185148514851485e-05, "loss": 0.0662, "step": 1843 }, { "epoch": 7.288537549407114, "grad_norm": 2.1362714767456055, "learning_rate": 3.184158415841584e-05, "loss": 0.0853, "step": 1844 }, { "epoch": 7.292490118577075, "grad_norm": 2.014970064163208, "learning_rate": 3.183168316831683e-05, "loss": 0.071, "step": 1845 }, { "epoch": 7.296442687747035, "grad_norm": 2.023700714111328, "learning_rate": 3.1821782178217824e-05, "loss": 0.0827, "step": 1846 }, { "epoch": 7.300395256916996, "grad_norm": 2.037409782409668, "learning_rate": 3.1811881188118815e-05, "loss": 0.0703, "step": 1847 }, { "epoch": 7.304347826086957, "grad_norm": 2.0059449672698975, "learning_rate": 3.1801980198019806e-05, "loss": 0.0834, "step": 1848 }, { "epoch": 7.308300395256917, "grad_norm": 2.086137056350708, "learning_rate": 3.17920792079208e-05, "loss": 0.0777, "step": 1849 }, { "epoch": 7.312252964426877, "grad_norm": 2.200608253479004, "learning_rate": 3.178217821782178e-05, "loss": 0.0708, "step": 1850 }, { "epoch": 7.316205533596838, "grad_norm": 2.275210380554199, "learning_rate": 3.177227722772277e-05, "loss": 0.0931, "step": 1851 }, { "epoch": 7.320158102766799, "grad_norm": 2.0692341327667236, "learning_rate": 3.1762376237623764e-05, "loss": 0.0766, "step": 1852 }, { "epoch": 7.324110671936759, "grad_norm": 1.943354845046997, "learning_rate": 3.1752475247524755e-05, "loss": 0.0648, "step": 1853 }, { "epoch": 7.32806324110672, "grad_norm": 2.432307004928589, "learning_rate": 3.1742574257425746e-05, "loss": 0.1151, "step": 1854 }, { "epoch": 7.33201581027668, "grad_norm": 2.313936948776245, "learning_rate": 3.173267326732674e-05, "loss": 0.0954, "step": 1855 }, { "epoch": 7.335968379446641, "grad_norm": 1.9747581481933594, "learning_rate": 3.172277227722773e-05, "loss": 0.0657, "step": 1856 }, { "epoch": 7.339920948616601, "grad_norm": 2.2645602226257324, "learning_rate": 3.171287128712871e-05, "loss": 0.0767, "step": 1857 }, { "epoch": 7.3438735177865615, "grad_norm": 2.0246191024780273, "learning_rate": 3.1702970297029703e-05, "loss": 0.0877, "step": 1858 }, { "epoch": 7.3478260869565215, "grad_norm": 2.0427699089050293, "learning_rate": 3.1693069306930694e-05, "loss": 0.0911, "step": 1859 }, { "epoch": 7.351778656126482, "grad_norm": 2.5756514072418213, "learning_rate": 3.1683168316831686e-05, "loss": 0.0925, "step": 1860 }, { "epoch": 7.355731225296442, "grad_norm": 2.4556028842926025, "learning_rate": 3.167326732673268e-05, "loss": 0.0783, "step": 1861 }, { "epoch": 7.359683794466403, "grad_norm": 2.0145773887634277, "learning_rate": 3.166336633663367e-05, "loss": 0.0657, "step": 1862 }, { "epoch": 7.363636363636363, "grad_norm": 1.9760226011276245, "learning_rate": 3.165346534653465e-05, "loss": 0.0887, "step": 1863 }, { "epoch": 7.367588932806324, "grad_norm": 1.9677963256835938, "learning_rate": 3.164356435643564e-05, "loss": 0.0663, "step": 1864 }, { "epoch": 7.371541501976284, "grad_norm": 2.2797627449035645, "learning_rate": 3.1633663366336634e-05, "loss": 0.0899, "step": 1865 }, { "epoch": 7.375494071146245, "grad_norm": 1.9144444465637207, "learning_rate": 3.1623762376237625e-05, "loss": 0.0655, "step": 1866 }, { "epoch": 7.379446640316205, "grad_norm": 2.6955976486206055, "learning_rate": 3.1613861386138616e-05, "loss": 0.1026, "step": 1867 }, { "epoch": 7.383399209486166, "grad_norm": 1.865747332572937, "learning_rate": 3.160396039603961e-05, "loss": 0.0639, "step": 1868 }, { "epoch": 7.387351778656127, "grad_norm": 2.758430242538452, "learning_rate": 3.15940594059406e-05, "loss": 0.086, "step": 1869 }, { "epoch": 7.391304347826087, "grad_norm": 2.5428738594055176, "learning_rate": 3.158415841584158e-05, "loss": 0.113, "step": 1870 }, { "epoch": 7.395256916996048, "grad_norm": 2.0843374729156494, "learning_rate": 3.1574257425742574e-05, "loss": 0.0687, "step": 1871 }, { "epoch": 7.399209486166008, "grad_norm": 2.1183483600616455, "learning_rate": 3.1564356435643565e-05, "loss": 0.0733, "step": 1872 }, { "epoch": 7.403162055335969, "grad_norm": 1.8515515327453613, "learning_rate": 3.1554455445544556e-05, "loss": 0.0685, "step": 1873 }, { "epoch": 7.407114624505929, "grad_norm": 1.9264715909957886, "learning_rate": 3.154455445544555e-05, "loss": 0.0764, "step": 1874 }, { "epoch": 7.41106719367589, "grad_norm": 2.817488431930542, "learning_rate": 3.153465346534654e-05, "loss": 0.0917, "step": 1875 }, { "epoch": 7.41501976284585, "grad_norm": 1.9016402959823608, "learning_rate": 3.152475247524753e-05, "loss": 0.0764, "step": 1876 }, { "epoch": 7.4189723320158105, "grad_norm": 1.8834534883499146, "learning_rate": 3.151485148514851e-05, "loss": 0.0672, "step": 1877 }, { "epoch": 7.4229249011857705, "grad_norm": 1.8532918691635132, "learning_rate": 3.1504950495049504e-05, "loss": 0.0794, "step": 1878 }, { "epoch": 7.426877470355731, "grad_norm": 2.0499825477600098, "learning_rate": 3.1495049504950495e-05, "loss": 0.0752, "step": 1879 }, { "epoch": 7.430830039525691, "grad_norm": 1.967946171760559, "learning_rate": 3.1485148514851487e-05, "loss": 0.0826, "step": 1880 }, { "epoch": 7.434782608695652, "grad_norm": 2.374180316925049, "learning_rate": 3.147524752475248e-05, "loss": 0.0898, "step": 1881 }, { "epoch": 7.438735177865612, "grad_norm": 3.425501823425293, "learning_rate": 3.146534653465347e-05, "loss": 0.118, "step": 1882 }, { "epoch": 7.442687747035573, "grad_norm": 2.6722114086151123, "learning_rate": 3.145544554455446e-05, "loss": 0.0966, "step": 1883 }, { "epoch": 7.446640316205533, "grad_norm": 2.2939226627349854, "learning_rate": 3.1445544554455444e-05, "loss": 0.0881, "step": 1884 }, { "epoch": 7.450592885375494, "grad_norm": 2.1664910316467285, "learning_rate": 3.1435643564356435e-05, "loss": 0.07, "step": 1885 }, { "epoch": 7.454545454545454, "grad_norm": 1.6963131427764893, "learning_rate": 3.1425742574257426e-05, "loss": 0.0621, "step": 1886 }, { "epoch": 7.458498023715415, "grad_norm": 2.930201292037964, "learning_rate": 3.141584158415842e-05, "loss": 0.1227, "step": 1887 }, { "epoch": 7.462450592885375, "grad_norm": 2.088059186935425, "learning_rate": 3.140594059405941e-05, "loss": 0.0819, "step": 1888 }, { "epoch": 7.466403162055336, "grad_norm": 2.301215648651123, "learning_rate": 3.13960396039604e-05, "loss": 0.0723, "step": 1889 }, { "epoch": 7.470355731225297, "grad_norm": 2.113079786300659, "learning_rate": 3.138613861386139e-05, "loss": 0.0973, "step": 1890 }, { "epoch": 7.474308300395257, "grad_norm": 2.3632912635803223, "learning_rate": 3.1376237623762375e-05, "loss": 0.0768, "step": 1891 }, { "epoch": 7.478260869565218, "grad_norm": 2.5391876697540283, "learning_rate": 3.1366336633663366e-05, "loss": 0.0863, "step": 1892 }, { "epoch": 7.482213438735178, "grad_norm": 1.9134883880615234, "learning_rate": 3.135643564356436e-05, "loss": 0.0699, "step": 1893 }, { "epoch": 7.486166007905139, "grad_norm": 2.1936020851135254, "learning_rate": 3.134653465346535e-05, "loss": 0.0742, "step": 1894 }, { "epoch": 7.490118577075099, "grad_norm": 1.9534437656402588, "learning_rate": 3.133663366336634e-05, "loss": 0.0798, "step": 1895 }, { "epoch": 7.4940711462450595, "grad_norm": 2.049443006515503, "learning_rate": 3.132673267326733e-05, "loss": 0.0802, "step": 1896 }, { "epoch": 7.4980237154150196, "grad_norm": 2.5238490104675293, "learning_rate": 3.131683168316832e-05, "loss": 0.0983, "step": 1897 }, { "epoch": 7.5019762845849804, "grad_norm": 2.1057331562042236, "learning_rate": 3.1306930693069305e-05, "loss": 0.0889, "step": 1898 }, { "epoch": 7.5059288537549405, "grad_norm": 2.19503116607666, "learning_rate": 3.1297029702970296e-05, "loss": 0.0766, "step": 1899 }, { "epoch": 7.509881422924901, "grad_norm": 2.296220302581787, "learning_rate": 3.128712871287129e-05, "loss": 0.0732, "step": 1900 }, { "epoch": 7.513833992094861, "grad_norm": 2.4914751052856445, "learning_rate": 3.127722772277228e-05, "loss": 0.084, "step": 1901 }, { "epoch": 7.517786561264822, "grad_norm": 2.1635220050811768, "learning_rate": 3.126732673267327e-05, "loss": 0.0805, "step": 1902 }, { "epoch": 7.521739130434782, "grad_norm": 2.4028971195220947, "learning_rate": 3.125742574257426e-05, "loss": 0.0908, "step": 1903 }, { "epoch": 7.525691699604743, "grad_norm": 2.5337584018707275, "learning_rate": 3.124752475247525e-05, "loss": 0.1009, "step": 1904 }, { "epoch": 7.529644268774703, "grad_norm": 1.6034029722213745, "learning_rate": 3.1237623762376236e-05, "loss": 0.0688, "step": 1905 }, { "epoch": 7.533596837944664, "grad_norm": 2.9323158264160156, "learning_rate": 3.122772277227723e-05, "loss": 0.1039, "step": 1906 }, { "epoch": 7.537549407114625, "grad_norm": 2.126162528991699, "learning_rate": 3.121782178217822e-05, "loss": 0.0747, "step": 1907 }, { "epoch": 7.541501976284585, "grad_norm": 2.247230291366577, "learning_rate": 3.120792079207921e-05, "loss": 0.0894, "step": 1908 }, { "epoch": 7.545454545454545, "grad_norm": 2.235623598098755, "learning_rate": 3.11980198019802e-05, "loss": 0.0824, "step": 1909 }, { "epoch": 7.549407114624506, "grad_norm": 2.5068511962890625, "learning_rate": 3.118811881188119e-05, "loss": 0.0851, "step": 1910 }, { "epoch": 7.553359683794467, "grad_norm": 1.894625186920166, "learning_rate": 3.117821782178218e-05, "loss": 0.0748, "step": 1911 }, { "epoch": 7.557312252964427, "grad_norm": 2.528669595718384, "learning_rate": 3.116831683168317e-05, "loss": 0.0945, "step": 1912 }, { "epoch": 7.561264822134388, "grad_norm": 2.7860591411590576, "learning_rate": 3.115841584158416e-05, "loss": 0.1065, "step": 1913 }, { "epoch": 7.565217391304348, "grad_norm": 2.199695348739624, "learning_rate": 3.114851485148515e-05, "loss": 0.0867, "step": 1914 }, { "epoch": 7.569169960474309, "grad_norm": 2.248028516769409, "learning_rate": 3.113861386138614e-05, "loss": 0.0846, "step": 1915 }, { "epoch": 7.573122529644269, "grad_norm": 1.7341463565826416, "learning_rate": 3.112871287128713e-05, "loss": 0.0577, "step": 1916 }, { "epoch": 7.5770750988142295, "grad_norm": 1.94870924949646, "learning_rate": 3.111881188118812e-05, "loss": 0.0768, "step": 1917 }, { "epoch": 7.5810276679841895, "grad_norm": 2.112471103668213, "learning_rate": 3.110891089108911e-05, "loss": 0.0762, "step": 1918 }, { "epoch": 7.58498023715415, "grad_norm": 2.2701523303985596, "learning_rate": 3.10990099009901e-05, "loss": 0.0805, "step": 1919 }, { "epoch": 7.58893280632411, "grad_norm": 2.0794973373413086, "learning_rate": 3.108910891089109e-05, "loss": 0.0742, "step": 1920 }, { "epoch": 7.592885375494071, "grad_norm": 2.236959934234619, "learning_rate": 3.107920792079208e-05, "loss": 0.0834, "step": 1921 }, { "epoch": 7.596837944664031, "grad_norm": 2.2500040531158447, "learning_rate": 3.106930693069307e-05, "loss": 0.0682, "step": 1922 }, { "epoch": 7.600790513833992, "grad_norm": 1.899522304534912, "learning_rate": 3.105940594059406e-05, "loss": 0.0727, "step": 1923 }, { "epoch": 7.604743083003952, "grad_norm": 2.9326465129852295, "learning_rate": 3.104950495049505e-05, "loss": 0.0978, "step": 1924 }, { "epoch": 7.608695652173913, "grad_norm": 2.100834608078003, "learning_rate": 3.1039603960396044e-05, "loss": 0.0896, "step": 1925 }, { "epoch": 7.612648221343873, "grad_norm": 2.061936378479004, "learning_rate": 3.102970297029703e-05, "loss": 0.0662, "step": 1926 }, { "epoch": 7.616600790513834, "grad_norm": 2.2188315391540527, "learning_rate": 3.101980198019802e-05, "loss": 0.0737, "step": 1927 }, { "epoch": 7.620553359683795, "grad_norm": 3.3783230781555176, "learning_rate": 3.100990099009901e-05, "loss": 0.127, "step": 1928 }, { "epoch": 7.624505928853755, "grad_norm": 2.727034568786621, "learning_rate": 3.1e-05, "loss": 0.0901, "step": 1929 }, { "epoch": 7.628458498023716, "grad_norm": 1.8792964220046997, "learning_rate": 3.099009900990099e-05, "loss": 0.0588, "step": 1930 }, { "epoch": 7.632411067193676, "grad_norm": 1.8605564832687378, "learning_rate": 3.0980198019801983e-05, "loss": 0.0727, "step": 1931 }, { "epoch": 7.636363636363637, "grad_norm": 2.1648929119110107, "learning_rate": 3.0970297029702975e-05, "loss": 0.0832, "step": 1932 }, { "epoch": 7.640316205533597, "grad_norm": 2.2906036376953125, "learning_rate": 3.096039603960396e-05, "loss": 0.1083, "step": 1933 }, { "epoch": 7.644268774703558, "grad_norm": 1.7894140481948853, "learning_rate": 3.095049504950495e-05, "loss": 0.0753, "step": 1934 }, { "epoch": 7.648221343873518, "grad_norm": 2.301386594772339, "learning_rate": 3.094059405940594e-05, "loss": 0.0715, "step": 1935 }, { "epoch": 7.6521739130434785, "grad_norm": 2.7486836910247803, "learning_rate": 3.093069306930693e-05, "loss": 0.1043, "step": 1936 }, { "epoch": 7.6561264822134385, "grad_norm": 3.0914204120635986, "learning_rate": 3.092079207920792e-05, "loss": 0.1068, "step": 1937 }, { "epoch": 7.660079051383399, "grad_norm": 2.3445394039154053, "learning_rate": 3.0910891089108914e-05, "loss": 0.0827, "step": 1938 }, { "epoch": 7.664031620553359, "grad_norm": 1.9056646823883057, "learning_rate": 3.0900990099009905e-05, "loss": 0.0634, "step": 1939 }, { "epoch": 7.66798418972332, "grad_norm": 1.8222278356552124, "learning_rate": 3.089108910891089e-05, "loss": 0.0687, "step": 1940 }, { "epoch": 7.67193675889328, "grad_norm": 2.551405429840088, "learning_rate": 3.088118811881188e-05, "loss": 0.0861, "step": 1941 }, { "epoch": 7.675889328063241, "grad_norm": 2.659294843673706, "learning_rate": 3.087128712871287e-05, "loss": 0.1035, "step": 1942 }, { "epoch": 7.679841897233201, "grad_norm": 2.4411120414733887, "learning_rate": 3.086138613861386e-05, "loss": 0.0811, "step": 1943 }, { "epoch": 7.683794466403162, "grad_norm": 3.4701881408691406, "learning_rate": 3.0851485148514854e-05, "loss": 0.1097, "step": 1944 }, { "epoch": 7.687747035573123, "grad_norm": 1.9612984657287598, "learning_rate": 3.0841584158415845e-05, "loss": 0.0712, "step": 1945 }, { "epoch": 7.691699604743083, "grad_norm": 2.435497999191284, "learning_rate": 3.0831683168316836e-05, "loss": 0.0875, "step": 1946 }, { "epoch": 7.695652173913043, "grad_norm": 2.1108126640319824, "learning_rate": 3.082178217821782e-05, "loss": 0.0995, "step": 1947 }, { "epoch": 7.699604743083004, "grad_norm": 2.516301155090332, "learning_rate": 3.081188118811881e-05, "loss": 0.089, "step": 1948 }, { "epoch": 7.703557312252965, "grad_norm": 2.3038480281829834, "learning_rate": 3.08019801980198e-05, "loss": 0.0837, "step": 1949 }, { "epoch": 7.707509881422925, "grad_norm": 2.101778984069824, "learning_rate": 3.0792079207920793e-05, "loss": 0.0693, "step": 1950 }, { "epoch": 7.711462450592886, "grad_norm": 2.474229574203491, "learning_rate": 3.0782178217821784e-05, "loss": 0.0849, "step": 1951 }, { "epoch": 7.715415019762846, "grad_norm": 2.0718514919281006, "learning_rate": 3.0772277227722776e-05, "loss": 0.0778, "step": 1952 }, { "epoch": 7.719367588932807, "grad_norm": 2.4683122634887695, "learning_rate": 3.0762376237623767e-05, "loss": 0.1013, "step": 1953 }, { "epoch": 7.723320158102767, "grad_norm": 3.0682730674743652, "learning_rate": 3.075247524752475e-05, "loss": 0.1066, "step": 1954 }, { "epoch": 7.7272727272727275, "grad_norm": 2.8526501655578613, "learning_rate": 3.074257425742574e-05, "loss": 0.0982, "step": 1955 }, { "epoch": 7.7312252964426875, "grad_norm": 2.544355869293213, "learning_rate": 3.073267326732673e-05, "loss": 0.0867, "step": 1956 }, { "epoch": 7.735177865612648, "grad_norm": 1.9162529706954956, "learning_rate": 3.0722772277227724e-05, "loss": 0.0787, "step": 1957 }, { "epoch": 7.739130434782608, "grad_norm": 1.9401508569717407, "learning_rate": 3.0712871287128715e-05, "loss": 0.087, "step": 1958 }, { "epoch": 7.743083003952569, "grad_norm": 2.3244242668151855, "learning_rate": 3.0702970297029706e-05, "loss": 0.0996, "step": 1959 }, { "epoch": 7.747035573122529, "grad_norm": 2.221698045730591, "learning_rate": 3.06930693069307e-05, "loss": 0.081, "step": 1960 }, { "epoch": 7.75098814229249, "grad_norm": 2.8440051078796387, "learning_rate": 3.068316831683168e-05, "loss": 0.113, "step": 1961 }, { "epoch": 7.75494071146245, "grad_norm": 2.2042148113250732, "learning_rate": 3.067326732673267e-05, "loss": 0.0871, "step": 1962 }, { "epoch": 7.758893280632411, "grad_norm": 2.250248670578003, "learning_rate": 3.0663366336633664e-05, "loss": 0.0852, "step": 1963 }, { "epoch": 7.762845849802371, "grad_norm": 2.214618682861328, "learning_rate": 3.0653465346534655e-05, "loss": 0.0827, "step": 1964 }, { "epoch": 7.766798418972332, "grad_norm": 2.473637104034424, "learning_rate": 3.0643564356435646e-05, "loss": 0.0928, "step": 1965 }, { "epoch": 7.770750988142293, "grad_norm": 2.6402039527893066, "learning_rate": 3.063366336633664e-05, "loss": 0.1012, "step": 1966 }, { "epoch": 7.774703557312253, "grad_norm": 2.234550952911377, "learning_rate": 3.062376237623763e-05, "loss": 0.0989, "step": 1967 }, { "epoch": 7.778656126482213, "grad_norm": 2.4424855709075928, "learning_rate": 3.061386138613861e-05, "loss": 0.0971, "step": 1968 }, { "epoch": 7.782608695652174, "grad_norm": 2.6651477813720703, "learning_rate": 3.06039603960396e-05, "loss": 0.0847, "step": 1969 }, { "epoch": 7.786561264822135, "grad_norm": 2.028580904006958, "learning_rate": 3.0594059405940594e-05, "loss": 0.0829, "step": 1970 }, { "epoch": 7.790513833992095, "grad_norm": 2.5752015113830566, "learning_rate": 3.0584158415841585e-05, "loss": 0.1038, "step": 1971 }, { "epoch": 7.794466403162056, "grad_norm": 2.5352060794830322, "learning_rate": 3.0574257425742577e-05, "loss": 0.0851, "step": 1972 }, { "epoch": 7.798418972332016, "grad_norm": 2.379042387008667, "learning_rate": 3.056435643564357e-05, "loss": 0.0742, "step": 1973 }, { "epoch": 7.8023715415019765, "grad_norm": 2.0376081466674805, "learning_rate": 3.055445544554456e-05, "loss": 0.0785, "step": 1974 }, { "epoch": 7.8063241106719365, "grad_norm": 2.3658785820007324, "learning_rate": 3.054455445544554e-05, "loss": 0.0946, "step": 1975 }, { "epoch": 7.810276679841897, "grad_norm": 2.276978015899658, "learning_rate": 3.0534653465346534e-05, "loss": 0.0916, "step": 1976 }, { "epoch": 7.8142292490118574, "grad_norm": 2.4611423015594482, "learning_rate": 3.0524752475247525e-05, "loss": 0.1072, "step": 1977 }, { "epoch": 7.818181818181818, "grad_norm": 3.030529499053955, "learning_rate": 3.0514851485148516e-05, "loss": 0.1039, "step": 1978 }, { "epoch": 7.822134387351778, "grad_norm": 2.3222742080688477, "learning_rate": 3.0504950495049507e-05, "loss": 0.0986, "step": 1979 }, { "epoch": 7.826086956521739, "grad_norm": 2.2313146591186523, "learning_rate": 3.0495049504950495e-05, "loss": 0.0804, "step": 1980 }, { "epoch": 7.830039525691699, "grad_norm": 2.533900499343872, "learning_rate": 3.0485148514851486e-05, "loss": 0.0974, "step": 1981 }, { "epoch": 7.83399209486166, "grad_norm": 1.8796467781066895, "learning_rate": 3.0475247524752477e-05, "loss": 0.0971, "step": 1982 }, { "epoch": 7.837944664031621, "grad_norm": 1.9251315593719482, "learning_rate": 3.0465346534653465e-05, "loss": 0.0797, "step": 1983 }, { "epoch": 7.841897233201581, "grad_norm": 1.9420123100280762, "learning_rate": 3.0455445544554456e-05, "loss": 0.0801, "step": 1984 }, { "epoch": 7.845849802371541, "grad_norm": 2.108203649520874, "learning_rate": 3.0445544554455447e-05, "loss": 0.0748, "step": 1985 }, { "epoch": 7.849802371541502, "grad_norm": 2.752800226211548, "learning_rate": 3.0435643564356438e-05, "loss": 0.0875, "step": 1986 }, { "epoch": 7.853754940711463, "grad_norm": 2.0807039737701416, "learning_rate": 3.0425742574257426e-05, "loss": 0.0887, "step": 1987 }, { "epoch": 7.857707509881423, "grad_norm": 2.751215934753418, "learning_rate": 3.0415841584158417e-05, "loss": 0.0971, "step": 1988 }, { "epoch": 7.861660079051384, "grad_norm": 2.135378122329712, "learning_rate": 3.0405940594059408e-05, "loss": 0.0809, "step": 1989 }, { "epoch": 7.865612648221344, "grad_norm": 2.3896353244781494, "learning_rate": 3.0396039603960395e-05, "loss": 0.0823, "step": 1990 }, { "epoch": 7.869565217391305, "grad_norm": 2.23494291305542, "learning_rate": 3.0386138613861386e-05, "loss": 0.0917, "step": 1991 }, { "epoch": 7.873517786561265, "grad_norm": 3.176953077316284, "learning_rate": 3.0376237623762378e-05, "loss": 0.1163, "step": 1992 }, { "epoch": 7.877470355731226, "grad_norm": 2.1850149631500244, "learning_rate": 3.036633663366337e-05, "loss": 0.0841, "step": 1993 }, { "epoch": 7.881422924901186, "grad_norm": 2.343153715133667, "learning_rate": 3.0356435643564356e-05, "loss": 0.0743, "step": 1994 }, { "epoch": 7.8853754940711465, "grad_norm": 1.7955726385116577, "learning_rate": 3.0346534653465347e-05, "loss": 0.0748, "step": 1995 }, { "epoch": 7.8893280632411065, "grad_norm": 2.471789598464966, "learning_rate": 3.033663366336634e-05, "loss": 0.0797, "step": 1996 }, { "epoch": 7.893280632411067, "grad_norm": 1.9187581539154053, "learning_rate": 3.0326732673267326e-05, "loss": 0.0625, "step": 1997 }, { "epoch": 7.897233201581027, "grad_norm": 1.9856213331222534, "learning_rate": 3.0316831683168317e-05, "loss": 0.0739, "step": 1998 }, { "epoch": 7.901185770750988, "grad_norm": 2.384261131286621, "learning_rate": 3.0306930693069308e-05, "loss": 0.0771, "step": 1999 }, { "epoch": 7.905138339920948, "grad_norm": 2.120793342590332, "learning_rate": 3.02970297029703e-05, "loss": 0.0831, "step": 2000 }, { "epoch": 7.909090909090909, "grad_norm": 2.454169511795044, "learning_rate": 3.0287128712871287e-05, "loss": 0.0864, "step": 2001 }, { "epoch": 7.913043478260869, "grad_norm": 2.5327823162078857, "learning_rate": 3.0277227722772278e-05, "loss": 0.1032, "step": 2002 }, { "epoch": 7.91699604743083, "grad_norm": 2.0178539752960205, "learning_rate": 3.026732673267327e-05, "loss": 0.0729, "step": 2003 }, { "epoch": 7.920948616600791, "grad_norm": 2.319223642349243, "learning_rate": 3.0257425742574257e-05, "loss": 0.0891, "step": 2004 }, { "epoch": 7.924901185770751, "grad_norm": 3.0591816902160645, "learning_rate": 3.0247524752475248e-05, "loss": 0.1062, "step": 2005 }, { "epoch": 7.928853754940711, "grad_norm": 2.132173538208008, "learning_rate": 3.023762376237624e-05, "loss": 0.081, "step": 2006 }, { "epoch": 7.932806324110672, "grad_norm": 2.1925642490386963, "learning_rate": 3.0227722772277227e-05, "loss": 0.0879, "step": 2007 }, { "epoch": 7.936758893280633, "grad_norm": 2.472123861312866, "learning_rate": 3.0217821782178218e-05, "loss": 0.1021, "step": 2008 }, { "epoch": 7.940711462450593, "grad_norm": 2.6943516731262207, "learning_rate": 3.020792079207921e-05, "loss": 0.1018, "step": 2009 }, { "epoch": 7.944664031620554, "grad_norm": 2.146381378173828, "learning_rate": 3.01980198019802e-05, "loss": 0.0794, "step": 2010 }, { "epoch": 7.948616600790514, "grad_norm": 2.303863763809204, "learning_rate": 3.0188118811881187e-05, "loss": 0.0929, "step": 2011 }, { "epoch": 7.952569169960475, "grad_norm": 2.1839396953582764, "learning_rate": 3.017821782178218e-05, "loss": 0.0759, "step": 2012 }, { "epoch": 7.956521739130435, "grad_norm": 2.3496556282043457, "learning_rate": 3.016831683168317e-05, "loss": 0.1132, "step": 2013 }, { "epoch": 7.9604743083003955, "grad_norm": 1.7513418197631836, "learning_rate": 3.0158415841584157e-05, "loss": 0.0744, "step": 2014 }, { "epoch": 7.9644268774703555, "grad_norm": 2.128807544708252, "learning_rate": 3.014851485148515e-05, "loss": 0.0798, "step": 2015 }, { "epoch": 7.968379446640316, "grad_norm": 2.300731658935547, "learning_rate": 3.013861386138614e-05, "loss": 0.0804, "step": 2016 }, { "epoch": 7.972332015810276, "grad_norm": 2.4656991958618164, "learning_rate": 3.012871287128713e-05, "loss": 0.0916, "step": 2017 }, { "epoch": 7.976284584980237, "grad_norm": 2.1707351207733154, "learning_rate": 3.0118811881188118e-05, "loss": 0.0801, "step": 2018 }, { "epoch": 7.980237154150197, "grad_norm": 1.9051332473754883, "learning_rate": 3.010891089108911e-05, "loss": 0.0752, "step": 2019 }, { "epoch": 7.984189723320158, "grad_norm": 2.9272541999816895, "learning_rate": 3.00990099009901e-05, "loss": 0.0966, "step": 2020 }, { "epoch": 7.988142292490118, "grad_norm": 2.758774995803833, "learning_rate": 3.0089108910891088e-05, "loss": 0.0979, "step": 2021 }, { "epoch": 7.992094861660079, "grad_norm": 2.204257011413574, "learning_rate": 3.007920792079208e-05, "loss": 0.1198, "step": 2022 }, { "epoch": 7.996047430830039, "grad_norm": 2.3086135387420654, "learning_rate": 3.006930693069307e-05, "loss": 0.0981, "step": 2023 }, { "epoch": 8.0, "grad_norm": 2.2083933353424072, "learning_rate": 3.005940594059406e-05, "loss": 0.0773, "step": 2024 }, { "epoch": 8.003952569169961, "grad_norm": 1.2820247411727905, "learning_rate": 3.004950495049505e-05, "loss": 0.0435, "step": 2025 }, { "epoch": 8.007905138339922, "grad_norm": 1.6155034303665161, "learning_rate": 3.003960396039604e-05, "loss": 0.0511, "step": 2026 }, { "epoch": 8.011857707509881, "grad_norm": 1.4556971788406372, "learning_rate": 3.002970297029703e-05, "loss": 0.0518, "step": 2027 }, { "epoch": 8.015810276679842, "grad_norm": 1.3475127220153809, "learning_rate": 3.001980198019802e-05, "loss": 0.0414, "step": 2028 }, { "epoch": 8.019762845849803, "grad_norm": 1.4684066772460938, "learning_rate": 3.000990099009901e-05, "loss": 0.0579, "step": 2029 }, { "epoch": 8.023715415019764, "grad_norm": 1.320636510848999, "learning_rate": 3e-05, "loss": 0.0379, "step": 2030 }, { "epoch": 8.027667984189723, "grad_norm": 1.622406244277954, "learning_rate": 2.9990099009900992e-05, "loss": 0.0502, "step": 2031 }, { "epoch": 8.031620553359684, "grad_norm": 2.583479166030884, "learning_rate": 2.998019801980198e-05, "loss": 0.0726, "step": 2032 }, { "epoch": 8.035573122529645, "grad_norm": 1.7958500385284424, "learning_rate": 2.997029702970297e-05, "loss": 0.0498, "step": 2033 }, { "epoch": 8.039525691699605, "grad_norm": 2.085667848587036, "learning_rate": 2.996039603960396e-05, "loss": 0.0582, "step": 2034 }, { "epoch": 8.043478260869565, "grad_norm": 1.9411925077438354, "learning_rate": 2.995049504950495e-05, "loss": 0.0684, "step": 2035 }, { "epoch": 8.047430830039525, "grad_norm": 1.8351064920425415, "learning_rate": 2.994059405940594e-05, "loss": 0.057, "step": 2036 }, { "epoch": 8.051383399209486, "grad_norm": 1.7462049722671509, "learning_rate": 2.993069306930693e-05, "loss": 0.0546, "step": 2037 }, { "epoch": 8.055335968379447, "grad_norm": 1.8774205446243286, "learning_rate": 2.9920792079207923e-05, "loss": 0.0605, "step": 2038 }, { "epoch": 8.059288537549406, "grad_norm": 3.3498106002807617, "learning_rate": 2.991089108910891e-05, "loss": 0.1086, "step": 2039 }, { "epoch": 8.063241106719367, "grad_norm": 1.3285433053970337, "learning_rate": 2.99009900990099e-05, "loss": 0.0438, "step": 2040 }, { "epoch": 8.067193675889328, "grad_norm": 1.860072374343872, "learning_rate": 2.9891089108910892e-05, "loss": 0.0638, "step": 2041 }, { "epoch": 8.071146245059289, "grad_norm": 1.8126575946807861, "learning_rate": 2.988118811881188e-05, "loss": 0.0618, "step": 2042 }, { "epoch": 8.075098814229248, "grad_norm": 1.4979151487350464, "learning_rate": 2.987128712871287e-05, "loss": 0.0481, "step": 2043 }, { "epoch": 8.079051383399209, "grad_norm": 1.75325345993042, "learning_rate": 2.9861386138613862e-05, "loss": 0.0547, "step": 2044 }, { "epoch": 8.08300395256917, "grad_norm": 2.1887335777282715, "learning_rate": 2.9851485148514853e-05, "loss": 0.0689, "step": 2045 }, { "epoch": 8.08695652173913, "grad_norm": 1.674923300743103, "learning_rate": 2.984158415841584e-05, "loss": 0.0638, "step": 2046 }, { "epoch": 8.090909090909092, "grad_norm": 1.5696450471878052, "learning_rate": 2.9831683168316832e-05, "loss": 0.0499, "step": 2047 }, { "epoch": 8.09486166007905, "grad_norm": 1.8903003931045532, "learning_rate": 2.9821782178217823e-05, "loss": 0.0628, "step": 2048 }, { "epoch": 8.098814229249012, "grad_norm": 1.904882550239563, "learning_rate": 2.981188118811881e-05, "loss": 0.0644, "step": 2049 }, { "epoch": 8.102766798418973, "grad_norm": 1.516392469406128, "learning_rate": 2.9801980198019802e-05, "loss": 0.0563, "step": 2050 }, { "epoch": 8.106719367588934, "grad_norm": 1.549643635749817, "learning_rate": 2.9792079207920793e-05, "loss": 0.0633, "step": 2051 }, { "epoch": 8.110671936758893, "grad_norm": 1.5683530569076538, "learning_rate": 2.9782178217821784e-05, "loss": 0.054, "step": 2052 }, { "epoch": 8.114624505928854, "grad_norm": 1.5425059795379639, "learning_rate": 2.977227722772277e-05, "loss": 0.0546, "step": 2053 }, { "epoch": 8.118577075098814, "grad_norm": 1.7313646078109741, "learning_rate": 2.9762376237623763e-05, "loss": 0.0634, "step": 2054 }, { "epoch": 8.122529644268775, "grad_norm": 1.660712480545044, "learning_rate": 2.9752475247524754e-05, "loss": 0.0579, "step": 2055 }, { "epoch": 8.126482213438734, "grad_norm": 2.208404064178467, "learning_rate": 2.974257425742574e-05, "loss": 0.0673, "step": 2056 }, { "epoch": 8.130434782608695, "grad_norm": 1.9389158487319946, "learning_rate": 2.9732673267326732e-05, "loss": 0.0607, "step": 2057 }, { "epoch": 8.134387351778656, "grad_norm": 1.7905778884887695, "learning_rate": 2.9722772277227724e-05, "loss": 0.059, "step": 2058 }, { "epoch": 8.138339920948617, "grad_norm": 1.9263213872909546, "learning_rate": 2.9712871287128715e-05, "loss": 0.0561, "step": 2059 }, { "epoch": 8.142292490118576, "grad_norm": 3.096728801727295, "learning_rate": 2.9702970297029702e-05, "loss": 0.0971, "step": 2060 }, { "epoch": 8.146245059288537, "grad_norm": 3.0652503967285156, "learning_rate": 2.9693069306930693e-05, "loss": 0.0951, "step": 2061 }, { "epoch": 8.150197628458498, "grad_norm": 1.8803935050964355, "learning_rate": 2.9683168316831684e-05, "loss": 0.0536, "step": 2062 }, { "epoch": 8.154150197628459, "grad_norm": 1.8182830810546875, "learning_rate": 2.9673267326732672e-05, "loss": 0.0626, "step": 2063 }, { "epoch": 8.15810276679842, "grad_norm": 2.705658197402954, "learning_rate": 2.9663366336633663e-05, "loss": 0.0773, "step": 2064 }, { "epoch": 8.162055335968379, "grad_norm": 1.9855663776397705, "learning_rate": 2.9653465346534654e-05, "loss": 0.0561, "step": 2065 }, { "epoch": 8.16600790513834, "grad_norm": 1.7251075506210327, "learning_rate": 2.9643564356435642e-05, "loss": 0.0547, "step": 2066 }, { "epoch": 8.1699604743083, "grad_norm": 1.8825830221176147, "learning_rate": 2.9633663366336633e-05, "loss": 0.064, "step": 2067 }, { "epoch": 8.173913043478262, "grad_norm": 1.922816276550293, "learning_rate": 2.9623762376237624e-05, "loss": 0.0678, "step": 2068 }, { "epoch": 8.17786561264822, "grad_norm": 2.061502456665039, "learning_rate": 2.9613861386138615e-05, "loss": 0.061, "step": 2069 }, { "epoch": 8.181818181818182, "grad_norm": 1.7140027284622192, "learning_rate": 2.9603960396039603e-05, "loss": 0.0566, "step": 2070 }, { "epoch": 8.185770750988143, "grad_norm": 1.5344748497009277, "learning_rate": 2.9594059405940594e-05, "loss": 0.0514, "step": 2071 }, { "epoch": 8.189723320158103, "grad_norm": 1.4780486822128296, "learning_rate": 2.9584158415841585e-05, "loss": 0.0484, "step": 2072 }, { "epoch": 8.193675889328063, "grad_norm": 1.4861502647399902, "learning_rate": 2.9574257425742573e-05, "loss": 0.0471, "step": 2073 }, { "epoch": 8.197628458498023, "grad_norm": 1.6154067516326904, "learning_rate": 2.9564356435643564e-05, "loss": 0.0537, "step": 2074 }, { "epoch": 8.201581027667984, "grad_norm": 1.5887720584869385, "learning_rate": 2.9554455445544555e-05, "loss": 0.0509, "step": 2075 }, { "epoch": 8.205533596837945, "grad_norm": 2.011563777923584, "learning_rate": 2.9544554455445546e-05, "loss": 0.0627, "step": 2076 }, { "epoch": 8.209486166007904, "grad_norm": 2.204807996749878, "learning_rate": 2.9534653465346533e-05, "loss": 0.0723, "step": 2077 }, { "epoch": 8.213438735177865, "grad_norm": 1.9599748849868774, "learning_rate": 2.9524752475247524e-05, "loss": 0.0645, "step": 2078 }, { "epoch": 8.217391304347826, "grad_norm": 1.3280692100524902, "learning_rate": 2.9514851485148516e-05, "loss": 0.0436, "step": 2079 }, { "epoch": 8.221343873517787, "grad_norm": 1.9152424335479736, "learning_rate": 2.9504950495049503e-05, "loss": 0.0615, "step": 2080 }, { "epoch": 8.225296442687746, "grad_norm": 1.9655647277832031, "learning_rate": 2.9495049504950494e-05, "loss": 0.0604, "step": 2081 }, { "epoch": 8.229249011857707, "grad_norm": 2.4758338928222656, "learning_rate": 2.9485148514851485e-05, "loss": 0.0677, "step": 2082 }, { "epoch": 8.233201581027668, "grad_norm": 2.014911651611328, "learning_rate": 2.9475247524752476e-05, "loss": 0.0627, "step": 2083 }, { "epoch": 8.237154150197629, "grad_norm": 2.1655447483062744, "learning_rate": 2.9465346534653464e-05, "loss": 0.0688, "step": 2084 }, { "epoch": 8.24110671936759, "grad_norm": 2.5103843212127686, "learning_rate": 2.9455445544554455e-05, "loss": 0.0739, "step": 2085 }, { "epoch": 8.245059288537549, "grad_norm": 1.9521121978759766, "learning_rate": 2.9445544554455446e-05, "loss": 0.0569, "step": 2086 }, { "epoch": 8.24901185770751, "grad_norm": 1.5631282329559326, "learning_rate": 2.9435643564356434e-05, "loss": 0.0571, "step": 2087 }, { "epoch": 8.25296442687747, "grad_norm": 1.966561198234558, "learning_rate": 2.9425742574257425e-05, "loss": 0.0612, "step": 2088 }, { "epoch": 8.256916996047432, "grad_norm": 1.712216854095459, "learning_rate": 2.9415841584158416e-05, "loss": 0.0547, "step": 2089 }, { "epoch": 8.26086956521739, "grad_norm": 1.8395965099334717, "learning_rate": 2.9405940594059407e-05, "loss": 0.0667, "step": 2090 }, { "epoch": 8.264822134387352, "grad_norm": 2.1637749671936035, "learning_rate": 2.9396039603960395e-05, "loss": 0.0625, "step": 2091 }, { "epoch": 8.268774703557312, "grad_norm": 1.8324702978134155, "learning_rate": 2.9386138613861386e-05, "loss": 0.0692, "step": 2092 }, { "epoch": 8.272727272727273, "grad_norm": 2.2161331176757812, "learning_rate": 2.9376237623762377e-05, "loss": 0.0726, "step": 2093 }, { "epoch": 8.276679841897232, "grad_norm": 1.7193669080734253, "learning_rate": 2.9366336633663365e-05, "loss": 0.0672, "step": 2094 }, { "epoch": 8.280632411067193, "grad_norm": 2.248197078704834, "learning_rate": 2.9356435643564356e-05, "loss": 0.0686, "step": 2095 }, { "epoch": 8.284584980237154, "grad_norm": 1.456818699836731, "learning_rate": 2.9346534653465347e-05, "loss": 0.0491, "step": 2096 }, { "epoch": 8.288537549407115, "grad_norm": 1.8384908437728882, "learning_rate": 2.9336633663366338e-05, "loss": 0.0625, "step": 2097 }, { "epoch": 8.292490118577074, "grad_norm": 1.771870493888855, "learning_rate": 2.9326732673267325e-05, "loss": 0.0581, "step": 2098 }, { "epoch": 8.296442687747035, "grad_norm": 2.0997579097747803, "learning_rate": 2.9316831683168317e-05, "loss": 0.0663, "step": 2099 }, { "epoch": 8.300395256916996, "grad_norm": 2.1802330017089844, "learning_rate": 2.9306930693069308e-05, "loss": 0.0693, "step": 2100 }, { "epoch": 8.304347826086957, "grad_norm": 2.5558583736419678, "learning_rate": 2.9297029702970295e-05, "loss": 0.0768, "step": 2101 }, { "epoch": 8.308300395256918, "grad_norm": 2.7538676261901855, "learning_rate": 2.9287128712871286e-05, "loss": 0.0945, "step": 2102 }, { "epoch": 8.312252964426877, "grad_norm": 2.165545701980591, "learning_rate": 2.9277227722772277e-05, "loss": 0.0718, "step": 2103 }, { "epoch": 8.316205533596838, "grad_norm": 1.424879789352417, "learning_rate": 2.926732673267327e-05, "loss": 0.056, "step": 2104 }, { "epoch": 8.320158102766799, "grad_norm": 1.934768557548523, "learning_rate": 2.9257425742574256e-05, "loss": 0.0653, "step": 2105 }, { "epoch": 8.32411067193676, "grad_norm": 1.6412161588668823, "learning_rate": 2.9247524752475247e-05, "loss": 0.0536, "step": 2106 }, { "epoch": 8.328063241106719, "grad_norm": 1.86897873878479, "learning_rate": 2.9237623762376238e-05, "loss": 0.0684, "step": 2107 }, { "epoch": 8.33201581027668, "grad_norm": 2.2633161544799805, "learning_rate": 2.9227722772277226e-05, "loss": 0.0728, "step": 2108 }, { "epoch": 8.33596837944664, "grad_norm": 2.2101898193359375, "learning_rate": 2.9217821782178217e-05, "loss": 0.0801, "step": 2109 }, { "epoch": 8.339920948616601, "grad_norm": 2.1677732467651367, "learning_rate": 2.9207920792079208e-05, "loss": 0.0704, "step": 2110 }, { "epoch": 8.34387351778656, "grad_norm": 1.7175079584121704, "learning_rate": 2.91980198019802e-05, "loss": 0.0548, "step": 2111 }, { "epoch": 8.347826086956522, "grad_norm": 2.6206259727478027, "learning_rate": 2.9188118811881187e-05, "loss": 0.0771, "step": 2112 }, { "epoch": 8.351778656126482, "grad_norm": 1.7399725914001465, "learning_rate": 2.9178217821782178e-05, "loss": 0.0633, "step": 2113 }, { "epoch": 8.355731225296443, "grad_norm": 2.0651071071624756, "learning_rate": 2.916831683168317e-05, "loss": 0.0638, "step": 2114 }, { "epoch": 8.359683794466402, "grad_norm": 1.7805978059768677, "learning_rate": 2.9158415841584157e-05, "loss": 0.0641, "step": 2115 }, { "epoch": 8.363636363636363, "grad_norm": 2.6731820106506348, "learning_rate": 2.9148514851485148e-05, "loss": 0.0809, "step": 2116 }, { "epoch": 8.367588932806324, "grad_norm": 1.7166242599487305, "learning_rate": 2.913861386138614e-05, "loss": 0.0479, "step": 2117 }, { "epoch": 8.371541501976285, "grad_norm": 1.8768218755722046, "learning_rate": 2.912871287128713e-05, "loss": 0.0679, "step": 2118 }, { "epoch": 8.375494071146244, "grad_norm": 1.6058732271194458, "learning_rate": 2.9118811881188118e-05, "loss": 0.0553, "step": 2119 }, { "epoch": 8.379446640316205, "grad_norm": 1.7579554319381714, "learning_rate": 2.910891089108911e-05, "loss": 0.0615, "step": 2120 }, { "epoch": 8.383399209486166, "grad_norm": 2.1073784828186035, "learning_rate": 2.90990099009901e-05, "loss": 0.0637, "step": 2121 }, { "epoch": 8.387351778656127, "grad_norm": 1.939737319946289, "learning_rate": 2.9089108910891087e-05, "loss": 0.0616, "step": 2122 }, { "epoch": 8.391304347826088, "grad_norm": 1.8796907663345337, "learning_rate": 2.907920792079208e-05, "loss": 0.0678, "step": 2123 }, { "epoch": 8.395256916996047, "grad_norm": 1.8378864526748657, "learning_rate": 2.906930693069307e-05, "loss": 0.0549, "step": 2124 }, { "epoch": 8.399209486166008, "grad_norm": 2.0537109375, "learning_rate": 2.9059405940594057e-05, "loss": 0.0642, "step": 2125 }, { "epoch": 8.403162055335969, "grad_norm": 1.9893449544906616, "learning_rate": 2.9049504950495048e-05, "loss": 0.0588, "step": 2126 }, { "epoch": 8.40711462450593, "grad_norm": 1.964008092880249, "learning_rate": 2.903960396039604e-05, "loss": 0.0662, "step": 2127 }, { "epoch": 8.411067193675889, "grad_norm": 1.9039689302444458, "learning_rate": 2.902970297029703e-05, "loss": 0.067, "step": 2128 }, { "epoch": 8.41501976284585, "grad_norm": 1.6828795671463013, "learning_rate": 2.9019801980198018e-05, "loss": 0.0536, "step": 2129 }, { "epoch": 8.41897233201581, "grad_norm": 1.2856837511062622, "learning_rate": 2.900990099009901e-05, "loss": 0.0451, "step": 2130 }, { "epoch": 8.422924901185771, "grad_norm": 2.3688838481903076, "learning_rate": 2.9e-05, "loss": 0.0692, "step": 2131 }, { "epoch": 8.42687747035573, "grad_norm": 1.792022705078125, "learning_rate": 2.8990099009900988e-05, "loss": 0.0565, "step": 2132 }, { "epoch": 8.430830039525691, "grad_norm": 1.6967265605926514, "learning_rate": 2.898019801980198e-05, "loss": 0.0595, "step": 2133 }, { "epoch": 8.434782608695652, "grad_norm": 2.171623468399048, "learning_rate": 2.897029702970297e-05, "loss": 0.073, "step": 2134 }, { "epoch": 8.438735177865613, "grad_norm": 2.2980382442474365, "learning_rate": 2.896039603960396e-05, "loss": 0.0639, "step": 2135 }, { "epoch": 8.442687747035572, "grad_norm": 2.255557060241699, "learning_rate": 2.895049504950495e-05, "loss": 0.0845, "step": 2136 }, { "epoch": 8.446640316205533, "grad_norm": 1.7793859243392944, "learning_rate": 2.894059405940594e-05, "loss": 0.0609, "step": 2137 }, { "epoch": 8.450592885375494, "grad_norm": 2.235987663269043, "learning_rate": 2.893069306930693e-05, "loss": 0.0965, "step": 2138 }, { "epoch": 8.454545454545455, "grad_norm": 1.8821008205413818, "learning_rate": 2.892079207920792e-05, "loss": 0.0647, "step": 2139 }, { "epoch": 8.458498023715414, "grad_norm": 2.27559232711792, "learning_rate": 2.891089108910891e-05, "loss": 0.0773, "step": 2140 }, { "epoch": 8.462450592885375, "grad_norm": 2.144597291946411, "learning_rate": 2.89009900990099e-05, "loss": 0.0742, "step": 2141 }, { "epoch": 8.466403162055336, "grad_norm": 1.8517370223999023, "learning_rate": 2.8891089108910892e-05, "loss": 0.0561, "step": 2142 }, { "epoch": 8.470355731225297, "grad_norm": 2.790147304534912, "learning_rate": 2.888118811881188e-05, "loss": 0.0818, "step": 2143 }, { "epoch": 8.474308300395258, "grad_norm": 1.7248456478118896, "learning_rate": 2.887128712871287e-05, "loss": 0.0608, "step": 2144 }, { "epoch": 8.478260869565217, "grad_norm": 1.803924322128296, "learning_rate": 2.886138613861386e-05, "loss": 0.0645, "step": 2145 }, { "epoch": 8.482213438735178, "grad_norm": 2.376180648803711, "learning_rate": 2.885148514851485e-05, "loss": 0.0878, "step": 2146 }, { "epoch": 8.486166007905139, "grad_norm": 2.218592882156372, "learning_rate": 2.884158415841584e-05, "loss": 0.0684, "step": 2147 }, { "epoch": 8.4901185770751, "grad_norm": 2.3117473125457764, "learning_rate": 2.883168316831683e-05, "loss": 0.0525, "step": 2148 }, { "epoch": 8.494071146245059, "grad_norm": 2.027350425720215, "learning_rate": 2.8821782178217822e-05, "loss": 0.0672, "step": 2149 }, { "epoch": 8.49802371541502, "grad_norm": 2.2401833534240723, "learning_rate": 2.881188118811881e-05, "loss": 0.0645, "step": 2150 }, { "epoch": 8.50197628458498, "grad_norm": 2.3592429161071777, "learning_rate": 2.88019801980198e-05, "loss": 0.0754, "step": 2151 }, { "epoch": 8.505928853754941, "grad_norm": 2.02468204498291, "learning_rate": 2.8792079207920792e-05, "loss": 0.064, "step": 2152 }, { "epoch": 8.5098814229249, "grad_norm": 1.9859503507614136, "learning_rate": 2.878217821782178e-05, "loss": 0.0672, "step": 2153 }, { "epoch": 8.513833992094861, "grad_norm": 1.9613122940063477, "learning_rate": 2.877227722772277e-05, "loss": 0.0654, "step": 2154 }, { "epoch": 8.517786561264822, "grad_norm": 1.9209879636764526, "learning_rate": 2.8762376237623762e-05, "loss": 0.0665, "step": 2155 }, { "epoch": 8.521739130434783, "grad_norm": 2.142094373703003, "learning_rate": 2.8752475247524753e-05, "loss": 0.0611, "step": 2156 }, { "epoch": 8.525691699604742, "grad_norm": 1.8080549240112305, "learning_rate": 2.874257425742574e-05, "loss": 0.0631, "step": 2157 }, { "epoch": 8.529644268774703, "grad_norm": 1.8271846771240234, "learning_rate": 2.8732673267326732e-05, "loss": 0.0674, "step": 2158 }, { "epoch": 8.533596837944664, "grad_norm": 1.6788458824157715, "learning_rate": 2.8722772277227723e-05, "loss": 0.0545, "step": 2159 }, { "epoch": 8.537549407114625, "grad_norm": 1.8324283361434937, "learning_rate": 2.871287128712871e-05, "loss": 0.0597, "step": 2160 }, { "epoch": 8.541501976284586, "grad_norm": 2.2549266815185547, "learning_rate": 2.87029702970297e-05, "loss": 0.0661, "step": 2161 }, { "epoch": 8.545454545454545, "grad_norm": 1.8250654935836792, "learning_rate": 2.8693069306930693e-05, "loss": 0.0594, "step": 2162 }, { "epoch": 8.549407114624506, "grad_norm": 1.6130670309066772, "learning_rate": 2.8683168316831684e-05, "loss": 0.0577, "step": 2163 }, { "epoch": 8.553359683794467, "grad_norm": 1.7462300062179565, "learning_rate": 2.867326732673267e-05, "loss": 0.0615, "step": 2164 }, { "epoch": 8.557312252964428, "grad_norm": 2.6029415130615234, "learning_rate": 2.8663366336633663e-05, "loss": 0.0877, "step": 2165 }, { "epoch": 8.561264822134387, "grad_norm": 1.7388004064559937, "learning_rate": 2.8653465346534654e-05, "loss": 0.0632, "step": 2166 }, { "epoch": 8.565217391304348, "grad_norm": 2.173640012741089, "learning_rate": 2.864356435643564e-05, "loss": 0.0681, "step": 2167 }, { "epoch": 8.569169960474309, "grad_norm": 2.15759539604187, "learning_rate": 2.8633663366336632e-05, "loss": 0.0687, "step": 2168 }, { "epoch": 8.57312252964427, "grad_norm": 2.238081932067871, "learning_rate": 2.8623762376237623e-05, "loss": 0.0765, "step": 2169 }, { "epoch": 8.577075098814229, "grad_norm": 2.358383893966675, "learning_rate": 2.8613861386138614e-05, "loss": 0.0709, "step": 2170 }, { "epoch": 8.58102766798419, "grad_norm": 2.113286018371582, "learning_rate": 2.8603960396039602e-05, "loss": 0.0746, "step": 2171 }, { "epoch": 8.58498023715415, "grad_norm": 2.401909112930298, "learning_rate": 2.8594059405940593e-05, "loss": 0.0691, "step": 2172 }, { "epoch": 8.588932806324111, "grad_norm": 1.56204354763031, "learning_rate": 2.8584158415841584e-05, "loss": 0.0516, "step": 2173 }, { "epoch": 8.59288537549407, "grad_norm": 1.9512985944747925, "learning_rate": 2.8574257425742572e-05, "loss": 0.0606, "step": 2174 }, { "epoch": 8.596837944664031, "grad_norm": 1.6974376440048218, "learning_rate": 2.8564356435643563e-05, "loss": 0.0609, "step": 2175 }, { "epoch": 8.600790513833992, "grad_norm": 2.388051748275757, "learning_rate": 2.8554455445544554e-05, "loss": 0.0804, "step": 2176 }, { "epoch": 8.604743083003953, "grad_norm": 1.7497717142105103, "learning_rate": 2.8544554455445545e-05, "loss": 0.0578, "step": 2177 }, { "epoch": 8.608695652173914, "grad_norm": 2.2503294944763184, "learning_rate": 2.8534653465346533e-05, "loss": 0.0751, "step": 2178 }, { "epoch": 8.612648221343873, "grad_norm": 2.230764389038086, "learning_rate": 2.8524752475247524e-05, "loss": 0.0773, "step": 2179 }, { "epoch": 8.616600790513834, "grad_norm": 1.9860472679138184, "learning_rate": 2.8514851485148515e-05, "loss": 0.0736, "step": 2180 }, { "epoch": 8.620553359683795, "grad_norm": 2.169675827026367, "learning_rate": 2.8504950495049503e-05, "loss": 0.0737, "step": 2181 }, { "epoch": 8.624505928853754, "grad_norm": 1.9016166925430298, "learning_rate": 2.8495049504950494e-05, "loss": 0.0757, "step": 2182 }, { "epoch": 8.628458498023715, "grad_norm": 1.9401284456253052, "learning_rate": 2.8485148514851485e-05, "loss": 0.0668, "step": 2183 }, { "epoch": 8.632411067193676, "grad_norm": 1.7244874238967896, "learning_rate": 2.8475247524752472e-05, "loss": 0.0592, "step": 2184 }, { "epoch": 8.636363636363637, "grad_norm": 2.6369059085845947, "learning_rate": 2.8465346534653464e-05, "loss": 0.0763, "step": 2185 }, { "epoch": 8.640316205533598, "grad_norm": 2.3759939670562744, "learning_rate": 2.8455445544554455e-05, "loss": 0.0792, "step": 2186 }, { "epoch": 8.644268774703557, "grad_norm": 2.084322690963745, "learning_rate": 2.8445544554455446e-05, "loss": 0.0706, "step": 2187 }, { "epoch": 8.648221343873518, "grad_norm": 1.7226157188415527, "learning_rate": 2.8435643564356433e-05, "loss": 0.0558, "step": 2188 }, { "epoch": 8.652173913043478, "grad_norm": 1.780396819114685, "learning_rate": 2.8425742574257424e-05, "loss": 0.0583, "step": 2189 }, { "epoch": 8.65612648221344, "grad_norm": 2.367802143096924, "learning_rate": 2.8415841584158415e-05, "loss": 0.0852, "step": 2190 }, { "epoch": 8.660079051383399, "grad_norm": 2.4490408897399902, "learning_rate": 2.8405940594059403e-05, "loss": 0.0826, "step": 2191 }, { "epoch": 8.66403162055336, "grad_norm": 1.8819794654846191, "learning_rate": 2.8396039603960394e-05, "loss": 0.0679, "step": 2192 }, { "epoch": 8.66798418972332, "grad_norm": 2.071528196334839, "learning_rate": 2.8386138613861385e-05, "loss": 0.0892, "step": 2193 }, { "epoch": 8.671936758893281, "grad_norm": 1.8685520887374878, "learning_rate": 2.8376237623762376e-05, "loss": 0.0582, "step": 2194 }, { "epoch": 8.67588932806324, "grad_norm": 1.8320471048355103, "learning_rate": 2.8366336633663364e-05, "loss": 0.0652, "step": 2195 }, { "epoch": 8.679841897233201, "grad_norm": 1.9017678499221802, "learning_rate": 2.8356435643564355e-05, "loss": 0.0622, "step": 2196 }, { "epoch": 8.683794466403162, "grad_norm": 2.227947235107422, "learning_rate": 2.8346534653465346e-05, "loss": 0.0667, "step": 2197 }, { "epoch": 8.687747035573123, "grad_norm": 2.120788812637329, "learning_rate": 2.8336633663366334e-05, "loss": 0.0787, "step": 2198 }, { "epoch": 8.691699604743082, "grad_norm": 1.4539971351623535, "learning_rate": 2.8326732673267325e-05, "loss": 0.0552, "step": 2199 }, { "epoch": 8.695652173913043, "grad_norm": 2.217979669570923, "learning_rate": 2.8316831683168316e-05, "loss": 0.0657, "step": 2200 }, { "epoch": 8.699604743083004, "grad_norm": 1.7687668800354004, "learning_rate": 2.8306930693069307e-05, "loss": 0.0619, "step": 2201 }, { "epoch": 8.703557312252965, "grad_norm": 1.532417893409729, "learning_rate": 2.8297029702970295e-05, "loss": 0.052, "step": 2202 }, { "epoch": 8.707509881422926, "grad_norm": 1.6064121723175049, "learning_rate": 2.8287128712871286e-05, "loss": 0.0604, "step": 2203 }, { "epoch": 8.711462450592885, "grad_norm": 1.8147326707839966, "learning_rate": 2.8277227722772277e-05, "loss": 0.0661, "step": 2204 }, { "epoch": 8.715415019762846, "grad_norm": 2.728243112564087, "learning_rate": 2.8267326732673265e-05, "loss": 0.054, "step": 2205 }, { "epoch": 8.719367588932807, "grad_norm": 1.5268528461456299, "learning_rate": 2.8257425742574256e-05, "loss": 0.0511, "step": 2206 }, { "epoch": 8.723320158102768, "grad_norm": 2.0269250869750977, "learning_rate": 2.8247524752475247e-05, "loss": 0.0603, "step": 2207 }, { "epoch": 8.727272727272727, "grad_norm": 2.630305290222168, "learning_rate": 2.8237623762376238e-05, "loss": 0.0844, "step": 2208 }, { "epoch": 8.731225296442688, "grad_norm": 2.1006577014923096, "learning_rate": 2.8227722772277225e-05, "loss": 0.0663, "step": 2209 }, { "epoch": 8.735177865612648, "grad_norm": 1.9232585430145264, "learning_rate": 2.8217821782178216e-05, "loss": 0.0631, "step": 2210 }, { "epoch": 8.73913043478261, "grad_norm": 2.4283599853515625, "learning_rate": 2.8207920792079208e-05, "loss": 0.0786, "step": 2211 }, { "epoch": 8.743083003952568, "grad_norm": 1.8329625129699707, "learning_rate": 2.8198019801980195e-05, "loss": 0.0631, "step": 2212 }, { "epoch": 8.74703557312253, "grad_norm": 2.143566846847534, "learning_rate": 2.8188118811881186e-05, "loss": 0.0701, "step": 2213 }, { "epoch": 8.75098814229249, "grad_norm": 1.968238353729248, "learning_rate": 2.8178217821782177e-05, "loss": 0.0699, "step": 2214 }, { "epoch": 8.754940711462451, "grad_norm": 2.1620233058929443, "learning_rate": 2.816831683168317e-05, "loss": 0.0688, "step": 2215 }, { "epoch": 8.75889328063241, "grad_norm": 1.9235700368881226, "learning_rate": 2.8158415841584156e-05, "loss": 0.0676, "step": 2216 }, { "epoch": 8.762845849802371, "grad_norm": 2.087008476257324, "learning_rate": 2.8148514851485147e-05, "loss": 0.0721, "step": 2217 }, { "epoch": 8.766798418972332, "grad_norm": 2.2739555835723877, "learning_rate": 2.8138613861386138e-05, "loss": 0.0879, "step": 2218 }, { "epoch": 8.770750988142293, "grad_norm": 2.0886847972869873, "learning_rate": 2.8128712871287126e-05, "loss": 0.0648, "step": 2219 }, { "epoch": 8.774703557312254, "grad_norm": 2.7450613975524902, "learning_rate": 2.8118811881188124e-05, "loss": 0.0877, "step": 2220 }, { "epoch": 8.778656126482213, "grad_norm": 2.2516679763793945, "learning_rate": 2.810891089108911e-05, "loss": 0.0747, "step": 2221 }, { "epoch": 8.782608695652174, "grad_norm": 2.5872914791107178, "learning_rate": 2.8099009900990102e-05, "loss": 0.0823, "step": 2222 }, { "epoch": 8.786561264822135, "grad_norm": 1.9789952039718628, "learning_rate": 2.8089108910891094e-05, "loss": 0.0714, "step": 2223 }, { "epoch": 8.790513833992096, "grad_norm": 2.561842441558838, "learning_rate": 2.8079207920792085e-05, "loss": 0.0732, "step": 2224 }, { "epoch": 8.794466403162055, "grad_norm": 2.0526421070098877, "learning_rate": 2.8069306930693072e-05, "loss": 0.0668, "step": 2225 }, { "epoch": 8.798418972332016, "grad_norm": 1.8163750171661377, "learning_rate": 2.8059405940594063e-05, "loss": 0.0654, "step": 2226 }, { "epoch": 8.802371541501977, "grad_norm": 2.070326089859009, "learning_rate": 2.8049504950495054e-05, "loss": 0.0737, "step": 2227 }, { "epoch": 8.806324110671937, "grad_norm": 2.309839963912964, "learning_rate": 2.8039603960396042e-05, "loss": 0.0835, "step": 2228 }, { "epoch": 8.810276679841897, "grad_norm": 1.6812690496444702, "learning_rate": 2.8029702970297033e-05, "loss": 0.0577, "step": 2229 }, { "epoch": 8.814229249011857, "grad_norm": 1.8067909479141235, "learning_rate": 2.8019801980198024e-05, "loss": 0.066, "step": 2230 }, { "epoch": 8.818181818181818, "grad_norm": 1.8315060138702393, "learning_rate": 2.8009900990099015e-05, "loss": 0.0745, "step": 2231 }, { "epoch": 8.82213438735178, "grad_norm": 1.9177123308181763, "learning_rate": 2.8000000000000003e-05, "loss": 0.0579, "step": 2232 }, { "epoch": 8.826086956521738, "grad_norm": 1.493560791015625, "learning_rate": 2.7990099009900994e-05, "loss": 0.0613, "step": 2233 }, { "epoch": 8.8300395256917, "grad_norm": 1.7907785177230835, "learning_rate": 2.7980198019801985e-05, "loss": 0.0581, "step": 2234 }, { "epoch": 8.83399209486166, "grad_norm": 2.810260772705078, "learning_rate": 2.7970297029702973e-05, "loss": 0.0836, "step": 2235 }, { "epoch": 8.837944664031621, "grad_norm": 1.983428716659546, "learning_rate": 2.7960396039603964e-05, "loss": 0.0735, "step": 2236 }, { "epoch": 8.841897233201582, "grad_norm": 1.9230412244796753, "learning_rate": 2.7950495049504955e-05, "loss": 0.0609, "step": 2237 }, { "epoch": 8.845849802371541, "grad_norm": 1.669242262840271, "learning_rate": 2.7940594059405946e-05, "loss": 0.0683, "step": 2238 }, { "epoch": 8.849802371541502, "grad_norm": 2.4905624389648438, "learning_rate": 2.7930693069306934e-05, "loss": 0.0921, "step": 2239 }, { "epoch": 8.853754940711463, "grad_norm": 1.8049850463867188, "learning_rate": 2.7920792079207925e-05, "loss": 0.0656, "step": 2240 }, { "epoch": 8.857707509881424, "grad_norm": 1.6094136238098145, "learning_rate": 2.7910891089108916e-05, "loss": 0.0555, "step": 2241 }, { "epoch": 8.861660079051383, "grad_norm": 1.8134102821350098, "learning_rate": 2.7900990099009903e-05, "loss": 0.0624, "step": 2242 }, { "epoch": 8.865612648221344, "grad_norm": 2.0376391410827637, "learning_rate": 2.7891089108910895e-05, "loss": 0.0599, "step": 2243 }, { "epoch": 8.869565217391305, "grad_norm": 2.2510569095611572, "learning_rate": 2.7881188118811886e-05, "loss": 0.0748, "step": 2244 }, { "epoch": 8.873517786561266, "grad_norm": 1.732476830482483, "learning_rate": 2.7871287128712877e-05, "loss": 0.0657, "step": 2245 }, { "epoch": 8.877470355731225, "grad_norm": 1.9895007610321045, "learning_rate": 2.7861386138613864e-05, "loss": 0.0655, "step": 2246 }, { "epoch": 8.881422924901186, "grad_norm": 1.7584738731384277, "learning_rate": 2.7851485148514855e-05, "loss": 0.0619, "step": 2247 }, { "epoch": 8.885375494071146, "grad_norm": 2.5343644618988037, "learning_rate": 2.7841584158415846e-05, "loss": 0.0806, "step": 2248 }, { "epoch": 8.889328063241107, "grad_norm": 2.2883429527282715, "learning_rate": 2.7831683168316834e-05, "loss": 0.0771, "step": 2249 }, { "epoch": 8.893280632411066, "grad_norm": 2.2479300498962402, "learning_rate": 2.7821782178217825e-05, "loss": 0.07, "step": 2250 }, { "epoch": 8.897233201581027, "grad_norm": 1.7287509441375732, "learning_rate": 2.7811881188118816e-05, "loss": 0.0534, "step": 2251 }, { "epoch": 8.901185770750988, "grad_norm": 2.4114739894866943, "learning_rate": 2.7801980198019807e-05, "loss": 0.0843, "step": 2252 }, { "epoch": 8.90513833992095, "grad_norm": 2.1718227863311768, "learning_rate": 2.7792079207920795e-05, "loss": 0.0725, "step": 2253 }, { "epoch": 8.909090909090908, "grad_norm": 1.761806607246399, "learning_rate": 2.7782178217821786e-05, "loss": 0.0713, "step": 2254 }, { "epoch": 8.91304347826087, "grad_norm": 1.481679081916809, "learning_rate": 2.7772277227722777e-05, "loss": 0.0454, "step": 2255 }, { "epoch": 8.91699604743083, "grad_norm": 1.8649990558624268, "learning_rate": 2.7762376237623765e-05, "loss": 0.0693, "step": 2256 }, { "epoch": 8.920948616600791, "grad_norm": 2.2529780864715576, "learning_rate": 2.7752475247524756e-05, "loss": 0.0743, "step": 2257 }, { "epoch": 8.92490118577075, "grad_norm": 1.5071758031845093, "learning_rate": 2.7742574257425747e-05, "loss": 0.0487, "step": 2258 }, { "epoch": 8.928853754940711, "grad_norm": 2.5997564792633057, "learning_rate": 2.7732673267326738e-05, "loss": 0.0741, "step": 2259 }, { "epoch": 8.932806324110672, "grad_norm": 2.572242259979248, "learning_rate": 2.7722772277227726e-05, "loss": 0.0776, "step": 2260 }, { "epoch": 8.936758893280633, "grad_norm": 2.599885940551758, "learning_rate": 2.7712871287128717e-05, "loss": 0.0896, "step": 2261 }, { "epoch": 8.940711462450594, "grad_norm": 2.559628963470459, "learning_rate": 2.7702970297029708e-05, "loss": 0.0683, "step": 2262 }, { "epoch": 8.944664031620553, "grad_norm": 3.259251356124878, "learning_rate": 2.7693069306930696e-05, "loss": 0.1187, "step": 2263 }, { "epoch": 8.948616600790514, "grad_norm": 1.5515271425247192, "learning_rate": 2.7683168316831687e-05, "loss": 0.0581, "step": 2264 }, { "epoch": 8.952569169960475, "grad_norm": 1.853208303451538, "learning_rate": 2.7673267326732678e-05, "loss": 0.0624, "step": 2265 }, { "epoch": 8.956521739130435, "grad_norm": 1.8357259035110474, "learning_rate": 2.766336633663367e-05, "loss": 0.0707, "step": 2266 }, { "epoch": 8.960474308300395, "grad_norm": 1.9865976572036743, "learning_rate": 2.7653465346534656e-05, "loss": 0.0588, "step": 2267 }, { "epoch": 8.964426877470355, "grad_norm": 2.2406439781188965, "learning_rate": 2.7643564356435647e-05, "loss": 0.0828, "step": 2268 }, { "epoch": 8.968379446640316, "grad_norm": 1.8917419910430908, "learning_rate": 2.763366336633664e-05, "loss": 0.0657, "step": 2269 }, { "epoch": 8.972332015810277, "grad_norm": 2.238126039505005, "learning_rate": 2.7623762376237626e-05, "loss": 0.0716, "step": 2270 }, { "epoch": 8.976284584980236, "grad_norm": 1.551047682762146, "learning_rate": 2.7613861386138617e-05, "loss": 0.0599, "step": 2271 }, { "epoch": 8.980237154150197, "grad_norm": 1.6375608444213867, "learning_rate": 2.760396039603961e-05, "loss": 0.0558, "step": 2272 }, { "epoch": 8.984189723320158, "grad_norm": 2.108804702758789, "learning_rate": 2.7594059405940596e-05, "loss": 0.0786, "step": 2273 }, { "epoch": 8.988142292490119, "grad_norm": 1.8367949724197388, "learning_rate": 2.7584158415841587e-05, "loss": 0.0659, "step": 2274 }, { "epoch": 8.992094861660078, "grad_norm": 2.4967141151428223, "learning_rate": 2.7574257425742578e-05, "loss": 0.081, "step": 2275 }, { "epoch": 8.996047430830039, "grad_norm": 1.7615149021148682, "learning_rate": 2.756435643564357e-05, "loss": 0.0635, "step": 2276 }, { "epoch": 9.0, "grad_norm": 2.454653263092041, "learning_rate": 2.7554455445544557e-05, "loss": 0.0815, "step": 2277 }, { "epoch": 9.003952569169961, "grad_norm": 1.698432445526123, "learning_rate": 2.7544554455445548e-05, "loss": 0.0514, "step": 2278 }, { "epoch": 9.007905138339922, "grad_norm": 1.4801515340805054, "learning_rate": 2.753465346534654e-05, "loss": 0.0478, "step": 2279 }, { "epoch": 9.011857707509881, "grad_norm": 1.8764443397521973, "learning_rate": 2.7524752475247527e-05, "loss": 0.0658, "step": 2280 }, { "epoch": 9.015810276679842, "grad_norm": 1.2528640031814575, "learning_rate": 2.7514851485148518e-05, "loss": 0.0447, "step": 2281 }, { "epoch": 9.019762845849803, "grad_norm": 1.3641486167907715, "learning_rate": 2.750495049504951e-05, "loss": 0.0475, "step": 2282 }, { "epoch": 9.023715415019764, "grad_norm": 1.0674891471862793, "learning_rate": 2.74950495049505e-05, "loss": 0.0328, "step": 2283 }, { "epoch": 9.027667984189723, "grad_norm": 1.1281448602676392, "learning_rate": 2.7485148514851488e-05, "loss": 0.0349, "step": 2284 }, { "epoch": 9.031620553359684, "grad_norm": 2.090841054916382, "learning_rate": 2.747524752475248e-05, "loss": 0.0624, "step": 2285 }, { "epoch": 9.035573122529645, "grad_norm": 1.7504425048828125, "learning_rate": 2.746534653465347e-05, "loss": 0.0529, "step": 2286 }, { "epoch": 9.039525691699605, "grad_norm": 1.756168007850647, "learning_rate": 2.7455445544554457e-05, "loss": 0.0491, "step": 2287 }, { "epoch": 9.043478260869565, "grad_norm": 1.6505038738250732, "learning_rate": 2.744554455445545e-05, "loss": 0.0467, "step": 2288 }, { "epoch": 9.047430830039525, "grad_norm": 1.3603137731552124, "learning_rate": 2.743564356435644e-05, "loss": 0.0375, "step": 2289 }, { "epoch": 9.051383399209486, "grad_norm": 1.2336719036102295, "learning_rate": 2.742574257425743e-05, "loss": 0.0411, "step": 2290 }, { "epoch": 9.055335968379447, "grad_norm": 1.6150755882263184, "learning_rate": 2.7415841584158418e-05, "loss": 0.0442, "step": 2291 }, { "epoch": 9.059288537549406, "grad_norm": 1.8003437519073486, "learning_rate": 2.740594059405941e-05, "loss": 0.0508, "step": 2292 }, { "epoch": 9.063241106719367, "grad_norm": 1.4279950857162476, "learning_rate": 2.73960396039604e-05, "loss": 0.0417, "step": 2293 }, { "epoch": 9.067193675889328, "grad_norm": 1.4173280000686646, "learning_rate": 2.7386138613861388e-05, "loss": 0.0413, "step": 2294 }, { "epoch": 9.071146245059289, "grad_norm": 2.2572765350341797, "learning_rate": 2.737623762376238e-05, "loss": 0.0663, "step": 2295 }, { "epoch": 9.075098814229248, "grad_norm": 1.2516882419586182, "learning_rate": 2.736633663366337e-05, "loss": 0.0367, "step": 2296 }, { "epoch": 9.079051383399209, "grad_norm": 1.96310293674469, "learning_rate": 2.735643564356436e-05, "loss": 0.0541, "step": 2297 }, { "epoch": 9.08300395256917, "grad_norm": 1.9568673372268677, "learning_rate": 2.734653465346535e-05, "loss": 0.058, "step": 2298 }, { "epoch": 9.08695652173913, "grad_norm": 1.7115801572799683, "learning_rate": 2.733663366336634e-05, "loss": 0.0532, "step": 2299 }, { "epoch": 9.090909090909092, "grad_norm": 1.4340516328811646, "learning_rate": 2.732673267326733e-05, "loss": 0.0391, "step": 2300 }, { "epoch": 9.09486166007905, "grad_norm": 1.738873839378357, "learning_rate": 2.731683168316832e-05, "loss": 0.0529, "step": 2301 }, { "epoch": 9.098814229249012, "grad_norm": 2.0912539958953857, "learning_rate": 2.730693069306931e-05, "loss": 0.066, "step": 2302 }, { "epoch": 9.102766798418973, "grad_norm": 1.6749258041381836, "learning_rate": 2.72970297029703e-05, "loss": 0.0482, "step": 2303 }, { "epoch": 9.106719367588934, "grad_norm": 2.200265645980835, "learning_rate": 2.7287128712871292e-05, "loss": 0.0656, "step": 2304 }, { "epoch": 9.110671936758893, "grad_norm": 1.9086296558380127, "learning_rate": 2.727722772277228e-05, "loss": 0.0586, "step": 2305 }, { "epoch": 9.114624505928854, "grad_norm": 1.4664795398712158, "learning_rate": 2.726732673267327e-05, "loss": 0.0452, "step": 2306 }, { "epoch": 9.118577075098814, "grad_norm": 1.438987135887146, "learning_rate": 2.7257425742574262e-05, "loss": 0.0432, "step": 2307 }, { "epoch": 9.122529644268775, "grad_norm": 1.403167486190796, "learning_rate": 2.724752475247525e-05, "loss": 0.0455, "step": 2308 }, { "epoch": 9.126482213438734, "grad_norm": 1.7621055841445923, "learning_rate": 2.723762376237624e-05, "loss": 0.05, "step": 2309 }, { "epoch": 9.130434782608695, "grad_norm": 2.0378241539001465, "learning_rate": 2.722772277227723e-05, "loss": 0.0787, "step": 2310 }, { "epoch": 9.134387351778656, "grad_norm": 1.6555578708648682, "learning_rate": 2.7217821782178223e-05, "loss": 0.0472, "step": 2311 }, { "epoch": 9.138339920948617, "grad_norm": 1.5165730714797974, "learning_rate": 2.720792079207921e-05, "loss": 0.0415, "step": 2312 }, { "epoch": 9.142292490118576, "grad_norm": 1.9997520446777344, "learning_rate": 2.71980198019802e-05, "loss": 0.0588, "step": 2313 }, { "epoch": 9.146245059288537, "grad_norm": 1.41425621509552, "learning_rate": 2.7188118811881192e-05, "loss": 0.0447, "step": 2314 }, { "epoch": 9.150197628458498, "grad_norm": 1.7173479795455933, "learning_rate": 2.717821782178218e-05, "loss": 0.0556, "step": 2315 }, { "epoch": 9.154150197628459, "grad_norm": 1.5232248306274414, "learning_rate": 2.716831683168317e-05, "loss": 0.0473, "step": 2316 }, { "epoch": 9.15810276679842, "grad_norm": 1.5574851036071777, "learning_rate": 2.7158415841584162e-05, "loss": 0.0503, "step": 2317 }, { "epoch": 9.162055335968379, "grad_norm": 2.0634102821350098, "learning_rate": 2.7148514851485153e-05, "loss": 0.0676, "step": 2318 }, { "epoch": 9.16600790513834, "grad_norm": 1.3666168451309204, "learning_rate": 2.713861386138614e-05, "loss": 0.0459, "step": 2319 }, { "epoch": 9.1699604743083, "grad_norm": 2.817124843597412, "learning_rate": 2.7128712871287132e-05, "loss": 0.0849, "step": 2320 }, { "epoch": 9.173913043478262, "grad_norm": 1.6400901079177856, "learning_rate": 2.7118811881188123e-05, "loss": 0.0541, "step": 2321 }, { "epoch": 9.17786561264822, "grad_norm": 1.229610800743103, "learning_rate": 2.710891089108911e-05, "loss": 0.0415, "step": 2322 }, { "epoch": 9.181818181818182, "grad_norm": 1.9802181720733643, "learning_rate": 2.7099009900990102e-05, "loss": 0.0535, "step": 2323 }, { "epoch": 9.185770750988143, "grad_norm": 1.6513592004776, "learning_rate": 2.7089108910891093e-05, "loss": 0.0515, "step": 2324 }, { "epoch": 9.189723320158103, "grad_norm": 1.6192011833190918, "learning_rate": 2.7079207920792084e-05, "loss": 0.0563, "step": 2325 }, { "epoch": 9.193675889328063, "grad_norm": 1.4913856983184814, "learning_rate": 2.706930693069307e-05, "loss": 0.041, "step": 2326 }, { "epoch": 9.197628458498023, "grad_norm": 1.1605242490768433, "learning_rate": 2.7059405940594063e-05, "loss": 0.0355, "step": 2327 }, { "epoch": 9.201581027667984, "grad_norm": 1.3722583055496216, "learning_rate": 2.7049504950495054e-05, "loss": 0.0477, "step": 2328 }, { "epoch": 9.205533596837945, "grad_norm": 1.521227478981018, "learning_rate": 2.703960396039604e-05, "loss": 0.0497, "step": 2329 }, { "epoch": 9.209486166007904, "grad_norm": 2.0504047870635986, "learning_rate": 2.7029702970297033e-05, "loss": 0.0639, "step": 2330 }, { "epoch": 9.213438735177865, "grad_norm": 2.064445734024048, "learning_rate": 2.7019801980198024e-05, "loss": 0.0565, "step": 2331 }, { "epoch": 9.217391304347826, "grad_norm": 1.2652782201766968, "learning_rate": 2.700990099009901e-05, "loss": 0.0407, "step": 2332 }, { "epoch": 9.221343873517787, "grad_norm": 1.5517112016677856, "learning_rate": 2.7000000000000002e-05, "loss": 0.052, "step": 2333 }, { "epoch": 9.225296442687746, "grad_norm": 1.618432879447937, "learning_rate": 2.6990099009900993e-05, "loss": 0.0517, "step": 2334 }, { "epoch": 9.229249011857707, "grad_norm": 1.469072937965393, "learning_rate": 2.6980198019801985e-05, "loss": 0.0551, "step": 2335 }, { "epoch": 9.233201581027668, "grad_norm": 1.5397303104400635, "learning_rate": 2.6970297029702972e-05, "loss": 0.0449, "step": 2336 }, { "epoch": 9.237154150197629, "grad_norm": 2.483172655105591, "learning_rate": 2.6960396039603963e-05, "loss": 0.0723, "step": 2337 }, { "epoch": 9.24110671936759, "grad_norm": 1.998645544052124, "learning_rate": 2.6950495049504954e-05, "loss": 0.0633, "step": 2338 }, { "epoch": 9.245059288537549, "grad_norm": 1.3025997877120972, "learning_rate": 2.6940594059405942e-05, "loss": 0.0417, "step": 2339 }, { "epoch": 9.24901185770751, "grad_norm": 1.6129134893417358, "learning_rate": 2.6930693069306933e-05, "loss": 0.0503, "step": 2340 }, { "epoch": 9.25296442687747, "grad_norm": 1.6566739082336426, "learning_rate": 2.6920792079207924e-05, "loss": 0.0533, "step": 2341 }, { "epoch": 9.256916996047432, "grad_norm": 1.544024109840393, "learning_rate": 2.6910891089108915e-05, "loss": 0.0505, "step": 2342 }, { "epoch": 9.26086956521739, "grad_norm": 1.3488221168518066, "learning_rate": 2.6900990099009903e-05, "loss": 0.0427, "step": 2343 }, { "epoch": 9.264822134387352, "grad_norm": 2.2433106899261475, "learning_rate": 2.6891089108910894e-05, "loss": 0.0839, "step": 2344 }, { "epoch": 9.268774703557312, "grad_norm": 1.532610535621643, "learning_rate": 2.6881188118811885e-05, "loss": 0.0497, "step": 2345 }, { "epoch": 9.272727272727273, "grad_norm": 1.4470645189285278, "learning_rate": 2.6871287128712873e-05, "loss": 0.0466, "step": 2346 }, { "epoch": 9.276679841897232, "grad_norm": 1.8333078622817993, "learning_rate": 2.6861386138613864e-05, "loss": 0.0536, "step": 2347 }, { "epoch": 9.280632411067193, "grad_norm": 2.167539119720459, "learning_rate": 2.6851485148514855e-05, "loss": 0.0618, "step": 2348 }, { "epoch": 9.284584980237154, "grad_norm": 1.700262188911438, "learning_rate": 2.6841584158415846e-05, "loss": 0.0545, "step": 2349 }, { "epoch": 9.288537549407115, "grad_norm": 1.893081545829773, "learning_rate": 2.6831683168316834e-05, "loss": 0.0586, "step": 2350 }, { "epoch": 9.292490118577074, "grad_norm": 1.2898362874984741, "learning_rate": 2.6821782178217825e-05, "loss": 0.0465, "step": 2351 }, { "epoch": 9.296442687747035, "grad_norm": 1.1584510803222656, "learning_rate": 2.6811881188118816e-05, "loss": 0.037, "step": 2352 }, { "epoch": 9.300395256916996, "grad_norm": 1.8713995218276978, "learning_rate": 2.6801980198019803e-05, "loss": 0.0563, "step": 2353 }, { "epoch": 9.304347826086957, "grad_norm": 1.8590437173843384, "learning_rate": 2.6792079207920794e-05, "loss": 0.0605, "step": 2354 }, { "epoch": 9.308300395256918, "grad_norm": 1.5611218214035034, "learning_rate": 2.6782178217821786e-05, "loss": 0.054, "step": 2355 }, { "epoch": 9.312252964426877, "grad_norm": 1.3695837259292603, "learning_rate": 2.6772277227722777e-05, "loss": 0.0392, "step": 2356 }, { "epoch": 9.316205533596838, "grad_norm": 1.911346673965454, "learning_rate": 2.6762376237623764e-05, "loss": 0.0456, "step": 2357 }, { "epoch": 9.320158102766799, "grad_norm": 2.092742681503296, "learning_rate": 2.6752475247524755e-05, "loss": 0.0602, "step": 2358 }, { "epoch": 9.32411067193676, "grad_norm": 1.3872787952423096, "learning_rate": 2.6742574257425746e-05, "loss": 0.0457, "step": 2359 }, { "epoch": 9.328063241106719, "grad_norm": 1.78934907913208, "learning_rate": 2.6732673267326734e-05, "loss": 0.0579, "step": 2360 }, { "epoch": 9.33201581027668, "grad_norm": 1.8842891454696655, "learning_rate": 2.6722772277227725e-05, "loss": 0.0581, "step": 2361 }, { "epoch": 9.33596837944664, "grad_norm": 2.0135138034820557, "learning_rate": 2.6712871287128716e-05, "loss": 0.0678, "step": 2362 }, { "epoch": 9.339920948616601, "grad_norm": 1.6698349714279175, "learning_rate": 2.6702970297029707e-05, "loss": 0.0601, "step": 2363 }, { "epoch": 9.34387351778656, "grad_norm": 1.4716256856918335, "learning_rate": 2.6693069306930695e-05, "loss": 0.042, "step": 2364 }, { "epoch": 9.347826086956522, "grad_norm": 1.8310630321502686, "learning_rate": 2.6683168316831686e-05, "loss": 0.0598, "step": 2365 }, { "epoch": 9.351778656126482, "grad_norm": 3.6523401737213135, "learning_rate": 2.6673267326732677e-05, "loss": 0.0599, "step": 2366 }, { "epoch": 9.355731225296443, "grad_norm": 2.1205625534057617, "learning_rate": 2.6663366336633665e-05, "loss": 0.0604, "step": 2367 }, { "epoch": 9.359683794466402, "grad_norm": 1.7760454416275024, "learning_rate": 2.6653465346534656e-05, "loss": 0.0522, "step": 2368 }, { "epoch": 9.363636363636363, "grad_norm": 1.3189547061920166, "learning_rate": 2.6643564356435647e-05, "loss": 0.0434, "step": 2369 }, { "epoch": 9.367588932806324, "grad_norm": 1.8187811374664307, "learning_rate": 2.6633663366336638e-05, "loss": 0.0522, "step": 2370 }, { "epoch": 9.371541501976285, "grad_norm": 1.8806556463241577, "learning_rate": 2.6623762376237626e-05, "loss": 0.064, "step": 2371 }, { "epoch": 9.375494071146244, "grad_norm": 1.6149438619613647, "learning_rate": 2.6613861386138617e-05, "loss": 0.0447, "step": 2372 }, { "epoch": 9.379446640316205, "grad_norm": 1.5469831228256226, "learning_rate": 2.6603960396039608e-05, "loss": 0.0446, "step": 2373 }, { "epoch": 9.383399209486166, "grad_norm": 1.8596569299697876, "learning_rate": 2.6594059405940595e-05, "loss": 0.0509, "step": 2374 }, { "epoch": 9.387351778656127, "grad_norm": 1.8015838861465454, "learning_rate": 2.6584158415841586e-05, "loss": 0.0523, "step": 2375 }, { "epoch": 9.391304347826088, "grad_norm": 1.7738046646118164, "learning_rate": 2.6574257425742578e-05, "loss": 0.0498, "step": 2376 }, { "epoch": 9.395256916996047, "grad_norm": 1.5252565145492554, "learning_rate": 2.656435643564357e-05, "loss": 0.0535, "step": 2377 }, { "epoch": 9.399209486166008, "grad_norm": 1.8083176612854004, "learning_rate": 2.6554455445544556e-05, "loss": 0.0561, "step": 2378 }, { "epoch": 9.403162055335969, "grad_norm": 1.8712735176086426, "learning_rate": 2.6544554455445547e-05, "loss": 0.0569, "step": 2379 }, { "epoch": 9.40711462450593, "grad_norm": 2.3082659244537354, "learning_rate": 2.653465346534654e-05, "loss": 0.0771, "step": 2380 }, { "epoch": 9.411067193675889, "grad_norm": 2.0071911811828613, "learning_rate": 2.6524752475247526e-05, "loss": 0.0632, "step": 2381 }, { "epoch": 9.41501976284585, "grad_norm": 1.2642353773117065, "learning_rate": 2.6514851485148517e-05, "loss": 0.0385, "step": 2382 }, { "epoch": 9.41897233201581, "grad_norm": 1.4350475072860718, "learning_rate": 2.6504950495049508e-05, "loss": 0.0504, "step": 2383 }, { "epoch": 9.422924901185771, "grad_norm": 1.63456130027771, "learning_rate": 2.64950495049505e-05, "loss": 0.0494, "step": 2384 }, { "epoch": 9.42687747035573, "grad_norm": 1.9926373958587646, "learning_rate": 2.6485148514851487e-05, "loss": 0.0642, "step": 2385 }, { "epoch": 9.430830039525691, "grad_norm": 1.9030565023422241, "learning_rate": 2.6475247524752478e-05, "loss": 0.0563, "step": 2386 }, { "epoch": 9.434782608695652, "grad_norm": 1.857448697090149, "learning_rate": 2.646534653465347e-05, "loss": 0.0579, "step": 2387 }, { "epoch": 9.438735177865613, "grad_norm": 1.9808776378631592, "learning_rate": 2.6455445544554457e-05, "loss": 0.0693, "step": 2388 }, { "epoch": 9.442687747035572, "grad_norm": 1.169295072555542, "learning_rate": 2.6445544554455448e-05, "loss": 0.0392, "step": 2389 }, { "epoch": 9.446640316205533, "grad_norm": 1.9987329244613647, "learning_rate": 2.643564356435644e-05, "loss": 0.0591, "step": 2390 }, { "epoch": 9.450592885375494, "grad_norm": 1.9153774976730347, "learning_rate": 2.6425742574257427e-05, "loss": 0.0567, "step": 2391 }, { "epoch": 9.454545454545455, "grad_norm": 2.1307449340820312, "learning_rate": 2.6415841584158418e-05, "loss": 0.0658, "step": 2392 }, { "epoch": 9.458498023715414, "grad_norm": 2.101557493209839, "learning_rate": 2.640594059405941e-05, "loss": 0.0626, "step": 2393 }, { "epoch": 9.462450592885375, "grad_norm": 1.3687713146209717, "learning_rate": 2.63960396039604e-05, "loss": 0.0442, "step": 2394 }, { "epoch": 9.466403162055336, "grad_norm": 1.5858515501022339, "learning_rate": 2.6386138613861387e-05, "loss": 0.0467, "step": 2395 }, { "epoch": 9.470355731225297, "grad_norm": 1.214250087738037, "learning_rate": 2.637623762376238e-05, "loss": 0.0395, "step": 2396 }, { "epoch": 9.474308300395258, "grad_norm": 1.6746623516082764, "learning_rate": 2.636633663366337e-05, "loss": 0.0561, "step": 2397 }, { "epoch": 9.478260869565217, "grad_norm": 1.8827106952667236, "learning_rate": 2.6356435643564357e-05, "loss": 0.0558, "step": 2398 }, { "epoch": 9.482213438735178, "grad_norm": 1.1269687414169312, "learning_rate": 2.634653465346535e-05, "loss": 0.0475, "step": 2399 }, { "epoch": 9.486166007905139, "grad_norm": 1.8919798135757446, "learning_rate": 2.633663366336634e-05, "loss": 0.0622, "step": 2400 }, { "epoch": 9.4901185770751, "grad_norm": 1.4680562019348145, "learning_rate": 2.632673267326733e-05, "loss": 0.0414, "step": 2401 }, { "epoch": 9.494071146245059, "grad_norm": 1.3922396898269653, "learning_rate": 2.6316831683168318e-05, "loss": 0.0466, "step": 2402 }, { "epoch": 9.49802371541502, "grad_norm": 1.142188549041748, "learning_rate": 2.630693069306931e-05, "loss": 0.037, "step": 2403 }, { "epoch": 9.50197628458498, "grad_norm": 2.170102834701538, "learning_rate": 2.62970297029703e-05, "loss": 0.0596, "step": 2404 }, { "epoch": 9.505928853754941, "grad_norm": 1.5014910697937012, "learning_rate": 2.6287128712871288e-05, "loss": 0.0482, "step": 2405 }, { "epoch": 9.5098814229249, "grad_norm": 1.5824153423309326, "learning_rate": 2.627722772277228e-05, "loss": 0.0495, "step": 2406 }, { "epoch": 9.513833992094861, "grad_norm": 1.746131181716919, "learning_rate": 2.626732673267327e-05, "loss": 0.0514, "step": 2407 }, { "epoch": 9.517786561264822, "grad_norm": 1.5390735864639282, "learning_rate": 2.625742574257426e-05, "loss": 0.0492, "step": 2408 }, { "epoch": 9.521739130434783, "grad_norm": 1.6376750469207764, "learning_rate": 2.624752475247525e-05, "loss": 0.0492, "step": 2409 }, { "epoch": 9.525691699604742, "grad_norm": 2.177889108657837, "learning_rate": 2.623762376237624e-05, "loss": 0.0723, "step": 2410 }, { "epoch": 9.529644268774703, "grad_norm": 1.4220775365829468, "learning_rate": 2.622772277227723e-05, "loss": 0.0453, "step": 2411 }, { "epoch": 9.533596837944664, "grad_norm": 2.188873291015625, "learning_rate": 2.621782178217822e-05, "loss": 0.0645, "step": 2412 }, { "epoch": 9.537549407114625, "grad_norm": 2.0508599281311035, "learning_rate": 2.620792079207921e-05, "loss": 0.056, "step": 2413 }, { "epoch": 9.541501976284586, "grad_norm": 1.5632494688034058, "learning_rate": 2.61980198019802e-05, "loss": 0.0516, "step": 2414 }, { "epoch": 9.545454545454545, "grad_norm": 1.8457725048065186, "learning_rate": 2.6188118811881192e-05, "loss": 0.0554, "step": 2415 }, { "epoch": 9.549407114624506, "grad_norm": 1.8504745960235596, "learning_rate": 2.617821782178218e-05, "loss": 0.0602, "step": 2416 }, { "epoch": 9.553359683794467, "grad_norm": 1.4297717809677124, "learning_rate": 2.616831683168317e-05, "loss": 0.0431, "step": 2417 }, { "epoch": 9.557312252964428, "grad_norm": 1.7277193069458008, "learning_rate": 2.615841584158416e-05, "loss": 0.0458, "step": 2418 }, { "epoch": 9.561264822134387, "grad_norm": 1.4764007329940796, "learning_rate": 2.614851485148515e-05, "loss": 0.0505, "step": 2419 }, { "epoch": 9.565217391304348, "grad_norm": 1.819220781326294, "learning_rate": 2.613861386138614e-05, "loss": 0.0556, "step": 2420 }, { "epoch": 9.569169960474309, "grad_norm": 1.847817301750183, "learning_rate": 2.612871287128713e-05, "loss": 0.0564, "step": 2421 }, { "epoch": 9.57312252964427, "grad_norm": 1.9011605978012085, "learning_rate": 2.6118811881188123e-05, "loss": 0.0544, "step": 2422 }, { "epoch": 9.577075098814229, "grad_norm": 1.859048843383789, "learning_rate": 2.610891089108911e-05, "loss": 0.0596, "step": 2423 }, { "epoch": 9.58102766798419, "grad_norm": 2.11446213722229, "learning_rate": 2.60990099009901e-05, "loss": 0.0647, "step": 2424 }, { "epoch": 9.58498023715415, "grad_norm": 1.3351813554763794, "learning_rate": 2.6089108910891092e-05, "loss": 0.0446, "step": 2425 }, { "epoch": 9.588932806324111, "grad_norm": 1.7789154052734375, "learning_rate": 2.607920792079208e-05, "loss": 0.0549, "step": 2426 }, { "epoch": 9.59288537549407, "grad_norm": 1.915568470954895, "learning_rate": 2.606930693069307e-05, "loss": 0.0594, "step": 2427 }, { "epoch": 9.596837944664031, "grad_norm": 1.6413213014602661, "learning_rate": 2.6059405940594062e-05, "loss": 0.0582, "step": 2428 }, { "epoch": 9.600790513833992, "grad_norm": 1.8496522903442383, "learning_rate": 2.6049504950495053e-05, "loss": 0.0637, "step": 2429 }, { "epoch": 9.604743083003953, "grad_norm": 1.4442410469055176, "learning_rate": 2.603960396039604e-05, "loss": 0.0462, "step": 2430 }, { "epoch": 9.608695652173914, "grad_norm": 2.3356995582580566, "learning_rate": 2.6029702970297032e-05, "loss": 0.0573, "step": 2431 }, { "epoch": 9.612648221343873, "grad_norm": 1.9251607656478882, "learning_rate": 2.6019801980198023e-05, "loss": 0.0648, "step": 2432 }, { "epoch": 9.616600790513834, "grad_norm": 1.7375855445861816, "learning_rate": 2.600990099009901e-05, "loss": 0.0521, "step": 2433 }, { "epoch": 9.620553359683795, "grad_norm": 1.7185657024383545, "learning_rate": 2.6000000000000002e-05, "loss": 0.0547, "step": 2434 }, { "epoch": 9.624505928853754, "grad_norm": 1.5124340057373047, "learning_rate": 2.5990099009900993e-05, "loss": 0.0432, "step": 2435 }, { "epoch": 9.628458498023715, "grad_norm": 1.7628847360610962, "learning_rate": 2.5980198019801984e-05, "loss": 0.0542, "step": 2436 }, { "epoch": 9.632411067193676, "grad_norm": 1.6523082256317139, "learning_rate": 2.597029702970297e-05, "loss": 0.0522, "step": 2437 }, { "epoch": 9.636363636363637, "grad_norm": 1.4191522598266602, "learning_rate": 2.5960396039603963e-05, "loss": 0.0524, "step": 2438 }, { "epoch": 9.640316205533598, "grad_norm": 1.5798585414886475, "learning_rate": 2.5950495049504954e-05, "loss": 0.0495, "step": 2439 }, { "epoch": 9.644268774703557, "grad_norm": 1.7892032861709595, "learning_rate": 2.594059405940594e-05, "loss": 0.0553, "step": 2440 }, { "epoch": 9.648221343873518, "grad_norm": 1.6450530290603638, "learning_rate": 2.5930693069306932e-05, "loss": 0.0471, "step": 2441 }, { "epoch": 9.652173913043478, "grad_norm": 1.8182644844055176, "learning_rate": 2.5920792079207924e-05, "loss": 0.0542, "step": 2442 }, { "epoch": 9.65612648221344, "grad_norm": 2.6392593383789062, "learning_rate": 2.5910891089108915e-05, "loss": 0.0704, "step": 2443 }, { "epoch": 9.660079051383399, "grad_norm": 1.484141230583191, "learning_rate": 2.5900990099009902e-05, "loss": 0.0427, "step": 2444 }, { "epoch": 9.66403162055336, "grad_norm": 3.0401995182037354, "learning_rate": 2.5891089108910893e-05, "loss": 0.1027, "step": 2445 }, { "epoch": 9.66798418972332, "grad_norm": 3.289292097091675, "learning_rate": 2.5881188118811884e-05, "loss": 0.0552, "step": 2446 }, { "epoch": 9.671936758893281, "grad_norm": 1.6513707637786865, "learning_rate": 2.5871287128712872e-05, "loss": 0.0491, "step": 2447 }, { "epoch": 9.67588932806324, "grad_norm": 2.30669903755188, "learning_rate": 2.5861386138613863e-05, "loss": 0.0596, "step": 2448 }, { "epoch": 9.679841897233201, "grad_norm": 1.8580228090286255, "learning_rate": 2.5851485148514854e-05, "loss": 0.0655, "step": 2449 }, { "epoch": 9.683794466403162, "grad_norm": 2.1748464107513428, "learning_rate": 2.5841584158415842e-05, "loss": 0.0592, "step": 2450 }, { "epoch": 9.687747035573123, "grad_norm": 2.0328047275543213, "learning_rate": 2.5831683168316833e-05, "loss": 0.0687, "step": 2451 }, { "epoch": 9.691699604743082, "grad_norm": 1.991131067276001, "learning_rate": 2.5821782178217824e-05, "loss": 0.0593, "step": 2452 }, { "epoch": 9.695652173913043, "grad_norm": 1.6861224174499512, "learning_rate": 2.5811881188118815e-05, "loss": 0.0534, "step": 2453 }, { "epoch": 9.699604743083004, "grad_norm": 1.8019297122955322, "learning_rate": 2.5801980198019803e-05, "loss": 0.0544, "step": 2454 }, { "epoch": 9.703557312252965, "grad_norm": 1.7626951932907104, "learning_rate": 2.5792079207920794e-05, "loss": 0.054, "step": 2455 }, { "epoch": 9.707509881422926, "grad_norm": 1.605804204940796, "learning_rate": 2.5782178217821785e-05, "loss": 0.0542, "step": 2456 }, { "epoch": 9.711462450592885, "grad_norm": 2.039774179458618, "learning_rate": 2.5772277227722773e-05, "loss": 0.0587, "step": 2457 }, { "epoch": 9.715415019762846, "grad_norm": 2.0253169536590576, "learning_rate": 2.5762376237623764e-05, "loss": 0.0558, "step": 2458 }, { "epoch": 9.719367588932807, "grad_norm": 1.3254649639129639, "learning_rate": 2.5752475247524755e-05, "loss": 0.0398, "step": 2459 }, { "epoch": 9.723320158102768, "grad_norm": 2.391296625137329, "learning_rate": 2.5742574257425746e-05, "loss": 0.0743, "step": 2460 }, { "epoch": 9.727272727272727, "grad_norm": 1.481223225593567, "learning_rate": 2.5732673267326733e-05, "loss": 0.0431, "step": 2461 }, { "epoch": 9.731225296442688, "grad_norm": 1.5423688888549805, "learning_rate": 2.5722772277227725e-05, "loss": 0.0486, "step": 2462 }, { "epoch": 9.735177865612648, "grad_norm": 1.7814785242080688, "learning_rate": 2.5712871287128716e-05, "loss": 0.0572, "step": 2463 }, { "epoch": 9.73913043478261, "grad_norm": 2.437194347381592, "learning_rate": 2.5702970297029703e-05, "loss": 0.0621, "step": 2464 }, { "epoch": 9.743083003952568, "grad_norm": 2.1086230278015137, "learning_rate": 2.5693069306930694e-05, "loss": 0.0642, "step": 2465 }, { "epoch": 9.74703557312253, "grad_norm": 2.026095390319824, "learning_rate": 2.5683168316831685e-05, "loss": 0.0608, "step": 2466 }, { "epoch": 9.75098814229249, "grad_norm": 1.6836949586868286, "learning_rate": 2.5673267326732676e-05, "loss": 0.0488, "step": 2467 }, { "epoch": 9.754940711462451, "grad_norm": 1.5908231735229492, "learning_rate": 2.5663366336633664e-05, "loss": 0.0523, "step": 2468 }, { "epoch": 9.75889328063241, "grad_norm": 2.4316842555999756, "learning_rate": 2.5653465346534655e-05, "loss": 0.0583, "step": 2469 }, { "epoch": 9.762845849802371, "grad_norm": 2.150331735610962, "learning_rate": 2.5643564356435646e-05, "loss": 0.056, "step": 2470 }, { "epoch": 9.766798418972332, "grad_norm": 1.384072184562683, "learning_rate": 2.5633663366336634e-05, "loss": 0.0468, "step": 2471 }, { "epoch": 9.770750988142293, "grad_norm": 1.5494447946548462, "learning_rate": 2.5623762376237625e-05, "loss": 0.0606, "step": 2472 }, { "epoch": 9.774703557312254, "grad_norm": 1.6640361547470093, "learning_rate": 2.5613861386138616e-05, "loss": 0.0542, "step": 2473 }, { "epoch": 9.778656126482213, "grad_norm": 2.283940076828003, "learning_rate": 2.5603960396039607e-05, "loss": 0.0624, "step": 2474 }, { "epoch": 9.782608695652174, "grad_norm": 1.8987025022506714, "learning_rate": 2.5594059405940595e-05, "loss": 0.055, "step": 2475 }, { "epoch": 9.786561264822135, "grad_norm": 1.7635688781738281, "learning_rate": 2.5584158415841586e-05, "loss": 0.0507, "step": 2476 }, { "epoch": 9.790513833992096, "grad_norm": 1.2326593399047852, "learning_rate": 2.5574257425742577e-05, "loss": 0.0416, "step": 2477 }, { "epoch": 9.794466403162055, "grad_norm": 1.620133876800537, "learning_rate": 2.5564356435643565e-05, "loss": 0.0495, "step": 2478 }, { "epoch": 9.798418972332016, "grad_norm": 2.419328212738037, "learning_rate": 2.5554455445544556e-05, "loss": 0.0791, "step": 2479 }, { "epoch": 9.802371541501977, "grad_norm": 1.5854742527008057, "learning_rate": 2.5544554455445547e-05, "loss": 0.0479, "step": 2480 }, { "epoch": 9.806324110671937, "grad_norm": 1.511894702911377, "learning_rate": 2.5534653465346538e-05, "loss": 0.047, "step": 2481 }, { "epoch": 9.810276679841897, "grad_norm": 1.5028235912322998, "learning_rate": 2.5524752475247526e-05, "loss": 0.0492, "step": 2482 }, { "epoch": 9.814229249011857, "grad_norm": 1.7423605918884277, "learning_rate": 2.5514851485148517e-05, "loss": 0.0548, "step": 2483 }, { "epoch": 9.818181818181818, "grad_norm": 1.4795522689819336, "learning_rate": 2.5504950495049508e-05, "loss": 0.0449, "step": 2484 }, { "epoch": 9.82213438735178, "grad_norm": 1.7023639678955078, "learning_rate": 2.5495049504950495e-05, "loss": 0.0483, "step": 2485 }, { "epoch": 9.826086956521738, "grad_norm": 2.729461193084717, "learning_rate": 2.5485148514851486e-05, "loss": 0.0931, "step": 2486 }, { "epoch": 9.8300395256917, "grad_norm": 1.5015708208084106, "learning_rate": 2.5475247524752477e-05, "loss": 0.0436, "step": 2487 }, { "epoch": 9.83399209486166, "grad_norm": 2.009730100631714, "learning_rate": 2.546534653465347e-05, "loss": 0.0618, "step": 2488 }, { "epoch": 9.837944664031621, "grad_norm": 1.9459285736083984, "learning_rate": 2.5455445544554456e-05, "loss": 0.0539, "step": 2489 }, { "epoch": 9.841897233201582, "grad_norm": 1.8838940858840942, "learning_rate": 2.5445544554455447e-05, "loss": 0.0679, "step": 2490 }, { "epoch": 9.845849802371541, "grad_norm": 1.9405587911605835, "learning_rate": 2.543564356435644e-05, "loss": 0.0599, "step": 2491 }, { "epoch": 9.849802371541502, "grad_norm": 1.7054545879364014, "learning_rate": 2.5425742574257426e-05, "loss": 0.0558, "step": 2492 }, { "epoch": 9.853754940711463, "grad_norm": 1.9355525970458984, "learning_rate": 2.5415841584158417e-05, "loss": 0.0539, "step": 2493 }, { "epoch": 9.857707509881424, "grad_norm": 1.5988727807998657, "learning_rate": 2.5405940594059408e-05, "loss": 0.0516, "step": 2494 }, { "epoch": 9.861660079051383, "grad_norm": 1.826970100402832, "learning_rate": 2.53960396039604e-05, "loss": 0.0463, "step": 2495 }, { "epoch": 9.865612648221344, "grad_norm": 1.8423352241516113, "learning_rate": 2.5386138613861387e-05, "loss": 0.0635, "step": 2496 }, { "epoch": 9.869565217391305, "grad_norm": 2.2359702587127686, "learning_rate": 2.5376237623762378e-05, "loss": 0.0768, "step": 2497 }, { "epoch": 9.873517786561266, "grad_norm": 1.8031789064407349, "learning_rate": 2.536633663366337e-05, "loss": 0.0548, "step": 2498 }, { "epoch": 9.877470355731225, "grad_norm": 1.9371914863586426, "learning_rate": 2.5356435643564357e-05, "loss": 0.0654, "step": 2499 }, { "epoch": 9.881422924901186, "grad_norm": 1.9082609415054321, "learning_rate": 2.5346534653465348e-05, "loss": 0.0544, "step": 2500 }, { "epoch": 9.885375494071146, "grad_norm": 1.7364088296890259, "learning_rate": 2.533663366336634e-05, "loss": 0.0622, "step": 2501 }, { "epoch": 9.889328063241107, "grad_norm": 2.2019705772399902, "learning_rate": 2.532673267326733e-05, "loss": 0.0616, "step": 2502 }, { "epoch": 9.893280632411066, "grad_norm": 1.495164394378662, "learning_rate": 2.5316831683168318e-05, "loss": 0.0552, "step": 2503 }, { "epoch": 9.897233201581027, "grad_norm": 1.7459869384765625, "learning_rate": 2.530693069306931e-05, "loss": 0.0492, "step": 2504 }, { "epoch": 9.901185770750988, "grad_norm": 1.2211105823516846, "learning_rate": 2.52970297029703e-05, "loss": 0.0407, "step": 2505 }, { "epoch": 9.90513833992095, "grad_norm": 1.4153066873550415, "learning_rate": 2.5287128712871287e-05, "loss": 0.0493, "step": 2506 }, { "epoch": 9.909090909090908, "grad_norm": 2.061140537261963, "learning_rate": 2.527722772277228e-05, "loss": 0.0584, "step": 2507 }, { "epoch": 9.91304347826087, "grad_norm": 1.6386499404907227, "learning_rate": 2.526732673267327e-05, "loss": 0.0499, "step": 2508 }, { "epoch": 9.91699604743083, "grad_norm": 1.7371035814285278, "learning_rate": 2.5257425742574257e-05, "loss": 0.0614, "step": 2509 }, { "epoch": 9.920948616600791, "grad_norm": 1.698156714439392, "learning_rate": 2.5247524752475248e-05, "loss": 0.0543, "step": 2510 }, { "epoch": 9.92490118577075, "grad_norm": 1.7899894714355469, "learning_rate": 2.523762376237624e-05, "loss": 0.0563, "step": 2511 }, { "epoch": 9.928853754940711, "grad_norm": 1.730225920677185, "learning_rate": 2.522772277227723e-05, "loss": 0.0533, "step": 2512 }, { "epoch": 9.932806324110672, "grad_norm": 2.2273473739624023, "learning_rate": 2.5217821782178218e-05, "loss": 0.0637, "step": 2513 }, { "epoch": 9.936758893280633, "grad_norm": 1.6896551847457886, "learning_rate": 2.520792079207921e-05, "loss": 0.0489, "step": 2514 }, { "epoch": 9.940711462450594, "grad_norm": 1.9916422367095947, "learning_rate": 2.51980198019802e-05, "loss": 0.0578, "step": 2515 }, { "epoch": 9.944664031620553, "grad_norm": 1.399458408355713, "learning_rate": 2.5188118811881188e-05, "loss": 0.0457, "step": 2516 }, { "epoch": 9.948616600790514, "grad_norm": 1.9079673290252686, "learning_rate": 2.517821782178218e-05, "loss": 0.0645, "step": 2517 }, { "epoch": 9.952569169960475, "grad_norm": 1.7246520519256592, "learning_rate": 2.516831683168317e-05, "loss": 0.0518, "step": 2518 }, { "epoch": 9.956521739130435, "grad_norm": 1.3559677600860596, "learning_rate": 2.515841584158416e-05, "loss": 0.0408, "step": 2519 }, { "epoch": 9.960474308300395, "grad_norm": 2.0350522994995117, "learning_rate": 2.514851485148515e-05, "loss": 0.0562, "step": 2520 }, { "epoch": 9.964426877470355, "grad_norm": 1.8791509866714478, "learning_rate": 2.513861386138614e-05, "loss": 0.0598, "step": 2521 }, { "epoch": 9.968379446640316, "grad_norm": 1.8241581916809082, "learning_rate": 2.512871287128713e-05, "loss": 0.0523, "step": 2522 }, { "epoch": 9.972332015810277, "grad_norm": 1.4811714887619019, "learning_rate": 2.511881188118812e-05, "loss": 0.0441, "step": 2523 }, { "epoch": 9.976284584980236, "grad_norm": 1.7711213827133179, "learning_rate": 2.510891089108911e-05, "loss": 0.0583, "step": 2524 }, { "epoch": 9.980237154150197, "grad_norm": 1.35908842086792, "learning_rate": 2.50990099009901e-05, "loss": 0.0489, "step": 2525 }, { "epoch": 9.984189723320158, "grad_norm": 1.9127227067947388, "learning_rate": 2.5089108910891092e-05, "loss": 0.0507, "step": 2526 }, { "epoch": 9.988142292490119, "grad_norm": 2.2715985774993896, "learning_rate": 2.507920792079208e-05, "loss": 0.0636, "step": 2527 }, { "epoch": 9.992094861660078, "grad_norm": 1.6719472408294678, "learning_rate": 2.506930693069307e-05, "loss": 0.054, "step": 2528 }, { "epoch": 9.996047430830039, "grad_norm": 1.95130455493927, "learning_rate": 2.505940594059406e-05, "loss": 0.0585, "step": 2529 }, { "epoch": 10.0, "grad_norm": 1.9763691425323486, "learning_rate": 2.504950495049505e-05, "loss": 0.0615, "step": 2530 }, { "epoch": 10.003952569169961, "grad_norm": 1.230286717414856, "learning_rate": 2.503960396039604e-05, "loss": 0.0346, "step": 2531 }, { "epoch": 10.007905138339922, "grad_norm": 1.8030517101287842, "learning_rate": 2.502970297029703e-05, "loss": 0.0567, "step": 2532 }, { "epoch": 10.011857707509881, "grad_norm": 1.0350579023361206, "learning_rate": 2.5019801980198022e-05, "loss": 0.0349, "step": 2533 }, { "epoch": 10.015810276679842, "grad_norm": 1.8892607688903809, "learning_rate": 2.500990099009901e-05, "loss": 0.0605, "step": 2534 }, { "epoch": 10.019762845849803, "grad_norm": 0.8503550291061401, "learning_rate": 2.5e-05, "loss": 0.0245, "step": 2535 }, { "epoch": 10.023715415019764, "grad_norm": 1.3090264797210693, "learning_rate": 2.4990099009900992e-05, "loss": 0.0423, "step": 2536 }, { "epoch": 10.027667984189723, "grad_norm": 0.9017503261566162, "learning_rate": 2.498019801980198e-05, "loss": 0.0293, "step": 2537 }, { "epoch": 10.031620553359684, "grad_norm": 1.3760610818862915, "learning_rate": 2.497029702970297e-05, "loss": 0.0418, "step": 2538 }, { "epoch": 10.035573122529645, "grad_norm": 1.0394277572631836, "learning_rate": 2.4960396039603962e-05, "loss": 0.0324, "step": 2539 }, { "epoch": 10.039525691699605, "grad_norm": 1.4052479267120361, "learning_rate": 2.4950495049504953e-05, "loss": 0.04, "step": 2540 }, { "epoch": 10.043478260869565, "grad_norm": 1.2445170879364014, "learning_rate": 2.494059405940594e-05, "loss": 0.035, "step": 2541 }, { "epoch": 10.047430830039525, "grad_norm": 1.208992838859558, "learning_rate": 2.4930693069306932e-05, "loss": 0.0337, "step": 2542 }, { "epoch": 10.051383399209486, "grad_norm": 1.4467679262161255, "learning_rate": 2.4920792079207923e-05, "loss": 0.0418, "step": 2543 }, { "epoch": 10.055335968379447, "grad_norm": 1.1472065448760986, "learning_rate": 2.491089108910891e-05, "loss": 0.0298, "step": 2544 }, { "epoch": 10.059288537549406, "grad_norm": 1.1275383234024048, "learning_rate": 2.49009900990099e-05, "loss": 0.0345, "step": 2545 }, { "epoch": 10.063241106719367, "grad_norm": 1.398996114730835, "learning_rate": 2.4891089108910893e-05, "loss": 0.0372, "step": 2546 }, { "epoch": 10.067193675889328, "grad_norm": 1.6347278356552124, "learning_rate": 2.4881188118811884e-05, "loss": 0.0418, "step": 2547 }, { "epoch": 10.071146245059289, "grad_norm": 1.8869022130966187, "learning_rate": 2.487128712871287e-05, "loss": 0.0569, "step": 2548 }, { "epoch": 10.075098814229248, "grad_norm": 1.8023834228515625, "learning_rate": 2.4861386138613863e-05, "loss": 0.0504, "step": 2549 }, { "epoch": 10.079051383399209, "grad_norm": 1.2219035625457764, "learning_rate": 2.4851485148514854e-05, "loss": 0.0314, "step": 2550 }, { "epoch": 10.08300395256917, "grad_norm": 1.3041025400161743, "learning_rate": 2.484158415841584e-05, "loss": 0.0364, "step": 2551 }, { "epoch": 10.08695652173913, "grad_norm": 1.1773312091827393, "learning_rate": 2.4831683168316832e-05, "loss": 0.0406, "step": 2552 }, { "epoch": 10.090909090909092, "grad_norm": 1.4438090324401855, "learning_rate": 2.4821782178217823e-05, "loss": 0.0406, "step": 2553 }, { "epoch": 10.09486166007905, "grad_norm": 1.3536142110824585, "learning_rate": 2.4811881188118815e-05, "loss": 0.0403, "step": 2554 }, { "epoch": 10.098814229249012, "grad_norm": 1.4740244150161743, "learning_rate": 2.4801980198019802e-05, "loss": 0.0513, "step": 2555 }, { "epoch": 10.102766798418973, "grad_norm": 1.626861810684204, "learning_rate": 2.4792079207920793e-05, "loss": 0.0497, "step": 2556 }, { "epoch": 10.106719367588934, "grad_norm": 1.6449626684188843, "learning_rate": 2.4782178217821784e-05, "loss": 0.0456, "step": 2557 }, { "epoch": 10.110671936758893, "grad_norm": 1.8075177669525146, "learning_rate": 2.4772277227722772e-05, "loss": 0.0495, "step": 2558 }, { "epoch": 10.114624505928854, "grad_norm": 1.2482951879501343, "learning_rate": 2.4762376237623763e-05, "loss": 0.035, "step": 2559 }, { "epoch": 10.118577075098814, "grad_norm": 1.3251415491104126, "learning_rate": 2.4752475247524754e-05, "loss": 0.038, "step": 2560 }, { "epoch": 10.122529644268775, "grad_norm": 1.460797905921936, "learning_rate": 2.4742574257425745e-05, "loss": 0.0437, "step": 2561 }, { "epoch": 10.126482213438734, "grad_norm": 1.520089030265808, "learning_rate": 2.4732673267326733e-05, "loss": 0.0444, "step": 2562 }, { "epoch": 10.130434782608695, "grad_norm": 1.6610418558120728, "learning_rate": 2.4722772277227724e-05, "loss": 0.0475, "step": 2563 }, { "epoch": 10.134387351778656, "grad_norm": 1.8135102987289429, "learning_rate": 2.4712871287128715e-05, "loss": 0.0481, "step": 2564 }, { "epoch": 10.138339920948617, "grad_norm": 1.5815588235855103, "learning_rate": 2.4702970297029703e-05, "loss": 0.0453, "step": 2565 }, { "epoch": 10.142292490118576, "grad_norm": 2.132692337036133, "learning_rate": 2.4693069306930694e-05, "loss": 0.062, "step": 2566 }, { "epoch": 10.146245059288537, "grad_norm": 2.0057616233825684, "learning_rate": 2.4683168316831685e-05, "loss": 0.0545, "step": 2567 }, { "epoch": 10.150197628458498, "grad_norm": 1.1556954383850098, "learning_rate": 2.4673267326732672e-05, "loss": 0.0366, "step": 2568 }, { "epoch": 10.154150197628459, "grad_norm": 1.2460317611694336, "learning_rate": 2.4663366336633664e-05, "loss": 0.0404, "step": 2569 }, { "epoch": 10.15810276679842, "grad_norm": 1.5843098163604736, "learning_rate": 2.4653465346534655e-05, "loss": 0.0493, "step": 2570 }, { "epoch": 10.162055335968379, "grad_norm": 1.197706937789917, "learning_rate": 2.4643564356435646e-05, "loss": 0.0366, "step": 2571 }, { "epoch": 10.16600790513834, "grad_norm": 1.002633810043335, "learning_rate": 2.4633663366336633e-05, "loss": 0.0302, "step": 2572 }, { "epoch": 10.1699604743083, "grad_norm": 1.5681037902832031, "learning_rate": 2.4623762376237624e-05, "loss": 0.0449, "step": 2573 }, { "epoch": 10.173913043478262, "grad_norm": 1.41460120677948, "learning_rate": 2.4613861386138616e-05, "loss": 0.0378, "step": 2574 }, { "epoch": 10.17786561264822, "grad_norm": 2.5204408168792725, "learning_rate": 2.4603960396039603e-05, "loss": 0.0672, "step": 2575 }, { "epoch": 10.181818181818182, "grad_norm": 1.7972041368484497, "learning_rate": 2.4594059405940594e-05, "loss": 0.0527, "step": 2576 }, { "epoch": 10.185770750988143, "grad_norm": 1.173115849494934, "learning_rate": 2.4584158415841585e-05, "loss": 0.0343, "step": 2577 }, { "epoch": 10.189723320158103, "grad_norm": 1.330645203590393, "learning_rate": 2.4574257425742576e-05, "loss": 0.0384, "step": 2578 }, { "epoch": 10.193675889328063, "grad_norm": 2.61321759223938, "learning_rate": 2.4564356435643564e-05, "loss": 0.0745, "step": 2579 }, { "epoch": 10.197628458498023, "grad_norm": 2.3448495864868164, "learning_rate": 2.4554455445544555e-05, "loss": 0.0524, "step": 2580 }, { "epoch": 10.201581027667984, "grad_norm": 1.9829871654510498, "learning_rate": 2.4544554455445546e-05, "loss": 0.0593, "step": 2581 }, { "epoch": 10.205533596837945, "grad_norm": 1.4846584796905518, "learning_rate": 2.4534653465346534e-05, "loss": 0.0451, "step": 2582 }, { "epoch": 10.209486166007904, "grad_norm": 1.4935392141342163, "learning_rate": 2.4524752475247525e-05, "loss": 0.0428, "step": 2583 }, { "epoch": 10.213438735177865, "grad_norm": 1.4938994646072388, "learning_rate": 2.4514851485148516e-05, "loss": 0.0425, "step": 2584 }, { "epoch": 10.217391304347826, "grad_norm": 1.6958317756652832, "learning_rate": 2.4504950495049507e-05, "loss": 0.0503, "step": 2585 }, { "epoch": 10.221343873517787, "grad_norm": 1.5017327070236206, "learning_rate": 2.4495049504950495e-05, "loss": 0.042, "step": 2586 }, { "epoch": 10.225296442687746, "grad_norm": 1.3924531936645508, "learning_rate": 2.4485148514851486e-05, "loss": 0.039, "step": 2587 }, { "epoch": 10.229249011857707, "grad_norm": 1.91053307056427, "learning_rate": 2.4475247524752477e-05, "loss": 0.0618, "step": 2588 }, { "epoch": 10.233201581027668, "grad_norm": 1.6566091775894165, "learning_rate": 2.4465346534653465e-05, "loss": 0.0556, "step": 2589 }, { "epoch": 10.237154150197629, "grad_norm": 1.522037148475647, "learning_rate": 2.4455445544554456e-05, "loss": 0.0464, "step": 2590 }, { "epoch": 10.24110671936759, "grad_norm": 1.833832859992981, "learning_rate": 2.4445544554455447e-05, "loss": 0.0509, "step": 2591 }, { "epoch": 10.245059288537549, "grad_norm": 1.2517982721328735, "learning_rate": 2.4435643564356438e-05, "loss": 0.0345, "step": 2592 }, { "epoch": 10.24901185770751, "grad_norm": 1.7737735509872437, "learning_rate": 2.4425742574257425e-05, "loss": 0.0508, "step": 2593 }, { "epoch": 10.25296442687747, "grad_norm": 1.4050178527832031, "learning_rate": 2.4415841584158416e-05, "loss": 0.0477, "step": 2594 }, { "epoch": 10.256916996047432, "grad_norm": 2.6510720252990723, "learning_rate": 2.4405940594059408e-05, "loss": 0.0646, "step": 2595 }, { "epoch": 10.26086956521739, "grad_norm": 1.0306216478347778, "learning_rate": 2.4396039603960395e-05, "loss": 0.0311, "step": 2596 }, { "epoch": 10.264822134387352, "grad_norm": 1.285875916481018, "learning_rate": 2.4386138613861386e-05, "loss": 0.0396, "step": 2597 }, { "epoch": 10.268774703557312, "grad_norm": 2.0907397270202637, "learning_rate": 2.4376237623762377e-05, "loss": 0.0607, "step": 2598 }, { "epoch": 10.272727272727273, "grad_norm": 1.2815570831298828, "learning_rate": 2.436633663366337e-05, "loss": 0.0408, "step": 2599 }, { "epoch": 10.276679841897232, "grad_norm": 1.2680206298828125, "learning_rate": 2.4356435643564356e-05, "loss": 0.039, "step": 2600 }, { "epoch": 10.280632411067193, "grad_norm": 1.463365077972412, "learning_rate": 2.4346534653465347e-05, "loss": 0.0441, "step": 2601 }, { "epoch": 10.284584980237154, "grad_norm": 1.5797768831253052, "learning_rate": 2.4336633663366338e-05, "loss": 0.0524, "step": 2602 }, { "epoch": 10.288537549407115, "grad_norm": 1.8467251062393188, "learning_rate": 2.4326732673267326e-05, "loss": 0.0558, "step": 2603 }, { "epoch": 10.292490118577074, "grad_norm": 1.4967631101608276, "learning_rate": 2.4316831683168317e-05, "loss": 0.0403, "step": 2604 }, { "epoch": 10.296442687747035, "grad_norm": 1.124082326889038, "learning_rate": 2.4306930693069308e-05, "loss": 0.0404, "step": 2605 }, { "epoch": 10.300395256916996, "grad_norm": 1.5682737827301025, "learning_rate": 2.42970297029703e-05, "loss": 0.0441, "step": 2606 }, { "epoch": 10.304347826086957, "grad_norm": 1.8255658149719238, "learning_rate": 2.4287128712871287e-05, "loss": 0.0592, "step": 2607 }, { "epoch": 10.308300395256918, "grad_norm": 1.4027087688446045, "learning_rate": 2.4277227722772278e-05, "loss": 0.0442, "step": 2608 }, { "epoch": 10.312252964426877, "grad_norm": 1.6431357860565186, "learning_rate": 2.426732673267327e-05, "loss": 0.0505, "step": 2609 }, { "epoch": 10.316205533596838, "grad_norm": 1.4613101482391357, "learning_rate": 2.4257425742574257e-05, "loss": 0.0462, "step": 2610 }, { "epoch": 10.320158102766799, "grad_norm": 1.2558099031448364, "learning_rate": 2.4247524752475248e-05, "loss": 0.0379, "step": 2611 }, { "epoch": 10.32411067193676, "grad_norm": 1.4268484115600586, "learning_rate": 2.423762376237624e-05, "loss": 0.0403, "step": 2612 }, { "epoch": 10.328063241106719, "grad_norm": 1.8302972316741943, "learning_rate": 2.422772277227723e-05, "loss": 0.0634, "step": 2613 }, { "epoch": 10.33201581027668, "grad_norm": 1.2978883981704712, "learning_rate": 2.4217821782178217e-05, "loss": 0.0432, "step": 2614 }, { "epoch": 10.33596837944664, "grad_norm": 1.3345191478729248, "learning_rate": 2.420792079207921e-05, "loss": 0.0361, "step": 2615 }, { "epoch": 10.339920948616601, "grad_norm": 1.5218837261199951, "learning_rate": 2.41980198019802e-05, "loss": 0.0474, "step": 2616 }, { "epoch": 10.34387351778656, "grad_norm": 1.2001280784606934, "learning_rate": 2.4188118811881187e-05, "loss": 0.0356, "step": 2617 }, { "epoch": 10.347826086956522, "grad_norm": 1.233418583869934, "learning_rate": 2.417821782178218e-05, "loss": 0.0368, "step": 2618 }, { "epoch": 10.351778656126482, "grad_norm": 3.1745145320892334, "learning_rate": 2.416831683168317e-05, "loss": 0.0824, "step": 2619 }, { "epoch": 10.355731225296443, "grad_norm": 1.4785910844802856, "learning_rate": 2.415841584158416e-05, "loss": 0.0413, "step": 2620 }, { "epoch": 10.359683794466402, "grad_norm": 1.1766948699951172, "learning_rate": 2.4148514851485148e-05, "loss": 0.0334, "step": 2621 }, { "epoch": 10.363636363636363, "grad_norm": 1.7353991270065308, "learning_rate": 2.413861386138614e-05, "loss": 0.0477, "step": 2622 }, { "epoch": 10.367588932806324, "grad_norm": 1.844902753829956, "learning_rate": 2.412871287128713e-05, "loss": 0.0486, "step": 2623 }, { "epoch": 10.371541501976285, "grad_norm": 1.453034520149231, "learning_rate": 2.4118811881188118e-05, "loss": 0.0412, "step": 2624 }, { "epoch": 10.375494071146244, "grad_norm": 1.9265364408493042, "learning_rate": 2.410891089108911e-05, "loss": 0.0509, "step": 2625 }, { "epoch": 10.379446640316205, "grad_norm": 1.4409905672073364, "learning_rate": 2.40990099009901e-05, "loss": 0.041, "step": 2626 }, { "epoch": 10.383399209486166, "grad_norm": 1.4052621126174927, "learning_rate": 2.4089108910891088e-05, "loss": 0.0374, "step": 2627 }, { "epoch": 10.387351778656127, "grad_norm": 1.4964357614517212, "learning_rate": 2.407920792079208e-05, "loss": 0.0504, "step": 2628 }, { "epoch": 10.391304347826088, "grad_norm": 1.3023245334625244, "learning_rate": 2.406930693069307e-05, "loss": 0.0373, "step": 2629 }, { "epoch": 10.395256916996047, "grad_norm": 1.9414830207824707, "learning_rate": 2.405940594059406e-05, "loss": 0.0523, "step": 2630 }, { "epoch": 10.399209486166008, "grad_norm": 1.5991520881652832, "learning_rate": 2.404950495049505e-05, "loss": 0.0489, "step": 2631 }, { "epoch": 10.403162055335969, "grad_norm": 1.7837496995925903, "learning_rate": 2.403960396039604e-05, "loss": 0.0491, "step": 2632 }, { "epoch": 10.40711462450593, "grad_norm": 1.8126829862594604, "learning_rate": 2.402970297029703e-05, "loss": 0.0487, "step": 2633 }, { "epoch": 10.411067193675889, "grad_norm": 1.9854708909988403, "learning_rate": 2.401980198019802e-05, "loss": 0.0578, "step": 2634 }, { "epoch": 10.41501976284585, "grad_norm": 1.2300968170166016, "learning_rate": 2.400990099009901e-05, "loss": 0.0387, "step": 2635 }, { "epoch": 10.41897233201581, "grad_norm": 2.038818836212158, "learning_rate": 2.4e-05, "loss": 0.0612, "step": 2636 }, { "epoch": 10.422924901185771, "grad_norm": 1.4362776279449463, "learning_rate": 2.399009900990099e-05, "loss": 0.0448, "step": 2637 }, { "epoch": 10.42687747035573, "grad_norm": 1.619510293006897, "learning_rate": 2.398019801980198e-05, "loss": 0.0413, "step": 2638 }, { "epoch": 10.430830039525691, "grad_norm": 1.6483683586120605, "learning_rate": 2.397029702970297e-05, "loss": 0.0499, "step": 2639 }, { "epoch": 10.434782608695652, "grad_norm": 1.8103317022323608, "learning_rate": 2.396039603960396e-05, "loss": 0.0576, "step": 2640 }, { "epoch": 10.438735177865613, "grad_norm": 1.1536823511123657, "learning_rate": 2.395049504950495e-05, "loss": 0.0378, "step": 2641 }, { "epoch": 10.442687747035572, "grad_norm": 1.513077735900879, "learning_rate": 2.394059405940594e-05, "loss": 0.0407, "step": 2642 }, { "epoch": 10.446640316205533, "grad_norm": 1.5335613489151, "learning_rate": 2.393069306930693e-05, "loss": 0.0445, "step": 2643 }, { "epoch": 10.450592885375494, "grad_norm": 0.9189258813858032, "learning_rate": 2.3920792079207922e-05, "loss": 0.0304, "step": 2644 }, { "epoch": 10.454545454545455, "grad_norm": 1.3073348999023438, "learning_rate": 2.391089108910891e-05, "loss": 0.0433, "step": 2645 }, { "epoch": 10.458498023715414, "grad_norm": 1.6946250200271606, "learning_rate": 2.39009900990099e-05, "loss": 0.052, "step": 2646 }, { "epoch": 10.462450592885375, "grad_norm": 1.925939679145813, "learning_rate": 2.3891089108910892e-05, "loss": 0.0758, "step": 2647 }, { "epoch": 10.466403162055336, "grad_norm": 1.4356261491775513, "learning_rate": 2.388118811881188e-05, "loss": 0.0445, "step": 2648 }, { "epoch": 10.470355731225297, "grad_norm": 2.4026238918304443, "learning_rate": 2.387128712871287e-05, "loss": 0.0974, "step": 2649 }, { "epoch": 10.474308300395258, "grad_norm": 1.045005440711975, "learning_rate": 2.3861386138613862e-05, "loss": 0.0351, "step": 2650 }, { "epoch": 10.478260869565217, "grad_norm": 1.621512532234192, "learning_rate": 2.3851485148514853e-05, "loss": 0.0446, "step": 2651 }, { "epoch": 10.482213438735178, "grad_norm": 1.4734852313995361, "learning_rate": 2.384158415841584e-05, "loss": 0.0459, "step": 2652 }, { "epoch": 10.486166007905139, "grad_norm": 1.4412637948989868, "learning_rate": 2.3831683168316832e-05, "loss": 0.0449, "step": 2653 }, { "epoch": 10.4901185770751, "grad_norm": 1.5563961267471313, "learning_rate": 2.3821782178217823e-05, "loss": 0.0454, "step": 2654 }, { "epoch": 10.494071146245059, "grad_norm": 1.2778637409210205, "learning_rate": 2.381188118811881e-05, "loss": 0.0378, "step": 2655 }, { "epoch": 10.49802371541502, "grad_norm": 1.1935839653015137, "learning_rate": 2.38019801980198e-05, "loss": 0.0352, "step": 2656 }, { "epoch": 10.50197628458498, "grad_norm": 1.2721806764602661, "learning_rate": 2.3792079207920793e-05, "loss": 0.037, "step": 2657 }, { "epoch": 10.505928853754941, "grad_norm": 1.5275217294692993, "learning_rate": 2.3782178217821784e-05, "loss": 0.0488, "step": 2658 }, { "epoch": 10.5098814229249, "grad_norm": 1.4390796422958374, "learning_rate": 2.377227722772277e-05, "loss": 0.0375, "step": 2659 }, { "epoch": 10.513833992094861, "grad_norm": 1.786362886428833, "learning_rate": 2.3762376237623762e-05, "loss": 0.0567, "step": 2660 }, { "epoch": 10.517786561264822, "grad_norm": 1.4250315427780151, "learning_rate": 2.3752475247524754e-05, "loss": 0.047, "step": 2661 }, { "epoch": 10.521739130434783, "grad_norm": 1.9806933403015137, "learning_rate": 2.374257425742574e-05, "loss": 0.0529, "step": 2662 }, { "epoch": 10.525691699604742, "grad_norm": 1.7346775531768799, "learning_rate": 2.3732673267326732e-05, "loss": 0.0588, "step": 2663 }, { "epoch": 10.529644268774703, "grad_norm": 1.242470622062683, "learning_rate": 2.3722772277227723e-05, "loss": 0.0354, "step": 2664 }, { "epoch": 10.533596837944664, "grad_norm": 1.3049713373184204, "learning_rate": 2.3712871287128714e-05, "loss": 0.0418, "step": 2665 }, { "epoch": 10.537549407114625, "grad_norm": 1.1761693954467773, "learning_rate": 2.3702970297029702e-05, "loss": 0.0324, "step": 2666 }, { "epoch": 10.541501976284586, "grad_norm": 1.1646145582199097, "learning_rate": 2.3693069306930693e-05, "loss": 0.0362, "step": 2667 }, { "epoch": 10.545454545454545, "grad_norm": 1.1377053260803223, "learning_rate": 2.3683168316831684e-05, "loss": 0.0378, "step": 2668 }, { "epoch": 10.549407114624506, "grad_norm": 1.4388465881347656, "learning_rate": 2.3673267326732672e-05, "loss": 0.0426, "step": 2669 }, { "epoch": 10.553359683794467, "grad_norm": 1.321148157119751, "learning_rate": 2.3663366336633663e-05, "loss": 0.0419, "step": 2670 }, { "epoch": 10.557312252964428, "grad_norm": 1.4911779165267944, "learning_rate": 2.3653465346534654e-05, "loss": 0.0421, "step": 2671 }, { "epoch": 10.561264822134387, "grad_norm": 1.2032262086868286, "learning_rate": 2.3643564356435645e-05, "loss": 0.0349, "step": 2672 }, { "epoch": 10.565217391304348, "grad_norm": 1.2327481508255005, "learning_rate": 2.3633663366336633e-05, "loss": 0.0391, "step": 2673 }, { "epoch": 10.569169960474309, "grad_norm": 1.6718777418136597, "learning_rate": 2.3623762376237624e-05, "loss": 0.0501, "step": 2674 }, { "epoch": 10.57312252964427, "grad_norm": 1.4705296754837036, "learning_rate": 2.3613861386138615e-05, "loss": 0.0429, "step": 2675 }, { "epoch": 10.577075098814229, "grad_norm": 1.757813811302185, "learning_rate": 2.3603960396039603e-05, "loss": 0.0544, "step": 2676 }, { "epoch": 10.58102766798419, "grad_norm": 1.2530285120010376, "learning_rate": 2.3594059405940594e-05, "loss": 0.0374, "step": 2677 }, { "epoch": 10.58498023715415, "grad_norm": 2.2816851139068604, "learning_rate": 2.3584158415841585e-05, "loss": 0.0667, "step": 2678 }, { "epoch": 10.588932806324111, "grad_norm": 1.5257638692855835, "learning_rate": 2.3574257425742576e-05, "loss": 0.0394, "step": 2679 }, { "epoch": 10.59288537549407, "grad_norm": 1.3651671409606934, "learning_rate": 2.3564356435643563e-05, "loss": 0.0402, "step": 2680 }, { "epoch": 10.596837944664031, "grad_norm": 1.5354106426239014, "learning_rate": 2.3554455445544555e-05, "loss": 0.0434, "step": 2681 }, { "epoch": 10.600790513833992, "grad_norm": 1.580075979232788, "learning_rate": 2.3544554455445546e-05, "loss": 0.0453, "step": 2682 }, { "epoch": 10.604743083003953, "grad_norm": 1.9237476587295532, "learning_rate": 2.3534653465346533e-05, "loss": 0.0572, "step": 2683 }, { "epoch": 10.608695652173914, "grad_norm": 1.7281787395477295, "learning_rate": 2.3524752475247524e-05, "loss": 0.0441, "step": 2684 }, { "epoch": 10.612648221343873, "grad_norm": 1.4923232793807983, "learning_rate": 2.3514851485148515e-05, "loss": 0.0473, "step": 2685 }, { "epoch": 10.616600790513834, "grad_norm": 1.650608777999878, "learning_rate": 2.3504950495049503e-05, "loss": 0.0455, "step": 2686 }, { "epoch": 10.620553359683795, "grad_norm": 1.3070241212844849, "learning_rate": 2.3495049504950494e-05, "loss": 0.0377, "step": 2687 }, { "epoch": 10.624505928853754, "grad_norm": 1.6660254001617432, "learning_rate": 2.3485148514851485e-05, "loss": 0.0476, "step": 2688 }, { "epoch": 10.628458498023715, "grad_norm": 2.0659825801849365, "learning_rate": 2.3475247524752476e-05, "loss": 0.0597, "step": 2689 }, { "epoch": 10.632411067193676, "grad_norm": 1.6319752931594849, "learning_rate": 2.3465346534653464e-05, "loss": 0.0547, "step": 2690 }, { "epoch": 10.636363636363637, "grad_norm": 1.954634428024292, "learning_rate": 2.3455445544554455e-05, "loss": 0.0566, "step": 2691 }, { "epoch": 10.640316205533598, "grad_norm": 1.277179479598999, "learning_rate": 2.3445544554455446e-05, "loss": 0.0391, "step": 2692 }, { "epoch": 10.644268774703557, "grad_norm": 1.4013490676879883, "learning_rate": 2.3435643564356437e-05, "loss": 0.0453, "step": 2693 }, { "epoch": 10.648221343873518, "grad_norm": 1.6931759119033813, "learning_rate": 2.3425742574257428e-05, "loss": 0.0517, "step": 2694 }, { "epoch": 10.652173913043478, "grad_norm": 1.41490638256073, "learning_rate": 2.341584158415842e-05, "loss": 0.0391, "step": 2695 }, { "epoch": 10.65612648221344, "grad_norm": 1.1496773958206177, "learning_rate": 2.3405940594059407e-05, "loss": 0.0403, "step": 2696 }, { "epoch": 10.660079051383399, "grad_norm": 1.6764216423034668, "learning_rate": 2.3396039603960398e-05, "loss": 0.0536, "step": 2697 }, { "epoch": 10.66403162055336, "grad_norm": 1.326882243156433, "learning_rate": 2.338613861386139e-05, "loss": 0.0399, "step": 2698 }, { "epoch": 10.66798418972332, "grad_norm": 1.4323588609695435, "learning_rate": 2.337623762376238e-05, "loss": 0.0414, "step": 2699 }, { "epoch": 10.671936758893281, "grad_norm": 1.388510823249817, "learning_rate": 2.3366336633663368e-05, "loss": 0.042, "step": 2700 }, { "epoch": 10.67588932806324, "grad_norm": 1.352927565574646, "learning_rate": 2.335643564356436e-05, "loss": 0.0362, "step": 2701 }, { "epoch": 10.679841897233201, "grad_norm": 1.5100293159484863, "learning_rate": 2.334653465346535e-05, "loss": 0.0455, "step": 2702 }, { "epoch": 10.683794466403162, "grad_norm": 1.4751832485198975, "learning_rate": 2.3336633663366338e-05, "loss": 0.0401, "step": 2703 }, { "epoch": 10.687747035573123, "grad_norm": 1.6399329900741577, "learning_rate": 2.332673267326733e-05, "loss": 0.0363, "step": 2704 }, { "epoch": 10.691699604743082, "grad_norm": 1.111297369003296, "learning_rate": 2.331683168316832e-05, "loss": 0.037, "step": 2705 }, { "epoch": 10.695652173913043, "grad_norm": 1.7264074087142944, "learning_rate": 2.330693069306931e-05, "loss": 0.051, "step": 2706 }, { "epoch": 10.699604743083004, "grad_norm": 2.0200514793395996, "learning_rate": 2.32970297029703e-05, "loss": 0.0509, "step": 2707 }, { "epoch": 10.703557312252965, "grad_norm": 2.0624306201934814, "learning_rate": 2.328712871287129e-05, "loss": 0.0542, "step": 2708 }, { "epoch": 10.707509881422926, "grad_norm": 1.0713139772415161, "learning_rate": 2.327722772277228e-05, "loss": 0.033, "step": 2709 }, { "epoch": 10.711462450592885, "grad_norm": 1.5157147645950317, "learning_rate": 2.326732673267327e-05, "loss": 0.0431, "step": 2710 }, { "epoch": 10.715415019762846, "grad_norm": 1.4376459121704102, "learning_rate": 2.325742574257426e-05, "loss": 0.0441, "step": 2711 }, { "epoch": 10.719367588932807, "grad_norm": 1.5410816669464111, "learning_rate": 2.324752475247525e-05, "loss": 0.0483, "step": 2712 }, { "epoch": 10.723320158102768, "grad_norm": 1.787246823310852, "learning_rate": 2.323762376237624e-05, "loss": 0.0466, "step": 2713 }, { "epoch": 10.727272727272727, "grad_norm": 1.8572641611099243, "learning_rate": 2.322772277227723e-05, "loss": 0.0607, "step": 2714 }, { "epoch": 10.731225296442688, "grad_norm": 1.3746086359024048, "learning_rate": 2.321782178217822e-05, "loss": 0.0419, "step": 2715 }, { "epoch": 10.735177865612648, "grad_norm": 1.7419465780258179, "learning_rate": 2.320792079207921e-05, "loss": 0.0481, "step": 2716 }, { "epoch": 10.73913043478261, "grad_norm": 1.6720646619796753, "learning_rate": 2.31980198019802e-05, "loss": 0.0502, "step": 2717 }, { "epoch": 10.743083003952568, "grad_norm": 1.8969156742095947, "learning_rate": 2.318811881188119e-05, "loss": 0.0604, "step": 2718 }, { "epoch": 10.74703557312253, "grad_norm": 1.4605003595352173, "learning_rate": 2.317821782178218e-05, "loss": 0.0483, "step": 2719 }, { "epoch": 10.75098814229249, "grad_norm": 1.192971110343933, "learning_rate": 2.3168316831683172e-05, "loss": 0.0362, "step": 2720 }, { "epoch": 10.754940711462451, "grad_norm": 1.25176203250885, "learning_rate": 2.315841584158416e-05, "loss": 0.0364, "step": 2721 }, { "epoch": 10.75889328063241, "grad_norm": 1.341831088066101, "learning_rate": 2.314851485148515e-05, "loss": 0.0389, "step": 2722 }, { "epoch": 10.762845849802371, "grad_norm": 1.6302502155303955, "learning_rate": 2.3138613861386142e-05, "loss": 0.0516, "step": 2723 }, { "epoch": 10.766798418972332, "grad_norm": 1.6141759157180786, "learning_rate": 2.312871287128713e-05, "loss": 0.0523, "step": 2724 }, { "epoch": 10.770750988142293, "grad_norm": 1.7800735235214233, "learning_rate": 2.311881188118812e-05, "loss": 0.0513, "step": 2725 }, { "epoch": 10.774703557312254, "grad_norm": 1.6695626974105835, "learning_rate": 2.3108910891089112e-05, "loss": 0.0492, "step": 2726 }, { "epoch": 10.778656126482213, "grad_norm": 1.041245698928833, "learning_rate": 2.30990099009901e-05, "loss": 0.0321, "step": 2727 }, { "epoch": 10.782608695652174, "grad_norm": 1.3573249578475952, "learning_rate": 2.308910891089109e-05, "loss": 0.0403, "step": 2728 }, { "epoch": 10.786561264822135, "grad_norm": 1.3664249181747437, "learning_rate": 2.307920792079208e-05, "loss": 0.0406, "step": 2729 }, { "epoch": 10.790513833992096, "grad_norm": 1.1999021768569946, "learning_rate": 2.3069306930693073e-05, "loss": 0.0337, "step": 2730 }, { "epoch": 10.794466403162055, "grad_norm": 1.5350315570831299, "learning_rate": 2.305940594059406e-05, "loss": 0.0446, "step": 2731 }, { "epoch": 10.798418972332016, "grad_norm": 1.7841217517852783, "learning_rate": 2.304950495049505e-05, "loss": 0.0579, "step": 2732 }, { "epoch": 10.802371541501977, "grad_norm": 1.6352752447128296, "learning_rate": 2.3039603960396043e-05, "loss": 0.049, "step": 2733 }, { "epoch": 10.806324110671937, "grad_norm": 1.5531632900238037, "learning_rate": 2.302970297029703e-05, "loss": 0.0461, "step": 2734 }, { "epoch": 10.810276679841897, "grad_norm": 1.3284839391708374, "learning_rate": 2.301980198019802e-05, "loss": 0.0363, "step": 2735 }, { "epoch": 10.814229249011857, "grad_norm": 1.5151387453079224, "learning_rate": 2.3009900990099012e-05, "loss": 0.0447, "step": 2736 }, { "epoch": 10.818181818181818, "grad_norm": 1.4957261085510254, "learning_rate": 2.3000000000000003e-05, "loss": 0.0474, "step": 2737 }, { "epoch": 10.82213438735178, "grad_norm": 1.060352087020874, "learning_rate": 2.299009900990099e-05, "loss": 0.0445, "step": 2738 }, { "epoch": 10.826086956521738, "grad_norm": 1.744613766670227, "learning_rate": 2.2980198019801982e-05, "loss": 0.0454, "step": 2739 }, { "epoch": 10.8300395256917, "grad_norm": 1.4505136013031006, "learning_rate": 2.2970297029702973e-05, "loss": 0.0404, "step": 2740 }, { "epoch": 10.83399209486166, "grad_norm": 1.7851382493972778, "learning_rate": 2.296039603960396e-05, "loss": 0.0516, "step": 2741 }, { "epoch": 10.837944664031621, "grad_norm": 1.7465606927871704, "learning_rate": 2.2950495049504952e-05, "loss": 0.0535, "step": 2742 }, { "epoch": 10.841897233201582, "grad_norm": 1.5415796041488647, "learning_rate": 2.2940594059405943e-05, "loss": 0.0461, "step": 2743 }, { "epoch": 10.845849802371541, "grad_norm": 1.358091115951538, "learning_rate": 2.2930693069306934e-05, "loss": 0.0376, "step": 2744 }, { "epoch": 10.849802371541502, "grad_norm": 1.1826964616775513, "learning_rate": 2.2920792079207922e-05, "loss": 0.034, "step": 2745 }, { "epoch": 10.853754940711463, "grad_norm": 1.793292760848999, "learning_rate": 2.2910891089108913e-05, "loss": 0.0695, "step": 2746 }, { "epoch": 10.857707509881424, "grad_norm": 1.5757602453231812, "learning_rate": 2.2900990099009904e-05, "loss": 0.0459, "step": 2747 }, { "epoch": 10.861660079051383, "grad_norm": 1.559159755706787, "learning_rate": 2.289108910891089e-05, "loss": 0.0482, "step": 2748 }, { "epoch": 10.865612648221344, "grad_norm": 2.000089168548584, "learning_rate": 2.2881188118811883e-05, "loss": 0.064, "step": 2749 }, { "epoch": 10.869565217391305, "grad_norm": 1.3605917692184448, "learning_rate": 2.2871287128712874e-05, "loss": 0.0382, "step": 2750 }, { "epoch": 10.873517786561266, "grad_norm": 1.6145367622375488, "learning_rate": 2.2861386138613865e-05, "loss": 0.0482, "step": 2751 }, { "epoch": 10.877470355731225, "grad_norm": 1.563657283782959, "learning_rate": 2.2851485148514852e-05, "loss": 0.0503, "step": 2752 }, { "epoch": 10.881422924901186, "grad_norm": 1.1347399950027466, "learning_rate": 2.2841584158415844e-05, "loss": 0.0393, "step": 2753 }, { "epoch": 10.885375494071146, "grad_norm": 1.5107812881469727, "learning_rate": 2.2831683168316835e-05, "loss": 0.0438, "step": 2754 }, { "epoch": 10.889328063241107, "grad_norm": 1.4366496801376343, "learning_rate": 2.2821782178217822e-05, "loss": 0.0424, "step": 2755 }, { "epoch": 10.893280632411066, "grad_norm": 1.275848627090454, "learning_rate": 2.2811881188118813e-05, "loss": 0.0397, "step": 2756 }, { "epoch": 10.897233201581027, "grad_norm": 1.6528239250183105, "learning_rate": 2.2801980198019804e-05, "loss": 0.046, "step": 2757 }, { "epoch": 10.901185770750988, "grad_norm": 1.9767048358917236, "learning_rate": 2.2792079207920795e-05, "loss": 0.0586, "step": 2758 }, { "epoch": 10.90513833992095, "grad_norm": 1.4716681241989136, "learning_rate": 2.2782178217821783e-05, "loss": 0.0385, "step": 2759 }, { "epoch": 10.909090909090908, "grad_norm": 1.5717904567718506, "learning_rate": 2.2772277227722774e-05, "loss": 0.0462, "step": 2760 }, { "epoch": 10.91304347826087, "grad_norm": 1.5703141689300537, "learning_rate": 2.2762376237623765e-05, "loss": 0.0416, "step": 2761 }, { "epoch": 10.91699604743083, "grad_norm": 1.5964816808700562, "learning_rate": 2.2752475247524753e-05, "loss": 0.0425, "step": 2762 }, { "epoch": 10.920948616600791, "grad_norm": 2.2897305488586426, "learning_rate": 2.2742574257425744e-05, "loss": 0.0762, "step": 2763 }, { "epoch": 10.92490118577075, "grad_norm": 1.7323611974716187, "learning_rate": 2.2732673267326735e-05, "loss": 0.042, "step": 2764 }, { "epoch": 10.928853754940711, "grad_norm": 1.4680981636047363, "learning_rate": 2.2722772277227726e-05, "loss": 0.0495, "step": 2765 }, { "epoch": 10.932806324110672, "grad_norm": 1.7502238750457764, "learning_rate": 2.2712871287128714e-05, "loss": 0.0533, "step": 2766 }, { "epoch": 10.936758893280633, "grad_norm": 1.6299428939819336, "learning_rate": 2.2702970297029705e-05, "loss": 0.0446, "step": 2767 }, { "epoch": 10.940711462450594, "grad_norm": 1.658179521560669, "learning_rate": 2.2693069306930696e-05, "loss": 0.0409, "step": 2768 }, { "epoch": 10.944664031620553, "grad_norm": 1.369956374168396, "learning_rate": 2.2683168316831684e-05, "loss": 0.0437, "step": 2769 }, { "epoch": 10.948616600790514, "grad_norm": 1.263959288597107, "learning_rate": 2.2673267326732675e-05, "loss": 0.0353, "step": 2770 }, { "epoch": 10.952569169960475, "grad_norm": 1.7411878108978271, "learning_rate": 2.2663366336633666e-05, "loss": 0.0503, "step": 2771 }, { "epoch": 10.956521739130435, "grad_norm": 1.4506467580795288, "learning_rate": 2.2653465346534657e-05, "loss": 0.0419, "step": 2772 }, { "epoch": 10.960474308300395, "grad_norm": 1.5551061630249023, "learning_rate": 2.2643564356435645e-05, "loss": 0.047, "step": 2773 }, { "epoch": 10.964426877470355, "grad_norm": 1.9233285188674927, "learning_rate": 2.2633663366336636e-05, "loss": 0.058, "step": 2774 }, { "epoch": 10.968379446640316, "grad_norm": 1.1635584831237793, "learning_rate": 2.2623762376237627e-05, "loss": 0.0341, "step": 2775 }, { "epoch": 10.972332015810277, "grad_norm": 1.9155261516571045, "learning_rate": 2.2613861386138614e-05, "loss": 0.0494, "step": 2776 }, { "epoch": 10.976284584980236, "grad_norm": 1.5236390829086304, "learning_rate": 2.2603960396039605e-05, "loss": 0.0474, "step": 2777 }, { "epoch": 10.980237154150197, "grad_norm": 1.3742378950119019, "learning_rate": 2.2594059405940596e-05, "loss": 0.039, "step": 2778 }, { "epoch": 10.984189723320158, "grad_norm": 1.5048751831054688, "learning_rate": 2.2584158415841588e-05, "loss": 0.0487, "step": 2779 }, { "epoch": 10.988142292490119, "grad_norm": 1.4429644346237183, "learning_rate": 2.2574257425742575e-05, "loss": 0.044, "step": 2780 }, { "epoch": 10.992094861660078, "grad_norm": 1.6993399858474731, "learning_rate": 2.2564356435643566e-05, "loss": 0.0501, "step": 2781 }, { "epoch": 10.996047430830039, "grad_norm": 1.2944649457931519, "learning_rate": 2.2554455445544557e-05, "loss": 0.0382, "step": 2782 }, { "epoch": 11.0, "grad_norm": 1.83441162109375, "learning_rate": 2.2544554455445545e-05, "loss": 0.0535, "step": 2783 }, { "epoch": 11.003952569169961, "grad_norm": 1.3442505598068237, "learning_rate": 2.2534653465346536e-05, "loss": 0.0378, "step": 2784 }, { "epoch": 11.007905138339922, "grad_norm": 0.9354000091552734, "learning_rate": 2.2524752475247527e-05, "loss": 0.0287, "step": 2785 }, { "epoch": 11.011857707509881, "grad_norm": 1.1196565628051758, "learning_rate": 2.2514851485148515e-05, "loss": 0.0327, "step": 2786 }, { "epoch": 11.015810276679842, "grad_norm": 1.013670802116394, "learning_rate": 2.2504950495049506e-05, "loss": 0.0281, "step": 2787 }, { "epoch": 11.019762845849803, "grad_norm": 1.1460349559783936, "learning_rate": 2.2495049504950497e-05, "loss": 0.0431, "step": 2788 }, { "epoch": 11.023715415019764, "grad_norm": 1.145628809928894, "learning_rate": 2.2485148514851488e-05, "loss": 0.0294, "step": 2789 }, { "epoch": 11.027667984189723, "grad_norm": 1.4844728708267212, "learning_rate": 2.2475247524752476e-05, "loss": 0.0407, "step": 2790 }, { "epoch": 11.031620553359684, "grad_norm": 1.8009521961212158, "learning_rate": 2.2465346534653467e-05, "loss": 0.0551, "step": 2791 }, { "epoch": 11.035573122529645, "grad_norm": 0.7155687808990479, "learning_rate": 2.2455445544554458e-05, "loss": 0.0233, "step": 2792 }, { "epoch": 11.039525691699605, "grad_norm": 1.3494378328323364, "learning_rate": 2.2445544554455446e-05, "loss": 0.0396, "step": 2793 }, { "epoch": 11.043478260869565, "grad_norm": 1.0340697765350342, "learning_rate": 2.2435643564356437e-05, "loss": 0.0306, "step": 2794 }, { "epoch": 11.047430830039525, "grad_norm": 1.2561464309692383, "learning_rate": 2.2425742574257428e-05, "loss": 0.0323, "step": 2795 }, { "epoch": 11.051383399209486, "grad_norm": 1.491833209991455, "learning_rate": 2.241584158415842e-05, "loss": 0.0334, "step": 2796 }, { "epoch": 11.055335968379447, "grad_norm": 1.4058970212936401, "learning_rate": 2.2405940594059406e-05, "loss": 0.0388, "step": 2797 }, { "epoch": 11.059288537549406, "grad_norm": 1.2275718450546265, "learning_rate": 2.2396039603960397e-05, "loss": 0.0365, "step": 2798 }, { "epoch": 11.063241106719367, "grad_norm": 1.0804698467254639, "learning_rate": 2.238613861386139e-05, "loss": 0.0287, "step": 2799 }, { "epoch": 11.067193675889328, "grad_norm": 1.6424906253814697, "learning_rate": 2.2376237623762376e-05, "loss": 0.05, "step": 2800 }, { "epoch": 11.071146245059289, "grad_norm": 1.1693159341812134, "learning_rate": 2.2366336633663367e-05, "loss": 0.0322, "step": 2801 }, { "epoch": 11.075098814229248, "grad_norm": 1.089708685874939, "learning_rate": 2.235643564356436e-05, "loss": 0.0325, "step": 2802 }, { "epoch": 11.079051383399209, "grad_norm": 0.9867976307868958, "learning_rate": 2.234653465346535e-05, "loss": 0.0269, "step": 2803 }, { "epoch": 11.08300395256917, "grad_norm": 1.6470941305160522, "learning_rate": 2.2336633663366337e-05, "loss": 0.0416, "step": 2804 }, { "epoch": 11.08695652173913, "grad_norm": 1.3164324760437012, "learning_rate": 2.2326732673267328e-05, "loss": 0.0331, "step": 2805 }, { "epoch": 11.090909090909092, "grad_norm": 1.210435390472412, "learning_rate": 2.231683168316832e-05, "loss": 0.0416, "step": 2806 }, { "epoch": 11.09486166007905, "grad_norm": 1.4648571014404297, "learning_rate": 2.2306930693069307e-05, "loss": 0.0469, "step": 2807 }, { "epoch": 11.098814229249012, "grad_norm": 1.9764188528060913, "learning_rate": 2.2297029702970298e-05, "loss": 0.0526, "step": 2808 }, { "epoch": 11.102766798418973, "grad_norm": 1.6987800598144531, "learning_rate": 2.228712871287129e-05, "loss": 0.0452, "step": 2809 }, { "epoch": 11.106719367588934, "grad_norm": 1.6511965990066528, "learning_rate": 2.227722772277228e-05, "loss": 0.0476, "step": 2810 }, { "epoch": 11.110671936758893, "grad_norm": 1.4550827741622925, "learning_rate": 2.2267326732673268e-05, "loss": 0.0325, "step": 2811 }, { "epoch": 11.114624505928854, "grad_norm": 1.259514570236206, "learning_rate": 2.225742574257426e-05, "loss": 0.0382, "step": 2812 }, { "epoch": 11.118577075098814, "grad_norm": 0.973564088344574, "learning_rate": 2.224752475247525e-05, "loss": 0.0265, "step": 2813 }, { "epoch": 11.122529644268775, "grad_norm": 1.324453592300415, "learning_rate": 2.2237623762376238e-05, "loss": 0.0338, "step": 2814 }, { "epoch": 11.126482213438734, "grad_norm": 1.0715748071670532, "learning_rate": 2.222772277227723e-05, "loss": 0.0293, "step": 2815 }, { "epoch": 11.130434782608695, "grad_norm": 1.6038068532943726, "learning_rate": 2.221782178217822e-05, "loss": 0.0485, "step": 2816 }, { "epoch": 11.134387351778656, "grad_norm": 0.9934631586074829, "learning_rate": 2.220792079207921e-05, "loss": 0.0317, "step": 2817 }, { "epoch": 11.138339920948617, "grad_norm": 1.345691442489624, "learning_rate": 2.21980198019802e-05, "loss": 0.0334, "step": 2818 }, { "epoch": 11.142292490118576, "grad_norm": 1.0704251527786255, "learning_rate": 2.218811881188119e-05, "loss": 0.0318, "step": 2819 }, { "epoch": 11.146245059288537, "grad_norm": 2.117433786392212, "learning_rate": 2.217821782178218e-05, "loss": 0.064, "step": 2820 }, { "epoch": 11.150197628458498, "grad_norm": 1.0934911966323853, "learning_rate": 2.2168316831683168e-05, "loss": 0.0296, "step": 2821 }, { "epoch": 11.154150197628459, "grad_norm": 1.4089946746826172, "learning_rate": 2.215841584158416e-05, "loss": 0.0361, "step": 2822 }, { "epoch": 11.15810276679842, "grad_norm": 1.1010053157806396, "learning_rate": 2.214851485148515e-05, "loss": 0.0317, "step": 2823 }, { "epoch": 11.162055335968379, "grad_norm": 1.2775205373764038, "learning_rate": 2.213861386138614e-05, "loss": 0.0374, "step": 2824 }, { "epoch": 11.16600790513834, "grad_norm": 1.2159770727157593, "learning_rate": 2.212871287128713e-05, "loss": 0.0408, "step": 2825 }, { "epoch": 11.1699604743083, "grad_norm": 1.2917265892028809, "learning_rate": 2.211881188118812e-05, "loss": 0.0404, "step": 2826 }, { "epoch": 11.173913043478262, "grad_norm": 1.0109119415283203, "learning_rate": 2.210891089108911e-05, "loss": 0.0277, "step": 2827 }, { "epoch": 11.17786561264822, "grad_norm": 1.3481512069702148, "learning_rate": 2.20990099009901e-05, "loss": 0.0378, "step": 2828 }, { "epoch": 11.181818181818182, "grad_norm": 1.4169065952301025, "learning_rate": 2.208910891089109e-05, "loss": 0.0443, "step": 2829 }, { "epoch": 11.185770750988143, "grad_norm": 0.8938689231872559, "learning_rate": 2.207920792079208e-05, "loss": 0.0255, "step": 2830 }, { "epoch": 11.189723320158103, "grad_norm": 1.7077336311340332, "learning_rate": 2.2069306930693072e-05, "loss": 0.0481, "step": 2831 }, { "epoch": 11.193675889328063, "grad_norm": 1.1939144134521484, "learning_rate": 2.205940594059406e-05, "loss": 0.0337, "step": 2832 }, { "epoch": 11.197628458498023, "grad_norm": 1.1766804456710815, "learning_rate": 2.204950495049505e-05, "loss": 0.0383, "step": 2833 }, { "epoch": 11.201581027667984, "grad_norm": 1.0796892642974854, "learning_rate": 2.2039603960396042e-05, "loss": 0.0315, "step": 2834 }, { "epoch": 11.205533596837945, "grad_norm": 1.357995867729187, "learning_rate": 2.202970297029703e-05, "loss": 0.0433, "step": 2835 }, { "epoch": 11.209486166007904, "grad_norm": 1.6146154403686523, "learning_rate": 2.201980198019802e-05, "loss": 0.0495, "step": 2836 }, { "epoch": 11.213438735177865, "grad_norm": 1.2158877849578857, "learning_rate": 2.2009900990099012e-05, "loss": 0.04, "step": 2837 }, { "epoch": 11.217391304347826, "grad_norm": 1.3769692182540894, "learning_rate": 2.2000000000000003e-05, "loss": 0.0358, "step": 2838 }, { "epoch": 11.221343873517787, "grad_norm": 1.7808622121810913, "learning_rate": 2.199009900990099e-05, "loss": 0.04, "step": 2839 }, { "epoch": 11.225296442687746, "grad_norm": 1.347683310508728, "learning_rate": 2.198019801980198e-05, "loss": 0.0396, "step": 2840 }, { "epoch": 11.229249011857707, "grad_norm": 1.1763908863067627, "learning_rate": 2.1970297029702973e-05, "loss": 0.0302, "step": 2841 }, { "epoch": 11.233201581027668, "grad_norm": 1.7768423557281494, "learning_rate": 2.196039603960396e-05, "loss": 0.0487, "step": 2842 }, { "epoch": 11.237154150197629, "grad_norm": 1.145398497581482, "learning_rate": 2.195049504950495e-05, "loss": 0.035, "step": 2843 }, { "epoch": 11.24110671936759, "grad_norm": 1.6971150636672974, "learning_rate": 2.1940594059405942e-05, "loss": 0.0476, "step": 2844 }, { "epoch": 11.245059288537549, "grad_norm": 1.565805196762085, "learning_rate": 2.193069306930693e-05, "loss": 0.0461, "step": 2845 }, { "epoch": 11.24901185770751, "grad_norm": 1.6725757122039795, "learning_rate": 2.192079207920792e-05, "loss": 0.0439, "step": 2846 }, { "epoch": 11.25296442687747, "grad_norm": 1.3251878023147583, "learning_rate": 2.1910891089108912e-05, "loss": 0.0373, "step": 2847 }, { "epoch": 11.256916996047432, "grad_norm": 1.1810892820358276, "learning_rate": 2.1900990099009903e-05, "loss": 0.0323, "step": 2848 }, { "epoch": 11.26086956521739, "grad_norm": 1.156447410583496, "learning_rate": 2.189108910891089e-05, "loss": 0.0336, "step": 2849 }, { "epoch": 11.264822134387352, "grad_norm": 1.0476000308990479, "learning_rate": 2.1881188118811882e-05, "loss": 0.0314, "step": 2850 }, { "epoch": 11.268774703557312, "grad_norm": 1.672797441482544, "learning_rate": 2.1871287128712873e-05, "loss": 0.0521, "step": 2851 }, { "epoch": 11.272727272727273, "grad_norm": 1.1781623363494873, "learning_rate": 2.186138613861386e-05, "loss": 0.0314, "step": 2852 }, { "epoch": 11.276679841897232, "grad_norm": 1.4425806999206543, "learning_rate": 2.1851485148514852e-05, "loss": 0.0417, "step": 2853 }, { "epoch": 11.280632411067193, "grad_norm": 0.9254428148269653, "learning_rate": 2.1841584158415843e-05, "loss": 0.0428, "step": 2854 }, { "epoch": 11.284584980237154, "grad_norm": 1.1138046979904175, "learning_rate": 2.1831683168316834e-05, "loss": 0.031, "step": 2855 }, { "epoch": 11.288537549407115, "grad_norm": 1.6439061164855957, "learning_rate": 2.182178217821782e-05, "loss": 0.0409, "step": 2856 }, { "epoch": 11.292490118577074, "grad_norm": 1.3690448999404907, "learning_rate": 2.1811881188118813e-05, "loss": 0.0488, "step": 2857 }, { "epoch": 11.296442687747035, "grad_norm": 1.435912847518921, "learning_rate": 2.1801980198019804e-05, "loss": 0.0445, "step": 2858 }, { "epoch": 11.300395256916996, "grad_norm": 1.7691850662231445, "learning_rate": 2.179207920792079e-05, "loss": 0.048, "step": 2859 }, { "epoch": 11.304347826086957, "grad_norm": 1.2059659957885742, "learning_rate": 2.1782178217821783e-05, "loss": 0.0347, "step": 2860 }, { "epoch": 11.308300395256918, "grad_norm": 1.5445821285247803, "learning_rate": 2.1772277227722774e-05, "loss": 0.046, "step": 2861 }, { "epoch": 11.312252964426877, "grad_norm": 1.0638519525527954, "learning_rate": 2.1762376237623765e-05, "loss": 0.0313, "step": 2862 }, { "epoch": 11.316205533596838, "grad_norm": 1.225050687789917, "learning_rate": 2.1752475247524752e-05, "loss": 0.0368, "step": 2863 }, { "epoch": 11.320158102766799, "grad_norm": 1.6228591203689575, "learning_rate": 2.1742574257425743e-05, "loss": 0.0449, "step": 2864 }, { "epoch": 11.32411067193676, "grad_norm": 1.4313108921051025, "learning_rate": 2.1732673267326734e-05, "loss": 0.0397, "step": 2865 }, { "epoch": 11.328063241106719, "grad_norm": 1.366606593132019, "learning_rate": 2.1722772277227722e-05, "loss": 0.0464, "step": 2866 }, { "epoch": 11.33201581027668, "grad_norm": 1.1494160890579224, "learning_rate": 2.1712871287128713e-05, "loss": 0.0345, "step": 2867 }, { "epoch": 11.33596837944664, "grad_norm": 1.0085707902908325, "learning_rate": 2.1702970297029704e-05, "loss": 0.0293, "step": 2868 }, { "epoch": 11.339920948616601, "grad_norm": 1.110657811164856, "learning_rate": 2.1693069306930695e-05, "loss": 0.0317, "step": 2869 }, { "epoch": 11.34387351778656, "grad_norm": 1.4241842031478882, "learning_rate": 2.1683168316831683e-05, "loss": 0.036, "step": 2870 }, { "epoch": 11.347826086956522, "grad_norm": 1.8218530416488647, "learning_rate": 2.1673267326732674e-05, "loss": 0.054, "step": 2871 }, { "epoch": 11.351778656126482, "grad_norm": 1.0528109073638916, "learning_rate": 2.1663366336633665e-05, "loss": 0.0254, "step": 2872 }, { "epoch": 11.355731225296443, "grad_norm": 1.872027039527893, "learning_rate": 2.1653465346534653e-05, "loss": 0.0564, "step": 2873 }, { "epoch": 11.359683794466402, "grad_norm": 2.24637508392334, "learning_rate": 2.1643564356435644e-05, "loss": 0.0571, "step": 2874 }, { "epoch": 11.363636363636363, "grad_norm": 1.5424247980117798, "learning_rate": 2.1633663366336635e-05, "loss": 0.0431, "step": 2875 }, { "epoch": 11.367588932806324, "grad_norm": 1.0103260278701782, "learning_rate": 2.1623762376237626e-05, "loss": 0.0307, "step": 2876 }, { "epoch": 11.371541501976285, "grad_norm": 0.7997999787330627, "learning_rate": 2.1613861386138614e-05, "loss": 0.0292, "step": 2877 }, { "epoch": 11.375494071146244, "grad_norm": 0.9349464774131775, "learning_rate": 2.1603960396039605e-05, "loss": 0.0274, "step": 2878 }, { "epoch": 11.379446640316205, "grad_norm": 1.3485474586486816, "learning_rate": 2.1594059405940596e-05, "loss": 0.0425, "step": 2879 }, { "epoch": 11.383399209486166, "grad_norm": 1.2241548299789429, "learning_rate": 2.1584158415841584e-05, "loss": 0.04, "step": 2880 }, { "epoch": 11.387351778656127, "grad_norm": 1.323196291923523, "learning_rate": 2.1574257425742575e-05, "loss": 0.0417, "step": 2881 }, { "epoch": 11.391304347826088, "grad_norm": 1.3765828609466553, "learning_rate": 2.1564356435643566e-05, "loss": 0.0388, "step": 2882 }, { "epoch": 11.395256916996047, "grad_norm": 1.1343966722488403, "learning_rate": 2.1554455445544557e-05, "loss": 0.0304, "step": 2883 }, { "epoch": 11.399209486166008, "grad_norm": 1.1014480590820312, "learning_rate": 2.1544554455445544e-05, "loss": 0.0346, "step": 2884 }, { "epoch": 11.403162055335969, "grad_norm": 2.5743091106414795, "learning_rate": 2.1534653465346535e-05, "loss": 0.0952, "step": 2885 }, { "epoch": 11.40711462450593, "grad_norm": 2.003746747970581, "learning_rate": 2.1524752475247527e-05, "loss": 0.0505, "step": 2886 }, { "epoch": 11.411067193675889, "grad_norm": 1.5372415781021118, "learning_rate": 2.1514851485148514e-05, "loss": 0.0461, "step": 2887 }, { "epoch": 11.41501976284585, "grad_norm": 0.9136408567428589, "learning_rate": 2.1504950495049505e-05, "loss": 0.0306, "step": 2888 }, { "epoch": 11.41897233201581, "grad_norm": 1.392761468887329, "learning_rate": 2.1495049504950496e-05, "loss": 0.0406, "step": 2889 }, { "epoch": 11.422924901185771, "grad_norm": 1.2603839635849, "learning_rate": 2.1485148514851487e-05, "loss": 0.0413, "step": 2890 }, { "epoch": 11.42687747035573, "grad_norm": 1.116632103919983, "learning_rate": 2.1475247524752475e-05, "loss": 0.0331, "step": 2891 }, { "epoch": 11.430830039525691, "grad_norm": 1.2441052198410034, "learning_rate": 2.1465346534653466e-05, "loss": 0.0367, "step": 2892 }, { "epoch": 11.434782608695652, "grad_norm": 1.7843313217163086, "learning_rate": 2.1455445544554457e-05, "loss": 0.055, "step": 2893 }, { "epoch": 11.438735177865613, "grad_norm": 1.6412118673324585, "learning_rate": 2.1445544554455445e-05, "loss": 0.0465, "step": 2894 }, { "epoch": 11.442687747035572, "grad_norm": 1.2968428134918213, "learning_rate": 2.1435643564356436e-05, "loss": 0.0371, "step": 2895 }, { "epoch": 11.446640316205533, "grad_norm": 1.5063434839248657, "learning_rate": 2.1425742574257427e-05, "loss": 0.0421, "step": 2896 }, { "epoch": 11.450592885375494, "grad_norm": 1.2456010580062866, "learning_rate": 2.1415841584158418e-05, "loss": 0.0351, "step": 2897 }, { "epoch": 11.454545454545455, "grad_norm": 1.5236023664474487, "learning_rate": 2.1405940594059406e-05, "loss": 0.0427, "step": 2898 }, { "epoch": 11.458498023715414, "grad_norm": 0.817237377166748, "learning_rate": 2.1396039603960397e-05, "loss": 0.0281, "step": 2899 }, { "epoch": 11.462450592885375, "grad_norm": 1.1599724292755127, "learning_rate": 2.1386138613861388e-05, "loss": 0.038, "step": 2900 }, { "epoch": 11.466403162055336, "grad_norm": 1.5484449863433838, "learning_rate": 2.1376237623762376e-05, "loss": 0.0471, "step": 2901 }, { "epoch": 11.470355731225297, "grad_norm": 1.9765840768814087, "learning_rate": 2.1366336633663367e-05, "loss": 0.0627, "step": 2902 }, { "epoch": 11.474308300395258, "grad_norm": 0.9364761114120483, "learning_rate": 2.1356435643564358e-05, "loss": 0.0283, "step": 2903 }, { "epoch": 11.478260869565217, "grad_norm": 1.6713351011276245, "learning_rate": 2.1346534653465345e-05, "loss": 0.0447, "step": 2904 }, { "epoch": 11.482213438735178, "grad_norm": 1.2112828493118286, "learning_rate": 2.1336633663366336e-05, "loss": 0.0305, "step": 2905 }, { "epoch": 11.486166007905139, "grad_norm": 1.0910948514938354, "learning_rate": 2.1326732673267328e-05, "loss": 0.0313, "step": 2906 }, { "epoch": 11.4901185770751, "grad_norm": 1.3914623260498047, "learning_rate": 2.131683168316832e-05, "loss": 0.0415, "step": 2907 }, { "epoch": 11.494071146245059, "grad_norm": 1.1036934852600098, "learning_rate": 2.1306930693069306e-05, "loss": 0.0301, "step": 2908 }, { "epoch": 11.49802371541502, "grad_norm": 1.556960105895996, "learning_rate": 2.1297029702970297e-05, "loss": 0.0424, "step": 2909 }, { "epoch": 11.50197628458498, "grad_norm": 1.5532281398773193, "learning_rate": 2.128712871287129e-05, "loss": 0.0516, "step": 2910 }, { "epoch": 11.505928853754941, "grad_norm": 1.1597799062728882, "learning_rate": 2.1277227722772276e-05, "loss": 0.0333, "step": 2911 }, { "epoch": 11.5098814229249, "grad_norm": 1.2698713541030884, "learning_rate": 2.1267326732673267e-05, "loss": 0.0338, "step": 2912 }, { "epoch": 11.513833992094861, "grad_norm": 1.0219616889953613, "learning_rate": 2.1257425742574258e-05, "loss": 0.031, "step": 2913 }, { "epoch": 11.517786561264822, "grad_norm": 1.7600486278533936, "learning_rate": 2.124752475247525e-05, "loss": 0.0486, "step": 2914 }, { "epoch": 11.521739130434783, "grad_norm": 1.5489503145217896, "learning_rate": 2.1237623762376237e-05, "loss": 0.0465, "step": 2915 }, { "epoch": 11.525691699604742, "grad_norm": 1.242188572883606, "learning_rate": 2.1227722772277228e-05, "loss": 0.039, "step": 2916 }, { "epoch": 11.529644268774703, "grad_norm": 1.5549061298370361, "learning_rate": 2.121782178217822e-05, "loss": 0.0412, "step": 2917 }, { "epoch": 11.533596837944664, "grad_norm": 1.0485998392105103, "learning_rate": 2.1207920792079207e-05, "loss": 0.0296, "step": 2918 }, { "epoch": 11.537549407114625, "grad_norm": 1.6895643472671509, "learning_rate": 2.1198019801980198e-05, "loss": 0.0504, "step": 2919 }, { "epoch": 11.541501976284586, "grad_norm": 1.785576581954956, "learning_rate": 2.118811881188119e-05, "loss": 0.0491, "step": 2920 }, { "epoch": 11.545454545454545, "grad_norm": 1.5142230987548828, "learning_rate": 2.117821782178218e-05, "loss": 0.0471, "step": 2921 }, { "epoch": 11.549407114624506, "grad_norm": 1.2696739435195923, "learning_rate": 2.1168316831683168e-05, "loss": 0.0377, "step": 2922 }, { "epoch": 11.553359683794467, "grad_norm": 1.373161792755127, "learning_rate": 2.115841584158416e-05, "loss": 0.0388, "step": 2923 }, { "epoch": 11.557312252964428, "grad_norm": 1.0092976093292236, "learning_rate": 2.114851485148515e-05, "loss": 0.0358, "step": 2924 }, { "epoch": 11.561264822134387, "grad_norm": 1.2542871236801147, "learning_rate": 2.1138613861386137e-05, "loss": 0.033, "step": 2925 }, { "epoch": 11.565217391304348, "grad_norm": 1.043573021888733, "learning_rate": 2.112871287128713e-05, "loss": 0.0316, "step": 2926 }, { "epoch": 11.569169960474309, "grad_norm": 1.4796109199523926, "learning_rate": 2.111881188118812e-05, "loss": 0.045, "step": 2927 }, { "epoch": 11.57312252964427, "grad_norm": 1.0325745344161987, "learning_rate": 2.110891089108911e-05, "loss": 0.0313, "step": 2928 }, { "epoch": 11.577075098814229, "grad_norm": 1.0866289138793945, "learning_rate": 2.10990099009901e-05, "loss": 0.0352, "step": 2929 }, { "epoch": 11.58102766798419, "grad_norm": 1.3573859930038452, "learning_rate": 2.108910891089109e-05, "loss": 0.037, "step": 2930 }, { "epoch": 11.58498023715415, "grad_norm": 1.161630392074585, "learning_rate": 2.107920792079208e-05, "loss": 0.0312, "step": 2931 }, { "epoch": 11.588932806324111, "grad_norm": 1.7953673601150513, "learning_rate": 2.1069306930693068e-05, "loss": 0.0441, "step": 2932 }, { "epoch": 11.59288537549407, "grad_norm": 1.3369241952896118, "learning_rate": 2.105940594059406e-05, "loss": 0.0413, "step": 2933 }, { "epoch": 11.596837944664031, "grad_norm": 1.3186440467834473, "learning_rate": 2.104950495049505e-05, "loss": 0.0427, "step": 2934 }, { "epoch": 11.600790513833992, "grad_norm": 1.54726243019104, "learning_rate": 2.103960396039604e-05, "loss": 0.05, "step": 2935 }, { "epoch": 11.604743083003953, "grad_norm": 1.6058597564697266, "learning_rate": 2.102970297029703e-05, "loss": 0.0447, "step": 2936 }, { "epoch": 11.608695652173914, "grad_norm": 1.4869719743728638, "learning_rate": 2.101980198019802e-05, "loss": 0.0432, "step": 2937 }, { "epoch": 11.612648221343873, "grad_norm": 1.3169610500335693, "learning_rate": 2.100990099009901e-05, "loss": 0.0376, "step": 2938 }, { "epoch": 11.616600790513834, "grad_norm": 1.4671881198883057, "learning_rate": 2.1e-05, "loss": 0.0423, "step": 2939 }, { "epoch": 11.620553359683795, "grad_norm": 1.6210981607437134, "learning_rate": 2.099009900990099e-05, "loss": 0.0402, "step": 2940 }, { "epoch": 11.624505928853754, "grad_norm": 1.2259242534637451, "learning_rate": 2.098019801980198e-05, "loss": 0.0358, "step": 2941 }, { "epoch": 11.628458498023715, "grad_norm": 1.2353581190109253, "learning_rate": 2.0970297029702972e-05, "loss": 0.0356, "step": 2942 }, { "epoch": 11.632411067193676, "grad_norm": 1.4206687211990356, "learning_rate": 2.096039603960396e-05, "loss": 0.037, "step": 2943 }, { "epoch": 11.636363636363637, "grad_norm": 1.2210633754730225, "learning_rate": 2.095049504950495e-05, "loss": 0.0335, "step": 2944 }, { "epoch": 11.640316205533598, "grad_norm": 1.282558798789978, "learning_rate": 2.0940594059405942e-05, "loss": 0.0445, "step": 2945 }, { "epoch": 11.644268774703557, "grad_norm": 1.3816938400268555, "learning_rate": 2.093069306930693e-05, "loss": 0.0371, "step": 2946 }, { "epoch": 11.648221343873518, "grad_norm": 1.1143611669540405, "learning_rate": 2.092079207920792e-05, "loss": 0.0372, "step": 2947 }, { "epoch": 11.652173913043478, "grad_norm": 1.5145795345306396, "learning_rate": 2.091089108910891e-05, "loss": 0.0475, "step": 2948 }, { "epoch": 11.65612648221344, "grad_norm": 1.2481929063796997, "learning_rate": 2.0900990099009903e-05, "loss": 0.0348, "step": 2949 }, { "epoch": 11.660079051383399, "grad_norm": 1.0681403875350952, "learning_rate": 2.089108910891089e-05, "loss": 0.0308, "step": 2950 }, { "epoch": 11.66403162055336, "grad_norm": 1.0421465635299683, "learning_rate": 2.088118811881188e-05, "loss": 0.0358, "step": 2951 }, { "epoch": 11.66798418972332, "grad_norm": 1.4123704433441162, "learning_rate": 2.0871287128712873e-05, "loss": 0.037, "step": 2952 }, { "epoch": 11.671936758893281, "grad_norm": 2.1731958389282227, "learning_rate": 2.086138613861386e-05, "loss": 0.0537, "step": 2953 }, { "epoch": 11.67588932806324, "grad_norm": 1.316388487815857, "learning_rate": 2.085148514851485e-05, "loss": 0.0402, "step": 2954 }, { "epoch": 11.679841897233201, "grad_norm": 1.0978331565856934, "learning_rate": 2.0841584158415842e-05, "loss": 0.0392, "step": 2955 }, { "epoch": 11.683794466403162, "grad_norm": 1.6348326206207275, "learning_rate": 2.0831683168316833e-05, "loss": 0.0465, "step": 2956 }, { "epoch": 11.687747035573123, "grad_norm": 1.281080722808838, "learning_rate": 2.082178217821782e-05, "loss": 0.0415, "step": 2957 }, { "epoch": 11.691699604743082, "grad_norm": 1.5223556756973267, "learning_rate": 2.0811881188118812e-05, "loss": 0.0403, "step": 2958 }, { "epoch": 11.695652173913043, "grad_norm": 1.0688214302062988, "learning_rate": 2.0801980198019803e-05, "loss": 0.0296, "step": 2959 }, { "epoch": 11.699604743083004, "grad_norm": 1.4109771251678467, "learning_rate": 2.079207920792079e-05, "loss": 0.0396, "step": 2960 }, { "epoch": 11.703557312252965, "grad_norm": 1.2959903478622437, "learning_rate": 2.0782178217821782e-05, "loss": 0.0388, "step": 2961 }, { "epoch": 11.707509881422926, "grad_norm": 1.4684207439422607, "learning_rate": 2.0772277227722773e-05, "loss": 0.0458, "step": 2962 }, { "epoch": 11.711462450592885, "grad_norm": 1.7478736639022827, "learning_rate": 2.076237623762376e-05, "loss": 0.0493, "step": 2963 }, { "epoch": 11.715415019762846, "grad_norm": 1.7139337062835693, "learning_rate": 2.0752475247524752e-05, "loss": 0.0406, "step": 2964 }, { "epoch": 11.719367588932807, "grad_norm": 1.211409091949463, "learning_rate": 2.0742574257425743e-05, "loss": 0.0334, "step": 2965 }, { "epoch": 11.723320158102768, "grad_norm": 1.4313125610351562, "learning_rate": 2.0732673267326734e-05, "loss": 0.0374, "step": 2966 }, { "epoch": 11.727272727272727, "grad_norm": 0.8586317300796509, "learning_rate": 2.072277227722772e-05, "loss": 0.0264, "step": 2967 }, { "epoch": 11.731225296442688, "grad_norm": 1.586061954498291, "learning_rate": 2.0712871287128713e-05, "loss": 0.0435, "step": 2968 }, { "epoch": 11.735177865612648, "grad_norm": 1.3905242681503296, "learning_rate": 2.0702970297029704e-05, "loss": 0.0398, "step": 2969 }, { "epoch": 11.73913043478261, "grad_norm": 1.483136773109436, "learning_rate": 2.069306930693069e-05, "loss": 0.0443, "step": 2970 }, { "epoch": 11.743083003952568, "grad_norm": 1.0777661800384521, "learning_rate": 2.0683168316831682e-05, "loss": 0.031, "step": 2971 }, { "epoch": 11.74703557312253, "grad_norm": 1.3226211071014404, "learning_rate": 2.0673267326732674e-05, "loss": 0.0364, "step": 2972 }, { "epoch": 11.75098814229249, "grad_norm": 1.4583830833435059, "learning_rate": 2.0663366336633665e-05, "loss": 0.05, "step": 2973 }, { "epoch": 11.754940711462451, "grad_norm": 1.4899508953094482, "learning_rate": 2.0653465346534652e-05, "loss": 0.0471, "step": 2974 }, { "epoch": 11.75889328063241, "grad_norm": 1.5548717975616455, "learning_rate": 2.0643564356435643e-05, "loss": 0.0385, "step": 2975 }, { "epoch": 11.762845849802371, "grad_norm": 1.2757103443145752, "learning_rate": 2.0633663366336634e-05, "loss": 0.0358, "step": 2976 }, { "epoch": 11.766798418972332, "grad_norm": 1.087137222290039, "learning_rate": 2.0623762376237622e-05, "loss": 0.0309, "step": 2977 }, { "epoch": 11.770750988142293, "grad_norm": 1.178876519203186, "learning_rate": 2.0613861386138613e-05, "loss": 0.0338, "step": 2978 }, { "epoch": 11.774703557312254, "grad_norm": 1.3368802070617676, "learning_rate": 2.0603960396039604e-05, "loss": 0.0388, "step": 2979 }, { "epoch": 11.778656126482213, "grad_norm": 1.2650136947631836, "learning_rate": 2.0594059405940595e-05, "loss": 0.0371, "step": 2980 }, { "epoch": 11.782608695652174, "grad_norm": 1.3025778532028198, "learning_rate": 2.0584158415841583e-05, "loss": 0.0362, "step": 2981 }, { "epoch": 11.786561264822135, "grad_norm": 1.9464139938354492, "learning_rate": 2.0574257425742574e-05, "loss": 0.0468, "step": 2982 }, { "epoch": 11.790513833992096, "grad_norm": 1.48578679561615, "learning_rate": 2.0564356435643565e-05, "loss": 0.0379, "step": 2983 }, { "epoch": 11.794466403162055, "grad_norm": 1.0006580352783203, "learning_rate": 2.0554455445544553e-05, "loss": 0.0315, "step": 2984 }, { "epoch": 11.798418972332016, "grad_norm": 1.3833731412887573, "learning_rate": 2.0544554455445544e-05, "loss": 0.0398, "step": 2985 }, { "epoch": 11.802371541501977, "grad_norm": 1.053807020187378, "learning_rate": 2.0534653465346535e-05, "loss": 0.031, "step": 2986 }, { "epoch": 11.806324110671937, "grad_norm": 1.84579598903656, "learning_rate": 2.0524752475247526e-05, "loss": 0.0469, "step": 2987 }, { "epoch": 11.810276679841897, "grad_norm": 0.9380949139595032, "learning_rate": 2.0514851485148514e-05, "loss": 0.0291, "step": 2988 }, { "epoch": 11.814229249011857, "grad_norm": 0.9448922872543335, "learning_rate": 2.0504950495049505e-05, "loss": 0.0259, "step": 2989 }, { "epoch": 11.818181818181818, "grad_norm": 1.1499254703521729, "learning_rate": 2.0495049504950496e-05, "loss": 0.0336, "step": 2990 }, { "epoch": 11.82213438735178, "grad_norm": 1.6964282989501953, "learning_rate": 2.0485148514851483e-05, "loss": 0.0544, "step": 2991 }, { "epoch": 11.826086956521738, "grad_norm": 1.2832599878311157, "learning_rate": 2.0475247524752475e-05, "loss": 0.0375, "step": 2992 }, { "epoch": 11.8300395256917, "grad_norm": 1.2776920795440674, "learning_rate": 2.0465346534653466e-05, "loss": 0.0412, "step": 2993 }, { "epoch": 11.83399209486166, "grad_norm": 1.200699806213379, "learning_rate": 2.0455445544554457e-05, "loss": 0.0336, "step": 2994 }, { "epoch": 11.837944664031621, "grad_norm": 1.7698127031326294, "learning_rate": 2.0445544554455444e-05, "loss": 0.0475, "step": 2995 }, { "epoch": 11.841897233201582, "grad_norm": 1.4895843267440796, "learning_rate": 2.0435643564356435e-05, "loss": 0.0401, "step": 2996 }, { "epoch": 11.845849802371541, "grad_norm": 1.458608627319336, "learning_rate": 2.0425742574257426e-05, "loss": 0.0431, "step": 2997 }, { "epoch": 11.849802371541502, "grad_norm": 1.3402233123779297, "learning_rate": 2.0415841584158414e-05, "loss": 0.0409, "step": 2998 }, { "epoch": 11.853754940711463, "grad_norm": 1.125826358795166, "learning_rate": 2.0405940594059405e-05, "loss": 0.0351, "step": 2999 }, { "epoch": 11.857707509881424, "grad_norm": 1.1845731735229492, "learning_rate": 2.0396039603960396e-05, "loss": 0.033, "step": 3000 }, { "epoch": 11.861660079051383, "grad_norm": 1.2256560325622559, "learning_rate": 2.0386138613861387e-05, "loss": 0.0427, "step": 3001 }, { "epoch": 11.865612648221344, "grad_norm": 1.3706421852111816, "learning_rate": 2.0376237623762375e-05, "loss": 0.0335, "step": 3002 }, { "epoch": 11.869565217391305, "grad_norm": 1.349879503250122, "learning_rate": 2.0366336633663366e-05, "loss": 0.0372, "step": 3003 }, { "epoch": 11.873517786561266, "grad_norm": 1.1587773561477661, "learning_rate": 2.0356435643564357e-05, "loss": 0.0358, "step": 3004 }, { "epoch": 11.877470355731225, "grad_norm": 1.1806353330612183, "learning_rate": 2.0346534653465345e-05, "loss": 0.0338, "step": 3005 }, { "epoch": 11.881422924901186, "grad_norm": 1.2671016454696655, "learning_rate": 2.0336633663366336e-05, "loss": 0.0343, "step": 3006 }, { "epoch": 11.885375494071146, "grad_norm": 1.0877180099487305, "learning_rate": 2.0326732673267327e-05, "loss": 0.0307, "step": 3007 }, { "epoch": 11.889328063241107, "grad_norm": 1.30611252784729, "learning_rate": 2.0316831683168318e-05, "loss": 0.0461, "step": 3008 }, { "epoch": 11.893280632411066, "grad_norm": 1.2684941291809082, "learning_rate": 2.030693069306931e-05, "loss": 0.0421, "step": 3009 }, { "epoch": 11.897233201581027, "grad_norm": 1.705476999282837, "learning_rate": 2.02970297029703e-05, "loss": 0.0465, "step": 3010 }, { "epoch": 11.901185770750988, "grad_norm": 1.244522213935852, "learning_rate": 2.0287128712871288e-05, "loss": 0.0346, "step": 3011 }, { "epoch": 11.90513833992095, "grad_norm": 1.333739995956421, "learning_rate": 2.027722772277228e-05, "loss": 0.0413, "step": 3012 }, { "epoch": 11.909090909090908, "grad_norm": 1.3015385866165161, "learning_rate": 2.026732673267327e-05, "loss": 0.0373, "step": 3013 }, { "epoch": 11.91304347826087, "grad_norm": 1.5979206562042236, "learning_rate": 2.025742574257426e-05, "loss": 0.045, "step": 3014 }, { "epoch": 11.91699604743083, "grad_norm": 1.3123173713684082, "learning_rate": 2.024752475247525e-05, "loss": 0.0413, "step": 3015 }, { "epoch": 11.920948616600791, "grad_norm": 1.1444761753082275, "learning_rate": 2.023762376237624e-05, "loss": 0.033, "step": 3016 }, { "epoch": 11.92490118577075, "grad_norm": 1.2669376134872437, "learning_rate": 2.022772277227723e-05, "loss": 0.0457, "step": 3017 }, { "epoch": 11.928853754940711, "grad_norm": 1.2476246356964111, "learning_rate": 2.021782178217822e-05, "loss": 0.0328, "step": 3018 }, { "epoch": 11.932806324110672, "grad_norm": 1.0192649364471436, "learning_rate": 2.020792079207921e-05, "loss": 0.043, "step": 3019 }, { "epoch": 11.936758893280633, "grad_norm": 0.9920166730880737, "learning_rate": 2.01980198019802e-05, "loss": 0.0301, "step": 3020 }, { "epoch": 11.940711462450594, "grad_norm": 1.5376096963882446, "learning_rate": 2.0188118811881192e-05, "loss": 0.0435, "step": 3021 }, { "epoch": 11.944664031620553, "grad_norm": 1.2956080436706543, "learning_rate": 2.017821782178218e-05, "loss": 0.0341, "step": 3022 }, { "epoch": 11.948616600790514, "grad_norm": 1.2682106494903564, "learning_rate": 2.016831683168317e-05, "loss": 0.0325, "step": 3023 }, { "epoch": 11.952569169960475, "grad_norm": 1.1597236394882202, "learning_rate": 2.015841584158416e-05, "loss": 0.0329, "step": 3024 }, { "epoch": 11.956521739130435, "grad_norm": 2.1015470027923584, "learning_rate": 2.014851485148515e-05, "loss": 0.0601, "step": 3025 }, { "epoch": 11.960474308300395, "grad_norm": 1.3985978364944458, "learning_rate": 2.013861386138614e-05, "loss": 0.0405, "step": 3026 }, { "epoch": 11.964426877470355, "grad_norm": 1.1066056489944458, "learning_rate": 2.012871287128713e-05, "loss": 0.0321, "step": 3027 }, { "epoch": 11.968379446640316, "grad_norm": 1.5145957469940186, "learning_rate": 2.0118811881188122e-05, "loss": 0.0484, "step": 3028 }, { "epoch": 11.972332015810277, "grad_norm": 1.2720836400985718, "learning_rate": 2.010891089108911e-05, "loss": 0.0377, "step": 3029 }, { "epoch": 11.976284584980236, "grad_norm": 1.2619870901107788, "learning_rate": 2.00990099009901e-05, "loss": 0.0355, "step": 3030 }, { "epoch": 11.980237154150197, "grad_norm": 1.5248780250549316, "learning_rate": 2.0089108910891092e-05, "loss": 0.0388, "step": 3031 }, { "epoch": 11.984189723320158, "grad_norm": 1.0843119621276855, "learning_rate": 2.007920792079208e-05, "loss": 0.0313, "step": 3032 }, { "epoch": 11.988142292490119, "grad_norm": 1.9179787635803223, "learning_rate": 2.006930693069307e-05, "loss": 0.0633, "step": 3033 }, { "epoch": 11.992094861660078, "grad_norm": 1.757576584815979, "learning_rate": 2.0059405940594062e-05, "loss": 0.041, "step": 3034 }, { "epoch": 11.996047430830039, "grad_norm": 1.2129508256912231, "learning_rate": 2.0049504950495053e-05, "loss": 0.0318, "step": 3035 }, { "epoch": 12.0, "grad_norm": 1.4925364255905151, "learning_rate": 2.003960396039604e-05, "loss": 0.0436, "step": 3036 }, { "epoch": 12.003952569169961, "grad_norm": 1.367187261581421, "learning_rate": 2.0029702970297032e-05, "loss": 0.0358, "step": 3037 }, { "epoch": 12.007905138339922, "grad_norm": 1.198751449584961, "learning_rate": 2.0019801980198023e-05, "loss": 0.0329, "step": 3038 }, { "epoch": 12.011857707509881, "grad_norm": 0.9797889590263367, "learning_rate": 2.000990099009901e-05, "loss": 0.0304, "step": 3039 }, { "epoch": 12.015810276679842, "grad_norm": 1.1576038599014282, "learning_rate": 2e-05, "loss": 0.0416, "step": 3040 }, { "epoch": 12.019762845849803, "grad_norm": 0.9224410057067871, "learning_rate": 1.9990099009900993e-05, "loss": 0.0269, "step": 3041 }, { "epoch": 12.023715415019764, "grad_norm": 1.091935157775879, "learning_rate": 1.9980198019801984e-05, "loss": 0.0349, "step": 3042 }, { "epoch": 12.027667984189723, "grad_norm": 0.6150736212730408, "learning_rate": 1.997029702970297e-05, "loss": 0.022, "step": 3043 }, { "epoch": 12.031620553359684, "grad_norm": 0.8205119371414185, "learning_rate": 1.9960396039603963e-05, "loss": 0.0286, "step": 3044 }, { "epoch": 12.035573122529645, "grad_norm": 1.004612684249878, "learning_rate": 1.9950495049504954e-05, "loss": 0.0336, "step": 3045 }, { "epoch": 12.039525691699605, "grad_norm": 1.1061493158340454, "learning_rate": 1.994059405940594e-05, "loss": 0.0319, "step": 3046 }, { "epoch": 12.043478260869565, "grad_norm": 1.3076575994491577, "learning_rate": 1.9930693069306932e-05, "loss": 0.0407, "step": 3047 }, { "epoch": 12.047430830039525, "grad_norm": 1.0459425449371338, "learning_rate": 1.9920792079207923e-05, "loss": 0.0319, "step": 3048 }, { "epoch": 12.051383399209486, "grad_norm": 1.1192636489868164, "learning_rate": 1.9910891089108914e-05, "loss": 0.0472, "step": 3049 }, { "epoch": 12.055335968379447, "grad_norm": 0.9237827658653259, "learning_rate": 1.9900990099009902e-05, "loss": 0.0519, "step": 3050 }, { "epoch": 12.059288537549406, "grad_norm": 0.9170507192611694, "learning_rate": 1.9891089108910893e-05, "loss": 0.0402, "step": 3051 }, { "epoch": 12.063241106719367, "grad_norm": 1.5172845125198364, "learning_rate": 1.9881188118811884e-05, "loss": 0.0416, "step": 3052 }, { "epoch": 12.067193675889328, "grad_norm": 0.846725583076477, "learning_rate": 1.9871287128712872e-05, "loss": 0.0238, "step": 3053 }, { "epoch": 12.071146245059289, "grad_norm": 1.2148655652999878, "learning_rate": 1.9861386138613863e-05, "loss": 0.036, "step": 3054 }, { "epoch": 12.075098814229248, "grad_norm": 1.0406270027160645, "learning_rate": 1.9851485148514854e-05, "loss": 0.0311, "step": 3055 }, { "epoch": 12.079051383399209, "grad_norm": 1.0052616596221924, "learning_rate": 1.9841584158415845e-05, "loss": 0.0267, "step": 3056 }, { "epoch": 12.08300395256917, "grad_norm": 1.1493887901306152, "learning_rate": 1.9831683168316833e-05, "loss": 0.0354, "step": 3057 }, { "epoch": 12.08695652173913, "grad_norm": 1.3325055837631226, "learning_rate": 1.9821782178217824e-05, "loss": 0.0324, "step": 3058 }, { "epoch": 12.090909090909092, "grad_norm": 1.1400082111358643, "learning_rate": 1.9811881188118815e-05, "loss": 0.0346, "step": 3059 }, { "epoch": 12.09486166007905, "grad_norm": 2.1808862686157227, "learning_rate": 1.9801980198019803e-05, "loss": 0.0498, "step": 3060 }, { "epoch": 12.098814229249012, "grad_norm": 1.3355307579040527, "learning_rate": 1.9792079207920794e-05, "loss": 0.0337, "step": 3061 }, { "epoch": 12.102766798418973, "grad_norm": 0.8770806193351746, "learning_rate": 1.9782178217821785e-05, "loss": 0.0258, "step": 3062 }, { "epoch": 12.106719367588934, "grad_norm": 0.9087365865707397, "learning_rate": 1.9772277227722776e-05, "loss": 0.0247, "step": 3063 }, { "epoch": 12.110671936758893, "grad_norm": 0.9083654880523682, "learning_rate": 1.9762376237623764e-05, "loss": 0.0261, "step": 3064 }, { "epoch": 12.114624505928854, "grad_norm": 1.0996989011764526, "learning_rate": 1.9752475247524755e-05, "loss": 0.0326, "step": 3065 }, { "epoch": 12.118577075098814, "grad_norm": 1.2136564254760742, "learning_rate": 1.9742574257425746e-05, "loss": 0.0418, "step": 3066 }, { "epoch": 12.122529644268775, "grad_norm": 1.078505039215088, "learning_rate": 1.9732673267326733e-05, "loss": 0.0351, "step": 3067 }, { "epoch": 12.126482213438734, "grad_norm": 1.1354198455810547, "learning_rate": 1.9722772277227724e-05, "loss": 0.0303, "step": 3068 }, { "epoch": 12.130434782608695, "grad_norm": 1.2238078117370605, "learning_rate": 1.9712871287128715e-05, "loss": 0.0319, "step": 3069 }, { "epoch": 12.134387351778656, "grad_norm": 1.6967483758926392, "learning_rate": 1.9702970297029703e-05, "loss": 0.0521, "step": 3070 }, { "epoch": 12.138339920948617, "grad_norm": 0.8087192177772522, "learning_rate": 1.9693069306930694e-05, "loss": 0.0231, "step": 3071 }, { "epoch": 12.142292490118576, "grad_norm": 0.7595519423484802, "learning_rate": 1.9683168316831685e-05, "loss": 0.0227, "step": 3072 }, { "epoch": 12.146245059288537, "grad_norm": 0.9552290439605713, "learning_rate": 1.9673267326732676e-05, "loss": 0.0258, "step": 3073 }, { "epoch": 12.150197628458498, "grad_norm": 0.9997588396072388, "learning_rate": 1.9663366336633664e-05, "loss": 0.0484, "step": 3074 }, { "epoch": 12.154150197628459, "grad_norm": 0.9730979204177856, "learning_rate": 1.9653465346534655e-05, "loss": 0.0334, "step": 3075 }, { "epoch": 12.15810276679842, "grad_norm": 1.398061752319336, "learning_rate": 1.9643564356435646e-05, "loss": 0.0395, "step": 3076 }, { "epoch": 12.162055335968379, "grad_norm": 0.9508703351020813, "learning_rate": 1.9633663366336634e-05, "loss": 0.0306, "step": 3077 }, { "epoch": 12.16600790513834, "grad_norm": 0.9174811244010925, "learning_rate": 1.9623762376237625e-05, "loss": 0.0279, "step": 3078 }, { "epoch": 12.1699604743083, "grad_norm": 1.0476396083831787, "learning_rate": 1.9613861386138616e-05, "loss": 0.0299, "step": 3079 }, { "epoch": 12.173913043478262, "grad_norm": 1.3681377172470093, "learning_rate": 1.9603960396039607e-05, "loss": 0.0364, "step": 3080 }, { "epoch": 12.17786561264822, "grad_norm": 1.1918669939041138, "learning_rate": 1.9594059405940595e-05, "loss": 0.0307, "step": 3081 }, { "epoch": 12.181818181818182, "grad_norm": 0.8304014801979065, "learning_rate": 1.9584158415841586e-05, "loss": 0.023, "step": 3082 }, { "epoch": 12.185770750988143, "grad_norm": 1.1701602935791016, "learning_rate": 1.9574257425742577e-05, "loss": 0.0363, "step": 3083 }, { "epoch": 12.189723320158103, "grad_norm": 1.137604832649231, "learning_rate": 1.9564356435643564e-05, "loss": 0.0298, "step": 3084 }, { "epoch": 12.193675889328063, "grad_norm": 1.0889915227890015, "learning_rate": 1.9554455445544556e-05, "loss": 0.0313, "step": 3085 }, { "epoch": 12.197628458498023, "grad_norm": 1.4675688743591309, "learning_rate": 1.9544554455445547e-05, "loss": 0.0359, "step": 3086 }, { "epoch": 12.201581027667984, "grad_norm": 1.019031286239624, "learning_rate": 1.9534653465346538e-05, "loss": 0.0341, "step": 3087 }, { "epoch": 12.205533596837945, "grad_norm": 1.5090476274490356, "learning_rate": 1.9524752475247525e-05, "loss": 0.0407, "step": 3088 }, { "epoch": 12.209486166007904, "grad_norm": 1.330480933189392, "learning_rate": 1.9514851485148516e-05, "loss": 0.039, "step": 3089 }, { "epoch": 12.213438735177865, "grad_norm": 1.0014712810516357, "learning_rate": 1.9504950495049508e-05, "loss": 0.0312, "step": 3090 }, { "epoch": 12.217391304347826, "grad_norm": 1.0672969818115234, "learning_rate": 1.9495049504950495e-05, "loss": 0.034, "step": 3091 }, { "epoch": 12.221343873517787, "grad_norm": 1.2428863048553467, "learning_rate": 1.9485148514851486e-05, "loss": 0.0358, "step": 3092 }, { "epoch": 12.225296442687746, "grad_norm": 1.024052381515503, "learning_rate": 1.9475247524752477e-05, "loss": 0.0291, "step": 3093 }, { "epoch": 12.229249011857707, "grad_norm": 0.9687153697013855, "learning_rate": 1.946534653465347e-05, "loss": 0.0293, "step": 3094 }, { "epoch": 12.233201581027668, "grad_norm": 0.9442889094352722, "learning_rate": 1.9455445544554456e-05, "loss": 0.0298, "step": 3095 }, { "epoch": 12.237154150197629, "grad_norm": 0.8521490693092346, "learning_rate": 1.9445544554455447e-05, "loss": 0.02, "step": 3096 }, { "epoch": 12.24110671936759, "grad_norm": 1.222674012184143, "learning_rate": 1.9435643564356438e-05, "loss": 0.0338, "step": 3097 }, { "epoch": 12.245059288537549, "grad_norm": 1.673208236694336, "learning_rate": 1.9425742574257426e-05, "loss": 0.0403, "step": 3098 }, { "epoch": 12.24901185770751, "grad_norm": 0.9996777772903442, "learning_rate": 1.9415841584158417e-05, "loss": 0.0292, "step": 3099 }, { "epoch": 12.25296442687747, "grad_norm": 1.397169589996338, "learning_rate": 1.9405940594059408e-05, "loss": 0.075, "step": 3100 }, { "epoch": 12.256916996047432, "grad_norm": 0.9996297955513, "learning_rate": 1.93960396039604e-05, "loss": 0.0254, "step": 3101 }, { "epoch": 12.26086956521739, "grad_norm": 1.2501285076141357, "learning_rate": 1.9386138613861387e-05, "loss": 0.0399, "step": 3102 }, { "epoch": 12.264822134387352, "grad_norm": 1.4116286039352417, "learning_rate": 1.9376237623762378e-05, "loss": 0.038, "step": 3103 }, { "epoch": 12.268774703557312, "grad_norm": 1.1823358535766602, "learning_rate": 1.936633663366337e-05, "loss": 0.0312, "step": 3104 }, { "epoch": 12.272727272727273, "grad_norm": 1.4181437492370605, "learning_rate": 1.9356435643564357e-05, "loss": 0.04, "step": 3105 }, { "epoch": 12.276679841897232, "grad_norm": 1.0753183364868164, "learning_rate": 1.9346534653465348e-05, "loss": 0.0316, "step": 3106 }, { "epoch": 12.280632411067193, "grad_norm": 1.4289147853851318, "learning_rate": 1.933663366336634e-05, "loss": 0.0385, "step": 3107 }, { "epoch": 12.284584980237154, "grad_norm": 1.1187119483947754, "learning_rate": 1.932673267326733e-05, "loss": 0.0293, "step": 3108 }, { "epoch": 12.288537549407115, "grad_norm": 0.8385266661643982, "learning_rate": 1.9316831683168317e-05, "loss": 0.0243, "step": 3109 }, { "epoch": 12.292490118577074, "grad_norm": 1.0450007915496826, "learning_rate": 1.930693069306931e-05, "loss": 0.0325, "step": 3110 }, { "epoch": 12.296442687747035, "grad_norm": 0.9134737849235535, "learning_rate": 1.92970297029703e-05, "loss": 0.035, "step": 3111 }, { "epoch": 12.300395256916996, "grad_norm": 0.9111508131027222, "learning_rate": 1.9287128712871287e-05, "loss": 0.0326, "step": 3112 }, { "epoch": 12.304347826086957, "grad_norm": 0.9908366203308105, "learning_rate": 1.9277227722772278e-05, "loss": 0.0268, "step": 3113 }, { "epoch": 12.308300395256918, "grad_norm": 0.8613056540489197, "learning_rate": 1.926732673267327e-05, "loss": 0.0267, "step": 3114 }, { "epoch": 12.312252964426877, "grad_norm": 0.9821538329124451, "learning_rate": 1.925742574257426e-05, "loss": 0.0423, "step": 3115 }, { "epoch": 12.316205533596838, "grad_norm": 0.954716145992279, "learning_rate": 1.9247524752475248e-05, "loss": 0.0326, "step": 3116 }, { "epoch": 12.320158102766799, "grad_norm": 0.9907593727111816, "learning_rate": 1.923762376237624e-05, "loss": 0.0281, "step": 3117 }, { "epoch": 12.32411067193676, "grad_norm": 0.9852539300918579, "learning_rate": 1.922772277227723e-05, "loss": 0.0342, "step": 3118 }, { "epoch": 12.328063241106719, "grad_norm": 1.1602667570114136, "learning_rate": 1.9217821782178218e-05, "loss": 0.0313, "step": 3119 }, { "epoch": 12.33201581027668, "grad_norm": 0.8248794078826904, "learning_rate": 1.920792079207921e-05, "loss": 0.0255, "step": 3120 }, { "epoch": 12.33596837944664, "grad_norm": 1.2565209865570068, "learning_rate": 1.91980198019802e-05, "loss": 0.0333, "step": 3121 }, { "epoch": 12.339920948616601, "grad_norm": 0.7930582165718079, "learning_rate": 1.918811881188119e-05, "loss": 0.0252, "step": 3122 }, { "epoch": 12.34387351778656, "grad_norm": 0.8789475560188293, "learning_rate": 1.917821782178218e-05, "loss": 0.0268, "step": 3123 }, { "epoch": 12.347826086956522, "grad_norm": 0.9201738238334656, "learning_rate": 1.916831683168317e-05, "loss": 0.026, "step": 3124 }, { "epoch": 12.351778656126482, "grad_norm": 0.9256227016448975, "learning_rate": 1.915841584158416e-05, "loss": 0.0301, "step": 3125 }, { "epoch": 12.355731225296443, "grad_norm": 1.4643151760101318, "learning_rate": 1.914851485148515e-05, "loss": 0.0417, "step": 3126 }, { "epoch": 12.359683794466402, "grad_norm": 0.9463131427764893, "learning_rate": 1.913861386138614e-05, "loss": 0.0279, "step": 3127 }, { "epoch": 12.363636363636363, "grad_norm": 1.2232880592346191, "learning_rate": 1.912871287128713e-05, "loss": 0.0306, "step": 3128 }, { "epoch": 12.367588932806324, "grad_norm": 1.120313048362732, "learning_rate": 1.911881188118812e-05, "loss": 0.0285, "step": 3129 }, { "epoch": 12.371541501976285, "grad_norm": 1.0808844566345215, "learning_rate": 1.910891089108911e-05, "loss": 0.0278, "step": 3130 }, { "epoch": 12.375494071146244, "grad_norm": 0.7870800495147705, "learning_rate": 1.90990099009901e-05, "loss": 0.0236, "step": 3131 }, { "epoch": 12.379446640316205, "grad_norm": 1.4770317077636719, "learning_rate": 1.908910891089109e-05, "loss": 0.0555, "step": 3132 }, { "epoch": 12.383399209486166, "grad_norm": 0.7480015158653259, "learning_rate": 1.907920792079208e-05, "loss": 0.0234, "step": 3133 }, { "epoch": 12.387351778656127, "grad_norm": 0.6288163065910339, "learning_rate": 1.906930693069307e-05, "loss": 0.0173, "step": 3134 }, { "epoch": 12.391304347826088, "grad_norm": 0.9102922677993774, "learning_rate": 1.905940594059406e-05, "loss": 0.0265, "step": 3135 }, { "epoch": 12.395256916996047, "grad_norm": 0.9833833575248718, "learning_rate": 1.904950495049505e-05, "loss": 0.0303, "step": 3136 }, { "epoch": 12.399209486166008, "grad_norm": 0.8541263937950134, "learning_rate": 1.903960396039604e-05, "loss": 0.0298, "step": 3137 }, { "epoch": 12.403162055335969, "grad_norm": 0.8470565676689148, "learning_rate": 1.902970297029703e-05, "loss": 0.0238, "step": 3138 }, { "epoch": 12.40711462450593, "grad_norm": 1.011746883392334, "learning_rate": 1.9019801980198022e-05, "loss": 0.0284, "step": 3139 }, { "epoch": 12.411067193675889, "grad_norm": 1.255311369895935, "learning_rate": 1.900990099009901e-05, "loss": 0.0326, "step": 3140 }, { "epoch": 12.41501976284585, "grad_norm": 2.007350206375122, "learning_rate": 1.9e-05, "loss": 0.0542, "step": 3141 }, { "epoch": 12.41897233201581, "grad_norm": 0.9729670882225037, "learning_rate": 1.8990099009900992e-05, "loss": 0.0322, "step": 3142 }, { "epoch": 12.422924901185771, "grad_norm": 0.7548816204071045, "learning_rate": 1.898019801980198e-05, "loss": 0.0428, "step": 3143 }, { "epoch": 12.42687747035573, "grad_norm": 1.048368215560913, "learning_rate": 1.897029702970297e-05, "loss": 0.0333, "step": 3144 }, { "epoch": 12.430830039525691, "grad_norm": 1.068520188331604, "learning_rate": 1.8960396039603962e-05, "loss": 0.0275, "step": 3145 }, { "epoch": 12.434782608695652, "grad_norm": 1.0025371313095093, "learning_rate": 1.8950495049504953e-05, "loss": 0.0286, "step": 3146 }, { "epoch": 12.438735177865613, "grad_norm": 1.170985460281372, "learning_rate": 1.894059405940594e-05, "loss": 0.0325, "step": 3147 }, { "epoch": 12.442687747035572, "grad_norm": 1.300419807434082, "learning_rate": 1.8930693069306932e-05, "loss": 0.0497, "step": 3148 }, { "epoch": 12.446640316205533, "grad_norm": 1.0652148723602295, "learning_rate": 1.8920792079207923e-05, "loss": 0.0279, "step": 3149 }, { "epoch": 12.450592885375494, "grad_norm": 1.2511491775512695, "learning_rate": 1.891089108910891e-05, "loss": 0.0413, "step": 3150 }, { "epoch": 12.454545454545455, "grad_norm": 1.2394198179244995, "learning_rate": 1.89009900990099e-05, "loss": 0.0355, "step": 3151 }, { "epoch": 12.458498023715414, "grad_norm": 1.0210580825805664, "learning_rate": 1.8891089108910893e-05, "loss": 0.0285, "step": 3152 }, { "epoch": 12.462450592885375, "grad_norm": 1.079872965812683, "learning_rate": 1.8881188118811884e-05, "loss": 0.0356, "step": 3153 }, { "epoch": 12.466403162055336, "grad_norm": 1.1941852569580078, "learning_rate": 1.887128712871287e-05, "loss": 0.0344, "step": 3154 }, { "epoch": 12.470355731225297, "grad_norm": 1.4849677085876465, "learning_rate": 1.8861386138613862e-05, "loss": 0.0457, "step": 3155 }, { "epoch": 12.474308300395258, "grad_norm": 1.2483214139938354, "learning_rate": 1.8851485148514853e-05, "loss": 0.0308, "step": 3156 }, { "epoch": 12.478260869565217, "grad_norm": 1.0716394186019897, "learning_rate": 1.884158415841584e-05, "loss": 0.0309, "step": 3157 }, { "epoch": 12.482213438735178, "grad_norm": 0.7751546502113342, "learning_rate": 1.8831683168316832e-05, "loss": 0.0224, "step": 3158 }, { "epoch": 12.486166007905139, "grad_norm": 1.1457443237304688, "learning_rate": 1.8821782178217823e-05, "loss": 0.0343, "step": 3159 }, { "epoch": 12.4901185770751, "grad_norm": 1.0132465362548828, "learning_rate": 1.8811881188118814e-05, "loss": 0.0289, "step": 3160 }, { "epoch": 12.494071146245059, "grad_norm": 0.9197384119033813, "learning_rate": 1.8801980198019802e-05, "loss": 0.0348, "step": 3161 }, { "epoch": 12.49802371541502, "grad_norm": 1.3338367938995361, "learning_rate": 1.8792079207920793e-05, "loss": 0.0378, "step": 3162 }, { "epoch": 12.50197628458498, "grad_norm": 1.1786954402923584, "learning_rate": 1.8782178217821784e-05, "loss": 0.0372, "step": 3163 }, { "epoch": 12.505928853754941, "grad_norm": 0.8761476278305054, "learning_rate": 1.8772277227722772e-05, "loss": 0.0234, "step": 3164 }, { "epoch": 12.5098814229249, "grad_norm": 1.2546488046646118, "learning_rate": 1.8762376237623763e-05, "loss": 0.0351, "step": 3165 }, { "epoch": 12.513833992094861, "grad_norm": 1.0254158973693848, "learning_rate": 1.8752475247524754e-05, "loss": 0.0344, "step": 3166 }, { "epoch": 12.517786561264822, "grad_norm": 1.2233061790466309, "learning_rate": 1.8742574257425745e-05, "loss": 0.0389, "step": 3167 }, { "epoch": 12.521739130434783, "grad_norm": 1.3618972301483154, "learning_rate": 1.8732673267326733e-05, "loss": 0.0409, "step": 3168 }, { "epoch": 12.525691699604742, "grad_norm": 1.352601408958435, "learning_rate": 1.8722772277227724e-05, "loss": 0.0395, "step": 3169 }, { "epoch": 12.529644268774703, "grad_norm": 0.869033932685852, "learning_rate": 1.8712871287128715e-05, "loss": 0.0236, "step": 3170 }, { "epoch": 12.533596837944664, "grad_norm": 1.2126970291137695, "learning_rate": 1.8702970297029703e-05, "loss": 0.0379, "step": 3171 }, { "epoch": 12.537549407114625, "grad_norm": 1.182604193687439, "learning_rate": 1.8693069306930694e-05, "loss": 0.0345, "step": 3172 }, { "epoch": 12.541501976284586, "grad_norm": 1.246134877204895, "learning_rate": 1.8683168316831685e-05, "loss": 0.0422, "step": 3173 }, { "epoch": 12.545454545454545, "grad_norm": 1.1336332559585571, "learning_rate": 1.8673267326732676e-05, "loss": 0.0267, "step": 3174 }, { "epoch": 12.549407114624506, "grad_norm": 1.0685977935791016, "learning_rate": 1.8663366336633663e-05, "loss": 0.0323, "step": 3175 }, { "epoch": 12.553359683794467, "grad_norm": 1.1409662961959839, "learning_rate": 1.8653465346534654e-05, "loss": 0.04, "step": 3176 }, { "epoch": 12.557312252964428, "grad_norm": 0.9637084603309631, "learning_rate": 1.8643564356435646e-05, "loss": 0.0334, "step": 3177 }, { "epoch": 12.561264822134387, "grad_norm": 1.347076654434204, "learning_rate": 1.8633663366336633e-05, "loss": 0.0351, "step": 3178 }, { "epoch": 12.565217391304348, "grad_norm": 1.1089354753494263, "learning_rate": 1.8623762376237624e-05, "loss": 0.0285, "step": 3179 }, { "epoch": 12.569169960474309, "grad_norm": 0.9154654145240784, "learning_rate": 1.8613861386138615e-05, "loss": 0.0305, "step": 3180 }, { "epoch": 12.57312252964427, "grad_norm": 1.44314706325531, "learning_rate": 1.8603960396039603e-05, "loss": 0.052, "step": 3181 }, { "epoch": 12.577075098814229, "grad_norm": 1.2596913576126099, "learning_rate": 1.8594059405940594e-05, "loss": 0.0331, "step": 3182 }, { "epoch": 12.58102766798419, "grad_norm": 1.178729772567749, "learning_rate": 1.8584158415841585e-05, "loss": 0.0375, "step": 3183 }, { "epoch": 12.58498023715415, "grad_norm": 1.2843221426010132, "learning_rate": 1.8574257425742576e-05, "loss": 0.0359, "step": 3184 }, { "epoch": 12.588932806324111, "grad_norm": 1.3128455877304077, "learning_rate": 1.8564356435643564e-05, "loss": 0.0335, "step": 3185 }, { "epoch": 12.59288537549407, "grad_norm": 1.0830817222595215, "learning_rate": 1.8554455445544555e-05, "loss": 0.0325, "step": 3186 }, { "epoch": 12.596837944664031, "grad_norm": 1.0995783805847168, "learning_rate": 1.8544554455445546e-05, "loss": 0.0331, "step": 3187 }, { "epoch": 12.600790513833992, "grad_norm": 1.2764323949813843, "learning_rate": 1.8534653465346534e-05, "loss": 0.0356, "step": 3188 }, { "epoch": 12.604743083003953, "grad_norm": 1.089827060699463, "learning_rate": 1.8524752475247525e-05, "loss": 0.0363, "step": 3189 }, { "epoch": 12.608695652173914, "grad_norm": 1.197160243988037, "learning_rate": 1.8514851485148516e-05, "loss": 0.0342, "step": 3190 }, { "epoch": 12.612648221343873, "grad_norm": 1.2197495698928833, "learning_rate": 1.8504950495049507e-05, "loss": 0.0324, "step": 3191 }, { "epoch": 12.616600790513834, "grad_norm": 1.2989892959594727, "learning_rate": 1.8495049504950495e-05, "loss": 0.0362, "step": 3192 }, { "epoch": 12.620553359683795, "grad_norm": 1.3440642356872559, "learning_rate": 1.8485148514851486e-05, "loss": 0.0469, "step": 3193 }, { "epoch": 12.624505928853754, "grad_norm": 1.1231417655944824, "learning_rate": 1.8475247524752477e-05, "loss": 0.0317, "step": 3194 }, { "epoch": 12.628458498023715, "grad_norm": 1.3787822723388672, "learning_rate": 1.8465346534653464e-05, "loss": 0.0369, "step": 3195 }, { "epoch": 12.632411067193676, "grad_norm": 1.533516764640808, "learning_rate": 1.8455445544554455e-05, "loss": 0.0422, "step": 3196 }, { "epoch": 12.636363636363637, "grad_norm": 0.6983890533447266, "learning_rate": 1.8445544554455447e-05, "loss": 0.0268, "step": 3197 }, { "epoch": 12.640316205533598, "grad_norm": 0.9148122072219849, "learning_rate": 1.8435643564356438e-05, "loss": 0.0274, "step": 3198 }, { "epoch": 12.644268774703557, "grad_norm": 1.7844573259353638, "learning_rate": 1.8425742574257425e-05, "loss": 0.0403, "step": 3199 }, { "epoch": 12.648221343873518, "grad_norm": 0.8514466285705566, "learning_rate": 1.8415841584158416e-05, "loss": 0.024, "step": 3200 }, { "epoch": 12.652173913043478, "grad_norm": 1.4467363357543945, "learning_rate": 1.8405940594059407e-05, "loss": 0.0466, "step": 3201 }, { "epoch": 12.65612648221344, "grad_norm": 1.3368598222732544, "learning_rate": 1.8396039603960395e-05, "loss": 0.0402, "step": 3202 }, { "epoch": 12.660079051383399, "grad_norm": 1.2260581254959106, "learning_rate": 1.8386138613861386e-05, "loss": 0.0349, "step": 3203 }, { "epoch": 12.66403162055336, "grad_norm": 1.5283174514770508, "learning_rate": 1.8376237623762377e-05, "loss": 0.0395, "step": 3204 }, { "epoch": 12.66798418972332, "grad_norm": 1.225465178489685, "learning_rate": 1.8366336633663368e-05, "loss": 0.0321, "step": 3205 }, { "epoch": 12.671936758893281, "grad_norm": 1.233218789100647, "learning_rate": 1.8356435643564356e-05, "loss": 0.0323, "step": 3206 }, { "epoch": 12.67588932806324, "grad_norm": 1.5278937816619873, "learning_rate": 1.8346534653465347e-05, "loss": 0.0406, "step": 3207 }, { "epoch": 12.679841897233201, "grad_norm": 1.2398098707199097, "learning_rate": 1.8336633663366338e-05, "loss": 0.0366, "step": 3208 }, { "epoch": 12.683794466403162, "grad_norm": 1.4367772340774536, "learning_rate": 1.8326732673267326e-05, "loss": 0.037, "step": 3209 }, { "epoch": 12.687747035573123, "grad_norm": 1.2649582624435425, "learning_rate": 1.8316831683168317e-05, "loss": 0.0357, "step": 3210 }, { "epoch": 12.691699604743082, "grad_norm": 1.041571855545044, "learning_rate": 1.8306930693069308e-05, "loss": 0.0282, "step": 3211 }, { "epoch": 12.695652173913043, "grad_norm": 1.249204397201538, "learning_rate": 1.82970297029703e-05, "loss": 0.0367, "step": 3212 }, { "epoch": 12.699604743083004, "grad_norm": 1.0690991878509521, "learning_rate": 1.8287128712871287e-05, "loss": 0.029, "step": 3213 }, { "epoch": 12.703557312252965, "grad_norm": 1.188254952430725, "learning_rate": 1.8277227722772278e-05, "loss": 0.0292, "step": 3214 }, { "epoch": 12.707509881422926, "grad_norm": 1.4866710901260376, "learning_rate": 1.826732673267327e-05, "loss": 0.0388, "step": 3215 }, { "epoch": 12.711462450592885, "grad_norm": 0.7669569253921509, "learning_rate": 1.8257425742574256e-05, "loss": 0.0213, "step": 3216 }, { "epoch": 12.715415019762846, "grad_norm": 1.0527762174606323, "learning_rate": 1.8247524752475248e-05, "loss": 0.0282, "step": 3217 }, { "epoch": 12.719367588932807, "grad_norm": 1.1242635250091553, "learning_rate": 1.823762376237624e-05, "loss": 0.0369, "step": 3218 }, { "epoch": 12.723320158102768, "grad_norm": 1.0151340961456299, "learning_rate": 1.822772277227723e-05, "loss": 0.0302, "step": 3219 }, { "epoch": 12.727272727272727, "grad_norm": 0.9825148582458496, "learning_rate": 1.8217821782178217e-05, "loss": 0.0301, "step": 3220 }, { "epoch": 12.731225296442688, "grad_norm": 1.3746803998947144, "learning_rate": 1.820792079207921e-05, "loss": 0.0393, "step": 3221 }, { "epoch": 12.735177865612648, "grad_norm": 0.9855623841285706, "learning_rate": 1.81980198019802e-05, "loss": 0.03, "step": 3222 }, { "epoch": 12.73913043478261, "grad_norm": 1.2263152599334717, "learning_rate": 1.8188118811881187e-05, "loss": 0.033, "step": 3223 }, { "epoch": 12.743083003952568, "grad_norm": 0.8484699130058289, "learning_rate": 1.8178217821782178e-05, "loss": 0.0246, "step": 3224 }, { "epoch": 12.74703557312253, "grad_norm": 1.0972453355789185, "learning_rate": 1.816831683168317e-05, "loss": 0.0397, "step": 3225 }, { "epoch": 12.75098814229249, "grad_norm": 1.0490835905075073, "learning_rate": 1.815841584158416e-05, "loss": 0.0305, "step": 3226 }, { "epoch": 12.754940711462451, "grad_norm": 1.4880040884017944, "learning_rate": 1.8148514851485148e-05, "loss": 0.0406, "step": 3227 }, { "epoch": 12.75889328063241, "grad_norm": 1.126147985458374, "learning_rate": 1.813861386138614e-05, "loss": 0.0347, "step": 3228 }, { "epoch": 12.762845849802371, "grad_norm": 1.1941932439804077, "learning_rate": 1.812871287128713e-05, "loss": 0.043, "step": 3229 }, { "epoch": 12.766798418972332, "grad_norm": 1.1798456907272339, "learning_rate": 1.8118811881188118e-05, "loss": 0.0355, "step": 3230 }, { "epoch": 12.770750988142293, "grad_norm": 1.3512955904006958, "learning_rate": 1.810891089108911e-05, "loss": 0.036, "step": 3231 }, { "epoch": 12.774703557312254, "grad_norm": 1.0146472454071045, "learning_rate": 1.80990099009901e-05, "loss": 0.0255, "step": 3232 }, { "epoch": 12.778656126482213, "grad_norm": 0.9368916749954224, "learning_rate": 1.808910891089109e-05, "loss": 0.0342, "step": 3233 }, { "epoch": 12.782608695652174, "grad_norm": 1.458371639251709, "learning_rate": 1.807920792079208e-05, "loss": 0.0411, "step": 3234 }, { "epoch": 12.786561264822135, "grad_norm": 1.150603175163269, "learning_rate": 1.806930693069307e-05, "loss": 0.0305, "step": 3235 }, { "epoch": 12.790513833992096, "grad_norm": 1.4688119888305664, "learning_rate": 1.805940594059406e-05, "loss": 0.0463, "step": 3236 }, { "epoch": 12.794466403162055, "grad_norm": 0.9768362641334534, "learning_rate": 1.804950495049505e-05, "loss": 0.0248, "step": 3237 }, { "epoch": 12.798418972332016, "grad_norm": 0.7948504090309143, "learning_rate": 1.803960396039604e-05, "loss": 0.0305, "step": 3238 }, { "epoch": 12.802371541501977, "grad_norm": 1.1944516897201538, "learning_rate": 1.802970297029703e-05, "loss": 0.0332, "step": 3239 }, { "epoch": 12.806324110671937, "grad_norm": 1.4716730117797852, "learning_rate": 1.801980198019802e-05, "loss": 0.0422, "step": 3240 }, { "epoch": 12.810276679841897, "grad_norm": 1.0211023092269897, "learning_rate": 1.800990099009901e-05, "loss": 0.0358, "step": 3241 }, { "epoch": 12.814229249011857, "grad_norm": 1.5208399295806885, "learning_rate": 1.8e-05, "loss": 0.0438, "step": 3242 }, { "epoch": 12.818181818181818, "grad_norm": 1.1413767337799072, "learning_rate": 1.799009900990099e-05, "loss": 0.028, "step": 3243 }, { "epoch": 12.82213438735178, "grad_norm": 1.6396913528442383, "learning_rate": 1.798019801980198e-05, "loss": 0.0474, "step": 3244 }, { "epoch": 12.826086956521738, "grad_norm": 1.6044002771377563, "learning_rate": 1.797029702970297e-05, "loss": 0.0459, "step": 3245 }, { "epoch": 12.8300395256917, "grad_norm": 0.9949911832809448, "learning_rate": 1.796039603960396e-05, "loss": 0.0289, "step": 3246 }, { "epoch": 12.83399209486166, "grad_norm": 1.3878989219665527, "learning_rate": 1.795049504950495e-05, "loss": 0.0404, "step": 3247 }, { "epoch": 12.837944664031621, "grad_norm": 1.9112662076950073, "learning_rate": 1.794059405940594e-05, "loss": 0.0633, "step": 3248 }, { "epoch": 12.841897233201582, "grad_norm": 1.187748908996582, "learning_rate": 1.793069306930693e-05, "loss": 0.0343, "step": 3249 }, { "epoch": 12.845849802371541, "grad_norm": 1.1207407712936401, "learning_rate": 1.7920792079207922e-05, "loss": 0.0382, "step": 3250 }, { "epoch": 12.849802371541502, "grad_norm": 1.2274922132492065, "learning_rate": 1.791089108910891e-05, "loss": 0.0363, "step": 3251 }, { "epoch": 12.853754940711463, "grad_norm": 1.2143886089324951, "learning_rate": 1.79009900990099e-05, "loss": 0.0323, "step": 3252 }, { "epoch": 12.857707509881424, "grad_norm": 1.389915108680725, "learning_rate": 1.7891089108910892e-05, "loss": 0.0362, "step": 3253 }, { "epoch": 12.861660079051383, "grad_norm": 0.9075316786766052, "learning_rate": 1.788118811881188e-05, "loss": 0.0305, "step": 3254 }, { "epoch": 12.865612648221344, "grad_norm": 0.9321634769439697, "learning_rate": 1.787128712871287e-05, "loss": 0.0272, "step": 3255 }, { "epoch": 12.869565217391305, "grad_norm": 0.9096898436546326, "learning_rate": 1.7861386138613862e-05, "loss": 0.0282, "step": 3256 }, { "epoch": 12.873517786561266, "grad_norm": 1.4015119075775146, "learning_rate": 1.7851485148514853e-05, "loss": 0.0387, "step": 3257 }, { "epoch": 12.877470355731225, "grad_norm": 1.5727510452270508, "learning_rate": 1.784158415841584e-05, "loss": 0.0418, "step": 3258 }, { "epoch": 12.881422924901186, "grad_norm": 0.8925953507423401, "learning_rate": 1.783168316831683e-05, "loss": 0.0266, "step": 3259 }, { "epoch": 12.885375494071146, "grad_norm": 0.9841691851615906, "learning_rate": 1.7821782178217823e-05, "loss": 0.0292, "step": 3260 }, { "epoch": 12.889328063241107, "grad_norm": 1.0374585390090942, "learning_rate": 1.781188118811881e-05, "loss": 0.0277, "step": 3261 }, { "epoch": 12.893280632411066, "grad_norm": 0.8873392343521118, "learning_rate": 1.78019801980198e-05, "loss": 0.0269, "step": 3262 }, { "epoch": 12.897233201581027, "grad_norm": 0.9561868906021118, "learning_rate": 1.7792079207920793e-05, "loss": 0.0261, "step": 3263 }, { "epoch": 12.901185770750988, "grad_norm": 1.058793544769287, "learning_rate": 1.7782178217821784e-05, "loss": 0.0328, "step": 3264 }, { "epoch": 12.90513833992095, "grad_norm": 0.9722856879234314, "learning_rate": 1.777227722772277e-05, "loss": 0.0279, "step": 3265 }, { "epoch": 12.909090909090908, "grad_norm": 1.3526966571807861, "learning_rate": 1.7762376237623762e-05, "loss": 0.0398, "step": 3266 }, { "epoch": 12.91304347826087, "grad_norm": 1.4510143995285034, "learning_rate": 1.7752475247524753e-05, "loss": 0.0444, "step": 3267 }, { "epoch": 12.91699604743083, "grad_norm": 1.3624268770217896, "learning_rate": 1.774257425742574e-05, "loss": 0.0391, "step": 3268 }, { "epoch": 12.920948616600791, "grad_norm": 1.03825843334198, "learning_rate": 1.7732673267326732e-05, "loss": 0.0324, "step": 3269 }, { "epoch": 12.92490118577075, "grad_norm": 0.9937554001808167, "learning_rate": 1.7722772277227723e-05, "loss": 0.0262, "step": 3270 }, { "epoch": 12.928853754940711, "grad_norm": 1.011080265045166, "learning_rate": 1.7712871287128714e-05, "loss": 0.0278, "step": 3271 }, { "epoch": 12.932806324110672, "grad_norm": 1.0789525508880615, "learning_rate": 1.7702970297029702e-05, "loss": 0.0304, "step": 3272 }, { "epoch": 12.936758893280633, "grad_norm": 1.368829607963562, "learning_rate": 1.7693069306930693e-05, "loss": 0.0385, "step": 3273 }, { "epoch": 12.940711462450594, "grad_norm": 1.0693432092666626, "learning_rate": 1.7683168316831684e-05, "loss": 0.0466, "step": 3274 }, { "epoch": 12.944664031620553, "grad_norm": 1.1010111570358276, "learning_rate": 1.7673267326732672e-05, "loss": 0.0307, "step": 3275 }, { "epoch": 12.948616600790514, "grad_norm": 0.6307644248008728, "learning_rate": 1.7663366336633663e-05, "loss": 0.0225, "step": 3276 }, { "epoch": 12.952569169960475, "grad_norm": 0.8736321926116943, "learning_rate": 1.7653465346534654e-05, "loss": 0.0256, "step": 3277 }, { "epoch": 12.956521739130435, "grad_norm": 1.3333877325057983, "learning_rate": 1.7643564356435645e-05, "loss": 0.0334, "step": 3278 }, { "epoch": 12.960474308300395, "grad_norm": 2.0538883209228516, "learning_rate": 1.7633663366336633e-05, "loss": 0.0658, "step": 3279 }, { "epoch": 12.964426877470355, "grad_norm": 1.5667617321014404, "learning_rate": 1.7623762376237624e-05, "loss": 0.0521, "step": 3280 }, { "epoch": 12.968379446640316, "grad_norm": 0.9057922959327698, "learning_rate": 1.7613861386138615e-05, "loss": 0.0252, "step": 3281 }, { "epoch": 12.972332015810277, "grad_norm": 1.3153730630874634, "learning_rate": 1.7603960396039602e-05, "loss": 0.034, "step": 3282 }, { "epoch": 12.976284584980236, "grad_norm": 1.3456804752349854, "learning_rate": 1.7594059405940594e-05, "loss": 0.0387, "step": 3283 }, { "epoch": 12.980237154150197, "grad_norm": 1.5599850416183472, "learning_rate": 1.7584158415841585e-05, "loss": 0.0482, "step": 3284 }, { "epoch": 12.984189723320158, "grad_norm": 1.2885959148406982, "learning_rate": 1.7574257425742576e-05, "loss": 0.0476, "step": 3285 }, { "epoch": 12.988142292490119, "grad_norm": 0.9898382425308228, "learning_rate": 1.7564356435643563e-05, "loss": 0.0314, "step": 3286 }, { "epoch": 12.992094861660078, "grad_norm": 1.09585440158844, "learning_rate": 1.7554455445544554e-05, "loss": 0.0287, "step": 3287 }, { "epoch": 12.996047430830039, "grad_norm": 0.9731926918029785, "learning_rate": 1.7544554455445545e-05, "loss": 0.0284, "step": 3288 }, { "epoch": 13.0, "grad_norm": 1.7188111543655396, "learning_rate": 1.7534653465346533e-05, "loss": 0.0452, "step": 3289 }, { "epoch": 13.003952569169961, "grad_norm": 0.6017376780509949, "learning_rate": 1.7524752475247524e-05, "loss": 0.0281, "step": 3290 }, { "epoch": 13.007905138339922, "grad_norm": 0.7886996269226074, "learning_rate": 1.7514851485148515e-05, "loss": 0.0247, "step": 3291 }, { "epoch": 13.011857707509881, "grad_norm": 1.058236002922058, "learning_rate": 1.7504950495049506e-05, "loss": 0.03, "step": 3292 }, { "epoch": 13.015810276679842, "grad_norm": 1.03701651096344, "learning_rate": 1.7495049504950494e-05, "loss": 0.0348, "step": 3293 }, { "epoch": 13.019762845849803, "grad_norm": 0.558673620223999, "learning_rate": 1.7485148514851485e-05, "loss": 0.0237, "step": 3294 }, { "epoch": 13.023715415019764, "grad_norm": 1.0949592590332031, "learning_rate": 1.7475247524752476e-05, "loss": 0.0339, "step": 3295 }, { "epoch": 13.027667984189723, "grad_norm": 0.6546443104743958, "learning_rate": 1.7465346534653464e-05, "loss": 0.0179, "step": 3296 }, { "epoch": 13.031620553359684, "grad_norm": 0.5907596349716187, "learning_rate": 1.7455445544554455e-05, "loss": 0.0214, "step": 3297 }, { "epoch": 13.035573122529645, "grad_norm": 0.8142299652099609, "learning_rate": 1.7445544554455446e-05, "loss": 0.0239, "step": 3298 }, { "epoch": 13.039525691699605, "grad_norm": 1.2065690755844116, "learning_rate": 1.7435643564356434e-05, "loss": 0.0315, "step": 3299 }, { "epoch": 13.043478260869565, "grad_norm": 0.9317627549171448, "learning_rate": 1.7425742574257425e-05, "loss": 0.026, "step": 3300 }, { "epoch": 13.047430830039525, "grad_norm": 1.1303623914718628, "learning_rate": 1.7415841584158416e-05, "loss": 0.0527, "step": 3301 }, { "epoch": 13.051383399209486, "grad_norm": 0.8671423196792603, "learning_rate": 1.7405940594059407e-05, "loss": 0.0306, "step": 3302 }, { "epoch": 13.055335968379447, "grad_norm": 0.8869814872741699, "learning_rate": 1.7396039603960394e-05, "loss": 0.0247, "step": 3303 }, { "epoch": 13.059288537549406, "grad_norm": 0.5696347951889038, "learning_rate": 1.7386138613861386e-05, "loss": 0.0153, "step": 3304 }, { "epoch": 13.063241106719367, "grad_norm": 0.8638686537742615, "learning_rate": 1.7376237623762377e-05, "loss": 0.0327, "step": 3305 }, { "epoch": 13.067193675889328, "grad_norm": 1.0841442346572876, "learning_rate": 1.7366336633663364e-05, "loss": 0.0323, "step": 3306 }, { "epoch": 13.071146245059289, "grad_norm": 0.5291824340820312, "learning_rate": 1.7356435643564355e-05, "loss": 0.0202, "step": 3307 }, { "epoch": 13.075098814229248, "grad_norm": 0.7705602645874023, "learning_rate": 1.7346534653465346e-05, "loss": 0.0239, "step": 3308 }, { "epoch": 13.079051383399209, "grad_norm": 0.7069066166877747, "learning_rate": 1.7336633663366338e-05, "loss": 0.0289, "step": 3309 }, { "epoch": 13.08300395256917, "grad_norm": 0.9034048914909363, "learning_rate": 1.7326732673267325e-05, "loss": 0.0346, "step": 3310 }, { "epoch": 13.08695652173913, "grad_norm": 0.86359041929245, "learning_rate": 1.7316831683168316e-05, "loss": 0.023, "step": 3311 }, { "epoch": 13.090909090909092, "grad_norm": 0.7813597321510315, "learning_rate": 1.7306930693069307e-05, "loss": 0.0285, "step": 3312 }, { "epoch": 13.09486166007905, "grad_norm": 0.8234113454818726, "learning_rate": 1.7297029702970295e-05, "loss": 0.0309, "step": 3313 }, { "epoch": 13.098814229249012, "grad_norm": 1.3863413333892822, "learning_rate": 1.7287128712871286e-05, "loss": 0.0377, "step": 3314 }, { "epoch": 13.102766798418973, "grad_norm": 0.7283810377120972, "learning_rate": 1.7277227722772277e-05, "loss": 0.0193, "step": 3315 }, { "epoch": 13.106719367588934, "grad_norm": 0.6732761263847351, "learning_rate": 1.7267326732673268e-05, "loss": 0.0175, "step": 3316 }, { "epoch": 13.110671936758893, "grad_norm": 0.7456334233283997, "learning_rate": 1.7257425742574256e-05, "loss": 0.0255, "step": 3317 }, { "epoch": 13.114624505928854, "grad_norm": 1.1658165454864502, "learning_rate": 1.7247524752475247e-05, "loss": 0.0327, "step": 3318 }, { "epoch": 13.118577075098814, "grad_norm": 1.3141635656356812, "learning_rate": 1.7237623762376238e-05, "loss": 0.0342, "step": 3319 }, { "epoch": 13.122529644268775, "grad_norm": 0.90336674451828, "learning_rate": 1.7227722772277226e-05, "loss": 0.0277, "step": 3320 }, { "epoch": 13.126482213438734, "grad_norm": 0.9471737146377563, "learning_rate": 1.7217821782178217e-05, "loss": 0.0242, "step": 3321 }, { "epoch": 13.130434782608695, "grad_norm": 1.2973151206970215, "learning_rate": 1.7207920792079208e-05, "loss": 0.0307, "step": 3322 }, { "epoch": 13.134387351778656, "grad_norm": 0.9391686916351318, "learning_rate": 1.71980198019802e-05, "loss": 0.0279, "step": 3323 }, { "epoch": 13.138339920948617, "grad_norm": 0.8350376486778259, "learning_rate": 1.7188118811881187e-05, "loss": 0.0268, "step": 3324 }, { "epoch": 13.142292490118576, "grad_norm": 0.9920552968978882, "learning_rate": 1.717821782178218e-05, "loss": 0.029, "step": 3325 }, { "epoch": 13.146245059288537, "grad_norm": 1.314151406288147, "learning_rate": 1.7168316831683172e-05, "loss": 0.0499, "step": 3326 }, { "epoch": 13.150197628458498, "grad_norm": 0.8533769249916077, "learning_rate": 1.715841584158416e-05, "loss": 0.0269, "step": 3327 }, { "epoch": 13.154150197628459, "grad_norm": 0.5367113351821899, "learning_rate": 1.714851485148515e-05, "loss": 0.0203, "step": 3328 }, { "epoch": 13.15810276679842, "grad_norm": 0.8376951217651367, "learning_rate": 1.7138613861386142e-05, "loss": 0.0236, "step": 3329 }, { "epoch": 13.162055335968379, "grad_norm": 0.8221657872200012, "learning_rate": 1.712871287128713e-05, "loss": 0.0289, "step": 3330 }, { "epoch": 13.16600790513834, "grad_norm": 0.883965253829956, "learning_rate": 1.711881188118812e-05, "loss": 0.0303, "step": 3331 }, { "epoch": 13.1699604743083, "grad_norm": 0.8150313496589661, "learning_rate": 1.710891089108911e-05, "loss": 0.0245, "step": 3332 }, { "epoch": 13.173913043478262, "grad_norm": 0.4126499593257904, "learning_rate": 1.7099009900990103e-05, "loss": 0.0217, "step": 3333 }, { "epoch": 13.17786561264822, "grad_norm": 0.563342273235321, "learning_rate": 1.708910891089109e-05, "loss": 0.0369, "step": 3334 }, { "epoch": 13.181818181818182, "grad_norm": 0.6740759611129761, "learning_rate": 1.707920792079208e-05, "loss": 0.0321, "step": 3335 }, { "epoch": 13.185770750988143, "grad_norm": 1.150596261024475, "learning_rate": 1.7069306930693073e-05, "loss": 0.0271, "step": 3336 }, { "epoch": 13.189723320158103, "grad_norm": 0.7164453268051147, "learning_rate": 1.705940594059406e-05, "loss": 0.0229, "step": 3337 }, { "epoch": 13.193675889328063, "grad_norm": 0.8708245754241943, "learning_rate": 1.704950495049505e-05, "loss": 0.0244, "step": 3338 }, { "epoch": 13.197628458498023, "grad_norm": 1.1507948637008667, "learning_rate": 1.7039603960396042e-05, "loss": 0.0379, "step": 3339 }, { "epoch": 13.201581027667984, "grad_norm": 1.1500343084335327, "learning_rate": 1.7029702970297033e-05, "loss": 0.0322, "step": 3340 }, { "epoch": 13.205533596837945, "grad_norm": 0.922844648361206, "learning_rate": 1.701980198019802e-05, "loss": 0.0247, "step": 3341 }, { "epoch": 13.209486166007904, "grad_norm": 1.0602214336395264, "learning_rate": 1.7009900990099012e-05, "loss": 0.0339, "step": 3342 }, { "epoch": 13.213438735177865, "grad_norm": 0.9777653813362122, "learning_rate": 1.7000000000000003e-05, "loss": 0.025, "step": 3343 }, { "epoch": 13.217391304347826, "grad_norm": 0.8936442136764526, "learning_rate": 1.699009900990099e-05, "loss": 0.0231, "step": 3344 }, { "epoch": 13.221343873517787, "grad_norm": 0.7998532652854919, "learning_rate": 1.6980198019801982e-05, "loss": 0.0173, "step": 3345 }, { "epoch": 13.225296442687746, "grad_norm": 0.9508734941482544, "learning_rate": 1.6970297029702973e-05, "loss": 0.0287, "step": 3346 }, { "epoch": 13.229249011857707, "grad_norm": 1.102386474609375, "learning_rate": 1.696039603960396e-05, "loss": 0.0255, "step": 3347 }, { "epoch": 13.233201581027668, "grad_norm": 1.1774332523345947, "learning_rate": 1.6950495049504952e-05, "loss": 0.0413, "step": 3348 }, { "epoch": 13.237154150197629, "grad_norm": 1.1394891738891602, "learning_rate": 1.6940594059405943e-05, "loss": 0.0313, "step": 3349 }, { "epoch": 13.24110671936759, "grad_norm": 0.6230199933052063, "learning_rate": 1.6930693069306934e-05, "loss": 0.0238, "step": 3350 }, { "epoch": 13.245059288537549, "grad_norm": 0.6841564774513245, "learning_rate": 1.692079207920792e-05, "loss": 0.0195, "step": 3351 }, { "epoch": 13.24901185770751, "grad_norm": 1.156973958015442, "learning_rate": 1.6910891089108913e-05, "loss": 0.0335, "step": 3352 }, { "epoch": 13.25296442687747, "grad_norm": 0.9996264576911926, "learning_rate": 1.6900990099009904e-05, "loss": 0.0259, "step": 3353 }, { "epoch": 13.256916996047432, "grad_norm": 1.546714425086975, "learning_rate": 1.689108910891089e-05, "loss": 0.0367, "step": 3354 }, { "epoch": 13.26086956521739, "grad_norm": 0.9186763167381287, "learning_rate": 1.6881188118811882e-05, "loss": 0.0347, "step": 3355 }, { "epoch": 13.264822134387352, "grad_norm": 2.050931930541992, "learning_rate": 1.6871287128712874e-05, "loss": 0.0458, "step": 3356 }, { "epoch": 13.268774703557312, "grad_norm": 1.0791443586349487, "learning_rate": 1.6861386138613865e-05, "loss": 0.0311, "step": 3357 }, { "epoch": 13.272727272727273, "grad_norm": 0.6483448147773743, "learning_rate": 1.6851485148514852e-05, "loss": 0.0238, "step": 3358 }, { "epoch": 13.276679841897232, "grad_norm": 0.6741937398910522, "learning_rate": 1.6841584158415843e-05, "loss": 0.0262, "step": 3359 }, { "epoch": 13.280632411067193, "grad_norm": 0.6524081826210022, "learning_rate": 1.6831683168316834e-05, "loss": 0.0206, "step": 3360 }, { "epoch": 13.284584980237154, "grad_norm": 0.9790909886360168, "learning_rate": 1.6821782178217822e-05, "loss": 0.0265, "step": 3361 }, { "epoch": 13.288537549407115, "grad_norm": 1.1808851957321167, "learning_rate": 1.6811881188118813e-05, "loss": 0.0597, "step": 3362 }, { "epoch": 13.292490118577074, "grad_norm": 0.9442753791809082, "learning_rate": 1.6801980198019804e-05, "loss": 0.0337, "step": 3363 }, { "epoch": 13.296442687747035, "grad_norm": 0.9247639775276184, "learning_rate": 1.6792079207920795e-05, "loss": 0.0238, "step": 3364 }, { "epoch": 13.300395256916996, "grad_norm": 0.9665504097938538, "learning_rate": 1.6782178217821783e-05, "loss": 0.0325, "step": 3365 }, { "epoch": 13.304347826086957, "grad_norm": 0.7801231741905212, "learning_rate": 1.6772277227722774e-05, "loss": 0.0271, "step": 3366 }, { "epoch": 13.308300395256918, "grad_norm": 0.7750723958015442, "learning_rate": 1.6762376237623765e-05, "loss": 0.0208, "step": 3367 }, { "epoch": 13.312252964426877, "grad_norm": 0.7564480304718018, "learning_rate": 1.6752475247524753e-05, "loss": 0.0232, "step": 3368 }, { "epoch": 13.316205533596838, "grad_norm": 0.8579455018043518, "learning_rate": 1.6742574257425744e-05, "loss": 0.0296, "step": 3369 }, { "epoch": 13.320158102766799, "grad_norm": 0.6938242316246033, "learning_rate": 1.6732673267326735e-05, "loss": 0.0253, "step": 3370 }, { "epoch": 13.32411067193676, "grad_norm": 0.9398701190948486, "learning_rate": 1.6722772277227726e-05, "loss": 0.0312, "step": 3371 }, { "epoch": 13.328063241106719, "grad_norm": 0.6591378450393677, "learning_rate": 1.6712871287128714e-05, "loss": 0.0183, "step": 3372 }, { "epoch": 13.33201581027668, "grad_norm": 1.7240321636199951, "learning_rate": 1.6702970297029705e-05, "loss": 0.035, "step": 3373 }, { "epoch": 13.33596837944664, "grad_norm": 0.9695301055908203, "learning_rate": 1.6693069306930696e-05, "loss": 0.0306, "step": 3374 }, { "epoch": 13.339920948616601, "grad_norm": 0.8749107122421265, "learning_rate": 1.6683168316831683e-05, "loss": 0.0258, "step": 3375 }, { "epoch": 13.34387351778656, "grad_norm": 1.0924855470657349, "learning_rate": 1.6673267326732675e-05, "loss": 0.0358, "step": 3376 }, { "epoch": 13.347826086956522, "grad_norm": 1.0155786275863647, "learning_rate": 1.6663366336633666e-05, "loss": 0.0292, "step": 3377 }, { "epoch": 13.351778656126482, "grad_norm": 0.8918850421905518, "learning_rate": 1.6653465346534657e-05, "loss": 0.0227, "step": 3378 }, { "epoch": 13.355731225296443, "grad_norm": 0.8268848061561584, "learning_rate": 1.6643564356435644e-05, "loss": 0.0237, "step": 3379 }, { "epoch": 13.359683794466402, "grad_norm": 0.8527054786682129, "learning_rate": 1.6633663366336635e-05, "loss": 0.023, "step": 3380 }, { "epoch": 13.363636363636363, "grad_norm": 0.9102256894111633, "learning_rate": 1.6623762376237626e-05, "loss": 0.0261, "step": 3381 }, { "epoch": 13.367588932806324, "grad_norm": 0.5867739319801331, "learning_rate": 1.6613861386138614e-05, "loss": 0.0238, "step": 3382 }, { "epoch": 13.371541501976285, "grad_norm": 2.0297601222991943, "learning_rate": 1.6603960396039605e-05, "loss": 0.0557, "step": 3383 }, { "epoch": 13.375494071146244, "grad_norm": 0.9494080543518066, "learning_rate": 1.6594059405940596e-05, "loss": 0.0279, "step": 3384 }, { "epoch": 13.379446640316205, "grad_norm": 2.316311836242676, "learning_rate": 1.6584158415841587e-05, "loss": 0.0667, "step": 3385 }, { "epoch": 13.383399209486166, "grad_norm": 0.7578169107437134, "learning_rate": 1.6574257425742575e-05, "loss": 0.0338, "step": 3386 }, { "epoch": 13.387351778656127, "grad_norm": 0.804232120513916, "learning_rate": 1.6564356435643566e-05, "loss": 0.0206, "step": 3387 }, { "epoch": 13.391304347826088, "grad_norm": 0.6963724493980408, "learning_rate": 1.6554455445544557e-05, "loss": 0.0208, "step": 3388 }, { "epoch": 13.395256916996047, "grad_norm": 0.9487035870552063, "learning_rate": 1.6544554455445545e-05, "loss": 0.0306, "step": 3389 }, { "epoch": 13.399209486166008, "grad_norm": 0.7726027369499207, "learning_rate": 1.6534653465346536e-05, "loss": 0.0323, "step": 3390 }, { "epoch": 13.403162055335969, "grad_norm": 0.7084749937057495, "learning_rate": 1.6524752475247527e-05, "loss": 0.0229, "step": 3391 }, { "epoch": 13.40711462450593, "grad_norm": 0.9561412334442139, "learning_rate": 1.6514851485148518e-05, "loss": 0.0271, "step": 3392 }, { "epoch": 13.411067193675889, "grad_norm": 1.0688868761062622, "learning_rate": 1.6504950495049506e-05, "loss": 0.0321, "step": 3393 }, { "epoch": 13.41501976284585, "grad_norm": 1.1098177433013916, "learning_rate": 1.6495049504950497e-05, "loss": 0.0336, "step": 3394 }, { "epoch": 13.41897233201581, "grad_norm": 0.7769515514373779, "learning_rate": 1.6485148514851488e-05, "loss": 0.0238, "step": 3395 }, { "epoch": 13.422924901185771, "grad_norm": 1.0765407085418701, "learning_rate": 1.6475247524752476e-05, "loss": 0.0313, "step": 3396 }, { "epoch": 13.42687747035573, "grad_norm": 0.9676995873451233, "learning_rate": 1.6465346534653467e-05, "loss": 0.0276, "step": 3397 }, { "epoch": 13.430830039525691, "grad_norm": 1.0737888813018799, "learning_rate": 1.6455445544554458e-05, "loss": 0.0332, "step": 3398 }, { "epoch": 13.434782608695652, "grad_norm": 0.7677600383758545, "learning_rate": 1.644554455445545e-05, "loss": 0.0278, "step": 3399 }, { "epoch": 13.438735177865613, "grad_norm": 0.9660235047340393, "learning_rate": 1.6435643564356436e-05, "loss": 0.0287, "step": 3400 }, { "epoch": 13.442687747035572, "grad_norm": 1.049368143081665, "learning_rate": 1.6425742574257427e-05, "loss": 0.0309, "step": 3401 }, { "epoch": 13.446640316205533, "grad_norm": 1.3674516677856445, "learning_rate": 1.641584158415842e-05, "loss": 0.042, "step": 3402 }, { "epoch": 13.450592885375494, "grad_norm": 1.0853265523910522, "learning_rate": 1.6405940594059406e-05, "loss": 0.0341, "step": 3403 }, { "epoch": 13.454545454545455, "grad_norm": 0.9523180723190308, "learning_rate": 1.6396039603960397e-05, "loss": 0.0307, "step": 3404 }, { "epoch": 13.458498023715414, "grad_norm": 1.0110280513763428, "learning_rate": 1.638613861386139e-05, "loss": 0.0291, "step": 3405 }, { "epoch": 13.462450592885375, "grad_norm": 1.7495434284210205, "learning_rate": 1.6376237623762376e-05, "loss": 0.0221, "step": 3406 }, { "epoch": 13.466403162055336, "grad_norm": 0.7190485596656799, "learning_rate": 1.6366336633663367e-05, "loss": 0.0238, "step": 3407 }, { "epoch": 13.470355731225297, "grad_norm": 0.8545430898666382, "learning_rate": 1.6356435643564358e-05, "loss": 0.0243, "step": 3408 }, { "epoch": 13.474308300395258, "grad_norm": 1.3282558917999268, "learning_rate": 1.634653465346535e-05, "loss": 0.0405, "step": 3409 }, { "epoch": 13.478260869565217, "grad_norm": 0.75214022397995, "learning_rate": 1.6336633663366337e-05, "loss": 0.0546, "step": 3410 }, { "epoch": 13.482213438735178, "grad_norm": 1.3999494314193726, "learning_rate": 1.6326732673267328e-05, "loss": 0.0388, "step": 3411 }, { "epoch": 13.486166007905139, "grad_norm": 0.864436686038971, "learning_rate": 1.631683168316832e-05, "loss": 0.0225, "step": 3412 }, { "epoch": 13.4901185770751, "grad_norm": 1.0790170431137085, "learning_rate": 1.6306930693069307e-05, "loss": 0.0328, "step": 3413 }, { "epoch": 13.494071146245059, "grad_norm": 0.7730204463005066, "learning_rate": 1.6297029702970298e-05, "loss": 0.0266, "step": 3414 }, { "epoch": 13.49802371541502, "grad_norm": 0.6589069962501526, "learning_rate": 1.628712871287129e-05, "loss": 0.0282, "step": 3415 }, { "epoch": 13.50197628458498, "grad_norm": 1.233088493347168, "learning_rate": 1.627722772277228e-05, "loss": 0.0355, "step": 3416 }, { "epoch": 13.505928853754941, "grad_norm": 0.8000485897064209, "learning_rate": 1.6267326732673268e-05, "loss": 0.0254, "step": 3417 }, { "epoch": 13.5098814229249, "grad_norm": 0.9592580199241638, "learning_rate": 1.625742574257426e-05, "loss": 0.0284, "step": 3418 }, { "epoch": 13.513833992094861, "grad_norm": 0.6181468963623047, "learning_rate": 1.624752475247525e-05, "loss": 0.0221, "step": 3419 }, { "epoch": 13.517786561264822, "grad_norm": 1.5289815664291382, "learning_rate": 1.6237623762376237e-05, "loss": 0.0485, "step": 3420 }, { "epoch": 13.521739130434783, "grad_norm": 1.0914250612258911, "learning_rate": 1.622772277227723e-05, "loss": 0.0375, "step": 3421 }, { "epoch": 13.525691699604742, "grad_norm": 0.8138375282287598, "learning_rate": 1.621782178217822e-05, "loss": 0.0257, "step": 3422 }, { "epoch": 13.529644268774703, "grad_norm": 0.7758058309555054, "learning_rate": 1.620792079207921e-05, "loss": 0.0291, "step": 3423 }, { "epoch": 13.533596837944664, "grad_norm": 0.926360547542572, "learning_rate": 1.6198019801980198e-05, "loss": 0.0233, "step": 3424 }, { "epoch": 13.537549407114625, "grad_norm": 1.3465940952301025, "learning_rate": 1.618811881188119e-05, "loss": 0.0331, "step": 3425 }, { "epoch": 13.541501976284586, "grad_norm": 0.7158646583557129, "learning_rate": 1.617821782178218e-05, "loss": 0.0348, "step": 3426 }, { "epoch": 13.545454545454545, "grad_norm": 1.0007764101028442, "learning_rate": 1.6168316831683168e-05, "loss": 0.031, "step": 3427 }, { "epoch": 13.549407114624506, "grad_norm": 1.0191679000854492, "learning_rate": 1.615841584158416e-05, "loss": 0.0344, "step": 3428 }, { "epoch": 13.553359683794467, "grad_norm": 0.9322157502174377, "learning_rate": 1.614851485148515e-05, "loss": 0.0271, "step": 3429 }, { "epoch": 13.557312252964428, "grad_norm": 1.346288800239563, "learning_rate": 1.613861386138614e-05, "loss": 0.0458, "step": 3430 }, { "epoch": 13.561264822134387, "grad_norm": 0.8077167868614197, "learning_rate": 1.612871287128713e-05, "loss": 0.0211, "step": 3431 }, { "epoch": 13.565217391304348, "grad_norm": 0.6137113571166992, "learning_rate": 1.611881188118812e-05, "loss": 0.0176, "step": 3432 }, { "epoch": 13.569169960474309, "grad_norm": 1.5701440572738647, "learning_rate": 1.610891089108911e-05, "loss": 0.0463, "step": 3433 }, { "epoch": 13.57312252964427, "grad_norm": 1.5394203662872314, "learning_rate": 1.60990099009901e-05, "loss": 0.0454, "step": 3434 }, { "epoch": 13.577075098814229, "grad_norm": 0.9082497358322144, "learning_rate": 1.608910891089109e-05, "loss": 0.0281, "step": 3435 }, { "epoch": 13.58102766798419, "grad_norm": 1.2318389415740967, "learning_rate": 1.607920792079208e-05, "loss": 0.0387, "step": 3436 }, { "epoch": 13.58498023715415, "grad_norm": 1.019819736480713, "learning_rate": 1.6069306930693072e-05, "loss": 0.027, "step": 3437 }, { "epoch": 13.588932806324111, "grad_norm": 0.9080525040626526, "learning_rate": 1.605940594059406e-05, "loss": 0.0277, "step": 3438 }, { "epoch": 13.59288537549407, "grad_norm": 0.9879224896430969, "learning_rate": 1.604950495049505e-05, "loss": 0.0362, "step": 3439 }, { "epoch": 13.596837944664031, "grad_norm": 0.788608968257904, "learning_rate": 1.6039603960396042e-05, "loss": 0.0257, "step": 3440 }, { "epoch": 13.600790513833992, "grad_norm": 0.821716845035553, "learning_rate": 1.602970297029703e-05, "loss": 0.0227, "step": 3441 }, { "epoch": 13.604743083003953, "grad_norm": 0.9425798654556274, "learning_rate": 1.601980198019802e-05, "loss": 0.0393, "step": 3442 }, { "epoch": 13.608695652173914, "grad_norm": 1.0194640159606934, "learning_rate": 1.600990099009901e-05, "loss": 0.0285, "step": 3443 }, { "epoch": 13.612648221343873, "grad_norm": 1.021409034729004, "learning_rate": 1.6000000000000003e-05, "loss": 0.0295, "step": 3444 }, { "epoch": 13.616600790513834, "grad_norm": 0.9105862379074097, "learning_rate": 1.599009900990099e-05, "loss": 0.0262, "step": 3445 }, { "epoch": 13.620553359683795, "grad_norm": 0.8552016019821167, "learning_rate": 1.598019801980198e-05, "loss": 0.0245, "step": 3446 }, { "epoch": 13.624505928853754, "grad_norm": 0.9359368085861206, "learning_rate": 1.5970297029702972e-05, "loss": 0.0276, "step": 3447 }, { "epoch": 13.628458498023715, "grad_norm": 1.013120412826538, "learning_rate": 1.596039603960396e-05, "loss": 0.0423, "step": 3448 }, { "epoch": 13.632411067193676, "grad_norm": 1.2945442199707031, "learning_rate": 1.595049504950495e-05, "loss": 0.0338, "step": 3449 }, { "epoch": 13.636363636363637, "grad_norm": 0.7762601375579834, "learning_rate": 1.5940594059405942e-05, "loss": 0.021, "step": 3450 }, { "epoch": 13.640316205533598, "grad_norm": 0.6575326323509216, "learning_rate": 1.5930693069306933e-05, "loss": 0.0228, "step": 3451 }, { "epoch": 13.644268774703557, "grad_norm": 1.0427922010421753, "learning_rate": 1.592079207920792e-05, "loss": 0.0366, "step": 3452 }, { "epoch": 13.648221343873518, "grad_norm": 1.513797640800476, "learning_rate": 1.5910891089108912e-05, "loss": 0.0615, "step": 3453 }, { "epoch": 13.652173913043478, "grad_norm": 1.5355252027511597, "learning_rate": 1.5900990099009903e-05, "loss": 0.0541, "step": 3454 }, { "epoch": 13.65612648221344, "grad_norm": 0.9464790225028992, "learning_rate": 1.589108910891089e-05, "loss": 0.0303, "step": 3455 }, { "epoch": 13.660079051383399, "grad_norm": 0.7372637987136841, "learning_rate": 1.5881188118811882e-05, "loss": 0.0233, "step": 3456 }, { "epoch": 13.66403162055336, "grad_norm": 0.575627863407135, "learning_rate": 1.5871287128712873e-05, "loss": 0.0208, "step": 3457 }, { "epoch": 13.66798418972332, "grad_norm": 0.9740623235702515, "learning_rate": 1.5861386138613864e-05, "loss": 0.0266, "step": 3458 }, { "epoch": 13.671936758893281, "grad_norm": 1.0887947082519531, "learning_rate": 1.5851485148514852e-05, "loss": 0.0375, "step": 3459 }, { "epoch": 13.67588932806324, "grad_norm": 1.359779715538025, "learning_rate": 1.5841584158415843e-05, "loss": 0.0346, "step": 3460 }, { "epoch": 13.679841897233201, "grad_norm": 1.1395184993743896, "learning_rate": 1.5831683168316834e-05, "loss": 0.0334, "step": 3461 }, { "epoch": 13.683794466403162, "grad_norm": 1.2110371589660645, "learning_rate": 1.582178217821782e-05, "loss": 0.0287, "step": 3462 }, { "epoch": 13.687747035573123, "grad_norm": 0.7835529446601868, "learning_rate": 1.5811881188118813e-05, "loss": 0.0254, "step": 3463 }, { "epoch": 13.691699604743082, "grad_norm": 0.5994678139686584, "learning_rate": 1.5801980198019804e-05, "loss": 0.0193, "step": 3464 }, { "epoch": 13.695652173913043, "grad_norm": 1.1314510107040405, "learning_rate": 1.579207920792079e-05, "loss": 0.0328, "step": 3465 }, { "epoch": 13.699604743083004, "grad_norm": 0.8516640067100525, "learning_rate": 1.5782178217821782e-05, "loss": 0.0233, "step": 3466 }, { "epoch": 13.703557312252965, "grad_norm": 0.7649357914924622, "learning_rate": 1.5772277227722773e-05, "loss": 0.0201, "step": 3467 }, { "epoch": 13.707509881422926, "grad_norm": 1.1704363822937012, "learning_rate": 1.5762376237623765e-05, "loss": 0.0381, "step": 3468 }, { "epoch": 13.711462450592885, "grad_norm": 0.7635601162910461, "learning_rate": 1.5752475247524752e-05, "loss": 0.0261, "step": 3469 }, { "epoch": 13.715415019762846, "grad_norm": 0.8921993970870972, "learning_rate": 1.5742574257425743e-05, "loss": 0.024, "step": 3470 }, { "epoch": 13.719367588932807, "grad_norm": 0.9765449166297913, "learning_rate": 1.5732673267326734e-05, "loss": 0.0277, "step": 3471 }, { "epoch": 13.723320158102768, "grad_norm": 0.8655273914337158, "learning_rate": 1.5722772277227722e-05, "loss": 0.0264, "step": 3472 }, { "epoch": 13.727272727272727, "grad_norm": 0.7765814065933228, "learning_rate": 1.5712871287128713e-05, "loss": 0.028, "step": 3473 }, { "epoch": 13.731225296442688, "grad_norm": 1.0973745584487915, "learning_rate": 1.5702970297029704e-05, "loss": 0.0351, "step": 3474 }, { "epoch": 13.735177865612648, "grad_norm": 0.7957023978233337, "learning_rate": 1.5693069306930695e-05, "loss": 0.0289, "step": 3475 }, { "epoch": 13.73913043478261, "grad_norm": 0.9714717864990234, "learning_rate": 1.5683168316831683e-05, "loss": 0.0297, "step": 3476 }, { "epoch": 13.743083003952568, "grad_norm": 0.9512803554534912, "learning_rate": 1.5673267326732674e-05, "loss": 0.0251, "step": 3477 }, { "epoch": 13.74703557312253, "grad_norm": 0.8976863026618958, "learning_rate": 1.5663366336633665e-05, "loss": 0.0269, "step": 3478 }, { "epoch": 13.75098814229249, "grad_norm": 1.0431920289993286, "learning_rate": 1.5653465346534653e-05, "loss": 0.038, "step": 3479 }, { "epoch": 13.754940711462451, "grad_norm": 1.2209020853042603, "learning_rate": 1.5643564356435644e-05, "loss": 0.0323, "step": 3480 }, { "epoch": 13.75889328063241, "grad_norm": 0.4930824339389801, "learning_rate": 1.5633663366336635e-05, "loss": 0.0175, "step": 3481 }, { "epoch": 13.762845849802371, "grad_norm": 1.1874768733978271, "learning_rate": 1.5623762376237626e-05, "loss": 0.0307, "step": 3482 }, { "epoch": 13.766798418972332, "grad_norm": 0.7669538855552673, "learning_rate": 1.5613861386138614e-05, "loss": 0.0205, "step": 3483 }, { "epoch": 13.770750988142293, "grad_norm": 0.5216184258460999, "learning_rate": 1.5603960396039605e-05, "loss": 0.0211, "step": 3484 }, { "epoch": 13.774703557312254, "grad_norm": 1.0948692560195923, "learning_rate": 1.5594059405940596e-05, "loss": 0.031, "step": 3485 }, { "epoch": 13.778656126482213, "grad_norm": 1.000885009765625, "learning_rate": 1.5584158415841583e-05, "loss": 0.032, "step": 3486 }, { "epoch": 13.782608695652174, "grad_norm": 0.8108469843864441, "learning_rate": 1.5574257425742574e-05, "loss": 0.0259, "step": 3487 }, { "epoch": 13.786561264822135, "grad_norm": 0.956916868686676, "learning_rate": 1.5564356435643566e-05, "loss": 0.029, "step": 3488 }, { "epoch": 13.790513833992096, "grad_norm": 0.9780763387680054, "learning_rate": 1.5554455445544557e-05, "loss": 0.0328, "step": 3489 }, { "epoch": 13.794466403162055, "grad_norm": 0.7476334571838379, "learning_rate": 1.5544554455445544e-05, "loss": 0.0256, "step": 3490 }, { "epoch": 13.798418972332016, "grad_norm": 0.7273722887039185, "learning_rate": 1.5534653465346535e-05, "loss": 0.0229, "step": 3491 }, { "epoch": 13.802371541501977, "grad_norm": 0.7341341376304626, "learning_rate": 1.5524752475247526e-05, "loss": 0.0363, "step": 3492 }, { "epoch": 13.806324110671937, "grad_norm": 1.125093698501587, "learning_rate": 1.5514851485148514e-05, "loss": 0.0376, "step": 3493 }, { "epoch": 13.810276679841897, "grad_norm": 0.735327959060669, "learning_rate": 1.5504950495049505e-05, "loss": 0.0281, "step": 3494 }, { "epoch": 13.814229249011857, "grad_norm": 1.193907618522644, "learning_rate": 1.5495049504950496e-05, "loss": 0.0377, "step": 3495 }, { "epoch": 13.818181818181818, "grad_norm": 0.5031324028968811, "learning_rate": 1.5485148514851487e-05, "loss": 0.0457, "step": 3496 }, { "epoch": 13.82213438735178, "grad_norm": 0.5973156690597534, "learning_rate": 1.5475247524752475e-05, "loss": 0.0201, "step": 3497 }, { "epoch": 13.826086956521738, "grad_norm": 0.7858068346977234, "learning_rate": 1.5465346534653466e-05, "loss": 0.0242, "step": 3498 }, { "epoch": 13.8300395256917, "grad_norm": 1.275333046913147, "learning_rate": 1.5455445544554457e-05, "loss": 0.0336, "step": 3499 }, { "epoch": 13.83399209486166, "grad_norm": 0.8698694705963135, "learning_rate": 1.5445544554455445e-05, "loss": 0.0223, "step": 3500 }, { "epoch": 13.837944664031621, "grad_norm": 1.6362009048461914, "learning_rate": 1.5435643564356436e-05, "loss": 0.0463, "step": 3501 }, { "epoch": 13.841897233201582, "grad_norm": 0.7010768055915833, "learning_rate": 1.5425742574257427e-05, "loss": 0.0257, "step": 3502 }, { "epoch": 13.845849802371541, "grad_norm": 1.0001814365386963, "learning_rate": 1.5415841584158418e-05, "loss": 0.0355, "step": 3503 }, { "epoch": 13.849802371541502, "grad_norm": 0.7338948249816895, "learning_rate": 1.5405940594059406e-05, "loss": 0.024, "step": 3504 }, { "epoch": 13.853754940711463, "grad_norm": 0.9132365584373474, "learning_rate": 1.5396039603960397e-05, "loss": 0.0433, "step": 3505 }, { "epoch": 13.857707509881424, "grad_norm": 0.7577654719352722, "learning_rate": 1.5386138613861388e-05, "loss": 0.0269, "step": 3506 }, { "epoch": 13.861660079051383, "grad_norm": 1.2270138263702393, "learning_rate": 1.5376237623762375e-05, "loss": 0.0534, "step": 3507 }, { "epoch": 13.865612648221344, "grad_norm": 0.8529956340789795, "learning_rate": 1.5366336633663367e-05, "loss": 0.0218, "step": 3508 }, { "epoch": 13.869565217391305, "grad_norm": 0.9383001327514648, "learning_rate": 1.5356435643564358e-05, "loss": 0.0316, "step": 3509 }, { "epoch": 13.873517786561266, "grad_norm": 0.9470325112342834, "learning_rate": 1.534653465346535e-05, "loss": 0.0318, "step": 3510 }, { "epoch": 13.877470355731225, "grad_norm": 1.3298810720443726, "learning_rate": 1.5336633663366336e-05, "loss": 0.0367, "step": 3511 }, { "epoch": 13.881422924901186, "grad_norm": 1.2301063537597656, "learning_rate": 1.5326732673267327e-05, "loss": 0.0417, "step": 3512 }, { "epoch": 13.885375494071146, "grad_norm": 0.9180741310119629, "learning_rate": 1.531683168316832e-05, "loss": 0.0234, "step": 3513 }, { "epoch": 13.889328063241107, "grad_norm": 0.782676637172699, "learning_rate": 1.5306930693069306e-05, "loss": 0.0251, "step": 3514 }, { "epoch": 13.893280632411066, "grad_norm": 0.8229760527610779, "learning_rate": 1.5297029702970297e-05, "loss": 0.0237, "step": 3515 }, { "epoch": 13.897233201581027, "grad_norm": 0.9918065667152405, "learning_rate": 1.5287128712871288e-05, "loss": 0.0335, "step": 3516 }, { "epoch": 13.901185770750988, "grad_norm": 0.7432458400726318, "learning_rate": 1.527722772277228e-05, "loss": 0.0331, "step": 3517 }, { "epoch": 13.90513833992095, "grad_norm": 0.9340378046035767, "learning_rate": 1.5267326732673267e-05, "loss": 0.0234, "step": 3518 }, { "epoch": 13.909090909090908, "grad_norm": 1.0651695728302002, "learning_rate": 1.5257425742574258e-05, "loss": 0.034, "step": 3519 }, { "epoch": 13.91304347826087, "grad_norm": 0.8750892281532288, "learning_rate": 1.5247524752475247e-05, "loss": 0.0278, "step": 3520 }, { "epoch": 13.91699604743083, "grad_norm": 0.6412554383277893, "learning_rate": 1.5237623762376239e-05, "loss": 0.0227, "step": 3521 }, { "epoch": 13.920948616600791, "grad_norm": 0.8236206769943237, "learning_rate": 1.5227722772277228e-05, "loss": 0.0252, "step": 3522 }, { "epoch": 13.92490118577075, "grad_norm": 0.8373007774353027, "learning_rate": 1.5217821782178219e-05, "loss": 0.0258, "step": 3523 }, { "epoch": 13.928853754940711, "grad_norm": 0.7383465766906738, "learning_rate": 1.5207920792079208e-05, "loss": 0.0263, "step": 3524 }, { "epoch": 13.932806324110672, "grad_norm": 1.276159405708313, "learning_rate": 1.5198019801980198e-05, "loss": 0.0369, "step": 3525 }, { "epoch": 13.936758893280633, "grad_norm": 1.2225431203842163, "learning_rate": 1.5188118811881189e-05, "loss": 0.0389, "step": 3526 }, { "epoch": 13.940711462450594, "grad_norm": 0.686886191368103, "learning_rate": 1.5178217821782178e-05, "loss": 0.0214, "step": 3527 }, { "epoch": 13.944664031620553, "grad_norm": 0.8234241008758545, "learning_rate": 1.516831683168317e-05, "loss": 0.0266, "step": 3528 }, { "epoch": 13.948616600790514, "grad_norm": 1.0629805326461792, "learning_rate": 1.5158415841584159e-05, "loss": 0.0345, "step": 3529 }, { "epoch": 13.952569169960475, "grad_norm": 1.1654633283615112, "learning_rate": 1.514851485148515e-05, "loss": 0.0462, "step": 3530 }, { "epoch": 13.956521739130435, "grad_norm": 0.9510064721107483, "learning_rate": 1.5138613861386139e-05, "loss": 0.0344, "step": 3531 }, { "epoch": 13.960474308300395, "grad_norm": 0.8088047504425049, "learning_rate": 1.5128712871287128e-05, "loss": 0.0228, "step": 3532 }, { "epoch": 13.964426877470355, "grad_norm": 0.9595481157302856, "learning_rate": 1.511881188118812e-05, "loss": 0.0308, "step": 3533 }, { "epoch": 13.968379446640316, "grad_norm": 1.439243197441101, "learning_rate": 1.5108910891089109e-05, "loss": 0.0332, "step": 3534 }, { "epoch": 13.972332015810277, "grad_norm": 1.2921990156173706, "learning_rate": 1.50990099009901e-05, "loss": 0.0407, "step": 3535 }, { "epoch": 13.976284584980236, "grad_norm": 1.1854232549667358, "learning_rate": 1.508910891089109e-05, "loss": 0.0353, "step": 3536 }, { "epoch": 13.980237154150197, "grad_norm": 0.7282184362411499, "learning_rate": 1.5079207920792079e-05, "loss": 0.0268, "step": 3537 }, { "epoch": 13.984189723320158, "grad_norm": 0.8035431504249573, "learning_rate": 1.506930693069307e-05, "loss": 0.0225, "step": 3538 }, { "epoch": 13.988142292490119, "grad_norm": 0.9193235635757446, "learning_rate": 1.5059405940594059e-05, "loss": 0.0269, "step": 3539 }, { "epoch": 13.992094861660078, "grad_norm": 0.5749298930168152, "learning_rate": 1.504950495049505e-05, "loss": 0.024, "step": 3540 }, { "epoch": 13.996047430830039, "grad_norm": 0.7868842482566833, "learning_rate": 1.503960396039604e-05, "loss": 0.0241, "step": 3541 }, { "epoch": 14.0, "grad_norm": 0.8459658026695251, "learning_rate": 1.502970297029703e-05, "loss": 0.0273, "step": 3542 }, { "epoch": 14.003952569169961, "grad_norm": 0.6285360455513, "learning_rate": 1.501980198019802e-05, "loss": 0.03, "step": 3543 }, { "epoch": 14.007905138339922, "grad_norm": 0.6612513065338135, "learning_rate": 1.500990099009901e-05, "loss": 0.0175, "step": 3544 }, { "epoch": 14.011857707509881, "grad_norm": 0.3971037268638611, "learning_rate": 1.5e-05, "loss": 0.0146, "step": 3545 }, { "epoch": 14.015810276679842, "grad_norm": 0.6367694735527039, "learning_rate": 1.499009900990099e-05, "loss": 0.023, "step": 3546 }, { "epoch": 14.019762845849803, "grad_norm": 0.747485339641571, "learning_rate": 1.498019801980198e-05, "loss": 0.0306, "step": 3547 }, { "epoch": 14.023715415019764, "grad_norm": 1.2573230266571045, "learning_rate": 1.497029702970297e-05, "loss": 0.0355, "step": 3548 }, { "epoch": 14.027667984189723, "grad_norm": 0.8924236297607422, "learning_rate": 1.4960396039603961e-05, "loss": 0.0242, "step": 3549 }, { "epoch": 14.031620553359684, "grad_norm": 0.6937573552131653, "learning_rate": 1.495049504950495e-05, "loss": 0.0214, "step": 3550 }, { "epoch": 14.035573122529645, "grad_norm": 0.3840979337692261, "learning_rate": 1.494059405940594e-05, "loss": 0.0153, "step": 3551 }, { "epoch": 14.039525691699605, "grad_norm": 0.6251746416091919, "learning_rate": 1.4930693069306931e-05, "loss": 0.028, "step": 3552 }, { "epoch": 14.043478260869565, "grad_norm": 0.2696545124053955, "learning_rate": 1.492079207920792e-05, "loss": 0.0146, "step": 3553 }, { "epoch": 14.047430830039525, "grad_norm": 0.6441566348075867, "learning_rate": 1.4910891089108912e-05, "loss": 0.022, "step": 3554 }, { "epoch": 14.051383399209486, "grad_norm": 0.5114539861679077, "learning_rate": 1.4900990099009901e-05, "loss": 0.0333, "step": 3555 }, { "epoch": 14.055335968379447, "grad_norm": 0.7823534607887268, "learning_rate": 1.4891089108910892e-05, "loss": 0.025, "step": 3556 }, { "epoch": 14.059288537549406, "grad_norm": 0.7560704946517944, "learning_rate": 1.4881188118811881e-05, "loss": 0.039, "step": 3557 }, { "epoch": 14.063241106719367, "grad_norm": 0.625924289226532, "learning_rate": 1.487128712871287e-05, "loss": 0.0202, "step": 3558 }, { "epoch": 14.067193675889328, "grad_norm": 1.0054829120635986, "learning_rate": 1.4861386138613862e-05, "loss": 0.0399, "step": 3559 }, { "epoch": 14.071146245059289, "grad_norm": 1.0049611330032349, "learning_rate": 1.4851485148514851e-05, "loss": 0.0338, "step": 3560 }, { "epoch": 14.075098814229248, "grad_norm": 0.6965822577476501, "learning_rate": 1.4841584158415842e-05, "loss": 0.0266, "step": 3561 }, { "epoch": 14.079051383399209, "grad_norm": 0.8681654930114746, "learning_rate": 1.4831683168316832e-05, "loss": 0.0471, "step": 3562 }, { "epoch": 14.08300395256917, "grad_norm": 0.4145543873310089, "learning_rate": 1.4821782178217821e-05, "loss": 0.03, "step": 3563 }, { "epoch": 14.08695652173913, "grad_norm": 0.9870001673698425, "learning_rate": 1.4811881188118812e-05, "loss": 0.0314, "step": 3564 }, { "epoch": 14.090909090909092, "grad_norm": 1.3092273473739624, "learning_rate": 1.4801980198019801e-05, "loss": 0.0299, "step": 3565 }, { "epoch": 14.09486166007905, "grad_norm": 0.4592971205711365, "learning_rate": 1.4792079207920792e-05, "loss": 0.0221, "step": 3566 }, { "epoch": 14.098814229249012, "grad_norm": 0.5552915334701538, "learning_rate": 1.4782178217821782e-05, "loss": 0.0264, "step": 3567 }, { "epoch": 14.102766798418973, "grad_norm": 0.8712630271911621, "learning_rate": 1.4772277227722773e-05, "loss": 0.0301, "step": 3568 }, { "epoch": 14.106719367588934, "grad_norm": 0.677077054977417, "learning_rate": 1.4762376237623762e-05, "loss": 0.0333, "step": 3569 }, { "epoch": 14.110671936758893, "grad_norm": 0.6550650596618652, "learning_rate": 1.4752475247524752e-05, "loss": 0.0214, "step": 3570 }, { "epoch": 14.114624505928854, "grad_norm": 0.5263770222663879, "learning_rate": 1.4742574257425743e-05, "loss": 0.0221, "step": 3571 }, { "epoch": 14.118577075098814, "grad_norm": 1.1121615171432495, "learning_rate": 1.4732673267326732e-05, "loss": 0.0293, "step": 3572 }, { "epoch": 14.122529644268775, "grad_norm": 1.035329818725586, "learning_rate": 1.4722772277227723e-05, "loss": 0.0288, "step": 3573 }, { "epoch": 14.126482213438734, "grad_norm": 0.5690826177597046, "learning_rate": 1.4712871287128712e-05, "loss": 0.0205, "step": 3574 }, { "epoch": 14.130434782608695, "grad_norm": 0.467568039894104, "learning_rate": 1.4702970297029704e-05, "loss": 0.0203, "step": 3575 }, { "epoch": 14.134387351778656, "grad_norm": 0.7593974471092224, "learning_rate": 1.4693069306930693e-05, "loss": 0.0266, "step": 3576 }, { "epoch": 14.138339920948617, "grad_norm": 0.5035028457641602, "learning_rate": 1.4683168316831682e-05, "loss": 0.0239, "step": 3577 }, { "epoch": 14.142292490118576, "grad_norm": 0.6196392178535461, "learning_rate": 1.4673267326732673e-05, "loss": 0.022, "step": 3578 }, { "epoch": 14.146245059288537, "grad_norm": 0.6958089470863342, "learning_rate": 1.4663366336633663e-05, "loss": 0.03, "step": 3579 }, { "epoch": 14.150197628458498, "grad_norm": 0.6183583736419678, "learning_rate": 1.4653465346534654e-05, "loss": 0.0183, "step": 3580 }, { "epoch": 14.154150197628459, "grad_norm": 0.9504097700119019, "learning_rate": 1.4643564356435643e-05, "loss": 0.0303, "step": 3581 }, { "epoch": 14.15810276679842, "grad_norm": 0.571793794631958, "learning_rate": 1.4633663366336634e-05, "loss": 0.0225, "step": 3582 }, { "epoch": 14.162055335968379, "grad_norm": 0.2546845078468323, "learning_rate": 1.4623762376237624e-05, "loss": 0.0307, "step": 3583 }, { "epoch": 14.16600790513834, "grad_norm": 0.5501284599304199, "learning_rate": 1.4613861386138613e-05, "loss": 0.0171, "step": 3584 }, { "epoch": 14.1699604743083, "grad_norm": 0.7117765545845032, "learning_rate": 1.4603960396039604e-05, "loss": 0.0233, "step": 3585 }, { "epoch": 14.173913043478262, "grad_norm": 0.7339401245117188, "learning_rate": 1.4594059405940593e-05, "loss": 0.0198, "step": 3586 }, { "epoch": 14.17786561264822, "grad_norm": 0.9419562220573425, "learning_rate": 1.4584158415841584e-05, "loss": 0.0406, "step": 3587 }, { "epoch": 14.181818181818182, "grad_norm": 0.6493412256240845, "learning_rate": 1.4574257425742574e-05, "loss": 0.0304, "step": 3588 }, { "epoch": 14.185770750988143, "grad_norm": 0.8191783428192139, "learning_rate": 1.4564356435643565e-05, "loss": 0.0247, "step": 3589 }, { "epoch": 14.189723320158103, "grad_norm": 0.6133555173873901, "learning_rate": 1.4554455445544554e-05, "loss": 0.0217, "step": 3590 }, { "epoch": 14.193675889328063, "grad_norm": 0.6625313758850098, "learning_rate": 1.4544554455445544e-05, "loss": 0.0303, "step": 3591 }, { "epoch": 14.197628458498023, "grad_norm": 0.650668203830719, "learning_rate": 1.4534653465346535e-05, "loss": 0.0241, "step": 3592 }, { "epoch": 14.201581027667984, "grad_norm": 0.8373653888702393, "learning_rate": 1.4524752475247524e-05, "loss": 0.0251, "step": 3593 }, { "epoch": 14.205533596837945, "grad_norm": 0.7018810510635376, "learning_rate": 1.4514851485148515e-05, "loss": 0.0257, "step": 3594 }, { "epoch": 14.209486166007904, "grad_norm": 1.112056851387024, "learning_rate": 1.4504950495049505e-05, "loss": 0.0298, "step": 3595 }, { "epoch": 14.213438735177865, "grad_norm": 0.33961766958236694, "learning_rate": 1.4495049504950494e-05, "loss": 0.0161, "step": 3596 }, { "epoch": 14.217391304347826, "grad_norm": 0.5958961248397827, "learning_rate": 1.4485148514851485e-05, "loss": 0.0339, "step": 3597 }, { "epoch": 14.221343873517787, "grad_norm": 0.5742468237876892, "learning_rate": 1.4475247524752474e-05, "loss": 0.0303, "step": 3598 }, { "epoch": 14.225296442687746, "grad_norm": 0.754533588886261, "learning_rate": 1.4465346534653465e-05, "loss": 0.0217, "step": 3599 }, { "epoch": 14.229249011857707, "grad_norm": 0.942776620388031, "learning_rate": 1.4455445544554455e-05, "loss": 0.0337, "step": 3600 }, { "epoch": 14.233201581027668, "grad_norm": 0.5864335894584656, "learning_rate": 1.4445544554455446e-05, "loss": 0.0329, "step": 3601 }, { "epoch": 14.237154150197629, "grad_norm": 0.5520839691162109, "learning_rate": 1.4435643564356435e-05, "loss": 0.024, "step": 3602 }, { "epoch": 14.24110671936759, "grad_norm": 0.9977213740348816, "learning_rate": 1.4425742574257425e-05, "loss": 0.0531, "step": 3603 }, { "epoch": 14.245059288537549, "grad_norm": 0.4953950047492981, "learning_rate": 1.4415841584158416e-05, "loss": 0.024, "step": 3604 }, { "epoch": 14.24901185770751, "grad_norm": 0.6448057293891907, "learning_rate": 1.4405940594059405e-05, "loss": 0.0263, "step": 3605 }, { "epoch": 14.25296442687747, "grad_norm": 0.46163463592529297, "learning_rate": 1.4396039603960396e-05, "loss": 0.0165, "step": 3606 }, { "epoch": 14.256916996047432, "grad_norm": 0.8082854151725769, "learning_rate": 1.4386138613861385e-05, "loss": 0.0235, "step": 3607 }, { "epoch": 14.26086956521739, "grad_norm": 0.6631015539169312, "learning_rate": 1.4376237623762377e-05, "loss": 0.0258, "step": 3608 }, { "epoch": 14.264822134387352, "grad_norm": 0.9251523613929749, "learning_rate": 1.4366336633663366e-05, "loss": 0.0323, "step": 3609 }, { "epoch": 14.268774703557312, "grad_norm": 0.8947650194168091, "learning_rate": 1.4356435643564355e-05, "loss": 0.0242, "step": 3610 }, { "epoch": 14.272727272727273, "grad_norm": 1.105173945426941, "learning_rate": 1.4346534653465346e-05, "loss": 0.0428, "step": 3611 }, { "epoch": 14.276679841897232, "grad_norm": 1.1088634729385376, "learning_rate": 1.4336633663366336e-05, "loss": 0.0313, "step": 3612 }, { "epoch": 14.280632411067193, "grad_norm": 0.8025822639465332, "learning_rate": 1.4326732673267327e-05, "loss": 0.0309, "step": 3613 }, { "epoch": 14.284584980237154, "grad_norm": 1.0850934982299805, "learning_rate": 1.4316831683168316e-05, "loss": 0.0344, "step": 3614 }, { "epoch": 14.288537549407115, "grad_norm": 0.6165494918823242, "learning_rate": 1.4306930693069307e-05, "loss": 0.0272, "step": 3615 }, { "epoch": 14.292490118577074, "grad_norm": 0.7101708054542542, "learning_rate": 1.4297029702970297e-05, "loss": 0.0267, "step": 3616 }, { "epoch": 14.296442687747035, "grad_norm": 0.7435272932052612, "learning_rate": 1.4287128712871286e-05, "loss": 0.0233, "step": 3617 }, { "epoch": 14.300395256916996, "grad_norm": 0.7173030376434326, "learning_rate": 1.4277227722772277e-05, "loss": 0.0351, "step": 3618 }, { "epoch": 14.304347826086957, "grad_norm": 0.9166218638420105, "learning_rate": 1.4267326732673266e-05, "loss": 0.0329, "step": 3619 }, { "epoch": 14.308300395256918, "grad_norm": 0.4430968165397644, "learning_rate": 1.4257425742574257e-05, "loss": 0.0304, "step": 3620 }, { "epoch": 14.312252964426877, "grad_norm": 0.5669541954994202, "learning_rate": 1.4247524752475247e-05, "loss": 0.0199, "step": 3621 }, { "epoch": 14.316205533596838, "grad_norm": 0.4081045389175415, "learning_rate": 1.4237623762376236e-05, "loss": 0.0211, "step": 3622 }, { "epoch": 14.320158102766799, "grad_norm": 0.4572242796421051, "learning_rate": 1.4227722772277227e-05, "loss": 0.0246, "step": 3623 }, { "epoch": 14.32411067193676, "grad_norm": 0.7696126699447632, "learning_rate": 1.4217821782178217e-05, "loss": 0.029, "step": 3624 }, { "epoch": 14.328063241106719, "grad_norm": 0.7123060822486877, "learning_rate": 1.4207920792079208e-05, "loss": 0.024, "step": 3625 }, { "epoch": 14.33201581027668, "grad_norm": 0.8479959964752197, "learning_rate": 1.4198019801980197e-05, "loss": 0.0288, "step": 3626 }, { "epoch": 14.33596837944664, "grad_norm": 0.5251991152763367, "learning_rate": 1.4188118811881188e-05, "loss": 0.0341, "step": 3627 }, { "epoch": 14.339920948616601, "grad_norm": 0.6902558207511902, "learning_rate": 1.4178217821782178e-05, "loss": 0.0328, "step": 3628 }, { "epoch": 14.34387351778656, "grad_norm": 0.6847053170204163, "learning_rate": 1.4168316831683167e-05, "loss": 0.0331, "step": 3629 }, { "epoch": 14.347826086956522, "grad_norm": 0.927829921245575, "learning_rate": 1.4158415841584158e-05, "loss": 0.0269, "step": 3630 }, { "epoch": 14.351778656126482, "grad_norm": 0.3792479336261749, "learning_rate": 1.4148514851485147e-05, "loss": 0.0285, "step": 3631 }, { "epoch": 14.355731225296443, "grad_norm": 0.9317724704742432, "learning_rate": 1.4138613861386138e-05, "loss": 0.0261, "step": 3632 }, { "epoch": 14.359683794466402, "grad_norm": 0.7387916445732117, "learning_rate": 1.4128712871287128e-05, "loss": 0.0254, "step": 3633 }, { "epoch": 14.363636363636363, "grad_norm": 0.8279317021369934, "learning_rate": 1.4118811881188119e-05, "loss": 0.0241, "step": 3634 }, { "epoch": 14.367588932806324, "grad_norm": 0.7013137936592102, "learning_rate": 1.4108910891089108e-05, "loss": 0.0402, "step": 3635 }, { "epoch": 14.371541501976285, "grad_norm": 0.6222538948059082, "learning_rate": 1.4099009900990098e-05, "loss": 0.0218, "step": 3636 }, { "epoch": 14.375494071146244, "grad_norm": 0.5429844260215759, "learning_rate": 1.4089108910891089e-05, "loss": 0.0204, "step": 3637 }, { "epoch": 14.379446640316205, "grad_norm": 0.5700187683105469, "learning_rate": 1.4079207920792078e-05, "loss": 0.022, "step": 3638 }, { "epoch": 14.383399209486166, "grad_norm": 0.9332280158996582, "learning_rate": 1.4069306930693069e-05, "loss": 0.0321, "step": 3639 }, { "epoch": 14.387351778656127, "grad_norm": 0.4397108256816864, "learning_rate": 1.4059405940594062e-05, "loss": 0.0169, "step": 3640 }, { "epoch": 14.391304347826088, "grad_norm": 0.40902724862098694, "learning_rate": 1.4049504950495051e-05, "loss": 0.0197, "step": 3641 }, { "epoch": 14.395256916996047, "grad_norm": 0.6526964902877808, "learning_rate": 1.4039603960396042e-05, "loss": 0.0229, "step": 3642 }, { "epoch": 14.399209486166008, "grad_norm": 0.6803340315818787, "learning_rate": 1.4029702970297032e-05, "loss": 0.0259, "step": 3643 }, { "epoch": 14.403162055335969, "grad_norm": 0.6051563024520874, "learning_rate": 1.4019801980198021e-05, "loss": 0.0194, "step": 3644 }, { "epoch": 14.40711462450593, "grad_norm": 0.8212450742721558, "learning_rate": 1.4009900990099012e-05, "loss": 0.0286, "step": 3645 }, { "epoch": 14.411067193675889, "grad_norm": 0.3461795747280121, "learning_rate": 1.4000000000000001e-05, "loss": 0.0157, "step": 3646 }, { "epoch": 14.41501976284585, "grad_norm": 0.5190447568893433, "learning_rate": 1.3990099009900993e-05, "loss": 0.0211, "step": 3647 }, { "epoch": 14.41897233201581, "grad_norm": 0.8233205676078796, "learning_rate": 1.3980198019801982e-05, "loss": 0.0269, "step": 3648 }, { "epoch": 14.422924901185771, "grad_norm": 0.6635833382606506, "learning_rate": 1.3970297029702973e-05, "loss": 0.019, "step": 3649 }, { "epoch": 14.42687747035573, "grad_norm": 0.9048696160316467, "learning_rate": 1.3960396039603962e-05, "loss": 0.0334, "step": 3650 }, { "epoch": 14.430830039525691, "grad_norm": 1.0150798559188843, "learning_rate": 1.3950495049504952e-05, "loss": 0.026, "step": 3651 }, { "epoch": 14.434782608695652, "grad_norm": 0.28314271569252014, "learning_rate": 1.3940594059405943e-05, "loss": 0.0184, "step": 3652 }, { "epoch": 14.438735177865613, "grad_norm": 1.2986301183700562, "learning_rate": 1.3930693069306932e-05, "loss": 0.0504, "step": 3653 }, { "epoch": 14.442687747035572, "grad_norm": 0.635288655757904, "learning_rate": 1.3920792079207923e-05, "loss": 0.0249, "step": 3654 }, { "epoch": 14.446640316205533, "grad_norm": 0.7654809355735779, "learning_rate": 1.3910891089108913e-05, "loss": 0.0234, "step": 3655 }, { "epoch": 14.450592885375494, "grad_norm": 0.8368544578552246, "learning_rate": 1.3900990099009904e-05, "loss": 0.0312, "step": 3656 }, { "epoch": 14.454545454545455, "grad_norm": 0.812411904335022, "learning_rate": 1.3891089108910893e-05, "loss": 0.0254, "step": 3657 }, { "epoch": 14.458498023715414, "grad_norm": 0.7393926978111267, "learning_rate": 1.3881188118811882e-05, "loss": 0.0182, "step": 3658 }, { "epoch": 14.462450592885375, "grad_norm": 0.805824875831604, "learning_rate": 1.3871287128712873e-05, "loss": 0.0247, "step": 3659 }, { "epoch": 14.466403162055336, "grad_norm": 0.6288550496101379, "learning_rate": 1.3861386138613863e-05, "loss": 0.0276, "step": 3660 }, { "epoch": 14.470355731225297, "grad_norm": 0.7367550730705261, "learning_rate": 1.3851485148514854e-05, "loss": 0.0213, "step": 3661 }, { "epoch": 14.474308300395258, "grad_norm": 0.6911674737930298, "learning_rate": 1.3841584158415843e-05, "loss": 0.0264, "step": 3662 }, { "epoch": 14.478260869565217, "grad_norm": 0.5238652229309082, "learning_rate": 1.3831683168316834e-05, "loss": 0.0163, "step": 3663 }, { "epoch": 14.482213438735178, "grad_norm": 1.0457496643066406, "learning_rate": 1.3821782178217824e-05, "loss": 0.036, "step": 3664 }, { "epoch": 14.486166007905139, "grad_norm": 0.5210838317871094, "learning_rate": 1.3811881188118813e-05, "loss": 0.0253, "step": 3665 }, { "epoch": 14.4901185770751, "grad_norm": 0.8919087648391724, "learning_rate": 1.3801980198019804e-05, "loss": 0.0324, "step": 3666 }, { "epoch": 14.494071146245059, "grad_norm": 0.39363908767700195, "learning_rate": 1.3792079207920794e-05, "loss": 0.0173, "step": 3667 }, { "epoch": 14.49802371541502, "grad_norm": 0.5984033942222595, "learning_rate": 1.3782178217821785e-05, "loss": 0.0185, "step": 3668 }, { "epoch": 14.50197628458498, "grad_norm": 0.5198653340339661, "learning_rate": 1.3772277227722774e-05, "loss": 0.019, "step": 3669 }, { "epoch": 14.505928853754941, "grad_norm": 0.588720977306366, "learning_rate": 1.3762376237623763e-05, "loss": 0.0199, "step": 3670 }, { "epoch": 14.5098814229249, "grad_norm": 0.9170841574668884, "learning_rate": 1.3752475247524754e-05, "loss": 0.033, "step": 3671 }, { "epoch": 14.513833992094861, "grad_norm": 0.7681698203086853, "learning_rate": 1.3742574257425744e-05, "loss": 0.0265, "step": 3672 }, { "epoch": 14.517786561264822, "grad_norm": 0.8406161069869995, "learning_rate": 1.3732673267326735e-05, "loss": 0.0286, "step": 3673 }, { "epoch": 14.521739130434783, "grad_norm": 0.7611433267593384, "learning_rate": 1.3722772277227724e-05, "loss": 0.0224, "step": 3674 }, { "epoch": 14.525691699604742, "grad_norm": 0.6476555466651917, "learning_rate": 1.3712871287128715e-05, "loss": 0.0294, "step": 3675 }, { "epoch": 14.529644268774703, "grad_norm": 0.9361061453819275, "learning_rate": 1.3702970297029705e-05, "loss": 0.031, "step": 3676 }, { "epoch": 14.533596837944664, "grad_norm": 0.5194649696350098, "learning_rate": 1.3693069306930694e-05, "loss": 0.0164, "step": 3677 }, { "epoch": 14.537549407114625, "grad_norm": 0.9243749380111694, "learning_rate": 1.3683168316831685e-05, "loss": 0.0243, "step": 3678 }, { "epoch": 14.541501976284586, "grad_norm": 0.45838600397109985, "learning_rate": 1.3673267326732674e-05, "loss": 0.0158, "step": 3679 }, { "epoch": 14.545454545454545, "grad_norm": 0.7420250773429871, "learning_rate": 1.3663366336633666e-05, "loss": 0.0223, "step": 3680 }, { "epoch": 14.549407114624506, "grad_norm": 0.5305086970329285, "learning_rate": 1.3653465346534655e-05, "loss": 0.0253, "step": 3681 }, { "epoch": 14.553359683794467, "grad_norm": 0.7819458246231079, "learning_rate": 1.3643564356435646e-05, "loss": 0.0303, "step": 3682 }, { "epoch": 14.557312252964428, "grad_norm": 0.5438023805618286, "learning_rate": 1.3633663366336635e-05, "loss": 0.0261, "step": 3683 }, { "epoch": 14.561264822134387, "grad_norm": 0.4493376910686493, "learning_rate": 1.3623762376237625e-05, "loss": 0.0181, "step": 3684 }, { "epoch": 14.565217391304348, "grad_norm": 0.7461639642715454, "learning_rate": 1.3613861386138616e-05, "loss": 0.0308, "step": 3685 }, { "epoch": 14.569169960474309, "grad_norm": 0.6112428903579712, "learning_rate": 1.3603960396039605e-05, "loss": 0.018, "step": 3686 }, { "epoch": 14.57312252964427, "grad_norm": 0.8107467293739319, "learning_rate": 1.3594059405940596e-05, "loss": 0.0264, "step": 3687 }, { "epoch": 14.577075098814229, "grad_norm": 0.9947041273117065, "learning_rate": 1.3584158415841586e-05, "loss": 0.0334, "step": 3688 }, { "epoch": 14.58102766798419, "grad_norm": 0.5895665884017944, "learning_rate": 1.3574257425742577e-05, "loss": 0.0177, "step": 3689 }, { "epoch": 14.58498023715415, "grad_norm": 0.538038432598114, "learning_rate": 1.3564356435643566e-05, "loss": 0.0199, "step": 3690 }, { "epoch": 14.588932806324111, "grad_norm": 0.7492208480834961, "learning_rate": 1.3554455445544555e-05, "loss": 0.0306, "step": 3691 }, { "epoch": 14.59288537549407, "grad_norm": 0.6933621168136597, "learning_rate": 1.3544554455445546e-05, "loss": 0.0152, "step": 3692 }, { "epoch": 14.596837944664031, "grad_norm": 0.657656729221344, "learning_rate": 1.3534653465346536e-05, "loss": 0.0234, "step": 3693 }, { "epoch": 14.600790513833992, "grad_norm": 0.4955858886241913, "learning_rate": 1.3524752475247527e-05, "loss": 0.0157, "step": 3694 }, { "epoch": 14.604743083003953, "grad_norm": 0.6174148321151733, "learning_rate": 1.3514851485148516e-05, "loss": 0.0225, "step": 3695 }, { "epoch": 14.608695652173914, "grad_norm": 0.5257104635238647, "learning_rate": 1.3504950495049506e-05, "loss": 0.0235, "step": 3696 }, { "epoch": 14.612648221343873, "grad_norm": 0.6156019568443298, "learning_rate": 1.3495049504950497e-05, "loss": 0.0252, "step": 3697 }, { "epoch": 14.616600790513834, "grad_norm": 0.8544965982437134, "learning_rate": 1.3485148514851486e-05, "loss": 0.0275, "step": 3698 }, { "epoch": 14.620553359683795, "grad_norm": 0.5411393046379089, "learning_rate": 1.3475247524752477e-05, "loss": 0.0236, "step": 3699 }, { "epoch": 14.624505928853754, "grad_norm": 0.951758861541748, "learning_rate": 1.3465346534653467e-05, "loss": 0.0247, "step": 3700 }, { "epoch": 14.628458498023715, "grad_norm": 0.5254088640213013, "learning_rate": 1.3455445544554458e-05, "loss": 0.0259, "step": 3701 }, { "epoch": 14.632411067193676, "grad_norm": 0.5250906944274902, "learning_rate": 1.3445544554455447e-05, "loss": 0.0172, "step": 3702 }, { "epoch": 14.636363636363637, "grad_norm": 0.6953026056289673, "learning_rate": 1.3435643564356436e-05, "loss": 0.024, "step": 3703 }, { "epoch": 14.640316205533598, "grad_norm": 0.8992546796798706, "learning_rate": 1.3425742574257427e-05, "loss": 0.0248, "step": 3704 }, { "epoch": 14.644268774703557, "grad_norm": 0.6585764288902283, "learning_rate": 1.3415841584158417e-05, "loss": 0.028, "step": 3705 }, { "epoch": 14.648221343873518, "grad_norm": 1.4599233865737915, "learning_rate": 1.3405940594059408e-05, "loss": 0.0579, "step": 3706 }, { "epoch": 14.652173913043478, "grad_norm": 0.8026504516601562, "learning_rate": 1.3396039603960397e-05, "loss": 0.0239, "step": 3707 }, { "epoch": 14.65612648221344, "grad_norm": 0.9745014905929565, "learning_rate": 1.3386138613861388e-05, "loss": 0.0311, "step": 3708 }, { "epoch": 14.660079051383399, "grad_norm": 0.8238971829414368, "learning_rate": 1.3376237623762378e-05, "loss": 0.0263, "step": 3709 }, { "epoch": 14.66403162055336, "grad_norm": 0.659579873085022, "learning_rate": 1.3366336633663367e-05, "loss": 0.023, "step": 3710 }, { "epoch": 14.66798418972332, "grad_norm": 0.48981571197509766, "learning_rate": 1.3356435643564358e-05, "loss": 0.0192, "step": 3711 }, { "epoch": 14.671936758893281, "grad_norm": 0.4572812616825104, "learning_rate": 1.3346534653465347e-05, "loss": 0.0259, "step": 3712 }, { "epoch": 14.67588932806324, "grad_norm": 0.9332716464996338, "learning_rate": 1.3336633663366339e-05, "loss": 0.0309, "step": 3713 }, { "epoch": 14.679841897233201, "grad_norm": 0.5266364812850952, "learning_rate": 1.3326732673267328e-05, "loss": 0.0207, "step": 3714 }, { "epoch": 14.683794466403162, "grad_norm": 1.1232261657714844, "learning_rate": 1.3316831683168319e-05, "loss": 0.0299, "step": 3715 }, { "epoch": 14.687747035573123, "grad_norm": 0.4660709500312805, "learning_rate": 1.3306930693069308e-05, "loss": 0.0281, "step": 3716 }, { "epoch": 14.691699604743082, "grad_norm": 0.9695098996162415, "learning_rate": 1.3297029702970298e-05, "loss": 0.0285, "step": 3717 }, { "epoch": 14.695652173913043, "grad_norm": 0.9506553411483765, "learning_rate": 1.3287128712871289e-05, "loss": 0.0241, "step": 3718 }, { "epoch": 14.699604743083004, "grad_norm": 0.6543096899986267, "learning_rate": 1.3277227722772278e-05, "loss": 0.0186, "step": 3719 }, { "epoch": 14.703557312252965, "grad_norm": 0.7381693124771118, "learning_rate": 1.326732673267327e-05, "loss": 0.0224, "step": 3720 }, { "epoch": 14.707509881422926, "grad_norm": 1.1591260433197021, "learning_rate": 1.3257425742574259e-05, "loss": 0.0494, "step": 3721 }, { "epoch": 14.711462450592885, "grad_norm": 0.9342645406723022, "learning_rate": 1.324752475247525e-05, "loss": 0.0263, "step": 3722 }, { "epoch": 14.715415019762846, "grad_norm": 0.6863256096839905, "learning_rate": 1.3237623762376239e-05, "loss": 0.0294, "step": 3723 }, { "epoch": 14.719367588932807, "grad_norm": 1.1497865915298462, "learning_rate": 1.3227722772277228e-05, "loss": 0.0341, "step": 3724 }, { "epoch": 14.723320158102768, "grad_norm": 0.6921968460083008, "learning_rate": 1.321782178217822e-05, "loss": 0.039, "step": 3725 }, { "epoch": 14.727272727272727, "grad_norm": 0.7693561911582947, "learning_rate": 1.3207920792079209e-05, "loss": 0.0359, "step": 3726 }, { "epoch": 14.731225296442688, "grad_norm": 0.9658418893814087, "learning_rate": 1.31980198019802e-05, "loss": 0.0343, "step": 3727 }, { "epoch": 14.735177865612648, "grad_norm": 0.4896537661552429, "learning_rate": 1.318811881188119e-05, "loss": 0.0224, "step": 3728 }, { "epoch": 14.73913043478261, "grad_norm": 0.8054685592651367, "learning_rate": 1.3178217821782179e-05, "loss": 0.0215, "step": 3729 }, { "epoch": 14.743083003952568, "grad_norm": 0.8967522382736206, "learning_rate": 1.316831683168317e-05, "loss": 0.0525, "step": 3730 }, { "epoch": 14.74703557312253, "grad_norm": 0.5398697257041931, "learning_rate": 1.3158415841584159e-05, "loss": 0.0329, "step": 3731 }, { "epoch": 14.75098814229249, "grad_norm": 0.6790114641189575, "learning_rate": 1.314851485148515e-05, "loss": 0.028, "step": 3732 }, { "epoch": 14.754940711462451, "grad_norm": 0.6450980305671692, "learning_rate": 1.313861386138614e-05, "loss": 0.0219, "step": 3733 }, { "epoch": 14.75889328063241, "grad_norm": 0.2705237567424774, "learning_rate": 1.312871287128713e-05, "loss": 0.0214, "step": 3734 }, { "epoch": 14.762845849802371, "grad_norm": 0.5961162447929382, "learning_rate": 1.311881188118812e-05, "loss": 0.0261, "step": 3735 }, { "epoch": 14.766798418972332, "grad_norm": 0.8702595829963684, "learning_rate": 1.310891089108911e-05, "loss": 0.0332, "step": 3736 }, { "epoch": 14.770750988142293, "grad_norm": 0.938378632068634, "learning_rate": 1.30990099009901e-05, "loss": 0.0336, "step": 3737 }, { "epoch": 14.774703557312254, "grad_norm": 0.6059134602546692, "learning_rate": 1.308910891089109e-05, "loss": 0.0209, "step": 3738 }, { "epoch": 14.778656126482213, "grad_norm": 0.4290800392627716, "learning_rate": 1.307920792079208e-05, "loss": 0.0177, "step": 3739 }, { "epoch": 14.782608695652174, "grad_norm": 0.44265109300613403, "learning_rate": 1.306930693069307e-05, "loss": 0.0197, "step": 3740 }, { "epoch": 14.786561264822135, "grad_norm": 0.7318822145462036, "learning_rate": 1.3059405940594061e-05, "loss": 0.0303, "step": 3741 }, { "epoch": 14.790513833992096, "grad_norm": 0.5536202788352966, "learning_rate": 1.304950495049505e-05, "loss": 0.0203, "step": 3742 }, { "epoch": 14.794466403162055, "grad_norm": 0.7669658064842224, "learning_rate": 1.303960396039604e-05, "loss": 0.023, "step": 3743 }, { "epoch": 14.798418972332016, "grad_norm": 0.5953296422958374, "learning_rate": 1.3029702970297031e-05, "loss": 0.0189, "step": 3744 }, { "epoch": 14.802371541501977, "grad_norm": 0.8491002917289734, "learning_rate": 1.301980198019802e-05, "loss": 0.023, "step": 3745 }, { "epoch": 14.806324110671937, "grad_norm": 0.44980388879776, "learning_rate": 1.3009900990099012e-05, "loss": 0.0169, "step": 3746 }, { "epoch": 14.810276679841897, "grad_norm": 0.6375241875648499, "learning_rate": 1.3000000000000001e-05, "loss": 0.0322, "step": 3747 }, { "epoch": 14.814229249011857, "grad_norm": 0.4267948567867279, "learning_rate": 1.2990099009900992e-05, "loss": 0.023, "step": 3748 }, { "epoch": 14.818181818181818, "grad_norm": 0.6000099778175354, "learning_rate": 1.2980198019801981e-05, "loss": 0.0238, "step": 3749 }, { "epoch": 14.82213438735178, "grad_norm": 0.3940739631652832, "learning_rate": 1.297029702970297e-05, "loss": 0.0212, "step": 3750 }, { "epoch": 14.826086956521738, "grad_norm": 0.5429204106330872, "learning_rate": 1.2960396039603962e-05, "loss": 0.0188, "step": 3751 }, { "epoch": 14.8300395256917, "grad_norm": 0.37150952219963074, "learning_rate": 1.2950495049504951e-05, "loss": 0.0205, "step": 3752 }, { "epoch": 14.83399209486166, "grad_norm": 0.7802686095237732, "learning_rate": 1.2940594059405942e-05, "loss": 0.0286, "step": 3753 }, { "epoch": 14.837944664031621, "grad_norm": 0.5310541987419128, "learning_rate": 1.2930693069306932e-05, "loss": 0.028, "step": 3754 }, { "epoch": 14.841897233201582, "grad_norm": 0.5370839834213257, "learning_rate": 1.2920792079207921e-05, "loss": 0.0185, "step": 3755 }, { "epoch": 14.845849802371541, "grad_norm": 0.603613018989563, "learning_rate": 1.2910891089108912e-05, "loss": 0.0255, "step": 3756 }, { "epoch": 14.849802371541502, "grad_norm": 0.4527356028556824, "learning_rate": 1.2900990099009901e-05, "loss": 0.0209, "step": 3757 }, { "epoch": 14.853754940711463, "grad_norm": 0.5991864800453186, "learning_rate": 1.2891089108910892e-05, "loss": 0.0201, "step": 3758 }, { "epoch": 14.857707509881424, "grad_norm": 0.5085933208465576, "learning_rate": 1.2881188118811882e-05, "loss": 0.0201, "step": 3759 }, { "epoch": 14.861660079051383, "grad_norm": 0.4281972646713257, "learning_rate": 1.2871287128712873e-05, "loss": 0.022, "step": 3760 }, { "epoch": 14.865612648221344, "grad_norm": 0.9016755819320679, "learning_rate": 1.2861386138613862e-05, "loss": 0.0297, "step": 3761 }, { "epoch": 14.869565217391305, "grad_norm": 0.47945937514305115, "learning_rate": 1.2851485148514852e-05, "loss": 0.0235, "step": 3762 }, { "epoch": 14.873517786561266, "grad_norm": 0.49569401144981384, "learning_rate": 1.2841584158415843e-05, "loss": 0.0167, "step": 3763 }, { "epoch": 14.877470355731225, "grad_norm": 0.7921637892723083, "learning_rate": 1.2831683168316832e-05, "loss": 0.0327, "step": 3764 }, { "epoch": 14.881422924901186, "grad_norm": 0.7625458240509033, "learning_rate": 1.2821782178217823e-05, "loss": 0.0316, "step": 3765 }, { "epoch": 14.885375494071146, "grad_norm": 0.5508131384849548, "learning_rate": 1.2811881188118813e-05, "loss": 0.0203, "step": 3766 }, { "epoch": 14.889328063241107, "grad_norm": 0.7009062170982361, "learning_rate": 1.2801980198019804e-05, "loss": 0.0235, "step": 3767 }, { "epoch": 14.893280632411066, "grad_norm": 0.6729530096054077, "learning_rate": 1.2792079207920793e-05, "loss": 0.0203, "step": 3768 }, { "epoch": 14.897233201581027, "grad_norm": 0.5044990181922913, "learning_rate": 1.2782178217821782e-05, "loss": 0.0191, "step": 3769 }, { "epoch": 14.901185770750988, "grad_norm": 0.5740998983383179, "learning_rate": 1.2772277227722773e-05, "loss": 0.0302, "step": 3770 }, { "epoch": 14.90513833992095, "grad_norm": 0.5180506706237793, "learning_rate": 1.2762376237623763e-05, "loss": 0.0202, "step": 3771 }, { "epoch": 14.909090909090908, "grad_norm": 0.5842786431312561, "learning_rate": 1.2752475247524754e-05, "loss": 0.0196, "step": 3772 }, { "epoch": 14.91304347826087, "grad_norm": 0.5002758502960205, "learning_rate": 1.2742574257425743e-05, "loss": 0.0194, "step": 3773 }, { "epoch": 14.91699604743083, "grad_norm": 1.5320725440979004, "learning_rate": 1.2732673267326734e-05, "loss": 0.0525, "step": 3774 }, { "epoch": 14.920948616600791, "grad_norm": 0.7174404859542847, "learning_rate": 1.2722772277227724e-05, "loss": 0.0251, "step": 3775 }, { "epoch": 14.92490118577075, "grad_norm": 0.6786478161811829, "learning_rate": 1.2712871287128713e-05, "loss": 0.0196, "step": 3776 }, { "epoch": 14.928853754940711, "grad_norm": 0.6391962170600891, "learning_rate": 1.2702970297029704e-05, "loss": 0.0205, "step": 3777 }, { "epoch": 14.932806324110672, "grad_norm": 0.29246872663497925, "learning_rate": 1.2693069306930693e-05, "loss": 0.0219, "step": 3778 }, { "epoch": 14.936758893280633, "grad_norm": 0.33569401502609253, "learning_rate": 1.2683168316831685e-05, "loss": 0.0179, "step": 3779 }, { "epoch": 14.940711462450594, "grad_norm": 0.7932928204536438, "learning_rate": 1.2673267326732674e-05, "loss": 0.0214, "step": 3780 }, { "epoch": 14.944664031620553, "grad_norm": 0.7715370655059814, "learning_rate": 1.2663366336633665e-05, "loss": 0.0284, "step": 3781 }, { "epoch": 14.948616600790514, "grad_norm": 0.5586667060852051, "learning_rate": 1.2653465346534654e-05, "loss": 0.0221, "step": 3782 }, { "epoch": 14.952569169960475, "grad_norm": 0.4383533000946045, "learning_rate": 1.2643564356435644e-05, "loss": 0.0211, "step": 3783 }, { "epoch": 14.956521739130435, "grad_norm": 0.6669021248817444, "learning_rate": 1.2633663366336635e-05, "loss": 0.0243, "step": 3784 }, { "epoch": 14.960474308300395, "grad_norm": 0.25973597168922424, "learning_rate": 1.2623762376237624e-05, "loss": 0.0365, "step": 3785 }, { "epoch": 14.964426877470355, "grad_norm": 0.5743598937988281, "learning_rate": 1.2613861386138615e-05, "loss": 0.0158, "step": 3786 }, { "epoch": 14.968379446640316, "grad_norm": 0.6987953186035156, "learning_rate": 1.2603960396039605e-05, "loss": 0.0286, "step": 3787 }, { "epoch": 14.972332015810277, "grad_norm": 0.7313985228538513, "learning_rate": 1.2594059405940594e-05, "loss": 0.0269, "step": 3788 }, { "epoch": 14.976284584980236, "grad_norm": 0.4488266706466675, "learning_rate": 1.2584158415841585e-05, "loss": 0.0155, "step": 3789 }, { "epoch": 14.980237154150197, "grad_norm": 0.8420386910438538, "learning_rate": 1.2574257425742574e-05, "loss": 0.021, "step": 3790 }, { "epoch": 14.984189723320158, "grad_norm": 0.8363755345344543, "learning_rate": 1.2564356435643565e-05, "loss": 0.0221, "step": 3791 }, { "epoch": 14.988142292490119, "grad_norm": 0.859943687915802, "learning_rate": 1.2554455445544555e-05, "loss": 0.0315, "step": 3792 }, { "epoch": 14.992094861660078, "grad_norm": 0.998887836933136, "learning_rate": 1.2544554455445546e-05, "loss": 0.0343, "step": 3793 }, { "epoch": 14.996047430830039, "grad_norm": 0.49239662289619446, "learning_rate": 1.2534653465346535e-05, "loss": 0.0208, "step": 3794 }, { "epoch": 15.0, "grad_norm": 0.545988142490387, "learning_rate": 1.2524752475247525e-05, "loss": 0.0236, "step": 3795 }, { "epoch": 15.003952569169961, "grad_norm": 0.21478904783725739, "learning_rate": 1.2514851485148516e-05, "loss": 0.0232, "step": 3796 }, { "epoch": 15.007905138339922, "grad_norm": 0.257678359746933, "learning_rate": 1.2504950495049505e-05, "loss": 0.0305, "step": 3797 }, { "epoch": 15.011857707509881, "grad_norm": 0.33608025312423706, "learning_rate": 1.2495049504950496e-05, "loss": 0.0139, "step": 3798 }, { "epoch": 15.015810276679842, "grad_norm": 0.3401578366756439, "learning_rate": 1.2485148514851486e-05, "loss": 0.0219, "step": 3799 }, { "epoch": 15.019762845849803, "grad_norm": 0.23548820614814758, "learning_rate": 1.2475247524752477e-05, "loss": 0.0115, "step": 3800 }, { "epoch": 15.023715415019764, "grad_norm": 0.30257171392440796, "learning_rate": 1.2465346534653466e-05, "loss": 0.0173, "step": 3801 }, { "epoch": 15.027667984189723, "grad_norm": 1.2783746719360352, "learning_rate": 1.2455445544554455e-05, "loss": 0.0407, "step": 3802 }, { "epoch": 15.031620553359684, "grad_norm": 0.23515494167804718, "learning_rate": 1.2445544554455446e-05, "loss": 0.014, "step": 3803 }, { "epoch": 15.035573122529645, "grad_norm": 0.5252200365066528, "learning_rate": 1.2435643564356436e-05, "loss": 0.0307, "step": 3804 }, { "epoch": 15.039525691699605, "grad_norm": 0.5314760804176331, "learning_rate": 1.2425742574257427e-05, "loss": 0.0198, "step": 3805 }, { "epoch": 15.043478260869565, "grad_norm": 0.5392735004425049, "learning_rate": 1.2415841584158416e-05, "loss": 0.0228, "step": 3806 }, { "epoch": 15.047430830039525, "grad_norm": 0.5527291297912598, "learning_rate": 1.2405940594059407e-05, "loss": 0.0168, "step": 3807 }, { "epoch": 15.051383399209486, "grad_norm": 0.3514786958694458, "learning_rate": 1.2396039603960397e-05, "loss": 0.0177, "step": 3808 }, { "epoch": 15.055335968379447, "grad_norm": 0.49954769015312195, "learning_rate": 1.2386138613861386e-05, "loss": 0.0252, "step": 3809 }, { "epoch": 15.059288537549406, "grad_norm": 0.31307169795036316, "learning_rate": 1.2376237623762377e-05, "loss": 0.0195, "step": 3810 }, { "epoch": 15.063241106719367, "grad_norm": 0.4472981095314026, "learning_rate": 1.2366336633663366e-05, "loss": 0.0157, "step": 3811 }, { "epoch": 15.067193675889328, "grad_norm": 0.3131963014602661, "learning_rate": 1.2356435643564358e-05, "loss": 0.019, "step": 3812 }, { "epoch": 15.071146245059289, "grad_norm": 0.49445632100105286, "learning_rate": 1.2346534653465347e-05, "loss": 0.0204, "step": 3813 }, { "epoch": 15.075098814229248, "grad_norm": 0.3126378655433655, "learning_rate": 1.2336633663366336e-05, "loss": 0.0187, "step": 3814 }, { "epoch": 15.079051383399209, "grad_norm": 0.33764222264289856, "learning_rate": 1.2326732673267327e-05, "loss": 0.0176, "step": 3815 }, { "epoch": 15.08300395256917, "grad_norm": 0.6490135788917542, "learning_rate": 1.2316831683168317e-05, "loss": 0.0377, "step": 3816 }, { "epoch": 15.08695652173913, "grad_norm": 0.18221792578697205, "learning_rate": 1.2306930693069308e-05, "loss": 0.0173, "step": 3817 }, { "epoch": 15.090909090909092, "grad_norm": 0.22651997208595276, "learning_rate": 1.2297029702970297e-05, "loss": 0.029, "step": 3818 }, { "epoch": 15.09486166007905, "grad_norm": 0.2887817919254303, "learning_rate": 1.2287128712871288e-05, "loss": 0.019, "step": 3819 }, { "epoch": 15.098814229249012, "grad_norm": 0.4877634048461914, "learning_rate": 1.2277227722772278e-05, "loss": 0.0219, "step": 3820 }, { "epoch": 15.102766798418973, "grad_norm": 0.32458046078681946, "learning_rate": 1.2267326732673267e-05, "loss": 0.0185, "step": 3821 }, { "epoch": 15.106719367588934, "grad_norm": 0.5022981762886047, "learning_rate": 1.2257425742574258e-05, "loss": 0.0168, "step": 3822 }, { "epoch": 15.110671936758893, "grad_norm": 0.3847731053829193, "learning_rate": 1.2247524752475247e-05, "loss": 0.0218, "step": 3823 }, { "epoch": 15.114624505928854, "grad_norm": 0.27849534153938293, "learning_rate": 1.2237623762376238e-05, "loss": 0.0275, "step": 3824 }, { "epoch": 15.118577075098814, "grad_norm": 0.3776545226573944, "learning_rate": 1.2227722772277228e-05, "loss": 0.0186, "step": 3825 }, { "epoch": 15.122529644268775, "grad_norm": 0.37147992849349976, "learning_rate": 1.2217821782178219e-05, "loss": 0.0166, "step": 3826 }, { "epoch": 15.126482213438734, "grad_norm": 0.8612362146377563, "learning_rate": 1.2207920792079208e-05, "loss": 0.0336, "step": 3827 }, { "epoch": 15.130434782608695, "grad_norm": 0.27339717745780945, "learning_rate": 1.2198019801980198e-05, "loss": 0.0266, "step": 3828 }, { "epoch": 15.134387351778656, "grad_norm": 0.6711599826812744, "learning_rate": 1.2188118811881189e-05, "loss": 0.0227, "step": 3829 }, { "epoch": 15.138339920948617, "grad_norm": 0.29427266120910645, "learning_rate": 1.2178217821782178e-05, "loss": 0.0121, "step": 3830 }, { "epoch": 15.142292490118576, "grad_norm": 0.6709986329078674, "learning_rate": 1.2168316831683169e-05, "loss": 0.0331, "step": 3831 }, { "epoch": 15.146245059288537, "grad_norm": 0.8062861561775208, "learning_rate": 1.2158415841584158e-05, "loss": 0.0152, "step": 3832 }, { "epoch": 15.150197628458498, "grad_norm": 0.2873133718967438, "learning_rate": 1.214851485148515e-05, "loss": 0.0199, "step": 3833 }, { "epoch": 15.154150197628459, "grad_norm": 0.3068978786468506, "learning_rate": 1.2138613861386139e-05, "loss": 0.0453, "step": 3834 }, { "epoch": 15.15810276679842, "grad_norm": 0.225699320435524, "learning_rate": 1.2128712871287128e-05, "loss": 0.0236, "step": 3835 }, { "epoch": 15.162055335968379, "grad_norm": 0.3559662103652954, "learning_rate": 1.211881188118812e-05, "loss": 0.022, "step": 3836 }, { "epoch": 15.16600790513834, "grad_norm": 0.16527846455574036, "learning_rate": 1.2108910891089109e-05, "loss": 0.0175, "step": 3837 }, { "epoch": 15.1699604743083, "grad_norm": 0.14490391314029694, "learning_rate": 1.20990099009901e-05, "loss": 0.0189, "step": 3838 }, { "epoch": 15.173913043478262, "grad_norm": 0.34588634967803955, "learning_rate": 1.208910891089109e-05, "loss": 0.0271, "step": 3839 }, { "epoch": 15.17786561264822, "grad_norm": 0.5971683263778687, "learning_rate": 1.207920792079208e-05, "loss": 0.0201, "step": 3840 }, { "epoch": 15.181818181818182, "grad_norm": 0.12139047682285309, "learning_rate": 1.206930693069307e-05, "loss": 0.0208, "step": 3841 }, { "epoch": 15.185770750988143, "grad_norm": 0.31537216901779175, "learning_rate": 1.2059405940594059e-05, "loss": 0.0299, "step": 3842 }, { "epoch": 15.189723320158103, "grad_norm": 0.9793798923492432, "learning_rate": 1.204950495049505e-05, "loss": 0.04, "step": 3843 }, { "epoch": 15.193675889328063, "grad_norm": 0.64158695936203, "learning_rate": 1.203960396039604e-05, "loss": 0.0206, "step": 3844 }, { "epoch": 15.197628458498023, "grad_norm": 0.2930070161819458, "learning_rate": 1.202970297029703e-05, "loss": 0.0369, "step": 3845 }, { "epoch": 15.201581027667984, "grad_norm": 0.3768121898174286, "learning_rate": 1.201980198019802e-05, "loss": 0.0246, "step": 3846 }, { "epoch": 15.205533596837945, "grad_norm": 0.9296312928199768, "learning_rate": 1.200990099009901e-05, "loss": 0.0485, "step": 3847 }, { "epoch": 15.209486166007904, "grad_norm": 0.5283029079437256, "learning_rate": 1.2e-05, "loss": 0.0251, "step": 3848 }, { "epoch": 15.213438735177865, "grad_norm": 0.3624534606933594, "learning_rate": 1.199009900990099e-05, "loss": 0.0253, "step": 3849 }, { "epoch": 15.217391304347826, "grad_norm": 0.23778383433818817, "learning_rate": 1.198019801980198e-05, "loss": 0.0164, "step": 3850 }, { "epoch": 15.221343873517787, "grad_norm": 0.2523452639579773, "learning_rate": 1.197029702970297e-05, "loss": 0.0188, "step": 3851 }, { "epoch": 15.225296442687746, "grad_norm": 0.41245633363723755, "learning_rate": 1.1960396039603961e-05, "loss": 0.0139, "step": 3852 }, { "epoch": 15.229249011857707, "grad_norm": 0.41162481904029846, "learning_rate": 1.195049504950495e-05, "loss": 0.0153, "step": 3853 }, { "epoch": 15.233201581027668, "grad_norm": 0.263547420501709, "learning_rate": 1.194059405940594e-05, "loss": 0.0157, "step": 3854 }, { "epoch": 15.237154150197629, "grad_norm": 0.7901440262794495, "learning_rate": 1.1930693069306931e-05, "loss": 0.0261, "step": 3855 }, { "epoch": 15.24110671936759, "grad_norm": 0.21986545622348785, "learning_rate": 1.192079207920792e-05, "loss": 0.0285, "step": 3856 }, { "epoch": 15.245059288537549, "grad_norm": 0.2297523468732834, "learning_rate": 1.1910891089108911e-05, "loss": 0.0301, "step": 3857 }, { "epoch": 15.24901185770751, "grad_norm": 0.379594087600708, "learning_rate": 1.19009900990099e-05, "loss": 0.0225, "step": 3858 }, { "epoch": 15.25296442687747, "grad_norm": 0.5585339665412903, "learning_rate": 1.1891089108910892e-05, "loss": 0.027, "step": 3859 }, { "epoch": 15.256916996047432, "grad_norm": 0.6036040186882019, "learning_rate": 1.1881188118811881e-05, "loss": 0.0283, "step": 3860 }, { "epoch": 15.26086956521739, "grad_norm": 0.6855416297912598, "learning_rate": 1.187128712871287e-05, "loss": 0.0287, "step": 3861 }, { "epoch": 15.264822134387352, "grad_norm": 0.4473704695701599, "learning_rate": 1.1861386138613862e-05, "loss": 0.0165, "step": 3862 }, { "epoch": 15.268774703557312, "grad_norm": 0.29710209369659424, "learning_rate": 1.1851485148514851e-05, "loss": 0.0258, "step": 3863 }, { "epoch": 15.272727272727273, "grad_norm": 0.5023214221000671, "learning_rate": 1.1841584158415842e-05, "loss": 0.0266, "step": 3864 }, { "epoch": 15.276679841897232, "grad_norm": 0.6661540865898132, "learning_rate": 1.1831683168316831e-05, "loss": 0.0258, "step": 3865 }, { "epoch": 15.280632411067193, "grad_norm": 0.7024422287940979, "learning_rate": 1.1821782178217823e-05, "loss": 0.0198, "step": 3866 }, { "epoch": 15.284584980237154, "grad_norm": 0.5760118961334229, "learning_rate": 1.1811881188118812e-05, "loss": 0.0229, "step": 3867 }, { "epoch": 15.288537549407115, "grad_norm": 0.3894463777542114, "learning_rate": 1.1801980198019801e-05, "loss": 0.0265, "step": 3868 }, { "epoch": 15.292490118577074, "grad_norm": 0.4161767363548279, "learning_rate": 1.1792079207920792e-05, "loss": 0.0169, "step": 3869 }, { "epoch": 15.296442687747035, "grad_norm": 0.2954789400100708, "learning_rate": 1.1782178217821782e-05, "loss": 0.0138, "step": 3870 }, { "epoch": 15.300395256916996, "grad_norm": 0.4153381586074829, "learning_rate": 1.1772277227722773e-05, "loss": 0.0242, "step": 3871 }, { "epoch": 15.304347826086957, "grad_norm": 0.3821921944618225, "learning_rate": 1.1762376237623762e-05, "loss": 0.024, "step": 3872 }, { "epoch": 15.308300395256918, "grad_norm": 0.18060734868049622, "learning_rate": 1.1752475247524752e-05, "loss": 0.0213, "step": 3873 }, { "epoch": 15.312252964426877, "grad_norm": 0.3451940715312958, "learning_rate": 1.1742574257425743e-05, "loss": 0.0143, "step": 3874 }, { "epoch": 15.316205533596838, "grad_norm": 0.37840673327445984, "learning_rate": 1.1732673267326732e-05, "loss": 0.0334, "step": 3875 }, { "epoch": 15.320158102766799, "grad_norm": 0.36275747418403625, "learning_rate": 1.1722772277227723e-05, "loss": 0.021, "step": 3876 }, { "epoch": 15.32411067193676, "grad_norm": 0.42950689792633057, "learning_rate": 1.1712871287128714e-05, "loss": 0.035, "step": 3877 }, { "epoch": 15.328063241106719, "grad_norm": 0.650473415851593, "learning_rate": 1.1702970297029703e-05, "loss": 0.0176, "step": 3878 }, { "epoch": 15.33201581027668, "grad_norm": 0.45645561814308167, "learning_rate": 1.1693069306930695e-05, "loss": 0.026, "step": 3879 }, { "epoch": 15.33596837944664, "grad_norm": 0.31588423252105713, "learning_rate": 1.1683168316831684e-05, "loss": 0.0187, "step": 3880 }, { "epoch": 15.339920948616601, "grad_norm": 0.6248782277107239, "learning_rate": 1.1673267326732675e-05, "loss": 0.0252, "step": 3881 }, { "epoch": 15.34387351778656, "grad_norm": 0.5033974051475525, "learning_rate": 1.1663366336633664e-05, "loss": 0.019, "step": 3882 }, { "epoch": 15.347826086956522, "grad_norm": 0.2043386697769165, "learning_rate": 1.1653465346534655e-05, "loss": 0.0174, "step": 3883 }, { "epoch": 15.351778656126482, "grad_norm": 0.22434967756271362, "learning_rate": 1.1643564356435645e-05, "loss": 0.0191, "step": 3884 }, { "epoch": 15.355731225296443, "grad_norm": 0.4280073046684265, "learning_rate": 1.1633663366336634e-05, "loss": 0.0188, "step": 3885 }, { "epoch": 15.359683794466402, "grad_norm": 0.4583016037940979, "learning_rate": 1.1623762376237625e-05, "loss": 0.0202, "step": 3886 }, { "epoch": 15.363636363636363, "grad_norm": 0.38481152057647705, "learning_rate": 1.1613861386138615e-05, "loss": 0.0248, "step": 3887 }, { "epoch": 15.367588932806324, "grad_norm": 0.4266909658908844, "learning_rate": 1.1603960396039606e-05, "loss": 0.0198, "step": 3888 }, { "epoch": 15.371541501976285, "grad_norm": 0.158377543091774, "learning_rate": 1.1594059405940595e-05, "loss": 0.018, "step": 3889 }, { "epoch": 15.375494071146244, "grad_norm": 0.3283778429031372, "learning_rate": 1.1584158415841586e-05, "loss": 0.0214, "step": 3890 }, { "epoch": 15.379446640316205, "grad_norm": 0.7898430824279785, "learning_rate": 1.1574257425742575e-05, "loss": 0.0357, "step": 3891 }, { "epoch": 15.383399209486166, "grad_norm": 0.23478594422340393, "learning_rate": 1.1564356435643565e-05, "loss": 0.0265, "step": 3892 }, { "epoch": 15.387351778656127, "grad_norm": 0.35603922605514526, "learning_rate": 1.1554455445544556e-05, "loss": 0.0146, "step": 3893 }, { "epoch": 15.391304347826088, "grad_norm": 0.9637416005134583, "learning_rate": 1.1544554455445545e-05, "loss": 0.0273, "step": 3894 }, { "epoch": 15.395256916996047, "grad_norm": 0.37616217136383057, "learning_rate": 1.1534653465346536e-05, "loss": 0.0174, "step": 3895 }, { "epoch": 15.399209486166008, "grad_norm": 0.837296187877655, "learning_rate": 1.1524752475247526e-05, "loss": 0.0242, "step": 3896 }, { "epoch": 15.403162055335969, "grad_norm": 0.48693567514419556, "learning_rate": 1.1514851485148515e-05, "loss": 0.0175, "step": 3897 }, { "epoch": 15.40711462450593, "grad_norm": 0.3840150833129883, "learning_rate": 1.1504950495049506e-05, "loss": 0.022, "step": 3898 }, { "epoch": 15.411067193675889, "grad_norm": 1.3014194965362549, "learning_rate": 1.1495049504950496e-05, "loss": 0.0222, "step": 3899 }, { "epoch": 15.41501976284585, "grad_norm": 0.37161508202552795, "learning_rate": 1.1485148514851487e-05, "loss": 0.016, "step": 3900 }, { "epoch": 15.41897233201581, "grad_norm": 0.37366023659706116, "learning_rate": 1.1475247524752476e-05, "loss": 0.0135, "step": 3901 }, { "epoch": 15.422924901185771, "grad_norm": 0.24977493286132812, "learning_rate": 1.1465346534653467e-05, "loss": 0.0206, "step": 3902 }, { "epoch": 15.42687747035573, "grad_norm": 0.3055574595928192, "learning_rate": 1.1455445544554456e-05, "loss": 0.0192, "step": 3903 }, { "epoch": 15.430830039525691, "grad_norm": 0.4926667809486389, "learning_rate": 1.1445544554455446e-05, "loss": 0.0218, "step": 3904 }, { "epoch": 15.434782608695652, "grad_norm": 0.44113072752952576, "learning_rate": 1.1435643564356437e-05, "loss": 0.0264, "step": 3905 }, { "epoch": 15.438735177865613, "grad_norm": 0.6281216144561768, "learning_rate": 1.1425742574257426e-05, "loss": 0.0195, "step": 3906 }, { "epoch": 15.442687747035572, "grad_norm": 0.25815731287002563, "learning_rate": 1.1415841584158417e-05, "loss": 0.0163, "step": 3907 }, { "epoch": 15.446640316205533, "grad_norm": 0.3738540709018707, "learning_rate": 1.1405940594059407e-05, "loss": 0.0168, "step": 3908 }, { "epoch": 15.450592885375494, "grad_norm": 0.3694983422756195, "learning_rate": 1.1396039603960398e-05, "loss": 0.0216, "step": 3909 }, { "epoch": 15.454545454545455, "grad_norm": 0.24378107488155365, "learning_rate": 1.1386138613861387e-05, "loss": 0.0202, "step": 3910 }, { "epoch": 15.458498023715414, "grad_norm": 0.5399953722953796, "learning_rate": 1.1376237623762376e-05, "loss": 0.0218, "step": 3911 }, { "epoch": 15.462450592885375, "grad_norm": 0.4094056189060211, "learning_rate": 1.1366336633663368e-05, "loss": 0.0193, "step": 3912 }, { "epoch": 15.466403162055336, "grad_norm": 0.1998009830713272, "learning_rate": 1.1356435643564357e-05, "loss": 0.0195, "step": 3913 }, { "epoch": 15.470355731225297, "grad_norm": 0.2376052737236023, "learning_rate": 1.1346534653465348e-05, "loss": 0.0176, "step": 3914 }, { "epoch": 15.474308300395258, "grad_norm": 0.44167035818099976, "learning_rate": 1.1336633663366337e-05, "loss": 0.0185, "step": 3915 }, { "epoch": 15.478260869565217, "grad_norm": 0.580353856086731, "learning_rate": 1.1326732673267328e-05, "loss": 0.0307, "step": 3916 }, { "epoch": 15.482213438735178, "grad_norm": 0.34564733505249023, "learning_rate": 1.1316831683168318e-05, "loss": 0.0174, "step": 3917 }, { "epoch": 15.486166007905139, "grad_norm": 0.22003570199012756, "learning_rate": 1.1306930693069307e-05, "loss": 0.0263, "step": 3918 }, { "epoch": 15.4901185770751, "grad_norm": 0.478954553604126, "learning_rate": 1.1297029702970298e-05, "loss": 0.0153, "step": 3919 }, { "epoch": 15.494071146245059, "grad_norm": 0.6658238172531128, "learning_rate": 1.1287128712871288e-05, "loss": 0.0317, "step": 3920 }, { "epoch": 15.49802371541502, "grad_norm": 0.663354218006134, "learning_rate": 1.1277227722772279e-05, "loss": 0.0244, "step": 3921 }, { "epoch": 15.50197628458498, "grad_norm": 0.39337411522865295, "learning_rate": 1.1267326732673268e-05, "loss": 0.0299, "step": 3922 }, { "epoch": 15.505928853754941, "grad_norm": 0.3883029818534851, "learning_rate": 1.1257425742574257e-05, "loss": 0.0204, "step": 3923 }, { "epoch": 15.5098814229249, "grad_norm": 0.16640141606330872, "learning_rate": 1.1247524752475248e-05, "loss": 0.0179, "step": 3924 }, { "epoch": 15.513833992094861, "grad_norm": 0.14372046291828156, "learning_rate": 1.1237623762376238e-05, "loss": 0.0226, "step": 3925 }, { "epoch": 15.517786561264822, "grad_norm": 0.6675691604614258, "learning_rate": 1.1227722772277229e-05, "loss": 0.0203, "step": 3926 }, { "epoch": 15.521739130434783, "grad_norm": 0.28566956520080566, "learning_rate": 1.1217821782178218e-05, "loss": 0.0222, "step": 3927 }, { "epoch": 15.525691699604742, "grad_norm": 0.3468085825443268, "learning_rate": 1.120792079207921e-05, "loss": 0.0203, "step": 3928 }, { "epoch": 15.529644268774703, "grad_norm": 0.30228281021118164, "learning_rate": 1.1198019801980199e-05, "loss": 0.0161, "step": 3929 }, { "epoch": 15.533596837944664, "grad_norm": 0.29542475938796997, "learning_rate": 1.1188118811881188e-05, "loss": 0.0197, "step": 3930 }, { "epoch": 15.537549407114625, "grad_norm": 1.229561448097229, "learning_rate": 1.117821782178218e-05, "loss": 0.05, "step": 3931 }, { "epoch": 15.541501976284586, "grad_norm": 0.33288198709487915, "learning_rate": 1.1168316831683169e-05, "loss": 0.0149, "step": 3932 }, { "epoch": 15.545454545454545, "grad_norm": 0.3907518684864044, "learning_rate": 1.115841584158416e-05, "loss": 0.0265, "step": 3933 }, { "epoch": 15.549407114624506, "grad_norm": 0.38764727115631104, "learning_rate": 1.1148514851485149e-05, "loss": 0.0219, "step": 3934 }, { "epoch": 15.553359683794467, "grad_norm": 0.24306076765060425, "learning_rate": 1.113861386138614e-05, "loss": 0.0178, "step": 3935 }, { "epoch": 15.557312252964428, "grad_norm": 0.32089412212371826, "learning_rate": 1.112871287128713e-05, "loss": 0.0283, "step": 3936 }, { "epoch": 15.561264822134387, "grad_norm": 0.42798247933387756, "learning_rate": 1.1118811881188119e-05, "loss": 0.0199, "step": 3937 }, { "epoch": 15.565217391304348, "grad_norm": 0.6173682808876038, "learning_rate": 1.110891089108911e-05, "loss": 0.0342, "step": 3938 }, { "epoch": 15.569169960474309, "grad_norm": 1.2680635452270508, "learning_rate": 1.10990099009901e-05, "loss": 0.0343, "step": 3939 }, { "epoch": 15.57312252964427, "grad_norm": 0.7359910011291504, "learning_rate": 1.108910891089109e-05, "loss": 0.0437, "step": 3940 }, { "epoch": 15.577075098814229, "grad_norm": 0.26753005385398865, "learning_rate": 1.107920792079208e-05, "loss": 0.0125, "step": 3941 }, { "epoch": 15.58102766798419, "grad_norm": 0.7229536175727844, "learning_rate": 1.106930693069307e-05, "loss": 0.0206, "step": 3942 }, { "epoch": 15.58498023715415, "grad_norm": 0.5124853849411011, "learning_rate": 1.105940594059406e-05, "loss": 0.027, "step": 3943 }, { "epoch": 15.588932806324111, "grad_norm": 0.2861965000629425, "learning_rate": 1.104950495049505e-05, "loss": 0.0182, "step": 3944 }, { "epoch": 15.59288537549407, "grad_norm": 0.4193480312824249, "learning_rate": 1.103960396039604e-05, "loss": 0.0201, "step": 3945 }, { "epoch": 15.596837944664031, "grad_norm": 0.24027983844280243, "learning_rate": 1.102970297029703e-05, "loss": 0.0169, "step": 3946 }, { "epoch": 15.600790513833992, "grad_norm": 0.4907889664173126, "learning_rate": 1.1019801980198021e-05, "loss": 0.015, "step": 3947 }, { "epoch": 15.604743083003953, "grad_norm": 0.1506580263376236, "learning_rate": 1.100990099009901e-05, "loss": 0.0188, "step": 3948 }, { "epoch": 15.608695652173914, "grad_norm": 0.5318458080291748, "learning_rate": 1.1000000000000001e-05, "loss": 0.0289, "step": 3949 }, { "epoch": 15.612648221343873, "grad_norm": 0.2659936845302582, "learning_rate": 1.099009900990099e-05, "loss": 0.0189, "step": 3950 }, { "epoch": 15.616600790513834, "grad_norm": 0.48690709471702576, "learning_rate": 1.098019801980198e-05, "loss": 0.0181, "step": 3951 }, { "epoch": 15.620553359683795, "grad_norm": 0.3975117802619934, "learning_rate": 1.0970297029702971e-05, "loss": 0.0337, "step": 3952 }, { "epoch": 15.624505928853754, "grad_norm": 0.41925132274627686, "learning_rate": 1.096039603960396e-05, "loss": 0.0231, "step": 3953 }, { "epoch": 15.628458498023715, "grad_norm": 0.5792222023010254, "learning_rate": 1.0950495049504952e-05, "loss": 0.0227, "step": 3954 }, { "epoch": 15.632411067193676, "grad_norm": 0.9861533045768738, "learning_rate": 1.0940594059405941e-05, "loss": 0.045, "step": 3955 }, { "epoch": 15.636363636363637, "grad_norm": 0.5323643684387207, "learning_rate": 1.093069306930693e-05, "loss": 0.0253, "step": 3956 }, { "epoch": 15.640316205533598, "grad_norm": 0.6773537397384644, "learning_rate": 1.0920792079207921e-05, "loss": 0.0205, "step": 3957 }, { "epoch": 15.644268774703557, "grad_norm": 0.47213515639305115, "learning_rate": 1.091089108910891e-05, "loss": 0.0211, "step": 3958 }, { "epoch": 15.648221343873518, "grad_norm": 0.3278203308582306, "learning_rate": 1.0900990099009902e-05, "loss": 0.0237, "step": 3959 }, { "epoch": 15.652173913043478, "grad_norm": 0.4708459675312042, "learning_rate": 1.0891089108910891e-05, "loss": 0.0271, "step": 3960 }, { "epoch": 15.65612648221344, "grad_norm": 0.8619837760925293, "learning_rate": 1.0881188118811882e-05, "loss": 0.0298, "step": 3961 }, { "epoch": 15.660079051383399, "grad_norm": 0.7651503682136536, "learning_rate": 1.0871287128712872e-05, "loss": 0.0264, "step": 3962 }, { "epoch": 15.66403162055336, "grad_norm": 0.3128441572189331, "learning_rate": 1.0861386138613861e-05, "loss": 0.0413, "step": 3963 }, { "epoch": 15.66798418972332, "grad_norm": 0.3915279507637024, "learning_rate": 1.0851485148514852e-05, "loss": 0.0227, "step": 3964 }, { "epoch": 15.671936758893281, "grad_norm": 0.340460866689682, "learning_rate": 1.0841584158415842e-05, "loss": 0.0214, "step": 3965 }, { "epoch": 15.67588932806324, "grad_norm": 0.47710084915161133, "learning_rate": 1.0831683168316833e-05, "loss": 0.0395, "step": 3966 }, { "epoch": 15.679841897233201, "grad_norm": 0.3024883568286896, "learning_rate": 1.0821782178217822e-05, "loss": 0.0173, "step": 3967 }, { "epoch": 15.683794466403162, "grad_norm": 0.36023378372192383, "learning_rate": 1.0811881188118813e-05, "loss": 0.013, "step": 3968 }, { "epoch": 15.687747035573123, "grad_norm": 0.4446355104446411, "learning_rate": 1.0801980198019802e-05, "loss": 0.0165, "step": 3969 }, { "epoch": 15.691699604743082, "grad_norm": 0.5827638506889343, "learning_rate": 1.0792079207920792e-05, "loss": 0.022, "step": 3970 }, { "epoch": 15.695652173913043, "grad_norm": 0.39204803109169006, "learning_rate": 1.0782178217821783e-05, "loss": 0.0177, "step": 3971 }, { "epoch": 15.699604743083004, "grad_norm": 0.6072433590888977, "learning_rate": 1.0772277227722772e-05, "loss": 0.0194, "step": 3972 }, { "epoch": 15.703557312252965, "grad_norm": 0.5263144373893738, "learning_rate": 1.0762376237623763e-05, "loss": 0.0248, "step": 3973 }, { "epoch": 15.707509881422926, "grad_norm": 0.624946653842926, "learning_rate": 1.0752475247524753e-05, "loss": 0.042, "step": 3974 }, { "epoch": 15.711462450592885, "grad_norm": 0.23140814900398254, "learning_rate": 1.0742574257425744e-05, "loss": 0.0191, "step": 3975 }, { "epoch": 15.715415019762846, "grad_norm": 0.818532407283783, "learning_rate": 1.0732673267326733e-05, "loss": 0.0362, "step": 3976 }, { "epoch": 15.719367588932807, "grad_norm": 0.38738036155700684, "learning_rate": 1.0722772277227722e-05, "loss": 0.0349, "step": 3977 }, { "epoch": 15.723320158102768, "grad_norm": 0.19291234016418457, "learning_rate": 1.0712871287128714e-05, "loss": 0.0159, "step": 3978 }, { "epoch": 15.727272727272727, "grad_norm": 0.879473090171814, "learning_rate": 1.0702970297029703e-05, "loss": 0.0263, "step": 3979 }, { "epoch": 15.731225296442688, "grad_norm": 0.510424017906189, "learning_rate": 1.0693069306930694e-05, "loss": 0.0214, "step": 3980 }, { "epoch": 15.735177865612648, "grad_norm": 0.15624366700649261, "learning_rate": 1.0683168316831683e-05, "loss": 0.0204, "step": 3981 }, { "epoch": 15.73913043478261, "grad_norm": 0.39847809076309204, "learning_rate": 1.0673267326732673e-05, "loss": 0.016, "step": 3982 }, { "epoch": 15.743083003952568, "grad_norm": 0.15956468880176544, "learning_rate": 1.0663366336633664e-05, "loss": 0.0301, "step": 3983 }, { "epoch": 15.74703557312253, "grad_norm": 0.47719860076904297, "learning_rate": 1.0653465346534653e-05, "loss": 0.0202, "step": 3984 }, { "epoch": 15.75098814229249, "grad_norm": 0.14814560115337372, "learning_rate": 1.0643564356435644e-05, "loss": 0.0183, "step": 3985 }, { "epoch": 15.754940711462451, "grad_norm": 0.5177909731864929, "learning_rate": 1.0633663366336634e-05, "loss": 0.0202, "step": 3986 }, { "epoch": 15.75889328063241, "grad_norm": 0.7555588483810425, "learning_rate": 1.0623762376237625e-05, "loss": 0.0259, "step": 3987 }, { "epoch": 15.762845849802371, "grad_norm": 0.352934330701828, "learning_rate": 1.0613861386138614e-05, "loss": 0.0253, "step": 3988 }, { "epoch": 15.766798418972332, "grad_norm": 0.23178963363170624, "learning_rate": 1.0603960396039603e-05, "loss": 0.0181, "step": 3989 }, { "epoch": 15.770750988142293, "grad_norm": 0.5159963369369507, "learning_rate": 1.0594059405940594e-05, "loss": 0.0284, "step": 3990 }, { "epoch": 15.774703557312254, "grad_norm": 0.48542124032974243, "learning_rate": 1.0584158415841584e-05, "loss": 0.017, "step": 3991 }, { "epoch": 15.778656126482213, "grad_norm": 0.5381563305854797, "learning_rate": 1.0574257425742575e-05, "loss": 0.0259, "step": 3992 }, { "epoch": 15.782608695652174, "grad_norm": 0.2505844533443451, "learning_rate": 1.0564356435643564e-05, "loss": 0.0247, "step": 3993 }, { "epoch": 15.786561264822135, "grad_norm": 0.3752920627593994, "learning_rate": 1.0554455445544555e-05, "loss": 0.0167, "step": 3994 }, { "epoch": 15.790513833992096, "grad_norm": 0.34744518995285034, "learning_rate": 1.0544554455445545e-05, "loss": 0.0207, "step": 3995 }, { "epoch": 15.794466403162055, "grad_norm": 0.27527621388435364, "learning_rate": 1.0534653465346534e-05, "loss": 0.0164, "step": 3996 }, { "epoch": 15.798418972332016, "grad_norm": 0.2954518795013428, "learning_rate": 1.0524752475247525e-05, "loss": 0.0194, "step": 3997 }, { "epoch": 15.802371541501977, "grad_norm": 0.14628306031227112, "learning_rate": 1.0514851485148515e-05, "loss": 0.0164, "step": 3998 }, { "epoch": 15.806324110671937, "grad_norm": 0.8714108467102051, "learning_rate": 1.0504950495049506e-05, "loss": 0.0297, "step": 3999 }, { "epoch": 15.810276679841897, "grad_norm": 0.528918981552124, "learning_rate": 1.0495049504950495e-05, "loss": 0.0247, "step": 4000 }, { "epoch": 15.814229249011857, "grad_norm": 0.7067568302154541, "learning_rate": 1.0485148514851486e-05, "loss": 0.0242, "step": 4001 }, { "epoch": 15.818181818181818, "grad_norm": 0.35606256127357483, "learning_rate": 1.0475247524752475e-05, "loss": 0.0159, "step": 4002 }, { "epoch": 15.82213438735178, "grad_norm": 0.3375142216682434, "learning_rate": 1.0465346534653465e-05, "loss": 0.027, "step": 4003 }, { "epoch": 15.826086956521738, "grad_norm": 0.6435807347297668, "learning_rate": 1.0455445544554456e-05, "loss": 0.0261, "step": 4004 }, { "epoch": 15.8300395256917, "grad_norm": 0.3674606680870056, "learning_rate": 1.0445544554455445e-05, "loss": 0.0324, "step": 4005 }, { "epoch": 15.83399209486166, "grad_norm": 0.2787797152996063, "learning_rate": 1.0435643564356436e-05, "loss": 0.0185, "step": 4006 }, { "epoch": 15.837944664031621, "grad_norm": 0.30911925435066223, "learning_rate": 1.0425742574257426e-05, "loss": 0.0208, "step": 4007 }, { "epoch": 15.841897233201582, "grad_norm": 0.44155290722846985, "learning_rate": 1.0415841584158417e-05, "loss": 0.0211, "step": 4008 }, { "epoch": 15.845849802371541, "grad_norm": 0.3604203760623932, "learning_rate": 1.0405940594059406e-05, "loss": 0.0336, "step": 4009 }, { "epoch": 15.849802371541502, "grad_norm": 0.529350221157074, "learning_rate": 1.0396039603960395e-05, "loss": 0.0237, "step": 4010 }, { "epoch": 15.853754940711463, "grad_norm": 0.35327771306037903, "learning_rate": 1.0386138613861387e-05, "loss": 0.0146, "step": 4011 }, { "epoch": 15.857707509881424, "grad_norm": 0.24805350601673126, "learning_rate": 1.0376237623762376e-05, "loss": 0.0251, "step": 4012 }, { "epoch": 15.861660079051383, "grad_norm": 0.5510994791984558, "learning_rate": 1.0366336633663367e-05, "loss": 0.0181, "step": 4013 }, { "epoch": 15.865612648221344, "grad_norm": 0.409164696931839, "learning_rate": 1.0356435643564356e-05, "loss": 0.024, "step": 4014 }, { "epoch": 15.869565217391305, "grad_norm": 0.34517961740493774, "learning_rate": 1.0346534653465346e-05, "loss": 0.0208, "step": 4015 }, { "epoch": 15.873517786561266, "grad_norm": 0.6004961729049683, "learning_rate": 1.0336633663366337e-05, "loss": 0.0277, "step": 4016 }, { "epoch": 15.877470355731225, "grad_norm": 0.27147376537323, "learning_rate": 1.0326732673267326e-05, "loss": 0.0164, "step": 4017 }, { "epoch": 15.881422924901186, "grad_norm": 0.5839587450027466, "learning_rate": 1.0316831683168317e-05, "loss": 0.0225, "step": 4018 }, { "epoch": 15.885375494071146, "grad_norm": 0.44475889205932617, "learning_rate": 1.0306930693069307e-05, "loss": 0.0139, "step": 4019 }, { "epoch": 15.889328063241107, "grad_norm": 0.24266059696674347, "learning_rate": 1.0297029702970298e-05, "loss": 0.0212, "step": 4020 }, { "epoch": 15.893280632411066, "grad_norm": 0.15000653266906738, "learning_rate": 1.0287128712871287e-05, "loss": 0.0182, "step": 4021 }, { "epoch": 15.897233201581027, "grad_norm": 0.3818146586418152, "learning_rate": 1.0277227722772276e-05, "loss": 0.0173, "step": 4022 }, { "epoch": 15.901185770750988, "grad_norm": 0.4534763991832733, "learning_rate": 1.0267326732673267e-05, "loss": 0.0229, "step": 4023 }, { "epoch": 15.90513833992095, "grad_norm": 0.6671978831291199, "learning_rate": 1.0257425742574257e-05, "loss": 0.0331, "step": 4024 }, { "epoch": 15.909090909090908, "grad_norm": 0.4264109432697296, "learning_rate": 1.0247524752475248e-05, "loss": 0.0219, "step": 4025 }, { "epoch": 15.91304347826087, "grad_norm": 0.41731002926826477, "learning_rate": 1.0237623762376237e-05, "loss": 0.0195, "step": 4026 }, { "epoch": 15.91699604743083, "grad_norm": 0.29848235845565796, "learning_rate": 1.0227722772277228e-05, "loss": 0.0202, "step": 4027 }, { "epoch": 15.920948616600791, "grad_norm": 0.2848087549209595, "learning_rate": 1.0217821782178218e-05, "loss": 0.031, "step": 4028 }, { "epoch": 15.92490118577075, "grad_norm": 0.3016265332698822, "learning_rate": 1.0207920792079207e-05, "loss": 0.0204, "step": 4029 }, { "epoch": 15.928853754940711, "grad_norm": 0.3336348235607147, "learning_rate": 1.0198019801980198e-05, "loss": 0.0209, "step": 4030 }, { "epoch": 15.932806324110672, "grad_norm": 0.6364120244979858, "learning_rate": 1.0188118811881188e-05, "loss": 0.0251, "step": 4031 }, { "epoch": 15.936758893280633, "grad_norm": 0.3789774775505066, "learning_rate": 1.0178217821782179e-05, "loss": 0.015, "step": 4032 }, { "epoch": 15.940711462450594, "grad_norm": 0.7191634178161621, "learning_rate": 1.0168316831683168e-05, "loss": 0.0233, "step": 4033 }, { "epoch": 15.944664031620553, "grad_norm": 0.32997453212738037, "learning_rate": 1.0158415841584159e-05, "loss": 0.0243, "step": 4034 }, { "epoch": 15.948616600790514, "grad_norm": 0.2634662389755249, "learning_rate": 1.014851485148515e-05, "loss": 0.0178, "step": 4035 }, { "epoch": 15.952569169960475, "grad_norm": 0.5701932311058044, "learning_rate": 1.013861386138614e-05, "loss": 0.025, "step": 4036 }, { "epoch": 15.956521739130435, "grad_norm": 0.2460014373064041, "learning_rate": 1.012871287128713e-05, "loss": 0.0251, "step": 4037 }, { "epoch": 15.960474308300395, "grad_norm": 0.29873085021972656, "learning_rate": 1.011881188118812e-05, "loss": 0.0193, "step": 4038 }, { "epoch": 15.964426877470355, "grad_norm": 0.21316447854042053, "learning_rate": 1.010891089108911e-05, "loss": 0.0204, "step": 4039 }, { "epoch": 15.968379446640316, "grad_norm": 0.37378209829330444, "learning_rate": 1.00990099009901e-05, "loss": 0.0235, "step": 4040 }, { "epoch": 15.972332015810277, "grad_norm": 0.9243780374526978, "learning_rate": 1.008910891089109e-05, "loss": 0.0262, "step": 4041 }, { "epoch": 15.976284584980236, "grad_norm": 0.48010534048080444, "learning_rate": 1.007920792079208e-05, "loss": 0.0221, "step": 4042 }, { "epoch": 15.980237154150197, "grad_norm": 0.4258614182472229, "learning_rate": 1.006930693069307e-05, "loss": 0.0171, "step": 4043 }, { "epoch": 15.984189723320158, "grad_norm": 0.4991080164909363, "learning_rate": 1.0059405940594061e-05, "loss": 0.0213, "step": 4044 }, { "epoch": 15.988142292490119, "grad_norm": 0.37145715951919556, "learning_rate": 1.004950495049505e-05, "loss": 0.0266, "step": 4045 }, { "epoch": 15.992094861660078, "grad_norm": 0.5489411354064941, "learning_rate": 1.003960396039604e-05, "loss": 0.0253, "step": 4046 }, { "epoch": 15.996047430830039, "grad_norm": 0.3076215386390686, "learning_rate": 1.0029702970297031e-05, "loss": 0.0201, "step": 4047 }, { "epoch": 16.0, "grad_norm": 0.11576864868402481, "learning_rate": 1.001980198019802e-05, "loss": 0.0185, "step": 4048 }, { "epoch": 16.00395256916996, "grad_norm": 0.24986760318279266, "learning_rate": 1.0009900990099011e-05, "loss": 0.0171, "step": 4049 }, { "epoch": 16.007905138339922, "grad_norm": 0.08012142032384872, "learning_rate": 1e-05, "loss": 0.0147, "step": 4050 }, { "epoch": 16.011857707509883, "grad_norm": 0.23237846791744232, "learning_rate": 9.990099009900992e-06, "loss": 0.0294, "step": 4051 }, { "epoch": 16.015810276679844, "grad_norm": 0.3199997544288635, "learning_rate": 9.980198019801981e-06, "loss": 0.0141, "step": 4052 }, { "epoch": 16.0197628458498, "grad_norm": 0.07866759598255157, "learning_rate": 9.97029702970297e-06, "loss": 0.0161, "step": 4053 }, { "epoch": 16.023715415019762, "grad_norm": 0.07244686782360077, "learning_rate": 9.960396039603962e-06, "loss": 0.0138, "step": 4054 }, { "epoch": 16.027667984189723, "grad_norm": 0.08883621543645859, "learning_rate": 9.950495049504951e-06, "loss": 0.0152, "step": 4055 }, { "epoch": 16.031620553359684, "grad_norm": 0.08542703837156296, "learning_rate": 9.940594059405942e-06, "loss": 0.0173, "step": 4056 }, { "epoch": 16.035573122529645, "grad_norm": 0.15095359086990356, "learning_rate": 9.930693069306932e-06, "loss": 0.0275, "step": 4057 }, { "epoch": 16.039525691699605, "grad_norm": 0.1130053773522377, "learning_rate": 9.920792079207923e-06, "loss": 0.0223, "step": 4058 }, { "epoch": 16.043478260869566, "grad_norm": 0.09003191441297531, "learning_rate": 9.910891089108912e-06, "loss": 0.0182, "step": 4059 }, { "epoch": 16.047430830039527, "grad_norm": 0.06832607090473175, "learning_rate": 9.900990099009901e-06, "loss": 0.0112, "step": 4060 }, { "epoch": 16.051383399209485, "grad_norm": 0.16024409234523773, "learning_rate": 9.891089108910892e-06, "loss": 0.0289, "step": 4061 }, { "epoch": 16.055335968379445, "grad_norm": 0.25271543860435486, "learning_rate": 9.881188118811882e-06, "loss": 0.0207, "step": 4062 }, { "epoch": 16.059288537549406, "grad_norm": 0.14919935166835785, "learning_rate": 9.871287128712873e-06, "loss": 0.0152, "step": 4063 }, { "epoch": 16.063241106719367, "grad_norm": 0.17505007982254028, "learning_rate": 9.861386138613862e-06, "loss": 0.0251, "step": 4064 }, { "epoch": 16.067193675889328, "grad_norm": 0.11101784557104111, "learning_rate": 9.851485148514852e-06, "loss": 0.0204, "step": 4065 }, { "epoch": 16.07114624505929, "grad_norm": 0.7953076958656311, "learning_rate": 9.841584158415843e-06, "loss": 0.0299, "step": 4066 }, { "epoch": 16.07509881422925, "grad_norm": 0.23288604617118835, "learning_rate": 9.831683168316832e-06, "loss": 0.0137, "step": 4067 }, { "epoch": 16.07905138339921, "grad_norm": 0.16941533982753754, "learning_rate": 9.821782178217823e-06, "loss": 0.0235, "step": 4068 }, { "epoch": 16.08300395256917, "grad_norm": 0.1363600343465805, "learning_rate": 9.811881188118812e-06, "loss": 0.0265, "step": 4069 }, { "epoch": 16.08695652173913, "grad_norm": 0.24240438640117645, "learning_rate": 9.801980198019804e-06, "loss": 0.0174, "step": 4070 }, { "epoch": 16.09090909090909, "grad_norm": 0.1265329122543335, "learning_rate": 9.792079207920793e-06, "loss": 0.0213, "step": 4071 }, { "epoch": 16.09486166007905, "grad_norm": 0.5480315685272217, "learning_rate": 9.782178217821782e-06, "loss": 0.0366, "step": 4072 }, { "epoch": 16.09881422924901, "grad_norm": 0.2371923178434372, "learning_rate": 9.772277227722773e-06, "loss": 0.0165, "step": 4073 }, { "epoch": 16.102766798418973, "grad_norm": 0.15473449230194092, "learning_rate": 9.762376237623763e-06, "loss": 0.0303, "step": 4074 }, { "epoch": 16.106719367588934, "grad_norm": 0.2704905867576599, "learning_rate": 9.752475247524754e-06, "loss": 0.0282, "step": 4075 }, { "epoch": 16.110671936758894, "grad_norm": 0.17479056119918823, "learning_rate": 9.742574257425743e-06, "loss": 0.0135, "step": 4076 }, { "epoch": 16.114624505928855, "grad_norm": 0.1033933237195015, "learning_rate": 9.732673267326734e-06, "loss": 0.0212, "step": 4077 }, { "epoch": 16.118577075098813, "grad_norm": 0.14604951441287994, "learning_rate": 9.722772277227724e-06, "loss": 0.0212, "step": 4078 }, { "epoch": 16.122529644268774, "grad_norm": 0.6081244349479675, "learning_rate": 9.712871287128713e-06, "loss": 0.0222, "step": 4079 }, { "epoch": 16.126482213438734, "grad_norm": 0.38193362951278687, "learning_rate": 9.702970297029704e-06, "loss": 0.0187, "step": 4080 }, { "epoch": 16.130434782608695, "grad_norm": 0.26413199305534363, "learning_rate": 9.693069306930693e-06, "loss": 0.0143, "step": 4081 }, { "epoch": 16.134387351778656, "grad_norm": 0.18585412204265594, "learning_rate": 9.683168316831684e-06, "loss": 0.0189, "step": 4082 }, { "epoch": 16.138339920948617, "grad_norm": 0.411183625459671, "learning_rate": 9.673267326732674e-06, "loss": 0.0193, "step": 4083 }, { "epoch": 16.142292490118578, "grad_norm": 0.13958127796649933, "learning_rate": 9.663366336633665e-06, "loss": 0.0143, "step": 4084 }, { "epoch": 16.14624505928854, "grad_norm": 0.3946831524372101, "learning_rate": 9.653465346534654e-06, "loss": 0.0189, "step": 4085 }, { "epoch": 16.150197628458496, "grad_norm": 0.42950373888015747, "learning_rate": 9.643564356435644e-06, "loss": 0.0229, "step": 4086 }, { "epoch": 16.154150197628457, "grad_norm": 0.14441066980361938, "learning_rate": 9.633663366336635e-06, "loss": 0.0192, "step": 4087 }, { "epoch": 16.158102766798418, "grad_norm": 0.26897329092025757, "learning_rate": 9.623762376237624e-06, "loss": 0.0164, "step": 4088 }, { "epoch": 16.16205533596838, "grad_norm": 0.3785199522972107, "learning_rate": 9.613861386138615e-06, "loss": 0.0258, "step": 4089 }, { "epoch": 16.16600790513834, "grad_norm": 0.30758941173553467, "learning_rate": 9.603960396039604e-06, "loss": 0.0184, "step": 4090 }, { "epoch": 16.1699604743083, "grad_norm": 0.2544354796409607, "learning_rate": 9.594059405940596e-06, "loss": 0.0156, "step": 4091 }, { "epoch": 16.17391304347826, "grad_norm": 0.21463795006275177, "learning_rate": 9.584158415841585e-06, "loss": 0.0271, "step": 4092 }, { "epoch": 16.177865612648223, "grad_norm": 0.08555018156766891, "learning_rate": 9.574257425742574e-06, "loss": 0.0152, "step": 4093 }, { "epoch": 16.181818181818183, "grad_norm": 0.39186304807662964, "learning_rate": 9.564356435643565e-06, "loss": 0.0208, "step": 4094 }, { "epoch": 16.18577075098814, "grad_norm": 0.1238565593957901, "learning_rate": 9.554455445544555e-06, "loss": 0.024, "step": 4095 }, { "epoch": 16.1897233201581, "grad_norm": 0.11853600293397903, "learning_rate": 9.544554455445546e-06, "loss": 0.0169, "step": 4096 }, { "epoch": 16.193675889328063, "grad_norm": 0.12091764807701111, "learning_rate": 9.534653465346535e-06, "loss": 0.0179, "step": 4097 }, { "epoch": 16.197628458498023, "grad_norm": 0.39361968636512756, "learning_rate": 9.524752475247525e-06, "loss": 0.0286, "step": 4098 }, { "epoch": 16.201581027667984, "grad_norm": 0.4037843942642212, "learning_rate": 9.514851485148516e-06, "loss": 0.019, "step": 4099 }, { "epoch": 16.205533596837945, "grad_norm": 0.256077378988266, "learning_rate": 9.504950495049505e-06, "loss": 0.0528, "step": 4100 }, { "epoch": 16.209486166007906, "grad_norm": 0.21473026275634766, "learning_rate": 9.495049504950496e-06, "loss": 0.0249, "step": 4101 }, { "epoch": 16.213438735177867, "grad_norm": 0.28430071473121643, "learning_rate": 9.485148514851485e-06, "loss": 0.0194, "step": 4102 }, { "epoch": 16.217391304347824, "grad_norm": 0.18489959836006165, "learning_rate": 9.475247524752476e-06, "loss": 0.013, "step": 4103 }, { "epoch": 16.221343873517785, "grad_norm": 0.13621190190315247, "learning_rate": 9.465346534653466e-06, "loss": 0.0253, "step": 4104 }, { "epoch": 16.225296442687746, "grad_norm": 0.12427568435668945, "learning_rate": 9.455445544554455e-06, "loss": 0.0294, "step": 4105 }, { "epoch": 16.229249011857707, "grad_norm": 0.15711766481399536, "learning_rate": 9.445544554455446e-06, "loss": 0.0299, "step": 4106 }, { "epoch": 16.233201581027668, "grad_norm": 0.3726131021976471, "learning_rate": 9.435643564356436e-06, "loss": 0.0204, "step": 4107 }, { "epoch": 16.23715415019763, "grad_norm": 0.0810607373714447, "learning_rate": 9.425742574257427e-06, "loss": 0.0159, "step": 4108 }, { "epoch": 16.24110671936759, "grad_norm": 0.09546815603971481, "learning_rate": 9.415841584158416e-06, "loss": 0.0183, "step": 4109 }, { "epoch": 16.24505928853755, "grad_norm": 0.21881286799907684, "learning_rate": 9.405940594059407e-06, "loss": 0.0218, "step": 4110 }, { "epoch": 16.24901185770751, "grad_norm": 0.1404864341020584, "learning_rate": 9.396039603960397e-06, "loss": 0.013, "step": 4111 }, { "epoch": 16.25296442687747, "grad_norm": 0.11336731910705566, "learning_rate": 9.386138613861386e-06, "loss": 0.0173, "step": 4112 }, { "epoch": 16.25691699604743, "grad_norm": 0.1439458727836609, "learning_rate": 9.376237623762377e-06, "loss": 0.0214, "step": 4113 }, { "epoch": 16.26086956521739, "grad_norm": 0.1796472668647766, "learning_rate": 9.366336633663366e-06, "loss": 0.0157, "step": 4114 }, { "epoch": 16.26482213438735, "grad_norm": 0.1071930080652237, "learning_rate": 9.356435643564357e-06, "loss": 0.0233, "step": 4115 }, { "epoch": 16.268774703557312, "grad_norm": 0.2963261008262634, "learning_rate": 9.346534653465347e-06, "loss": 0.025, "step": 4116 }, { "epoch": 16.272727272727273, "grad_norm": 0.7227439880371094, "learning_rate": 9.336633663366338e-06, "loss": 0.0211, "step": 4117 }, { "epoch": 16.276679841897234, "grad_norm": 0.7633346319198608, "learning_rate": 9.326732673267327e-06, "loss": 0.0402, "step": 4118 }, { "epoch": 16.280632411067195, "grad_norm": 0.5752243995666504, "learning_rate": 9.316831683168317e-06, "loss": 0.0229, "step": 4119 }, { "epoch": 16.284584980237153, "grad_norm": 0.4336450695991516, "learning_rate": 9.306930693069308e-06, "loss": 0.0206, "step": 4120 }, { "epoch": 16.288537549407113, "grad_norm": 0.30222442746162415, "learning_rate": 9.297029702970297e-06, "loss": 0.0247, "step": 4121 }, { "epoch": 16.292490118577074, "grad_norm": 0.11855607479810715, "learning_rate": 9.287128712871288e-06, "loss": 0.0232, "step": 4122 }, { "epoch": 16.296442687747035, "grad_norm": 0.10608139634132385, "learning_rate": 9.277227722772277e-06, "loss": 0.0148, "step": 4123 }, { "epoch": 16.300395256916996, "grad_norm": 0.2788778245449066, "learning_rate": 9.267326732673267e-06, "loss": 0.0156, "step": 4124 }, { "epoch": 16.304347826086957, "grad_norm": 0.44659215211868286, "learning_rate": 9.257425742574258e-06, "loss": 0.0356, "step": 4125 }, { "epoch": 16.308300395256918, "grad_norm": 0.573864758014679, "learning_rate": 9.247524752475247e-06, "loss": 0.0207, "step": 4126 }, { "epoch": 16.31225296442688, "grad_norm": 0.2804989516735077, "learning_rate": 9.237623762376238e-06, "loss": 0.014, "step": 4127 }, { "epoch": 16.31620553359684, "grad_norm": 0.7048799991607666, "learning_rate": 9.227722772277228e-06, "loss": 0.0415, "step": 4128 }, { "epoch": 16.320158102766797, "grad_norm": 0.10146640241146088, "learning_rate": 9.217821782178219e-06, "loss": 0.0196, "step": 4129 }, { "epoch": 16.324110671936758, "grad_norm": 0.30598604679107666, "learning_rate": 9.207920792079208e-06, "loss": 0.0328, "step": 4130 }, { "epoch": 16.32806324110672, "grad_norm": 0.17372798919677734, "learning_rate": 9.198019801980198e-06, "loss": 0.0345, "step": 4131 }, { "epoch": 16.33201581027668, "grad_norm": 0.13411682844161987, "learning_rate": 9.188118811881189e-06, "loss": 0.0201, "step": 4132 }, { "epoch": 16.33596837944664, "grad_norm": 0.27506691217422485, "learning_rate": 9.178217821782178e-06, "loss": 0.0145, "step": 4133 }, { "epoch": 16.3399209486166, "grad_norm": 0.211775004863739, "learning_rate": 9.168316831683169e-06, "loss": 0.0146, "step": 4134 }, { "epoch": 16.343873517786562, "grad_norm": 0.1287011057138443, "learning_rate": 9.158415841584158e-06, "loss": 0.0239, "step": 4135 }, { "epoch": 16.347826086956523, "grad_norm": 0.31582334637641907, "learning_rate": 9.14851485148515e-06, "loss": 0.0197, "step": 4136 }, { "epoch": 16.35177865612648, "grad_norm": 0.1865476369857788, "learning_rate": 9.138613861386139e-06, "loss": 0.0274, "step": 4137 }, { "epoch": 16.35573122529644, "grad_norm": 0.0841803103685379, "learning_rate": 9.128712871287128e-06, "loss": 0.0175, "step": 4138 }, { "epoch": 16.359683794466402, "grad_norm": 0.32546985149383545, "learning_rate": 9.11881188118812e-06, "loss": 0.0202, "step": 4139 }, { "epoch": 16.363636363636363, "grad_norm": 0.5931854844093323, "learning_rate": 9.108910891089109e-06, "loss": 0.0272, "step": 4140 }, { "epoch": 16.367588932806324, "grad_norm": 0.5117402672767639, "learning_rate": 9.0990099009901e-06, "loss": 0.0163, "step": 4141 }, { "epoch": 16.371541501976285, "grad_norm": 0.4330423176288605, "learning_rate": 9.089108910891089e-06, "loss": 0.0199, "step": 4142 }, { "epoch": 16.375494071146246, "grad_norm": 0.18616345524787903, "learning_rate": 9.07920792079208e-06, "loss": 0.0138, "step": 4143 }, { "epoch": 16.379446640316207, "grad_norm": 0.09388396888971329, "learning_rate": 9.06930693069307e-06, "loss": 0.0182, "step": 4144 }, { "epoch": 16.383399209486164, "grad_norm": 0.1572658270597458, "learning_rate": 9.059405940594059e-06, "loss": 0.0326, "step": 4145 }, { "epoch": 16.387351778656125, "grad_norm": 0.15576355159282684, "learning_rate": 9.04950495049505e-06, "loss": 0.0167, "step": 4146 }, { "epoch": 16.391304347826086, "grad_norm": 0.1984003186225891, "learning_rate": 9.03960396039604e-06, "loss": 0.0226, "step": 4147 }, { "epoch": 16.395256916996047, "grad_norm": 5.429525852203369, "learning_rate": 9.02970297029703e-06, "loss": 0.0236, "step": 4148 }, { "epoch": 16.399209486166008, "grad_norm": 0.3921261727809906, "learning_rate": 9.01980198019802e-06, "loss": 0.0214, "step": 4149 }, { "epoch": 16.40316205533597, "grad_norm": 0.1234780102968216, "learning_rate": 9.00990099009901e-06, "loss": 0.0185, "step": 4150 }, { "epoch": 16.40711462450593, "grad_norm": 0.4870821535587311, "learning_rate": 9e-06, "loss": 0.0226, "step": 4151 }, { "epoch": 16.41106719367589, "grad_norm": 0.4358524680137634, "learning_rate": 8.99009900990099e-06, "loss": 0.0211, "step": 4152 }, { "epoch": 16.41501976284585, "grad_norm": 0.26159220933914185, "learning_rate": 8.98019801980198e-06, "loss": 0.0099, "step": 4153 }, { "epoch": 16.41897233201581, "grad_norm": 0.14635029435157776, "learning_rate": 8.97029702970297e-06, "loss": 0.0274, "step": 4154 }, { "epoch": 16.42292490118577, "grad_norm": 0.13570208847522736, "learning_rate": 8.960396039603961e-06, "loss": 0.0249, "step": 4155 }, { "epoch": 16.42687747035573, "grad_norm": 0.13498082756996155, "learning_rate": 8.95049504950495e-06, "loss": 0.0203, "step": 4156 }, { "epoch": 16.43083003952569, "grad_norm": 0.23252758383750916, "learning_rate": 8.94059405940594e-06, "loss": 0.0158, "step": 4157 }, { "epoch": 16.434782608695652, "grad_norm": 0.18174104392528534, "learning_rate": 8.930693069306931e-06, "loss": 0.0163, "step": 4158 }, { "epoch": 16.438735177865613, "grad_norm": 0.09684844315052032, "learning_rate": 8.92079207920792e-06, "loss": 0.0189, "step": 4159 }, { "epoch": 16.442687747035574, "grad_norm": 0.4668383300304413, "learning_rate": 8.910891089108911e-06, "loss": 0.0193, "step": 4160 }, { "epoch": 16.446640316205535, "grad_norm": 0.09127598255872726, "learning_rate": 8.9009900990099e-06, "loss": 0.0172, "step": 4161 }, { "epoch": 16.450592885375492, "grad_norm": 0.09435838460922241, "learning_rate": 8.891089108910892e-06, "loss": 0.0199, "step": 4162 }, { "epoch": 16.454545454545453, "grad_norm": 0.07973987609148026, "learning_rate": 8.881188118811881e-06, "loss": 0.0153, "step": 4163 }, { "epoch": 16.458498023715414, "grad_norm": 0.2688567340373993, "learning_rate": 8.87128712871287e-06, "loss": 0.0199, "step": 4164 }, { "epoch": 16.462450592885375, "grad_norm": 0.22572244703769684, "learning_rate": 8.861386138613862e-06, "loss": 0.0155, "step": 4165 }, { "epoch": 16.466403162055336, "grad_norm": 0.1056339219212532, "learning_rate": 8.851485148514851e-06, "loss": 0.0106, "step": 4166 }, { "epoch": 16.470355731225297, "grad_norm": 0.39108726382255554, "learning_rate": 8.841584158415842e-06, "loss": 0.0214, "step": 4167 }, { "epoch": 16.474308300395258, "grad_norm": 0.09038659185171127, "learning_rate": 8.831683168316831e-06, "loss": 0.0182, "step": 4168 }, { "epoch": 16.47826086956522, "grad_norm": 0.16739755868911743, "learning_rate": 8.821782178217822e-06, "loss": 0.0178, "step": 4169 }, { "epoch": 16.48221343873518, "grad_norm": 0.2831993103027344, "learning_rate": 8.811881188118812e-06, "loss": 0.017, "step": 4170 }, { "epoch": 16.486166007905137, "grad_norm": 0.3935111165046692, "learning_rate": 8.801980198019801e-06, "loss": 0.0335, "step": 4171 }, { "epoch": 16.490118577075098, "grad_norm": 0.35388898849487305, "learning_rate": 8.792079207920792e-06, "loss": 0.0227, "step": 4172 }, { "epoch": 16.49407114624506, "grad_norm": 0.19930867850780487, "learning_rate": 8.782178217821782e-06, "loss": 0.0165, "step": 4173 }, { "epoch": 16.49802371541502, "grad_norm": 0.43218377232551575, "learning_rate": 8.772277227722773e-06, "loss": 0.0193, "step": 4174 }, { "epoch": 16.50197628458498, "grad_norm": 0.12009885907173157, "learning_rate": 8.762376237623762e-06, "loss": 0.0216, "step": 4175 }, { "epoch": 16.50592885375494, "grad_norm": 0.14550235867500305, "learning_rate": 8.752475247524753e-06, "loss": 0.0125, "step": 4176 }, { "epoch": 16.509881422924902, "grad_norm": 0.14113189280033112, "learning_rate": 8.742574257425743e-06, "loss": 0.0242, "step": 4177 }, { "epoch": 16.513833992094863, "grad_norm": 0.4160834848880768, "learning_rate": 8.732673267326732e-06, "loss": 0.0228, "step": 4178 }, { "epoch": 16.51778656126482, "grad_norm": 0.28901275992393494, "learning_rate": 8.722772277227723e-06, "loss": 0.0135, "step": 4179 }, { "epoch": 16.52173913043478, "grad_norm": 0.23270180821418762, "learning_rate": 8.712871287128712e-06, "loss": 0.0502, "step": 4180 }, { "epoch": 16.525691699604742, "grad_norm": 0.2315024584531784, "learning_rate": 8.702970297029703e-06, "loss": 0.0174, "step": 4181 }, { "epoch": 16.529644268774703, "grad_norm": 0.2065029740333557, "learning_rate": 8.693069306930693e-06, "loss": 0.0283, "step": 4182 }, { "epoch": 16.533596837944664, "grad_norm": 0.24288985133171082, "learning_rate": 8.683168316831682e-06, "loss": 0.0279, "step": 4183 }, { "epoch": 16.537549407114625, "grad_norm": 0.12854309380054474, "learning_rate": 8.673267326732673e-06, "loss": 0.0272, "step": 4184 }, { "epoch": 16.541501976284586, "grad_norm": 0.33031734824180603, "learning_rate": 8.663366336633663e-06, "loss": 0.0228, "step": 4185 }, { "epoch": 16.545454545454547, "grad_norm": 0.08728615194559097, "learning_rate": 8.653465346534654e-06, "loss": 0.0158, "step": 4186 }, { "epoch": 16.549407114624508, "grad_norm": 0.1724659651517868, "learning_rate": 8.643564356435643e-06, "loss": 0.0163, "step": 4187 }, { "epoch": 16.553359683794465, "grad_norm": 0.14346711337566376, "learning_rate": 8.633663366336634e-06, "loss": 0.0131, "step": 4188 }, { "epoch": 16.557312252964426, "grad_norm": 0.641791820526123, "learning_rate": 8.623762376237623e-06, "loss": 0.0363, "step": 4189 }, { "epoch": 16.561264822134387, "grad_norm": 0.23550304770469666, "learning_rate": 8.613861386138613e-06, "loss": 0.0325, "step": 4190 }, { "epoch": 16.565217391304348, "grad_norm": 0.3594205379486084, "learning_rate": 8.603960396039604e-06, "loss": 0.0329, "step": 4191 }, { "epoch": 16.56916996047431, "grad_norm": 0.23732666671276093, "learning_rate": 8.594059405940593e-06, "loss": 0.0254, "step": 4192 }, { "epoch": 16.57312252964427, "grad_norm": 0.09958739578723907, "learning_rate": 8.584158415841586e-06, "loss": 0.0208, "step": 4193 }, { "epoch": 16.57707509881423, "grad_norm": 0.28758832812309265, "learning_rate": 8.574257425742575e-06, "loss": 0.0204, "step": 4194 }, { "epoch": 16.58102766798419, "grad_norm": 0.14528107643127441, "learning_rate": 8.564356435643565e-06, "loss": 0.0316, "step": 4195 }, { "epoch": 16.58498023715415, "grad_norm": 0.11189550161361694, "learning_rate": 8.554455445544556e-06, "loss": 0.0203, "step": 4196 }, { "epoch": 16.58893280632411, "grad_norm": 0.19660070538520813, "learning_rate": 8.544554455445545e-06, "loss": 0.0174, "step": 4197 }, { "epoch": 16.59288537549407, "grad_norm": 0.3226984143257141, "learning_rate": 8.534653465346536e-06, "loss": 0.024, "step": 4198 }, { "epoch": 16.59683794466403, "grad_norm": 0.20195060968399048, "learning_rate": 8.524752475247526e-06, "loss": 0.0107, "step": 4199 }, { "epoch": 16.600790513833992, "grad_norm": 0.48350876569747925, "learning_rate": 8.514851485148517e-06, "loss": 0.0202, "step": 4200 }, { "epoch": 16.604743083003953, "grad_norm": 0.46012556552886963, "learning_rate": 8.504950495049506e-06, "loss": 0.0219, "step": 4201 }, { "epoch": 16.608695652173914, "grad_norm": 0.12655732035636902, "learning_rate": 8.495049504950495e-06, "loss": 0.0264, "step": 4202 }, { "epoch": 16.612648221343875, "grad_norm": 0.24045343697071075, "learning_rate": 8.485148514851487e-06, "loss": 0.0221, "step": 4203 }, { "epoch": 16.616600790513836, "grad_norm": 0.6192004084587097, "learning_rate": 8.475247524752476e-06, "loss": 0.0264, "step": 4204 }, { "epoch": 16.620553359683793, "grad_norm": 0.5062432289123535, "learning_rate": 8.465346534653467e-06, "loss": 0.0228, "step": 4205 }, { "epoch": 16.624505928853754, "grad_norm": 0.342330664396286, "learning_rate": 8.455445544554456e-06, "loss": 0.0273, "step": 4206 }, { "epoch": 16.628458498023715, "grad_norm": 0.1568595916032791, "learning_rate": 8.445544554455446e-06, "loss": 0.0161, "step": 4207 }, { "epoch": 16.632411067193676, "grad_norm": 0.2172902673482895, "learning_rate": 8.435643564356437e-06, "loss": 0.0235, "step": 4208 }, { "epoch": 16.636363636363637, "grad_norm": 0.7940640449523926, "learning_rate": 8.425742574257426e-06, "loss": 0.0317, "step": 4209 }, { "epoch": 16.640316205533598, "grad_norm": 0.10678622126579285, "learning_rate": 8.415841584158417e-06, "loss": 0.0202, "step": 4210 }, { "epoch": 16.64426877470356, "grad_norm": 0.3412444591522217, "learning_rate": 8.405940594059407e-06, "loss": 0.0174, "step": 4211 }, { "epoch": 16.64822134387352, "grad_norm": 0.0863598957657814, "learning_rate": 8.396039603960398e-06, "loss": 0.0165, "step": 4212 }, { "epoch": 16.652173913043477, "grad_norm": 0.3308269679546356, "learning_rate": 8.386138613861387e-06, "loss": 0.0173, "step": 4213 }, { "epoch": 16.656126482213438, "grad_norm": 0.15239432454109192, "learning_rate": 8.376237623762376e-06, "loss": 0.0164, "step": 4214 }, { "epoch": 16.6600790513834, "grad_norm": 0.14283309876918793, "learning_rate": 8.366336633663367e-06, "loss": 0.0207, "step": 4215 }, { "epoch": 16.66403162055336, "grad_norm": 0.13322024047374725, "learning_rate": 8.356435643564357e-06, "loss": 0.0175, "step": 4216 }, { "epoch": 16.66798418972332, "grad_norm": 0.29618045687675476, "learning_rate": 8.346534653465348e-06, "loss": 0.0182, "step": 4217 }, { "epoch": 16.67193675889328, "grad_norm": 0.3982749581336975, "learning_rate": 8.336633663366337e-06, "loss": 0.0194, "step": 4218 }, { "epoch": 16.675889328063242, "grad_norm": 0.1627119481563568, "learning_rate": 8.326732673267328e-06, "loss": 0.0277, "step": 4219 }, { "epoch": 16.679841897233203, "grad_norm": 0.3073245882987976, "learning_rate": 8.316831683168318e-06, "loss": 0.0212, "step": 4220 }, { "epoch": 16.683794466403164, "grad_norm": 0.21005167067050934, "learning_rate": 8.306930693069307e-06, "loss": 0.0124, "step": 4221 }, { "epoch": 16.68774703557312, "grad_norm": 0.2235172986984253, "learning_rate": 8.297029702970298e-06, "loss": 0.0248, "step": 4222 }, { "epoch": 16.691699604743082, "grad_norm": 0.17963793873786926, "learning_rate": 8.287128712871288e-06, "loss": 0.0356, "step": 4223 }, { "epoch": 16.695652173913043, "grad_norm": 0.13324092328548431, "learning_rate": 8.277227722772279e-06, "loss": 0.0138, "step": 4224 }, { "epoch": 16.699604743083004, "grad_norm": 0.2641155421733856, "learning_rate": 8.267326732673268e-06, "loss": 0.0154, "step": 4225 }, { "epoch": 16.703557312252965, "grad_norm": 0.1003686934709549, "learning_rate": 8.257425742574259e-06, "loss": 0.0166, "step": 4226 }, { "epoch": 16.707509881422926, "grad_norm": 0.22244547307491302, "learning_rate": 8.247524752475248e-06, "loss": 0.021, "step": 4227 }, { "epoch": 16.711462450592887, "grad_norm": 0.4314354658126831, "learning_rate": 8.237623762376238e-06, "loss": 0.0294, "step": 4228 }, { "epoch": 16.715415019762847, "grad_norm": 0.1638021171092987, "learning_rate": 8.227722772277229e-06, "loss": 0.0106, "step": 4229 }, { "epoch": 16.719367588932805, "grad_norm": 0.09067068994045258, "learning_rate": 8.217821782178218e-06, "loss": 0.0182, "step": 4230 }, { "epoch": 16.723320158102766, "grad_norm": 0.14976957440376282, "learning_rate": 8.20792079207921e-06, "loss": 0.0116, "step": 4231 }, { "epoch": 16.727272727272727, "grad_norm": 0.5319260358810425, "learning_rate": 8.198019801980199e-06, "loss": 0.026, "step": 4232 }, { "epoch": 16.731225296442688, "grad_norm": 0.10501709580421448, "learning_rate": 8.188118811881188e-06, "loss": 0.0178, "step": 4233 }, { "epoch": 16.73517786561265, "grad_norm": 0.5431573987007141, "learning_rate": 8.178217821782179e-06, "loss": 0.0269, "step": 4234 }, { "epoch": 16.73913043478261, "grad_norm": 0.23208168148994446, "learning_rate": 8.168316831683168e-06, "loss": 0.0232, "step": 4235 }, { "epoch": 16.74308300395257, "grad_norm": 0.1094033494591713, "learning_rate": 8.15841584158416e-06, "loss": 0.0241, "step": 4236 }, { "epoch": 16.74703557312253, "grad_norm": 0.10016578435897827, "learning_rate": 8.148514851485149e-06, "loss": 0.0202, "step": 4237 }, { "epoch": 16.75098814229249, "grad_norm": 0.2318798303604126, "learning_rate": 8.13861386138614e-06, "loss": 0.0227, "step": 4238 }, { "epoch": 16.75494071146245, "grad_norm": 0.1561136543750763, "learning_rate": 8.12871287128713e-06, "loss": 0.022, "step": 4239 }, { "epoch": 16.75889328063241, "grad_norm": 0.2699989974498749, "learning_rate": 8.118811881188119e-06, "loss": 0.0144, "step": 4240 }, { "epoch": 16.76284584980237, "grad_norm": 0.13087517023086548, "learning_rate": 8.10891089108911e-06, "loss": 0.0268, "step": 4241 }, { "epoch": 16.766798418972332, "grad_norm": 0.2388116717338562, "learning_rate": 8.099009900990099e-06, "loss": 0.0134, "step": 4242 }, { "epoch": 16.770750988142293, "grad_norm": 0.6969513893127441, "learning_rate": 8.08910891089109e-06, "loss": 0.0244, "step": 4243 }, { "epoch": 16.774703557312254, "grad_norm": 0.6552041172981262, "learning_rate": 8.07920792079208e-06, "loss": 0.0324, "step": 4244 }, { "epoch": 16.778656126482215, "grad_norm": 0.7132962346076965, "learning_rate": 8.06930693069307e-06, "loss": 0.0219, "step": 4245 }, { "epoch": 16.782608695652176, "grad_norm": 0.14074979722499847, "learning_rate": 8.05940594059406e-06, "loss": 0.0306, "step": 4246 }, { "epoch": 16.786561264822133, "grad_norm": 0.14661741256713867, "learning_rate": 8.04950495049505e-06, "loss": 0.0357, "step": 4247 }, { "epoch": 16.790513833992094, "grad_norm": 0.3361961841583252, "learning_rate": 8.03960396039604e-06, "loss": 0.0171, "step": 4248 }, { "epoch": 16.794466403162055, "grad_norm": 0.15100903809070587, "learning_rate": 8.02970297029703e-06, "loss": 0.0274, "step": 4249 }, { "epoch": 16.798418972332016, "grad_norm": 0.229981929063797, "learning_rate": 8.019801980198021e-06, "loss": 0.0211, "step": 4250 }, { "epoch": 16.802371541501977, "grad_norm": 0.2952955365180969, "learning_rate": 8.00990099009901e-06, "loss": 0.0355, "step": 4251 }, { "epoch": 16.806324110671937, "grad_norm": 0.07801418751478195, "learning_rate": 8.000000000000001e-06, "loss": 0.0169, "step": 4252 }, { "epoch": 16.8102766798419, "grad_norm": 0.12060002982616425, "learning_rate": 7.99009900990099e-06, "loss": 0.0156, "step": 4253 }, { "epoch": 16.81422924901186, "grad_norm": 0.15552133321762085, "learning_rate": 7.98019801980198e-06, "loss": 0.0271, "step": 4254 }, { "epoch": 16.818181818181817, "grad_norm": 0.15040303766727448, "learning_rate": 7.970297029702971e-06, "loss": 0.0273, "step": 4255 }, { "epoch": 16.822134387351777, "grad_norm": 0.7464338541030884, "learning_rate": 7.96039603960396e-06, "loss": 0.0233, "step": 4256 }, { "epoch": 16.82608695652174, "grad_norm": 0.15363778173923492, "learning_rate": 7.950495049504952e-06, "loss": 0.0248, "step": 4257 }, { "epoch": 16.8300395256917, "grad_norm": 0.28624433279037476, "learning_rate": 7.940594059405941e-06, "loss": 0.0136, "step": 4258 }, { "epoch": 16.83399209486166, "grad_norm": 0.08501491695642471, "learning_rate": 7.930693069306932e-06, "loss": 0.0172, "step": 4259 }, { "epoch": 16.83794466403162, "grad_norm": 0.11711373925209045, "learning_rate": 7.920792079207921e-06, "loss": 0.0237, "step": 4260 }, { "epoch": 16.841897233201582, "grad_norm": 0.08293797075748444, "learning_rate": 7.91089108910891e-06, "loss": 0.0175, "step": 4261 }, { "epoch": 16.845849802371543, "grad_norm": 0.3010933995246887, "learning_rate": 7.900990099009902e-06, "loss": 0.0192, "step": 4262 }, { "epoch": 16.8498023715415, "grad_norm": 0.2796427309513092, "learning_rate": 7.891089108910891e-06, "loss": 0.018, "step": 4263 }, { "epoch": 16.85375494071146, "grad_norm": 0.13663852214813232, "learning_rate": 7.881188118811882e-06, "loss": 0.0243, "step": 4264 }, { "epoch": 16.857707509881422, "grad_norm": 0.18883472681045532, "learning_rate": 7.871287128712872e-06, "loss": 0.0264, "step": 4265 }, { "epoch": 16.861660079051383, "grad_norm": 0.8831784129142761, "learning_rate": 7.861386138613861e-06, "loss": 0.0436, "step": 4266 }, { "epoch": 16.865612648221344, "grad_norm": 0.09422732144594193, "learning_rate": 7.851485148514852e-06, "loss": 0.0197, "step": 4267 }, { "epoch": 16.869565217391305, "grad_norm": 0.382027804851532, "learning_rate": 7.841584158415841e-06, "loss": 0.0212, "step": 4268 }, { "epoch": 16.873517786561266, "grad_norm": 0.10886300355195999, "learning_rate": 7.831683168316833e-06, "loss": 0.022, "step": 4269 }, { "epoch": 16.877470355731226, "grad_norm": 0.3368642032146454, "learning_rate": 7.821782178217822e-06, "loss": 0.0156, "step": 4270 }, { "epoch": 16.881422924901187, "grad_norm": 0.2875737249851227, "learning_rate": 7.811881188118813e-06, "loss": 0.0346, "step": 4271 }, { "epoch": 16.885375494071145, "grad_norm": 1.1082501411437988, "learning_rate": 7.801980198019802e-06, "loss": 0.0294, "step": 4272 }, { "epoch": 16.889328063241106, "grad_norm": 0.07581355422735214, "learning_rate": 7.792079207920792e-06, "loss": 0.0146, "step": 4273 }, { "epoch": 16.893280632411066, "grad_norm": 0.14402788877487183, "learning_rate": 7.782178217821783e-06, "loss": 0.0113, "step": 4274 }, { "epoch": 16.897233201581027, "grad_norm": 0.09494798630475998, "learning_rate": 7.772277227722772e-06, "loss": 0.0157, "step": 4275 }, { "epoch": 16.90118577075099, "grad_norm": 0.12890876829624176, "learning_rate": 7.762376237623763e-06, "loss": 0.0166, "step": 4276 }, { "epoch": 16.90513833992095, "grad_norm": 0.0861034244298935, "learning_rate": 7.752475247524753e-06, "loss": 0.016, "step": 4277 }, { "epoch": 16.90909090909091, "grad_norm": 0.23555700480937958, "learning_rate": 7.742574257425744e-06, "loss": 0.0242, "step": 4278 }, { "epoch": 16.91304347826087, "grad_norm": 0.2678074240684509, "learning_rate": 7.732673267326733e-06, "loss": 0.0143, "step": 4279 }, { "epoch": 16.91699604743083, "grad_norm": 0.4950369596481323, "learning_rate": 7.722772277227722e-06, "loss": 0.0175, "step": 4280 }, { "epoch": 16.92094861660079, "grad_norm": 0.3049291968345642, "learning_rate": 7.712871287128713e-06, "loss": 0.02, "step": 4281 }, { "epoch": 16.92490118577075, "grad_norm": 0.44325828552246094, "learning_rate": 7.702970297029703e-06, "loss": 0.0157, "step": 4282 }, { "epoch": 16.92885375494071, "grad_norm": 0.13629592955112457, "learning_rate": 7.693069306930694e-06, "loss": 0.0264, "step": 4283 }, { "epoch": 16.932806324110672, "grad_norm": 0.24774757027626038, "learning_rate": 7.683168316831683e-06, "loss": 0.0197, "step": 4284 }, { "epoch": 16.936758893280633, "grad_norm": 0.21256321668624878, "learning_rate": 7.673267326732674e-06, "loss": 0.0187, "step": 4285 }, { "epoch": 16.940711462450594, "grad_norm": 0.7193269729614258, "learning_rate": 7.663366336633664e-06, "loss": 0.0368, "step": 4286 }, { "epoch": 16.944664031620555, "grad_norm": 0.10200168192386627, "learning_rate": 7.653465346534653e-06, "loss": 0.0209, "step": 4287 }, { "epoch": 16.948616600790515, "grad_norm": 0.12019100040197372, "learning_rate": 7.643564356435644e-06, "loss": 0.0154, "step": 4288 }, { "epoch": 16.952569169960473, "grad_norm": 0.23715560138225555, "learning_rate": 7.633663366336634e-06, "loss": 0.0327, "step": 4289 }, { "epoch": 16.956521739130434, "grad_norm": 0.17137587070465088, "learning_rate": 7.623762376237624e-06, "loss": 0.0281, "step": 4290 }, { "epoch": 16.960474308300395, "grad_norm": 0.3610689043998718, "learning_rate": 7.613861386138614e-06, "loss": 0.0221, "step": 4291 }, { "epoch": 16.964426877470355, "grad_norm": 0.11016876250505447, "learning_rate": 7.603960396039604e-06, "loss": 0.0228, "step": 4292 }, { "epoch": 16.968379446640316, "grad_norm": 0.10525012016296387, "learning_rate": 7.594059405940594e-06, "loss": 0.0152, "step": 4293 }, { "epoch": 16.972332015810277, "grad_norm": 0.1386335790157318, "learning_rate": 7.584158415841585e-06, "loss": 0.0198, "step": 4294 }, { "epoch": 16.976284584980238, "grad_norm": 0.21110780537128448, "learning_rate": 7.574257425742575e-06, "loss": 0.0165, "step": 4295 }, { "epoch": 16.9802371541502, "grad_norm": 0.14104916155338287, "learning_rate": 7.564356435643564e-06, "loss": 0.0222, "step": 4296 }, { "epoch": 16.984189723320156, "grad_norm": 0.500043511390686, "learning_rate": 7.554455445544554e-06, "loss": 0.0216, "step": 4297 }, { "epoch": 16.988142292490117, "grad_norm": 0.3443065583705902, "learning_rate": 7.544554455445545e-06, "loss": 0.0206, "step": 4298 }, { "epoch": 16.992094861660078, "grad_norm": 0.1256883293390274, "learning_rate": 7.534653465346535e-06, "loss": 0.02, "step": 4299 }, { "epoch": 16.99604743083004, "grad_norm": 0.38197946548461914, "learning_rate": 7.524752475247525e-06, "loss": 0.0235, "step": 4300 }, { "epoch": 17.0, "grad_norm": 0.06003795564174652, "learning_rate": 7.514851485148515e-06, "loss": 0.0122, "step": 4301 }, { "epoch": 17.00395256916996, "grad_norm": 0.08561106026172638, "learning_rate": 7.504950495049505e-06, "loss": 0.0194, "step": 4302 }, { "epoch": 17.007905138339922, "grad_norm": 0.07685429602861404, "learning_rate": 7.495049504950495e-06, "loss": 0.017, "step": 4303 }, { "epoch": 17.011857707509883, "grad_norm": 0.08368678390979767, "learning_rate": 7.485148514851485e-06, "loss": 0.0208, "step": 4304 }, { "epoch": 17.015810276679844, "grad_norm": 0.08058632910251617, "learning_rate": 7.475247524752475e-06, "loss": 0.0203, "step": 4305 }, { "epoch": 17.0197628458498, "grad_norm": 0.10826626420021057, "learning_rate": 7.4653465346534655e-06, "loss": 0.0263, "step": 4306 }, { "epoch": 17.023715415019762, "grad_norm": 0.11969059705734253, "learning_rate": 7.455445544554456e-06, "loss": 0.0298, "step": 4307 }, { "epoch": 17.027667984189723, "grad_norm": 0.06484473496675491, "learning_rate": 7.445544554455446e-06, "loss": 0.0156, "step": 4308 }, { "epoch": 17.031620553359684, "grad_norm": 0.27713918685913086, "learning_rate": 7.435643564356435e-06, "loss": 0.0184, "step": 4309 }, { "epoch": 17.035573122529645, "grad_norm": 0.0682242140173912, "learning_rate": 7.4257425742574256e-06, "loss": 0.0143, "step": 4310 }, { "epoch": 17.039525691699605, "grad_norm": 0.10864315181970596, "learning_rate": 7.415841584158416e-06, "loss": 0.0132, "step": 4311 }, { "epoch": 17.043478260869566, "grad_norm": 0.08563873171806335, "learning_rate": 7.405940594059406e-06, "loss": 0.016, "step": 4312 }, { "epoch": 17.047430830039527, "grad_norm": 0.09449490904808044, "learning_rate": 7.396039603960396e-06, "loss": 0.022, "step": 4313 }, { "epoch": 17.051383399209485, "grad_norm": 0.11153861880302429, "learning_rate": 7.3861386138613864e-06, "loss": 0.0236, "step": 4314 }, { "epoch": 17.055335968379445, "grad_norm": 0.13597802817821503, "learning_rate": 7.376237623762376e-06, "loss": 0.0308, "step": 4315 }, { "epoch": 17.059288537549406, "grad_norm": 0.07415709644556046, "learning_rate": 7.366336633663366e-06, "loss": 0.0156, "step": 4316 }, { "epoch": 17.063241106719367, "grad_norm": 0.07147336006164551, "learning_rate": 7.356435643564356e-06, "loss": 0.0157, "step": 4317 }, { "epoch": 17.067193675889328, "grad_norm": 0.06224348768591881, "learning_rate": 7.3465346534653465e-06, "loss": 0.0137, "step": 4318 }, { "epoch": 17.07114624505929, "grad_norm": 0.18249787390232086, "learning_rate": 7.336633663366337e-06, "loss": 0.0208, "step": 4319 }, { "epoch": 17.07509881422925, "grad_norm": 0.13768602907657623, "learning_rate": 7.326732673267327e-06, "loss": 0.03, "step": 4320 }, { "epoch": 17.07905138339921, "grad_norm": 0.07214892655611038, "learning_rate": 7.316831683168317e-06, "loss": 0.0143, "step": 4321 }, { "epoch": 17.08300395256917, "grad_norm": 0.06717263162136078, "learning_rate": 7.3069306930693065e-06, "loss": 0.0151, "step": 4322 }, { "epoch": 17.08695652173913, "grad_norm": 0.08252083510160446, "learning_rate": 7.297029702970297e-06, "loss": 0.0173, "step": 4323 }, { "epoch": 17.09090909090909, "grad_norm": 0.09129945188760757, "learning_rate": 7.287128712871287e-06, "loss": 0.0197, "step": 4324 }, { "epoch": 17.09486166007905, "grad_norm": 0.12937861680984497, "learning_rate": 7.277227722772277e-06, "loss": 0.0279, "step": 4325 }, { "epoch": 17.09881422924901, "grad_norm": 0.1928369551897049, "learning_rate": 7.267326732673267e-06, "loss": 0.0179, "step": 4326 }, { "epoch": 17.102766798418973, "grad_norm": 0.10406550765037537, "learning_rate": 7.257425742574258e-06, "loss": 0.0217, "step": 4327 }, { "epoch": 17.106719367588934, "grad_norm": 0.0942697823047638, "learning_rate": 7.247524752475247e-06, "loss": 0.0145, "step": 4328 }, { "epoch": 17.110671936758894, "grad_norm": 0.09337180107831955, "learning_rate": 7.237623762376237e-06, "loss": 0.0179, "step": 4329 }, { "epoch": 17.114624505928855, "grad_norm": 0.09112568944692612, "learning_rate": 7.227722772277227e-06, "loss": 0.0182, "step": 4330 }, { "epoch": 17.118577075098813, "grad_norm": 0.06735286116600037, "learning_rate": 7.217821782178218e-06, "loss": 0.0138, "step": 4331 }, { "epoch": 17.122529644268774, "grad_norm": 0.390716016292572, "learning_rate": 7.207920792079208e-06, "loss": 0.0128, "step": 4332 }, { "epoch": 17.126482213438734, "grad_norm": 0.15505212545394897, "learning_rate": 7.198019801980198e-06, "loss": 0.0209, "step": 4333 }, { "epoch": 17.130434782608695, "grad_norm": 0.07473071664571762, "learning_rate": 7.188118811881188e-06, "loss": 0.0155, "step": 4334 }, { "epoch": 17.134387351778656, "grad_norm": 0.08261846005916595, "learning_rate": 7.178217821782178e-06, "loss": 0.0161, "step": 4335 }, { "epoch": 17.138339920948617, "grad_norm": 0.10901064425706863, "learning_rate": 7.168316831683168e-06, "loss": 0.0192, "step": 4336 }, { "epoch": 17.142292490118578, "grad_norm": 0.1513066589832306, "learning_rate": 7.158415841584158e-06, "loss": 0.0138, "step": 4337 }, { "epoch": 17.14624505928854, "grad_norm": 0.0833035409450531, "learning_rate": 7.148514851485148e-06, "loss": 0.0167, "step": 4338 }, { "epoch": 17.150197628458496, "grad_norm": 0.106849804520607, "learning_rate": 7.1386138613861385e-06, "loss": 0.0203, "step": 4339 }, { "epoch": 17.154150197628457, "grad_norm": 0.08654819428920746, "learning_rate": 7.128712871287129e-06, "loss": 0.0168, "step": 4340 }, { "epoch": 17.158102766798418, "grad_norm": 0.14175891876220703, "learning_rate": 7.118811881188118e-06, "loss": 0.0292, "step": 4341 }, { "epoch": 17.16205533596838, "grad_norm": 0.3468667268753052, "learning_rate": 7.108910891089108e-06, "loss": 0.0532, "step": 4342 }, { "epoch": 17.16600790513834, "grad_norm": 0.08659147471189499, "learning_rate": 7.0990099009900986e-06, "loss": 0.0181, "step": 4343 }, { "epoch": 17.1699604743083, "grad_norm": 0.10130812227725983, "learning_rate": 7.089108910891089e-06, "loss": 0.0208, "step": 4344 }, { "epoch": 17.17391304347826, "grad_norm": 0.06992212682962418, "learning_rate": 7.079207920792079e-06, "loss": 0.0149, "step": 4345 }, { "epoch": 17.177865612648223, "grad_norm": 0.08047778904438019, "learning_rate": 7.069306930693069e-06, "loss": 0.015, "step": 4346 }, { "epoch": 17.181818181818183, "grad_norm": 0.1487332284450531, "learning_rate": 7.0594059405940594e-06, "loss": 0.0282, "step": 4347 }, { "epoch": 17.18577075098814, "grad_norm": 0.17126406729221344, "learning_rate": 7.049504950495049e-06, "loss": 0.0354, "step": 4348 }, { "epoch": 17.1897233201581, "grad_norm": 0.17127791047096252, "learning_rate": 7.039603960396039e-06, "loss": 0.0369, "step": 4349 }, { "epoch": 17.193675889328063, "grad_norm": 0.15056514739990234, "learning_rate": 7.029702970297031e-06, "loss": 0.0321, "step": 4350 }, { "epoch": 17.197628458498023, "grad_norm": 0.11148091405630112, "learning_rate": 7.019801980198021e-06, "loss": 0.0231, "step": 4351 }, { "epoch": 17.201581027667984, "grad_norm": 0.08004866540431976, "learning_rate": 7.0099009900990105e-06, "loss": 0.016, "step": 4352 }, { "epoch": 17.205533596837945, "grad_norm": 0.13770712912082672, "learning_rate": 7.000000000000001e-06, "loss": 0.0138, "step": 4353 }, { "epoch": 17.209486166007906, "grad_norm": 0.0932900682091713, "learning_rate": 6.990099009900991e-06, "loss": 0.0192, "step": 4354 }, { "epoch": 17.213438735177867, "grad_norm": 0.10534641146659851, "learning_rate": 6.980198019801981e-06, "loss": 0.0215, "step": 4355 }, { "epoch": 17.217391304347824, "grad_norm": 0.07112771272659302, "learning_rate": 6.970297029702971e-06, "loss": 0.0142, "step": 4356 }, { "epoch": 17.221343873517785, "grad_norm": 0.13422010838985443, "learning_rate": 6.960396039603962e-06, "loss": 0.0266, "step": 4357 }, { "epoch": 17.225296442687746, "grad_norm": 0.11412366479635239, "learning_rate": 6.950495049504952e-06, "loss": 0.0134, "step": 4358 }, { "epoch": 17.229249011857707, "grad_norm": 0.152086079120636, "learning_rate": 6.940594059405941e-06, "loss": 0.0302, "step": 4359 }, { "epoch": 17.233201581027668, "grad_norm": 0.21407432854175568, "learning_rate": 6.9306930693069314e-06, "loss": 0.0207, "step": 4360 }, { "epoch": 17.23715415019763, "grad_norm": 0.08378229290246964, "learning_rate": 6.920792079207922e-06, "loss": 0.0154, "step": 4361 }, { "epoch": 17.24110671936759, "grad_norm": 0.0637480691075325, "learning_rate": 6.910891089108912e-06, "loss": 0.0125, "step": 4362 }, { "epoch": 17.24505928853755, "grad_norm": 0.09339303523302078, "learning_rate": 6.900990099009902e-06, "loss": 0.0175, "step": 4363 }, { "epoch": 17.24901185770751, "grad_norm": 0.13023485243320465, "learning_rate": 6.891089108910892e-06, "loss": 0.0271, "step": 4364 }, { "epoch": 17.25296442687747, "grad_norm": 0.18006746470928192, "learning_rate": 6.881188118811882e-06, "loss": 0.0352, "step": 4365 }, { "epoch": 17.25691699604743, "grad_norm": 0.18851958215236664, "learning_rate": 6.871287128712872e-06, "loss": 0.0149, "step": 4366 }, { "epoch": 17.26086956521739, "grad_norm": 0.11703674495220184, "learning_rate": 6.861386138613862e-06, "loss": 0.0226, "step": 4367 }, { "epoch": 17.26482213438735, "grad_norm": 0.1664036065340042, "learning_rate": 6.851485148514852e-06, "loss": 0.0118, "step": 4368 }, { "epoch": 17.268774703557312, "grad_norm": 0.09806764125823975, "learning_rate": 6.8415841584158426e-06, "loss": 0.0194, "step": 4369 }, { "epoch": 17.272727272727273, "grad_norm": 0.16291888058185577, "learning_rate": 6.831683168316833e-06, "loss": 0.0362, "step": 4370 }, { "epoch": 17.276679841897234, "grad_norm": 0.09055842459201813, "learning_rate": 6.821782178217823e-06, "loss": 0.0173, "step": 4371 }, { "epoch": 17.280632411067195, "grad_norm": 0.21838222444057465, "learning_rate": 6.811881188118812e-06, "loss": 0.0313, "step": 4372 }, { "epoch": 17.284584980237153, "grad_norm": 0.0728110820055008, "learning_rate": 6.801980198019803e-06, "loss": 0.0153, "step": 4373 }, { "epoch": 17.288537549407113, "grad_norm": 0.09016624093055725, "learning_rate": 6.792079207920793e-06, "loss": 0.0189, "step": 4374 }, { "epoch": 17.292490118577074, "grad_norm": 0.1159011721611023, "learning_rate": 6.782178217821783e-06, "loss": 0.022, "step": 4375 }, { "epoch": 17.296442687747035, "grad_norm": 0.14058226346969604, "learning_rate": 6.772277227722773e-06, "loss": 0.0275, "step": 4376 }, { "epoch": 17.300395256916996, "grad_norm": 0.45038461685180664, "learning_rate": 6.7623762376237635e-06, "loss": 0.0185, "step": 4377 }, { "epoch": 17.304347826086957, "grad_norm": 0.17978446185588837, "learning_rate": 6.752475247524753e-06, "loss": 0.0222, "step": 4378 }, { "epoch": 17.308300395256918, "grad_norm": 0.08165811747312546, "learning_rate": 6.742574257425743e-06, "loss": 0.0152, "step": 4379 }, { "epoch": 17.31225296442688, "grad_norm": 0.10143844038248062, "learning_rate": 6.732673267326733e-06, "loss": 0.0187, "step": 4380 }, { "epoch": 17.31620553359684, "grad_norm": 0.08564656227827072, "learning_rate": 6.7227722772277235e-06, "loss": 0.0177, "step": 4381 }, { "epoch": 17.320158102766797, "grad_norm": 0.06620433181524277, "learning_rate": 6.712871287128714e-06, "loss": 0.0146, "step": 4382 }, { "epoch": 17.324110671936758, "grad_norm": 0.24403271079063416, "learning_rate": 6.702970297029704e-06, "loss": 0.0467, "step": 4383 }, { "epoch": 17.32806324110672, "grad_norm": 0.15951114892959595, "learning_rate": 6.693069306930694e-06, "loss": 0.0336, "step": 4384 }, { "epoch": 17.33201581027668, "grad_norm": 0.07333330810070038, "learning_rate": 6.6831683168316835e-06, "loss": 0.0161, "step": 4385 }, { "epoch": 17.33596837944664, "grad_norm": 0.07032448798418045, "learning_rate": 6.673267326732674e-06, "loss": 0.0141, "step": 4386 }, { "epoch": 17.3399209486166, "grad_norm": 0.23346160352230072, "learning_rate": 6.663366336633664e-06, "loss": 0.0162, "step": 4387 }, { "epoch": 17.343873517786562, "grad_norm": 0.13185279071331024, "learning_rate": 6.653465346534654e-06, "loss": 0.028, "step": 4388 }, { "epoch": 17.347826086956523, "grad_norm": 0.06911715120077133, "learning_rate": 6.643564356435644e-06, "loss": 0.0143, "step": 4389 }, { "epoch": 17.35177865612648, "grad_norm": 0.10566288232803345, "learning_rate": 6.633663366336635e-06, "loss": 0.022, "step": 4390 }, { "epoch": 17.35573122529644, "grad_norm": 0.09696114808320999, "learning_rate": 6.623762376237625e-06, "loss": 0.0209, "step": 4391 }, { "epoch": 17.359683794466402, "grad_norm": 0.11320017278194427, "learning_rate": 6.613861386138614e-06, "loss": 0.0243, "step": 4392 }, { "epoch": 17.363636363636363, "grad_norm": 0.10249149799346924, "learning_rate": 6.603960396039604e-06, "loss": 0.0199, "step": 4393 }, { "epoch": 17.367588932806324, "grad_norm": 0.08971542119979858, "learning_rate": 6.594059405940595e-06, "loss": 0.0194, "step": 4394 }, { "epoch": 17.371541501976285, "grad_norm": 0.0906844511628151, "learning_rate": 6.584158415841585e-06, "loss": 0.0192, "step": 4395 }, { "epoch": 17.375494071146246, "grad_norm": 0.1071322113275528, "learning_rate": 6.574257425742575e-06, "loss": 0.0224, "step": 4396 }, { "epoch": 17.379446640316207, "grad_norm": 0.4384217858314514, "learning_rate": 6.564356435643565e-06, "loss": 0.0215, "step": 4397 }, { "epoch": 17.383399209486164, "grad_norm": 0.1430341750383377, "learning_rate": 6.554455445544555e-06, "loss": 0.0276, "step": 4398 }, { "epoch": 17.387351778656125, "grad_norm": 0.33239081501960754, "learning_rate": 6.544554455445545e-06, "loss": 0.0177, "step": 4399 }, { "epoch": 17.391304347826086, "grad_norm": 0.06797671318054199, "learning_rate": 6.534653465346535e-06, "loss": 0.0136, "step": 4400 }, { "epoch": 17.395256916996047, "grad_norm": 0.13017240166664124, "learning_rate": 6.524752475247525e-06, "loss": 0.0299, "step": 4401 }, { "epoch": 17.399209486166008, "grad_norm": 0.09706712514162064, "learning_rate": 6.5148514851485155e-06, "loss": 0.0189, "step": 4402 }, { "epoch": 17.40316205533597, "grad_norm": 0.07423125952482224, "learning_rate": 6.504950495049506e-06, "loss": 0.0145, "step": 4403 }, { "epoch": 17.40711462450593, "grad_norm": 0.23500563204288483, "learning_rate": 6.495049504950496e-06, "loss": 0.0514, "step": 4404 }, { "epoch": 17.41106719367589, "grad_norm": 0.139862060546875, "learning_rate": 6.485148514851485e-06, "loss": 0.0298, "step": 4405 }, { "epoch": 17.41501976284585, "grad_norm": 0.060626767575740814, "learning_rate": 6.4752475247524756e-06, "loss": 0.0124, "step": 4406 }, { "epoch": 17.41897233201581, "grad_norm": 0.14867304265499115, "learning_rate": 6.465346534653466e-06, "loss": 0.0303, "step": 4407 }, { "epoch": 17.42292490118577, "grad_norm": 0.097183957695961, "learning_rate": 6.455445544554456e-06, "loss": 0.0165, "step": 4408 }, { "epoch": 17.42687747035573, "grad_norm": 0.08188674598932266, "learning_rate": 6.445544554455446e-06, "loss": 0.0176, "step": 4409 }, { "epoch": 17.43083003952569, "grad_norm": 0.09238259494304657, "learning_rate": 6.4356435643564364e-06, "loss": 0.0195, "step": 4410 }, { "epoch": 17.434782608695652, "grad_norm": 0.4010498523712158, "learning_rate": 6.425742574257426e-06, "loss": 0.0335, "step": 4411 }, { "epoch": 17.438735177865613, "grad_norm": 0.20239560306072235, "learning_rate": 6.415841584158416e-06, "loss": 0.0154, "step": 4412 }, { "epoch": 17.442687747035574, "grad_norm": 0.15141966938972473, "learning_rate": 6.405940594059406e-06, "loss": 0.0301, "step": 4413 }, { "epoch": 17.446640316205535, "grad_norm": 0.08622755110263824, "learning_rate": 6.3960396039603965e-06, "loss": 0.0179, "step": 4414 }, { "epoch": 17.450592885375492, "grad_norm": 0.08369860053062439, "learning_rate": 6.386138613861387e-06, "loss": 0.0164, "step": 4415 }, { "epoch": 17.454545454545453, "grad_norm": 0.18716922402381897, "learning_rate": 6.376237623762377e-06, "loss": 0.0196, "step": 4416 }, { "epoch": 17.458498023715414, "grad_norm": 0.08483263850212097, "learning_rate": 6.366336633663367e-06, "loss": 0.0153, "step": 4417 }, { "epoch": 17.462450592885375, "grad_norm": 0.08729425072669983, "learning_rate": 6.3564356435643565e-06, "loss": 0.0178, "step": 4418 }, { "epoch": 17.466403162055336, "grad_norm": 0.1177845150232315, "learning_rate": 6.346534653465347e-06, "loss": 0.0202, "step": 4419 }, { "epoch": 17.470355731225297, "grad_norm": 0.06448335945606232, "learning_rate": 6.336633663366337e-06, "loss": 0.013, "step": 4420 }, { "epoch": 17.474308300395258, "grad_norm": 0.10523983091115952, "learning_rate": 6.326732673267327e-06, "loss": 0.0214, "step": 4421 }, { "epoch": 17.47826086956522, "grad_norm": 0.0932609811425209, "learning_rate": 6.316831683168317e-06, "loss": 0.0164, "step": 4422 }, { "epoch": 17.48221343873518, "grad_norm": 0.11722265183925629, "learning_rate": 6.306930693069308e-06, "loss": 0.0245, "step": 4423 }, { "epoch": 17.486166007905137, "grad_norm": 0.09465071558952332, "learning_rate": 6.297029702970297e-06, "loss": 0.0188, "step": 4424 }, { "epoch": 17.490118577075098, "grad_norm": 0.06931892782449722, "learning_rate": 6.287128712871287e-06, "loss": 0.0116, "step": 4425 }, { "epoch": 17.49407114624506, "grad_norm": 0.09157134592533112, "learning_rate": 6.277227722772277e-06, "loss": 0.0193, "step": 4426 }, { "epoch": 17.49802371541502, "grad_norm": 0.29892972111701965, "learning_rate": 6.267326732673268e-06, "loss": 0.0237, "step": 4427 }, { "epoch": 17.50197628458498, "grad_norm": 0.6017363667488098, "learning_rate": 6.257425742574258e-06, "loss": 0.0239, "step": 4428 }, { "epoch": 17.50592885375494, "grad_norm": 0.09069151431322098, "learning_rate": 6.247524752475248e-06, "loss": 0.0173, "step": 4429 }, { "epoch": 17.509881422924902, "grad_norm": 0.11489500105381012, "learning_rate": 6.237623762376238e-06, "loss": 0.0238, "step": 4430 }, { "epoch": 17.513833992094863, "grad_norm": 0.07129693776369095, "learning_rate": 6.227722772277228e-06, "loss": 0.0153, "step": 4431 }, { "epoch": 17.51778656126482, "grad_norm": 0.12418951839208603, "learning_rate": 6.217821782178218e-06, "loss": 0.0163, "step": 4432 }, { "epoch": 17.52173913043478, "grad_norm": 0.061060115694999695, "learning_rate": 6.207920792079208e-06, "loss": 0.0127, "step": 4433 }, { "epoch": 17.525691699604742, "grad_norm": 0.08818425238132477, "learning_rate": 6.198019801980198e-06, "loss": 0.0174, "step": 4434 }, { "epoch": 17.529644268774703, "grad_norm": 0.09138414263725281, "learning_rate": 6.1881188118811885e-06, "loss": 0.0184, "step": 4435 }, { "epoch": 17.533596837944664, "grad_norm": 0.10886795073747635, "learning_rate": 6.178217821782179e-06, "loss": 0.0233, "step": 4436 }, { "epoch": 17.537549407114625, "grad_norm": 0.0993211567401886, "learning_rate": 6.168316831683168e-06, "loss": 0.0199, "step": 4437 }, { "epoch": 17.541501976284586, "grad_norm": 0.055329907685518265, "learning_rate": 6.158415841584158e-06, "loss": 0.0111, "step": 4438 }, { "epoch": 17.545454545454547, "grad_norm": 0.11244935542345047, "learning_rate": 6.1485148514851486e-06, "loss": 0.0258, "step": 4439 }, { "epoch": 17.549407114624508, "grad_norm": 0.10077984631061554, "learning_rate": 6.138613861386139e-06, "loss": 0.0205, "step": 4440 }, { "epoch": 17.553359683794465, "grad_norm": 0.21969063580036163, "learning_rate": 6.128712871287129e-06, "loss": 0.0204, "step": 4441 }, { "epoch": 17.557312252964426, "grad_norm": 0.08475147932767868, "learning_rate": 6.118811881188119e-06, "loss": 0.0168, "step": 4442 }, { "epoch": 17.561264822134387, "grad_norm": 0.07661508023738861, "learning_rate": 6.1089108910891094e-06, "loss": 0.0173, "step": 4443 }, { "epoch": 17.565217391304348, "grad_norm": 0.11320629715919495, "learning_rate": 6.099009900990099e-06, "loss": 0.0222, "step": 4444 }, { "epoch": 17.56916996047431, "grad_norm": 0.10110759735107422, "learning_rate": 6.089108910891089e-06, "loss": 0.0195, "step": 4445 }, { "epoch": 17.57312252964427, "grad_norm": 0.22751984000205994, "learning_rate": 6.079207920792079e-06, "loss": 0.0269, "step": 4446 }, { "epoch": 17.57707509881423, "grad_norm": 0.10587583482265472, "learning_rate": 6.0693069306930695e-06, "loss": 0.0225, "step": 4447 }, { "epoch": 17.58102766798419, "grad_norm": 0.06958726793527603, "learning_rate": 6.05940594059406e-06, "loss": 0.0154, "step": 4448 }, { "epoch": 17.58498023715415, "grad_norm": 0.10610753297805786, "learning_rate": 6.04950495049505e-06, "loss": 0.0231, "step": 4449 }, { "epoch": 17.58893280632411, "grad_norm": 0.10857275873422623, "learning_rate": 6.03960396039604e-06, "loss": 0.0239, "step": 4450 }, { "epoch": 17.59288537549407, "grad_norm": 0.13145434856414795, "learning_rate": 6.0297029702970295e-06, "loss": 0.0291, "step": 4451 }, { "epoch": 17.59683794466403, "grad_norm": 0.26128390431404114, "learning_rate": 6.01980198019802e-06, "loss": 0.0391, "step": 4452 }, { "epoch": 17.600790513833992, "grad_norm": 0.0948048084974289, "learning_rate": 6.00990099009901e-06, "loss": 0.0191, "step": 4453 }, { "epoch": 17.604743083003953, "grad_norm": 0.09174801409244537, "learning_rate": 6e-06, "loss": 0.0221, "step": 4454 }, { "epoch": 17.608695652173914, "grad_norm": 0.04810671880841255, "learning_rate": 5.99009900990099e-06, "loss": 0.01, "step": 4455 }, { "epoch": 17.612648221343875, "grad_norm": 0.08879343420267105, "learning_rate": 5.980198019801981e-06, "loss": 0.0121, "step": 4456 }, { "epoch": 17.616600790513836, "grad_norm": 0.09187214076519012, "learning_rate": 5.97029702970297e-06, "loss": 0.0133, "step": 4457 }, { "epoch": 17.620553359683793, "grad_norm": 0.5850670337677002, "learning_rate": 5.96039603960396e-06, "loss": 0.0185, "step": 4458 }, { "epoch": 17.624505928853754, "grad_norm": 0.20422111451625824, "learning_rate": 5.95049504950495e-06, "loss": 0.0163, "step": 4459 }, { "epoch": 17.628458498023715, "grad_norm": 1.0020602941513062, "learning_rate": 5.940594059405941e-06, "loss": 0.0205, "step": 4460 }, { "epoch": 17.632411067193676, "grad_norm": 0.19633428752422333, "learning_rate": 5.930693069306931e-06, "loss": 0.0377, "step": 4461 }, { "epoch": 17.636363636363637, "grad_norm": 0.14987234771251678, "learning_rate": 5.920792079207921e-06, "loss": 0.0328, "step": 4462 }, { "epoch": 17.640316205533598, "grad_norm": 0.2607419490814209, "learning_rate": 5.910891089108911e-06, "loss": 0.0204, "step": 4463 }, { "epoch": 17.64426877470356, "grad_norm": 0.2606803774833679, "learning_rate": 5.900990099009901e-06, "loss": 0.0156, "step": 4464 }, { "epoch": 17.64822134387352, "grad_norm": 0.10839425027370453, "learning_rate": 5.891089108910891e-06, "loss": 0.0219, "step": 4465 }, { "epoch": 17.652173913043477, "grad_norm": 0.1941925585269928, "learning_rate": 5.881188118811881e-06, "loss": 0.0187, "step": 4466 }, { "epoch": 17.656126482213438, "grad_norm": 0.12735332548618317, "learning_rate": 5.871287128712871e-06, "loss": 0.03, "step": 4467 }, { "epoch": 17.6600790513834, "grad_norm": 0.3260721266269684, "learning_rate": 5.8613861386138615e-06, "loss": 0.0196, "step": 4468 }, { "epoch": 17.66403162055336, "grad_norm": 0.4856257140636444, "learning_rate": 5.851485148514852e-06, "loss": 0.025, "step": 4469 }, { "epoch": 17.66798418972332, "grad_norm": 0.11980253458023071, "learning_rate": 5.841584158415842e-06, "loss": 0.0144, "step": 4470 }, { "epoch": 17.67193675889328, "grad_norm": 0.31592240929603577, "learning_rate": 5.831683168316832e-06, "loss": 0.0184, "step": 4471 }, { "epoch": 17.675889328063242, "grad_norm": 0.11566602438688278, "learning_rate": 5.821782178217822e-06, "loss": 0.0263, "step": 4472 }, { "epoch": 17.679841897233203, "grad_norm": 0.07889257371425629, "learning_rate": 5.811881188118813e-06, "loss": 0.0152, "step": 4473 }, { "epoch": 17.683794466403164, "grad_norm": 0.292725533246994, "learning_rate": 5.801980198019803e-06, "loss": 0.0232, "step": 4474 }, { "epoch": 17.68774703557312, "grad_norm": 0.151706263422966, "learning_rate": 5.792079207920793e-06, "loss": 0.0335, "step": 4475 }, { "epoch": 17.691699604743082, "grad_norm": 0.12027816474437714, "learning_rate": 5.7821782178217824e-06, "loss": 0.0259, "step": 4476 }, { "epoch": 17.695652173913043, "grad_norm": 0.14188127219676971, "learning_rate": 5.772277227722773e-06, "loss": 0.0278, "step": 4477 }, { "epoch": 17.699604743083004, "grad_norm": 0.05881866067647934, "learning_rate": 5.762376237623763e-06, "loss": 0.0126, "step": 4478 }, { "epoch": 17.703557312252965, "grad_norm": 0.09773662686347961, "learning_rate": 5.752475247524753e-06, "loss": 0.0216, "step": 4479 }, { "epoch": 17.707509881422926, "grad_norm": 0.10499925911426544, "learning_rate": 5.742574257425743e-06, "loss": 0.0218, "step": 4480 }, { "epoch": 17.711462450592887, "grad_norm": 0.06858788430690765, "learning_rate": 5.7326732673267335e-06, "loss": 0.0138, "step": 4481 }, { "epoch": 17.715415019762847, "grad_norm": 0.06887722760438919, "learning_rate": 5.722772277227723e-06, "loss": 0.0132, "step": 4482 }, { "epoch": 17.719367588932805, "grad_norm": 0.11645437777042389, "learning_rate": 5.712871287128713e-06, "loss": 0.0238, "step": 4483 }, { "epoch": 17.723320158102766, "grad_norm": 0.0915154367685318, "learning_rate": 5.702970297029703e-06, "loss": 0.0205, "step": 4484 }, { "epoch": 17.727272727272727, "grad_norm": 0.11392847448587418, "learning_rate": 5.6930693069306936e-06, "loss": 0.0219, "step": 4485 }, { "epoch": 17.731225296442688, "grad_norm": 0.22711199522018433, "learning_rate": 5.683168316831684e-06, "loss": 0.0272, "step": 4486 }, { "epoch": 17.73517786561265, "grad_norm": 0.11765439808368683, "learning_rate": 5.673267326732674e-06, "loss": 0.026, "step": 4487 }, { "epoch": 17.73913043478261, "grad_norm": 0.07968644797801971, "learning_rate": 5.663366336633664e-06, "loss": 0.0131, "step": 4488 }, { "epoch": 17.74308300395257, "grad_norm": 0.08743298053741455, "learning_rate": 5.653465346534654e-06, "loss": 0.0184, "step": 4489 }, { "epoch": 17.74703557312253, "grad_norm": 0.29735442996025085, "learning_rate": 5.643564356435644e-06, "loss": 0.0257, "step": 4490 }, { "epoch": 17.75098814229249, "grad_norm": 0.07986912876367569, "learning_rate": 5.633663366336634e-06, "loss": 0.017, "step": 4491 }, { "epoch": 17.75494071146245, "grad_norm": 0.07551529258489609, "learning_rate": 5.623762376237624e-06, "loss": 0.0168, "step": 4492 }, { "epoch": 17.75889328063241, "grad_norm": 0.07317304611206055, "learning_rate": 5.6138613861386145e-06, "loss": 0.0163, "step": 4493 }, { "epoch": 17.76284584980237, "grad_norm": 0.09270171821117401, "learning_rate": 5.603960396039605e-06, "loss": 0.0196, "step": 4494 }, { "epoch": 17.766798418972332, "grad_norm": 0.0829479843378067, "learning_rate": 5.594059405940594e-06, "loss": 0.018, "step": 4495 }, { "epoch": 17.770750988142293, "grad_norm": 0.19506104290485382, "learning_rate": 5.584158415841584e-06, "loss": 0.0474, "step": 4496 }, { "epoch": 17.774703557312254, "grad_norm": 0.0977305993437767, "learning_rate": 5.5742574257425745e-06, "loss": 0.0215, "step": 4497 }, { "epoch": 17.778656126482215, "grad_norm": 0.10010389238595963, "learning_rate": 5.564356435643565e-06, "loss": 0.0219, "step": 4498 }, { "epoch": 17.782608695652176, "grad_norm": 0.1320183128118515, "learning_rate": 5.554455445544555e-06, "loss": 0.0164, "step": 4499 }, { "epoch": 17.786561264822133, "grad_norm": 0.13019560277462006, "learning_rate": 5.544554455445545e-06, "loss": 0.0237, "step": 4500 }, { "epoch": 17.790513833992094, "grad_norm": 0.08445321768522263, "learning_rate": 5.534653465346535e-06, "loss": 0.0184, "step": 4501 }, { "epoch": 17.794466403162055, "grad_norm": 0.09389789402484894, "learning_rate": 5.524752475247525e-06, "loss": 0.0191, "step": 4502 }, { "epoch": 17.798418972332016, "grad_norm": 0.2428017258644104, "learning_rate": 5.514851485148515e-06, "loss": 0.0237, "step": 4503 }, { "epoch": 17.802371541501977, "grad_norm": 0.06421608477830887, "learning_rate": 5.504950495049505e-06, "loss": 0.0136, "step": 4504 }, { "epoch": 17.806324110671937, "grad_norm": 0.08237195014953613, "learning_rate": 5.495049504950495e-06, "loss": 0.0189, "step": 4505 }, { "epoch": 17.8102766798419, "grad_norm": 0.23471198976039886, "learning_rate": 5.485148514851486e-06, "loss": 0.0478, "step": 4506 }, { "epoch": 17.81422924901186, "grad_norm": 0.09500358998775482, "learning_rate": 5.475247524752476e-06, "loss": 0.0221, "step": 4507 }, { "epoch": 17.818181818181817, "grad_norm": 0.09819921106100082, "learning_rate": 5.465346534653465e-06, "loss": 0.0221, "step": 4508 }, { "epoch": 17.822134387351777, "grad_norm": 0.5426698327064514, "learning_rate": 5.455445544554455e-06, "loss": 0.0197, "step": 4509 }, { "epoch": 17.82608695652174, "grad_norm": 0.11651584506034851, "learning_rate": 5.445544554455446e-06, "loss": 0.0209, "step": 4510 }, { "epoch": 17.8300395256917, "grad_norm": 0.06901956349611282, "learning_rate": 5.435643564356436e-06, "loss": 0.0158, "step": 4511 }, { "epoch": 17.83399209486166, "grad_norm": 0.0681718960404396, "learning_rate": 5.425742574257426e-06, "loss": 0.0159, "step": 4512 }, { "epoch": 17.83794466403162, "grad_norm": 0.22152681648731232, "learning_rate": 5.415841584158416e-06, "loss": 0.0169, "step": 4513 }, { "epoch": 17.841897233201582, "grad_norm": 0.39504361152648926, "learning_rate": 5.4059405940594065e-06, "loss": 0.0282, "step": 4514 }, { "epoch": 17.845849802371543, "grad_norm": 0.24436047673225403, "learning_rate": 5.396039603960396e-06, "loss": 0.0292, "step": 4515 }, { "epoch": 17.8498023715415, "grad_norm": 0.06409277021884918, "learning_rate": 5.386138613861386e-06, "loss": 0.0141, "step": 4516 }, { "epoch": 17.85375494071146, "grad_norm": 0.3022923469543457, "learning_rate": 5.376237623762376e-06, "loss": 0.0165, "step": 4517 }, { "epoch": 17.857707509881422, "grad_norm": 0.11901092529296875, "learning_rate": 5.3663366336633665e-06, "loss": 0.0265, "step": 4518 }, { "epoch": 17.861660079051383, "grad_norm": 0.09014872461557388, "learning_rate": 5.356435643564357e-06, "loss": 0.0196, "step": 4519 }, { "epoch": 17.865612648221344, "grad_norm": 0.20630097389221191, "learning_rate": 5.346534653465347e-06, "loss": 0.0225, "step": 4520 }, { "epoch": 17.869565217391305, "grad_norm": 0.09917695820331573, "learning_rate": 5.336633663366336e-06, "loss": 0.0218, "step": 4521 }, { "epoch": 17.873517786561266, "grad_norm": 0.08082331717014313, "learning_rate": 5.3267326732673266e-06, "loss": 0.0188, "step": 4522 }, { "epoch": 17.877470355731226, "grad_norm": 0.07425107806921005, "learning_rate": 5.316831683168317e-06, "loss": 0.017, "step": 4523 }, { "epoch": 17.881422924901187, "grad_norm": 0.07668577134609222, "learning_rate": 5.306930693069307e-06, "loss": 0.0172, "step": 4524 }, { "epoch": 17.885375494071145, "grad_norm": 0.07960010319948196, "learning_rate": 5.297029702970297e-06, "loss": 0.0172, "step": 4525 }, { "epoch": 17.889328063241106, "grad_norm": 0.07505331933498383, "learning_rate": 5.2871287128712874e-06, "loss": 0.0179, "step": 4526 }, { "epoch": 17.893280632411066, "grad_norm": 0.26071545481681824, "learning_rate": 5.277227722772278e-06, "loss": 0.0249, "step": 4527 }, { "epoch": 17.897233201581027, "grad_norm": 0.1586107313632965, "learning_rate": 5.267326732673267e-06, "loss": 0.0176, "step": 4528 }, { "epoch": 17.90118577075099, "grad_norm": 0.06614018976688385, "learning_rate": 5.257425742574257e-06, "loss": 0.0151, "step": 4529 }, { "epoch": 17.90513833992095, "grad_norm": 0.1017146036028862, "learning_rate": 5.2475247524752475e-06, "loss": 0.0231, "step": 4530 }, { "epoch": 17.90909090909091, "grad_norm": 0.12340854108333588, "learning_rate": 5.237623762376238e-06, "loss": 0.0176, "step": 4531 }, { "epoch": 17.91304347826087, "grad_norm": 0.1261013001203537, "learning_rate": 5.227722772277228e-06, "loss": 0.0262, "step": 4532 }, { "epoch": 17.91699604743083, "grad_norm": 0.05789937824010849, "learning_rate": 5.217821782178218e-06, "loss": 0.013, "step": 4533 }, { "epoch": 17.92094861660079, "grad_norm": 0.08536413311958313, "learning_rate": 5.207920792079208e-06, "loss": 0.0196, "step": 4534 }, { "epoch": 17.92490118577075, "grad_norm": 0.09243008494377136, "learning_rate": 5.198019801980198e-06, "loss": 0.0222, "step": 4535 }, { "epoch": 17.92885375494071, "grad_norm": 0.12908107042312622, "learning_rate": 5.188118811881188e-06, "loss": 0.0235, "step": 4536 }, { "epoch": 17.932806324110672, "grad_norm": 0.16983526945114136, "learning_rate": 5.178217821782178e-06, "loss": 0.0361, "step": 4537 }, { "epoch": 17.936758893280633, "grad_norm": 0.07735263556241989, "learning_rate": 5.168316831683168e-06, "loss": 0.016, "step": 4538 }, { "epoch": 17.940711462450594, "grad_norm": 0.071987085044384, "learning_rate": 5.158415841584159e-06, "loss": 0.0162, "step": 4539 }, { "epoch": 17.944664031620555, "grad_norm": 0.06897421181201935, "learning_rate": 5.148514851485149e-06, "loss": 0.0145, "step": 4540 }, { "epoch": 17.948616600790515, "grad_norm": 0.0706239640712738, "learning_rate": 5.138613861386138e-06, "loss": 0.0151, "step": 4541 }, { "epoch": 17.952569169960473, "grad_norm": 0.07290058583021164, "learning_rate": 5.128712871287128e-06, "loss": 0.0154, "step": 4542 }, { "epoch": 17.956521739130434, "grad_norm": 0.0573497973382473, "learning_rate": 5.118811881188119e-06, "loss": 0.0131, "step": 4543 }, { "epoch": 17.960474308300395, "grad_norm": 0.13600096106529236, "learning_rate": 5.108910891089109e-06, "loss": 0.0277, "step": 4544 }, { "epoch": 17.964426877470355, "grad_norm": 0.12543384730815887, "learning_rate": 5.099009900990099e-06, "loss": 0.0277, "step": 4545 }, { "epoch": 17.968379446640316, "grad_norm": 0.07211162894964218, "learning_rate": 5.089108910891089e-06, "loss": 0.0148, "step": 4546 }, { "epoch": 17.972332015810277, "grad_norm": 0.08077354729175568, "learning_rate": 5.0792079207920795e-06, "loss": 0.0187, "step": 4547 }, { "epoch": 17.976284584980238, "grad_norm": 0.058231011033058167, "learning_rate": 5.06930693069307e-06, "loss": 0.0123, "step": 4548 }, { "epoch": 17.9802371541502, "grad_norm": 0.10743081569671631, "learning_rate": 5.05940594059406e-06, "loss": 0.0163, "step": 4549 }, { "epoch": 17.984189723320156, "grad_norm": 0.08976811915636063, "learning_rate": 5.04950495049505e-06, "loss": 0.0185, "step": 4550 }, { "epoch": 17.988142292490117, "grad_norm": 0.08673840016126633, "learning_rate": 5.03960396039604e-06, "loss": 0.0198, "step": 4551 }, { "epoch": 17.992094861660078, "grad_norm": 0.0923733189702034, "learning_rate": 5.029702970297031e-06, "loss": 0.0203, "step": 4552 }, { "epoch": 17.99604743083004, "grad_norm": 0.20303665101528168, "learning_rate": 5.01980198019802e-06, "loss": 0.0447, "step": 4553 }, { "epoch": 18.0, "grad_norm": 0.42744511365890503, "learning_rate": 5.00990099009901e-06, "loss": 0.0209, "step": 4554 }, { "epoch": 18.00395256916996, "grad_norm": 0.06428878754377365, "learning_rate": 5e-06, "loss": 0.0157, "step": 4555 }, { "epoch": 18.007905138339922, "grad_norm": 0.0634344220161438, "learning_rate": 4.990099009900991e-06, "loss": 0.0145, "step": 4556 }, { "epoch": 18.011857707509883, "grad_norm": 0.06198235973715782, "learning_rate": 4.980198019801981e-06, "loss": 0.0136, "step": 4557 }, { "epoch": 18.015810276679844, "grad_norm": 0.0970991998910904, "learning_rate": 4.970297029702971e-06, "loss": 0.0247, "step": 4558 }, { "epoch": 18.0197628458498, "grad_norm": 0.14284563064575195, "learning_rate": 4.960396039603961e-06, "loss": 0.0341, "step": 4559 }, { "epoch": 18.023715415019762, "grad_norm": 0.060979168862104416, "learning_rate": 4.950495049504951e-06, "loss": 0.0144, "step": 4560 }, { "epoch": 18.027667984189723, "grad_norm": 0.07451700419187546, "learning_rate": 4.940594059405941e-06, "loss": 0.0166, "step": 4561 }, { "epoch": 18.031620553359684, "grad_norm": 0.09503410011529922, "learning_rate": 4.930693069306931e-06, "loss": 0.0203, "step": 4562 }, { "epoch": 18.035573122529645, "grad_norm": 0.08138763159513474, "learning_rate": 4.920792079207921e-06, "loss": 0.0175, "step": 4563 }, { "epoch": 18.039525691699605, "grad_norm": 0.07171603292226791, "learning_rate": 4.9108910891089115e-06, "loss": 0.0166, "step": 4564 }, { "epoch": 18.043478260869566, "grad_norm": 0.06181710585951805, "learning_rate": 4.900990099009902e-06, "loss": 0.0151, "step": 4565 }, { "epoch": 18.047430830039527, "grad_norm": 0.09416337311267853, "learning_rate": 4.891089108910891e-06, "loss": 0.0222, "step": 4566 }, { "epoch": 18.051383399209485, "grad_norm": 0.09318000078201294, "learning_rate": 4.881188118811881e-06, "loss": 0.0201, "step": 4567 }, { "epoch": 18.055335968379445, "grad_norm": 0.06864951550960541, "learning_rate": 4.8712871287128716e-06, "loss": 0.0157, "step": 4568 }, { "epoch": 18.059288537549406, "grad_norm": 0.0793004184961319, "learning_rate": 4.861386138613862e-06, "loss": 0.0192, "step": 4569 }, { "epoch": 18.063241106719367, "grad_norm": 0.09409666806459427, "learning_rate": 4.851485148514852e-06, "loss": 0.0214, "step": 4570 }, { "epoch": 18.067193675889328, "grad_norm": 0.08815417438745499, "learning_rate": 4.841584158415842e-06, "loss": 0.0197, "step": 4571 }, { "epoch": 18.07114624505929, "grad_norm": 0.07333401590585709, "learning_rate": 4.8316831683168324e-06, "loss": 0.0171, "step": 4572 }, { "epoch": 18.07509881422925, "grad_norm": 0.07872568070888519, "learning_rate": 4.821782178217822e-06, "loss": 0.0187, "step": 4573 }, { "epoch": 18.07905138339921, "grad_norm": 0.115149587392807, "learning_rate": 4.811881188118812e-06, "loss": 0.0237, "step": 4574 }, { "epoch": 18.08300395256917, "grad_norm": 0.11247420310974121, "learning_rate": 4.801980198019802e-06, "loss": 0.0249, "step": 4575 }, { "epoch": 18.08695652173913, "grad_norm": 0.07307711243629456, "learning_rate": 4.7920792079207925e-06, "loss": 0.0166, "step": 4576 }, { "epoch": 18.09090909090909, "grad_norm": 0.3838382065296173, "learning_rate": 4.782178217821783e-06, "loss": 0.0263, "step": 4577 }, { "epoch": 18.09486166007905, "grad_norm": 0.09118162840604782, "learning_rate": 4.772277227722773e-06, "loss": 0.0199, "step": 4578 }, { "epoch": 18.09881422924901, "grad_norm": 0.08479972183704376, "learning_rate": 4.762376237623762e-06, "loss": 0.0193, "step": 4579 }, { "epoch": 18.102766798418973, "grad_norm": 0.08059870451688766, "learning_rate": 4.7524752475247525e-06, "loss": 0.0175, "step": 4580 }, { "epoch": 18.106719367588934, "grad_norm": 0.14460915327072144, "learning_rate": 4.742574257425743e-06, "loss": 0.0323, "step": 4581 }, { "epoch": 18.110671936758894, "grad_norm": 0.07031207531690598, "learning_rate": 4.732673267326733e-06, "loss": 0.014, "step": 4582 }, { "epoch": 18.114624505928855, "grad_norm": 0.09666726738214493, "learning_rate": 4.722772277227723e-06, "loss": 0.0216, "step": 4583 }, { "epoch": 18.118577075098813, "grad_norm": 0.19726380705833435, "learning_rate": 4.712871287128713e-06, "loss": 0.0439, "step": 4584 }, { "epoch": 18.122529644268774, "grad_norm": 0.061520837247371674, "learning_rate": 4.702970297029704e-06, "loss": 0.0126, "step": 4585 }, { "epoch": 18.126482213438734, "grad_norm": 0.09368573129177094, "learning_rate": 4.693069306930693e-06, "loss": 0.0214, "step": 4586 }, { "epoch": 18.130434782608695, "grad_norm": 0.09103845059871674, "learning_rate": 4.683168316831683e-06, "loss": 0.0195, "step": 4587 }, { "epoch": 18.134387351778656, "grad_norm": 0.15128833055496216, "learning_rate": 4.673267326732673e-06, "loss": 0.0137, "step": 4588 }, { "epoch": 18.138339920948617, "grad_norm": 0.14402815699577332, "learning_rate": 4.663366336633664e-06, "loss": 0.0267, "step": 4589 }, { "epoch": 18.142292490118578, "grad_norm": 0.08224126696586609, "learning_rate": 4.653465346534654e-06, "loss": 0.0149, "step": 4590 }, { "epoch": 18.14624505928854, "grad_norm": 0.10302513092756271, "learning_rate": 4.643564356435644e-06, "loss": 0.0221, "step": 4591 }, { "epoch": 18.150197628458496, "grad_norm": 0.09976250678300858, "learning_rate": 4.6336633663366334e-06, "loss": 0.0216, "step": 4592 }, { "epoch": 18.154150197628457, "grad_norm": 0.08072274178266525, "learning_rate": 4.623762376237624e-06, "loss": 0.017, "step": 4593 }, { "epoch": 18.158102766798418, "grad_norm": 0.07912739366292953, "learning_rate": 4.613861386138614e-06, "loss": 0.017, "step": 4594 }, { "epoch": 18.16205533596838, "grad_norm": 0.1479351967573166, "learning_rate": 4.603960396039604e-06, "loss": 0.0281, "step": 4595 }, { "epoch": 18.16600790513834, "grad_norm": 0.16142329573631287, "learning_rate": 4.594059405940594e-06, "loss": 0.0352, "step": 4596 }, { "epoch": 18.1699604743083, "grad_norm": 0.08844535797834396, "learning_rate": 4.5841584158415845e-06, "loss": 0.0191, "step": 4597 }, { "epoch": 18.17391304347826, "grad_norm": 0.07092805206775665, "learning_rate": 4.574257425742575e-06, "loss": 0.0158, "step": 4598 }, { "epoch": 18.177865612648223, "grad_norm": 0.07968784123659134, "learning_rate": 4.564356435643564e-06, "loss": 0.0159, "step": 4599 }, { "epoch": 18.181818181818183, "grad_norm": 0.09544012695550919, "learning_rate": 4.554455445544554e-06, "loss": 0.0206, "step": 4600 }, { "epoch": 18.18577075098814, "grad_norm": 0.09112362563610077, "learning_rate": 4.5445544554455446e-06, "loss": 0.0193, "step": 4601 }, { "epoch": 18.1897233201581, "grad_norm": 0.09787477552890778, "learning_rate": 4.534653465346535e-06, "loss": 0.0195, "step": 4602 }, { "epoch": 18.193675889328063, "grad_norm": 0.09266125410795212, "learning_rate": 4.524752475247525e-06, "loss": 0.0198, "step": 4603 }, { "epoch": 18.197628458498023, "grad_norm": 0.0787879154086113, "learning_rate": 4.514851485148515e-06, "loss": 0.0163, "step": 4604 }, { "epoch": 18.201581027667984, "grad_norm": 0.19169843196868896, "learning_rate": 4.504950495049505e-06, "loss": 0.0405, "step": 4605 }, { "epoch": 18.205533596837945, "grad_norm": 0.10360418260097504, "learning_rate": 4.495049504950495e-06, "loss": 0.0223, "step": 4606 }, { "epoch": 18.209486166007906, "grad_norm": 0.05530301854014397, "learning_rate": 4.485148514851485e-06, "loss": 0.0116, "step": 4607 }, { "epoch": 18.213438735177867, "grad_norm": 0.09453761577606201, "learning_rate": 4.475247524752475e-06, "loss": 0.0199, "step": 4608 }, { "epoch": 18.217391304347824, "grad_norm": 0.08177429437637329, "learning_rate": 4.4653465346534655e-06, "loss": 0.0165, "step": 4609 }, { "epoch": 18.221343873517785, "grad_norm": 0.06628158688545227, "learning_rate": 4.455445544554456e-06, "loss": 0.0156, "step": 4610 }, { "epoch": 18.225296442687746, "grad_norm": 0.12673664093017578, "learning_rate": 4.445544554455446e-06, "loss": 0.029, "step": 4611 }, { "epoch": 18.229249011857707, "grad_norm": 0.09021451324224472, "learning_rate": 4.435643564356435e-06, "loss": 0.0187, "step": 4612 }, { "epoch": 18.233201581027668, "grad_norm": 0.09661680459976196, "learning_rate": 4.4257425742574255e-06, "loss": 0.0195, "step": 4613 }, { "epoch": 18.23715415019763, "grad_norm": 0.060823675245046616, "learning_rate": 4.415841584158416e-06, "loss": 0.013, "step": 4614 }, { "epoch": 18.24110671936759, "grad_norm": 0.11025483161211014, "learning_rate": 4.405940594059406e-06, "loss": 0.0218, "step": 4615 }, { "epoch": 18.24505928853755, "grad_norm": 0.08044340461492538, "learning_rate": 4.396039603960396e-06, "loss": 0.0169, "step": 4616 }, { "epoch": 18.24901185770751, "grad_norm": 0.09354065358638763, "learning_rate": 4.386138613861386e-06, "loss": 0.0196, "step": 4617 }, { "epoch": 18.25296442687747, "grad_norm": 0.07635927200317383, "learning_rate": 4.376237623762377e-06, "loss": 0.0162, "step": 4618 }, { "epoch": 18.25691699604743, "grad_norm": 0.0802261009812355, "learning_rate": 4.366336633663366e-06, "loss": 0.0167, "step": 4619 }, { "epoch": 18.26086956521739, "grad_norm": 0.09504250437021255, "learning_rate": 4.356435643564356e-06, "loss": 0.0194, "step": 4620 }, { "epoch": 18.26482213438735, "grad_norm": 0.09892295300960541, "learning_rate": 4.346534653465346e-06, "loss": 0.0205, "step": 4621 }, { "epoch": 18.268774703557312, "grad_norm": 0.1158684715628624, "learning_rate": 4.336633663366337e-06, "loss": 0.0236, "step": 4622 }, { "epoch": 18.272727272727273, "grad_norm": 0.09733343124389648, "learning_rate": 4.326732673267327e-06, "loss": 0.0215, "step": 4623 }, { "epoch": 18.276679841897234, "grad_norm": 0.07804633677005768, "learning_rate": 4.316831683168317e-06, "loss": 0.0168, "step": 4624 }, { "epoch": 18.280632411067195, "grad_norm": 0.08710518479347229, "learning_rate": 4.306930693069306e-06, "loss": 0.019, "step": 4625 }, { "epoch": 18.284584980237153, "grad_norm": 0.1010795384645462, "learning_rate": 4.297029702970297e-06, "loss": 0.0222, "step": 4626 }, { "epoch": 18.288537549407113, "grad_norm": 0.05708782374858856, "learning_rate": 4.287128712871288e-06, "loss": 0.0118, "step": 4627 }, { "epoch": 18.292490118577074, "grad_norm": 0.13330306112766266, "learning_rate": 4.277227722772278e-06, "loss": 0.0276, "step": 4628 }, { "epoch": 18.296442687747035, "grad_norm": 0.09566564112901688, "learning_rate": 4.267326732673268e-06, "loss": 0.0198, "step": 4629 }, { "epoch": 18.300395256916996, "grad_norm": 0.08801818639039993, "learning_rate": 4.257425742574258e-06, "loss": 0.018, "step": 4630 }, { "epoch": 18.304347826086957, "grad_norm": 0.09293770790100098, "learning_rate": 4.247524752475248e-06, "loss": 0.0205, "step": 4631 }, { "epoch": 18.308300395256918, "grad_norm": 0.11491464823484421, "learning_rate": 4.237623762376238e-06, "loss": 0.0222, "step": 4632 }, { "epoch": 18.31225296442688, "grad_norm": 0.6014373898506165, "learning_rate": 4.227722772277228e-06, "loss": 0.049, "step": 4633 }, { "epoch": 18.31620553359684, "grad_norm": 0.09594627469778061, "learning_rate": 4.217821782178218e-06, "loss": 0.0156, "step": 4634 }, { "epoch": 18.320158102766797, "grad_norm": 0.08056531846523285, "learning_rate": 4.207920792079209e-06, "loss": 0.0167, "step": 4635 }, { "epoch": 18.324110671936758, "grad_norm": 0.07412175089120865, "learning_rate": 4.198019801980199e-06, "loss": 0.0163, "step": 4636 }, { "epoch": 18.32806324110672, "grad_norm": 0.10852610319852829, "learning_rate": 4.188118811881188e-06, "loss": 0.0226, "step": 4637 }, { "epoch": 18.33201581027668, "grad_norm": 0.14057545363903046, "learning_rate": 4.178217821782178e-06, "loss": 0.0278, "step": 4638 }, { "epoch": 18.33596837944664, "grad_norm": 0.09499368816614151, "learning_rate": 4.168316831683169e-06, "loss": 0.0204, "step": 4639 }, { "epoch": 18.3399209486166, "grad_norm": 0.21772192418575287, "learning_rate": 4.158415841584159e-06, "loss": 0.0425, "step": 4640 }, { "epoch": 18.343873517786562, "grad_norm": 0.13216857612133026, "learning_rate": 4.148514851485149e-06, "loss": 0.0286, "step": 4641 }, { "epoch": 18.347826086956523, "grad_norm": 0.0779455229640007, "learning_rate": 4.138613861386139e-06, "loss": 0.0173, "step": 4642 }, { "epoch": 18.35177865612648, "grad_norm": 0.08620145916938782, "learning_rate": 4.1287128712871295e-06, "loss": 0.0183, "step": 4643 }, { "epoch": 18.35573122529644, "grad_norm": 0.0964633896946907, "learning_rate": 4.118811881188119e-06, "loss": 0.0197, "step": 4644 }, { "epoch": 18.359683794466402, "grad_norm": 0.08035824447870255, "learning_rate": 4.108910891089109e-06, "loss": 0.0182, "step": 4645 }, { "epoch": 18.363636363636363, "grad_norm": 0.24137753248214722, "learning_rate": 4.099009900990099e-06, "loss": 0.0498, "step": 4646 }, { "epoch": 18.367588932806324, "grad_norm": 0.049549538642168045, "learning_rate": 4.0891089108910895e-06, "loss": 0.0106, "step": 4647 }, { "epoch": 18.371541501976285, "grad_norm": 0.07585926353931427, "learning_rate": 4.07920792079208e-06, "loss": 0.0162, "step": 4648 }, { "epoch": 18.375494071146246, "grad_norm": 0.6094945669174194, "learning_rate": 4.06930693069307e-06, "loss": 0.0165, "step": 4649 }, { "epoch": 18.379446640316207, "grad_norm": 0.06740503758192062, "learning_rate": 4.059405940594059e-06, "loss": 0.0124, "step": 4650 }, { "epoch": 18.383399209486164, "grad_norm": 0.17695990204811096, "learning_rate": 4.0495049504950496e-06, "loss": 0.036, "step": 4651 }, { "epoch": 18.387351778656125, "grad_norm": 0.11498956382274628, "learning_rate": 4.03960396039604e-06, "loss": 0.0255, "step": 4652 }, { "epoch": 18.391304347826086, "grad_norm": 0.08927769213914871, "learning_rate": 4.02970297029703e-06, "loss": 0.0197, "step": 4653 }, { "epoch": 18.395256916996047, "grad_norm": 0.05139454826712608, "learning_rate": 4.01980198019802e-06, "loss": 0.0117, "step": 4654 }, { "epoch": 18.399209486166008, "grad_norm": 0.09289796650409698, "learning_rate": 4.0099009900990104e-06, "loss": 0.0207, "step": 4655 }, { "epoch": 18.40316205533597, "grad_norm": 0.05939631536602974, "learning_rate": 4.000000000000001e-06, "loss": 0.0143, "step": 4656 }, { "epoch": 18.40711462450593, "grad_norm": 0.09954679012298584, "learning_rate": 3.99009900990099e-06, "loss": 0.0211, "step": 4657 }, { "epoch": 18.41106719367589, "grad_norm": 0.08324762433767319, "learning_rate": 3.98019801980198e-06, "loss": 0.0181, "step": 4658 }, { "epoch": 18.41501976284585, "grad_norm": 0.060197535902261734, "learning_rate": 3.9702970297029705e-06, "loss": 0.0132, "step": 4659 }, { "epoch": 18.41897233201581, "grad_norm": 0.12450237572193146, "learning_rate": 3.960396039603961e-06, "loss": 0.026, "step": 4660 }, { "epoch": 18.42292490118577, "grad_norm": 0.07680385559797287, "learning_rate": 3.950495049504951e-06, "loss": 0.0164, "step": 4661 }, { "epoch": 18.42687747035573, "grad_norm": 0.0747704952955246, "learning_rate": 3.940594059405941e-06, "loss": 0.016, "step": 4662 }, { "epoch": 18.43083003952569, "grad_norm": 0.10421628504991531, "learning_rate": 3.9306930693069305e-06, "loss": 0.0233, "step": 4663 }, { "epoch": 18.434782608695652, "grad_norm": 0.06608473509550095, "learning_rate": 3.920792079207921e-06, "loss": 0.0132, "step": 4664 }, { "epoch": 18.438735177865613, "grad_norm": 0.08618957549333572, "learning_rate": 3.910891089108911e-06, "loss": 0.0184, "step": 4665 }, { "epoch": 18.442687747035574, "grad_norm": 0.24050664901733398, "learning_rate": 3.900990099009901e-06, "loss": 0.0556, "step": 4666 }, { "epoch": 18.446640316205535, "grad_norm": 0.13225236535072327, "learning_rate": 3.891089108910891e-06, "loss": 0.0259, "step": 4667 }, { "epoch": 18.450592885375492, "grad_norm": 0.0973949059844017, "learning_rate": 3.881188118811882e-06, "loss": 0.0215, "step": 4668 }, { "epoch": 18.454545454545453, "grad_norm": 0.09168301522731781, "learning_rate": 3.871287128712872e-06, "loss": 0.0193, "step": 4669 }, { "epoch": 18.458498023715414, "grad_norm": 0.4083468019962311, "learning_rate": 3.861386138613861e-06, "loss": 0.0205, "step": 4670 }, { "epoch": 18.462450592885375, "grad_norm": 0.14040742814540863, "learning_rate": 3.851485148514851e-06, "loss": 0.0329, "step": 4671 }, { "epoch": 18.466403162055336, "grad_norm": 0.07015987485647202, "learning_rate": 3.841584158415842e-06, "loss": 0.0155, "step": 4672 }, { "epoch": 18.470355731225297, "grad_norm": 0.11282292753458023, "learning_rate": 3.831683168316832e-06, "loss": 0.0254, "step": 4673 }, { "epoch": 18.474308300395258, "grad_norm": 0.09705977141857147, "learning_rate": 3.821782178217822e-06, "loss": 0.0217, "step": 4674 }, { "epoch": 18.47826086956522, "grad_norm": 0.10362553596496582, "learning_rate": 3.811881188118812e-06, "loss": 0.022, "step": 4675 }, { "epoch": 18.48221343873518, "grad_norm": 0.10510674118995667, "learning_rate": 3.801980198019802e-06, "loss": 0.0223, "step": 4676 }, { "epoch": 18.486166007905137, "grad_norm": 0.0604848749935627, "learning_rate": 3.7920792079207923e-06, "loss": 0.0126, "step": 4677 }, { "epoch": 18.490118577075098, "grad_norm": 0.06301448494195938, "learning_rate": 3.782178217821782e-06, "loss": 0.0142, "step": 4678 }, { "epoch": 18.49407114624506, "grad_norm": 0.1129738837480545, "learning_rate": 3.7722772277227723e-06, "loss": 0.0257, "step": 4679 }, { "epoch": 18.49802371541502, "grad_norm": 0.0792364701628685, "learning_rate": 3.7623762376237625e-06, "loss": 0.0159, "step": 4680 }, { "epoch": 18.50197628458498, "grad_norm": 0.1397581845521927, "learning_rate": 3.7524752475247523e-06, "loss": 0.0298, "step": 4681 }, { "epoch": 18.50592885375494, "grad_norm": 0.07535760849714279, "learning_rate": 3.7425742574257425e-06, "loss": 0.0163, "step": 4682 }, { "epoch": 18.509881422924902, "grad_norm": 0.0925644263625145, "learning_rate": 3.7326732673267328e-06, "loss": 0.0191, "step": 4683 }, { "epoch": 18.513833992094863, "grad_norm": 0.1408582478761673, "learning_rate": 3.722772277227723e-06, "loss": 0.0276, "step": 4684 }, { "epoch": 18.51778656126482, "grad_norm": 0.1671958714723587, "learning_rate": 3.7128712871287128e-06, "loss": 0.0359, "step": 4685 }, { "epoch": 18.52173913043478, "grad_norm": 0.12372199445962906, "learning_rate": 3.702970297029703e-06, "loss": 0.0261, "step": 4686 }, { "epoch": 18.525691699604742, "grad_norm": 0.06320289522409439, "learning_rate": 3.6930693069306932e-06, "loss": 0.0144, "step": 4687 }, { "epoch": 18.529644268774703, "grad_norm": 0.08539468050003052, "learning_rate": 3.683168316831683e-06, "loss": 0.019, "step": 4688 }, { "epoch": 18.533596837944664, "grad_norm": 0.09643939137458801, "learning_rate": 3.6732673267326732e-06, "loss": 0.0205, "step": 4689 }, { "epoch": 18.537549407114625, "grad_norm": 0.10043076425790787, "learning_rate": 3.6633663366336635e-06, "loss": 0.0218, "step": 4690 }, { "epoch": 18.541501976284586, "grad_norm": 0.06785158812999725, "learning_rate": 3.6534653465346532e-06, "loss": 0.0146, "step": 4691 }, { "epoch": 18.545454545454547, "grad_norm": 0.0947439894080162, "learning_rate": 3.6435643564356435e-06, "loss": 0.0194, "step": 4692 }, { "epoch": 18.549407114624508, "grad_norm": 0.0953734815120697, "learning_rate": 3.6336633663366337e-06, "loss": 0.0206, "step": 4693 }, { "epoch": 18.553359683794465, "grad_norm": 0.14941661059856415, "learning_rate": 3.6237623762376235e-06, "loss": 0.0302, "step": 4694 }, { "epoch": 18.557312252964426, "grad_norm": 0.06193288788199425, "learning_rate": 3.6138613861386137e-06, "loss": 0.0133, "step": 4695 }, { "epoch": 18.561264822134387, "grad_norm": 0.08040556311607361, "learning_rate": 3.603960396039604e-06, "loss": 0.0164, "step": 4696 }, { "epoch": 18.565217391304348, "grad_norm": 0.08464600145816803, "learning_rate": 3.594059405940594e-06, "loss": 0.0173, "step": 4697 }, { "epoch": 18.56916996047431, "grad_norm": 0.11398971825838089, "learning_rate": 3.584158415841584e-06, "loss": 0.0238, "step": 4698 }, { "epoch": 18.57312252964427, "grad_norm": 0.07481592148542404, "learning_rate": 3.574257425742574e-06, "loss": 0.0139, "step": 4699 }, { "epoch": 18.57707509881423, "grad_norm": 0.14174163341522217, "learning_rate": 3.5643564356435644e-06, "loss": 0.0289, "step": 4700 }, { "epoch": 18.58102766798419, "grad_norm": 0.06147484481334686, "learning_rate": 3.554455445544554e-06, "loss": 0.0132, "step": 4701 }, { "epoch": 18.58498023715415, "grad_norm": 0.057895537465810776, "learning_rate": 3.5445544554455444e-06, "loss": 0.0134, "step": 4702 }, { "epoch": 18.58893280632411, "grad_norm": 0.10581756383180618, "learning_rate": 3.5346534653465346e-06, "loss": 0.0223, "step": 4703 }, { "epoch": 18.59288537549407, "grad_norm": 0.06617599725723267, "learning_rate": 3.5247524752475244e-06, "loss": 0.0142, "step": 4704 }, { "epoch": 18.59683794466403, "grad_norm": 0.065773144364357, "learning_rate": 3.5148514851485155e-06, "loss": 0.0139, "step": 4705 }, { "epoch": 18.600790513833992, "grad_norm": 0.11534105986356735, "learning_rate": 3.5049504950495053e-06, "loss": 0.0253, "step": 4706 }, { "epoch": 18.604743083003953, "grad_norm": 0.06479186564683914, "learning_rate": 3.4950495049504955e-06, "loss": 0.0143, "step": 4707 }, { "epoch": 18.608695652173914, "grad_norm": 0.07274450361728668, "learning_rate": 3.4851485148514857e-06, "loss": 0.0146, "step": 4708 }, { "epoch": 18.612648221343875, "grad_norm": 0.09270237386226654, "learning_rate": 3.475247524752476e-06, "loss": 0.021, "step": 4709 }, { "epoch": 18.616600790513836, "grad_norm": 0.10195731371641159, "learning_rate": 3.4653465346534657e-06, "loss": 0.0208, "step": 4710 }, { "epoch": 18.620553359683793, "grad_norm": 0.08300893753767014, "learning_rate": 3.455445544554456e-06, "loss": 0.0176, "step": 4711 }, { "epoch": 18.624505928853754, "grad_norm": 0.1667424440383911, "learning_rate": 3.445544554455446e-06, "loss": 0.0363, "step": 4712 }, { "epoch": 18.628458498023715, "grad_norm": 0.11803440749645233, "learning_rate": 3.435643564356436e-06, "loss": 0.0233, "step": 4713 }, { "epoch": 18.632411067193676, "grad_norm": 0.08449536561965942, "learning_rate": 3.425742574257426e-06, "loss": 0.0164, "step": 4714 }, { "epoch": 18.636363636363637, "grad_norm": 0.06951592117547989, "learning_rate": 3.4158415841584164e-06, "loss": 0.0142, "step": 4715 }, { "epoch": 18.640316205533598, "grad_norm": 0.10341779887676239, "learning_rate": 3.405940594059406e-06, "loss": 0.0214, "step": 4716 }, { "epoch": 18.64426877470356, "grad_norm": 0.04906265437602997, "learning_rate": 3.3960396039603964e-06, "loss": 0.0094, "step": 4717 }, { "epoch": 18.64822134387352, "grad_norm": 0.11896172165870667, "learning_rate": 3.3861386138613866e-06, "loss": 0.0247, "step": 4718 }, { "epoch": 18.652173913043477, "grad_norm": 0.09158039093017578, "learning_rate": 3.3762376237623764e-06, "loss": 0.0187, "step": 4719 }, { "epoch": 18.656126482213438, "grad_norm": 0.10068244487047195, "learning_rate": 3.3663366336633666e-06, "loss": 0.0205, "step": 4720 }, { "epoch": 18.6600790513834, "grad_norm": 0.058893926441669464, "learning_rate": 3.356435643564357e-06, "loss": 0.0122, "step": 4721 }, { "epoch": 18.66403162055336, "grad_norm": 0.08914996683597565, "learning_rate": 3.346534653465347e-06, "loss": 0.0193, "step": 4722 }, { "epoch": 18.66798418972332, "grad_norm": 0.05999904125928879, "learning_rate": 3.336633663366337e-06, "loss": 0.0134, "step": 4723 }, { "epoch": 18.67193675889328, "grad_norm": 0.14326529204845428, "learning_rate": 3.326732673267327e-06, "loss": 0.0322, "step": 4724 }, { "epoch": 18.675889328063242, "grad_norm": 0.0890645980834961, "learning_rate": 3.3168316831683173e-06, "loss": 0.0195, "step": 4725 }, { "epoch": 18.679841897233203, "grad_norm": 0.07452435046434402, "learning_rate": 3.306930693069307e-06, "loss": 0.0166, "step": 4726 }, { "epoch": 18.683794466403164, "grad_norm": 0.08744929730892181, "learning_rate": 3.2970297029702973e-06, "loss": 0.0188, "step": 4727 }, { "epoch": 18.68774703557312, "grad_norm": 0.08316735178232193, "learning_rate": 3.2871287128712875e-06, "loss": 0.0138, "step": 4728 }, { "epoch": 18.691699604743082, "grad_norm": 0.06520634889602661, "learning_rate": 3.2772277227722773e-06, "loss": 0.015, "step": 4729 }, { "epoch": 18.695652173913043, "grad_norm": 0.06126917526125908, "learning_rate": 3.2673267326732676e-06, "loss": 0.0125, "step": 4730 }, { "epoch": 18.699604743083004, "grad_norm": 0.07390894740819931, "learning_rate": 3.2574257425742578e-06, "loss": 0.0158, "step": 4731 }, { "epoch": 18.703557312252965, "grad_norm": 0.11460433155298233, "learning_rate": 3.247524752475248e-06, "loss": 0.0249, "step": 4732 }, { "epoch": 18.707509881422926, "grad_norm": 0.11062317341566086, "learning_rate": 3.2376237623762378e-06, "loss": 0.024, "step": 4733 }, { "epoch": 18.711462450592887, "grad_norm": 0.0952666699886322, "learning_rate": 3.227722772277228e-06, "loss": 0.0198, "step": 4734 }, { "epoch": 18.715415019762847, "grad_norm": 0.13101255893707275, "learning_rate": 3.2178217821782182e-06, "loss": 0.0295, "step": 4735 }, { "epoch": 18.719367588932805, "grad_norm": 0.10590437799692154, "learning_rate": 3.207920792079208e-06, "loss": 0.023, "step": 4736 }, { "epoch": 18.723320158102766, "grad_norm": 0.11056187003850937, "learning_rate": 3.1980198019801982e-06, "loss": 0.0234, "step": 4737 }, { "epoch": 18.727272727272727, "grad_norm": 0.07427699118852615, "learning_rate": 3.1881188118811885e-06, "loss": 0.0152, "step": 4738 }, { "epoch": 18.731225296442688, "grad_norm": 0.09998095035552979, "learning_rate": 3.1782178217821783e-06, "loss": 0.0212, "step": 4739 }, { "epoch": 18.73517786561265, "grad_norm": 0.09469113498926163, "learning_rate": 3.1683168316831685e-06, "loss": 0.0206, "step": 4740 }, { "epoch": 18.73913043478261, "grad_norm": 0.17773567140102386, "learning_rate": 3.1584158415841587e-06, "loss": 0.0368, "step": 4741 }, { "epoch": 18.74308300395257, "grad_norm": 0.08677980303764343, "learning_rate": 3.1485148514851485e-06, "loss": 0.0184, "step": 4742 }, { "epoch": 18.74703557312253, "grad_norm": 0.11605938524007797, "learning_rate": 3.1386138613861387e-06, "loss": 0.0239, "step": 4743 }, { "epoch": 18.75098814229249, "grad_norm": 0.06622570753097534, "learning_rate": 3.128712871287129e-06, "loss": 0.0133, "step": 4744 }, { "epoch": 18.75494071146245, "grad_norm": 0.09067898243665695, "learning_rate": 3.118811881188119e-06, "loss": 0.0191, "step": 4745 }, { "epoch": 18.75889328063241, "grad_norm": 0.05877280235290527, "learning_rate": 3.108910891089109e-06, "loss": 0.0131, "step": 4746 }, { "epoch": 18.76284584980237, "grad_norm": 0.06578218936920166, "learning_rate": 3.099009900990099e-06, "loss": 0.0142, "step": 4747 }, { "epoch": 18.766798418972332, "grad_norm": 0.06463664770126343, "learning_rate": 3.0891089108910894e-06, "loss": 0.0139, "step": 4748 }, { "epoch": 18.770750988142293, "grad_norm": 0.0763903260231018, "learning_rate": 3.079207920792079e-06, "loss": 0.0159, "step": 4749 }, { "epoch": 18.774703557312254, "grad_norm": 0.1409105360507965, "learning_rate": 3.0693069306930694e-06, "loss": 0.0157, "step": 4750 }, { "epoch": 18.778656126482215, "grad_norm": 0.12925446033477783, "learning_rate": 3.0594059405940596e-06, "loss": 0.0295, "step": 4751 }, { "epoch": 18.782608695652176, "grad_norm": 0.12830856442451477, "learning_rate": 3.0495049504950494e-06, "loss": 0.0261, "step": 4752 }, { "epoch": 18.786561264822133, "grad_norm": 0.07950760424137115, "learning_rate": 3.0396039603960396e-06, "loss": 0.0165, "step": 4753 }, { "epoch": 18.790513833992094, "grad_norm": 0.08831553906202316, "learning_rate": 3.02970297029703e-06, "loss": 0.0187, "step": 4754 }, { "epoch": 18.794466403162055, "grad_norm": 0.05881474167108536, "learning_rate": 3.01980198019802e-06, "loss": 0.0123, "step": 4755 }, { "epoch": 18.798418972332016, "grad_norm": 0.08305840939283371, "learning_rate": 3.00990099009901e-06, "loss": 0.0162, "step": 4756 }, { "epoch": 18.802371541501977, "grad_norm": 0.10726728290319443, "learning_rate": 3e-06, "loss": 0.022, "step": 4757 }, { "epoch": 18.806324110671937, "grad_norm": 0.17576538026332855, "learning_rate": 2.9900990099009903e-06, "loss": 0.0367, "step": 4758 }, { "epoch": 18.8102766798419, "grad_norm": 0.07591678202152252, "learning_rate": 2.98019801980198e-06, "loss": 0.0158, "step": 4759 }, { "epoch": 18.81422924901186, "grad_norm": 0.06758023053407669, "learning_rate": 2.9702970297029703e-06, "loss": 0.0145, "step": 4760 }, { "epoch": 18.818181818181817, "grad_norm": 0.09827972203493118, "learning_rate": 2.9603960396039605e-06, "loss": 0.0207, "step": 4761 }, { "epoch": 18.822134387351777, "grad_norm": 0.16200478374958038, "learning_rate": 2.9504950495049503e-06, "loss": 0.0351, "step": 4762 }, { "epoch": 18.82608695652174, "grad_norm": 0.10513581335544586, "learning_rate": 2.9405940594059405e-06, "loss": 0.0225, "step": 4763 }, { "epoch": 18.8300395256917, "grad_norm": 0.11619819700717926, "learning_rate": 2.9306930693069308e-06, "loss": 0.0221, "step": 4764 }, { "epoch": 18.83399209486166, "grad_norm": 0.08046438544988632, "learning_rate": 2.920792079207921e-06, "loss": 0.0177, "step": 4765 }, { "epoch": 18.83794466403162, "grad_norm": 0.08606404066085815, "learning_rate": 2.910891089108911e-06, "loss": 0.0201, "step": 4766 }, { "epoch": 18.841897233201582, "grad_norm": 0.06685566902160645, "learning_rate": 2.9009900990099014e-06, "loss": 0.0138, "step": 4767 }, { "epoch": 18.845849802371543, "grad_norm": 0.11893579363822937, "learning_rate": 2.8910891089108912e-06, "loss": 0.026, "step": 4768 }, { "epoch": 18.8498023715415, "grad_norm": 0.11395417153835297, "learning_rate": 2.8811881188118814e-06, "loss": 0.023, "step": 4769 }, { "epoch": 18.85375494071146, "grad_norm": 0.08299757540225983, "learning_rate": 2.8712871287128717e-06, "loss": 0.0171, "step": 4770 }, { "epoch": 18.857707509881422, "grad_norm": 0.08622147887945175, "learning_rate": 2.8613861386138614e-06, "loss": 0.0192, "step": 4771 }, { "epoch": 18.861660079051383, "grad_norm": 0.14935800433158875, "learning_rate": 2.8514851485148517e-06, "loss": 0.0321, "step": 4772 }, { "epoch": 18.865612648221344, "grad_norm": 0.291666716337204, "learning_rate": 2.841584158415842e-06, "loss": 0.0162, "step": 4773 }, { "epoch": 18.869565217391305, "grad_norm": 0.24634121358394623, "learning_rate": 2.831683168316832e-06, "loss": 0.0192, "step": 4774 }, { "epoch": 18.873517786561266, "grad_norm": 0.09296803176403046, "learning_rate": 2.821782178217822e-06, "loss": 0.0184, "step": 4775 }, { "epoch": 18.877470355731226, "grad_norm": 0.06016135960817337, "learning_rate": 2.811881188118812e-06, "loss": 0.0135, "step": 4776 }, { "epoch": 18.881422924901187, "grad_norm": 0.08753107488155365, "learning_rate": 2.8019801980198023e-06, "loss": 0.0178, "step": 4777 }, { "epoch": 18.885375494071145, "grad_norm": 0.09958222508430481, "learning_rate": 2.792079207920792e-06, "loss": 0.0225, "step": 4778 }, { "epoch": 18.889328063241106, "grad_norm": 0.09058510512113571, "learning_rate": 2.7821782178217824e-06, "loss": 0.0189, "step": 4779 }, { "epoch": 18.893280632411066, "grad_norm": 0.057984981685876846, "learning_rate": 2.7722772277227726e-06, "loss": 0.0127, "step": 4780 }, { "epoch": 18.897233201581027, "grad_norm": 0.11154164373874664, "learning_rate": 2.7623762376237624e-06, "loss": 0.0237, "step": 4781 }, { "epoch": 18.90118577075099, "grad_norm": 0.09499694406986237, "learning_rate": 2.7524752475247526e-06, "loss": 0.0195, "step": 4782 }, { "epoch": 18.90513833992095, "grad_norm": 0.06522903591394424, "learning_rate": 2.742574257425743e-06, "loss": 0.0137, "step": 4783 }, { "epoch": 18.90909090909091, "grad_norm": 0.12736207246780396, "learning_rate": 2.7326732673267326e-06, "loss": 0.0266, "step": 4784 }, { "epoch": 18.91304347826087, "grad_norm": 0.0985836610198021, "learning_rate": 2.722772277227723e-06, "loss": 0.0223, "step": 4785 }, { "epoch": 18.91699604743083, "grad_norm": 0.08190957456827164, "learning_rate": 2.712871287128713e-06, "loss": 0.0174, "step": 4786 }, { "epoch": 18.92094861660079, "grad_norm": 0.08380056172609329, "learning_rate": 2.7029702970297033e-06, "loss": 0.0183, "step": 4787 }, { "epoch": 18.92490118577075, "grad_norm": 0.07738867402076721, "learning_rate": 2.693069306930693e-06, "loss": 0.0174, "step": 4788 }, { "epoch": 18.92885375494071, "grad_norm": 0.13254186511039734, "learning_rate": 2.6831683168316833e-06, "loss": 0.0281, "step": 4789 }, { "epoch": 18.932806324110672, "grad_norm": 0.07335705310106277, "learning_rate": 2.6732673267326735e-06, "loss": 0.0151, "step": 4790 }, { "epoch": 18.936758893280633, "grad_norm": 0.06914924085140228, "learning_rate": 2.6633663366336633e-06, "loss": 0.0154, "step": 4791 }, { "epoch": 18.940711462450594, "grad_norm": 0.07040832936763763, "learning_rate": 2.6534653465346535e-06, "loss": 0.0153, "step": 4792 }, { "epoch": 18.944664031620555, "grad_norm": 0.06452862173318863, "learning_rate": 2.6435643564356437e-06, "loss": 0.0134, "step": 4793 }, { "epoch": 18.948616600790515, "grad_norm": 0.060402970761060715, "learning_rate": 2.6336633663366335e-06, "loss": 0.0136, "step": 4794 }, { "epoch": 18.952569169960473, "grad_norm": 0.08574655652046204, "learning_rate": 2.6237623762376237e-06, "loss": 0.0187, "step": 4795 }, { "epoch": 18.956521739130434, "grad_norm": 0.08375293761491776, "learning_rate": 2.613861386138614e-06, "loss": 0.0168, "step": 4796 }, { "epoch": 18.960474308300395, "grad_norm": 0.07224112004041672, "learning_rate": 2.603960396039604e-06, "loss": 0.0151, "step": 4797 }, { "epoch": 18.964426877470355, "grad_norm": 0.07791998237371445, "learning_rate": 2.594059405940594e-06, "loss": 0.0175, "step": 4798 }, { "epoch": 18.968379446640316, "grad_norm": 0.07818736135959625, "learning_rate": 2.584158415841584e-06, "loss": 0.0167, "step": 4799 }, { "epoch": 18.972332015810277, "grad_norm": 0.24131862819194794, "learning_rate": 2.5742574257425744e-06, "loss": 0.0556, "step": 4800 }, { "epoch": 18.976284584980238, "grad_norm": 0.16252610087394714, "learning_rate": 2.564356435643564e-06, "loss": 0.0302, "step": 4801 }, { "epoch": 18.9802371541502, "grad_norm": 0.14663942158222198, "learning_rate": 2.5544554455445544e-06, "loss": 0.0307, "step": 4802 }, { "epoch": 18.984189723320156, "grad_norm": 0.09116839617490768, "learning_rate": 2.5445544554455446e-06, "loss": 0.0187, "step": 4803 }, { "epoch": 18.988142292490117, "grad_norm": 0.09319161623716354, "learning_rate": 2.534653465346535e-06, "loss": 0.0189, "step": 4804 }, { "epoch": 18.992094861660078, "grad_norm": 0.11148131638765335, "learning_rate": 2.524752475247525e-06, "loss": 0.0228, "step": 4805 }, { "epoch": 18.99604743083004, "grad_norm": 0.09920572489500046, "learning_rate": 2.5148514851485153e-06, "loss": 0.0207, "step": 4806 }, { "epoch": 19.0, "grad_norm": 0.09807518869638443, "learning_rate": 2.504950495049505e-06, "loss": 0.0221, "step": 4807 }, { "epoch": 19.00395256916996, "grad_norm": 0.08094090968370438, "learning_rate": 2.4950495049504953e-06, "loss": 0.0186, "step": 4808 }, { "epoch": 19.007905138339922, "grad_norm": 0.05945340543985367, "learning_rate": 2.4851485148514855e-06, "loss": 0.0127, "step": 4809 }, { "epoch": 19.011857707509883, "grad_norm": 0.051307082176208496, "learning_rate": 2.4752475247524753e-06, "loss": 0.0119, "step": 4810 }, { "epoch": 19.015810276679844, "grad_norm": 0.11299656331539154, "learning_rate": 2.4653465346534655e-06, "loss": 0.0261, "step": 4811 }, { "epoch": 19.0197628458498, "grad_norm": 0.09056013822555542, "learning_rate": 2.4554455445544558e-06, "loss": 0.0214, "step": 4812 }, { "epoch": 19.023715415019762, "grad_norm": 0.1103508397936821, "learning_rate": 2.4455445544554456e-06, "loss": 0.0242, "step": 4813 }, { "epoch": 19.027667984189723, "grad_norm": 0.07804236561059952, "learning_rate": 2.4356435643564358e-06, "loss": 0.0185, "step": 4814 }, { "epoch": 19.031620553359684, "grad_norm": 0.0793350338935852, "learning_rate": 2.425742574257426e-06, "loss": 0.0188, "step": 4815 }, { "epoch": 19.035573122529645, "grad_norm": 0.1370858997106552, "learning_rate": 2.4158415841584162e-06, "loss": 0.0324, "step": 4816 }, { "epoch": 19.039525691699605, "grad_norm": 0.06073697283864021, "learning_rate": 2.405940594059406e-06, "loss": 0.014, "step": 4817 }, { "epoch": 19.043478260869566, "grad_norm": 0.1068793386220932, "learning_rate": 2.3960396039603962e-06, "loss": 0.0237, "step": 4818 }, { "epoch": 19.047430830039527, "grad_norm": 0.11655991524457932, "learning_rate": 2.3861386138613865e-06, "loss": 0.0264, "step": 4819 }, { "epoch": 19.051383399209485, "grad_norm": 0.09777363389730453, "learning_rate": 2.3762376237623762e-06, "loss": 0.0242, "step": 4820 }, { "epoch": 19.055335968379445, "grad_norm": 0.1008680984377861, "learning_rate": 2.3663366336633665e-06, "loss": 0.0223, "step": 4821 }, { "epoch": 19.059288537549406, "grad_norm": 0.054144252091646194, "learning_rate": 2.3564356435643567e-06, "loss": 0.0127, "step": 4822 }, { "epoch": 19.063241106719367, "grad_norm": 0.16055144369602203, "learning_rate": 2.3465346534653465e-06, "loss": 0.0359, "step": 4823 }, { "epoch": 19.067193675889328, "grad_norm": 0.06130475923418999, "learning_rate": 2.3366336633663367e-06, "loss": 0.0142, "step": 4824 }, { "epoch": 19.07114624505929, "grad_norm": 0.09688978642225266, "learning_rate": 2.326732673267327e-06, "loss": 0.0214, "step": 4825 }, { "epoch": 19.07509881422925, "grad_norm": 0.08701872080564499, "learning_rate": 2.3168316831683167e-06, "loss": 0.0204, "step": 4826 }, { "epoch": 19.07905138339921, "grad_norm": 0.052399251610040665, "learning_rate": 2.306930693069307e-06, "loss": 0.0122, "step": 4827 }, { "epoch": 19.08300395256917, "grad_norm": 0.09825041890144348, "learning_rate": 2.297029702970297e-06, "loss": 0.0221, "step": 4828 }, { "epoch": 19.08695652173913, "grad_norm": 0.07572689652442932, "learning_rate": 2.2871287128712874e-06, "loss": 0.018, "step": 4829 }, { "epoch": 19.09090909090909, "grad_norm": 0.0840848758816719, "learning_rate": 2.277227722772277e-06, "loss": 0.0183, "step": 4830 }, { "epoch": 19.09486166007905, "grad_norm": 0.22514699399471283, "learning_rate": 2.2673267326732674e-06, "loss": 0.0476, "step": 4831 }, { "epoch": 19.09881422924901, "grad_norm": 0.06651589274406433, "learning_rate": 2.2574257425742576e-06, "loss": 0.0161, "step": 4832 }, { "epoch": 19.102766798418973, "grad_norm": 0.10062935203313828, "learning_rate": 2.2475247524752474e-06, "loss": 0.023, "step": 4833 }, { "epoch": 19.106719367588934, "grad_norm": 0.07120473682880402, "learning_rate": 2.2376237623762376e-06, "loss": 0.0168, "step": 4834 }, { "epoch": 19.110671936758894, "grad_norm": 0.11063100397586823, "learning_rate": 2.227722772277228e-06, "loss": 0.0252, "step": 4835 }, { "epoch": 19.114624505928855, "grad_norm": 0.06272239983081818, "learning_rate": 2.2178217821782176e-06, "loss": 0.0143, "step": 4836 }, { "epoch": 19.118577075098813, "grad_norm": 0.06803707033395767, "learning_rate": 2.207920792079208e-06, "loss": 0.0159, "step": 4837 }, { "epoch": 19.122529644268774, "grad_norm": 0.10538329184055328, "learning_rate": 2.198019801980198e-06, "loss": 0.0241, "step": 4838 }, { "epoch": 19.126482213438734, "grad_norm": 0.1018119677901268, "learning_rate": 2.1881188118811883e-06, "loss": 0.0235, "step": 4839 }, { "epoch": 19.130434782608695, "grad_norm": 0.06962017714977264, "learning_rate": 2.178217821782178e-06, "loss": 0.0157, "step": 4840 }, { "epoch": 19.134387351778656, "grad_norm": 0.08790461719036102, "learning_rate": 2.1683168316831683e-06, "loss": 0.0211, "step": 4841 }, { "epoch": 19.138339920948617, "grad_norm": 0.109166219830513, "learning_rate": 2.1584158415841585e-06, "loss": 0.0238, "step": 4842 }, { "epoch": 19.142292490118578, "grad_norm": 0.10021164268255234, "learning_rate": 2.1485148514851483e-06, "loss": 0.0214, "step": 4843 }, { "epoch": 19.14624505928854, "grad_norm": 0.05808226391673088, "learning_rate": 2.138613861386139e-06, "loss": 0.0132, "step": 4844 }, { "epoch": 19.150197628458496, "grad_norm": 0.11528687924146652, "learning_rate": 2.128712871287129e-06, "loss": 0.0249, "step": 4845 }, { "epoch": 19.154150197628457, "grad_norm": 0.11448343098163605, "learning_rate": 2.118811881188119e-06, "loss": 0.0233, "step": 4846 }, { "epoch": 19.158102766798418, "grad_norm": 0.04905841872096062, "learning_rate": 2.108910891089109e-06, "loss": 0.0108, "step": 4847 }, { "epoch": 19.16205533596838, "grad_norm": 0.0638711154460907, "learning_rate": 2.0990099009900994e-06, "loss": 0.0139, "step": 4848 }, { "epoch": 19.16600790513834, "grad_norm": 0.0787539929151535, "learning_rate": 2.089108910891089e-06, "loss": 0.0177, "step": 4849 }, { "epoch": 19.1699604743083, "grad_norm": 0.10402161628007889, "learning_rate": 2.0792079207920794e-06, "loss": 0.0241, "step": 4850 }, { "epoch": 19.17391304347826, "grad_norm": 0.0924585610628128, "learning_rate": 2.0693069306930696e-06, "loss": 0.0201, "step": 4851 }, { "epoch": 19.177865612648223, "grad_norm": 0.08653881400823593, "learning_rate": 2.0594059405940594e-06, "loss": 0.0179, "step": 4852 }, { "epoch": 19.181818181818183, "grad_norm": 0.13048668205738068, "learning_rate": 2.0495049504950497e-06, "loss": 0.0278, "step": 4853 }, { "epoch": 19.18577075098814, "grad_norm": 0.10192209482192993, "learning_rate": 2.03960396039604e-06, "loss": 0.0223, "step": 4854 }, { "epoch": 19.1897233201581, "grad_norm": 0.09663470834493637, "learning_rate": 2.0297029702970297e-06, "loss": 0.0208, "step": 4855 }, { "epoch": 19.193675889328063, "grad_norm": 0.08021564781665802, "learning_rate": 2.01980198019802e-06, "loss": 0.0178, "step": 4856 }, { "epoch": 19.197628458498023, "grad_norm": 0.05874083191156387, "learning_rate": 2.00990099009901e-06, "loss": 0.0127, "step": 4857 }, { "epoch": 19.201581027667984, "grad_norm": 0.06407742202281952, "learning_rate": 2.0000000000000003e-06, "loss": 0.0147, "step": 4858 }, { "epoch": 19.205533596837945, "grad_norm": 0.11026555299758911, "learning_rate": 1.99009900990099e-06, "loss": 0.0229, "step": 4859 }, { "epoch": 19.209486166007906, "grad_norm": 0.12622922658920288, "learning_rate": 1.9801980198019803e-06, "loss": 0.0262, "step": 4860 }, { "epoch": 19.213438735177867, "grad_norm": 0.16153490543365479, "learning_rate": 1.9702970297029706e-06, "loss": 0.0371, "step": 4861 }, { "epoch": 19.217391304347824, "grad_norm": 0.09624055027961731, "learning_rate": 1.9603960396039604e-06, "loss": 0.0215, "step": 4862 }, { "epoch": 19.221343873517785, "grad_norm": 0.08086435496807098, "learning_rate": 1.9504950495049506e-06, "loss": 0.017, "step": 4863 }, { "epoch": 19.225296442687746, "grad_norm": 0.06533783674240112, "learning_rate": 1.940594059405941e-06, "loss": 0.0148, "step": 4864 }, { "epoch": 19.229249011857707, "grad_norm": 0.09404882788658142, "learning_rate": 1.9306930693069306e-06, "loss": 0.0199, "step": 4865 }, { "epoch": 19.233201581027668, "grad_norm": 0.07474951446056366, "learning_rate": 1.920792079207921e-06, "loss": 0.0159, "step": 4866 }, { "epoch": 19.23715415019763, "grad_norm": 0.1022556871175766, "learning_rate": 1.910891089108911e-06, "loss": 0.022, "step": 4867 }, { "epoch": 19.24110671936759, "grad_norm": 0.07190510630607605, "learning_rate": 1.900990099009901e-06, "loss": 0.0148, "step": 4868 }, { "epoch": 19.24505928853755, "grad_norm": 0.07469405233860016, "learning_rate": 1.891089108910891e-06, "loss": 0.0147, "step": 4869 }, { "epoch": 19.24901185770751, "grad_norm": 0.07940872013568878, "learning_rate": 1.8811881188118813e-06, "loss": 0.0177, "step": 4870 }, { "epoch": 19.25296442687747, "grad_norm": 0.05883575603365898, "learning_rate": 1.8712871287128713e-06, "loss": 0.0132, "step": 4871 }, { "epoch": 19.25691699604743, "grad_norm": 0.1469162106513977, "learning_rate": 1.8613861386138615e-06, "loss": 0.0325, "step": 4872 }, { "epoch": 19.26086956521739, "grad_norm": 0.11239240318536758, "learning_rate": 1.8514851485148515e-06, "loss": 0.024, "step": 4873 }, { "epoch": 19.26482213438735, "grad_norm": 0.15634000301361084, "learning_rate": 1.8415841584158415e-06, "loss": 0.0353, "step": 4874 }, { "epoch": 19.268774703557312, "grad_norm": 0.08266416192054749, "learning_rate": 1.8316831683168317e-06, "loss": 0.0177, "step": 4875 }, { "epoch": 19.272727272727273, "grad_norm": 0.08713242411613464, "learning_rate": 1.8217821782178217e-06, "loss": 0.0187, "step": 4876 }, { "epoch": 19.276679841897234, "grad_norm": 0.05467486009001732, "learning_rate": 1.8118811881188117e-06, "loss": 0.0115, "step": 4877 }, { "epoch": 19.280632411067195, "grad_norm": 0.06904654204845428, "learning_rate": 1.801980198019802e-06, "loss": 0.0147, "step": 4878 }, { "epoch": 19.284584980237153, "grad_norm": 0.07986623048782349, "learning_rate": 1.792079207920792e-06, "loss": 0.0165, "step": 4879 }, { "epoch": 19.288537549407113, "grad_norm": 0.08735392987728119, "learning_rate": 1.7821782178217822e-06, "loss": 0.018, "step": 4880 }, { "epoch": 19.292490118577074, "grad_norm": 0.06086980924010277, "learning_rate": 1.7722772277227722e-06, "loss": 0.0132, "step": 4881 }, { "epoch": 19.296442687747035, "grad_norm": 0.07801414281129837, "learning_rate": 1.7623762376237622e-06, "loss": 0.0161, "step": 4882 }, { "epoch": 19.300395256916996, "grad_norm": 0.06905828416347504, "learning_rate": 1.7524752475247526e-06, "loss": 0.0134, "step": 4883 }, { "epoch": 19.304347826086957, "grad_norm": 0.1067027747631073, "learning_rate": 1.7425742574257428e-06, "loss": 0.0223, "step": 4884 }, { "epoch": 19.308300395256918, "grad_norm": 0.10176712274551392, "learning_rate": 1.7326732673267329e-06, "loss": 0.0232, "step": 4885 }, { "epoch": 19.31225296442688, "grad_norm": 0.07439107447862625, "learning_rate": 1.722772277227723e-06, "loss": 0.017, "step": 4886 }, { "epoch": 19.31620553359684, "grad_norm": 0.15840519964694977, "learning_rate": 1.712871287128713e-06, "loss": 0.0361, "step": 4887 }, { "epoch": 19.320158102766797, "grad_norm": 0.11561636626720428, "learning_rate": 1.702970297029703e-06, "loss": 0.0235, "step": 4888 }, { "epoch": 19.324110671936758, "grad_norm": 0.0623919814825058, "learning_rate": 1.6930693069306933e-06, "loss": 0.0142, "step": 4889 }, { "epoch": 19.32806324110672, "grad_norm": 0.11590003967285156, "learning_rate": 1.6831683168316833e-06, "loss": 0.025, "step": 4890 }, { "epoch": 19.33201581027668, "grad_norm": 0.08240262418985367, "learning_rate": 1.6732673267326735e-06, "loss": 0.018, "step": 4891 }, { "epoch": 19.33596837944664, "grad_norm": 0.07131791859865189, "learning_rate": 1.6633663366336635e-06, "loss": 0.0168, "step": 4892 }, { "epoch": 19.3399209486166, "grad_norm": 0.11021753400564194, "learning_rate": 1.6534653465346535e-06, "loss": 0.0218, "step": 4893 }, { "epoch": 19.343873517786562, "grad_norm": 0.09916914254426956, "learning_rate": 1.6435643564356438e-06, "loss": 0.0212, "step": 4894 }, { "epoch": 19.347826086956523, "grad_norm": 0.12380921840667725, "learning_rate": 1.6336633663366338e-06, "loss": 0.0264, "step": 4895 }, { "epoch": 19.35177865612648, "grad_norm": 0.12512126564979553, "learning_rate": 1.623762376237624e-06, "loss": 0.0256, "step": 4896 }, { "epoch": 19.35573122529644, "grad_norm": 0.10217534005641937, "learning_rate": 1.613861386138614e-06, "loss": 0.0219, "step": 4897 }, { "epoch": 19.359683794466402, "grad_norm": 0.08075512945652008, "learning_rate": 1.603960396039604e-06, "loss": 0.0166, "step": 4898 }, { "epoch": 19.363636363636363, "grad_norm": 0.10646136850118637, "learning_rate": 1.5940594059405942e-06, "loss": 0.023, "step": 4899 }, { "epoch": 19.367588932806324, "grad_norm": 0.11744405329227448, "learning_rate": 1.5841584158415842e-06, "loss": 0.0253, "step": 4900 }, { "epoch": 19.371541501976285, "grad_norm": 0.08526954054832458, "learning_rate": 1.5742574257425742e-06, "loss": 0.0158, "step": 4901 }, { "epoch": 19.375494071146246, "grad_norm": 0.08729571849107742, "learning_rate": 1.5643564356435645e-06, "loss": 0.0181, "step": 4902 }, { "epoch": 19.379446640316207, "grad_norm": 0.08164673298597336, "learning_rate": 1.5544554455445545e-06, "loss": 0.0169, "step": 4903 }, { "epoch": 19.383399209486164, "grad_norm": 0.048974767327308655, "learning_rate": 1.5445544554455447e-06, "loss": 0.0102, "step": 4904 }, { "epoch": 19.387351778656125, "grad_norm": 0.07859613746404648, "learning_rate": 1.5346534653465347e-06, "loss": 0.0176, "step": 4905 }, { "epoch": 19.391304347826086, "grad_norm": 0.10157660394906998, "learning_rate": 1.5247524752475247e-06, "loss": 0.022, "step": 4906 }, { "epoch": 19.395256916996047, "grad_norm": 0.05585305020213127, "learning_rate": 1.514851485148515e-06, "loss": 0.012, "step": 4907 }, { "epoch": 19.399209486166008, "grad_norm": 0.09481612592935562, "learning_rate": 1.504950495049505e-06, "loss": 0.0209, "step": 4908 }, { "epoch": 19.40316205533597, "grad_norm": 0.10926420986652374, "learning_rate": 1.4950495049504951e-06, "loss": 0.0206, "step": 4909 }, { "epoch": 19.40711462450593, "grad_norm": 0.0834321677684784, "learning_rate": 1.4851485148514852e-06, "loss": 0.0188, "step": 4910 }, { "epoch": 19.41106719367589, "grad_norm": 0.15589074790477753, "learning_rate": 1.4752475247524752e-06, "loss": 0.0365, "step": 4911 }, { "epoch": 19.41501976284585, "grad_norm": 0.0826529935002327, "learning_rate": 1.4653465346534654e-06, "loss": 0.018, "step": 4912 }, { "epoch": 19.41897233201581, "grad_norm": 0.07027612626552582, "learning_rate": 1.4554455445544556e-06, "loss": 0.0143, "step": 4913 }, { "epoch": 19.42292490118577, "grad_norm": 0.0828486979007721, "learning_rate": 1.4455445544554456e-06, "loss": 0.0176, "step": 4914 }, { "epoch": 19.42687747035573, "grad_norm": 0.09636640548706055, "learning_rate": 1.4356435643564358e-06, "loss": 0.0204, "step": 4915 }, { "epoch": 19.43083003952569, "grad_norm": 0.07195334881544113, "learning_rate": 1.4257425742574258e-06, "loss": 0.0165, "step": 4916 }, { "epoch": 19.434782608695652, "grad_norm": 0.08266804367303848, "learning_rate": 1.415841584158416e-06, "loss": 0.0175, "step": 4917 }, { "epoch": 19.438735177865613, "grad_norm": 0.08565854281187057, "learning_rate": 1.405940594059406e-06, "loss": 0.0171, "step": 4918 }, { "epoch": 19.442687747035574, "grad_norm": 0.1739148497581482, "learning_rate": 1.396039603960396e-06, "loss": 0.0331, "step": 4919 }, { "epoch": 19.446640316205535, "grad_norm": 0.0879354253411293, "learning_rate": 1.3861386138613863e-06, "loss": 0.0188, "step": 4920 }, { "epoch": 19.450592885375492, "grad_norm": 0.09928828477859497, "learning_rate": 1.3762376237623763e-06, "loss": 0.0217, "step": 4921 }, { "epoch": 19.454545454545453, "grad_norm": 0.07251608371734619, "learning_rate": 1.3663366336633663e-06, "loss": 0.0162, "step": 4922 }, { "epoch": 19.458498023715414, "grad_norm": 0.0921507477760315, "learning_rate": 1.3564356435643565e-06, "loss": 0.0193, "step": 4923 }, { "epoch": 19.462450592885375, "grad_norm": 0.09153872728347778, "learning_rate": 1.3465346534653465e-06, "loss": 0.0202, "step": 4924 }, { "epoch": 19.466403162055336, "grad_norm": 0.06729371100664139, "learning_rate": 1.3366336633663367e-06, "loss": 0.0144, "step": 4925 }, { "epoch": 19.470355731225297, "grad_norm": 0.22581547498703003, "learning_rate": 1.3267326732673268e-06, "loss": 0.045, "step": 4926 }, { "epoch": 19.474308300395258, "grad_norm": 0.08230797201395035, "learning_rate": 1.3168316831683168e-06, "loss": 0.0174, "step": 4927 }, { "epoch": 19.47826086956522, "grad_norm": 0.11937136948108673, "learning_rate": 1.306930693069307e-06, "loss": 0.0246, "step": 4928 }, { "epoch": 19.48221343873518, "grad_norm": 0.15270064771175385, "learning_rate": 1.297029702970297e-06, "loss": 0.0301, "step": 4929 }, { "epoch": 19.486166007905137, "grad_norm": 0.08668936789035797, "learning_rate": 1.2871287128712872e-06, "loss": 0.0159, "step": 4930 }, { "epoch": 19.490118577075098, "grad_norm": 0.09638085216283798, "learning_rate": 1.2772277227722772e-06, "loss": 0.0202, "step": 4931 }, { "epoch": 19.49407114624506, "grad_norm": 0.08451702445745468, "learning_rate": 1.2673267326732674e-06, "loss": 0.0184, "step": 4932 }, { "epoch": 19.49802371541502, "grad_norm": 0.10864075273275375, "learning_rate": 1.2574257425742576e-06, "loss": 0.022, "step": 4933 }, { "epoch": 19.50197628458498, "grad_norm": 0.1374835968017578, "learning_rate": 1.2475247524752477e-06, "loss": 0.0275, "step": 4934 }, { "epoch": 19.50592885375494, "grad_norm": 0.06336801499128342, "learning_rate": 1.2376237623762377e-06, "loss": 0.0136, "step": 4935 }, { "epoch": 19.509881422924902, "grad_norm": 0.08489874750375748, "learning_rate": 1.2277227722772279e-06, "loss": 0.0184, "step": 4936 }, { "epoch": 19.513833992094863, "grad_norm": 0.07670220732688904, "learning_rate": 1.2178217821782179e-06, "loss": 0.0156, "step": 4937 }, { "epoch": 19.51778656126482, "grad_norm": 0.052467335015535355, "learning_rate": 1.2079207920792081e-06, "loss": 0.0114, "step": 4938 }, { "epoch": 19.52173913043478, "grad_norm": 0.14413702487945557, "learning_rate": 1.1980198019801981e-06, "loss": 0.0301, "step": 4939 }, { "epoch": 19.525691699604742, "grad_norm": 0.06137281656265259, "learning_rate": 1.1881188118811881e-06, "loss": 0.0132, "step": 4940 }, { "epoch": 19.529644268774703, "grad_norm": 0.09480182081460953, "learning_rate": 1.1782178217821783e-06, "loss": 0.0195, "step": 4941 }, { "epoch": 19.533596837944664, "grad_norm": 0.0715889185667038, "learning_rate": 1.1683168316831683e-06, "loss": 0.0147, "step": 4942 }, { "epoch": 19.537549407114625, "grad_norm": 0.10746315866708755, "learning_rate": 1.1584158415841584e-06, "loss": 0.0228, "step": 4943 }, { "epoch": 19.541501976284586, "grad_norm": 0.05719698593020439, "learning_rate": 1.1485148514851486e-06, "loss": 0.0118, "step": 4944 }, { "epoch": 19.545454545454547, "grad_norm": 0.10665254294872284, "learning_rate": 1.1386138613861386e-06, "loss": 0.0202, "step": 4945 }, { "epoch": 19.549407114624508, "grad_norm": 0.08776424825191498, "learning_rate": 1.1287128712871288e-06, "loss": 0.0193, "step": 4946 }, { "epoch": 19.553359683794465, "grad_norm": 0.08800166100263596, "learning_rate": 1.1188118811881188e-06, "loss": 0.0178, "step": 4947 }, { "epoch": 19.557312252964426, "grad_norm": 0.07843968272209167, "learning_rate": 1.1089108910891088e-06, "loss": 0.0166, "step": 4948 }, { "epoch": 19.561264822134387, "grad_norm": 0.07594018429517746, "learning_rate": 1.099009900990099e-06, "loss": 0.0157, "step": 4949 }, { "epoch": 19.565217391304348, "grad_norm": 0.09111127257347107, "learning_rate": 1.089108910891089e-06, "loss": 0.0176, "step": 4950 }, { "epoch": 19.56916996047431, "grad_norm": 0.06790529191493988, "learning_rate": 1.0792079207920793e-06, "loss": 0.0141, "step": 4951 }, { "epoch": 19.57312252964427, "grad_norm": 0.09833591431379318, "learning_rate": 1.0693069306930695e-06, "loss": 0.0213, "step": 4952 }, { "epoch": 19.57707509881423, "grad_norm": 0.08010268956422806, "learning_rate": 1.0594059405940595e-06, "loss": 0.0164, "step": 4953 }, { "epoch": 19.58102766798419, "grad_norm": 0.0664481520652771, "learning_rate": 1.0495049504950497e-06, "loss": 0.0137, "step": 4954 }, { "epoch": 19.58498023715415, "grad_norm": 0.0900559201836586, "learning_rate": 1.0396039603960397e-06, "loss": 0.0186, "step": 4955 }, { "epoch": 19.58893280632411, "grad_norm": 0.07506775856018066, "learning_rate": 1.0297029702970297e-06, "loss": 0.0151, "step": 4956 }, { "epoch": 19.59288537549407, "grad_norm": 0.07078597694635391, "learning_rate": 1.01980198019802e-06, "loss": 0.0134, "step": 4957 }, { "epoch": 19.59683794466403, "grad_norm": 0.06190524250268936, "learning_rate": 1.00990099009901e-06, "loss": 0.0124, "step": 4958 }, { "epoch": 19.600790513833992, "grad_norm": 0.09289881587028503, "learning_rate": 1.0000000000000002e-06, "loss": 0.0198, "step": 4959 }, { "epoch": 19.604743083003953, "grad_norm": 0.08702456206083298, "learning_rate": 9.900990099009902e-07, "loss": 0.0183, "step": 4960 }, { "epoch": 19.608695652173914, "grad_norm": 0.07483638823032379, "learning_rate": 9.801980198019802e-07, "loss": 0.0146, "step": 4961 }, { "epoch": 19.612648221343875, "grad_norm": 0.09252170473337173, "learning_rate": 9.702970297029704e-07, "loss": 0.0189, "step": 4962 }, { "epoch": 19.616600790513836, "grad_norm": 0.07708915323019028, "learning_rate": 9.603960396039604e-07, "loss": 0.0151, "step": 4963 }, { "epoch": 19.620553359683793, "grad_norm": 0.07711998373270035, "learning_rate": 9.504950495049505e-07, "loss": 0.0152, "step": 4964 }, { "epoch": 19.624505928853754, "grad_norm": 0.05478300899267197, "learning_rate": 9.405940594059406e-07, "loss": 0.0115, "step": 4965 }, { "epoch": 19.628458498023715, "grad_norm": 0.07956750690937042, "learning_rate": 9.306930693069307e-07, "loss": 0.0161, "step": 4966 }, { "epoch": 19.632411067193676, "grad_norm": 0.08878491818904877, "learning_rate": 9.207920792079208e-07, "loss": 0.0198, "step": 4967 }, { "epoch": 19.636363636363637, "grad_norm": 0.09108737111091614, "learning_rate": 9.108910891089109e-07, "loss": 0.0204, "step": 4968 }, { "epoch": 19.640316205533598, "grad_norm": 0.1006757989525795, "learning_rate": 9.00990099009901e-07, "loss": 0.0206, "step": 4969 }, { "epoch": 19.64426877470356, "grad_norm": 0.11767980456352234, "learning_rate": 8.910891089108911e-07, "loss": 0.024, "step": 4970 }, { "epoch": 19.64822134387352, "grad_norm": 0.05596867576241493, "learning_rate": 8.811881188118811e-07, "loss": 0.0117, "step": 4971 }, { "epoch": 19.652173913043477, "grad_norm": 0.09111103415489197, "learning_rate": 8.712871287128714e-07, "loss": 0.0198, "step": 4972 }, { "epoch": 19.656126482213438, "grad_norm": 0.08109857887029648, "learning_rate": 8.613861386138615e-07, "loss": 0.0172, "step": 4973 }, { "epoch": 19.6600790513834, "grad_norm": 0.0911722183227539, "learning_rate": 8.514851485148515e-07, "loss": 0.0192, "step": 4974 }, { "epoch": 19.66403162055336, "grad_norm": 0.10958622395992279, "learning_rate": 8.415841584158417e-07, "loss": 0.0227, "step": 4975 }, { "epoch": 19.66798418972332, "grad_norm": 0.10069285333156586, "learning_rate": 8.316831683168318e-07, "loss": 0.0216, "step": 4976 }, { "epoch": 19.67193675889328, "grad_norm": 0.07060444355010986, "learning_rate": 8.217821782178219e-07, "loss": 0.0144, "step": 4977 }, { "epoch": 19.675889328063242, "grad_norm": 0.26114872097969055, "learning_rate": 8.11881188118812e-07, "loss": 0.052, "step": 4978 }, { "epoch": 19.679841897233203, "grad_norm": 0.0888320729136467, "learning_rate": 8.01980198019802e-07, "loss": 0.0196, "step": 4979 }, { "epoch": 19.683794466403164, "grad_norm": 0.11444205790758133, "learning_rate": 7.920792079207921e-07, "loss": 0.026, "step": 4980 }, { "epoch": 19.68774703557312, "grad_norm": 0.06368529051542282, "learning_rate": 7.821782178217822e-07, "loss": 0.0138, "step": 4981 }, { "epoch": 19.691699604743082, "grad_norm": 0.07329872250556946, "learning_rate": 7.722772277227723e-07, "loss": 0.0146, "step": 4982 }, { "epoch": 19.695652173913043, "grad_norm": 0.09171003848314285, "learning_rate": 7.623762376237624e-07, "loss": 0.0192, "step": 4983 }, { "epoch": 19.699604743083004, "grad_norm": 0.09695214033126831, "learning_rate": 7.524752475247525e-07, "loss": 0.0202, "step": 4984 }, { "epoch": 19.703557312252965, "grad_norm": 0.1151820719242096, "learning_rate": 7.425742574257426e-07, "loss": 0.0251, "step": 4985 }, { "epoch": 19.707509881422926, "grad_norm": 0.11710719019174576, "learning_rate": 7.326732673267327e-07, "loss": 0.0233, "step": 4986 }, { "epoch": 19.711462450592887, "grad_norm": 0.0775756910443306, "learning_rate": 7.227722772277228e-07, "loss": 0.0175, "step": 4987 }, { "epoch": 19.715415019762847, "grad_norm": 0.07473313063383102, "learning_rate": 7.128712871287129e-07, "loss": 0.0151, "step": 4988 }, { "epoch": 19.719367588932805, "grad_norm": 0.05931631475687027, "learning_rate": 7.02970297029703e-07, "loss": 0.0131, "step": 4989 }, { "epoch": 19.723320158102766, "grad_norm": 0.13358810544013977, "learning_rate": 6.930693069306931e-07, "loss": 0.0282, "step": 4990 }, { "epoch": 19.727272727272727, "grad_norm": 0.06934131681919098, "learning_rate": 6.831683168316831e-07, "loss": 0.0144, "step": 4991 }, { "epoch": 19.731225296442688, "grad_norm": 0.09296256303787231, "learning_rate": 6.732673267326733e-07, "loss": 0.0198, "step": 4992 }, { "epoch": 19.73517786561265, "grad_norm": 0.05820873752236366, "learning_rate": 6.633663366336634e-07, "loss": 0.0119, "step": 4993 }, { "epoch": 19.73913043478261, "grad_norm": 0.12499816715717316, "learning_rate": 6.534653465346535e-07, "loss": 0.0262, "step": 4994 }, { "epoch": 19.74308300395257, "grad_norm": 0.0824083536863327, "learning_rate": 6.435643564356436e-07, "loss": 0.0177, "step": 4995 }, { "epoch": 19.74703557312253, "grad_norm": 0.10897348821163177, "learning_rate": 6.336633663366337e-07, "loss": 0.0234, "step": 4996 }, { "epoch": 19.75098814229249, "grad_norm": 0.11322897672653198, "learning_rate": 6.237623762376238e-07, "loss": 0.0241, "step": 4997 }, { "epoch": 19.75494071146245, "grad_norm": 0.07748400419950485, "learning_rate": 6.138613861386139e-07, "loss": 0.0178, "step": 4998 }, { "epoch": 19.75889328063241, "grad_norm": 0.07376900315284729, "learning_rate": 6.039603960396041e-07, "loss": 0.0147, "step": 4999 }, { "epoch": 19.76284584980237, "grad_norm": 0.12123782932758331, "learning_rate": 5.940594059405941e-07, "loss": 0.0249, "step": 5000 }, { "epoch": 19.766798418972332, "grad_norm": 0.13230913877487183, "learning_rate": 5.841584158415842e-07, "loss": 0.0278, "step": 5001 }, { "epoch": 19.770750988142293, "grad_norm": 0.1593136191368103, "learning_rate": 5.742574257425743e-07, "loss": 0.0348, "step": 5002 }, { "epoch": 19.774703557312254, "grad_norm": 0.08499763906002045, "learning_rate": 5.643564356435644e-07, "loss": 0.0182, "step": 5003 }, { "epoch": 19.778656126482215, "grad_norm": 0.15347464382648468, "learning_rate": 5.544554455445544e-07, "loss": 0.0326, "step": 5004 }, { "epoch": 19.782608695652176, "grad_norm": 0.12244894355535507, "learning_rate": 5.445544554455445e-07, "loss": 0.0279, "step": 5005 }, { "epoch": 19.786561264822133, "grad_norm": 0.12859375774860382, "learning_rate": 5.346534653465347e-07, "loss": 0.0258, "step": 5006 }, { "epoch": 19.790513833992094, "grad_norm": 0.15549464523792267, "learning_rate": 5.247524752475249e-07, "loss": 0.0348, "step": 5007 }, { "epoch": 19.794466403162055, "grad_norm": 0.07840237021446228, "learning_rate": 5.148514851485149e-07, "loss": 0.0156, "step": 5008 }, { "epoch": 19.798418972332016, "grad_norm": 0.06027500331401825, "learning_rate": 5.04950495049505e-07, "loss": 0.0131, "step": 5009 }, { "epoch": 19.802371541501977, "grad_norm": 0.09037033468484879, "learning_rate": 4.950495049504951e-07, "loss": 0.0196, "step": 5010 }, { "epoch": 19.806324110671937, "grad_norm": 0.06372921168804169, "learning_rate": 4.851485148514852e-07, "loss": 0.0134, "step": 5011 }, { "epoch": 19.8102766798419, "grad_norm": 0.07141555845737457, "learning_rate": 4.7524752475247526e-07, "loss": 0.014, "step": 5012 }, { "epoch": 19.81422924901186, "grad_norm": 0.14800450205802917, "learning_rate": 4.6534653465346537e-07, "loss": 0.0301, "step": 5013 }, { "epoch": 19.818181818181817, "grad_norm": 0.09713122993707657, "learning_rate": 4.5544554455445543e-07, "loss": 0.0147, "step": 5014 }, { "epoch": 19.822134387351777, "grad_norm": 0.09772741794586182, "learning_rate": 4.4554455445544555e-07, "loss": 0.0221, "step": 5015 }, { "epoch": 19.82608695652174, "grad_norm": 0.07063353806734085, "learning_rate": 4.356435643564357e-07, "loss": 0.0143, "step": 5016 }, { "epoch": 19.8300395256917, "grad_norm": 0.07360374927520752, "learning_rate": 4.2574257425742577e-07, "loss": 0.016, "step": 5017 }, { "epoch": 19.83399209486166, "grad_norm": 0.07476194947957993, "learning_rate": 4.158415841584159e-07, "loss": 0.0167, "step": 5018 }, { "epoch": 19.83794466403162, "grad_norm": 0.11618003994226456, "learning_rate": 4.05940594059406e-07, "loss": 0.0228, "step": 5019 }, { "epoch": 19.841897233201582, "grad_norm": 0.08956097811460495, "learning_rate": 3.9603960396039606e-07, "loss": 0.0188, "step": 5020 }, { "epoch": 19.845849802371543, "grad_norm": 0.08096391707658768, "learning_rate": 3.8613861386138617e-07, "loss": 0.0154, "step": 5021 }, { "epoch": 19.8498023715415, "grad_norm": 0.12467169761657715, "learning_rate": 3.7623762376237623e-07, "loss": 0.0286, "step": 5022 }, { "epoch": 19.85375494071146, "grad_norm": 0.07591530680656433, "learning_rate": 3.6633663366336635e-07, "loss": 0.0175, "step": 5023 }, { "epoch": 19.857707509881422, "grad_norm": 0.12216554582118988, "learning_rate": 3.5643564356435646e-07, "loss": 0.0264, "step": 5024 }, { "epoch": 19.861660079051383, "grad_norm": 0.07317045331001282, "learning_rate": 3.4653465346534657e-07, "loss": 0.0147, "step": 5025 }, { "epoch": 19.865612648221344, "grad_norm": 0.08463995903730392, "learning_rate": 3.3663366336633663e-07, "loss": 0.0188, "step": 5026 }, { "epoch": 19.869565217391305, "grad_norm": 0.06504493206739426, "learning_rate": 3.2673267326732674e-07, "loss": 0.013, "step": 5027 }, { "epoch": 19.873517786561266, "grad_norm": 0.07946094870567322, "learning_rate": 3.1683168316831686e-07, "loss": 0.0153, "step": 5028 }, { "epoch": 19.877470355731226, "grad_norm": 0.08311747759580612, "learning_rate": 3.0693069306930697e-07, "loss": 0.0178, "step": 5029 }, { "epoch": 19.881422924901187, "grad_norm": 0.07019700855016708, "learning_rate": 2.9702970297029703e-07, "loss": 0.0147, "step": 5030 }, { "epoch": 19.885375494071145, "grad_norm": 0.08537231385707855, "learning_rate": 2.8712871287128714e-07, "loss": 0.0192, "step": 5031 }, { "epoch": 19.889328063241106, "grad_norm": 0.07101425528526306, "learning_rate": 2.772277227722772e-07, "loss": 0.0155, "step": 5032 }, { "epoch": 19.893280632411066, "grad_norm": 0.07197488844394684, "learning_rate": 2.6732673267326737e-07, "loss": 0.0152, "step": 5033 }, { "epoch": 19.897233201581027, "grad_norm": 0.12315721809864044, "learning_rate": 2.5742574257425743e-07, "loss": 0.0272, "step": 5034 }, { "epoch": 19.90118577075099, "grad_norm": 0.08128491044044495, "learning_rate": 2.4752475247524754e-07, "loss": 0.017, "step": 5035 }, { "epoch": 19.90513833992095, "grad_norm": 0.0826103687286377, "learning_rate": 2.3762376237623763e-07, "loss": 0.0187, "step": 5036 }, { "epoch": 19.90909090909091, "grad_norm": 0.12392979860305786, "learning_rate": 2.2772277227722772e-07, "loss": 0.0226, "step": 5037 }, { "epoch": 19.91304347826087, "grad_norm": 0.08524937927722931, "learning_rate": 2.1782178217821786e-07, "loss": 0.0177, "step": 5038 }, { "epoch": 19.91699604743083, "grad_norm": 0.08728362619876862, "learning_rate": 2.0792079207920794e-07, "loss": 0.0174, "step": 5039 }, { "epoch": 19.92094861660079, "grad_norm": 0.06953892111778259, "learning_rate": 1.9801980198019803e-07, "loss": 0.0147, "step": 5040 }, { "epoch": 19.92490118577075, "grad_norm": 0.08887167274951935, "learning_rate": 1.8811881188118812e-07, "loss": 0.0187, "step": 5041 }, { "epoch": 19.92885375494071, "grad_norm": 0.31473830342292786, "learning_rate": 1.7821782178217823e-07, "loss": 0.0617, "step": 5042 }, { "epoch": 19.932806324110672, "grad_norm": 0.062241774052381516, "learning_rate": 1.6831683168316832e-07, "loss": 0.0135, "step": 5043 }, { "epoch": 19.936758893280633, "grad_norm": 0.1607532650232315, "learning_rate": 1.5841584158415843e-07, "loss": 0.0348, "step": 5044 }, { "epoch": 19.940711462450594, "grad_norm": 0.09338298439979553, "learning_rate": 1.4851485148514852e-07, "loss": 0.0205, "step": 5045 }, { "epoch": 19.944664031620555, "grad_norm": 0.09630955010652542, "learning_rate": 1.386138613861386e-07, "loss": 0.0203, "step": 5046 }, { "epoch": 19.948616600790515, "grad_norm": 0.08341194689273834, "learning_rate": 1.2871287128712872e-07, "loss": 0.0191, "step": 5047 }, { "epoch": 19.952569169960473, "grad_norm": 0.12010954320430756, "learning_rate": 1.1881188118811882e-07, "loss": 0.0259, "step": 5048 }, { "epoch": 19.956521739130434, "grad_norm": 0.07380735129117966, "learning_rate": 1.0891089108910893e-07, "loss": 0.0152, "step": 5049 }, { "epoch": 19.960474308300395, "grad_norm": 0.14853960275650024, "learning_rate": 9.900990099009901e-08, "loss": 0.0314, "step": 5050 }, { "epoch": 19.964426877470355, "grad_norm": 0.08187050372362137, "learning_rate": 8.910891089108911e-08, "loss": 0.0167, "step": 5051 }, { "epoch": 19.968379446640316, "grad_norm": 0.08192218095064163, "learning_rate": 7.920792079207921e-08, "loss": 0.017, "step": 5052 }, { "epoch": 19.972332015810277, "grad_norm": 0.08760989457368851, "learning_rate": 6.93069306930693e-08, "loss": 0.0196, "step": 5053 }, { "epoch": 19.976284584980238, "grad_norm": 0.12986694276332855, "learning_rate": 5.940594059405941e-08, "loss": 0.0272, "step": 5054 }, { "epoch": 19.9802371541502, "grad_norm": 0.09919023513793945, "learning_rate": 4.950495049504951e-08, "loss": 0.0219, "step": 5055 }, { "epoch": 19.984189723320156, "grad_norm": 0.07380606979131699, "learning_rate": 3.960396039603961e-08, "loss": 0.014, "step": 5056 }, { "epoch": 19.988142292490117, "grad_norm": 0.08057574927806854, "learning_rate": 2.9702970297029704e-08, "loss": 0.0171, "step": 5057 }, { "epoch": 19.992094861660078, "grad_norm": 0.13772280514240265, "learning_rate": 1.9801980198019804e-08, "loss": 0.03, "step": 5058 }, { "epoch": 19.99604743083004, "grad_norm": 0.12736505270004272, "learning_rate": 9.900990099009902e-09, "loss": 0.0278, "step": 5059 }, { "epoch": 20.0, "grad_norm": 0.12622465193271637, "learning_rate": 0.0, "loss": 0.0277, "step": 5060 } ], "logging_steps": 1, "max_steps": 5060, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.9621772453966295e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }